Development of the Neural Network Based Recognition Methods at V.L. Arlazarov’s Scientific School

A. V. Sheshkusa,b,*, A. N. Kondrashovaa,b,**, and D. P. Nikolaeva,c,***

a Smart Engines Service LLC, Moscow, 117312 Russian Federation

b Federal Research Center “Computer Science and Control” of the Russian Academy of Sciences, Moscow, 119333 Russian Federation

c Institute for Information Transmission Problems of the Russian Academy of Sciences, Moscow, 127051 Russian Federation

Correspondence to: * e-mail: asheshkus@smartengines.com
Correspondence to: ** e-mail: nastyachirvonaya@smartengines.com
Correspondence to: *** e-mail: dimonstr@iitp.ru

Received 19 October, 2022

Abstract—This work is devoted to methods for improving the quality of artificial neural networks, developed by a group of scientists led by V.L. Arlazarov. It describes various approaches both to training the networks themselves, from applying regularization to using new types of layers, and to preparing training data. This paper presents an effective method of data augmentation, as well as a neural network method for generating text data for training, which allows taking into account the properties of the language model in text recognition tasks. The applicability of metric networks to solving the classification problem and methods for generating and balancing training data for training such networks by creating Siamese and triplet schemes using contrastive loss and triplet loss functions are described. A separate section is devoted to research on the use of the fast Hough transform in neural networks as one of the methods for extracting nontrivial features. The paper also describes studies of a block convolutional layer, a method for constructing estimates of the presence of cuts between characters in the text segmentation problem, and a method for increasing the generalization ability of networks using orthogonalization.

Keywords: scientific school, neural network training, Hough transform, text recognition, metric networks, data generation

DOI: 10.1134/S1054661823040417