CONTENTS
- Vasiliev E., Bolotov D., Bolotov M., Smirnov L. Neural network approach to solving the problem of self-action of wave fields in nonlinear media
- Isupov K, Knyazkov V, Korzhavina A. Implementation and Convergence Evaluation of Multiple-Precision Iterative CG and PCG Solvers for GPUs
- Pirova A. Yu. Hybrid MPI + OpenMP algorithm for symmetric spare matrix reordering and its application to the solving systems of linear equations
- Kholkin S., Filimonov A. Finding graph chromatic number through deep learning
- Starostin N., Shtanyuk A.. Godovitsyn M., Zhivchikova J. Parallel algorithm for implementing logical operations on sets of orthogonal polygons
- Kaledina E.A., Kaledin O.E., Kulyagina T. I. Applying machine learning for prediction of cardiovascular diseases on small data sets
- Rodionov D.M., Karchkov D.A., Moskalenko V. A.. Nikolsky A. V., Osipov G. V, Zolotykh N.Yu. Diagnosis of sinus rhythm and atrial fibrillation using artificial intelligence
E. Vasiliev, D. Bolotov, M. Bolotov, L. Smirnov
Lobachevsky State University, 603022, Nizhny Novgorod, Russia
NEURAL NETWORK APPROACH TO SOLVING THE PROBLEM OF SELF-ACTION OF WAVE FIELDS IN NONLINEAR MEDIA
UDC: 517.957
DOI:10.24412/2073-0667-2022-1-5-16
We consider the problem of propagation of optical impulses in media with Kerr nonlinearity. As a mathematical model describing an optical pulse propagation process, we chose a generalized parabolic equation, which in dimensionless variables has the form of a one-dimensional modified Nonlinear Schrödinger Equation. We trained a fully connected neural network with various optimization functions and did experiments with network configuration and hyperparameters optimization. The conducted experiments have shown the promise of using the quasi-Newtonian L-BGFS optimization function over first-order optimization functions in this problem.
The article presents the results of experiments on training the model using various optimization functions: SDG, RMSProp, Adam, L-BGFS, the last optimizer allows you to train the network about an order faster. We also consider the size of the training sample required to train the model. From the obtained training results, we can conclude that due to the randomly uniform selection of points from the area using the Latin hypercube, it is enough to make a train sample of 10-15% of the dataset, this will correspond to a step of about 0.12 in z and 0.022 in τ compared to 0.039 in z and 0.008 in τ in a regular grid obtained by numerical methods.
In low-dimensional problems, the use of machine learning is not always appropriate, since training takes much more time than solving the problem using direct numerical simulation. However, as the complexity of the system increases, due to the increase in the number of unknown variables, a huge superiority of machine learning methods is expected due to fast calculation using an already trained network. Also open and interesting for future research is the issue of fast retraining of an already trained model for a problem with new parameters.
Key words: nonlinear Schrödinger equation, neural networks, deep learning, optimization functions