Участник:Strijov/Drafts
Материал из MachineLearning.
Строка 2: | Строка 2: | ||
{{TOCright}} | {{TOCright}} | ||
- | + | =2021= | |
==Bayesian model selection and multimodeling== | ==Bayesian model selection and multimodeling== | ||
Строка 32: | Строка 32: | ||
# 1.12 Bayesian PCA, GLM and NN | # 1.12 Bayesian PCA, GLM and NN | ||
# 8.12 Gaussian processes | # 8.12 Gaussian processes | ||
- | |||
- | |||
==References== | ==References== | ||
Строка 52: | Строка 50: | ||
# Bakhteev O.Y., Strijov V.V. Deep learning model selection of suboptimal complexity // Automation and Remote Control, 2018, 79(8) : 1474–1488, [https://link.springer.com/content/pdf/10.1134%2FS000511791808009X.pdf PDF]. | # Bakhteev O.Y., Strijov V.V. Deep learning model selection of suboptimal complexity // Automation and Remote Control, 2018, 79(8) : 1474–1488, [https://link.springer.com/content/pdf/10.1134%2FS000511791808009X.pdf PDF]. | ||
# Bakhteev O.Y., Strijov V.V. Comprehensive analysis of gradient-based hyperparameter optimization algorithmss // Annals of Operations Research, 2020 : 1-15, [http://strijov.com/papers/Bakhteev2017Hypergrad.pdf PDF]. | # Bakhteev O.Y., Strijov V.V. Comprehensive analysis of gradient-based hyperparameter optimization algorithmss // Annals of Operations Research, 2020 : 1-15, [http://strijov.com/papers/Bakhteev2017Hypergrad.pdf PDF]. | ||
+ | |||
+ | |||
+ | ==Mathematical methods of forecasting== | ||
+ | |||
+ | The lecture course and seminar introduces and applies methods of modern physics to the problems of machine learning. | ||
+ | |||
+ | Minimum topics to discuss: [https://arxiv.org/abs/2104.13478 Geometric deep learning] approach. | ||
+ | |||
+ | Optimum topics to discuss are: tensors, differential forms, Riemannian and differential geometry, metrics, differential operators in various spaces, embeddings, manifolds, bundles. We investigate scalar, vector and tensor fields (as well as jets, fibers and shiefs, tensor bundles, sheaf bundles etc.). The fields and spaces are one-dimensional, multidimensional and continuously dimensional. | ||
+ | |||
+ | ==Grading== | ||
+ | * Questionnaires during lectures (4) | ||
+ | * Two application projects (2+2) | ||
+ | * The final exam: problems with discussion (1) | ||
+ | * Bonus for active participation (2) | ||
+ | |||
+ | ===Deadlines=== | ||
+ | * October 14th, (October 7th one-week before preliminary show is strongly advised) | ||
+ | * December 2nd, (November 25th one-week before preliminary show is strongly advised) | ||
+ | |||
+ | ===Lab work report and talk=== | ||
+ | # Title and motivated abstract | ||
+ | # Problem statement | ||
+ | # Problem solution | ||
+ | # Link to the code | ||
+ | # Analysis and illustrative plots | ||
+ | # References | ||
+ | |||
+ | The report template [https://www.overleaf.com/read/gbtfqypksjsr is here]. Please follow the instructions in the template. | ||
+ | |||
+ | ==Themes== | ||
+ | ===BCI, Matrix and tensor approximation=== | ||
+ | # Коренев, Г.В. Тензорное исчисление, 2000, 240 с., [https://lib.mipt.ru/book/7235/Korenev-GV-Tenzornoe-ischislenie.djvu lib.mipt.ru]. | ||
+ | # Roger Penrose, "Applications of negative dimensional tensors," in Combinatorial Mathematics and its Applications, Academic Press (1971). See Vladimir Turaev, Quantum invariants of knots and 3-manifolds (1994), De Gruyter, p. 71 for a brief commentary [https://www.mscs.dal.ca/~selinger/papers/graphical-bib/public/Penrose-applications-of-negative-dimensional-tensors.pdf PDF]. | ||
+ | # Tai-Danae Bradley, At the Interface of Algebra and Statistics, 2020, [https://arxiv.org/abs/2004.05631 ArXiv]. | ||
+ | # Oseledets, I.V. Tensor-Train Decomposition //SIAM Journal on Scientific Computing, 2011, 33(5): 2295–2317, [https://doi.org/10.1137/090752286 DOI], [https://www.researchgate.net/publication/220412263_Tensor-Train_Decomposition RG], [https://www-labs.iro.umontreal.ca/~grabus/courses/ift6760_files/lecture-11.pdf lecture], [https://github.com/oseledets/TT-Toolbox GitHub], [https://www.ifi.uzh.ch/dam/jcr:846e4588-673e-4f55-b531-544a2e1f602e/TA_Tutorial_Part2.pdf Tutoiral]. | ||
+ | # Wikipedia: [https://en.wikipedia.org/wiki/Singular_value_decomposition#Matrix_approximation SVD], [https://en.wikipedia.org/wiki/Multilinear_subspace_learning Multilinear subspace learning], [https://en.wikipedia.org/wiki/Higher-order_singular_value_decomposition HOSVD]. | ||
+ | |||
+ | ===BCI, Feature selection=== | ||
+ | # Мотренко А.П. Выбор моделей прогнозирования мультикоррелирующих временных рядов (диссертация), 2019 [https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/PhDThesis/Motrenko/doc/Motrenko2018Thesis.pdf PDF] | ||
+ | # Исаченко Р.В. Снижение размерности пространства в задачах декодирования сигналов (дисссертация), 2021 [https://github.com/r-isachenko/PhDThesis/blob/master/doc/Isachenko2021PhDThesis.pdf PDF] | ||
+ | |||
+ | ===High order partial least squares=== | ||
+ | # Qibin Zhao, et al. and A. Cichocki, Higher Order Partial Least Squares (HOPLS): A Generalized Multilinear Regression Method // IEEE Transactions on Pattern Analysis and Machine Intelligence, July 2013, pp. 1660-1673, vol. 35, [https://10.1109/TPAMI.2012.254 DOI], [https://arxiv.org/abs/1207.1230 ArXiv]. | ||
+ | |||
+ | ===Neural ODEs and Continuous normalizing flows === | ||
+ | # Ricky T. Q. Chen et al., Neural Ordinary Differential Equations // NIPS, 2018, [https://arxiv.org/abs/1806.07366 ArXiv], [https://papers.nips.cc/paper/2018/hash/69386f6bb1dfed68692a24c8686939b9-Abstract.html source paper and code] | ||
+ | # Johann Brehmera and Kyle Cranmera, Flows for simultaneous manifold learning and density estimation // NIPS, 2020, [https://arxiv.org/pdf/2003.13913.pdf ArXiv] | ||
+ | #[https://deepgenerativemodels.github.io/notes/flow/ Flows at deepgenerativemodels.github.io] | ||
+ | #[https://lilianweng.github.io/lil-log/2018/10/13/flow-based-deep-generative-models.html Flow-based deep generative models] | ||
+ | #[https://arxiv.org/pdf/1505.05770.pdf Variational Inference with Normalizing Flows (source paper, Goes to BME)] | ||
+ | #[https://habr.com/ru/company/ods/blog/442002/ Знакомство с Neural ODE на хабре], [https://en.wikipedia.org/wiki/Flow-based_generative_model W: Flow-based generative model] | ||
+ | |||
+ | ===Continous time representation=== | ||
+ | # Самохина Алина, Непрерывное представление времени в задачах декодирования сигналов (магистерская диссертация): 2021 [http://www.machinelearning.ru/wiki/images/6/62/Samokhina2021MSThesis.pdf PDF], [https://github.com/Alina-Samokhina/MasterThesis GitHub] | ||
+ | # Aaron R Voelker, Ivana Kajić, Chris Eliasmith, Legendre Memory Units: Continuous-Time Representation in Recurrent Neural Networks // NIPS, 2019, [https://openreview.net/forum?id=HyxlRHBlUB PDF],[https://papers.nips.cc/paper/2019/file/952285b9b7e7a1be5aa7849f32ffff05-Paper.pdf PDF]. | ||
+ | # Functional data analysis: splines | ||
+ | |||
+ | |||
+ | ===Вопросы первого семестра=== | ||
+ | # Проблемы прогнозирования и постановки задач прогнозирования | ||
+ | # Авторегрессионные модели | ||
+ | # Гусеница, выбор компонент | ||
+ | # Модели прогнозирования временных рядов с высокой дисперсией ошибки | ||
+ | # Учет ошибки при прогнозировании | ||
+ | <!-- # Иерархическое прогнозирование --> | ||
+ | # Снижение размерности, SVD, PCA | ||
+ | # Метод Сугихары ССМ | ||
+ | # Метод PLD | ||
+ | # Тензорные разложения | ||
+ | # Метод HOPLS | ||
+ | # LSTM и авторегрессионное модели | ||
+ | # NeurODE | ||
+ | |||
+ | ===Navier-Stokes equations and viscous flow=== | ||
+ | # Neural PDE | ||
+ | # Newtonian and Non-Newtonian Fluids in Pipe Flows Using Neural Networks [https://doi.org/10.2202/1556-3758.1079], [https://doi.org/10.1103/PhysRevE.102.043309] | ||
+ | ===Metric tensors and kernels=== | ||
+ | # Lynn Houthuys and Johan A. K. Suykens, Tensor Learning in Multi-view Kernel PCA // ICANN 2018, pp 205-215, [https://10.1007/978-3-030-01421-6_21 DOI]. | ||
+ | |||
+ | ===fRMI, Riemannian geometry on shapes=== | ||
+ | # Xavier Pennec, Stefan Sommer, and Tom Fletcher, Riemannian Geometric Statistics in Medical Image Analysis, 2019 [https://www.elsevier.com/books/riemannian-geometric-statistics-in-medical-image-analysis/pennec/978-0-12-814725-2?format=print&utm_source=google_ads&utm_medium=paid_search&utm_campaign=frdsa-en&gclid=Cj0KCQjw-NaJBhDsARIsAAja6dPsj06W8cZ5mD15TS4MYPHqP4kbjExHfm55gxgQRzeDif2rmcWZEw4aApreEALw_wcB&gclsrc=aw.ds book] | ||
+ | # Surface differential geometry [https://www.coursera.org/lecture/image-processing/3-surface-differential-geometry-duration-11-43-vtuJ1 Coursera code video] for Image and Video Processing | ||
+ | |||
+ | ===Spatial time series alignment=== | ||
+ | # Titouan Vayer et al., Time Series Alignment with Global Invariances, 2020,[https://arxiv.org/abs/2002.03848 ArXiv] | ||
+ | # Marco Cuturi and Mathieu Blondel, Soft-DTW: a Differentiable Loss Function for Time-Series, [https://arxiv.org/pdf/1703.01541.pdf ArXiv] | ||
+ | # Marcel Campen et al., Scale-Invariant Directional Alignment of Surface Parametrizations // Eurographics Symposium on Geometry Processing, 2016, 35(5), [https://10.1111/cgf.12958 DOI] | ||
+ | # Helmut Pottmann et al. Geodesic Patterns // ACM Transactions on Graphics, 29(4), [https://10.1145/1833349.1778780 DOI], [https://www.researchgate.net/publication/220184673_Geodesic_Patterns PDF] | ||
+ | |||
+ | === Reproducing kernel Hilbert space === | ||
+ | # Mauricio A. Alvarez, Lorenzo Rosasco, Neil D. Lawrence, Kernels for Vector-Valued Functions: a Review, 2012, [https://arxiv.org/abs/1106.6251 ArXiv] | ||
+ | # Pedro Domingos, Every Model Learned by Gradient Descent Is Approximately a Kernel Machine, 2020, [https://arxiv.org/pdf/2012.00152.pdf ArXiv] | ||
+ | # Wikipedia: [https://en.wikipedia.org/wiki/Reproducing_kernel_Hilbert_space RKHS] | ||
+ | # Aronszajn, Nachman (1950). "Theory of Reproducing Kernels". Transactions of the American Mathematical Society. 68 (3): 337–404. [https://10.1090/S0002-9947-1950-0051437-7 DOI]. | ||
+ | |||
+ | ===Convolutions and Graphs=== | ||
+ | # Gama, F. et al. Graphs, Convolutions, and Neural Networks: From Graph Filters to Graph Neural Networks // IEEE Signal Processing Magazine, 2020, 37(6), 128-138, [https://doi.org/10.1109/MSP.2020.3016143 DOI]. | ||
+ | # Zhou, J. et al. Graph neural networks: A review of methods and applications // AI Open, 2020, 1: 57-81, [https://doi.org/10.1016j.aiopen.2021.01.001 DOI], [https://arxiv.org/pdf/1812.08434.pdf ArXiv]. | ||
+ | # Zonghan, W. et al. A Comprehensive Survey on Graph Neural Networks // IEEE Transactions on Neural Networks and Learning Systems, 2021, 32(1): 4-24, [https://10.1109/TNNLS.2020.2978386 DOI], [https://arxiv.org/pdf/1901.00596.pdf). ArXiv]. | ||
+ | # Zhang, S. et al. Graph convolutional networks: a comprehensive review // Computational Social Networks, 2019, 6(11), [https://doi.org/10.1186/s40649-019-0069-y DOI]. | ||
+ | # Xie, Y. et al. Self-Supervised Learning of Graph Neural Networks: A Unified Review // [https://arxiv.org/pdf/2102.10757.pdf ArXiv]. | ||
+ | # Wikipedia: [https://en.wikipedia.org/wiki/Laplacian_matrix Laplacian matrix], [https://en.wikipedia.org/wiki/Discrete_Poisson_equation Discrete] [https://en.wikipedia.org/wiki/Poisson%27s_equation Poisson's equation], [https://en.wikipedia.org/wiki/Graph_Fourier_transform Graph FT] | ||
+ | # [https://github.com/thunlp/GNNPapers GNN papers collection] | ||
+ | |||
+ | ===Higher order Fourier transform=== | ||
+ | # Zongyi Li et al., Fourier Neural Operator for Parametric Partial Differential Equations // ICLR, 2021, [https://arxiv.org/abs/2010.08895 ArXiv] | ||
+ | # Fourier for fun and practice 1D [https://morioh.com/p/18b3158eab36?f=5c21fb01c16e2556b555ab32&fbclid=IwAR0FBF6IfmEaaedMEDzxiSclxGgNVweHgDQZympUZ-4doeKGNwQwUvp9upo Fourier Code] | ||
+ | # Fourier for fun and practice nD | ||
+ | #* Fourier analysis on Manifolds 5G page 49 | ||
+ | #* Spectral analysis on meshes | ||
+ | |||
+ | ===Spherical Regression=== | ||
+ | # Shuai Liao, Efstratios Gavves, Cees G. M. Snoek, Spherical Regression: Learning Viewpoints, Surface Normals and 3D Rotations on N-Spheres // CVPR, 2019, 9759-9767, [https://openaccess.thecvf.com/content_CVPR_2019/html/Liao_Spherical_Regression_Learning_Viewpoints_Surface_Normals_and_3D_Rotations_on_CVPR_2019_paper.html ArXiv] | ||
+ | |||
+ | ===Category theory=== | ||
+ | # Tai-Danae Bradley, What is Applied Category Theory?, 2018, [https://arxiv.org/pdf/1809.05923.pdf ArXiv], [https://www.math3ma.com/ demo]. | ||
+ | # F. William Lawvere, Conceptual Mathematics: A First Introduction to Categories, 2011, [https://img.4plebs.org/boards/tg/image/1460/05/1460059215690.pdf PDF]. | ||
+ | # Картан А. Дифференциальное исчисление. Дифференциальные формы, 1971 [https://lib.mipt.ru/book/21021/Kartan-A-Differentsialnoe-ischislenie-Differentsialnye-formy.djvu lib.mipt.ru] | ||
+ | # Wikipedia: [https://en.wikipedia.org/wiki/Homology_(mathematics) Homology], Topological data analysis | ||
+ | === Geometric algebra === | ||
+ | # experior product and quaternions | ||
+ | # Nick Lucid, Advanced Theoretical Physics, 2019, [https://www.scienceasylum.com/AP_Sample.pdf sample]. | ||
+ | |||
+ | |||
+ | ==Lab works: Serie I== | ||
+ | Report consists of | ||
+ | # Problem motivation in details | ||
+ | # Formal statement and description | ||
+ | # Experiment description | ||
+ | # Plot and its analysis with explanations | ||
+ | |||
+ | The style: use a toolbox to illustrate a practical example with a Figure and its analysis. | ||
+ | |||
+ | One condition is the variety of toolboxes. | ||
+ | |||
+ | ===Lab work 1=== | ||
+ | Tensor approximation. Approximate the 2, 3 and 4 index matrices using low-rank decompositions, linear and nonlinear. The data sets are: a picture, a short animation movie (basic variant), a sound spectrogram, an fMRI. Plot the sequence of data approximations with ranks 1,..,n. Plot the error: x-axis is the rank, y-axis is the approximation error. Plot the variance of the error for various samples of data, if possible. | ||
+ | * Code [Tucker, PARAFAC-CANDECOMP, Tensor-train, and variants] | ||
+ | * Data sets [Cartoon, Sound, fMRI, and variants] | ||
+ | |||
+ | ===Lab work 2=== | ||
+ | PCA on higher orders. Construct a linear map of pairwise distance matrix to a space of lower dimensionality and plot the data. There are two possibilities: high-rank dimensional reduction and order reduction. If you could demonstrate the order reduction, it would be great. The pictures are appreciated. | ||
+ | * Code [Tucker, Tensor-train, and variants] | ||
+ | * Data sets [Cartoon, Sound, fMRI, and variants] | ||
+ | |||
+ | ===Lab work 3=== | ||
+ | Three-dimensional image reconstruction. Use sinograms of the computed tomography scans. Show the sequential approximation like in the lab work 1 with plots. | ||
+ | * Code [linear and NN] | ||
+ | * Data [Sinogram to fMRI] | ||
+ | |||
+ | ===Lab work 4=== | ||
+ | Formalize the set of low-rank approximation quality criterions (for multilinear operators with pseudo-inverses): precision, stability, computational complexity. Compare the decomposition methods from hottbox. | ||
+ | * Code [Hottbox] | ||
+ | * Data [Synthetic data with a simple visual/geometric visualization, like TF logo] | ||
+ | |||
+ | ===Lab work 5=== | ||
+ | Construct a phase trajectory of a spatial time series (a video, cartoon) and make a forecast with the higher order singular structure analysis. | ||
+ | * Code [SSA, HOSVD] | ||
+ | * Data [Cartoon with a walking hero] | ||
+ | |||
+ | ===Lab work 6=== | ||
+ | Construct two phase trajectories of spatial time series and discover a casualty with the higher order convergent cross mapping. | ||
+ | * Code [CCM, HOSVD] | ||
+ | * Data [Cartoon with a walking hero] | ||
+ | |||
+ | ===Lab work 7=== | ||
+ | Reduction an index on the tensor representation of the data, along with dimensionality reduction. The data is synthetic video, in more than 4D space, visualised by 3D projections. | ||
+ | * Code [HOSVD, HOPCA] | ||
+ | * Data [Synthetic 4D+ video] | ||
+ | |||
+ | ===Lab work 8=== | ||
+ | Higher order spectra analysis versus (higher order) Fourier transform on six-dimensional gyrokinetic data. Reduce number of indexes in tensor data. | ||
+ | * Code [HOSVD] | ||
+ | * Data [Gyrokinetic data 10.1016/j.jcp.2012.02.007] | ||
+ | |||
+ | ===Lab work 9=== | ||
+ | Combine higher-order PLS with CCN to predict eye-movement. | ||
+ | * Code [HOPLS] | ||
+ | * Data [Systhesic video and eye-movement] | ||
+ | |||
+ | ===Lab work 10=== | ||
+ | Higher order PLS for multiway data | ||
+ | * Code [HOPLS] | ||
+ | * Data [Accelerometer under various conditions to predict gyro data] | ||
+ | |||
+ | |||
+ | |||
+ | ==Lab works, Serie II== | ||
+ | As previously, the report consists of | ||
+ | # Problem motivation in details | ||
+ | # Formal statement and description | ||
+ | # Experiment description | ||
+ | # Plot and its analysis with explanations | ||
+ | The style: use a toolbox to illustrate a practical example with a Figure and its analysis. | ||
+ | |||
+ | ===The datasets for all lab works=== | ||
+ | The datasets for any lab work (if it is not specified) are 1) synthetic, are generated from the explicit solution of some differential equation, 2) from existing datasets, but not too long, 3) the self-made measurements are welcome!!! All of us have accelerometer, microphone, photo-video camera in the mobile phone. | ||
+ | |||
+ | ===References=== | ||
+ | * Neural Ordinary Differential Equations by Ricky T. Q. Chen, Yulia Rubanova, Jesse Bettencourt, David K. Duvenaud // NeurIPS 2018 [https://arxiv.org/pdf/1806.07366.pdf] | ||
+ | * Learning neural event functions for ordinary differential equations by Ricky T. Q. Chen, Brandon Amos, Maximilian Nickel // ICLR 2021, arXiv:2011.03902v4 | ||
+ | * Interpolation Technique to Speed Up Gradients Propagation in Neural ODEs by Talgat Daulbaev, Alexandr Katrutsa, Larisa Markeeva, Julia Gusak, Andrzej Cichocki, Ivan Oseledets // 2020, arXiv:2003.05271v2 | ||
+ | * Dynamical Systems with Applications using Python by Stephen Lynch, 2018 | ||
+ | * How To Train Your Neural ODE ChrisFinlay et al. [https://www.researchgate.net/profile/Levon-Nurbekyan/publication/339139726_How_to_train_your_neural_ODE/links/5e93a7ed299bf13079945d07/How-to-train-your-neural-ODE.pdf] or [http://proceedings.mlr.press/v119/finlay20a.html] | ||
+ | |||
+ | ===Code=== | ||
+ | * PyTorch Implementation of Differentiable ODE Solvers by Ricky Chen [https://github.com/rtqichen/torchdiffeq] | ||
+ | * A PyTorch library dedicated to neural differential equations and implicit models (the simplest implement, 100 lines). Maintained by DiffEqML [https://github.com/DiffEqML/torchdyn] | ||
+ | * Neural Ordinary Differential Equations by Mikhail Surtsukov [https://github.com/msurtsukov/neural-ode/blob/master/Neural%20ODEs%20(Russian).ipynb (Russian explanation)] | ||
+ | * Meta-Solver for Neural Ordinary Differential Equations by Julia Gusak [https://github.com/juliagusak/neural-ode-norm], [https://github.com/juliagusak/neural-ode-metasolver] | ||
+ | * Neural ODE Processes [https://github.com/crisbodnar/ndp code] [https://openreview.net/forum?id=27acGyyI1BY paper] | ||
+ | * PyTorch Implementation of Differentiable ODE Solvers [github.com/rtqichen/torchdiffeq] | ||
+ | * Jupyter notebook with Pytorch implementation of Neural Ordinary Differential Equations [github.com/msurtsukov/neural-ode] | ||
+ | * DiffEqFlux.jl – A Julia Library for Neural Differential Equations Julia [https://julialang.org/blog/2019/01/fluxdiffeq/#what_do_differential_equations_have_to_do_with_machine_learning] | ||
+ | |||
+ | ===Problem 1=== | ||
+ | Plot the [https://en.wikipedia.org/wiki/Slope_field direction field] and approximation by ODE-Net near the special points for simple ODEs like y’=y/x (uniform/узел), y’=-y/x saddle, y’=-x/y center, y’=(x+y)/(x-y) spiral. | ||
+ | |||
+ | ===Problem 2 === | ||
+ | Plot the stable and unstable [https://en.wikipedia.org/wiki/Autonomous_system_(mathematics) phase portraits] and approximations for simple ODE. | ||
+ | |||
+ | ===Problem 3=== | ||
+ | Find the ODE and plot the solution and ODE-Net approximation to compare and show difference between ODEsolvers: Euler, RK4 and others. | ||
+ | |||
+ | ===Problem 4=== | ||
+ | Approximate and plot the [http://www.machinelearning.ru/wiki/index.php?title=TS electricity consumption] time series. Compare the ODE-Net and LSTM or another model you like. | ||
+ | |||
+ | ===Problem 5=== | ||
+ | Plot and approximate by ODE-Net the phase portrait of a pendulum with decay. | ||
+ | |||
+ | ===Problem 5+=== | ||
+ | Plot and approximate the phase portrait of [https://en.wikipedia.org/wiki/Pendulum_(mechanics)#Coupled_pendula coupled pendula]. | ||
+ | |||
+ | ===Problem 6=== | ||
+ | Approximate an accelerometer (walking time series) by a solution of the pendulum ODE. | ||
+ | |||
+ | ===Problem 6+=== | ||
+ | Approximate an accelerometer by a solution of the double-pendulum ODE. | ||
+ | |||
+ | ===Problem 6++=== | ||
+ | Approximate an accelerometer and a gyroscope (acceleration and velocity) by a solution of the pendulum ODE. | ||
+ | |||
+ | ===Problem 7=== | ||
+ | Lorentz attractor, [https://en.wikipedia.org/wiki/Lorenz_system#Python_simulation three ODEs]. Compare the forecast LSTM (or any model you like) how it differs from the forecast ODE-Net under different initial conditions and under different noise conditions. An option is to immerse the trajectory in a space of a higher dimension by random rotation. | ||
+ | |||
+ | ===Problem 8=== | ||
+ | LSTM и ODE-RNN от [http://www.machinelearning.ru/wiki/images/6/62/Samokhina2021MSThesis.pdf от Алины Самохиной] раздел 2.2.2, только вместо EEG для упрощения временные ряды потребления электроэнергии. | ||
+ | |||
+ | ===Problem 9=== | ||
+ | Compare generated and forecaster electricity [http://www.machinelearning.ru/wiki/index.php?title=TS electricity consumption consumption time] series, see section 5 the arXiv:1806.07366v5 and the [ code] | ||
+ | |||
+ | ===Problem 10=== | ||
+ | Re-run the normalizing flows from [https://arxiv.org/abs/1806.07366 the paper] for various examples and analyse properties. | ||
+ | |||
+ | ==Lab works, Serie III== | ||
+ | Various scales, algebraic structures of the design space to construct rating (rank-regression models). | ||
+ | The results are published in [Link Colab] | ||
+ | |||
+ | ==Lab works, Serie IV== | ||
+ | The model generation and model forecasting. The goal is to hold an experiment to forecast a structure of an optimal model, given dataset. | ||
+ | |||
+ | ===List of modules (Problems)=== | ||
+ | # Class Model carries trivial methods | ||
+ | # Select optimal model from the generated set | ||
+ | # Check the model | ||
+ | # Propose and compute model complexity and distance between model structures | ||
+ | # Tune parameters of the model and keep them | ||
+ | # Get and put subtrees to the structure | ||
+ | # Generate the data and the model set | ||
+ | # Recover structure with Prims | ||
+ | # Recover structure with PCST | ||
+ | # Plot the trees | ||
+ | # Plot the distance maps | ||
+ | # Collect and analyse the errors | ||
+ | |||
+ | ===Each module is responsible for=== | ||
+ | # communication with neighbour models (input and output arguments) | ||
+ | # acting on the whole system. | ||
+ | # Each module runs the system! | ||
+ | |||
+ | ===The report consists of=== | ||
+ | # Module description | ||
+ | # Module test | ||
+ | # Results of the module (highly depends on the nature of the module ) | ||
+ | |||
+ | ===References=== | ||
+ | # Kulunchakov A.S., Strijov V.V. Generation of simple structured Information Retrieval functions by genetic algorithm without stagnation // Expert Systems with Applications, 2017, 85 : 221-230. [http://strijov.com/papers/Kulunchakov2014RankingBySimpleFun.pdf] | ||
+ | # Бочкарев А.М., Софронов И.Л., Стрижов В.В. Порождение экспертно-интерпретируемых моделей для прогноза проницаемости горной породы // Системы и средства информатики, 2017, 27(3) : 74-87.[http://strijov.com/papers/Bochkarev2017PermeabilityEstimation.pdf] | ||
+ | # Сологуб Р.А. [https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/PhDThesis/Sologub2013GenerationFinal/diss/my_input_2e.pdf?format=raw Методы трансформации моделей в задачах нелинейной регрессии], [https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/PhDThesis/Sologub2013GenerationFinal/diss/PHDpresentation.pdf?format=raw презентация PDF], [https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/PhDThesis/Sologub2013GenerationFinal/diss/PHDpresentationlog.pdf?format=raw полная версия PDF]. 2013. МФТИ. | ||
+ | # Kuznetsov M., Strijov V. SVM Composer, Matlab code, 2012. [https://sourceforge.net/p/mvr/code/HEAD/tree/mvr.stable/] | ||
+ | # Kuznetsov M., Strijov V. Notes on structure learning, 2012.[https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/StructureLearning/] | ||
+ | # [[Участник:Varf Ann|Варфоломеева Анна Андреевна]]. [[Media:Varfolomeeva2013Diploma.pdf|Методы структурного обучения для построения прогностических моделей]], [[Media:Varfolomeeva2013Presentation.pdf|презентация (PDF)]]. 2013.[[Media:Varfolomeeva2015MsThesis.pdf|Методы структурного обучения в задаче обобщения структур прогностических моделей]]'', [[Media:Varfolomeeva2015MsPresentation.pdf|презентация (PDF)]]. 2015. | ||
+ | # [[Участник:IShibaev|Шибаев Иннокентий Андреевич]] Бакалаврская диссертация: ''[[Media:Shibaev2018BThesis.pdf|Прогнозирование оптимальных суперпозиций в задачах регрессии]]'' ([[Media:Shibaev2018Presentation.pdf|Презентация]]). Магистерская диссертация: ''[[Media:Rudoy2014ModelsSelection.pdf|Алгоритмы индуктивного порождения и упрощения и критерии выбора оптимальной существенно нелинейной регрессионной модели для аппроксимации измеряемых данных]]'', [[Media:Rudoy2014ModelsSelectionSlides.pdf|презентация (PDF)]] | ||
+ | # [[Участник:Кулунчаков|Кулунчаков Андрей Сергеевич]] Бакалаврская диссертация: ''[[Media:Kulunchakov2015RankingBySimpleFun.pdf| Порождение структурно простых ранжирующих функций для задач информационного поиска]]'', [[Media:PresentationKulunchakov2015Ranking.pdf|презентация (PDF)]] | ||
+ | # Adam Gaier, David Ha. Weight Agnostic Neural Networks, ArXiv, 2019. [https://arxiv.org/abs/1906.04358] | ||
+ | # [https://en.wikipedia.org/wiki/Symbolic_regression Symbolic regression, wikipedia.] | ||
+ | # Wouter Minnebo and Sean Stijven. Empowering Knowledge Computing with Variable Selection [https://www.yumpu.com/en/document/view/20149728/empowering-knowledge-computing-with-variable-selection-on- not sure] | ||
+ | # Tyler William Hughes. Accelerating Symbolic Regression with Deep Learning, 2017. [http://twhughes.github.io/projects.html] | ||
+ | # EJ Vladislavleva, GF Smits, D Den Hertog. Order of nonlinearity as a complexity measure for models generated by symbolic regression via pareto genetic programming // IEEE Transactions on Evolutionary Computation 13 (2), 333-349 [https://www.researchgate.net/profile/Ekaterina-katya-Vladislavleva/publication/224330345_Order_of_Nonlinearity_as_a_Complexity_Measure_for_Models_Generated_by_Symbolic_Regression_via_Pareto_Genetic_Programming/links/00b7d5306967756b1d000000/Order-of-Nonlinearity-as-a-Complexity-Measure-for-Models-Generated-by-Symbolic-Regression-via-Pareto-Genetic-Programming.pdf] | ||
+ | # Tommi Jaakkola. Scaling structured prediction, 2012. [https://www.youtube.com/watch?v=LbsBguCUFEc] | ||
+ | # [[Media:MMP_2022_Symbolic_ModelGen.pdf|Слайды лекции "Прогнозирование прогностических моделей", часть "Символьная регрессия"]] | ||
+ | |||
+ | == Lab works Serie V== | ||
+ | |||
+ | The goal of the lab set is to join several, possible simple modules using joint efforts and information exchange. The following modules make the system work. | ||
+ | [[Media:PLSassemble.png|PLS scheme]] | ||
+ | |||
+ | '''Lab 0 ''' | ||
+ | * Make the documented demo of the system. List to run all possible models. Make an example of the most interesting models (Linear, Autoencoders, Neur-ODE, Tensor, Graph). | ||
+ | * In: All labs, Lab 4,5 | ||
+ | * Out: github.io page with examples | ||
+ | |||
+ | '''Lab 1 ''' | ||
+ | * Load data and put the data to PLS. Two sets of data: synthetic and real (X is time series times sensors, Y is time series times 3D coordinates) | ||
+ | * In: no | ||
+ | * Out: the formats for X, Y, U, V | ||
+ | |||
+ | '''Lab 2''' | ||
+ | * Make a set of the error functions. The reconstruction of (X, X), of (Y,Y), and of (Y, X) | ||
+ | * In: Lab 1 | ||
+ | * Out: Four functions Qall, QXX, QYY, QXY | ||
+ | |||
+ | '''Lab 3''' | ||
+ | * Check the admissibility of any given module in the system. Run the system in all variants | ||
+ | * In: All Labs | ||
+ | * Out: it works in the system, yes/no | ||
+ | |||
+ | '''Lab 4''' | ||
+ | * Make the simplest system, linear transformations | ||
+ | * In: Lab 1, Lab 2 | ||
+ | * Out: The shell with .predict, .fit stubs, which include interfaces to the models XX, YY, XY | ||
+ | |||
+ | '''Lab 5''' | ||
+ | * Select an optimisation algorithm from the set to optimise error functions | ||
+ | * In: Lab 2, Lab 3 | ||
+ | * Out: The optimal parameters for the system and predictions | ||
+ | |||
+ | '''Lab 6''' | ||
+ | * Plot the predictions and error analysis | ||
+ | * In: Lab 5 | ||
+ | * Out: Series of plots | ||
+ | |||
+ | '''Lab 7''' | ||
+ | Make YY Neur-ODE model | ||
+ | * In: Lab 5 | ||
+ | * Out: Series of plots | ||
+ | |||
+ | '''Lab 8''' | ||
+ | * Make XX, YY, stack-auto-encoder model | ||
+ | * In: Lab 5 | ||
+ | * Out: The tuned model and forecast | ||
+ | |||
+ | '''Lab 9''' | ||
+ | * Make XX, tensor decomposition model | ||
+ | * In: Lab 5 | ||
+ | * Out: The tuned model and forecast | ||
+ | |||
+ | '''Lab 10''' | ||
+ | * Make XX Graph convolution model | ||
+ | * In: Lab 5 | ||
+ | * Out: The tuned model and forecast | ||
+ | |||
+ | |||
+ | |||
+ | '''Lab 11''' | ||
+ | * Make XX Graph LSTM model | ||
+ | * In: Lab 5 | ||
+ | * Out: The tuned model and forecast | ||
+ | |||
+ | '''Lab 12''' | ||
+ | * Make XX Graph GRAND model | ||
+ | * In: Lab 5 | ||
+ | * Out: The tuned model and forecast | ||
+ | |||
+ | ===References=== | ||
+ | # [http://www.www.neurotycho.org/epidural-ecog-food-tracking-task neurotycho.org] | ||
+ | # [http://strijov.com/papers/Isachenko2020CanonicCorrelation.pdf Яушев Ф.Ю., Исаченко Р.В., Стрижов В.В. Модели согласования скрытого пространства в задаче прогнозирования // Системы и средства информатики, 2021, 31(1) : 4-16.] | ||
+ | # [http://strijov.com/papers/MotrenkoStrijov2017ECoG_HL_2.pdf Motrenko A.P., Strijov V.V. Multi-way feature selection for ECoG-based brain-computer interface // Expert Systems with Applications, 2018, 114(30) : 402-413.] | ||
+ | # [http://strijov.com/papers/Grabovoy2019QuasiPeriodicTimeSeries.pdf Grabovoy A.V., Strijov V.V. Quasi-periodic time series clustering for human activity recognition // Lobachevskii Journal of Mathematics, 2020, 41 : 333-339.] | ||
+ | # [[Участник:Isachenkoroma|Роман Владимирович Исаченко]] [http://www.frccsc.ru/sites/default/files/docs/ds/002-073-05/diss/31-isachenko/ds05_31-isachenko_main.pdf?18 Снижение размерности пространства в задачах декодирования сигналов] [https://github.com/r-isachenko/PhDThesis git], [https://github.com/r-isachenko/PhDThesis/raw/master/doc/Isachenko2021AutoAbstract.pdf автореферат], [https://github.com/r-isachenko/PhDThesis/raw/master/pres/Isachenko2021PhDThesisPres.pdf презентация (PDF)]. 2021. МФТИ | ||
+ | # [[Участник:Anastasiya|Мотренко Анастасия Петровна]]. [https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/PhDThesis/Motrenko/doc/Motrenko2018Thesis.pdf Выбор моделей прогнозирования мультикоррелирующих временных рядов], [https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/PhDThesis/Motrenko/doc/autoteferat/Motrenko2019Autoreferat.pdf автореферат], [https://sourceforge.net/p/mlalgorithms/code/HEAD/tree/PhDThesis/Motrenko/presentation/Motrenko2018PhDSlides.pdf презентация (PDF)] |
Версия 12:37, 6 февраля 2023
2021
Bayesian model selection and multimodeling
Course page: https://github.com/Intelligent-Systems-Phystech/BMM-21
The lecture course delivers the main problem of machine learning, the problem of model selection. One can set a heuristic model and optimise its parameters, or select a model from a class, or make a teacher model to transform its knowledge to a student model, or even make an ensemble from a models. Behind all these strategies there is a fundamental technique: the Bayesian inference. It assumes hypotheses about the measured data set, about the model parameters and even about the model structure. And it deduce the error function to optimise. This is called the Minimum Description Length principle. It selects simple, stable and precise models. This course joins the theory and the practical lab works of the model selection and multimodeling.
Grading
- Labs: 6 in total
- Forms: 1 in total
- Reports: 2 in total
The maximum score is 11, so the final score is MIN(10, score)
Syllabus
- 8.09 Intro
- 15.09 Distributions, expectation, likelihood
- 22.09 Bayesian inference
- 29.09 MDL, Minimum description length principle
- 6.10 Probabilistic metric spaces
- 13.10 Generative and discriminative models
- 20.10 Data generation, VAE, GAN
- 27.10 Probabilistic graphical models
- 3.11 Variational inference
- 10.11 Variational inference 2
- 17.11 Hyperparameter optimization
- 24.11 Meta-optimization
- 1.12 Bayesian PCA, GLM and NN
- 8.12 Gaussian processes
References
Books
- Bishop
- Barber
- Murphy
- Rasmussen and Williams, of course!
- Taboga(to catch up)
Theses
- Грабововй А.В. Диссертация.
- Бахтеев О.Ю.. Выбор моделей глубокого обучения субоптимальной сложности git, автореферат, презентация (PDF), видео. 2020. МФТИ.
- Адуенко А.А. Выбор мультимоделей в задачах классификации, презентация (PDF), видео. 2017. МФТИ.
- Кузьмин А.А. | Построение иерархических тематических моделей коллекций коротких текстов, | презентация (PDF), видео. 2017. МФТИ.
Papers
- Kuznetsov M.P., Tokmakova A.A., Strijov V.V. Analytic and stochastic methods of structure parameter estimation // Informatica, 2016, 27(3) : 607-624, PDF.
- Bakhteev O.Y., Strijov V.V. Deep learning model selection of suboptimal complexity // Automation and Remote Control, 2018, 79(8) : 1474–1488, PDF.
- Bakhteev O.Y., Strijov V.V. Comprehensive analysis of gradient-based hyperparameter optimization algorithmss // Annals of Operations Research, 2020 : 1-15, PDF.
Mathematical methods of forecasting
The lecture course and seminar introduces and applies methods of modern physics to the problems of machine learning.
Minimum topics to discuss: Geometric deep learning approach.
Optimum topics to discuss are: tensors, differential forms, Riemannian and differential geometry, metrics, differential operators in various spaces, embeddings, manifolds, bundles. We investigate scalar, vector and tensor fields (as well as jets, fibers and shiefs, tensor bundles, sheaf bundles etc.). The fields and spaces are one-dimensional, multidimensional and continuously dimensional.
Grading
- Questionnaires during lectures (4)
- Two application projects (2+2)
- The final exam: problems with discussion (1)
- Bonus for active participation (2)
Deadlines
- October 14th, (October 7th one-week before preliminary show is strongly advised)
- December 2nd, (November 25th one-week before preliminary show is strongly advised)
Lab work report and talk
- Title and motivated abstract
- Problem statement
- Problem solution
- Link to the code
- Analysis and illustrative plots
- References
The report template is here. Please follow the instructions in the template.
Themes
BCI, Matrix and tensor approximation
- Коренев, Г.В. Тензорное исчисление, 2000, 240 с., lib.mipt.ru.
- Roger Penrose, "Applications of negative dimensional tensors," in Combinatorial Mathematics and its Applications, Academic Press (1971). See Vladimir Turaev, Quantum invariants of knots and 3-manifolds (1994), De Gruyter, p. 71 for a brief commentary PDF.
- Tai-Danae Bradley, At the Interface of Algebra and Statistics, 2020, ArXiv.
- Oseledets, I.V. Tensor-Train Decomposition //SIAM Journal on Scientific Computing, 2011, 33(5): 2295–2317, DOI, RG, lecture, GitHub, Tutoiral.
- Wikipedia: SVD, Multilinear subspace learning, HOSVD.
BCI, Feature selection
- Мотренко А.П. Выбор моделей прогнозирования мультикоррелирующих временных рядов (диссертация), 2019 PDF
- Исаченко Р.В. Снижение размерности пространства в задачах декодирования сигналов (дисссертация), 2021 PDF
High order partial least squares
- Qibin Zhao, et al. and A. Cichocki, Higher Order Partial Least Squares (HOPLS): A Generalized Multilinear Regression Method // IEEE Transactions on Pattern Analysis and Machine Intelligence, July 2013, pp. 1660-1673, vol. 35, DOI, ArXiv.
Neural ODEs and Continuous normalizing flows
- Ricky T. Q. Chen et al., Neural Ordinary Differential Equations // NIPS, 2018, ArXiv, source paper and code
- Johann Brehmera and Kyle Cranmera, Flows for simultaneous manifold learning and density estimation // NIPS, 2020, ArXiv
- Flows at deepgenerativemodels.github.io
- Flow-based deep generative models
- Variational Inference with Normalizing Flows (source paper, Goes to BME)
- Знакомство с Neural ODE на хабре, W: Flow-based generative model
Continous time representation
- Самохина Алина, Непрерывное представление времени в задачах декодирования сигналов (магистерская диссертация): 2021 PDF, GitHub
- Aaron R Voelker, Ivana Kajić, Chris Eliasmith, Legendre Memory Units: Continuous-Time Representation in Recurrent Neural Networks // NIPS, 2019, PDF,PDF.
- Functional data analysis: splines
Вопросы первого семестра
- Проблемы прогнозирования и постановки задач прогнозирования
- Авторегрессионные модели
- Гусеница, выбор компонент
- Модели прогнозирования временных рядов с высокой дисперсией ошибки
- Учет ошибки при прогнозировании
- Снижение размерности, SVD, PCA
- Метод Сугихары ССМ
- Метод PLD
- Тензорные разложения
- Метод HOPLS
- LSTM и авторегрессионное модели
- NeurODE
Navier-Stokes equations and viscous flow
Metric tensors and kernels
- Lynn Houthuys and Johan A. K. Suykens, Tensor Learning in Multi-view Kernel PCA // ICANN 2018, pp 205-215, DOI.
fRMI, Riemannian geometry on shapes
- Xavier Pennec, Stefan Sommer, and Tom Fletcher, Riemannian Geometric Statistics in Medical Image Analysis, 2019 book
- Surface differential geometry Coursera code video for Image and Video Processing
Spatial time series alignment
- Titouan Vayer et al., Time Series Alignment with Global Invariances, 2020,ArXiv
- Marco Cuturi and Mathieu Blondel, Soft-DTW: a Differentiable Loss Function for Time-Series, ArXiv
- Marcel Campen et al., Scale-Invariant Directional Alignment of Surface Parametrizations // Eurographics Symposium on Geometry Processing, 2016, 35(5), DOI
- Helmut Pottmann et al. Geodesic Patterns // ACM Transactions on Graphics, 29(4), DOI, PDF
Reproducing kernel Hilbert space
- Mauricio A. Alvarez, Lorenzo Rosasco, Neil D. Lawrence, Kernels for Vector-Valued Functions: a Review, 2012, ArXiv
- Pedro Domingos, Every Model Learned by Gradient Descent Is Approximately a Kernel Machine, 2020, ArXiv
- Wikipedia: RKHS
- Aronszajn, Nachman (1950). "Theory of Reproducing Kernels". Transactions of the American Mathematical Society. 68 (3): 337–404. DOI.
Convolutions and Graphs
- Gama, F. et al. Graphs, Convolutions, and Neural Networks: From Graph Filters to Graph Neural Networks // IEEE Signal Processing Magazine, 2020, 37(6), 128-138, DOI.
- Zhou, J. et al. Graph neural networks: A review of methods and applications // AI Open, 2020, 1: 57-81, DOI, ArXiv.
- Zonghan, W. et al. A Comprehensive Survey on Graph Neural Networks // IEEE Transactions on Neural Networks and Learning Systems, 2021, 32(1): 4-24, DOI, ArXiv.
- Zhang, S. et al. Graph convolutional networks: a comprehensive review // Computational Social Networks, 2019, 6(11), DOI.
- Xie, Y. et al. Self-Supervised Learning of Graph Neural Networks: A Unified Review // ArXiv.
- Wikipedia: Laplacian matrix, Discrete Poisson's equation, Graph FT
- GNN papers collection
Higher order Fourier transform
- Zongyi Li et al., Fourier Neural Operator for Parametric Partial Differential Equations // ICLR, 2021, ArXiv
- Fourier for fun and practice 1D Fourier Code
- Fourier for fun and practice nD
- Fourier analysis on Manifolds 5G page 49
- Spectral analysis on meshes
Spherical Regression
- Shuai Liao, Efstratios Gavves, Cees G. M. Snoek, Spherical Regression: Learning Viewpoints, Surface Normals and 3D Rotations on N-Spheres // CVPR, 2019, 9759-9767, ArXiv
Category theory
- Tai-Danae Bradley, What is Applied Category Theory?, 2018, ArXiv, demo.
- F. William Lawvere, Conceptual Mathematics: A First Introduction to Categories, 2011, PDF.
- Картан А. Дифференциальное исчисление. Дифференциальные формы, 1971 lib.mipt.ru
- Wikipedia: Homology, Topological data analysis
Geometric algebra
- experior product and quaternions
- Nick Lucid, Advanced Theoretical Physics, 2019, sample.
Lab works: Serie I
Report consists of
- Problem motivation in details
- Formal statement and description
- Experiment description
- Plot and its analysis with explanations
The style: use a toolbox to illustrate a practical example with a Figure and its analysis.
One condition is the variety of toolboxes.
Lab work 1
Tensor approximation. Approximate the 2, 3 and 4 index matrices using low-rank decompositions, linear and nonlinear. The data sets are: a picture, a short animation movie (basic variant), a sound spectrogram, an fMRI. Plot the sequence of data approximations with ranks 1,..,n. Plot the error: x-axis is the rank, y-axis is the approximation error. Plot the variance of the error for various samples of data, if possible.
- Code [Tucker, PARAFAC-CANDECOMP, Tensor-train, and variants]
- Data sets [Cartoon, Sound, fMRI, and variants]
Lab work 2
PCA on higher orders. Construct a linear map of pairwise distance matrix to a space of lower dimensionality and plot the data. There are two possibilities: high-rank dimensional reduction and order reduction. If you could demonstrate the order reduction, it would be great. The pictures are appreciated.
- Code [Tucker, Tensor-train, and variants]
- Data sets [Cartoon, Sound, fMRI, and variants]
Lab work 3
Three-dimensional image reconstruction. Use sinograms of the computed tomography scans. Show the sequential approximation like in the lab work 1 with plots.
- Code [linear and NN]
- Data [Sinogram to fMRI]
Lab work 4
Formalize the set of low-rank approximation quality criterions (for multilinear operators with pseudo-inverses): precision, stability, computational complexity. Compare the decomposition methods from hottbox.
- Code [Hottbox]
- Data [Synthetic data with a simple visual/geometric visualization, like TF logo]
Lab work 5
Construct a phase trajectory of a spatial time series (a video, cartoon) and make a forecast with the higher order singular structure analysis.
- Code [SSA, HOSVD]
- Data [Cartoon with a walking hero]
Lab work 6
Construct two phase trajectories of spatial time series and discover a casualty with the higher order convergent cross mapping.
- Code [CCM, HOSVD]
- Data [Cartoon with a walking hero]
Lab work 7
Reduction an index on the tensor representation of the data, along with dimensionality reduction. The data is synthetic video, in more than 4D space, visualised by 3D projections.
- Code [HOSVD, HOPCA]
- Data [Synthetic 4D+ video]
Lab work 8
Higher order spectra analysis versus (higher order) Fourier transform on six-dimensional gyrokinetic data. Reduce number of indexes in tensor data.
- Code [HOSVD]
- Data [Gyrokinetic data 10.1016/j.jcp.2012.02.007]
Lab work 9
Combine higher-order PLS with CCN to predict eye-movement.
- Code [HOPLS]
- Data [Systhesic video and eye-movement]
Lab work 10
Higher order PLS for multiway data
- Code [HOPLS]
- Data [Accelerometer under various conditions to predict gyro data]
Lab works, Serie II
As previously, the report consists of
- Problem motivation in details
- Formal statement and description
- Experiment description
- Plot and its analysis with explanations
The style: use a toolbox to illustrate a practical example with a Figure and its analysis.
The datasets for all lab works
The datasets for any lab work (if it is not specified) are 1) synthetic, are generated from the explicit solution of some differential equation, 2) from existing datasets, but not too long, 3) the self-made measurements are welcome!!! All of us have accelerometer, microphone, photo-video camera in the mobile phone.
References
- Neural Ordinary Differential Equations by Ricky T. Q. Chen, Yulia Rubanova, Jesse Bettencourt, David K. Duvenaud // NeurIPS 2018 [3]
- Learning neural event functions for ordinary differential equations by Ricky T. Q. Chen, Brandon Amos, Maximilian Nickel // ICLR 2021, arXiv:2011.03902v4
- Interpolation Technique to Speed Up Gradients Propagation in Neural ODEs by Talgat Daulbaev, Alexandr Katrutsa, Larisa Markeeva, Julia Gusak, Andrzej Cichocki, Ivan Oseledets // 2020, arXiv:2003.05271v2
- Dynamical Systems with Applications using Python by Stephen Lynch, 2018
- How To Train Your Neural ODE ChrisFinlay et al. [4] or [5]
Code
- PyTorch Implementation of Differentiable ODE Solvers by Ricky Chen [6]
- A PyTorch library dedicated to neural differential equations and implicit models (the simplest implement, 100 lines). Maintained by DiffEqML [7]
- Neural Ordinary Differential Equations by Mikhail Surtsukov (Russian explanation)
- Meta-Solver for Neural Ordinary Differential Equations by Julia Gusak [8], [9]
- Neural ODE Processes code paper
- PyTorch Implementation of Differentiable ODE Solvers [github.com/rtqichen/torchdiffeq]
- Jupyter notebook with Pytorch implementation of Neural Ordinary Differential Equations [github.com/msurtsukov/neural-ode]
- DiffEqFlux.jl – A Julia Library for Neural Differential Equations Julia [10]
Problem 1
Plot the direction field and approximation by ODE-Net near the special points for simple ODEs like y’=y/x (uniform/узел), y’=-y/x saddle, y’=-x/y center, y’=(x+y)/(x-y) spiral.
Problem 2
Plot the stable and unstable phase portraits and approximations for simple ODE.
Problem 3
Find the ODE and plot the solution and ODE-Net approximation to compare and show difference between ODEsolvers: Euler, RK4 and others.
Problem 4
Approximate and plot the electricity consumption time series. Compare the ODE-Net and LSTM or another model you like.
Problem 5
Plot and approximate by ODE-Net the phase portrait of a pendulum with decay.
Problem 5+
Plot and approximate the phase portrait of coupled pendula.
Problem 6
Approximate an accelerometer (walking time series) by a solution of the pendulum ODE.
Problem 6+
Approximate an accelerometer by a solution of the double-pendulum ODE.
Problem 6++
Approximate an accelerometer and a gyroscope (acceleration and velocity) by a solution of the pendulum ODE.
Problem 7
Lorentz attractor, three ODEs. Compare the forecast LSTM (or any model you like) how it differs from the forecast ODE-Net under different initial conditions and under different noise conditions. An option is to immerse the trajectory in a space of a higher dimension by random rotation.
Problem 8
LSTM и ODE-RNN от от Алины Самохиной раздел 2.2.2, только вместо EEG для упрощения временные ряды потребления электроэнергии.
Problem 9
Compare generated and forecaster electricity electricity consumption consumption time series, see section 5 the arXiv:1806.07366v5 and the [ code]
Problem 10
Re-run the normalizing flows from the paper for various examples and analyse properties.
Lab works, Serie III
Various scales, algebraic structures of the design space to construct rating (rank-regression models). The results are published in [Link Colab]
Lab works, Serie IV
The model generation and model forecasting. The goal is to hold an experiment to forecast a structure of an optimal model, given dataset.
List of modules (Problems)
- Class Model carries trivial methods
- Select optimal model from the generated set
- Check the model
- Propose and compute model complexity and distance between model structures
- Tune parameters of the model and keep them
- Get and put subtrees to the structure
- Generate the data and the model set
- Recover structure with Prims
- Recover structure with PCST
- Plot the trees
- Plot the distance maps
- Collect and analyse the errors
Each module is responsible for
- communication with neighbour models (input and output arguments)
- acting on the whole system.
- Each module runs the system!
The report consists of
- Module description
- Module test
- Results of the module (highly depends on the nature of the module )
References
- Kulunchakov A.S., Strijov V.V. Generation of simple structured Information Retrieval functions by genetic algorithm without stagnation // Expert Systems with Applications, 2017, 85 : 221-230. [11]
- Бочкарев А.М., Софронов И.Л., Стрижов В.В. Порождение экспертно-интерпретируемых моделей для прогноза проницаемости горной породы // Системы и средства информатики, 2017, 27(3) : 74-87.[12]
- Сологуб Р.А. Методы трансформации моделей в задачах нелинейной регрессии, презентация PDF, полная версия PDF. 2013. МФТИ.
- Kuznetsov M., Strijov V. SVM Composer, Matlab code, 2012. [13]
- Kuznetsov M., Strijov V. Notes on structure learning, 2012.[14]
- Варфоломеева Анна Андреевна. Методы структурного обучения для построения прогностических моделей, презентация (PDF). 2013.Методы структурного обучения в задаче обобщения структур прогностических моделей, презентация (PDF). 2015.
- Шибаев Иннокентий Андреевич Бакалаврская диссертация: Прогнозирование оптимальных суперпозиций в задачах регрессии (Презентация). Магистерская диссертация: Алгоритмы индуктивного порождения и упрощения и критерии выбора оптимальной существенно нелинейной регрессионной модели для аппроксимации измеряемых данных, презентация (PDF)
- Кулунчаков Андрей Сергеевич Бакалаврская диссертация: Порождение структурно простых ранжирующих функций для задач информационного поиска, презентация (PDF)
- Adam Gaier, David Ha. Weight Agnostic Neural Networks, ArXiv, 2019. [15]
- Symbolic regression, wikipedia.
- Wouter Minnebo and Sean Stijven. Empowering Knowledge Computing with Variable Selection not sure
- Tyler William Hughes. Accelerating Symbolic Regression with Deep Learning, 2017. [16]
- EJ Vladislavleva, GF Smits, D Den Hertog. Order of nonlinearity as a complexity measure for models generated by symbolic regression via pareto genetic programming // IEEE Transactions on Evolutionary Computation 13 (2), 333-349 [17]
- Tommi Jaakkola. Scaling structured prediction, 2012. [18]
- Слайды лекции "Прогнозирование прогностических моделей", часть "Символьная регрессия"
Lab works Serie V
The goal of the lab set is to join several, possible simple modules using joint efforts and information exchange. The following modules make the system work. PLS scheme
Lab 0
- Make the documented demo of the system. List to run all possible models. Make an example of the most interesting models (Linear, Autoencoders, Neur-ODE, Tensor, Graph).
- In: All labs, Lab 4,5
- Out: github.io page with examples
Lab 1
- Load data and put the data to PLS. Two sets of data: synthetic and real (X is time series times sensors, Y is time series times 3D coordinates)
- In: no
- Out: the formats for X, Y, U, V
Lab 2
- Make a set of the error functions. The reconstruction of (X, X), of (Y,Y), and of (Y, X)
- In: Lab 1
- Out: Four functions Qall, QXX, QYY, QXY
Lab 3
- Check the admissibility of any given module in the system. Run the system in all variants
- In: All Labs
- Out: it works in the system, yes/no
Lab 4
- Make the simplest system, linear transformations
- In: Lab 1, Lab 2
- Out: The shell with .predict, .fit stubs, which include interfaces to the models XX, YY, XY
Lab 5
- Select an optimisation algorithm from the set to optimise error functions
- In: Lab 2, Lab 3
- Out: The optimal parameters for the system and predictions
Lab 6
- Plot the predictions and error analysis
- In: Lab 5
- Out: Series of plots
Lab 7 Make YY Neur-ODE model
- In: Lab 5
- Out: Series of plots
Lab 8
- Make XX, YY, stack-auto-encoder model
- In: Lab 5
- Out: The tuned model and forecast
Lab 9
- Make XX, tensor decomposition model
- In: Lab 5
- Out: The tuned model and forecast
Lab 10
- Make XX Graph convolution model
- In: Lab 5
- Out: The tuned model and forecast
Lab 11
- Make XX Graph LSTM model
- In: Lab 5
- Out: The tuned model and forecast
Lab 12
- Make XX Graph GRAND model
- In: Lab 5
- Out: The tuned model and forecast
References
- neurotycho.org
- Яушев Ф.Ю., Исаченко Р.В., Стрижов В.В. Модели согласования скрытого пространства в задаче прогнозирования // Системы и средства информатики, 2021, 31(1) : 4-16.
- Motrenko A.P., Strijov V.V. Multi-way feature selection for ECoG-based brain-computer interface // Expert Systems with Applications, 2018, 114(30) : 402-413.
- Grabovoy A.V., Strijov V.V. Quasi-periodic time series clustering for human activity recognition // Lobachevskii Journal of Mathematics, 2020, 41 : 333-339.
- Роман Владимирович Исаченко Снижение размерности пространства в задачах декодирования сигналов git, автореферат, презентация (PDF). 2021. МФТИ
- Мотренко Анастасия Петровна. Выбор моделей прогнозирования мультикоррелирующих временных рядов, автореферат, презентация (PDF)