Modern Methods for Dynamics and Trajectory Tracking Modeling of High-Degree-of-Freedom Systems
DOI:
https://doi.org/10.17072/1993-0550-2026-1-100-117Keywords:
dynamic analysis, multidimensional systems, high-degree-of-freedom systems, computer vision, deep learning, neural networks, markerless object tracking, state classification, transition probability matrices, Markov modelsAbstract
The study is devoted to analyzing a key methodological shift in the quantitative assessment of bio-object behavior: the transition from classifying individual behavioral acts to analyzing their dynamic structure. The problem statement lies in the existence of a methodological gap: traditional tracking methods simplify behavior to the center-of-mass trajectory, while modern deep learning algorithms, providing high pose recognition accuracy, often ignore the temporal context and function as computationally expensive "black boxes". The aim of the work is to analyze existing approaches to video processing and substantiate a hybrid methodology that combines the detail of neural network analysis with a systemic approach to dynamics. The objectives include reviewing methods ranging from manual annotation to transformer architectures and identifying their limitations in long-term forecasting tasks. The research methods are based on a comparative analysis of computer vision algorithms, deep learning, and stochastic modeling in the context of processing video data from laboratory animal experiments. The main results show that neural networks are optimized primarily for local classification. The paper proposes a methodological framework integrating frame-level classification with the construction of interpretable stochastic models (transition probability matrices). The main conclusions indicate that using Markov representations in discrete state spaces allows for the effective identification of stable behavioral modes (attractors) and anomalies, creating a reliable basis for decision support systems without excessive computational requirements.References
Schiehlen W. Multibody system dynamics: roots and perspectives // Multibody system dynamics. 1997. V. 1, no 2. P. 149−188.
Strogatz S. H. Nonlinear dynamics and chaos: with applications to physics, biology, chemistry, and engineering. Chapman and Hall/CRC, 2024.
Rabiner L. R. A tutorial on hidden Markov models and selected applications in speech recognition // Proceedings of the IEEE. 2002. V. 77, no 2. P. 257−286.
Berman G. J. Measuring behavior across scales // BMC biology. 2018. V. 16, no 1. P. 23.
Noldus L. P. J. J., Spink A. J., Tegelenbosch R. A. J. EthoVision: a versatile video tracking system for automation of behavioral experiments // Behavior Research Methods, Instruments, & Computers. 2001. V. 33, no 3. P. 398−414.
Mathis A. et al. DeepLabCut: markerless pose estimation of user-defined body parts with deep learning // Nature neuroscience. 2018. V. 21, no 9. P. 1281−1289.
Nath T. et al. Using DeepLabCut for 3D markerless pose estimation across species and behaviors / /Nature protocols. 2019. V. 14, no 7. P. 2152−2176.
Dawkins M. S., Martin P., Bateson P. Measuring Behaviour. An Introductory Guide // Journal of Animal Ecology. 1994. V. 63, no 3. P. 746.
Ушакова А. А. и др. Некоторые поведенческие особенности сенильных животных 5XFAD, выявляемые при длительных наблюдениях в домашних клетках // Вестник Пермской государственной фармацевтической академии "Кромеровские чтения 2023". 2023. С. 95−98.
Tuyttens F. A. M. et al. Observer bias in animal behaviour research: can we believe what we score, if we score what we believe? // Animal Behaviour. 2014. V. 90. P. 273−280.
Fort K., Nazarenko A., Rosset S. Modeling the complexity of manual annotation tasks: a grid of analysis // International conference on computational linguistics. 2012. P. 895−910.
Ahmadzadeh A. et al. A Guide for Manual Annotation of Scientific Imagery: How to Prepare for Large Projects // arXiv preprint arXiv:2508.14801. 2025.
Thakur A., Arya S. Comparative analysis of manual and programmed annotations for crowd assessment and classification using artificial intelligence // Data Science and Management. 2024. V. 7, no 4. P. 340−348.
Dell A. I. et al. Automated image-based tracking and its application in ecology // Trends in ecology & evolution. 2014. V. 29, no 7. P. 417−428.
Seibenhener M. L., Wooten M. C. Use of the open field maze to measure locomotor and anxiety-like behavior in mice // Journal of visualized experiments: JoVE. 2015. No 96. P. 52434.
Lim C. J. M. et al. Comparison of automated video tracking systems in the open field test: ANY-Maze versus EthoVision XT // Journal of Neuroscience Methods. 2023. V. 397. P. 109940.
Escalona J. L., Urda P., Muñoz S. A track geometry measuring system based on multi-body kinematics, inertial sensors and computer vision // Sensors. 2021. V. 21, no 3. P. 683.
Piccardi M. Background subtraction techniques: a review // 2004 IEEE international conference on systems, man and cybernetics (IEEE Cat. No. 04CH37583). IEEE, 2004. V. 4. P. 3099−3104.
Bouwmans T. Traditional and recent approaches in background modeling for foreground detection: An overview // Computer science review. 2014. V. 11. P. 31−66.
Bobick A. F., Davis J. W. The recognition of human movement using temporal templates // IEEE Transactions on pattern analysis and machine intelligence. 2002. V. 23, no 3. P. 257−267.
Haro G. Shape from silhouette consensus and photo-consistency // 2014 IEEE International Conference on Image Processing (ICIP). IEEE, 2014. P. 4837−4841.
Barron J. L., Fleet D. J., Beauchemin S. S. Performance of optical flow techniques // International journal of computer vision. 1994. V. 12, no 1. P. 43−77.
Baker S. et al. A database and evaluation methodology for optical flow // International journal of computer vision. 2011. V. 92, no 1. P. 1−31.
Brunton S. L., Noack B. R., Koumoutsakos P. Machine learning for fluid mechanics // Annual review of fluid mechanics. 2020. V. 52, no 1. P. 477−508.
Pereira T. D. et al. SLEAP: A deep learning system for multi-animal pose tracking // Nature methods. 2022. V. 19, no 4. P. 486−495.
LeCun Y. et al. Gradient-based learning applied to document recognition // Proceedings of the IEEE. 2002. V. 86, no 11. P. 2278−2324.
He K. et al. Deep residual learning for image recognition // Proceedings of the IEEE conference on computer vision and pattern recognition. 2016. P. 770−778.
Deng J. et al. Imagenet: A large-scale hierarchical image database // 2009 IEEE conference on computer vision and pattern recognition (CVPR 2009). 2009. P. 248−255.
Lauer J. et al. Multi-animal pose estimation, identification and tracking with DeepLabCut // Nature Methods. 2022. V. 19, no 4. P. 496−504.
Pereira T. D. et al. Fast animal pose estimation using deep neural networks // Nature methods. 2019. V. 16, no 1. P. 117−125.
Wiltschko A. B. et al. Mapping sub-second structure in mouse behavior // Neuron. 2015. V. 88, no 6. P. 1121−1135.
Cortes C., Vapnik V. Support-vector networks // Machine learning. 1995. V. 20, no 3. P. 273−297.
Breiman L. Random forests // Machine learning. 2001. V. 45, no 1. P. 5−32.
Friedman J. H. Greedy function approximation: a gradient boosting machine // Annals of statistics. 2001. P. 1189−1232.
Kabra M. et al. JAABA: interactive machine learning for automatic annotation of animal behavior // Nature methods. 2013. V. 10, no 1. P. 64−67.
Zhang L. et al. A nearly end-to-end deep learning approach to fault diagnosis of wind turbine gearboxes under nonstationary conditions // Engineering applications of artificial intelligence. 2023. V. 119. P. 105735.
Sturman O. et al. Deep learning-based behavioral analysis reaches human accuracy and is capable of outperforming commercial solutions // Neuropsychopharmacology. 2020. V. 45, no 11. P. 1942−1952.
Sherstinsky A. Fundamentals of recurrent neural network (RNN) and long short-term memory (LSTM) network // Physica D: Nonlinear Phenomena. 2020. V. 404. P. 132306.
Graves A. Long short-term memory // Supervised sequence labelling with recurrent neural networks. 2012. P. 37−45.
Graves A., Mohamed A., Hinton G. Speech recognition with deep recurrent neural networks // 2013 IEEE international conference on acoustics, speech and signal processing. Ieee, 2013. P. 6645−6649.
Ashish V. Attention is all you need // Advances in neural information processing systems. 2017. V. 30. P. 1.
Bertasius G., Wang H., Torresani L. Is space-time attention all you need for video understanding? // Icml. 2021. V. 2, no 3. P. 4.
Nascimento E. G. S. et al. T4PdM: A deep neural network based on the transformer architecture for fault diagnosis of rotating machinery // arXiv preprint arXiv:2204.03725. 2022.
Ushakova A. et al. Assessment of Gnawing Behavior in 5xFAD Mice // International Perm Forum Science and Global Challenges of the 21st Century. Cham : Springer Nature Switzerland, 2022. P. 493−498.
Barulina M. et al. Method for preprocessing video data for training deep-learning models for identifying behavioral events in bio-objects // Mathematics. 2024. V. 12. no 24. P. 3978.
Rudin C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead // Nature machine intelligence. 2019. V. 1, no 5. P. 206−215.
Carreira J., Zisserman A. Quo vadis, action recognition? a new model and the kinetics dataset // proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2017. P. 6299−6308.
Tran D. et al. Learning spatiotemporal features with 3d convolutional networks // Proceedings of the IEEE international conference on computer vision. 2015. P. 4489−4497.
Shirobokova S. et al. Application of Transition Matrices as Selective Descriptors for Behavioral Phenotyping of Young 5xFAD Animals // International Perm Forum Science and Global Challenges of the 21st Century. Cham: Springer Nature Switzerland, 2022. P. 499−506.
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2026 Илья Иванович Коваленко, Евгения Александровна Ахременко, Александр Игоревич Андреев, Марина Александровна Барулина

This work is licensed under a Creative Commons Attribution 4.0 International License.
Articles are published under license Creative Commons Attribution 4.0 International (CC BY 4.0).
