Speech and Multimodal Interfaces Laboratory

Markitantov Maxim Viktorovich

Markitantov Maxim Viktorovich
Position
Junior researcher
Qualification
PhD student

Publications

2020

Markitantov M., Dresvyanskiy D., Mamontov D., Kaya H., Minker W., Karpov A. Ensembling End-to-End Deep Models for Computational Paralinguistics Tasks: ComParE 2020 Mask and Breathing Sub-Challenges // Proceedings of INTERSPEECH-2020, ISCA, 2020. pp. 2072-2076.
Kaya H., Verkholyak O., Markitantov M., Karpov A. Combining Clustering and Functionals based Acoustic Feature Representations for Classification of Baby Sounds. In Companion Publication of the 2020 International Conference on Multimodal Interaction ICMI’20 Companion, WoCBU’20 Workshop Proceedings, ACM, 2020, pp. 509-513.
Markitantov M. Transfer Learning in Speaker’s Age and Gender Recognition // Lecture Notes in Computer Science, Springer LNAI 12335, SPECOM 2020. 2020. pp. 326-335.
Markitantov M., Karpov А. Automatic human age and gender recognition using time-delay neural networks based on acoustic features // Proceedings of III All-Russian Acoustic Conference, St. Petersburg, 2020, pp. 374-380.
Markitantov M. Analytical survey of audiovisual speech corpora for automatic speaker’s age recognition // Almanac of scientific works of young scientists of the University ITMO. 2020, vol. 3, pp. 124-128.

2019

Kaya H., Fedotov D., Dresvyanskiy D., Doyran M., Mamontov D., Markitantov M., Akdag Salah A., Kavcar E., Karpov A., Salah A.A. Predicting depression and emotions in the cross-roads of cultures, para-linguistics, and non-linguistics. In Proc. 9th ACM International Audio/Visual Emotion Challenge and Workshop AVEC’19, Nice, France, 2019, ACM, New York, NY, USA, 9 pages.
More
Markitantov M., Verkholyak O. Automatic Recognition of Speaker Age and Gender Based on Deep Neural Networks // Lecture Notes in Computer Science, Springer LNAI 11658, SPECOM 2019, 2019, pp. 327-336.