Лаборатория речевых и многомодальных интерфейсов

Публикации

2019

Verkholyak O., Fedotov D., Kaya H., Zhang Y., Karpov A. Hierarchical Two-Level Modelling of Emotional States in Spoken Dialog Systems. In Proc. 44th IEEE International Conference on Acoustics, Speech, and Signal Processing ICASSP-2019, Brighton, UK, 2019, pp. 6700-6704.
Подробнее
Kaya H., Fedotov D., Dresvyanskiy D., Doyran M., Mamontov D., Markitantov M., Akdag Salah A., Kavcar E., Karpov A., Salah A.A. Predicting depression and emotions in the cross-roads of cultures, para-linguistics, and non-linguistics. In Proc. 9th International Audio/Visual Emotion Challenge and Workshop AVEC’19, co-located with ACM Multimedia 2019, Nice, France, 2019, pp. 27-35.
Подробнее
Ryumin D., Ivanko D., Kagirov I., Axyonov A., Karpov A., Zelezny M. Human-Robot Interaction with Smart Shopping Trolley using Sign Language: Data Collection. In Proc. 2019 IEEE International Conference on Pervasive Computing and Communications Workshops, PerCom Workshops 2019, Kyoto, Japan, 2019, pp. 949-954.
Подробнее
Akhtiamov O., Siegert I., Karpov A., Minker W. Cross-Corpus Data Augmentation for Acoustic Addressee Detection. In Proc. 20th ACL International Conference on Discourse and Dialogue SIGDial-2019, Stockholm, Sweden, 2019, pp. 274-283.
Подробнее
Fedotov D., Kim B., Karpov A., Minker W. Time-Continuous Emotion Recognition Using Spectrogram Based CNN-RNN Modelling // Lecture Notes in Computer Science, Springer LNAI 11658, SPECOM 2019, 2019, pp. 93-102.
Подробнее
Yu J., Markov K., Karpov A. Speaking Style Based Apparent Personality Recognition // Lecture Notes in Computer Science, Springer LNAI 11658, SPECOM 2019, 2019, pp. 540-548.
Подробнее
Verkholyak O.V., Kaya H., Karpov A.A. Modeling short-term and long-term dependencies of the speech signal for paralinguistic emotion classification // Труды СПИИРАН (SPIIRAS Proceedings), Вып. 62, № 1, 2019, С. 30-56.
Подробнее
Иванько Д.В., Рюмин Д.А., Карпов А.А., Железны М. Исследование влияния высокоскоростных видеоданных на точность распознавания аудиовизуальной речи // Информационно-управляющие системы (Informatsionno-Upravliaiushchie Sistemy) [Information and Control Systems], № 2, 2019, C. 26-34.
Подробнее
Федотов Д.В., Верхоляк О.В., Карпов А.А. Контекстное непрерывное распознавание эмоций в русской речи с использованием рекуррентных нейронных сетей. Труды 8-го междисциплинарного семинара «Анализ разговорной русской речи» АР3-2019, Санкт-Петербург, СПбГУ, 2019, С. 96-99.
Ryumin D., Kagirov I., Ivanko D., Axyonov A., Karpov A. Automatic detection and recognition of 3D manual gestures for human-machine interaction. International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences ISPRS Archives 42(2/W12), 2019, pp. 179-183.
Подробнее
Ivanko D., Ryumin D., Karpov A. Automatic lip-reading of hearing impaired people // International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences - ISPRS Archives 42(2/W12), 2019, pp. 97-101.
Подробнее
Kipyatkova I. LSTM-Based Language Models for Very Large Vocabulary Continuous Russian Speech Recognition System // Lecture Notes in Computer Science, Springer LNAI 11658, SPECOM 2019, 2019, pp. 219-226.
Подробнее
Markovnikov N., Kipyatkova I. Investigating Joint CTC-Attention Models for End-to-End Russian Speech Recognition // Lecture Notes in Computer Science, Springer LNAI 11658, SPECOM 2019, 2019, pp. 337-347.
Markitantov M., Verkholyak O. Automatic Recognition of Speaker Age and Gender Based on Deep Neural Networks // Lecture Notes in Computer Science, Springer LNAI 11658, SPECOM 2019, 2019, pp. 327-336.
Kagirov I., Ryumin D., Axyonov A. Method for Multimodal Recognition of One-Handed Sign Language Gestures Through 3D Convolution and LSTM Neural Networks // Lecture Notes in Computer Science, Springer LNAI 11658, SPECOM 2019, 2019, pp. 191-200.