I am an AI Research Engineer at Eyes, JAPAN, Co. Ltd and hold a Ph.D. in Computer Science and Engineering from the University of Aizu, Japan (2024).
My research focuses on the intersection of speech technology, machine learning, and language education. My doctoral work centered on computational approaches to speech prosody analysis and computer-assisted pronunciation training (CAPT) with personalized assessment for learners from diverse first-language backgrounds. Currently, I work on voice conversion for pronunciation training and domain-specific speech recognition for aviation communication. I am always open to conversations and collaborations that cross disciplinary boundaries.
Previously, I earned a BSc in Air Traffic Control and Flight Support from the Belarusian State Academy of Aviation (2017).
Note: name changed in March 2023 from Veranika Mikhailava to Veronica Khaustova.
Blake, J., Bogach, N., Kusakari, A., Lezhenin, I., Khaustova, V., Xuan, S. L., & Pyshkin, E. (2024). An Open CAPT System for Prosody Practice: Practical Steps towards Multilingual Setup. Languages, 9(1), 27. DOI
Mikhailava, V., Lesnichaia, M., Bogach, N., Lezhenin, I., Blake, J., & Pyshkin, E. (2022). Language Accent Detection with CNN Using Sparse Data from a Crowd-Sourced Speech Archive. Mathematics, 10(16), 2913. DOI
Tang, L.; Khaustova, V.; Villegas, J. (2025, October). Construction of a Japanese air traffic control communication corpus assisted with automatic speech recognition. In Proc. 159 Audio Eng. Soc. Conv. [Link] (https://aesshow2025lb.sched.com/event/294Pt/construction-of-a-japanese-air-traffic-control-communication-corpus-assisted-with-automatic-speech-recognition)
Pyshkin, E., Blake, J., Khaustova, V., Lezhenin, I., Svechnikov, R., Efimov, D., Bogach, N. (2024, March). Multimodal Contextualizing and Targeting Exercises in iCAPT Systems. INTED2024. DOI
Khaustova, V., Pyshkin, E., Khaustov, V., Blake, J., & Bogach, N. (2023, November). CAPTuring Accents: An Approach to Personalize Pronunciation Training for Learners with Different L1 Backgrounds. In International Conference on Speech and Computer (pp. 59–70). Springer Nature Switzerland. DOI
Lesnichaia, M., Mikhailava, V., Bogach, N., Lezhenin, I., Blake, J., & Pyshkin, E. (2022, September). Classification of Accented English Using CNN Model Trained on Amplitude Mel-Spectrograms. Proc. Interspeech 2022, 3669–3673. DOI
Mikhailava, V., Blake, J., Pyshkin, E., Bogach, N., Chernonog, S., Zhuikov, A., Lesnichaya, M., Lezhenin, I., & Svechnikov, R. (2022, May). Dynamic Assessment during Suprasegmental Training with Mobile CAPT. In Proc. Speech Prosody (Vol. 2022, pp. 430–434). DOI
Mikhailava, V., Pyshkin, E., Blake, J., Chernonog, S., Lezhenin, I., Svechnikov, R., & Bogach, N. (2022, March). Tailoring computer-assisted pronunciation teaching: Mixing and matching the mode and manner of feedback to learners. In INTED2022 Proceedings (pp. 767–773). IATED. DOI
Mikhailava, V., Pyshkin, E., & Klyuev, V. (2020, February). Aesthetic evaluation of food plate images using deep learning. In 2020 22nd International Conference on Advanced Communication Technology (ICACT) (pp. 285–289). IEEE. DOI
Mikhailava, V., Khaustov, V., & Klyuev, V. (2018, November). Overview and Categorization of Recent Approaches to Microblog Classification. In Proceedings of the 3rd International Conference on Applications in Information Technology (pp. 127–130). DOI