avatar

Veronica Khaustova

Ph.D. in Computer Science and Engineering from the University of Aizu, Japan
AI Research Engineer @ Eyes, JAPAN Co. Ltd
veronica.khaustova (at) gmail.com


About Me

I am an AI Research Engineer at Eyes, JAPAN, Co. Ltd and hold a Ph.D. in Computer Science and Engineering from the University of Aizu, Japan (2024).

My research focuses on the intersection of speech technology, machine learning, and language education. My doctoral work centered on computational approaches to speech prosody analysis and computer-assisted pronunciation training (CAPT) with personalized assessment for learners from diverse first-language backgrounds. Currently, I work on voice conversion for pronunciation training and domain-specific speech recognition for aviation communication. I am always open to conversations and collaborations that cross disciplinary boundaries.

Previously, I earned a BSc in Air Traffic Control and Flight Support from the Belarusian State Academy of Aviation (2017).

Research Interests

News

Scholarships & Awards

Service & Activities

Publications

Note: name changed in March 2023 from Veranika Mikhailava to Veronica Khaustova.

Journal Publications

  1. Blake, J., Bogach, N., Kusakari, A., Lezhenin, I., Khaustova, V., Xuan, S. L., & Pyshkin, E. (2024). An Open CAPT System for Prosody Practice: Practical Steps towards Multilingual Setup. Languages, 9(1), 27. DOI

  2. Mikhailava, V., Lesnichaia, M., Bogach, N., Lezhenin, I., Blake, J., & Pyshkin, E. (2022). Language Accent Detection with CNN Using Sparse Data from a Crowd-Sourced Speech Archive. Mathematics, 10(16), 2913. DOI

Conference Publications

  1. Tang, L.; Khaustova, V.; Villegas, J. (2025, October). Construction of a Japanese air traffic control communication corpus assisted with automatic speech recognition. In Proc. 159 Audio Eng. Soc. Conv. [Link] (https://aesshow2025lb.sched.com/event/294Pt/construction-of-a-japanese-air-traffic-control-communication-corpus-assisted-with-automatic-speech-recognition)

  2. Pyshkin, E., Blake, J., Khaustova, V., Lezhenin, I., Svechnikov, R., Efimov, D., Bogach, N. (2024, March). Multimodal Contextualizing and Targeting Exercises in iCAPT Systems. INTED2024. DOI

  3. Khaustova, V., Pyshkin, E., Khaustov, V., Blake, J., & Bogach, N. (2023, November). CAPTuring Accents: An Approach to Personalize Pronunciation Training for Learners with Different L1 Backgrounds. In International Conference on Speech and Computer (pp. 59–70). Springer Nature Switzerland. DOI

  4. Lesnichaia, M., Mikhailava, V., Bogach, N., Lezhenin, I., Blake, J., & Pyshkin, E. (2022, September). Classification of Accented English Using CNN Model Trained on Amplitude Mel-Spectrograms. Proc. Interspeech 2022, 3669–3673. DOI

  5. Mikhailava, V., Blake, J., Pyshkin, E., Bogach, N., Chernonog, S., Zhuikov, A., Lesnichaya, M., Lezhenin, I., & Svechnikov, R. (2022, May). Dynamic Assessment during Suprasegmental Training with Mobile CAPT. In Proc. Speech Prosody (Vol. 2022, pp. 430–434). DOI

  6. Mikhailava, V., Pyshkin, E., Blake, J., Chernonog, S., Lezhenin, I., Svechnikov, R., & Bogach, N. (2022, March). Tailoring computer-assisted pronunciation teaching: Mixing and matching the mode and manner of feedback to learners. In INTED2022 Proceedings (pp. 767–773). IATED. DOI

  7. Mikhailava, V., Pyshkin, E., & Klyuev, V. (2020, February). Aesthetic evaluation of food plate images using deep learning. In 2020 22nd International Conference on Advanced Communication Technology (ICACT) (pp. 285–289). IEEE. DOI

  8. Mikhailava, V., Khaustov, V., & Klyuev, V. (2018, November). Overview and Categorization of Recent Approaches to Microblog Classification. In Proceedings of the 3rd International Conference on Applications in Information Technology (pp. 127–130). DOI