Mehedi Hasan Bijoy

I am a PhD student in the AaltoASR research group, supervised by Prof. Mikko Kurimo, working in speech and language processing and specializing in dialectal/accented automatic speech recognition. My research interests include speech-focused LLMs, multimodality, and RL.

I am open to collaborations and new opportunities.

When it comes to leisure, I enjoy playing cricket \|/🏏, cooking without following a recipe πŸ§‘β€πŸ³πŸ₯£, and wandering deep into forests πŸŒ³πŸšΆπŸ»β€βž‘οΈπŸŒ΄.

Email  /  CV  /  LinkedIn  /  Google Scholar  /  GitHub

profile photo
Recent News
  • (September 2025): Started my PhD in Speech and Language Processing πŸŽ‰.
  • (August 2025): Presented a first-authored paper at Interspeech 2025 πŸŽ‰.
  • (June 2025): Completed my master’s degree with honours πŸŽ‰.
  • (January 2025): A first-authored journal article was published in Computer Speech & Language πŸŽ‰.
  • (August 2024): A first-authored paper was published in ACMM 2024 / MuSe 2024 πŸŽ‰.
  • (July 2024): Our team (AaltoASR) placed 3rd in both shared tasks at ACMM 2024 / MuSe 2024 πŸ₯‰πŸ†.
  • (April 2024): A first-authored journal article was published in Neural Computing and Applications πŸŽ‰.
  • (March 2024): Joined the Automatic Speech Recognition research group as a Research Assistant πŸ•΅.
  • (February 2024): A first-authored journal article was published in IEEE Access πŸŽ‰.
  • (December 2023): A first-authored paper was accepted at EMNLP 2023 / BLP 2023 πŸŽ‰.
  • (October 2023): Started an MSc in Speech and Language Processing at Aalto University πŸ‘¨β€πŸ’».
  • (September 2023): Served as a Reviewer πŸ•΅οΈβ€β™‚οΈ for the BLP workshop at EMNLP 2023.
  • (January 2023): Joined Bangladesh University of Business and Technology as a Lecturer πŸ‘¨β€πŸ«.



Blogs
clean-usnob Teaching Speech Recognition with a Team of Experts: How Multi-Teacher Knowledge Distillation Helps Machines Learn to Listen Carefully
Keywords: Multi-Teacher Knowledge Distillation, Accented ASR, Multilingual SER

Multi-teacher knowledge distillation is a training approach where a student model learns from the combined guidance of multiple teacher models, capturing diverse knowledge to improve speech recognition performance. Learn More β†’



Tutorials
clean-usnob Handritten Character Recognition
It showcases Bangla Handwritten Character Recognition using various Convolutional Neural Network architectures, implemented with both Keras and PyTorch.
Video / GitHub Repo.


Oral Presentations
clean-usnob Multi-Teacher Knowledge Distillation for Accented English Speech Recognition
Publications: @Interspeech 2025 and @----

This thesis develops a dual-adaptive multi-teacher knowledge distillation framework that fuses expertise from multiple accent-specialist models into a single compact student, enabling robust and efficient speech recognition across diverse and unseen English accents.
Video / GitHub Repo.