BioEmre received the B.Sc. degree in electrical and electronics engineering from the Middle East Technical University (METU), Turkey in 2008 and the M.Sc. degree in electrical engineering from the Royal Institute of Technology (KTH), Sweden in 2010. Then, he worked as a part-time researcher in the Institute of Communication Systems and Data Processing (IND), RWTH Aachen, Germany. He joined the Department of Electrical Engineering (ESAT), KU Leuven, Belgium as a Ph.D. candidate in January 2011 and received the Ph.D. degree in May 2015. From July 2015 to May 2018, he was a postdoctoral researcher at the CLST group in Radboud University Nijmegen, Netherlands. He was a research fellow in the Department of Electrical and Computer Engineering at the National University of Singapore from June 2018 to March 2020. Currently, he is an automatic speech recognition (ASR) engineer at the Speech Technology & Research (STAR) Lab in SRI International, CA, USA. His research interests are ASR of code-switching speech, noise robust ASR, ASR of low-resourced languages, ASR of children and pathological speech, medical applications of ASR, speech enhancement and speaker & language recognition.
Contact details
Work address: 333 Ravenswood Ave, Menlo Park, CA 94024, USA E-mail address: emre.yilmaz (at) sri.com, emrey (at) kth.se
Skype: schemreier1
2020 - E.
Yılmaz, Ö. B. Gevrek, J. Wu, Y. Chen, X. Meng and H. Li, “Deep Convolutional Spiking Neural Networks for Keyword Spotting,” in Proc. INTERSPEECH 2020, pp. 2257-2561, Shanghai, China, Oct. 2020. [Postprint]
- X. Zhou, E.
Yılmaz, Y. Long, Y. Li and H. Li, “Multi-Encoder-Decoder Transformer for Code-Switching Speech Recognition,” in Proc. INTERSPEECH 2020, pp. 1042-1046, Shanghai, China, Oct. 2020. [arXiv] [Postprint]
- X. Zhou, G. Lee, E.
Yılmaz, Y. Long, J. Liang and H. Li, “Self-and-Mixed Attention Decoder with Deep Acoustic Structure for Transformer-based LVCSR,” in Proc. INTERSPEECH 2020, pp. 5016-5020, Shanghai, China, Oct. 2020. [arXiv] [Postprint]
- J. van Hout, V. Mitra, H. Franco, E. Yılmaz, “Time-Frequency Convolutional Neural Networks with Bottleneck Architecture for Query-by-Example Processing,” US Patent 10,777,188 B2, issued on Nov. 14, 2018, granted on Sept. 15, 2020. [Postprint]
- C. Gupta, E.
Yılmaz and H. Li, “Automatic Lyrics Alignment and Transcription in Polyphonic Music: Does Background Music Help?,” in Proc. ICASSP 2020, pp. 496-500, Barcelona, Spain, May 2020. [arXiv] [Postprint]
- C. Gupta, E.
Yılmaz and H. Li, “NUS Auto Lyrix Align,” ICASSP 2020, Show & Tell, Barcelona, Spain, May 2020.
- N. Wilkinson, A. Biswas, E.
Yılmaz, F. De Wet, E. Van der Westhuizen and T. Niesler, “Semi-Supervised Acoustic Modeling for Five-lingual Code-switched ASR Using Automatically-Segmented Soap Opera Speech,” in Proc. SLTU-CCURL, pp. 70-78, Marseille, France, May 2020. [arXiv] [Postprint]
- A. Biswas, E.
Yılmaz, F. De Wet, E. Van der Westhuizen and T. Niesler, “Semi-Supervised Development of ASR Systems for Multilingual Code-Switched Speech in Under-Resourced Languages,” in Proc. LREC, pp. 3468-3474, Marseille, France, May 2020. [arXiv] [Postprint]
- J. Wu, E.
Yılmaz, M. Zhang, H. Li and K. C. Tan, “Deep Spiking Neural Networks for Large Vocabulary Automatic Speech Recognition,” Frontiers in Neuroscience, vol. 14 (199), pp. 1-14, Mar. 2020. [arXiv] [Postprint]
2019 - X. Yue, G. Lee, E.
Yılmaz, F. Deng and H. Li, “End-to-End Code-Switching ASR for Low-Resourced Language Pairs,” in Proc. IEEE ASRU Workshop, pp. 972-979, Singapore, Dec. 2019. [arXiv] [Postprint]
- Y. Zhou, X. Tian, E.
Yılmaz, R. K. Das and H. Li, “A Modularized Neural Network with Language-Specific Output Layers for Cross-lingual Voice Conversion,” in Proc. IEEE ASRU Workshop, pp. 160-167, Singapore, Dec. 2019. [arXiv] [Postprint]
- C. Gupta, E.
Yılmaz and H. Li, “Lyrics-to-Audio Alignment with Music-Aware Acoustic Models,” in International Music Information Retrieval Evaluation Exchange (MIREX), Nov. 2019. [Abstract] [Poster]
- E.
Yılmaz, V. Mitra, G. Sivaraman and H. Franco, “Articulatory and Bottleneck Features for Speaker-Independent ASR of Dysarthric Speech,” Computer Speech & Language, vol. 58, pp. 319-334, Nov. 2019. [Preprint] [arXiv] [Postprint]
- E.
Yılmaz, S. Cohen, X. Yue, D. van Leeuwen and H. Li, “Multi-Graph Decoding for Code-Switching ASR,” in Proc. INTERSPEECH, pp. 3750-3754, Graz, Austria, Sept. 2019. [arXiv] [Postprint]
- Q. Wang, E.
Yılmaz, A. Derinel and H. Li, “Code-Switching Detection Using ASR-Generated Language Posteriors,” in Proc. INTERSPEECH, pp. 3740-3744, Graz, Austria, Sept. 2019. [arXiv] [Postprint]
- E.
Yılmaz, A. Derinel, Z. Kun, H. van den Heuvel, N. Brummer, H. Li and D. van Leeuwen, “Large-Scale Speaker Diarization of Radio Broadcast Archives,” in Proc. INTERSPEECH, pp. 411-415, Graz, Austria, Sept. 2019. [arXiv] [Postprint]
- A. Biswas, E.
Yılmaz, F. De Wet, E. Van der Westhuizen and T. Niesler, “Semi-Supervised Acoustic Model Training for Five-lingual Code-Switched ASR,” in Proc. INTERSPEECH, pp. 3745-3749, Graz, Austria, Sept. 2019. [arXiv] [Postprint]
- C. Gupta, E.
Yılmaz and H. Li, “Acoustic Modeling for Lyrics-to-Audio Alignment,” in Proc. INTERSPEECH, pp. 2040-2044, Graz, Austria, Sept. 2019. [arXiv] [Postprint]
- K. A. Lee et al., “I4U Submission to NIST SRE 2018: Leveraging from a Decade of Shared Experiences,” in Proc. INTERSPEECH, pp. 1497-1501, Graz, Austria, Sept. 2019. [arXiv] [Postprint]
2018 - E.
Yılmaz, M. McLaren, H. van den Heuvel and D. van Leeuwen, “Semi-Supervised Acoustic Model Training for Speech with Code-Switching,” Speech
Communication, vol. 105, pp. 12-22, Dec. 2018. [Preprint] [arXiv] [Postprint]
- L. Xu, R. K. Das, E.
Yılmaz, J. Yang and H. Li, “Generative X-vectors for Text-Independent Speaker Verification,” in Workshop on Spoken Language Technologies (SLT), pp. 1014-1020, Athens, Greece, Dec. 2018. [Preprint] [arXiv] [Postprint]
- E.
Yılmaz, H. van den Heuvel and D. van Leeuwen, “Code-Switching Detection with Data-Augmented Acoustic and Language Models,” in 6th Workshop on Spoken Language Technologies for Under-resourced Languages (SLTU), pp. 127-131, India, Sept. 2018. [Preprint] [arXiv] [
|
|