The Audio and Speech Processing with Deep Learning (ASPDL) Group is a team of young researchers focused on basic and applied research within the area of Digital Speech Processing. The group is mainly focused on modern deep-learning models for processing single and multimodal speech signals, observed through a single sensor. From the application point of view, the group is mainly interested in the processing of audio and biomedical signals for various real-time speech applications.
Research Grants and Collaborations:
Co-Principle Investigator with Princess Nourah Bint Abdulrahman University on an AI-Driven Real-Time Processing in Application to Speech Enhancement \& Joint Optimization for ASR Systems Project (Grant No: RPFAP-30-1445)
Co-Principle Investigator with Princess Nourah Bint Abdulrahman University on a Deep Neural Networks-based Multimodal Prediction of Alzheimer’s Disease and Its Progression Systems Project (Grant No: RPFAP-75-1445).
Supervised the Deep Neural Network-based COVID-19 Mask and Temperature Detection, Funded by National ICT RD, Ministry of IT, Government of Pakistan.
Collaborating with the National Engineering Laboratory for Speech and Language Information Processing, University of Science and Technology of China, Hefei, China, on real-time speech enhancement systems on embedded systems using compact neural models.
Collaborating with the Texas Instruments DSP Joint Laboratory, School of Electrical and Information Engineering, Tianjin University, China, on speech enhancement systems using novel neural models.
Research Fellow in Towards Cognitively Inspired 5G-IoT Enabled, Multi-Modal Hearing Aids (COG-MHEAR, EPSRC Programme Grant (2021-2026) (EP/T021063/1).
Team Members
Dr. Nasir Saleem (Ph.D/Postdoc)
Dr. Muhammad Irfan (Ph.D)
Dr. Sami Bourous (Ph.D)
Fazal E Wahab (Ph.D Student)
Jawad Ali (Ph.D Student)
Junaid Mushtaq (Ph.D Student)
Fahad Khalil Peracha (Ph.D Student)
Iqra Batool (MS Student)
Sher Muhammad (MS Student)
Muhammad Uzair (MS Student)
Ibad Ullah (MS Student)
Shehzad Raheem (MS Student)
Muhammad Sami (MS Student)
Current Research Focus
Audio Only Speech Enhancement for Wireless Communication Systems
Compact Deep Learning Models for Speech Enhancement on Low-Resource Devices
Visually-Driven Speech Enhancement for Consumer Electronics
Dysarthric Speech Recognition Systems
Cross-Attentional Multimodal Audiovideo Speech Enhancement
(International Hearing-Aid Research Conference)
(INTERSPEECH 2024 Conference)