Conference

Enhancing Child Vocalization Classification with Phonetically-Tuned Embeddings for Assisting Autism Diagnosis

Jialu Li, Mark Hasegawa-Johnson, and Karrie Karahalios

Accepted to Interspeech 2024, June 2024 


Sound Tagging in Infant-centric Home Soundscapes

Mohammad Nur Hossain Khan, Jialu Li, Nancy L. McElwain, Mark Hasegawa-Johnson, and Bashima Islam

Accepted to IEEE/ACM International Conference on Connected Health: Applications, Systems and Engineering Technologies (CHASE 2024), February 2024


Towards Robust Family-Infant Audio Analysis Based on Unsupervised Pretraining of Wav2vec 2.0 on Large-Scale Unlabeled Family Audio 

Jialu Li, Mark Hasegawa-Johnson, and Nancy L. McElwain

Published in the Proceedings of Interspeech, Dublin, Ireland, August 2023


Listen, Decipher and Sign: Toward Unsupervised Speech-to-Sign Language Recognition 

[PDF]

Liming Wang, Junrui Ni, Heting Gao, Jialu Li, Kai Chieh Chang, Xulin Fan, Junkai Wu, Mark Hasegawa-Johnson, and Chang D. Yoo

Published in the Findings of Association for Computational Linguistics (Findings of ACL’23), July 2023


Autosegmental Neural Nets: Should Phones and Tones be Synchronous or Asynchronous? 

Jialu Li and Mark Hasegawa-Johnson

Published in the Proceedings of Interspeech, Shanghai, China, October 2020

Journal

Preliminary Technical Validation of LittleBeats™: A Multimodal Sensing Platform to Capture Cardiac Physiology, Motion, and Vocalizations

[PDF]

Bashima Islam, Nancy L. McElwain, Jialu Li, Maria Davila, Yannan Hu, Kexin Hu, Jordan Bodway, Ashutosh Dhekne, Romit Roy Choudhury, and Mark Hasegawa-Johnson

Published in Journal of Sensors, January 2024


Autosegmental neural nets 2.0: An extensive study of training synchronous and asynchronous phones and tones for under-resourced tonal languages 

[PDF]

Jialu Li and Mark Hasegawa-Johnson

Published in Journal of IEEE Transactions on Audio, Speech, and Language Processing, May 2022


Analysis of acoustic and voice quality features for the classification of infant and mother vocalizations

[PDF]

Jialu Li, Mark Hasegawa-Johnson, and Nancy L. McElwain

Published in Journal of Speech Communication, July 2021


An embodied, platform-invariant architecture for connecting high-level spatial commands to platform articulation 

[PDF]

Anum Jang Sher, Umer Huzaifa, Jialu Li, Varun Jain, Alex Zurawski, and Amy LaViers

Published in Journal of Robotics and Autonomous Systems, July 2019

Workshop

Analysis of Self-Supervised Speech Models on Children's Speech and Infant Vocalizations

Jialu Li, Mark Hasegawa-Johnson, and Nancy L. McElwain

Accepted to the IEEE ICASSP 2024 workshop Self-supervision in Audio, Speech and Beyond (SASB), January 2024


A Comparable Phone Set for the TIMIT Dataset Discovered in Clustering of Listen, Attend and Spell 

Jialu Li and Mark Hasegawa-Johnson

Published in the Workshop on Interpretability and Robustness in Audio, Speech, and Language (IRASL), NeurIPS, Montreal, Canada, December 2018 

Preprint

Visualizations of complex sequences of family-infant vocalizations using bag-of-audio-words approach based on wav2vec 2.0 features 

[PDF]

Jialu Li, Mark Hasegawa-Johnson, and Nancy L. McElwain

Preprint, March 2022


Accent-robust automatic speech recognition using supervised and unsupervised wav2vec embeddings 

[PDF]

Jialu Li, Vimal Manohar, Pooja Chitkara, Andros Tjandra, Michael Picheny, Frank Zhang, Xiaohui Zhang, Yatharth Saraf

Preprint, October 2021