Conference
Enhancing Child Vocalization Classification with Phonetically-Tuned Embeddings for Assisting Autism Diagnosis
Jialu Li, Mark Hasegawa-Johnson, and Karrie Karahalios
Accepted to Interspeech 2024, June 2024
Sound Tagging in Infant-centric Home Soundscapes
Mohammad Nur Hossain Khan, Jialu Li, Nancy L. McElwain, Mark Hasegawa-Johnson, and Bashima Islam
Accepted to IEEE/ACM International Conference on Connected Health: Applications, Systems and Engineering Technologies (CHASE 2024), February 2024
Towards Robust Family-Infant Audio Analysis Based on Unsupervised Pretraining of Wav2vec 2.0 on Large-Scale Unlabeled Family Audio
Jialu Li, Mark Hasegawa-Johnson, and Nancy L. McElwain
Published in the Proceedings of Interspeech, Dublin, Ireland, August 2023
Listen, Decipher and Sign: Toward Unsupervised Speech-to-Sign Language Recognition
Liming Wang, Junrui Ni, Heting Gao, Jialu Li, Kai Chieh Chang, Xulin Fan, Junkai Wu, Mark Hasegawa-Johnson, and Chang D. Yoo
Published in the Findings of Association for Computational Linguistics (Findings of ACL’23), July 2023
Autosegmental Neural Nets: Should Phones and Tones be Synchronous or Asynchronous?
Jialu Li and Mark Hasegawa-Johnson
Published in the Proceedings of Interspeech, Shanghai, China, October 2020
Journal
Preliminary Technical Validation of LittleBeats™: A Multimodal Sensing Platform to Capture Cardiac Physiology, Motion, and Vocalizations
Bashima Islam, Nancy L. McElwain, Jialu Li, Maria Davila, Yannan Hu, Kexin Hu, Jordan Bodway, Ashutosh Dhekne, Romit Roy Choudhury, and Mark Hasegawa-Johnson
Published in Journal of Sensors, January 2024
Autosegmental neural nets 2.0: An extensive study of training synchronous and asynchronous phones and tones for under-resourced tonal languages
Jialu Li and Mark Hasegawa-Johnson
Published in Journal of IEEE Transactions on Audio, Speech, and Language Processing, May 2022
Analysis of acoustic and voice quality features for the classification of infant and mother vocalizations
Jialu Li, Mark Hasegawa-Johnson, and Nancy L. McElwain
Published in Journal of Speech Communication, July 2021
An embodied, platform-invariant architecture for connecting high-level spatial commands to platform articulation
Anum Jang Sher, Umer Huzaifa, Jialu Li, Varun Jain, Alex Zurawski, and Amy LaViers
Published in Journal of Robotics and Autonomous Systems, July 2019
Workshop
Analysis of Self-Supervised Speech Models on Children's Speech and Infant Vocalizations
Jialu Li, Mark Hasegawa-Johnson, and Nancy L. McElwain
Accepted to the IEEE ICASSP 2024 workshop Self-supervision in Audio, Speech and Beyond (SASB), January 2024
A Comparable Phone Set for the TIMIT Dataset Discovered in Clustering of Listen, Attend and Spell
Jialu Li and Mark Hasegawa-Johnson
Published in the Workshop on Interpretability and Robustness in Audio, Speech, and Language (IRASL), NeurIPS, Montreal, Canada, December 2018
Preprint
Visualizations of complex sequences of family-infant vocalizations using bag-of-audio-words approach based on wav2vec 2.0 features
Jialu Li, Mark Hasegawa-Johnson, and Nancy L. McElwain
Preprint, March 2022
Accent-robust automatic speech recognition using supervised and unsupervised wav2vec embeddings
Jialu Li, Vimal Manohar, Pooja Chitkara, Andros Tjandra, Michael Picheny, Frank Zhang, Xiaohui Zhang, Yatharth Saraf
Preprint, October 2021