(2025) We are delighted to announce that four of our papers have been accepted to EMNLP 2025. Congratulations to all authors on this excellent achievement!
"CoBA: Counterbias Text Augmentation for Mitigating Various Spurious Correlations via Semantic Triples" (Main Conference)
"LLM Agents at the Roundtable: A Multi-Perspective and Dialectical Reasoning Framework for Essay Scoring" (Findings of EMNLP 2025)
"Beyond Single-User Dialogue: Assessing Multi-User Dialogue State Tracking Capabilities of Large Language Models" (Findings of EMNLP 2025)
"From Ground Trust to Truth: Disparities in Offensive Language Judgments on Contemporary Korean Political Discourse" (Findings of EMNLP 2025)
(2025) Our paper, "Query, Decompose, Compress: Structured Query Expansion for Efficient Multi-Hop Retrieval," has been accepted to CIKM 2025.
(2025) Our papers have been accepted to the following workshops:
ICCV 2025 Workshop on Curated Data for Efficient Learning (CDEL): "VolDoGer: LLM-assisted Datasets for Domain Generalization in Vision-Language Tasks"
COLM 2025 First Workshop on Optimal Reliance and Accountability in Interactions with Generative Language Models (ORIGen): "Medal Matters: Probing LLMs’ Failure Cases Through Olympic Rankings"
MICCAI 2025 Workshop on Applications of Medical AI (AMAI): "XTag-CLIP: Robust and Reliable Thyroid Scar Analysis with Limited Data via Cross-Attention"
(2025) We are delighted to announce that two of our papers have been accepted to the ACL 2025 Main Conference. Congratulations to all authors for this excellent achievement! The accepted papers are:
"Delving into Multilingual Ethical Bias: The MSQAD with Statistical Hypothesis Tests for Large Language Models"
"Plug-in and Fine-tuning: Bridging the Gap between Small Language Models and Large Language Models"
(2025) We are pleased to share that our paper, "Making Sense of Korean Sentences: A Comprehensive Evaluation of LLMs through KoSEnd Dataset," has been accepted to the ACL 2025 Student Research Workshop (SRW).
(2025) Prof. YoungBin Kim has been awarded the Best Reviewer Award at AISTATS 2025. (link)
(2025) We are pleased to announce that our research project, Neural Journalism++: 논리·윤리·공정성 강화 언어 모델 기반 지능형 저널리즘 연구, has been selected for funding under the Consolidator Grants (중견연구) program by the National Research Foundation of Korea (한국연구재단, NRF).
(2025) We are delighted to announce that our paper, led by Junehyoung Kwon, has been accepted to NAACL 2025 Main Conference as an ORAL PRESENTATION: "See-Saw Modality Balance: See Gradient, and Sew Impaired Vision-Language Balance to Mitigate Dominant Modality Bias"
(2025) Prof. YoungBin Kim was honored for excellence in the Fall 2024 Course Evaluation.
(2024) Prof. YoungBin Kim will serve as an Area Chair for ACL Rolling Review (ARR).
(2024) We are pleased to share that our paper has been accepted to the AAAI 2025 (Demo Track): "SummPilot: Bridging Efficiency and Customization for Interactive Summarization System" (News)
(2024) We are delighted to announce that three of our papers have been accepted to the EMNLP 2024 Main Conference. Congratulations to all the authors on this outstanding achievement! The accepted papers are (News):
(2024) We are delighted to announce that our paper, led by Soojin Jang, has been accepted to ECCV 2024. Congratulations to Soojin on this outstanding achievement! - "DIAL: Dense Image-text ALignment for Weakly Supervised Semantic Segmentation" (Link)
(2024) We are delighted to announce that our paper, led by Seunguk Yu, has been accepted to the Findings of NAACL 2024. Congratulations, Seunguk, on this excellent work! - "Don't be a Fool: Pooling Strategies in Offensive Language Detection from User-Intended Adversarial Attacks" (Link)
(2024) Our paper has been accepted to COLING 2024. - "Enhancing Effectiveness and Robustness in a Low-Resource Regime via Decision-Boundary-aware Data Augmentation" (Link)
(2024) We are pleased to inform that our paper, led by Juhwan Choi, has been accepted to the Findings of EACL 2024. Congratulations to Juhwan on this achievement! - "GPTs Are Multilingual Annotators for Sequence Generation Tasks" (Link)
(2024) We'd like to share the acceptance of a paper by Juhwan Choi to the EACL 2024 Student Research Workshop. - "AutoAugment Is What You Need: Enhancing Rule-based Augmentation Methods in Low-resource Regimes" (Link)
(2024) We are happy to share that our paper, led by JuneHyoung Kwon, has been accepted to WACV 2024. Well done, JuneHyoung! - "Learning to Detour: Shortcut Mitigating Augmentation for Weakly Supervised Semantic Segmentation" (Link)
(2023) We are pleased to announce that our paper, led by Jungmin Yun, has been accepted to the Findings of EMNLP 2023 and will be presented in-person at the SpLU-RoboNLP workshop. Congratulations to Jungmin on this achievement! - "Focus on the Core: Efficient Attention via Pruned Token Compression for Document Classification" (Link)
우리 연구팀은 AI 연구를 통해 저자원 환경에서 발생하는 다양한 문제를 해결하는 데 주력하고 있습니다. 데이터 증강, 자연어 처리(NLP), 멀티모달 AI 등 다양한 연구 주제를 다루며, EMNLP, NAACL, COLING, ECCV, EACL, WACV 등 국제 학회에서 인정받은 연구 성과를 보유하고 있습니다. 현재 학부 연구생과 대학원생을 모집 중입니다:
AI학과: 자연어 처리(NLP)를 중심으로 감정 분석, 공격성 언어 탐지, 다국어 주석, 문법 오류 수정, 대화형 요약 시스템 등 다양한 언어 기반 AI 기술을 연구합니다.
영상학과: 텍스트와 이미지를 동시에 이해하는 멀티모달 AI를 중심으로 약지도 기반 이미지 분석, 데이터 증강 기법 개발, 모델 성능 개선 등 다양한 연구를 진행합니다.
NLP와 멀티모달 AI 연구에 관심이 있는 분들의 많은 지원 바랍니다. 간단한 자기소개와 이력서를 ybkim85@cau.ac.kr(김영빈 교수)에게 보내주세요.
Our research team is actively engaged in addressing the unique challenges presented by low-resource environments through AI research. Under the guidance of experienced researchers, we focus on areas like data augmentation, language processing, and enhancing the efficiency of AI systems. Our collaborative efforts have led to recognition at several international conferences, including EMNLP, NAACL, COLING, ICLR, EACL, and WACV.
To apply, please send a short introduction and your Curriculum Vitae to Professor Youngbin Kim at ybkim85@cau.ac.kr.
Our lab is dedicated to solving challenges and exploring opportunities within low-resource environments. By focusing on efficient and innovative approaches in data augmentation, natural language processing (NLP), and multimodal AI, we aim to make AI technologies more accessible and impactful. Our mission is to develop solutions that are practical, inclusive, and capable of addressing real-world constraints, particularly in scenarios with limited resources.
- We are advancing the frontiers of multilingual and low-resource language processing by developing inclusive technologies that enhance accessibility and usability across diverse languages and dialects. Our work includes reducing linguistic bias, improving grammatical accuracy, and enabling more equitable language technologies.
- To overcome the challenges of data scarcity, we design innovative data augmentation techniques for both text and images. These strategies include methods that progressively improve model robustness and performance, ensuring adaptability to low-resource environments while minimizing resource consumption.
- Our research focuses on optimizing the efficiency of AI systems through robust training methods and model compression techniques. By addressing computational and data limitations, we aim to ensure that AI systems remain sustainable and practical in real-world applications, even with constrained resources.
Through these focus areas, our lab contributes to making AI technologies more accessible and equitable, fostering solutions that address the global challenges of inclusivity and resource scarcity. By combining innovation with practical impact, we aim to drive the development of AI that benefits communities worldwide and meets the pressing needs of diverse, real-world applications.