I aim to develop scalable reinforcement learning (RL) algorithms that (1) improve performance as more data becomes available, (2) maintain robust performance in long-horizon episodes, and (3) generalize well across diverse tasks.
To achieve this, I believe it is essential to leverage the previously collected data and advance multi-task learning. With this perspective, I have been deeply engaged in data-driven RL, particularly offline RL. Additionally, I am interested in learning approaches such as preference-based RL, goal-conditioned RL, and imitation learning.
M.S./Ph.D. [2023.03~present] (Advisor: Taesup Moon)
Dept. of Electrical and Computer Engineering, Seoul National University
Machine Intelligence and Data Science Lab (M.IN.D Lab)
B.S. [2017.03~2023.02]
Dept. of Electrical and Computer Engineering, Seoul National University
Leave of absence for military service: 2019.03~2021.02
Option-aware Temporally Abstracted Value for Offline Goal-Conditioned Reinforcement Learning
Hongjoon Ahn*, Heewoong Choi*, Jisu Han*, Taesup Moon
NeurIPS 2025 (Spotlight)
Listwise Reward Estimation for Offline Preference-based Reinforcement Learning
Heewoong Choi, Sangwon Jung, Hongjoon Ahn, and Taesup Moon
ICML 2024
paper | code | project page
NCIS: Neural Contextual Iterative Smoothing for Purifying Adversarial Perturbations
Sungmin Cha, Naeun Ko, Heewoong Choi, Youngjoon Yoo, and Taesup Moon
WACV 2024
[2017.3~2023.02] Presidential Science Scholarship
Full enrollment fee from Korea Student Aid Foundation (KOSAF)
[2022.06~2022.11] Intern at Naver Clova, Image Vision