Education


Conference Publications

ICPAD
In-Context Policy Adaptation via Cross-Domain Skill Diffusion
Minjong Yoo*, Woo Kyung Kim, Honguk Woo
AAAI, 2025.02, Philadelphia, United States

In this work, we present an in-context policy adaptation (ICPAD) framework designed for long-horizon multi-task environments, exploring diffusion-based skill learning techniques in cross-domain settings.

LDuS
LLM-based Skill Diffusion for Zero-shot Policy Adaptation
Woo Kyung Kim*, Youngseok Lee, Jooyoung Kim, Honguk Woo
NeurIPS, 2024.12, Vancouver, Canada

In this paper, we present a novel LLM-based policy adaptation framework LDuS which leverages an LLM to guide the generation process of a skill diffusion model upon contexts specified in language, facilitating zero-shot skill-based policy adaptation to different contexts.

IsCiL
Incremental Learning of Retrievable Skills for Efficient Continual Task Adaptation
NeurIPS, 2024.12, Vancouver, Canada

We introduce IsCiL, an adapter-based continual imitation learning framework that incrementally learns sharable skills from different demonstrations, enabling sample efficient task adaptation using the skills.

ParIRL
Pareto Inverse Reinforcement Learning for Diverse Expert Policy Generation
Woo Kyung Kim*, Minjong Yoo, Honguk Woo
IJCAI, 2024.08, Jeju, Korea

In this paper, we present Pareto inverse reinforcement learning (ParIRL) framework in which a Pareto policy set corresponding to the best compromise solutions over multi-objectives can be induced.

DEDER
Embodied CoT Distillation From LLM To Off-the-shelf Agents
ICML, 2024.07, Wien, Austria

We present DEDER, a framework for decomposing and distilling the embodied reasoning capabilities from large language models (LLMs) to efficient, small language model (sLM)-based policies.

DuSkill
Robust Policy Learning via Offline Skill Diffusion
Woo Kyung Kim*, Minjong Yoo, Honguk Woo
AAAI, 2024.02, Vancouver, Canada

We present a novel offline skill learning (DuSkill) framework which employs a guided Diffusion model to generate versatile skills extended from the limited skills in datasets, thereby enhancing the robustness of policy learning for tasks in different domains.

ConPE
Efficient Policy Adaptation with Contrastive Prompt Ensemble for Embodied Agents
Wonje Choi*, Woo Kyung Kim, SeungHyun Kim, Honguk Woo
NeurIPS, 2023.12, New Orleans, United States

We present a novel contrastive prompt ensemble (ConPE) framework which utilizes a pretrained vision-language model and a set of visual prompts, thus enables efficient policy learning and adaptation upon environmental and physical changes encountered by embodied agents.

OnIS
One-shot Imitation in a Non-stationary Environment via Multi-modal Skill
ICML, 2023.07, Honolulu, United States

In this paper, we explore the compositionality of complex tasks, and present a novel skill-based imitation learning (OnIS) framework enabling one-shot imitation and zero-shot adaptation.


Journal Publications

A2D2
Aspect-Augmented Distillation of Task-Oriented Dialogues to Small Language Models
Jongmoon Jun, Woo Kyung Kim, Hyunseong Na, Honguk Woo, Jeehyeong Kim
Expert Systems with Applications, 2025.03. Volume 302, Page 130494

We present A2D2, an aspect-augmented dialogue distillation framework designed to transfer capabilities from larger language models to smaller ones for task-oriented dialogue systems, incorporating human aspect-aware capabilities while maintaining task requirements.

Repot
Repot: Transferable Reinforcement Learning for Quality-Centric Networked Monitoring in Various Environments
Youngseok Lee*, Woo Kyung Kim, Sung Hyun Choi, Honguk Woo
IEEE Access, 2023.11. Volume 9, Page 147280-147294

In this paper, we present a transferable RL model Repot in which a policy trained in an easy-to-learn network environment can be readily adjusted in various target network environments.