Research
I have a broad interest in vision-and-language and video modeling.
Recently, I have been researching methods for Large Language Models (LLMs) to understand videos and effectively represent temporal information.
|
Publications
|
PGA: Personalizing Grasping Agents with Single Human-Robot Interaction
Junghyun Kim, Gi-Cheon Kang, Jaein Kim, Seoyun Yang, Minjoon Jung,
Byoung-Tak Zhang*
arXiv, 2023
paper /
code
|
|
Overcoming Weak Visual-Textual Alignment for Video Moment Retrieval
Minjoon Jung,
Youwon Jang,
Seongho Choi,
Joochan Kim,
Jin-Hwa Kim*,
Byoung-Tak Zhang*
arXiv, 2023
paper /
code
|
|
Modal-specific Pseudo Query Generation for Video Corpus Moment Retrieval
Minjoon Jung,
Seongho Choi,
Joochan Kim,
Jin-Hwa Kim*,
Byoung-Tak Zhang*
EMNLP, 2022
paper /
code
|
|
Stagemix video generation using face and body keypoints detection
Minjoon Jung,
Seung-Hyun Lee,
Eunseon Sim,
Minho Jo,
Yujin Lee,
Hyebin Choi,
Junseok Kwon*
Multimedia Tools and Applications, 2022
paper /
code
|
|
Toward a Human-Level Video Understanding Intelligence
Yu-Jung Heo,
Minsu Lee,
Seongho Choi,
Woo Suk Choi,
Minjung Shin,
Minjoon Jung,
Jeh-Kwang Ryu,
Byoung-Tak Zhang*
AAAI 2021 Fall Symposium Series on Artificial Intelligence for Human-Robot Interaction, 2021
paper /
code
|
Service
- Conference Reviewer: EMNLP 2022, ACL 2023, EMNLP 2023
|
|