(* Equal Contribution, † Corresponding Authors)
FlowFixer: Towards Detail-Preserving Subject-Driven Generation
Jinyoung Jun, Won-Dong Jang, Wenbin Ouyang, Raghudeep Gadde, and Jungbeom Lee,in CVPR 2026
Focus, Don’t Prune: Identifying Instruction-Relevant Regions for Information-Rich Image Understanding
Mincheol Kwon, Minseung Lee, Seonga Choi, Miso Choi, Kyeongjin Oh, Hyunyoung Lee, Cheonyoung Park, Yongho Song, Seunghyun Park†, and Jinkyu Kim†,in CVPR 2026
MambaEye: A Size-Agnostic Visual Encoder with Causal Sequential Processing
Changho Choi, Minho Kim, and Jinkyu Kim,in CVPR 2026 (Findings)
Image-Guided Semantic Pseudo-LiDAR Point Generation for 3D Object Detection
Minseung Lee, Seokha Moon, Seung Joon Lee, Reza Mahjourian, and Jinkyu Kim,in WACV 2026
Style-Friendly SNR Sampler for Style-Driven Generation
Jooyoung Choi*, Chaehun Shin*, Yeongtak Oh, Heeseung Kim, Jungbeom Lee†, and Sungroh Yoon†,in WACV 2026
Towards Failure-Resilient Lifelong Learning Agents through Scene Graph-Guided Proactive Replanning
Che Rin Yu, Daewon Chae, Dabin Seo, Sangwon Lee, Hyeongwoo Im, and Jinkyu Kim,in AAMAS 2026 (Oral)
3D Occupancy Prediction with Low-Resolution Queries via Prototype-aware View Transformation
Gyeongrok Oh*, Sungjune Kim*, Heeju Ko, Hyung-gun Chi, Jinkyu Kim, Dongwook Lee, Daehyun Ji, Sungjoon Choi, Sujin Jang†, and Sangpil Kim†,in CVPR 2025
DiffExp: Efficient Exploration in Reward Fine-tuning for Text-to-Image Diffusion Models
Daewon Chae*, June Suk Choi*, Jinkyu Kim†, and Kimin Lee†,in AAAI 2025 (Oral)
Encoder-only Next Token Prediction
Ethan Ewer*, Daewon Chae*, Thomas Zeng*, Jinkyu Kim, and Kangwook Lee,in TMLR 2025
Fast and Precise Multimodal Spatiotemporal Calibration via Periodic-Activated 2D Gaussian Splatting
Hongbeen Park*, Minjeong Park*, Sunpil Kim, Jinkyu Kim, and Jung Hyun Lee,in ICCV 2025 (Workshop)
Fpanet: Frequency-based video demoireing using frame-level post alignment
Gyeongrok Oh, Sungjune Kim, Heon Gu, Sang Ho Yoon, Jinkyu Kim†, and Sangpil Kim†,in NN 2025
GUIDE-CoT: Goal-driven and user-informed dynamic estimation for pedestrian trajectory using chain-of-thought
Sungsik Kim, Janghyun Baek, Jinkyu Kim, and Jaekoo Lee,in AAMAS 2025
Hard-Negative Prototype-Based Regularization for Few-Shot Class-Incremental Learning
Seongbeom Park, Hyunju Yun, Daewon Chae, Sungyoon Kim, Suhong Moon, Minwoo Kang, Seunghyun Park†, and Jinkyu Kim†,in TMLR 2025
Know "No" Better: A Data-Driven Approach for Enhancing Negation Awareness in CLIP
Junsung Park, Jungbeom Lee, Jongyoon Song, Sangwon Yu, Dahuin Jung, and Sungroh Yoon,in ICCV 2025
Parameter-Efficient 12-Lead ECG Reconstruction from a Single Lead
Junseok Lee, Yeonho Yoo, Jinkyu Kim, Dosun Lim, Gyeongsik Yang, and Chuck Yoo,in MICCAI 2025
Querying Labeled Time Series Data with Scenario Programs
Edward Kim*, Devan Shanker*, Varun Bharadwaj, Hongbeen Park, Jinkyu Kim, Hazem Torfah, Daniel Fremont, and Sanjit Seshia,in NFM 2025
Scene Graph-Guided Proactive Replanning for Failure-Resilient Embodied Agents
Che Rin Yu, Daewon Chae, Dabin Seo, Yoonha Jang, Sangwon Lee, Hyeongwoo Im, and Jinkyu Kim,in RSS 2025
SemanticControl: A Training-free Approach for Handling Loosely Aligned Visual Conditions in ControlNet
Woosung Joung*, Daewon Chae*, and Jinkyu Kim,in BMVC 2025
ViTA-PAR: Visual and Textual Attribute Alignment with Attribute Prompting for Pedestrian Attribute Recognition
Minjeong Park, Hongbeen Park, and Jinkyu Kim,in ICIP 2025 (Oral)
Watermarking for Factuality: Guiding Vision-Language Models Toward Truth via Tri-layer Contrastive Decoding
Kyungryul Back, Seongbeom Park, Milim Kim, Mincheol Kwon, SangHyeok Lee, Hyunyoung Lee, Junhee Cho, Seunghyun Park, and Jinkyu Kim,in EMNLP 2025 (Findings)
Advancing Cross-Domain Generalizability in Face Anti-Spoofing: Insights, Design, and Metrics
Hyojin Kim*, Jiyoon Lee*, Yonghyun Jeong, Heeyoung Jang, Jinkyu Kim†, and Jae-Young Yoo†,in CVPR 2024 (Workshop)
Audio-guided implicit neural representation for local image stylization
Seung Hyun Lee*, Sieun Kim*, Wonmin Byeon, Gyeongrok Oh, Sumin In, Hyeongcheol Park, Sang Ho Yoon, Sung-Hee Hong, Jinkyu Kim, and Sangpil Kim,in CVM 2024
BEVMap: Map-Aware BEV Modeling for 3D Perception
Mincheol Chang, Seokha Moon, Reza Mahjourian, and Jinkyu Kim,in WACV 2024
CMDA: Cross-Modal and Domain Adversarial Adaptation for LiDAR-Based 3D Object Detection
Gyusam Chang*, Wonseok Roh*, Sujin Jang, Dongwook Lee, Daehyun Ji, Gyeongrok Oh, Jinsun Park, Jinkyu Kim†, and Sangpil Kim†,in AAAI 2024
Clustering-based Image-Text Graph Matching for Domain Generalization
Nokyung Park, Daewon Chae, Jeongyong Shim, Sangpil Kim, Eun-Sol Kim†, and Jinkyu Kim†,in ICPR 2024
Enhanced Motion Forecasting with Visual Relation Reasoning
Sungjune Kim, Hadam Baek, Seunggwan Lee, Hyung Gun Chi, Hyerin Lim, Jinkyu Kim†, and Sangpil Kim†,in ECCV 2024
Finetuning Pre-trained Model with Limited Data for LiDAR-based 3D Object Detection by Bridging Domain Gaps
Jiyun Jang, Mincheol Chang, Jongwon Park, and Jinkyu Kim,in IROS 2024
FoMu-SSL: Foundation Model-Guided Multi-Sensor Self-Supervised Learning for Remote Sensing
Dabin Seo, Haeji Jung, and Jinkyu Kim,in ICML 2024 (Workshop)
Higher-order Relational Reasoning for Pedestrian Trajectory Prediction
Sungjune Kim, Hyung-gun Chi, Hyerin Lim, Karthik Ramani, Jinkyu Kim†, and Sangpil Kim†,in CVPR 2024
InstructBooth: Instruction-following Personalized Text-to-Image Generation
Daewon Chae, Nokyung Park, Jinkyu Kim†, and Kimin Lee†,in ICML 2024 (Workshop)
Just Add $100 More: Augmenting Pseudo-LiDAR Point Cloud for Resolving Class-imbalance Problem
Mincheol Chang*, Siyeong Lee*, Jinkyu Kim, and Namil Kim,in NeurIPS 2024
LRSLAM: Low-Rank Representation of Signed Distance Fields in Dense Visual SLAM System
Hongbeen Park, Minjeong Park, Giljoo Nam, and Jinkyu Kim,in ECCV 2024
Learning Temporal Cues by Predicting Objects Move for Multi-camera 3D Object Detection
Seokha Moon, Hongbeen Park, Jungphil Kwon, Jaekoo Lee, and Jinkyu Kim,in ICRA 2024
Localization and Manipulation of Immoral Visual Cues for Safe Text-to-Image Generation
Seongbeom Park, Suhong Moon, Seunghyun Park†, and Jinkyu Kim†,in WACV 2024
Mitigating the Linguistic Gap with Phonemic Representations for Robust Cross-lingual Transfer
Haeji Jung, Changdae Oh, Jooeon Kang, Jimin Sohn, Kyungwoo Song, Jinkyu Kim, and David R Mortensen,in EMNLP 2024 (Workshop)
Robust sound-guided image manipulation
Seung Hyun Lee, Gyeongrok Oh, Wonmin Byeon, Sang Ho Yoon, Jinkyu Kim†, and Sangpil Kim†,in NN 2024
Text-Driven Prototype Learning for Few-Shot Class-Incremental Learning
Seongbeom Park*, Haeji Jung*, Daewon Chae, Hyunju Yun, Sungyoon Kim, Suhong Moon, Jinkyu Kim, and Seunghyun Park,in ICPR 2024
Toward Interactive Regional Understanding in Vision-Large Language Models
Jungbeom Lee, Sanghyuk Chun†, and Sangdoo Yun†,in NAACL 2024
Unified Domain Generalization and Adaptation for Multi-View 3D Object Detection
Gyusam Chang*, Jiwon Lee*, Donghyun Kim, Jinkyu Kim, Dongwook Lee, Daehyun Ji, Sujin Jang†, and Sangpil Kim†,in NeurIPS 2024
VisionTrap: Vision-Augmented Trajectory Prediction Guided by Textual Descriptions
Seokha Moon, Hyun Woo, Hongbeen Park, Haeji Jung, Reza Mahjourian, Hyung-gun Chi, Hyerin Lim, Sangpil Kim, and Jinkyu Kim,in ECCV 2024
Who Should Have Been Focused: Transferring Attention-Based Knowledge from Future Observations for Trajectory Prediction
Seokha Moon, Kyuhwan Yeon, Hayoung Kim, Seong-Gyun Jeong, and Jinkyu Kim,in ICPR 2024
An Embedding-Dynamic Approach to Self-Supervised Learning
Suhong Moon, Domas Buracas, Seunghyun Park, Jinkyu Kim, and John Canny,in WACV 2023
CloudNet: A LiDAR-based face anti-spoofing model that is robust against light variation
Yongrae Kim, Hyunmin Gwak, Jaehoon Oh, Minho Kang, Jinkyu Kim, Hyun Kwon†, and Sunghwan Kim†,in IEEE Access 2023
Ensuring visual commonsense morality for text-to-image generation
Seongbeom Park, Suhong Moon, and Jinkyu Kim,in ICML 2023 (Workshop)
Improving Visual Prompt Tuning for Self-supervised Vision Transformers
Seungryong Yoo, Eunji Kim, Dahuin Jung, Jungbeom Lee, and Sungroh Yoon,in ICML 2023
Resolving class imbalance for lidar-based object detector by dynamic weight average and contextual ground truth sampling
Daeun Lee, Jongwon Park, and Jinkyu Kim,in WACV 2023
The Power of Sound (TPoS): Audio Reactive Video Generation with Stable Diffusion
Yujin Jeong, Wonjeong Ryoo, Seunghyun Lee, Dabin Seo, Wonmin Byeon, Sangpil Kim†, and Jinkyu Kim†,in ICCV 2023
Weakly Supervised Referring Image Segmentation with Intra-Chunk and Inter-Chunk Consistency
Jungbeom Lee, Sungjin Lee, Jinseok Nam, Seunghak Yu, Jaeyoung Do, and Tara Taghavi,in ICCV 2023
Anti-Adversarially Manipulated Attributions for Weakly Supervised Semantic Segmentation and Object Localization
Jungbeom Lee, Eunji Kim, Jisoo Mok, and Sungroh Yoon,in IEEE TPAMI 2022
Bridging the Gap between Classification and Localization for Weakly Supervised Object Localization
Eunji Kim, Siwon Kim, Jungbeom Lee, Hyunwoo Kim, and Sungroh Yoon,in CVPR 2022
Bridging the domain gap towards generalization in automatic colorization
Hyejin Lee*, Daehee Kim*, Daeun Lee, Jinkyu Kim†, and Jaekoo Lee†,in ECCV 2022
Grounding visual representations with texts for domain generalization
Seonwoo Min, Nokyung Park, Siwon Kim, Seunghyun Park, and Jinkyu Kim,in ECCV 2022
Inter-domain curriculum learning for domain generalization
Daehee Kim, Jinkyu Kim, and Jaekoo Lee*,in ICT Express 2022
Occupancy flow fields for motion forecasting in autonomous driving
Reza Mahjourian*, Jinkyu Kim*, Yuning Chai, Mingxing Tan, Ben Sapp, and Dragomir Anguelov,in IEEE R&A Letters 2022
Ora3d: Overlap region aware multi-view 3d object detection
Wonseok Roh, Gyusam Chang, Seokha Moon, Giljoo Nam, Chanyoung Kim, Younghyun Kim, Jinkyu Kim†, and Sangpil Kim†,in BMVC 2022
Perception Prioritized Training of Diffusion Models
Jooyoung Choi, Jungbeom Lee, Chaehun Shin, Sungwon Kim, Hyunwoo Kim, and Sungroh Yoon,in CVPR 2022
Sound-guided semantic image manipulation
Seung Hyun Lee, Wonseok Roh, Wonmin Byeon, Sang Ho Yoon, Chan Young Kim, Jinkyu Kim†, and Sangpil Kim†,in CVPR 2022
Sound-guided semantic video generation
Seung Hyun Lee, Gyeongrok Oh, Wonmin Byeon, Chanyoung Kim, Won Jeong Ryoo, Sang Ho Yoon, Hyunjun Cho, Jihyun Bae, Jinkyu Kim†, and Sangpil Kim†,in ECCV 2022
Stopnet: Scalable trajectory and occupancy prediction for urban autonomous driving
Jinkyu Kim*, Reza Mahjourian*, Scott Ettinger, Mayank Bansal, Brandyn White, Ben Sapp, and Dragomir Anguelov,in ICRA 2022
Weakly Supervised Semantic Segmentation using Out-of-Distribution Data
Jungbeom Lee, Seong Joon Oh, Sangdoo Yun, Junsuk Choe, Eunji Kim, and Sungroh Yoon,in CVPR 2022
Zero-shot visual commonsense immorality prediction
Yujin Jeong, Seongbeom Park, Suhong Moon, and Jinkyu Kim,in BMVC 2022
A Scenario-Based Platform for Testing Autonomous Vehicle Behavior Prediction Models in Simulation
Francis Indaheng, Edward Kim, Kesav Viswanadha, Jay Shenoy, Jinkyu Kim, Daniel J. Fremont, and Sanjit A. Seshia,in NeurIPS 2021 (Workshop)
Anti-Adversarially Manipulated Attributions for Weakly and Semi-Supervised Semantic Segmentation
Jungbeom Lee, Eunji Kim, and Sungroh Yoon,in CVPR 2021
Audio-Guided Image Manipulation for Artistic Paintings
Seung Hyun Lee, Nahyuk Lee, Chanyoung Kim, Wonjeong Ryoo, Jinkyu Kim, Sang Ho Yoon†, and Sangpil Kim†,in NeurIPS 2021 (Workshop)
BBAM: Bounding Box Attribution Map for Weakly Supervised Semantic and Instance Segmentation
Jungbeom Lee, Jihun Yi, Chaehun Shin, and Sungroh Yoon,in CVPR 2021
Reducing Information Bottleneck for Weakly Supervised Semantic Segmentation
Jungbeom Lee, Jooyoung Choi, Jisoo Mok, and Sungroh Yoon,in NeurIPS 2021
Selfreg: Self-supervised contrastive regularization for domain generalization
Daehee Kim, Seunghyun Park, Jinkyu Kim, and Jaekoo Lee,in ICCV 2021
Toward Spatially Unbiased Generative Models
Jooyoung Choi, Jungbeom Lee, Yonghyun Jeong, and Sungroh Yoon,in ICCV 2021
Toward explainable and advisable model for self‐driving cars
Jinkyu Kim, Anna Rohrbach, Zeynep Akata, Suhong Moon, Teruhisa Misu, Yi-Ting Chen, Trevor Darrell, and John Canny,in Applied AI Letters 2021
Towards an Interpretable Deep Driving Network by Attentional Bottleneck
Jinkyu Kim and Mayank Bansal,in IEEE R&A Letters 2021
Advisable learning for self-driving vehicles by internalizing observation-to-action rules
Jinkyu Kim, Suhong Moon, Anna Rohrbach, Trevor Darrell, and John Canny,in CVPR 2020
Attentional bottleneck: Towards an interpretable deep driving network
Jinkyu Kim and Mayank Bansal,in CVPR 2020
Periphery-fovea multi-resolution driving model guided by human attention
Ye Xia, Jinkyu Kim, John Canny, Karl Zipser, and David Whitney,in WACV 2020
FickleNet: Weakly and Semi-supervised Semantic Image Segmentation using Stochastic Inference
Jungbeom Lee, Eunji Kim, Sungmin Lee, Jangho Lee, and Sungroh Yoon,in CVPR 2019
Frame-to-Frame Aggregation of Active Regions in Web Videos for Weakly Supervised Semantic Segmentation
Jungbeom Lee, Eunji Kim, Sungmin Lee, Jangho Lee, and Sungroh Yoon,in ICCV 2019
Grounding human-to-vehicle advice for self-driving vehicles
Jinkyu Kim, Teruhisa Misu, Yi-Ting Chen, Ashish Tawari, and John Canny,in CVPR 2019
Msnet: Mutual suppression network for disentangled video representations
Jungbeom Lee, Jangho Lee, Sungmin Lee, and Sungroh Yoon,in BMVC 2019 (Oral)
Deep traffic light detection for self-driving cars from a large-scale dataset
Jinkyu Kim, Jiyoon Lee, Jaekoo Lee, and John Canny,in ITSC 2018
Explainable deep driving by visualizing causal attention
Jinkyu Kim and John Canny,in Explainable and Interpretable Models in CV&ML 2018
Predicting driver attention in critical situations
Ye Xia, Danqing Zhang, Jinkyu Kim, Ken Nakayama, Karl Zipser, and David Whitney,in ACCV 2018
Textual explanations for self-driving vehicles
Jinkyu Kim, Anna Rohrbach, Trevor Darrell, John Canny, and Zeynep Akata,in ECCV 2018
Interpretable learning for self-driving cars by visualizing causal attention
Jinkyu Kim and John Canny,in ICCV 2017