Research Topics
Advances in machine learning are reshaping the foundations of science and technology. Despite its successes and profound impact, a rigorous theoretical understanding of why machine learning algorithms succeed in some cases and fail in others remains limited. My research addresses this gap by developing theoretical foundations that explain when and why machine learning works and guide the design of more effective algorithms. My current research includes theoretical foundations of machine learning for:
Language Models
Representation Learning
Synthetic Data Generation
Ongoing Work
How to Correctly Report LLM-as-a-Judge Evaluations [arxiv] [github]
Chungpa Lee, Thomas Zeng, Jongwon Jeong, Jy-yong Sohn, Kangwook Lee
Poisson Regression with Additive Exponential Mean: Statistical Modeling and Insurance Applications (under review)
Chungpa Lee, Joseph H.T. Kim
A Theoretical Analysis of Zero-Shot and Few-Shot Inference in Transformers (work in progress)
Chungpa Lee, Jy-yong Sohn, Kangwook Lee
Publications
On the Similarities of Embeddings in Contrastive Learning [paper] [arxiv] [github]
Chungpa Lee, Sehee Lim, Kibok Lee, Jy-yong Sohn
In Proceedings of the 42nd International Conference on Machine Learning (ICML), 2025
A Generalized Theory of Mixup for Structure-Preserving Synthetic Data [paper] [arxiv] [github]
Chungpa Lee, Jongho Im, Joseph H.T. Kim
In Proceedings of the 28th International Conference on Artificial Intelligence and Statistics (AISTATS), 2025
A Theoretical Framework for Preventing Class Collapse in Supervised Contrastive Learning [paper] [arxiv] [github]
Chungpa Lee, Jeongheon Oh, Kibok Lee, Jy-yong Sohn
In Proceedings of the 28th International Conference on Artificial Intelligence and Statistics (AISTATS), 2025
Analysis of Using Sigmoid Loss for Contrastive Learning [paper] [arxiv] [github]
Chungpa Lee, Joonhwan Chang, Jy-yong Sohn
In Proceedings of the 27th International Conference on Artificial Intelligence and Statistics (AISTATS), 2024