Jungeum Kim
Machine Learning for everyone
About me
I am an Assistant Professor in the Statistics Department at North Carolina State University Currently, I am interested in bridging statistical principles with
cutting-edge advancements in machine learning and AI. I was a Principal Researcher at the University of Chicago Booth School of Business, where I worked with Professor Veronika Rockova. I received my Ph.D. degree from Purdue University under the supervision of Dr. Xiao Wang, and I obtained my Master’s and Bachelor’s degrees from Seoul National University under the supervision of Dr. Hee-seok Oh.
Interests
- Artificial Intelligence for Bayesian Statistics
- Robust and Principled Deep Learning for Science
- Statistical Uncertainty Quantification for Artificial Intelligence
- Manifold Learning and Data Visualization
- AI Data Synthesis
- Nonparametric Bayes
Education
- PhD in Statistics, 2022, Purdue University
- MSc in Statistics, 2017, Seoul National University
- BSc in Statistics, 2015, Seoul National University
- BSc in Social Welfare, 2015, Seoul National University

NEWS
March 1 2026 JASA accepted our paper on the Tree Bandits for Generative Bayes (with Sean and Veronika)-!
May 21 2025 EJS accepted our paper on the mixing rate of Bayesian CART (with Veronika Rockova)-!
Jan 22 2025 AISTATS 2025 accepted our paper on the Generative Bayesian posterior sampler (with Percy and Veronika).
Dec 18 2024 TMLR accepted our paper on Inductive Global and Local Manifold Approximation and Projection (with Xiao Wang).
Oct 14 2024 (NEW PAPER) Check out our draft on the Generative Bayesian posterior sampler (with Percy and Veronika).
Aug 19 2024 (NEW PAPER) Check out our Uncertainty Quantification for Generative AI (with Sean and Veronika).
June 26 2024 (TALK) Invited to present at ISBA satellite 2024.
Apr 17 2024 (NEW PAPER) Check out our draft on the generative tree bandits (with Sean and Veronika).
Mar 8 2024 🏆Accepted to Rising Stars in Computational and Data Sciences at UT- Austin in April 2024!
Dec 17 2023 (TALK) Invited to present at CMStatistics 2023.
Dec 06 2023 (NEW PAPER) Check out our draft of deep learning for Bayesian inference (with Veronika Rockova).
Oct 04 2023 (TALK) Presented at the Statistics Department Colloquium at the University of Wisconsin–Madison.
Aug 10 2023 (TALK) Invited to present in a BART section at JSM 2023.
Jul 19 2023 (TALK) Invited to present at the International Statistical Institute (ISI) 2023 conference.
May 6 2023 (NEW PAPER) Check out our draft on the mixing rate of Bayesian CART (with Veronika Rockova).
Mar 31 2023 (TALK) Presented at the Statistics Department Colloquium at Auburn University.
RESEARCH

Tree Bandits for Generative Bayes
Sean O’Hagan, Jungeum Kim, and Veronika Rockova (2026), JASA
We develop a self-aware framework that can be simplified to a binary bandit problem. This framework treats ABC acceptance as a reward. Each arm is a box from sequential recursive partitioning classifiers on the ABC lookup table. This new bandit approach accelerates ABC rejection! (draft, slide)

Jungeum Kim, Percy Zhai, and Veronika Rockova (2024), AISTATS
The vector quantile notion can be used to obtain a generative Bayesian posterior sampler-! We have features like automatic summary statistics and support shrinkage contraction of our posterior approximation) as in the image. We utilize Monge-Kantorovich depth in multivariate quantiles to directly sample from Bayesian credible sets. (draft, slide, code)

Adaptive Uncertainty Quantification for Generative AI
Jungeum Kim*, Sean O’Hagan*, and Veronika Rockova (2024)
Our empirical results show that conformalizing ChatGPT predictions using our adaptive approach (1) yields tighter sets (compared to split-conformal) for most test observations, and (2) is able to communicate “I do not know” through wide sets in cases when the language model is likely to “hallucinate” its query response. Our paper is not meant to encourage users to rely on ChatGPT for predictions. It is only meant to illustrate that if one were to do so, proper accounting for uncertainty is needed. (draft, slide, code)


On Mixing Rates for Bayesian CART
Jungeum Kim and Veronika Rockova (2023), EJS
We found that Bayesian CART can mix slow or fast depending on the intrinsic data structure. Instead of the standard local moves, we propose to use new and more aggressive movements, called twiggy movements. The new movements remove such data structural dependency. (draft, slide, video)

Jungeum Kim and Xiao Wang (2023), TMLR
Manifold learning for visualization purposes by constructing and preserving locally adaptive global distances. Our algorithm shows a clear progression from global formation (with random initialization) to local details in a single optimization process!(draft, code)

Robust Sensible Adversarial Learning of Deep Neural Networks for Image Classification
Jungeum Kim and Xiao Wang (2022), AOAS
The Bayes classifier in fact can be the most robust classifier. Therefore, adversarial training for robust classification with deep neural networks should still aim to learn the Bayes classifier-!

Purdue University Graduate School PhD Thesis.
AWARDS
2024 Rising Stars in Computational and Data Sciences at UT- Austin
2022 I. W. Burr Award, 2022, Purdue Department of Statistics
2021 Virgil Anderson and Gloria Fischer Graduate Fellowship, Purdue
contact
SAS Hall, 2311 Stinson Dr, Raleigh, NC 27607