Postdoc at Princeton University
email cs1095@princeton.eduI am a postdoctoral researcher in the Department of Computer Science at Princeton University, where I work with Brenden Lake. I completed my Ph.D. in Computer Science at the University of Wisconsin–Madison under the supervision of Frederic Sala. Prior to that, I studied psychology and computer science at Seoul National University.
My research centers on data-centric AI, focusing on methods for learning from imperfect supervision and improving the reliability of modern ML systems.
Learning from Weak Signals: Data-Centric Methods for Foundation Models
Ph.D. Dissertation, University of Wisconsin–Madison (2025).
[PDF]
[C7] Weak-to-Strong Generalization Through the Data-Centric Lens, ICLR 2025
Changho Shin, John Cooper, Frederic Sala
[C6] Personalize Your LLM: Fake it then Align it, NAACL 2025 Findings
Yijing Zhang, Dyah Adila, Changho Shin, Frederic Sala
[C5] OTTER: Improving Zero-Shot Classification via Optimal Transport, NeurIPS 2024
Changho Shin, Jitian Zhao, Sonia Cromp, Harit Vishwakarma, Frederic Sala
[C4] Zero-Shot Robustification of Zero-Shot Models, ICLR 2024
Also presented in NeurIPS 2023 R0-FoMo Workshop (Best Paper Award Honorable Mention)
Dyah Adila*, Changho Shin*, Linrong Cai, Frederic Sala
[C3] Mitigating Source Bias for Fairer Weak Supervision, NeurIPS 2023
Changho Shin, Sonia Cromp, Dyah Adila, Frederic Sala
[C2] Universalizing Weak Supervision, ICLR 2022
Changho Shin, Winfred Li, Harit Vishwakarma, Nicholas Roberts, Frederic Sala
[C1] Subtask Gated Networks for Non-Intrusive Load Monitoring, AAAI 2019
Changho Shin, Sunghwan Joo, Jaeryun Yim, Hyoseop Lee, Taesup Moon, and Wonjong Rhee
[J2] The ENERTALK dataset, 15 Hz electricity consumption data from 22 houses in
Korea, Scientific Data
Changho Shin, Eunjung Lee, Jeongyun Han, Jaeryun Yim, Hyoseop Lee, and Wonjong Rhee
[J1] Data Requirements for Applying Machine Learning to Energy Disaggregation, Energies
Changho Shin, Seungeun Rho, Hyoseop Lee, and Wonjong Rhee
[W8] Curriculum Learning as Transport: Training Along Wasserstein Geodesics, NeurIPS 2025 CCFM Workshop
Changho Shin, David Alvarez-Melis
[W7] From Many Voices to One: A Statistically Principled Aggregation of LLM Judges, NeurIPS 2025 LLM Evaluation Workshop; Reliable ML Workshop
Jitian Zhao*, Changho Shin*, Tzu-Heng Huang, Srinath Namburi, Frederic Sala
[W6] LLM-Integrated Bayesian State Space Models for Multimodal Time-Series Forecasting, NeurIPS 2025 BERT2S Workshop
Sungjun Cho, Changho Shin, Suenggwan Jo, Xinya Yan, Shourjo Aditya Chaudhuri, Frederic Sala
[W5] Is Free Self-Alignment Possible?, NeurIPS 2024 MINT Workshop
Dyah Adila, Changho Shin, Yijing Zhang, Frederic Sala
[W4] Foundation Models Can Robustify Themselves, For Free, NeurIPS 2023 R0-FoMo Workshop (Best Paper Award Honorable Mention)
Dyah Adila*, Changho Shin*, Linrong Cai, Frederic Sala
[W3] Pool-Search-Demonstrate: Improving Data-wrangling LLMs via better in-context examples, NeurIPS 2023 TRL Workshop (Oral)
Changho Shin*, Joon Suk Huh*, Elina Choi
[W2] Multimodal Data Curation via Object Detection and Filter Ensembles, ICCV 2023 DataComp Workshop (Filtering Track Rank #1 (Small))
Changho Shin*, Tzu-heng Huang*, Sui Jiet Tay, Dyah Adila, Frederic Sala
[W1] Can we get smarter than majority vote? Efficient use of individual rater’s labels for content moderation, NeurIPS 2022 ENLSP Workshop
Changho Shin, Alice Schoenauer Sebag