Jongha Jon Ryu
Postdoctoral Associate at MIT

Room 36-677
50 Vassar St
Cambridge, MA 02139
I am currently a postdoctoral associate at MIT, hosted by Gregory W. Wornell. Prior to joining MIT, I received my Ph.D. in Electrical Engineering from UC San Diego, where I was fortunate to be advised by Young-Han Kim and Sanjoy Dasgupta. My graduate study was generously supported by the Kwanjeong Educational Foundation. Before the graduate study, I received my B.S. in Electrical and Computer Engineering and B.S. in Mathematical Sciences (with minor in Physics) with the highest distinction from Seoul National University in 2015.
In general, I aim to develop principled and practical algorithms for machine learning and data science. My recent research topics include:
- Scalable parametric spectral decomposition methods
- efficient parametric operator SVD (NeuralSVD [ICML2024a])
- New generative modeling techniques
- new training framework for one-step generative models (Score-of-Mixture Training [arXiv2025a])
- unifying principles for fitting unnormalized distributions [arXiv2024]
- Efficient & reliable uncertainty quantification techniques
- universal-gambling-based time-uniform confidence sets [TIT2024], [ICML2024b], and applications [arXiv2025b]
- identifying pitfalls of evidential deep learning [NeurIPS2024]
- Nonparametric methods
- unified view on density functional estimation with fixed-k-NNs [TIT2022]
- efficient small-k-nearest-neighbors algorithms [arXiv2022]
As an information theorist by training, I enjoy doing research by simplifying intricate ideas, unifying concepts, and generalizing them to address complex problems.
Check out my resume for more information.
news
Feb 19, 2025 | New paper alerts! We propose: (1) a new simple and stable training scheme for one-step generative models (paper1); (2) new techniques for efficient off-policy contextual bandits (paper2). |
---|---|
Oct 23, 2024 | In this fall, I have given talks on NeuralSVD at MERL, KAIST, KIAS, and Flatiron Institute. |
Sep 25, 2024 | Our paper on demystifying the sucess of evidential deep learning methods got accepted at NeurIPS 2024! |
Sep 06, 2024 | I have posted a substantially revised version of the arXiv preprint on minimax optimal learning with fixed-k-nearest neighbors, now including new results on density estimation. |
Aug 21, 2024 | One paper on gambling-based confidence sequences has been accepted at IEEE Transactions on Information Theory! |
selected publications
- ICMLOperator SVD with Neural Networks via Nested Low-Rank ApproximationIn Proc. Int. Conf. Mach. Learn. (ICML) , July 2024