research
* denotes shared first authorship
2024
- arXiv
LLM training & eval What is Wrong with Perplexity for Long-context Language Modeling?arXiv preprint arXiv:2410.23771, 2024We proposed a long-context perplexity measure that emphasizes long-context relevant tokens at training and evaluation, improving benchmark scores on LongBench, LongEval, and RULER by up to 22%. - NeurIPS
Best Paper Award
at ICML-W’24A Theoretical Understanding of Self-Correction through In-context AlignmentIn NeurIPS, 2024Best Paper Award at ICML 2024 ICL Workshop
We proposed the first theoretical explanation of how LLM self-correction works (as in OpenAI o1) and showed its effectiveness against social bias and jailbreak attacks. - NeurIPS
Oral at NeurIPS-W’24 In-Context Symmetries: Self-Supervised Learning through Contextual World ModelsIn NeurIPS, 2024Oral Presentation (top 4) at NeurIPS 2024 SSL Workshop
We introduced unsupervised test-time adaptation ability to self-supervised learning through a contextual world model designed for joint embedding (JEPA) models. - NeurIPS
- NeurIPS WorkshopReasoning in Reasoning: A Hierarchical Framework for Better and Faster Neural Theorem ProvingIn NeurIPS 2024 Workshop on Mathematical Reasoning and AI, 2024
- NeurIPS WorkshopThe Multi-faceted Monosemanticity in Multimodal RepresentationsIn NeurIPS 2024 Workshop on Responsibly Building the Next Generation of Multimodal Foundational Models, 2024
- ICML WorkshopRethinking Invariance in In-context LearningIn ICML Workshop on Theoretical Foundations of Foundation Models (TF2M), 2024
2023
- arXiv
Featured by Anthropic Jailbreak and guard aligned language models with only few in-context demonstrationsarXiv preprint arXiv:2310.06387, 2023Cited over 140 times. Featured and scaled up in Anthropic’s research blog, where it successfully demonstrated jailbreaking prominent LLMs, including GPT and Claude. - ICML
- TIPEquilibrium Image Denoising with Implicit DifferentiationIEEE Transactions on Image Processing (IEEE TIP), 2023
- ICLRUnbiased Stochastic Proximal Solver for Graph Neural Networks with Equilibrium StatesIn ICLR, 2023
- AAAI
Oral On the Connection between Invariant Learning and Adversarial Training for Out-of-Distribution GeneralizationIn AAAI, 2023
2022
- NeurIPS Workshop
Oral - ICML
- ICLRChaos is a Ladder: A New Theoretical Understanding of Contrastive Learning via Augmentation OverlapIn ICLR, 2022Cited over 120 times. We derived tight generalization bounds for contrastive learning with a new realistic theoretical framework. It derived unsupervised evaluation metrics with 97% correlation to downstream performance.
- ICLR
Silver Best Paper
at ICML-W’21A Unified Contrastive Energy-based Model for Understanding the Generative Ability of Adversarial TrainingIn ICLR, 2022Silver Best Paper Award at ICML 2021 AdvML workshop
From an energy-based perspective, we formulated contrastive learning as a generative model, and established the connection between adversarial training and maximum likelihood, thus briding generative and discriminative models together.
2021
- ECML-PKDD
Best ML Paper Award Reparameterized Sampling for Generative Adversarial NetworksIn ECML-PKDD, 2021Best ML Paper Award (1/685), invited to Machine Learning
We explored using GAN discriminator (as a good reward model) to bootstrap sample quality through an efficient MCMC algorithm, which not only guarantees theoretical convergence but also improves sample efficiency and quality in practice.