Research

My recent research interests include

  • Automated theorem proving with LLMs

  • Out-of-Distribution (OOD) generalization

  • Factor Models

Publications

  1. Haoyu Zhao, Yihan Geng, Shange Tang, Yong Lin, Bohan Lyu, Hongzhou Lin, Chi Jin, Sanjeev Arora. “Ineq-Comp: Benchmarking Human-Intuitive Compositional Reasoning in Automated Theorem Proving on Inequalities.”, The Thirty-Ninth Annual Conference on Neural Information Processing Systems (NeurIPS) 2025. [arXiv]

  2. Yong Lin*, Shange Tang*, Bohan Lyu, Jiayun Wu, Hongzhou Lin, Kaiyu Yang, Jia Li, Mengzhou Xia, Danqi Chen, Sanjeev Arora, Chi Jin. “Goedel-Prover: A Frontier Model for Open-Source Automated Theorem Proving.”, Conference on Language Modeling (COLM) 2025. [arXiv]

  3. Kaixuan Huang, Jiacheng Guo, Zihao Li, Xiang Ji, Jiawei Ge, Wenzhe Li, Yingqing Guo, Tianle Cai, Hui Yuan, Runzhe Wang, Yue Wu, Ming Yin, Shange Tang, Yangsibo Huang, Chi Jin, Xinyun Chen, Chiyuan Zhang, Mengdi Wang. “MATH-Perturb: Benchmarking LLMs’ Math Reasoning Abilities against Hard Perturbations.”, International Conference on Machine Learning (ICML) 2025. [arXiv]

  4. Shange Tang*, Jiayun Wu*, Jianqing Fan, Chi Jin. “Benign Overfitting in Out-of-Distribution Generalization of Linear Models.”, International Conference on Learning Representations (ICLR) 2025. [arXiv]

  5. Jiawei Ge*, Shange Tang*, Jianqing Fan, Cong Ma, Chi Jin. “Maximum Likelihood Estimation is All You Need for Well-Specified Covariate Shift.”, International Conference on Learning Representations (ICLR) 2024. [arXiv]

  6. Jiawei Ge*, Shange Tang*, Jianqing Fan, Chi Jin. “On the Provable Advantage of Unsupervised Pretraining.”, International Conference on Learning Representations (ICLR) 2024, spotlight. [arXiv]

Preprints

  1. Yong Lin, Shange Tang, Bohan Lyu, Ziran Yang, Jui-Hui Chung, Haoyu Zhao, Lai Jiang, Yihan Geng, Jiawei Ge, Jingruo Sun, Jiayun Wu, Jiri Gesi, Ximing Lu, David Acuna, Kaiyu Yang, Hongzhou Lin, Yejin Choi, Danqi Chen, Sanjeev Arora, Chi Jin. “Goedel-Prover-V2: Scaling Formal Theorem Proving with Scaffolded Data Synthesis and Self-Correction.”, arXiv preprint arXiv:2508.03613 (2025). [arXiv]

  2. Jiawei Ge, Amanda Wang, Shange Tang, Chi Jin. “Principled Out-of-Distribution Generalization via Simplicity.”, arXiv preprint arXiv:2505.22622 (2025). [arXiv]

  3. Shange Tang, Yuanhao Wang, Chi Jin. “Is Elo Rating Reliable? A Study Under Model Misspecification.”, arXiv preprint arXiv:2502.10985 (2025). [arXiv]

  4. Shange Tang, Soham Jana, Jianqing Fan. “Factor Adjusted Spectral Clustering for Mixture Models.”, arXiv preprint arXiv:2408.12564 (2024). [arXiv]

* denotes equal contribution.

Invited Talks

  1. “Goedel-Prover-V2: State-of-the-art performance in Automated Mathematical Theorem Proving”, Centaur AI Institute, Neural-Symbolic AI Summer School, Aug 2025. [Youtube]

  2. “Goedel-Prover: A Frontier Model for Open-Source Automated Theorem Proving”, TTIC Machine Learning Seminar, Mar 2025.

  3. “Benign Overfitting in Out-of-Distribution Generalization of Linear Models”, Simons Institute for the Theory of Computing, Domain Adaptation and Related Areas workshop poster session, Nov 2024.

  4. “Maximum Likelihood Estimation is All You Need for Well-Specified Covariate Shift”, UIUC Machine Learning Seminar, Mar 2024.