Arora Research Lab @ Princeton
Team / Publications / Blogs
Our research lab develops conceptual understanding of AI models, including training techniques, datasets, model interpretability and model evaluations. Recent works involve new ways of training and evaluating models using “skills”, which are themselves elicited from powerful AI models, and in turn can be used to generate very effective pipelines for improving AI models using synthetic training data. Many of our papers involve mathematical analysis as well as experiments.
[09/2024] | Congrats to Sadhika Malladi on being named 2025 Siebel Scholar! |
[09/2024] | New paper Instruct-SkillMix: A Powerful Pipeline for LLM Instruction Tuning was accepted to Compositional Learning Workshop at NeurIPS 2024 (Oral), FITML Workshop at NeurIPS 2024! |
[09/2024] | New paper Can Models Learn Skill Composition from Examples? was accepted to NeurIPS 2024! |
[09/2024] | New paper Keeping LLMs Aligned After Fine-tuning: The Crucial Role of Prompt Templates was accepted to NeurIPS 2024! |
[09/2024] | New paper AI-Assisted Generation of Difficult Math Questions was accepted to MATH-AI Workshop at NeurIPS 2024! |
[05/2024] | New paper Trainable transformer in transformer was accepted to ICML 2024! |
[05/2024] | New paper LESS: Selecting Influential Data for Targeted Instruction Tuning was accepted to ICML 2024! |
[01/2024] | Our work A Theory for Emergence of Complex Skills in Language Models was featured in Quanta Magazine! |
[01/2024] | New paper Skill-Mix: A flexible and expandable family of evaluations for AI models was accepted to ICLR 2024! |
LLM Metacognition: Understanding and Leveraging LLMs' "Thinking about Thinking" (Simons Institute, 2024)
Why do large language models display new and complex skills? (University of Michigan Computer Science and Engineering, 2023)
A Theory for Emergence of Complex Skills in Language Models (Simons Institute, 2023)