One of the most remarkable abilities of recent systems created with machine learning is their flexibility. Large pre-trained models (so-called “foundation models”) can be adapted with relative ease to a wide range of tasks. Sometimes this adaptation can happen with no training examples at all. How is this ability achieved? What are its limits? And what should we do when it fails? This seminar course will survey recent research on these topics, including pre-training and transfer learning, instruction tuning, reinforcement learning from human and AI feedback, few-shot and zero-shot learning, weak supervision, and synthetic data generation. Students will lead discussions on recent research papers and develop final research projects.
For questions, discussion, and other course-related posts, use Canvas.
If you have an atypical question that you are certain does not belong on Canvas, email the instructor.
Suplemental reading:
- On the Opportunities and Risks of Foundation Models. Rishi Bommasani, Drew A. Hudson, Ehsan Adeli, Russ Altman, Simran Arora, Sydney von Arx, Michael S. Bernstein, Jeannette Bohg, Antoine Bosselut, Emma Brunskill, Erik Brynjolfsson, Shyamal Buch, Dallas Card, Rodrigo Castellon, Niladri Chatterji, Annie Chen, Kathleen Creel, Jared Quincy Davis, Dora Demszky, Chris Donahue, Moussa Doumbouya, Esin Durmus, Stefano Ermon, John Etchemendy, Kawin Ethayarajh, Li Fei-Fei, Chelsea Finn, Trevor Gale, Lauren Gillespie, Karan Goel, Noah Goodman, Shelby Grossman, Neel Guha, Tatsunori Hashimoto, Peter Henderson, John Hewitt, Daniel E. Ho, Jenny Hong, Kyle Hsu, Jing Huang, Thomas Icard, Saahil Jain, Dan Jurafsky, Pratyusha Kalluri, Siddharth Karamcheti, Geoff Keeling, Fereshte Khani, Omar Khattab, Pang Wei Koh, Mark Krass, Ranjay Krishna, Rohith Kuditipudi, Ananya Kumar, Faisal Ladhak, Mina Lee, Tony Lee, Jure Leskovec, Isabelle Levent, Xiang Lisa Li, Xuechen Li, Tengyu Ma, Ali Malik, Christopher D. Manning, Suvir Mirchandani, Eric Mitchell, Zanele Munyikwa, Suraj Nair, Avanika Narayan, Deepak Narayanan, Ben Newman, Allen Nie, Juan Carlos Niebles, Hamed Nilforoshan, Julian Nyarko, Giray Ogut, Laurel Orr, Isabel Papadimitriou, Joon Sung Park, Chris Piech, Eva Portelance, Christopher Potts, Aditi Raghunathan, Rob Reich, Hongyu Ren, Frieda Rong, Yusuf Roohani, Camilo Ruiz, Jack Ryan, Christopher Ré, Dorsa Sadigh, Shiori Sagawa, Keshav Santhanam, Andy Shih, Krishnan Srinivasan, Alex Tamkin, Rohan Taori, Armin W. Thomas, Florian Tramèr, Rose E. Wang, William Wang , Bohan Wu, Jiajun Wu, Yuhuai Wu, Sang Michael Xie, Michihiro Yasunaga, Jiaxuan You, Matei Zaharia, Michael Zhang, Tianyi Zhang, Xikun Zhang, Yuhui Zhang, Lucia Zheng, Kaitlyn Zhou, and Percy Liang. [PDF]
- How to Read a CS Research Paper? Philip W. L. Wong. [PDF]
- How to Read a Technical Paper. Jason Eisner. [Online]
- How to Read a Paper. S. Keshav. [PDF]
[PDF] [Reviews]
Suplemental reading:
[PDF] [Code] [Data] [Reviews]
Suplemental reading:
- Benchmarking Generalization via In-Context Instructions on 1,600+ Language Tasks. Yizhong Wang, Swaroop Mishra, Pegah Alipoormolabashi, Yeganeh Kordi, Amirreza Mirzaei, Anjana Arunkumar, Arjun Ashok, Arut Selvan Dhanasekaran, Atharva Naik, David Stap, Eshaan Pathak, Giannis Karamanolakis, Haizhi Gary Lai, Ishan Purohit, Ishani Mondal, Jacob Anderson, Kirby Kuznia, Krima Doshi, Maitreya Patel, Kuntal Kumar Pal, Mehrad Moradshahi, Mihir Parmar, Mirali Purohit, Neeraj Varshney, Phani Rohitha Kaza, Pulkit Verma, Ravsehaj Singh Puri, Rushang Karia, Shailaja Keyur Sampat, Savan Doshi, Siddhartha Mishra, Sujan Reddy, Sumanta Patro, Tanay Dixit, Xudong Shen, Chitta Baral, Yejin Choi, Noah A. Smith, Hannaneh Hajishirzi, and Daniel Khashabi. ArXiv 2204.07705 2022. [PDF] [Data]
[PDF] [Supplemental] [Video]
Suplemental reading:
[PDF]
Supplemental reading:
- Low-Resource Languages Jailbreak GPT-4. Zheng-Xin Yong, Cristina Menghini, and Stephen H. Bach. NeurIPS Workshop on Socially Responsible Language Modelling Research (SoLaR) 2023. [PDF]
[PDF]
Supplemental reading:
[PDF]
Supplemental reading:
[PDF] [Blog]
Supplemental reading:
[PDF]
Supplemental reading:
- Flamingo: a Visual Language Model for Few-Shot Learning. Jean-Baptiste Alayrac, Jeff Donahue, Pauline Luc, Antoine Miech, Iain Barr, Yana Hasson, Karel Lenc, Arthur Mensch, Katie Millican, Malcolm Reynolds, Roman Ring, Eliza Rutherford, Serkan Cabi, Tengda Han, Zhitao Gong, Sina Samangooei, Marianne Monteiro, Jacob Menick, Sebastian Borgeaud, Andrew Brock, Aida Nematzadeh, Sahand Sharifzadeh, Mikolaj Binkowski, Ricardo Barreira, Oriol Vinyals, Andrew Zisserman, and Karen Simonyan. Neural Information Processing Systems (NeurIPS) 2022. [PDF] [Reviews] [Blog]
InstructBLIP: Towards General-purpose Vision-Language Models with Instruction Tuning. Wenliang Dai, Junnan Li, Dongxu Li, Anthony Meng Huat Tiong, Junqi Zhao, Weisheng Wang, Boyang Li, Pascale Fung, and Steven Hoi. Neural Information Processing Systems (NeurIPS) 2023.
[PDF] [Reviews] [Code]
Supplemental reading:
[PDF] [Project]
Supplemental reading:
- Cambrian-1: A Fully Open, Vision-Centric Exploration of Multimodal LLMs. Shengbang Tong, Ellis Brown, Penghao Wu, Sanghyun Woo, Manoj Middepogu, Sai Charitha Akula, Jihan Yang, Shusheng Yang, Adithya Iyer, Xichen Pan, Austin Wang, Rob Fergus, Yann LeCun, and Saining Xie. Neural Information Processing Systems (NeurIPS) 2024. [PDF] [Reviews] [Project]
[PDF] [Reviews]
Supplemental reading:
- LAION-5B: An open large-scale dataset for training next generation image-text models. Christoph Schuhmann, Romain Beaumont, Richard Vencu, Cade Gordon, Ross Wightman, Mehdi Cherti, Theo Coombes, Aarush Katta, Clayton Mullis, Mitchell Wortsman, Patrick Schramowski, Srivatsa Kundurthy, Katherine Crowson, Ludwig Schmidt, Robert Kaczmarczyk, and Jenia Jitsev. Neural Information Processing Systems (NeurIPS) Datasets and Benchmarks 2022. [PDF] [Reviews] [Project]
[PDF] [Project]
Supplemental reading:
- TextGrad: Automatic "Differentiation" via Text. Mert Yuksekgonul, Federico Bianchi, Joseph Boen, Sheng Liu, Zhi Huang, Carlos Guestrin, and James Zou. Nature 2025. [PDF (Preprint)] [Code]
[PDF] [Reviews] [Code]
Supplemental reading:
- LLaMA-Adapter V2: Parameter-Efficient Visual Instruction Model. Peng Gao, Jiaming Han, Renrui Zhang, Ziyi Lin, Shijie Geng, Aojun Zhou, Wei Zhang, Pan Lu, Conghui He, Xiangyu Yue, Hongsheng Li, and Yu Qiao. ArXiv 2304.15010 2023. [PDF] [Code (Same as V1)]
[PDF] [Reviews]
Supplemental reading:
- Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model? Yang Yue, Zhiqi Chen, Rui Lu, Andrew Zhao, Zhaokai Wang, Yang Yue, Shiji Song, and Gao Huang. ArXiv 2504.13837 2025. [PDF]
Does Math Reasoning Improve General LLM Capabilities? Understanding Transferability of LLM Reasoning. Maggie Huan, Yuetai Li, Tuney Zheng, Xiaoyu Xu, Seungone Kim, Minxin Du, Radha Poovendran, Graham Neubig, and Xiang Yue. ArXiv 2507.00432 2025.
[PDF]
Supplemental reading:
[PDF] [Code]
Supplemental reading:
- Super-NaturalInstructions: Generalization via Declarative Instructions on 1600+ NLP Tasks. Yizhong Wang, Swaroop Mishra, Pegah Alipoormolabashi, Yeganeh Kordi, Amirreza Mirzaei, Anjana Arunkumar, Arjun Ashok, Arut Selvan Dhanasekaran, Atharva Naik, David Stap, Eshaan Pathak, Giannis Karamanolakis, Haizhi Gary Lai, Ishan Purohit, Ishani Mondal, Jacob Anderson, Kirby Kuznia, Krima Doshi, Maitreya Patel, Kuntal Kumar Pal, Mehrad Moradshahi, Mihir Parmar, Mirali Purohit, Neeraj Varshney, Phani Rohitha Kaza, Pulkit Verma, Ravsehaj Singh Puri, Rushang Karia, Shailaja Keyur Sampat, Savan Doshi, Siddhartha Mishra, Sujan Reddy, Sumanta Patro, Tanay Dixit, Xudong Shen, Chitta Baral, Yejin Choi, Noah A. Smith, Hannaneh Hajishirzi, and Daniel Khashabi. Conference on Empirical Methods in Natural Language Processing (EMNLP) 2022. [PDF] [Code]
(No class)
[PDF] [Code]
Supplemental reading:
- VideoPoet: A Large Language Model for Zero-Shot Video Generation. Dan Kondratyuk, Lijun Yu, Xiuye Gu, José Lezama, Jonathan Huang, Grant Schindler, Rachel Hornung, Vighnesh Birodkar, Jimmy Yan, Ming-Chang Chiu, Krishna Somandepalli, Hassan Akbari, Yair Alon, Yong Cheng, Josh Dillon, Agrim Gupta, Meera Hahn, Anja Hauth, David Hendon, Alonso Martinez, David Minnen, Mikhail Sirotenko, Kihyuk Sohn, Xuan Yang, Hartwig Adam, Ming-Hsuan Yang, Irfan Essa, Huisheng Wang, David A. Ross, Bryan Seybold, and Lu Jiang. International Conference on Machine Learning (ICML) 2024. [PDF] [Project] [Blog]
(No class)
Students who complete this course will:
- Acquire a working knowledge of the landscape of research on machine learning with limited labeled data.
- Practice identifying and critically assessing the claims, contributions, and supporting evidence in machine learning research papers.
- Develop their ability to share scientific ideas via writing and discussion.
- Gain practical experience with the course's subject matter by applying and extending it to their own research interests though an open-ended project.
The following standards will be used to assign grades. Anyone who doesn't complete the standards to earn a B will receive NC.
- Participate actively in class discussions by asking questions, sharing opinions, and listening carefully to others.
- Meet all deadlines in the course schedule related to the research project.
- Submit two discussion questions to Canvas by 6 PM the evening before class for assigned readings, missing no more than 3 readings.
- Attend class meetings, missing no more than 3 meetings.
- Fulfill the requirements below to earn a B.
- Conduct an original research project related to course materials and submit a written report meeting the assignment guidelines.
- Lead the assigned class discussion demonstrating preparation and inclusion.
- Submit two discussion questions to Canvas by 6 PM the evening before class for assigned readings, missing no more than 6 readings.
- Attend class meetings, missing no more than 6 meetings.
| Activity | Hours |
| Class Meetings | 28 |
| Readings | 65 |
| Submitting Discussion Questions | 10 |
| Preparing to Lead Discussion | 2 |
| Project Research | 60+ |
| Project Proposal / Status | 10 |
| Project Final Report | 5 |
| Total | 180+ |
The Brown computer science department has made it its mission to create and sustain a diverse and inclusive environment in which all students, faculty, and staff can thrive. In this course, that responsibility falls on us all, students and teaching staff alike. In particular, Brown's Nondiscrimination and Anti-Harassment Policy applies to all participants.
If you have not been treated in an inclusive manner by any of the course members, please contact either me (Stephen) or the department chair (Prof. Tamassia). The Diversity and Inclusion Student Advocates are also available as a resource for members of underrepresented groups. Additional resources are listed on the department's website. We, the computer science department, take all complaints about discrimination, harassment, and other unprofessional behavior seriously.
In addition, Brown welcomes students from all around the country and the world, and their unique perspectives enrich our learning community. To empower students whose first language is not English, an array of support is available on campus, including language and culture workshops and individual appointments. For more information, contact the English Language Learning Specialists at ellwriting@brown.edu.
Academic dishonesty will not be tolerated. This includes cheating, lying about course matters, plagiarism, or helping others commit a violation. Plagiarism includes reproducing the words of others without both the use of quotation marks and citation. Students are reminded of the obligations and expectations associated with the Brown Academic Code and Brown Code of Student Conduct. For project work, feel free to build on third-party software, datasets, or other resources, as long as you credit them in your report(s) and clearly state what work is solely your own. As a general policy (for this course and for the rest of your academic career): if you use any idea, text, code, or data that you did not create, then cite it.
Brown University is committed to full inclusion of all students. Please inform me if you have a disability or other condition that might require accommodations or modification of any of these course procedures. You may email me, come to office hours, or speak with me after class, and your confidentiality is respected. I will do whatever I can to support accommodations recommended by SAS. For more information contact Student Accessibility Services (SAS) at 401-863-9588 or SAS@brown.edu.
Being a student can be very stressful. If you feel you are under too much pressure or there are psychological issues that are keeping you from performing well at Brown, I encourage you to contact Brown’s Counseling and Psychological Services CAPS. They provide confidential counseling and can provide notes supporting accommodations for health reasons.
I expect everyone to complete the course on time. However, I understand that there may be factors beyond your control, such as health problems and family crises, that prevent you from finishing the course on time. If you feel you cannot complete the course on time, please discuss with me (Stephen) the possibility of being given a grade of Incomplete for the course and setting a schedule for completing the course in the upcoming year.
Thanks to Tom Doeppner, Laura Dobler, and Daniel Ritchie for borrowed text.
CSCI 2952-C