In this talk, I will introduce work on fundamental techniques for building and deploying effective natural language processing (NLP) systems that are also efficient and reliable. Specifically, I will address three interconnected challenges for modern machine learning in NLP: how to quickly adapt foundation models to new tasks with limited data, how to dynamically reconfigure large architectures for more efficient computation, and how to develop powerful theoretical tools for rigorous, yet practical, uncertainty quantification. To conclude, I will highlight a number of my future research directions, as well as extensions to interesting applications beyond natural language.
Bio: Adam Fisch is a PhD candidate at MIT working with Regina Barzilay and Tommi Jaakkola, and a recipient of an NSF Graduate Research Fellowship. His research centers around principled methods for efficient and reliable machine learning systems that work effectively in realistic scenarios, and has appeared in top-tier venues such as *ACL, ICLR, ICML, and NeurIPS. Adam also served as a co-instructor for the tutorial on Uncertainty Estimation for NLP at COLING 2022, and as a co-organizer of the Machine Reading for Question Answering workshops at EMNLP 2019 and 2021. Prior to MIT, Adam was a research engineer at Meta (Facebook) AI Research for two years, and studied mechanical engineering as an undergraduate at Princeton University.
To request accommodations for a disability please contact Emily Lawrence, emilyl@cs.princeton.edu, at least one week prior to the event.