Emma Strubell

Assistant Professor, Language Technologies Institute Carnegie Mellon University

  • Pittsburgh PA

Emma Strubell's research focuses on efficient and equitable natural language processing (NLP).

Contact

Carnegie Mellon University

View more experts managed by Carnegie Mellon University

Biography

Emma Strubell is an Assistant Professor at Carnegie Mellon University's Language Technologies Institute (LTI). Their research lies at the intersection of natural language processing (NLP) and machine learning, and their broad research objective is bridging the gap between state-of-the-art NLP methods, and the wide variety of users who stand to benefit from that technology, but for whom that technology does not yet work in practice.

Their work has been recognized with a Madrona AI Impact Award, best paper awards at ACL and EMNLP, and in 2024 they were named one of the most powerful people in AI by Business Insider.

Areas of Expertise

Green AI
Natural Language Processing (NLP)
Machine Learning
Artificiial Intelligence

Media Appearances

Emma Strubell | AI Power List

Business Insider  online

2024-10-24

In front of a whiteboard from a classroom at Carnegie Mellon University, Strubell explains the Jevons effect, in which the gains from increased efficiency of a technological tool could be negated as use increases. The concept has come into focus as the conversation shifts to efficiency, AI, and the environment. Though they are a proponent for the advancement of AI, Strubell told Business Insider. "GenAI training is a nightmare for energy providers." Their work as an assistant professor at Carnegie Mellon's Language Technologies Institute asks students and researchers to examine the systems that power AI to discover more efficient and environmentally friendly raw materials that power AI.

View More

Greater, newer AI models come with environmental impacts

Marketplace  online

2024-06-13

Emma Strubell of Carnegie Mellon University explains why carbon emissions increase with more AI data centers and more powerful AI features.

View More

An AI's Carbon Footprint Is 5 Times Bigger Than a Car's

Popular Mechanics  online

2019-06-06

The act of training a neural network, according to the study led by Emma Strubell of the University of Massachusetts Amherst, creates a carbon dioxide footprint of 284 tonnes—five times the lifetime emissions of an average car.

View More

Social

Industry Expertise

Education/Learning

Education

University of Maine

B.S.

Computer Science

UMass Amherst

Ph.D.

Event Appearances

AI and the Environment: Sustaining the Common Good

2024 | Markkula Center for Applied Ethics and Next 10  Santa Clara University

Articles

Efficient and equitable natural language processing in the age of deep learning (dagstuhl seminar 22232)

Dagstuhl Reports

2023

This report documents the program and the outcomes of Dagstuhl Seminar 22232" Efficient and Equitable Natural Language Processing in the Age of Deep Learning". Since 2012, the field of artificial intelligence (AI) has reported remarkable progress on a broad range of capabilities including object recognition, game playing, speech recognition, and machine translation. Much of this progress has been achieved by increasingly large and computationally intensive deep learning models: training costs for state-of-the-art deep learning models have increased 300,000 times between 2012 and 2018.

View more

Efficient methods for natural language processing: A survey

Transactions of the Association for Computational Linguistics

2023

Recent work in natural language processing (NLP) has yielded appealing results from scaling model parameters and training data; however, using only scale to improve performance means that resource consumption also grows. Such resources include data, time, storage, or energy, all of which are naturally limited and unevenly distributed. This motivates research into efficient methods that require fewer resources to achieve similar results. This survey synthesizes and relates current methods and findings in efficient NLP. We aim to provide both guidance for conducting NLP under limited resources, and point towards promising research directions for developing more efficient methods.

View more

Making scalable meta learning practical

Advances in Neural Information Processing Systems

2023

Despite its flexibility to learn diverse inductive biases in machine learning programs, meta learning (ie,\learning to learn) has long been recognized to suffer from poor scalability due to its tremendous compute/memory costs, training instability, and a lack of efficient distributed training support. In this work, we focus on making scalable meta learning practical by introducing SAMA, which combines advances in both implicit differentiation algorithms and systems. Specifically, SAMA is designed to flexibly support a broad range of adaptive optimizers in the base level of meta learning programs, while reducing computational burden by avoiding explicit computation of second-order gradient information, and exploiting efficient distributed training techniques implemented for first-order gradients.

View more

Show All +