hero image
Ameet Talwalkar - Carnegie Mellon University. Pittsburgh, PA, US

Ameet Talwalkar

Associate Professor, Machine Learning | Carnegie Mellon University

Pittsburgh, PA, UNITED STATES

Ameet Talwalkar's work is motivated by the goal of democratizing machine learning.

Biography

Ameet Talwalkar is an assistant professor in the Machine Learning Department at Carnegie Mellon University, and also co-founder and chief scientist at Determined AI. He led the initial development of the MLlib project in Apache Spark, is a co-author of the textbook Foundations of Machine Learning (MIT Press), and created an award-winning edX MOOC on distributed machine learning.

Areas of Expertise (5)

Machine Learning

Algorithms

Computational Biology

Bioinformatics

Data Mining

Media Appearances (4)

Pittsburgh’s AI-Powered Renaissance

CMU News  online

2024-10-09

"CMU has an unparalleled degree of expertise in AI among its faculty and students. In the context of human-centric AI, the fact that we have distinct departments in machine learning, human-computer interaction, and language technologies, coupled with a highly collaborative research environment, gives CMU and Pittsburgh a technical advantage. We have a burgeoning startup scene, in part based on academic spinouts, including two of the fastest growing AI startups in the world: Abridge AI and Skild AI."

view more

Researchers outline promises, challenges of understanding AI for biological discovery

Medical Xpress  online

2024-08-09

"Interpretable machine learning has generated significant excitement as machine learning and artificial intelligence tools are being applied to increasingly important problems," said Ameet Talwalkar, an associate professor in CMU's Machine Learning Department (MLD).

view more

Determined AI makes its machine learning infrastructure free and open source

TechCrunch  online

2020-04-29

“They’re using things like TensorFlow and PyTorch,” said Chief Scientist Ameet Talwalkar. “A lot of the way that work is done is just conventions: How do the models get trained? Where do I write down the data on which is best? How do I transform data to a good format? All these are bread and butter tasks. There’s tech to do it, but it’s really the Wild West. And the amount of work you have to do to get it set up… there’s a reason big tech companies build out these internal infrastructures.”

view more

AI in the 2020s Must Get Greener—and Here’s How The push for energy efficient “Green AI” requires new strategies | Opinion

IEEE Spectrum  online

2020-02-14

The environmental impact of artificial intelligence (AI) has been a hot topic as of late—and I believe it will be a defining issue for AI this decade. The conversation began with a recent study from the Allen Institute for AI that argued for the prioritization of “Green AI" efforts that focus on the energy efficiency of AI systems.

view more

Spotlight

Media

Publications:

Documents:

Photos:

loading image

Videos:

Automating Architecture Transfer feat. Ameet Talwalkar | Stanford MLSys Seminar Episode 19

Audio/Podcasts:

Social

Accomplishments (3)

Best Paper Award, Conference on Human Computation and Crowdsourcing (HCOMP) (professional)

2023

Okawa Foundation Research Grant (professional)

2018

Google Faculty Research Award (professional)

2015

Education (1)

New York University: Ph.D., Computer Science - Machine Learning 2010

Event Appearances (1)

Speaker: AI in Financial Services: Transforming the Sector for a Better World

AI Horizons Pittsburgh Summit  Pittsburgh, PA

2024-10-14

Articles (5)

The Impact of Element Ordering on LM Agent Performance

arXiv preprint

2024 There has been a surge of interest in language model agents that can navigate virtual environments such as the web or desktop. To navigate such environments, agents benefit from information on the various elements (e.g., buttons, text, or images) present. It remains unclear which element attributes have the greatest impact on agent performance, especially in environments that only provide a graphical representation (i.e., pixels). Here we find that the ordering in which elements are presented to the language model is surprisingly impactful--randomizing element ordering in a webpage degrades agent performance comparably to removing all visible text from an agent's state representation.

view more

Do llms exhibit human-like response biases? a case study in survey design

Transactions of the Association for Computational Linguistics

2024 One widely cited barrier to the adoption of LLMs as proxies for humans in subjective tasks is their sensitivity to prompt wording—but interestingly, humans also display sensitivities to instruction changes in the form of response biases. We investigate the extent to which LLMs reflect human response biases, if at all. We look to survey design, where human response biases caused by changes in the wordings of “prompts” have been extensively explored in social psychology literature. Drawing from these works, we design a dataset and framework to evaluate whether LLMs exhibit human-like response biases in survey questionnaires

view more

Applying interpretable machine learning in computational biology—pitfalls, recommendations and opportunities for new developments

Nature Methods

2024 Recent advances in machine learning have enabled the development of next-generation predictive models for complex computational biology problems, thereby spurring the use of interpretable machine learning (IML) to unveil biological insights. However, guidelines for using IML in computational biology are generally underdeveloped. We provide an overview of IML methods and evaluation techniques and discuss common pitfalls encountered when applying IML methods to computational biology problems. We also highlight open questions, especially in the era of large language models, and call for collaboration between IML and computational biology researchers.

view more

Revisiting Cascaded Ensembles for Efficient Inference

arXiv preprint

2024 A common approach to make machine learning inference more efficient is to use example-specific adaptive schemes, which route or select models for each example at inference time. In this work we study a simple scheme for adaptive inference. We build a cascade of ensembles (CoE), beginning with resource-efficient models and growing to larger, more expressive models, where ensemble agreement serves as a data-dependent routing criterion. This scheme is easy to incorporate into existing inference pipelines, requires no additional training, and can be used to place models across multiple resource tiers--for instance, serving efficient models at the edge and invoking larger models in the cloud only when necessary.

view more

Modulating Language Model Experiences through Frictions

arXiv preprint

2024 Language models are transforming the ways that their users engage with the world. Despite impressive capabilities, over-consumption of language model outputs risks propagating unchecked errors in the short-term and damaging human capabilities for critical thinking in the long-term, particularly in knowledge-based tasks. How can we develop scaffolding around language models to curate more appropriate use? We propose selective frictions for language model experiences, inspired by behavioral science interventions, to dampen misuse. Frictions involve small modifications to a user's experience, e.g., the addition of a button impeding model access and reminding a user of their expertise relative to the model.

view more