
Tom Mitchell
Founders University Professor Carnegie Mellon University
- Pittsburgh PA
Tom Mitchell is a longtime researcher in AI, including new systems that can improve healthcare, education, climate and more.
Biography
Areas of Expertise
Media Appearances
How AI is infiltrating labor union contracts, in Pittsburgh and beyond
Technical.ly online
2025-03-26
As unions consider how they want to cover AI in contracts, Tom Mitchell (School of Computer Science) believes white-collar, college-educated, knowledge-based professions might be the most vulnerable to disruptions from AI tools.
As artificial intelligence surges across daily life, so do concerns AI
TribLIVE.com online
2024-09-08
Tom Mitchell, who founded Carnegie Mellon’s Department of Machine Learning in 2006, gave a simple example: You could show a computer program photos of your mother and then photos of people who are not your mother. With the gained experience, the program would be able to identify the features that distinguish who is a positive example of your mother and who is not.
Why a YouTube Chat About Chess Got Flagged for Hate Speech
WIRED online
2021-03-01
“Fundamentally, language is still a very subtle thing,” says Tom Mitchell, a CMU professor who has previously worked with KhudaBukhsh. “These kinds of trained classifiers are not soon going to be 100 percent accurate.”
Social
Accomplishments
Best Dataset Paper Award at the Learning at Scale Conference
2024
Best Paper Award at the User Interface Software and Technology (UIST) Conference
2020
President’s Medal, Stevens Institute of Technology
2018
Alan Perlis Award for Imagination in Computer Science, Carnegie Mellon University
2018
10-Year Outstanding Research Contributions Award, Brain Informatics Conference
2017
Education
Stanford University
Ph.D.
Electrical Engineering
1979
Massachusetts Institute of Technology
S.B.
Electrical Engineering
1973
Affiliations
- Generative AI Task Force : Chair
Articles
Spring: Studying papers and reasoning to play games
Advances in Neural Information Processing Systems2024
Open-world survival games pose significant challenges for AI algorithms due to their multi-tasking, deep exploration, and goal prioritization requirements. Despite reinforcement learning (RL) being popular for solving games, its high sample complexity limits its effectiveness in complex open-world games like Crafter or Minecraft. We propose a novel approach, SPRING, to read Crafter's original academic paper and use the knowledge learned to reason and play the game through a large language model (LLM). Prompted with the LaTeX source as game context and a description of the agent's current observation, our SPRING framework employs a directed acyclic graph (DAG) with game-related questions as nodes and dependencies as edges. We identify the optimal action to take in the environment by traversing the DAG and calculating LLM responses for each node in topological order, with the LLM's answer to final node directly translating to environment actions.
Read and reap the rewards: Learning to play atari with the help of instruction manuals
Advances in Neural Information Processing Systems2024
High sample complexity has long been a challenge for RL. On the other hand, humans learn to perform tasks not only from interaction or demonstrations, but also by reading unstructured text documents, eg, instruction manuals. Instruction manuals and wiki pages are among the most abundant data that could inform agents of valuable features and policies or task-specific environmental dynamics and reward structures. Therefore, we hypothesize that the ability to utilize human-written instruction manuals to assist learning policies for specific tasks should lead to a more efficient and better-performing agent. We propose the Read and Reward framework. Read and Reward speeds up RL algorithms on Atari games by reading manuals released by the Atari game developers. Our framework consists of a QA Extraction module that extracts and summarizes relevant information from the manual and a Reasoning module that evaluates object-agent interactions based on information from the manual.
Combining computational controls with natural text reveals aspects of meaning composition
Nature Computational Science2022
To study a core component of human intelligence—our ability to combine the meaning of words—neuroscientists have looked to linguistics. However, linguistic theories are insufficient to account for all brain responses reflecting linguistic composition. In contrast, we adopt a data-driven approach to study the composed meaning of words beyond their individual meaning, which we term ‘supra-word meaning’. We construct a computational representation for supra-word meaning and study its brain basis through brain recordings from two complementary imaging modalities. Using functional magnetic resonance imaging, we reveal that hubs that are thought to process lexical meaning also maintain supra-word meaning, suggesting a common substrate for lexical and combinatorial semantics.