Hoda Heidari

Assistant Professor Carnegie Mellon University

  • Pittsburgh PA

Hoda Heidari is broadly interested in the Ethical and Societal Aspects of Artificial Intelligence and Machine Learning.

Contact

Carnegie Mellon University

View more experts managed by Carnegie Mellon University

Biography

Hoda Heidari is the K&L Gates Career Development Assistant Professor in Ethics and Computational Technologies at Carnegie Mellon University with joint appointments in the Machine Learning Department and the Institute for Software, Systems, and Society. She is also affiliated with the Human-Computer Interaction Institute, CyLab, and the Block Center for Technology and Society at CMU, and she co-leads the university-wide Responsible AI Initiative.

Hoda is broadly interested in the Ethical and Societal Aspects of Artificial Intelligence and Machine Learning. In particular, her research has addressed issues of Fairness and Accountability.

Hoda's work has been generously supported by the NSF Program on Fairness in AI in Collaboration with Amazon, PwC, CyLab, Meta, and J. P. Morgan. Hoda is a senior personnel at AI-SDM: the NSF AI Institute for Societal Decision Making.

Areas of Expertise

Elections
Fairness and Accountability
Artificial Intelligence
Ethics
Machine Learning
Algorithmic Economics

Media Appearances

CMU Researchers Win NSF-Amazon Fairness in AI Awards

Carnegie Mellon University News  online

2021-02-16

Fair AI in Public Policy — Achieving Fair Societal Outcomes in ML Applications to Education, Criminal Justice, and Health & Human Services. Led by Hoda Heidari, an assistant professor in the Machine Learning Department (MLD) and Institute for Software Research, researchers in MLD and the Heinz College of Information Systems and Public Policy will help translate fairness goals in public policy into computationally tractable measures. They will focus on factors along the development life cycle, from data collection through evaluation of tools, to identify sources of unfair outcomes in systems related to education, child welfare and justice.

View More

CMU Launches Responsible AI Initiative To Direct Technology Toward Social Responsibility

Carnegie Mellon University News  online

2022-04-01

Housed at the Block Center for Technology and Society, the Responsible AI Initiative is spearheaded by faculty in the School of Computer Science (SCS) and the Heinz College of Information Systems and Public Policy. The initiative's leaders include Jodi Forlizzi, the Herbert A. Simon Professor in Computer Science and Human-Computer Interaction and the associate dean for diversity, equity and inclusion in SCS; Rayid Ghani, a professor in the Machine Learning Department (MLD) and the Heinz College; and Hoda Heidari, an assistant professor in MLD and the Institute for Software Research.

View More

Responsible AI Initiative launches at Carnegie Mellon University following panel discussion including government, industry leaders

PittsburghInno  online

2022-04-04

As artificial intelligence systems become more prevalent throughout all ways of life, Carnegie Mellon University wants to be at the forefront of ensuring that such AI technologies are being deployed in an ethical manner, limiting the potential for types of negligence and prejudice that have come to exist from the adoption of some automated systems.

View More

Show All +

Social

Industry Expertise

Research
Education/Learning
Computer Software

Accomplishments

Exemplary Track Award

2021

The ACM Conference on Economics and Computation (EC)

Best Paper Award

2021

ACM Conference on Fairness, Accountability, and Transparency (FAccT)

J. P. Morgan and Chase Individual Faculty Award

2021

Show All +

Education

Sharif University of Technology

B.Sc.

Computer Engineering

2011

Wharton School of Business

M.Sc.

Statistics

2017

University of Pennsylvania

Ph.D.

Computer and Information Science

2017

Event Appearances

Foundations of Algorithmic Fairness

ELLIS  

Roundtable on Data Privacy in Black Communities

Joint Center for Political and Economic Studies  

On human-AI collaboration

IDEAS Summer Program  

Research Grants

On the Impact of Algorithmic Fairness Metrics and Methods on Trust in Machine Learning Systems

CMU CyLab grant $50000

2021

Fair AI in Public Policy: Achieving Fair Societal Outcomes in ML Applications to Education, Criminal Justice, and Health and Human Services

NSF FAI grant $600000

2021

Robust and Fair AI Systems in Dynamic Environments

PwC Research Grant $300000

2022

Articles

Strategic Instrumental Variable Regression: Recovering Causal Relationships From Strategic Responses

Proceedings of the 39th International Conference on Machine Learning

2022

In settings where Machine Learning (ML) algorithms automate or inform consequential decisions about people, individual decision subjects are often incentivized to strategically modify their observable attributes to receive more favorable predictions. As a result, the distribution the assessment rule is trained on may differ from the one it operates on in deployment.

View more

A Validity Perspective on Evaluating the Justified Use of Data-driven Decision-making Algorithms

2023 IEEE Conference on Secure and Trustworthy Machine Learning (SaTML)

2023

Recent research increasingly brings to question the appropriateness of using predictive tools in complex, real-world tasks. While a growing body of work has explored ways to improve value alignment in these tools, comparatively less work has centered concerns around the fundamental justifiability of using these tools. This work seeks to center validity considerations in deliberations around whether and how to build data-driven algorithms in high-stakes domains.

View more

Local Justice and Machine Learning: Modeling and Inferring Dynamic Ethical Preferences toward Allocations

Proceedings of the AAAI Conference on Artificial Intelligence

2023

We consider a setting in which a social planner has to make a sequence of decisions to allocate scarce resources in a high-stakes domain. Our goal is to understand stakeholders' dynamic moral preferences toward such allocational policies. In particular, we evaluate the sensitivity of moral preferences to the history of allocations and their perceived future impact on various socially salient groups.

View more

Show All +