M. Ehsan Hoque

Asaro Biggar Family Fellow in Data Science, Assistant Professor of Computer Science, and faculty member in the Goergen Institute for Data Science and Artificial Intelligence University of Rochester

  • Rochester NY

M. Ehsan Hoque is designing and implementing new algorithms to sense subtle human nonverbal behavior

Contact

University of Rochester

View more experts managed by University of Rochester

Areas of Expertise

Data Science
Human Nonverbal Behavior
Interactive Machine Learning
Human-Computer Interaction
Computer Vision
Experimental Psychology
Social Skills Training

Social

Biography

M. Ehsan Hoque directs the Rochester Human-Computer Interaction Lab.

His research focuses on designing and implementing new algorithms to sense subtle human nonverbal behavior; enabling new behavior sensing and modelling for human-computer interaction; inventing new applications of emotion technology in high-impact social domains such as social skills training, public speaking; and assisting individuals who experience difficulties with social interactions.

Education

Massachusetts Institute of Technology

Ph.D.

Media Arts and Sciences (Media Lab)

2013

The University of Memphis

Electrical and Computer Engineering

M.Eng

2007

Penn State University

B.S.

Computer Engineering

2004

Affiliations

  • ACM Future of Computing Academy (ACM FCA)
  • Association for the Advancement of Artificial Intelligence (AAAI)
  • Association for Computing Machinery (ACM)
  • Institute of Electrical and Electronics Engineers (IEEE)
  • American Association for the Advancement of Science (AAAS)
Show All +

Selected Media Appearances

Connections: Can AI help us become more fair as a society?

WXXI  radio

2020-01-09

Can artificial intelligence -- or AI -- help us become more fair as a society?

How can we make sure the technology we create does not simply serve the most powerful in society? Our guests explore it:

Matt Kelly, independent journalist

Ehsan Hoque, Asaro-Biggar Family Assistant Professor of Computer Science at the University of Rochester

Jonathan Herington, lecturer in the Department of Philosophy, and assistant director of graduate education in the College of Arts, Sciences, and Engineering at the University of Rochester

View More

Federal award establishes Parkinson’s research center at URMC

WXXI  online

2018-10-04

The University of Rochester Medical Center has received a multimillion-dollar federal grant to study Parkinson’s disease, the university announced Wednesday. The $9.2 million award will fund the creation of a new research center, officials said.

Because of the volume of data expected to be generated from these experiments, Dorsey said, the center will involve URMC faculty disciplines beyond neurology, including biostatistics and computer science. The new research center “represents the novel convergence of medicine and data science,” said Ehsan Hoque, an assistant professor with the university’s institute for data science.

View More

How to Spot a Liar: Experts Uncover the Signs of Deception-Can you see them?

Newsweek  

2018-05-24

“A lot of times people tend to look a certain way or show some kind of facial expression when they’re remembering things,” commented Tay Sen, a PhD student working in the lab of Ehsan Hoque, an assistant professor of computer science. “When they are given a computational question, they have another kind of facial expression.” Using a machine learning tool, the researchers found patterns...

View More

Show All +

Selected Event Appearances

Automated Dyadic Data Recorder (ADDR) Framework and Analysis of Facial Cues in Deceptive Communication

Proceedings of ACM on Interactive, Mobile, Warble, and Ubiquitous Computing (IMWUT)  UbiComp 2018

CoCo: Collaboration Coach for Understanding Team Dynamics during Video Conferencing

Proceedings of ACM on Interactive, Mobile, Warble, and Ubiquitous Computing (IMWUT)  UbiComp 2018

The What, When, and Why of Facial Expressions: An Objective Analysis of Conversational Skills in Speed-Dating Videos

IEEE International Conference on Automated Face and Gesture Recognition  FG 2018

Show All +

Selected Articles

Buildup of speaking skills in an online learning community: A network-analytic exploration

Nature journal for Social Sciences

R. Shafipour, R. A. Baten, M. K. Hasan, G. Ghoshal, G. Mateos, and M. E. Hoque

2018

Studies in learning communities have consistently found evidence that peer-interactions contribute to students’ performance outcomes. A particularly important competence in the modern context is the ability to communicate ideas effectively. One metric of this is speaking, which is an important skill in professional and casual settings. In this study, we explore peer-interaction effects in online networks on speaking skill development. In particular, we present an evidence for gradual buildup of skills in a small-group setting that has not been reported in the literature. Evaluating the development of such skills requires studying objective evidence, for which purpose, we introduce a novel dataset of six online communities consisting of 158 participants focusing on improving their speaking skills. They video-record speeches for 5 prompts in 10 days and exchange comments and performance-ratings with their peers. We ask (i) whether the participants’ ratings are affected by their interaction patterns with peers, and (ii) whether there is any gradual buildup of speaking skills in the communities towards homogeneity. To analyze the data, we employ tools from the emerging field of Graph Signal Processing (GSP). GSP enjoys a distinction from Social Network Analysis in that the latter is concerned primarily with the connection structures of graphs, while the former studies signals on top of graphs. We study the performance ratings of the participants as graph signals atop underlying interaction topologies. Total variation analysis of the graph signals show that the participants’ rating differences decrease with time (slope = −0.04, p 

View more

Automated Dydadic Data Recorder (ADDR) Framework and Analysis of Facial Cues in Deceptive Communication

PACM on Interactive, Mobile, Wearable, and Ubiquitous Computing (IMWUT)

T. Sen, K. Hasan, Z. Teicher, M. E. Hoque

2018

We developed an online framework that can automatically pair two crowd-sourced participants, prompt them to follow a research protocol, and record their audio and video on a remote server. The framework comprises two web applications: an Automatic Quality Gatekeeper for ensuring only high quality crowd-sourced participants are recruited for the study, and a Session Controller which directs participants to play a research protocol, such as an interrogation game. This framework was used to run a research study for analyzing facial expressions during honest and deceptive communication using a novel interrogation protocol. The protocol gathers two sets of nonverbal facial cues in participants: features expressed during questions relating to the interrogation topic and features expressed during control questions. The framework and protocol were used to gather 151 dyadic conversations (1.3 million video frames). Interrogators who were lied to expressed the smile-related lip corner puller cue more often than interrogators who were being told the truth, suggesting that facial cues from interrogators may be useful in evaluating the honesty of witnesses in some contexts. Overall, these results demonstrate that this framework is capable of gathering high quality data which can identify statistically significant results in a communication study.

View more

CoCo: Collaboration Coach for Understanding Team Dynamics during Video Conferencing

PACM on Interactive, Mobile, Wearable, and Ubiquitous Computing (IMWUT)

S. Samrose, R. Zhao, J. White, V. Li, L. Nova, Y. Lu, M. R. Ali, M. E. Hoque

2018

We present and discuss a fully-automated collaboration system, CoCo, that allows multiple participants to video chat and receive feedback through custom video conferencing software. After a conferencing session, a virtual feedback assistant provides insights on the conversation to participants. CoCo automatically pulls audial and visual data during conversations and analyzes the extracted streams for affective features, including smiles, engagement, attention, as well as speech overlap and turn-taking. We validated CoCo with 39 participants split into 10 groups. Participants played two back-to-back teambuilding games, Lost at Sea and Survival on the Moon, with the system providing feedback between the two. With feedback, we found a statistically significant change in balanced participation—that is, everyone spoke for an equal amount of time. There was also statistically significant improvement in participants’ self-evaluations of conversational skills awareness, including how often they let others speak, as well as of teammates’ conversational skills.

View more

Show All +