Can AI Spot Liars?
September 5, 2019 | University of Southern CaliforniaEstimated reading time: 3 minutes
Most algorithms have probably never heard the Eagles' song, "Lyin' Eyes." Otherwise, they'd do a better job of recognizing duplicity.
Computers aren't very good at discerning misrepresentation, and that's a problem as the technologies are increasingly deployed in society to render decisions that shape public policy, business and people's lives.
Turns out that algorithms fail basic tests as truth detectors, according to researchers who study theoretical factors of expression and the complexities of reading emotions at the USC Institute for Creative Technologies. The research team completed a pair of studies using science that undermines popular psychology and AI expression understanding techniques, both of which assume facial expressions reveal what people are thinking.
"Both people and so-called 'emotion reading' algorithms rely on a folk wisdom that our emotions are written on our face," said Jonathan Gratch, director for virtual human research at ICT and a professor of computer science at the USC Viterbi School of Engineering. "This is far from the truth. People smile when they are angry or upset, they mask their true feelings, and many expressions have nothing to do with inner feelings, but reflect conversational or cultural conventions."
Gratch and colleagues presented the findings today at the 8th International Conference on Affective Computing and Intelligent Interaction in Cambridge, England.
Of course, people know that people can lie with a straight face. Poker players bluff. Job applicants fake interviews. Unfaithful spouses cheat. And politicians can cheerfully utter false statements.
Yet, algorithms aren't so good at catching duplicity, even as machines are increasingly deployed to read human emotions and inform life-changing decisions. For example, the Department of Homeland Security invests in such algorithms to predict potential threats. Some nations use mass surveillance to monitor communications data. Algorithms are used in focus groups, marketing campaigns, to screen loan applicants or hire people for jobs.
"We're trying to undermine the folk psychology view that people have that if we could recognize people's facial expressions, we could tell what they're thinking," said Gratch, who is also a professor of psychology. "Think about how people used polygraphs back in the day to see if people were lying. There were misuses of the technology then, just like misuses of facial expression technology today. We're using naïve assumptions about these techniques because there's no association between expressions and what people are really feeling based on these tests."
To prove it, Gratch and fellow researchers Su Lei and Rens Hoegen at ICT, along with Brian Parkinson and Danielle Shore at the University of Oxford, examined spontaneous facial expressions in social situations. In one study, they developed a game that 700 people played for money and then captured how people's expressions impacted their decisions and how much they earned. Next, they allowed subjects to review their behavior and provide insights into how they were using expressions to gain advantage and if their expressions matched their feelings.
Using several novel approaches, the team examined the relationships between spontaneous facial expressions and key events during the game. They adopted a technique from psychophysiology called "event-related potentials" to address the extreme variability in facial expressions and used computer vision techniques to analyze those expressions. To represent facial movements, they used a recently proposed method called facial factors, which captures many nuances of facial expressions without the difficulties modern analysis techniques provide.
The scientists found that smiles were the only expressions consistently provoked, regardless of the reward or fairness of outcomes. Additionally, participants were fairly inaccurate in perceiving facial emotion and particularly poor at recognizing when expressions were regulated. The findings show people smile for lots of reasons, not just happiness, a context important in the evaluation of facial expressions.
"These discoveries emphasize the limits of technology use to predict feelings and intentions," Gratch said. "When companies and governments claim these capabilities, the buyer should beware because often these techniques have simplistic assumptions built into them that have not been tested scientifically."
Prior research shows that people will make conclusions about other's intentions and likely actions simply based off of the other's expressions. While past studies exist using automatic expression analysis to make inferences, such as boredom, depression and rapport, less is known about the extent to which perceptions of expression are accurate. These recent findings highlight the importance of contextual information when reading other's emotions and support the view that facial expressions communicate more than we might believe.
Subscribe
Stay ahead of the technologies shaping the future of electronics with our latest newsletter, Advanced Electronics Packaging Digest. Get expert insights on advanced packaging, materials, and system-level innovation, delivered straight to your inbox.
Subscribe now to stay informed, competitive, and connected.
Suggested Items
ASC’s John Johnson Bullish on the U.S. and High-tech PCBs
04/28/2026 | Marcy LaRont, I-Connect007It was a good couple of days at the SMTA UHDI Symposium in Avondale, Arizona, in early April, where John Johnson, head of technology at American Standard Circuits (ASC) and resident PCB expert on UHDI in the real-world of manufacturing, was a presenter. As the symposium ended, I visited with John, who reflected on what he considered most important and what had made the greatest impression on him.
Cleaning Is Critical
04/28/2026 | Nolan Johnson, SMT007 MagazineMike Bixenman, president of Magnalytix, and his colleagues Vladimir Sitko, founder of PBT Works, and Adam Klett, PhD, director of science at KYZEN, discussed how cleaning can affect reliability, in a Professional Development Course at APEX EXPO 2026. We spoke to Mike after the session to better understand the context for the current activity in cleaning technology.
A Mic, a Tiny Cam, and the Show Floor: The Story Behind 'Take the Mic!'
04/23/2026 | I-Connect007 Editorial TeamI-Connect007’s Take the Mic! program at APEX EXPO 2026 in Anaheim is proving to be a must-watch feature for this year’s event coverage, as we invited company representatives to participate in a sponsored conversation about their businesses and products in a whole new way: in front of a camera at their booths. While our managing editors and guest editors were conducting slightly longer Real Time with… APEX EXPO interviews in our booth, three staff members and guests acted as roaming reporters, conducting short, prearranged interviews with several companies exhibiting on the show floor.
PHOTO GALLERY: A Walk on the Lighter Side
04/23/2026 | I-Connect007APEX EXPO isn’t all business! To wrap up our last photo gallery, we’re sharing some of our favorites from the show—a walk on the lighter side. We’re highlighting those in-between moments: quick selfies, familiar faces reconnecting, robot dogs, booth-side laughs, and the kind of candid snapshots you only get when people are genuinely enjoying themselves. At the end of the day, it’s not just about what we do, it’s about the people we do it with.
Koh Young Highlighting its Smart AI Solutions at the SMTA Monterrey Expo and Tech Forum
04/15/2026 | Koh YoungKoh Young, the industry leader in True 3D measurement-based inspection solutions, will exhibit at the upcoming SMTA Monterrey Expo & Tech Forum on Thursday, April 23, 2026, at Cintermex, located at Av. Fundidora 501, Colonia Obrera, Monterrey, Nuevo León, Mexico.