AAPL logo

Professional Capabilities

How AI Assessment Tools Affect Job Candidates’ Behavior

Jonas Goergen | Emanuel de Bellis | Anne-Kathrin Klesse

August 25, 2025


Summary:

When considering AI assessment tools, most companies focus on the gains the tools bring in terms of efficiency and quality. But what they don’t factor in is how AI assessment may change candidates’ behavior during the process.





According to the World Economic Forum, more than 90% of employers use automated systems to filter or rank job applications, and 88% of companies already employ some form of AI for initial candidate screening. Take Unilever, for example. The consumer goods giant uses AI-driven tools from HireVue to assess early-career applicants, saving 50,000 hours and more than $1 million in the process.

Most companies, when considering AI assessment tools, focus on the gains the tools bring in terms of efficiency and quality. But what they don’t factor in is how AI assessment may change candidates’ behavior during the assessment. Our new research, examining over 13,000 participants across 12 studies, reveals that this is a crucial blind spot. We looked at simulation of a variety of assessment situations in both the laboratory and the field, and we collaborated with a startup platform offering game-based hiring solutions called Equalture.

The results show that job candidates consistently emphasized analytical traits when they believed AI was evaluating them, while downplaying the very human qualities—empathy, creativity, intuition—that often distinguish outstanding employees from merely competent ones. This drove candidates to present a different and potentially more homogeneous version of themselves, in turn affecting who was likely to succeed in an AI-enabled hiring process, with implications for organizations using AI in hiring, promotions, or admission decisions.

Why This Matters for Your Organization

The implications of our findings extend beyond individual hiring decisions. When candidates systematically misrepresent themselves, organizations face several critical challenges:

  • Talent pool distortion: While AI is sometimes blamed for making biased hiring decisions (for example, discriminating against women in the selection process), our research suggests that knowing that one is assessed by AI also biases candidates, making them believe that they should prioritize their analytical capabilities. As a result, companies may be screening out exactly the candidates they need simply by using AI: that innovative thinker or emotionally intelligent leader you’re looking for might present themselves as a rule-following analyst because they believe that is what the AI wants to see.

  • Validity compromise: Assessment tools are only as good as the data they collect. When candidates strategically alter their responses, the fundamental validity of the assessment process might be undermined. Organizations may no longer measure authentic capabilities—instead, they may measure what candidates think AI will value the most.

  • Unintended homogenization: If most candidates believe AI favors analytical traits, the talent pipeline may become increasingly uniform, potentially undermining diversity initiatives and limiting the range of perspectives in organizations. Companies like IBM and Hilton, which integrate AI into both hiring and internal promotion systems, must now contend with whether such tools nudge employees toward formulaic self-presentation.

New transparency regulations like the EU’s AI Act, which require organizations to disclose AI use in high-stakes decisions, make these outcomes all the more likely. When candidates are aware that an AI is assessing them, they are more likely to change their behavior.

What Leaders Can Do

Based on our findings, organizations can take several concrete steps to address the AI assessment effect:

  1. Radical transparency: Do not just disclose AI assessment—be explicit about what it actually evaluates. Clearly communicate that your AI can and does value diverse traits, including creativity, emotional intelligence, and intuitive problem-solving. This might include providing examples of successful candidates who demonstrated strong intuitive or creative capabilities. Currently, few companies seem to be transparent about what exactly it is that AI assesses—at least this information is not easily accessible when clicking through career page information on the websites of many major companies. That said, applicants discuss and share their intuitions on blogs and videos, which may be counterproductive because it may or may not align with actual practices. We advise companies not to leave their candidates to speculate.

  2. Regular behavioral audits: Implement systematic reviews of your AI assessment outcomes. For instance, New York City has enacted Local Law 144, requiring employers to conduct annual bias audits of AI-based hiring. In response, one of the market leaders in AI-based hiring, HireVue reports their recent audits for race or gender bias across jobs and use cases. In addition to examining biases regarding demographics, we suggest using these audits to look for patterns indicating behavioral adaptation: Are candidates’ responses becoming more homogeneous over time? Are you seeing a shift toward analytical presentations at the expense of other valuable traits?

  3. Hybrid assessment: Some organizations combine human and AI assessments. For example, Salesforce notes that besides technology, a human will review applications. Nvidia and Philip Morris International guarantee ultimate assessment and decision-making through a human. One of our studies shows that while this hybrid human assessment does reduce candidates’ tendency to highlight analytical capabilities, it does not eliminate it. To close the gap, you need to train your human hirers to compensate for the AI effect.

The Path Forward

As AI becomes increasingly embedded in organizational decision-making, we must recognize that these tools do not just change processes—they change people. The efficiency gains from AI assessment may come at the cost of authentic candidate presentation and, ultimately, the human diversity that makes organizations innovative and resilient. The irony is striking: In our quest to remove human bias from hiring, we may have created a system where AI introduces a new form of bias. The solution is not to abandon AI, but to design assessment systems that account for and counteract these behavioral shifts. Only by keeping humans—not just metrics—at the heart of our assessment strategies can we build hiring systems that truly identify and nurture the diverse talent our organizations need.

Copyright 2025 Harvard Business School Publishing Corporation. Distributed by The New York Times Syndicate.

Explore AAPL Membership benefits.

Jonas Goergen

Jonas Goergen is a PhD candidate at the Institute of Behavioral Science and Technology at the University of St.Gallen. He studies consumer interactions with technologies.


Emanuel de Bellis

Emanuel de Bellis is an Associate Professor at the University of St.Gallen and Director of the Institute of Behavioral Science and Technology. His research focuses on consumers’ perception and use of autonomous products.


Anne-Kathrin Klesse

Anne-Kathrin Klesse is a Professor of Consumer Behavior & Technology in the Marketing Management Department of the Rotterdam School of Management, Erasmus University, and the academic director of the Psychology of AI lab at the Erasmus Center for Data Analytics (ECDA). Her research centers on understanding the impact of technology on human judgment and decision making, with a particular focus on elucidating human lay beliefs surrounding AI as well as gen AI usage behavior.

Interested in sharing leadership insights? Contribute



For over 50 years.

The American Association for Physician Leadership has helped physicians develop their leadership skills through education, career development, thought leadership and community building.

The American Association for Physician Leadership (AAPL) changed its name from the American College of Physician Executives (ACPE) in 2014. We may have changed our name, but we are the same organization that has been serving physician leaders since 1975.

CONTACT US

Mail Processing Address
PO Box 96503 I BMB 97493
Washington, DC 20090-6503

Payment Remittance Address
PO Box 745725
Atlanta, GA 30374-5725
(800) 562-8088
(813) 287-8993 Fax
customerservice@physicianleaders.org

CONNECT WITH US

LOOKING TO ENGAGE YOUR STAFF?

AAPL provides leadership development programs designed to retain valuable team members and improve patient outcomes.

American Association for Physician Leadership©

formerly known as the American College of Physician Executives (ACPE)