By Aimee Picchi, CBS News
If you're on the hunt for a new job, there's a new twist to the hiring process that's increasingly common: Getting judged by artificial intelligence.
Major businesses like Unilever and Ikea are turning to AI programs to sniff out suitable applicants — and predict future job performance — partly as a way to cut down on the time needed to hire employees.
Supporters say the tech can help companies find better job candidates and eliminate some of the inherent bias that executives may not even be aware of, such as leaning toward candidates who are white or male.
But critics say there are issues with relying on algorithms to make complex decisions. One Princeton professor, Arvind Narayanan, is questioning whether AI can accurately predict job success, calling such claims "snake oil."
And Electronic Privacy Information Center counsel John Davisson says the AI programs are a "nightmarish combination of privacy violations and outcomes that aren't based on any easily understood criteria." His group earlier this month petitioned the Federal Trade Commission to start regulating AI's use in the job market and other business applications.
At stake are the roughly 7 million job openings that employers seek to fill each month. Traditional gatekeepers such as human resources professionals are increasingly turning to AI to review resumes — no small task when employment site Glassdoor says the average job opening receives about 250 resumes. Some are also relying on AI programs to score applicants' video interviews, with the goal of skimming the cream from the top.
Some companies say they see results, with one case study from Unilever finding that an AI analysis from a company called HireVue saved the company $1 million by cutting down on the time needed to sort through 250,000 applications. Relying on AI to cull the applicant pool also helped diversify new hires by 16%, it found.
Yet the lack of transparency about the inner workings of these AI applications, as well as the vast amount of data they collect, are raising concerns among lawmakers and privacy experts, with some worried that these programs could unfairly judge certain types of applicants, such as minorities and women.
"It's like anything else — wine, music and food. Some of it is really good and some of it is really bad," said Tomas Chamorro-Premuzic, chief talent scientist at staffing and recruitment company ManpowerGroup, who is also an adviser to HireVue. "The big challenge and question of AI in the world of talent is: Will algorithms be better at knowing us than we do ourselves?"
Gaining traction in workplace
So far, there appears to be more questions than answers, even as the technology gains traction in the workplace. While there's no federal government tracking or oversight of AI, surveys of human resource executives provide insight into the usage of AI in the job market.
Almost 7 of 10 recruiters and hiring managers told LinkedIn in 2018 that AI applications help them save time, according to a survey of 8,800 executives, while about 4 of 10 said that AI tools help remove bias in the hiring process. And employment site Glassdoor says it's increasingly likely that candidates will encounter AI when they search for jobs today.
It's a growing field with dozens of companies offering different AI flavors to businesses. It's also attracting venture capital, with hundreds of millions of investments plugged into firms such as HireVue, Pymetrics and Mya, according to CrunchBase.
How it works
Some job applicants might not even realize they've been given the once-over by artificial intelligence. At the most basic level, AI programs analyze resumes and job applications to screen candidates, winnowing down the field to those considered the best choices.
Other AI programs are more interactive, although some applicants still might not know they're dealing with an algorithm. Consignment shop ThredUp, for example, uses an AI program to text job candidates to schedule job interviews, according to Glassdoor. Other companies ask candidates to play online games, like one game that asks applicants to click on objects that are red balls. The data are then fed through an AI program that analyzes responses to assess a candidate's abilities.
But the AI program that's perhaps sparking the most concern is the video interview, where candidates are taped as they provide responses to recruitment questions via webcam. An AI program then examines the video, analyzing everything from facial gestures to vocabulary, to determine whether the candidate is a good fit.
HireVue, one of the biggest companies in the field, says its AI assessment is partially based on data from companies' top performers: By analyzing vocabulary used by top workers, for instance, the AI program can pick out candidates whose word choices and expressions best match those top workers. It also relies on years of research into industrial and organizational psychology, a branch of psychology that focuses on how people behave within organizations.
"We are judging the answers that the interviewee is giving," said Kevin Parker, CEO of HireVue. "In a 20-minute interview, we can collect a lot of data. We have no idea who the candidate is, but we can collect a lot of data about vocal variation, personal pronouns" and other information to make an assessment, he added.
HireVue's AI programs, Parker said, are screened for bias, which helps level the playing field for job applicants. The AI models tend to be most effective for jobs where there's a lot of data to draw from, he noted.
In other words, AI programs aren't designed for finding the best CEO — since C-suite executives are relatively uncommon in the workforce — but for more rank-and-file workers like salespeople.
Is it accurate?
Companies in the AI space say that their programs save time by weeding out low-quality applicants and singling out the best potential hires. But artificial intelligence has its limits — and probably isn't more accurate at predicting job performance than tried-and-true methods, according to Arvind Narayanan, associate professor of computer science at Princeton University.
Narayanan points to an academic study that relied on family data to predict future outcomes in a number of areas, such as a child's likely academic performance and future material hardship. But the study's AI analysis failed to outperform old-fashioned linear regression in predicting future struggles or successes, he said.
While AI can prove useful for simpler perception-based tasks, such as speech-to-text translation or identifying plagiarized material, it "can't predict social outcomes," Narayanan concluded.
At the same time, he warned, AI has the potential to create harm, such as transferring massive amounts of personal data to tech firms and placing a greater reliance on applications that lack transparency.
Those concerns are echoed by EPIC, which in November asked the Federal Trade Commission to investigate HireVue. It claims HireVue's algorithms are "opaque" and that its "secret analysis of biometric data thus causes substantial privacy harms to job candidates."
The FTC told CBS MoneyWatch it had received both the letter and the petition, but declined additional comment.
HireVue said it believes the complaint is without merit. "The company upholds the highest levels of rigor and ethics as its employees work every day to increase fairness and objectivity in the hiring process," it told CBS MoneyWatch.
"Disappointed"
HireVue CEO Parker said he was "disappointed" with Narayanan's assessment, and said his company reached out to the Princeton professor to discuss his concerns but was "rebuffed."
Asked about whether he rebuffed HireVue's overture, Narayanan forwarded his email exchange with the company to CBS MoneyWatch. Instead of a private discussion, he wrote, he urged HireVue to publish its research, "ideally in a peer-reviewed forum, and discuss it openly." He added in the email, "I'm not sure a private conversation is the best medium. There's a high level of public interest in AI bias since the issue affects everyone."
In a statement, HireVue said it "was surprised that Dr. Narayanan would not welcome the opportunity for a one-on-one discussion on the facts and established science supporting HireVue's approach to assessments and AI before making public claims about efficacy and validity."
The issue of AI in hiring is gaining more attention from lawmakers. In January, a new Illinois law went into effect that requires companies to disclose that artificial intelligence will be used to assess candidates' video interviews. Employers also have to explain how the AI works, ask for consent from the job candidate, and destroy the video and any copies within 30 days, if a job applicant requests it.
Illinois State Rep. Jaime Andrade Jr., who sponsored the bill, told the Chicago Tribune last month: "The technology hasn't been vetted fully yet. There are some concerns regarding bias. [The software] is only as good as the data it is given."
Is it less biased?
HireVue and other hiring experts say AI has the potential to be less biased than human interviewers. Algorithms can strip out gender, race or any other protected group when analyzing candidates. Yet when algorithms go wrong, they can go very wrong, such as in the case of Amazon's recruitment engine.
In that now-infamous instance, Amazon scrapped the AI application in 2018 after realizing that the program was excluding women applicants from the pool — because the model was based on 10 years of data from job applicants, who tended to be men. The AI, in other words, "learned" the biased view that men were a better fit than women, reinforcing gender stereotypes in tech.
"Even if it doesn't look at sensitive attributes, like names that can be a proxy for gender, programs can pick up on this unknowingly," said Glen Cathey, senior vice president and head of digital strategy and innovation at Randstad, a recruitment agency. "Being responsible and ethical in AI, not just in recruitment, is the measurement element."
EEOC standard
Companies that rely on AI need to measure and test for bias, and that's something HireVue studies closely, said Lindsey Anderson Zuloaga, director of data science at the company.
"Whatever biases it has it could affect a lot of people — and that's true of who gets credit, who gets parole," she said, referring to other fields where AI has been used to make determinations. "In any pre-hire assessments, you need to look at how different demographic groups are being scored, and if there are differences, that may be evidence of adverse impact."
HireVue screens for results where any group, such as different genders or races, score less than 80% of the top-scoring group, which is a yardstick for bias used by the U.S. Equal Employment Opportunity Commission. For example, that would happen if an AI program singles out 50 men as solid candidates, but only 30 women.
"We're lucky in this space in that our notion of what is fair is defined by law," Zuloaga said. "We build the model, then we train the algorithm, then look at how is it scoring people within the group."
Hacks for AI?
Job applicants are perhaps justifiably nervous about facing an AI in a job interview, which can be nerve-wracking even with other humans. It's no surprise that YouTube tutorials have sprung up to help job candidates learn how to ace pre-recorded interviews by HireVue and others.
The tips include how to set up your webcam, how to look good on the video and how to keep cool under pressure — not exactly different advice than what a job hunter would hear about an interview with a real person.
In the end, there may be no secret sauce for acing an interview with an AI, partially because the algorithm is somewhat of a black box. Even so, human judgement is typically involved in the hiring process after AI programs have screened for the best candidates, HireVue's Parker said.
"We're the first step in a process — the next step is an in-person interview," he added. "It's our desire that there is always a human involved."
See more at CBS News
COMMENTS