November 23, 2024


recruitment
Credit: Pixabay/CC0 Public Domain

Imagine being the most qualified person for a job and not getting a call-back or not being selected for an interview.

As recruitment decisions shift from being based on assessing applicants’ traditional CVs to relying on information gleaned from technology, there is a risk qualified candidates will be filtered out because of their class, background or disability.

And perhaps even more worryingly, in some instances people may never know.

Employers are increasingly using technology to make, or assist with, recruitment decisions. Woolworths, Qantas and Afterpay have been reported to use bots to interview candidates while Hilton Australasia requires some applicants to create a TikTok video.

Other tools being used in modern day hiring include contextual recruitment systems, asynchronous video interviewing, gamification and .

Are we expecting too much of algorithms?

In my new book, I expose how some new recruitment technologies create unprecedented risks to equality in the workplace, with potential for discrimination.

Contextual recruitment systems is an algorithm used by a number of major employers, including law firms Allens Linklaters and Herbert Smith Freehills. It mines the demographic data of job candidates in an attempt to develop diverse workplaces. It helps recruiters identify so-called “hidden gems” from less privileged areas.

However, the algorithm can only assess data that is publicly available or volunteered. So it may not get an accurate picture of candidates who are in reality—as opposed to on paper—disadvantaged.

Instead, it may simply favor the stand out performers in settings it determines are less fortunate, to the detriment of those whose disadvantage cannot be measured by an algorithm.

Replacing face-to-face with face-to-video interviews

Human resources company HireVue’s asynchronous video interviewing tool, which uses AI to automate job interviews, is now being used by more than a hundred employers including Telstra.

Using the tool, prospective employers produce pre-recorded videos with questions for job applicants. Candidates can watch the video and then respond in their own time.

But this interviewing tool may disadvantage job candidates whose manner of speech and choice of words is associated with the working class. This is because the algorithm might favor candidates with cultivated language skills from professional households.

Gamification sounds like fun but can eliminate candidates

Then there is the increasing popularity of gamification, which involves filtering job candidates based on the results they get from playing online games. A number of employers, including ANZ, use this technique as part of the recruiting process.

But there are which may affect a person’s ability to complete . For example, in the Pymetrics lengths game candidates are shown two similar images which have very subtle differences (for example, one picture of a cat with a short mustache and one picture of a cat with a long mustache). Candidates are asked to press a certain key on their keyboard when they see a particular cat.

This may test for attention to detail as it is intended to do, but it also disadvantages people with disabilities which may hinder their visual perception or their physical or mental reaction speeds.

Given many people with invisible disabilities are unlikely to identify as disabled and to ask for reasonable adjustments, they are unlikely to speak up in this new digitalised environment.

For some job seekers disability and class may also intersect to compound their disadvantage. It’s worth considering whether being able to spot subtle differences in a cat’s mustache is relevant to some jobs such as being a lawyer, accountant or banker (as opposed to a pilot or bus driver).

Cybervetting gives employers a fuller picture

Social media is also increasingly used by employers to screen or “cybervet” job candidates, enabling both public and private profiles to be screened. In some cases, cybervetting is tantamount to having an visit your home to assess your family or determine if you’re a right “cultural fit” before they decide to interview or hire you.

Through cybervetting employers can access “digitized indicators” of a person’s class, social background and family. This makes them vulnerable to powerful cognitive biases which may lead them to discriminate.

Weighing up the benefits of technology

Using technology has many benefits, including automation to produce time and , but it may also come with a very human cost.

Employers need to take care the they are using to automate recruitment decisions does not also create risks of discrimination based on protected attributes.

Many of these technologies are touted on paper to reduce human bias and improve diversity in workplaces. But in practice they may have the opposite effect and create barriers to employment based on “social origin”, disability and age, or at the synergy of these attributes.

This may have scope to irreparably change the composition of our future professional workforce and favor the richer, more cultivated, more able and younger. This would see us move backwards, not forward, in our efforts for workplace equality.

Provided by
The Conversation


This article is republished from The Conversation under a Creative Commons license. Read the original article.The Conversation

Citation:
‘Computer says no’: More employers are using AI to recruit, increasing the risk of discrimination (2023, December 13)
retrieved 13 December 2023
from https://phys.org/news/2023-12-employers-ai-discrimination.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.





Source link

Leave a Reply

Your email address will not be published. Required fields are marked *