The Human Factor Is Essential To Eliminating Bias in Artificial Intelligence

It is not enough to open the ‘black box’ of machine learning. Direct human evaluation is the only way to ensure biases are not perpetuated through AI.

More and more technology and digital services are built upon, and driven, by AI and machine learning. But as we are beginning to see, these programmes are starting to replicate the biases which are fed into them, notably biases around gender. It is therefore imperative that the machine learning process is managed from input to output – including data, algorithms, models, training, testing and predictions – to assure that this bias is not perpetuated.

Bahar Gholipour notes this bias as AI’s so-called ‘black box’ problem — our inability to see the inside of an algorithm and therefore understand how it arrives at a decision. He claims that ‘left unsolved, it can devastate our societies by ensuring that historical discrimination, which many have worked hard to leave behind, is hard-coded into our future.’

Technological expertise is not enough to scrutinize, monitor and safeguard each stage of the machine learning process. The experience and perspective of people of all ages and all walks of life is needed to identify both obvious and subliminal social and linguistic biases, and make recommendations for adjustments to build accuracy and trust. Even more important than having an opportunity to evaluate gender bias in the ‘black box’ is having the freedom to correct the biases discovered.

The first step is to open the ‘black box’. Users are increasingly demanding that AI be honest, fair, transparent, accountable and human-centric. But proprietary interests and security issues have too often precluded transparency. However, positive initiatives are now being developed to accelerate open-sourcing code and create transparency standards. AI Now, a nonprofit at New York University advocating for algorithmic fairness, has a simple principle worth following: ‘When it comes to services for people, if designers can’t explain an algorithm’s decision, you shouldn’t be able to use it.’

Now there are a number of public and private organizations who are beginning to take this seriously. Google AI has several projects to push the business world, and society, to consider the biases in AI, including GlassBox, Active Question Answering and its PAIR initiative (People + AI Research) which add manual restrictions to machine learning systems to make their outputs more accurate and understandable.

The US Defense Advanced Research Projects Agency is also funding a big effort called XAI (Explainable AI) to make systems controlled by artificial intelligence more accountable to their users.

Microsoft CEO Satya Nadella has also gone on the record defending the need for ‘algorithmic accountability’ so that humans can undo any unintended harm.

But laudable as these efforts are, opening the box and establishing regulations and policies to ensure transparency is of little value until you have a human agent examining what’s inside to evaluate if the data is fair and unbiased. Automated natural language processing alone cannot do it because language is historically biased – not just basic vocabulary, but associations between words, and relationships between words and images.

Semantics matter. Casey Miller and Kate Swift, two women who in 1980 wrote The Handbook of Nonsexist Writing – the first handbook of its kind – dedicated their lives to promoting gender equity in language. That was almost 40 years ago and, while technology has advanced exponentially in that time period, we've made little progress removing gender bias from our lexicon.

The challenge for AI is in programming a changing vocabulary into a binary numerical system. Human intervention is necessary to adjudicate the bias in the programmer, the context and the language itself. But gender bias is not just in the algorithms. It lies within the outcomes – predictions and recommendations – powered by the algorithms.

Common stereotypes are even being reinforced by AI's virtual assistants: those tasked with addressing simple questions (e.g. Apple’s Siri and Amazon’s Alexa) have female voices while more sophisticated problem-solving bots (e.g. IBM’s Watson and Microsoft’s Einstein) have male voices.

Gender bias is further exacerbated by the paucity of women working in the field. AI Now’s 2017 report (opens in new window) identifies the lack of women, and ethnic minorities, working in AI as a foundational problem that is most likely having a material impact on AI systems and shaping their effects in society.

Human agents must question each stage of the process, and every question requires the perspective of a diverse, cross-disciplinary team, representing both the public and private sectors and inclusive of race, gender, culture, education, age and socioeconomic status to audit and monitor the system and what it generates. They don't need to know the answers – just how to ask the questions.

In some ways, 21st century machine learning needs to circle back to the ancient Socratic method of learning based on asking and answering questions to stimulate critical thinking, draw out ideas and challenge underlying presumptions. Developers should understand that this scrutiny and reformulation helps them clean identified biases from their training data, run ongoing simulations based on empirical evidence and fine tune their algorithms accordingly. This human audit would strengthen the reliability and accountability of AI and ultimately people’s trust in it.

By  Elizabeth Isele.  Associate Fellow, Global Economy and Finance, Royal Institute of International Affairs

Chatham House

You Might Also Read: 

Real Dangers of Artificial Intelligence:

Do Companies Need A Chief AI Officer?:

Don't Leave AI Governance To The Machines:

 

« Keeping Young People Off The Dark Web
UK Fallout From The Massive Breach At Equifax »

Infosecurity Europe
CyberSecurity Jobsite
Perimeter 81

Directory of Suppliers

Practice Labs

Practice Labs

Practice Labs is an IT competency hub, where live-lab environments give access to real equipment for hands-on practice of essential cybersecurity skills.

BackupVault

BackupVault

BackupVault is a leading provider of automatic cloud backup and critical data protection against ransomware, insider attacks and hackers for businesses and organisations worldwide.

Alvacomm

Alvacomm

Alvacomm offers holistic VIP cybersecurity services, providing comprehensive protection against cyber threats. Our solutions include risk assessment, threat detection, incident response.

CYRIN

CYRIN

CYRIN® Cyber Range. Real Tools, Real Attacks, Real Scenarios. See why leading educational institutions and companies in the U.S. have begun to adopt the CYRIN® system.

Infosecurity Europe, 3-5 June 2025, ExCel London

Infosecurity Europe, 3-5 June 2025, ExCel London

This year, Infosecurity Europe marks 30 years of bringing the global cybersecurity community together to further our joint mission of Building a Safer Cyber World.

TitanFile

TitanFile

TitanFile is an award-winning, easy and secure way for professionals to communicate without having to worry about security and privacy.

Versasec

Versasec

Versasec is a leader in identity and access management, providing customers with security solutions for managing digital identities.

CERT.hr

CERT.hr

CERT.hr is the national authority competent for prevention and protection from computer threats to public information systems in the Republic of Croatia.

Military Cyber Professionals Association (MCPA)

Military Cyber Professionals Association (MCPA)

MCPA are a team of Soldiers, Sailors, Airmen, Marines, Veterans and others interested in the development of the American military cyber profession.

Labs/02

Labs/02

Labs/02 is a seed-stage incubator with a mission to advance cutting-edge technology in innovative areas including AI, deep learning, autonomous transportation, and smart cities.

CYBRScore

CYBRScore

CYBRScore is a premium, performance-based cyber skills training and assessment provider that quantifies a user’s ability to defend a network.

Unit21

Unit21

Unit21 helps protect businesses against adversaries through a simple API and dashboard for detecting and managing money laundering, fraud, and other sophisticated risks across multiple industries.

Hybrid Identity Protection Conference (HIP)

Hybrid Identity Protection Conference (HIP)

Hybrid Identity Protection (HIP) is the premier educational forum for identity-centric cybersecurity practitioners charged with defending hybrid cloud environments.

Nitrokey

Nitrokey

Nitrokey is the world-leading company in open source security hardware. Nitrokey develops IT security hardware for data encryption, key management and user authentication.

Input Output (IOHK)

Input Output (IOHK)

IOHK is one of the world's pre-eminent blockchain infrastructure research and engineering companies.

Kennedys

Kennedys

Kennedys is a global law firm with expertise in litigation/dispute resolution and advisory services, particularly in the insurance/reinsurance and liability sectors, including cyber risk.

Siometrix

Siometrix

Siometrix addresses digital identity fraud. It steals your attacker's time and prevents many prevalent attack vectors.

AI Security Institute (AISI)

AI Security Institute (AISI)

The AI Security Institute’s mission is to minimise surprise to the UK and humanity from rapid and unexpected advances in AI.

Bluerydge

Bluerydge

Bluerydge specialises in cyber security and technology, focusing on the delivery of innovative sovereign solutions through trusted, cleared and experienced professionals.

Quantum Squint

Quantum Squint

Quantum Squint is a cutting-edge cybersecurity company specializing in the use of advanced regression management techniques to detect, analyze, and prevent vulnerabilities in digital systems.

Black Cipher Security

Black Cipher Security

Black Cipher is a New Jersey-based cybersecurity and incident response consulting firm.