Why it's important to build fair and ethical ML models

- 2 mins

For many years, others made vital decisions for us, and now, machine learning often takes the lead.

Within justice, defendants get scores predicting future crime risks, guiding decisions on bail, sentencing, and parole. In finance, statistical tools dictate our mortgage approvals and credit ratings. In business, ML sifts through job applications. In education, it determines our school admissions.

Beyond the critical, ML subtly shapes our culture daily. From search engines to news feeds to ads, it shapes our knowledge and views.

However, erratic or flawed decisions can stifle our ambitions and opportunities. We trust that admissions, job, and loan decisions hinge on truly pertinent factors.

Algorithmic decision making carries the risk of discrimination. Given the reliance on algorithms, it is important to ensure:


ML cycle — how ML predictions change the world

Imgur

Cycle of ML models

Historically, our records often mirror society’s biases, stereotypes, and unequal demographics. Analyzing such data risks perpetuating these very issues.

Thus, it’s paramount that our models don’t amplify these prejudices, especially related to race, gender, or other sensitive traits. Simply put, they shouldn’t just echo society’s ingrained biases, nor should they introduce new ones.

The reliability of evidence-based decision-making hinges heavily on the quality of the evidence itself. Just because machine learning thrives on data doesn’t guarantee its decisions will be accurate, consistent, or just.


What can we do about it?

So, how can we ensure decisions are made with integrity and for the right motives?

The answer lies in rigorous data standards: ample examples to reveal nuanced patterns; a wide variety of examples to represent diverse scenarios; well-documented examples to provide machine learning with dependable ground truth, and so forth.

To maintain these standards, we should consistently review the model’s decisions, ensuring fairness metrics are met—similar to how we monitor the model’s overall performance over time.


Viktoriia Oliinyk

Viktoriia Oliinyk

Building Responsible AI

comments powered by Disqus