Eightfold AI scored over a billion workers for Microsoft, Morgan Stanley, and PayPal. Nobody got a receipt.
The score you never saw
Erin Kistler has a computer science degree from Ohio State and 19 years of product management experience. She applied for a job at PayPal. Before a human recruiter ever opened her application, an algorithm had already scored her on a 0 to 5 scale -- star ratings, half-point increments -- and decided whether she was worth a look.
She did not know the score existed. Neither did over a billion other workers in Eightfold AI's database. The ones who scored low were filtered out silently -- discarded before anyone even looked.

On January 20, 2026, Kistler and co-plaintiff Sruti Bhaumik filed a class action in a California court. Their lawyer is Jenny R. Yang, former chair of the U.S. Equal Employment Opportunity Commission. The client list behind Eightfold reads like a Fortune 500 roll call: Microsoft, Morgan Stanley, Starbucks, PayPal, Chevron, Bayer, BNY. The system collected data from LinkedIn, GitHub, Stack Overflow, and other public sources to build profiles that candidates never consented to and had no way to challenge. (Eightfold disputes that it scrapes social media directly. The complaint cites the company's own privacy policy listing those platforms as data sources.)
The headlines called it an AI bias lawsuit. That framing is valid -- I am not dismissing it. But from where I sit, it misses the more useful story.
I spent two years at a national credit bureau building ML models that scored real people for income estimation and loan screening. Every model cleared a governance process before it touched a single applicant. Looking at the Eightfold case through that lens, this is not primarily a bias problem. It is a governance vacuum -- and the governance already exists in another industry.
Everyone is having the wrong conversation
Every time an AI hiring tool gets sued, the discourse follows the same arc. Algorithms are biased. We need fairness audits. Humans should stay in the loop. Rinse, repeat.
These are real concerns. They are also the concerns of someone looking at the problem from the outside -- like diagnosing a building collapse by talking about the weather when the foundation had no rebar.
The foundation is governance. Credit scoring has been assigning numbers to people for high-stakes decisions since the Fair Credit Reporting Act of 1970. The industry got it wrong plenty of times. But through decades of lawsuits, regulation, and hard operational lessons, it built a framework with five components: documentation, validation, discrimination testing, disclosure, and dispute rights. If you score people, you owe them the receipt. Simple enough to fit on a sticky note. Eightfold allegedly violated every part of it.

Former EEOC Commissioner Keith Sonderling has argued this for years: existing employment law already covers algorithmic hiring -- the problem is enforcement, not missing statute. The "black box" defense that AI hiring vendors rely on would not survive the scrutiny applied to consumer finance. And the Kistler plaintiffs are testing this directly -- their legal theory maps Eightfold's scores onto the consumer reporting framework that has governed credit bureaus for over 50 years. The governance exists. AI hiring skipped it.
(Deploying an unvalidated model that secretly scores millions would be career-ending in banking. In AI hiring, it is a Series C. Eightfold has raised over $400 million.)
What credit scoring requires -- and what Eightfold skipped
I had to clear five governance controls before any model I built touched an applicant at a credit bureau. Let me walk through them.
A model cannot be a black box. The Federal Reserve requires documentation detailed enough that someone outside the team can audit it -- think of it as a recipe a stranger could follow. Eightfold's methodology is proprietary. Kistler did not know the recipe existed, let alone what was in it. And the developers do not get to grade their own homework -- in credit risk, a separate team stress-tests every model before deployment. You would not trust a bridge designed and inspected by the same engineer. There is no public evidence that Eightfold's model has been independently validated by anyone outside the company.

Discrimination testing is where this gets personal for me. The harder problem is not explicit bias -- it is proxy discrimination. Variables that look neutral but correlate with protected classes. An insurance actuary can use your driving record but cannot use your zip code as a stand-in for your ethnicity. I spent enough time building scoring models to know that proxy effects hide in the features you least expect -- a seemingly innocent variable like "years since last job change" can penalize people who took parental leave or recovered from illness. The rules for catching this are specific. The testing is mandatory. Eightfold claims its AI reduces bias. The lawsuit alleges the opposite. Without access to the model's features and weights, there is no way to know which is true -- and that opacity alone would be disqualifying in credit risk.
Disclosure is the most straightforward rule and the most clearly violated. Before using a consumer report for employment decisions, the employer must provide a clear disclosure and get written authorization. Eightfold provided neither. This is the part that should make employers nervous. If Eightfold's output qualifies as a consumer report -- and the statutory definition is broad enough to include a hiring fitness score -- then every company using the tool has been violating federal disclosure law. Microsoft. Morgan Stanley. All of them.
Dispute rights close the loop. Under FCRA, consumers can see what is in their file, challenge inaccuracies, and have errors investigated within 30 days. A candidate scored 1 out of 5 because the system ingested an outdated LinkedIn profile has no path to correction. The score filters them out silently before a recruiter opens the application.
Imagine getting rejected from a bank loan because of a credit score you could not see, built from data you did not provide, with no way to challenge it. That is what happened here -- just in hiring instead of lending. In credit scoring, this would be a violation on day one. In AI hiring, it ran for years.

What this means if you build or buy these tools
The Kistler lawsuit is not isolated.
New York City already requires annual bias audits for automated hiring tools. Compliance research from Citizens and Tech found most companies are barely meeting the bar -- "extreme discretion" and "strong incentives to avoid transparency" were the researchers' words. But the law is active, auditors are engaged, and other cities are watching.
The EEOC issued guidance in 2023 saying AI hiring tools should be held to the same standard as any other hiring test. That guidance was pulled from the EEOC website in January 2025 under the new administration. The legal theory did not disappear with the webpage.
The federal AI safety playbook -- NIST's risk management framework -- reads like a compressed version of what credit bureaus have been doing for years. Different vocabulary, same architecture.
If you score people for consequential decisions, governance is coming whether you build it proactively or a court orders it. The companies that build it first avoid the scramble when the rules catch up. The rules always catch up. (Ask anyone who was in consumer lending before 2008.)

Back to the billion
Erin Kistler did not know she had a score. Over a billion other profiled workers still do not.
Every one of those failures maps to a governance control that has existed in credit scoring for years -- some for over half a century. The AI hiring industry did not discover a new problem. It skipped solutions that were already sitting on the Federal Reserve's website, in the operational playbooks of every credit bureau I worked at.

The governance is not hard to find. The will to apply it is.
An actuary would have stopped Eightfold at the door. The courts are now deciding whether they will do what the industry would not.