The Invisible AI: Episode 1 - The Day AI Rejected You (And You Didn't Even Know It)
Sun Feb 08 2026
What if an algorithm rejected you, and you never even knew it happened?
In this explosive first episode of our four-part "Invisible AI" series, Dr. JR teams up with AI research assistant Ada to expose the hidden algorithmic systems making life-changing decisions about millions of Americans every day—in hiring, housing, credit, and beyond.
In This Episode: • How 99% of Fortune 500 companies use AI to screen job applicants before humans ever see resumes • Why AI hiring tools prefer white-sounding names 85% of the time (University of Washington study) • The shocking truth about tenant screening algorithms that deny housing to qualified applicants • SafeRent's $2.28M settlement for algorithmic discrimination • How humans mirror AI bias 90% of the time—even when they know it's biased • Why "computer says no" is becoming impossible to appeal
Featured Research: University of Washington studies on AI hiring bias, SafeRent discrimination lawsuit, expert insights from Professor Aylin Caliskan
Algorithm of the Week: SafeRent—the tenant screening AI that couldn't be overridden
Next week: "The Fine Print You Never Signed"—exploring the consent illusion in algorithmic decision-making.
Subscribe for the full Invisible AI series and learn how to fight back against algorithmic discrimination.
Support the show
More
What if an algorithm rejected you, and you never even knew it happened? In this explosive first episode of our four-part "Invisible AI" series, Dr. JR teams up with AI research assistant Ada to expose the hidden algorithmic systems making life-changing decisions about millions of Americans every day—in hiring, housing, credit, and beyond. In This Episode: • How 99% of Fortune 500 companies use AI to screen job applicants before humans ever see resumes • Why AI hiring tools prefer white-sounding names 85% of the time (University of Washington study) • The shocking truth about tenant screening algorithms that deny housing to qualified applicants • SafeRent's $2.28M settlement for algorithmic discrimination • How humans mirror AI bias 90% of the time—even when they know it's biased • Why "computer says no" is becoming impossible to appeal Featured Research: University of Washington studies on AI hiring bias, SafeRent discrimination lawsuit, expert insights from Professor Aylin Caliskan Algorithm of the Week: SafeRent—the tenant screening AI that couldn't be overridden Next week: "The Fine Print You Never Signed"—exploring the consent illusion in algorithmic decision-making. Subscribe for the full Invisible AI series and learn how to fight back against algorithmic discrimination. Support the show