On facial recognition, surveillance, privacy, and the art project that tests whether we've given up.
A digital art experiment that sends people a fake facial recognition result using their own photo — and nobody questions where it came from.
Passwords can be reset. Credit cards can be cancelled. Your face is forever. When companies scrape billions of photos, that data never expires.
The EU has the AI Act. Illinois has BIPA. At the federal level? Nothing. Here's why your face has fewer legal protections than your email.
AI models are trained to make beautiful images. But for a surveillance experiment to work, the photos need to look like they came from someone's Facebook in 2013.
A 34% error rate for dark-skinned women. A 1% error rate for white men. The systems get deployed anyway.
Clearview AI built the world's largest facial recognition database by scraping every public photo it could find. You never consented. You probably can't opt out.
Airport gates scan your face. Stores track your shopping. Concert venues match you against databases. Nobody opted in. Almost nobody opts out.
From CV Dazzle face paint to AI-generated protest art, a growing movement of artists is turning surveillance tools into weapons of resistance.