Everyone knows the bit in The Matrix when Neo achieves digital messiah status and suddenly sees “reality” for what it really is: lines of trailing green code.

Well, thanks to an outage currently affecting Facebook, users of the social network have been given a similar peek behind the digital curtain, with many images on the site now replaced with the tags they’ve been assigned by the company’s machine vision systems.

So if you browse through your uploaded photos, instead of seeing holiday snaps or pictures of food and friends, you’ll be shown text saying things like “image may contain: people smiling, people dancing, wedding and indoor” or just “image may contain: cat.”

In short: this is how your life looks to a computer. This is how Facebook’s AI is judging you. Do you feel ashamed before the all-seeing digital eye?!

Oh yeah! I forgot Facebook uses machine learning to tag our photos with what it sees in the picture.



To be fair, "one person, beard" is pretty much a spot-on description of me. pic.twitter.com/fCpydUxtpz — Zack Whittaker (@zackwhittaker) July 3, 2019

The same image tags are showing up on Instagram, and as well as detailing general scene and object descriptions, they also suggest who is in a photo based on Facebook’s facial recognition. (The company has been doing this for photos you’re not tagged in since 2017.)

Facebook has been using machine learning to “read” images in this way since at least April 2016, and the project is a big part of the company’s accessibility efforts. Such tags are used to describe photos and videos to users with sight impairments.

There’s a lot of personal data contained in your photos that AI can extract

What’s not clear is whether Facebook also uses this information to target ads. There’s a lot of data about users’ lives that they might otherwise shield from Facebook contained in these images: whether you’ve got a pet, what your hobbies are, where you like going on holiday, or if you’re really into vintage cars, or swords, or whatever.

Back in 2017, one programmer was motivated by these questions to make a Chrome extension that revealed these tags. As they wrote at the time: “I think a lot of internet users don’t realize the amount of information that is now routinely extracted from photographs.”

Judging by the reactions on Twitter to this outage, this is certainly new information to a lot of people. We’ve reached out to Facebook to confirm whether they use this data for ad targeting and will update this story accordingly.

One nice thing about Facebook being utterly borked right now is it gives you a chance to see what its (very nice!) accessibility feature auto-alting images does.https://t.co/a3w9PcOosR pic.twitter.com/dnTXOih8Hp — Dieter Bohn (@backlon) July 3, 2019

Regardless of how this information is being used, though, it’s a fascinating peek behind the scenes at one of the world’s biggest data gathering operations. It also shows to what degree the visual world has become machine readable.

Improvements in deep learning in recent years have truly revolutionized the world of machine vision, and visual content online is often as legible to machines as text. But once something is legible, of course, it becomes easy to store, analyze, and extract data from. It’s only when the system breaks down, like today, that we realize it’s happening at all.