In July, a cutesy face-morphing app threw most Americans with a smartphone into utter panic. The goofy photo filter, appropriately (if unimaginatively) dubbed FaceApp, attracted hordes of users with its ability to add years of eerily-realistic age onto photographed faces. It rocketed into Internet virality, its edited photos sweeping across social media platforms and photo-sharing sites. It was a blast -- until some suspicious internet investigator realized that the firm behind FaceApp was Russian. Chaos erupted. Across social sites and message boards, people began fretting: who are these Russians, why do they want our faces, and what could they do with the data they took? ; one Democratic politician . All around the country, people wondered whether shady Russian foreign intelligence agents were using their duck-faced selfies to hone their nefarious facial-recognition programs. Media newscasters issued warnings called for an investigation As it turns out, the connection between the firm behind FaceApp and the Russian government was a tad overblown -- that is to say, nonexistent. An investigation conducted by the Washington Post reported that the firm has , does not enable unauthorized surveillance, and is not using the photos it stores to hone a facial-recognition program. The whole episode was a tempest in a digital teapot. no ties to international intelligence The uproar, however, is hardly surprising. The FaceApp debacle is only the latest in a series of concerns stemming from online services that seem fun, convenient, and exploitative. We wonder, is the ability to automatically tag ourselves in group photos on Facebook letting the company profile, track, and take advantage of us? When we walk into a cashier-less Amazon store, will the retail giant somehow use the surveillance footage it gathers against us? Mostly, we don’t know -- and that’s a problem. It raises the question: what can current AI programs do with our faces, exactly? How many of our fears are based in myth, rather than truth? Do the technology’s actual capabilities stack up to our concerns? Let’s consider. Facial Recognition -- What Is It, Exactly? Before we delve into the risks of facial recognition, it’s crucial to understand how such programs work. Facial recognition technology (FRT) functions primarily by based on measurements of our nodal points; these include but are not limited to the width of the nose, the length of the jaw, the distance between our eyes, etc. These measurements are transcribed into a template with a unique code, which can then be compared to and potentially matched to known photographs within a database. Casinos, for example, might use FRT to cross-reference photos of incoming guests to internal files that . Other investigators can source their photo references from driver’s license databases, mug shots, government records, or even social media platforms like Facebook. However, the breadth of their photo pool -- and, as we’ll get into later, the accuracy of their search -- depends on how many sources they have the authority to use. identifying and creating templates list problem gamblers What Can (and Can’t) FRT Do? Contrary to what a binge-watch of almost all spy dramas would suggest, FRT rarely returns accurate results instantaneously or within complex environments. The cliche of a government group being able to find a person of interest in grainy street surveillance footage within seconds is mostly overstated. While facial recognition surveillance is indeed possible, it is nevertheless fraught with inaccuracies and subject to environmental factors. Camera quality, algorithm, time of day, distance, database size, and even demographic points like race and gender can . Memorably, one test conducted by the ACLU in 2018 reported that Amazon’s FRT service, Rekognition, with police mugshots. A different investigation deployed by the MIT Media Lab found that the same program mistaking women for men 19% of the time and misidentifying women with darker skin for men over a third of the time. Amazon as the reason behind the inaccuracies. impact the effectiveness of FRT searches falsely matched 28 members of Congress struggled to identify gender, pointed to poor calibration Amazon may be right, at least in part. After all, researchers have found that than the FBI, given that the social media platform has considerably more photos to reference and actively asks its users to refine its algorithm by verifying its auto-tagging results. This leads us to the misunderstanding behind Rekognition, and indeed all FRT software: it’s only ever as good as the reference pool it uses. Facebook is likely more accurate in its facial identification As one Amazon spokesperson , “While 80% confidence is an acceptable threshold for photos of hot dogs, chairs, animals, or other social media use cases, it wouldn’t be appropriate for identifying individuals with a reasonable level of certainty.” A fair point, although the comment is somewhat ironic given that the company has shopped Rekognition to law enforcement and ICE and yet not instituted a mandatory degree of FRT confidence. The consistent inaccuracy has made some barriers to how FRT can be deployed in enforcement necessary. For example, the technology , but it can be used to develop leads during investigations. told the Verge cannot yet be used as the sole basis for an arrest Make no mistake, though; while FRT in the United States is not currently capable of picking out a target on a dark street via grainy surveillance photos, it is likely only a matter of time before it gains the ability to do so. China, a country known for its focus on citizen surveillance, recently made headlines for tracking down a BBC reporter via FRT . This sets a precedent for how the technology may be used in enforcement in the United States someday. within a mere seven minutes One point is clear: while facial recognition technology in the United States is not yet consistently accurate, there are entities within the country that are dedicated to improving and using it. For now, the problem we face is twofold; first, that people may face undeserved censure or detainment due to a faulty identification from unreliable technology. Second, that a surveillance trend may, as in China, develop as a result of the technology becoming both more accurate and widespread. Where Does This Leave Us, the FaceApp Users? The development of FRT programs is inevitable. While they may not be entirely accurate or well-used now, they will be one day. In the meantime, we need to take steps to establish how, when, and to what degree these technologies can be used -- and set hard limits now, before the usage of FRT spreads beyond the point of easy regulatory restriction. As put the matter: “Today, facial recognition may be pleasantly useful when it can admit you to a baseball game, but may seem far less so if a distant database thinks, on the basis of preprogrammed visual assumptions, that you are likely to be criminally violent.” one writer for the Washington Post Thus, the place we, as FaceApp users, stand is more philosophical than anything else. A goofy app that adds realistic wrinkles to our faces might be fun, but the data it collects could ultimately be used against us. Sure, those photos might languish on a Russian server somewhere, untouched and unused -- or they could be used as reference points to make a more effective and accurate FRT system that can be used to surveil civilians. This episode points at the heart of our digital insecurities and pushes us to ask; are we harming ourselves in the long run by being so trusting with even our most innocuous data? The lesson here? The next time a goofy photo filter or online game or digital trend goes viral, take a beat to consider whether you want to participate. Even if you think your duck-face selfie is languishing on a server in digital -- or literal -- Siberia, assume that someone’s using it. Because, in all honesty, unless someone sparks a viral panic over it, you’ll really never know.