Hello! We are officially launching a THING. It’s going to be a weekly roundup about what’s happening in artificial intelligence and how it affects you.
It’s no secret that Silicon Valley’s business model revolves around hoovering up a disgusting amount of consumer data and selling it off to the highest bidder (usually our own government). If you use the internet, you are the product—this is “surveillance capitalism” 101. But, after Zoom’s big terms-of-service debacle earlier this week, there are some signs that surveillance capitalism may be shape-shifting into some terrible new beast—thanks largely to AI.
Zoom was brutally pilloried earlier this week for a change to its terms of service. That change actually happened back in March, but people didn’t really notice the new policy until this week, when a blogger pointed out the shift in a post that went viral on Hacker News. The change, which came at the height of AI’s hype frenzy, gave Zoom an exclusive right to use user data to train future AI modules. More specifically, Zoom claimed a right to a “perpetual, worldwide, non-exclusive, royalty-free, sublicensable, and transferable license” to users’ data which, it was interpreted, included the contents of videoconferencing data. Suffice it to say, the backlash was swift and thunderous, and the internet really spanked the company.
Since the initial storm clouds have passed, Zoom has promised that it isn’t, in fact, using videoconferencing data to train AI and has even updated its terms of service (again) to make this explicitly clear. But whether Zoom is gobbling up your data or not, this week’s controversy clearly indicates an alarming new trend in which companies are now using all the data they’ve collected via “surveillance capitalism” to train nascent artificial intelligence products.
They’re then turning around and selling those AI services back to the very same users whose data helped build the products in the first place, thus creating an endless, self-propagating loop. It makes sense that companies are doing this, since any fleeting mention of the term “AI” now sends tech company investors and shareholders into a tizzy. Still, the biggest offenders here are companies that already own vast swaths of the world’s information, making it a particularly creepy and legally weird situation. Google, for instance, has recently made it known that it is scraping the web to train its new AI algorithms. Big AI vendors like OpenAI and MidJourney, meanwhile, have also vacuumed up most of the internet in an effort to amass enough data. Helpfully, the Harvard Business Review just published a “how-to” guide for companies who want to transform their collected data troves into new AI algorithm juice, so I’m sure we can expect more offenders in the future.
So, uh, just how worried should we be about this noxious brew of digital privacy violations and automation? Katharine Trendacosta, director of policy and advocacy at the Electronic Frontier Foundation (and a former Gizmodo employee), told Gizmodo she doesn’t necessarily think that generative AI is accelerating surveillance capitalism. That said, it’s not de-accelerating it, either.
“I don’t know if it [surveillance capitalism] can be more turbocharged, quite frankly—what more can Google possibly have access to?” she says. Instead, AI is just giving companies like Google one more way to monetize and utilize all the data they’ve amassed.
“The problems with AI have nothing to do with AI,” Trendacosta says. The real problem is the regulatory vacuum around these new technologies, which allows companies to wield them in a blindly profit-driven, obviously unethical way. “If we had a privacy law, we wouldn’t have to worry about AI. If we had labor protections, we would not have to worry about AI. All AI is a pattern recognition machine. So it’s not the specifics of the technology that is the problem. It is how it is used and what is fed into it.”
As often as possible, we’re going to try to update readers on the state of AI regulation (or lack thereof). Given the hugely disruptive potential of this technology, it just makes sense that governments should pass some new laws. Will they do that? Eh...
The Federal Election Commission can’t decide whether AI generated content in political advertising is a problem or not. A petition sent to the agency by the advocacy group Public Citizen has asked it to consider regulating “deepfake” media in political ads. This week, the FEC decided to advance the group’s petition, opening up the potential rule-making to a public comment period. In June, the FEC deadlocked on a similar petition from Public Citizen, with some regulators “expressing skepticism that they had the authority to regulate AI ads,” the Associated Press reports. The advocacy group was then forced to come back with a new petition that laid out to the federal agency why it did in fact have the jurisdiction to do so. Some Republican regulators remain unconvinced of their own authority—maybe because the GOP has, itself, been having a field day with AI in political ads. If you think AI shouldn’t be used in political advertising, you can write to the FEC via its website.
Last week, a small consortium of big players in the AI space—namely, OpenAI, Anthropic, Google, and Microsoft—launched the Frontier Model Forum, an industry body designed to guide the AI boom while also offering up watered down regulatory suggestions to governments. The forum, which says it wants to “advance AI safety research to promote responsible development of frontier models and minimize potential risks,” is based upon a weak regulatory vision promulgated by OpenAI itself. The so-called “frontier AI” model, which was outlined in a recently published study, focuses on AI “safety” issues and makes some mild suggestions for how governments can mitigate the potential impact of automated programs that “could exhibit dangerous capabilities.” Given how well Silicon Valley’s self-regulation model has worked for us so far, you’d certainly hope that our designated lawmakers would wake up and override this self-serving, profit-driven legal roadmap.
You can compare the U.S.’s predictably sleepy-eyed acquiescence to corporate power to what’s happening across the pond where Britain is in the process of prepping for a global summit on AI that it’ll be hosting. The summit also follows on the fast-paced development of the European Union’s “AI Act,” a proposed regulatory framework that carves out modest guardrails for commercial artificial intelligence systems. Hey America, take note!
This week, a number of media conglomerates penned an open letter urging that regulations be passed. The letter, signed by Gannet, the Associated Press, and a number of other U.S. and European media companies, says they “support the responsible advancement and deployment of generative AI technology, while believing that a legal framework must be developed to protect the content that powers AI applications as well as maintain public trust in the media that promotes facts and fuels our democracies.” Those in the media have good reason to be wary of new automated technologies. News orgs (including the ones who signed this letter) have been trying to position themselves as best they can to a new industry that seems liable to eat traditional news media.
We’ve all heard that the robots are coming to steal our jobs and there’s been a lot of chatter about whose head will be on the chopping block first. But another question worth asking is: who is least likely to be laid off and replaced by a corporate algorithm? The answer apparently is: barbers. That answer comes from a recently published Pew Research report that looked at the jobs considered most “exposed” to artificial intelligence (meaning they’re most likely to be automated). In addition to barbers, the people most unlikely to be replaced by a chatbot include dishwashers, child care workers, firefighters, and pipe layers, according to the report. Web developers and budget analysts, meanwhile, are at the top of AI’s hit list.
Occasionally, we’re going to include an interview with a notable AI proponent, critic, wonk, kook, entrepreneur, or other such person who is connect to the field. We thought we’d start off with Sarah Myers West, who has led a very decorated career in artificial intelligence research. In between academic careers, she recently served as a consultant on AI for the Federal Trade Commission and, these days, serves as managing director of the AI Now Institute, which advocates for industry regulation. This week, West and others released a new strategy for AI regulation dubbed the “Zero Trust” model, which advocates for strong federal action to safeguard against the more harmful impacts of AI. This interview has been lightly edited for brevity and clarity.
You’ve been researching artificial intelligence for quite some time. How did you first get interested in this subject? What was appealing (or alarming) about it? What got you hooked?
My background is as a researcher studying the political economy of the tech industry. That’s been the primary focus of my core work over the last decade, tracking how these big tech companies behave. My earlier work focused on the advent of commercial surveillance as a business model of networked technologies. The sorta “Cambrian” moment of AI is in many ways a byproduct of those dynamics of commercial surveillance—it sorta flows from there.
I also heard that you were a big fan of Jurassic Park when you were younger. I feel like that story’s themes definitely relate a lot to what’s going on with Silicon Valley these days. Relatedly, are you Source: Gizmodo