Satya Nadella says the explicit Taylor Swift AI fakes are ‘alarming and terrible’ – The Verge

Microsoft CEO Satya Nadella has responded to a controversy over sexually explicit AI-made fake images of Taylor Swift. In an interview with NBC Nightly News that will air next Tuesday, Nadella calls the proliferation of nonconsensual simulated nudes alarming and terrible, telling interviewer Lester Holt that I think it behooves us to move fast on this.

In a transcript distributed by NBC ahead of the January 30th show, Holt asks Nadella to react to the internet exploding with fake, and I emphasize fake, sexually explicit images of Taylor Swift. Nadellas response manages to crack open several cans of tech policy worms while saying remarkably little about them which isnt surprising when theres no surefire fix in sight.

I would say two things: One, is again I go back to what I thinks our responsibility, which is all of the guardrails that we need to place around the technology so that theres more safe content thats being produced. And theres a lot to be done and a lot being done there. But it is about global, societal you know, Ill say, convergence on certain norms. And we can do especially when you have law and law enforcement and tech platforms that can come together I think we can govern a lot more than we think we give ourselves credit for.

Microsoft might have a connection to the faked Swift pictures. A 404 Media report indicates they came from a Telegram-based nonconsensual porn-making community that recommends using the Microsoft Designer image generator. Designer theoretically refuses to produce images of famous people, but AI generators are easy to bamboozle, and 404 found you could break its rules with small tweaks to prompts. While that doesnt prove Designer was used for the Swift pictures, its the kind of technical shortcoming Microsoft can tackle.

But AI tools have massively simplified the process of creating fake nudes of real people, causing turmoil for women who have far less power and celebrity than Swift. And controlling their production isnt as simple as making huge companies bolster their guardrails. Even if major Big Tech platforms like Microsofts are locked down, people can retrain open tools like Stable Diffusion to produce NSFW pictures despite attempts to make that harder. Far fewer users might access these generators, but the Swift incident demonstrates how widely a small communitys work can spread.

There are other stopgap options like social networks limiting the reach of nonconsensual imagery or, apparently, Swiftie-imposed vigilante justice against people who spread them. (Does that count as convergence on certain norms?) For now, though, Nadellas only clear plan is putting Microsofts own AI house in order.

Go here to read the rest:

Satya Nadella says the explicit Taylor Swift AI fakes are 'alarming and terrible' - The Verge

Related Posts

Comments are closed.