Fight against nonconsensual sexual deepfakes is slowly getting somewhere


Public concern over the issue exploded earlier this year when people used generative AI tools to create explicit content featuring Taylor Swift. — The New York Times

People worry a lot about the various flavours of havoc that AI might wreak in the future, but one of the technology’s here-and-now problems is that of deepfakes, in particular those that depict actual women and girls in sexualised ways without their consent.

Public concern over the issue exploded earlier this year when people used generative AI tools to create explicit content featuring Taylor Swift. This kind of thing wasn’t really new – faked pornographic images have been a problem ever since Photoshop became a thing – but AI apps can make it particularly easy to do, and Swift’s unusually high profile ensured widespread consciousness.

Today, Meta’s independent (but Meta-funded) Oversight Board announced its findings on two deepfake cases that it took on in April. One was about an image on Facebook depicting a female American public figure in the nude, being groped – it is not clear if this figure was Swift, as the board didn’t name the people concerned. The other image, on Instagram, showed a female Indian public figure in the nude, depicted from behind.

The Board, which is sometimes referred to as Meta’s “Supreme Court”, gets the final say when it comes to specific cases such as these. It said today that Meta had done the right thing by taking down the picture of the American figure, despite the protestations of the poster.

However, Meta had initially ignored users’ complaints about the image of the Indian figure, only taking it down when the Board took on the case. Here, the Board decided Meta had been in the wrong.

Beyond specific cases, the Board can only make non-binding recommendations to Meta about its future content moderation. So, Meta will now have to make up its own mind about the Board’s deepfake recommendations, which mostly center on the company’s policy banning “derogatory sexualised photoshop”.

The Board reckons the word “derogatory” should be replaced by “non-consensual”, and the work “photoshop” by something that reflects the changing nature of the tech that people are using to make these images. The Board also wants Meta to start using the fact that content is AI-generated or AI-manipulated as a signal for lack of consent in its "Adult Sexual Exploitation" policy.

Meta said it would review the recommendations.

Of course, a social media company can’t do much more to those creating or sharing such content other than to ban them. So it’s also notable that, yesterday, the US Senate unanimously passed a bill that would let people portrayed in sexually explicit deepfakes sue the creators.

The Disrupt Explicit Forged Images and Non-Consensual Edits Act (DEFIANCE) Act would give victims the ability to claim up to US$150,000 in damages, plus an extra US$100,000 if the deepfake was connected to “actual or attempted sexual assault, stalking, or harassment”. It has a companion bill waiting in the House, and Senate Majority Leader Chuck Schumer (D-N.Y.) urged his counterparts there to take it up.

“By passing this bill, we are telling victims of explicit nonconsensual deepfakes that we hear them and we are taking action,” he said. But the House doesn’t have much time, as the August recess is just over a week away.

Over in the UK, the sharing of such content was already criminalised in last year’s Online Safety Act. The former Conservative government was set to make creation a crime too, but all new bills were dropped when the Conservatives called the election that led to their ouster at the start of this month.

Nonetheless, the new Labour government has also said it wants to criminalise the making of nonconsensual sexualised deepfakes – Deputy Prime Minister Angela Rayner has herself been a victim. Labour’s advisers have also recommended banning apps that are dedicated to making such imagery.

It's unclear how easy these laws will be to enforce, especially as the proliferation of open-source AI models makes it easy for people to set up “nudification” tools without the guardrails that AI providers like OpenAI put in place. But the threat of severe consequences, and action by social media companies, will at the very least make it clearer to people that this kind of behavior is socially unacceptable. – Fortune.com/The New York Times

Follow us on our official WhatsApp channel for breaking news alerts and key updates!

Deepfake

   

Next In Tech News

Sirius XM found liable in New York lawsuit over subscription cancellations
US Supreme Court tosses case involving securities fraud suit against Facebook
Amazon doubles down on AI startup Anthropic with $4 billion investment
Factbox-Who are bankrupt Northvolt's creditors?
UK should use new powers to probe Apple-Google mobile browser duopoly, report says
EU regulators scrap probe into Apple's e-book rules after complaint was withdrawn
Hyundai recalls over 145,000 electrified US vehicles on loss of drive power
'World of Warcraft' still going strong as it celebrates 20 years
Northvolt CEO steps down, saying group needs up to $1.2 billion
Bitcoin at record highs, sets sights on $100,000

Others Also Read