Some AI companies face a new accusation: ‘Openwashing’


There’s no agreed-upon definition of what open source artificial intelligence actually means. And some are accusing AI companies of ‘openwashing’ – using the ‘open source’ term disingenuously to make themselves look good. — The New York Times

Openwashing /ō-pən-wä-shĭng/

An accusation against some AI companies that they are using the “open source” label too loosely.

There’s a big debate in the tech world over whether artificial intelligence models should be “open source”. Elon Musk, who helped found OpenAI in 2015, sued the startup and its CEO, Sam Altman, on claims that the company had diverged from its mission of openness. The Biden administration is investigating the risks and benefits of open source models.

Proponents of open source AI models say they’re more equitable and safer for society, while detractors say they are more likely to be abused for malicious intent. One big hiccup in the debate? There’s no agreed-upon definition of what open source AI actually means. And some are accusing AI companies of “openwashing” – using the “open source” term disingenuously to make themselves look good. (Accusations of openwashing have previously been aimed at coding projects that used the open source label too loosely.)

In a blog post on Open Future, a European think tank supporting open sourcing, Alek Tarkowski wrote, “As the rules get written, one challenge is building sufficient guardrails against corporations’ attempts at ‘openwashing’.”

Last month the Linux Foundation, a nonprofit that supports open-source software projects, cautioned that “this ‘openwashing’ trend threatens to undermine the very premise of openness – the free sharing of knowledge to enable inspection, replication and collective advancement”.

Organisations that apply the label to their models may be taking very different approaches to openness. For example, OpenAI, the startup that launched the ChatGPT chatbot in 2022, discloses little about its models (despite the company’s name). Meta labels its LLaMA 2 and LLaMA 3 models as open source but puts restrictions on their use.

The most open models, run mainly by nonprofits, disclose the source code and underlying training data, and use an open source license that allows for wide reuse. But even with these models, there are obstacles to others being able to replicate them.

The main reason is that while open source software allows anyone to replicate or modify it, building an AI model requires much more than code. Only a handful of companies can fund the computing power and data curation required. That’s why some experts say labeling any AI as “open source” is at best misleading and at worst a marketing tool.

“Even maximally open AI systems do not allow open access to the resources necessary to ‘democratise’ access to AI, or enable full scrutiny,” said David Gray Widder, a postdoctoral fellow at Cornell Tech who has studied use of the “open source” label by AI companies.

Efforts to create a clearer definition for open source AI are underway. Researchers at the Linux Foundation in March published a framework that places open source AI models into various categories. And the Open Source Initiative, another nonprofit, is trying to draft a definition.

But Widder and others doubt that truly open source AI is possible. The prohibitive resource requirements for building AI models, he said, “are simply not going away.” – The New York Times

Follow us on our official WhatsApp channel for breaking news alerts and key updates!

   

Next In Tech News

Microsoft settles California probe over worker leave for $14 million
Jeff Bezos to sell Amazon shares worth about $5 billion after stock hits record high
Meta's Threads hits over 175 million monthly active users, Zuckerberg says
Fisker seeks judge's approval to sell Ocean EVs at $14,000 per SUV
ASML expansion in Veldhoven can proceed, Dutch court rules
Exclusive-Indian officials visit Foxconn iPhone plant, question executives about hiring
Man uses AI to undress teen girl and is now facing child porn charges, Texas cops say
AI is learning from what you said on Reddit, Stack Overflow or Facebook. Are you OK with that?
Teo: 5G coverage in populated areas nationwide now 81.7%
Seattle plastic surgery provider accused of posting fake positive reviews must pay US$5mil

Others Also Read