June 11

Employees And Experts Warn Of The Dangers Of AI

AI, Web3

An open letter signed by more than a dozen current and former employees warns of the dangers of AI and implores the government and companies to do something about it. Titled "A Right To Warn About Advanced Artificial Intelligence," the letter features thirteen signatures from current and former employees of OpenAI, Google DeepMind, and Anthropic.

Six of signers chose to remain anonymous, and four of those still work at OpenAI. According to The Washington Post, the letter was also endorsed by Yoshuo Bengio and Geoffrey Hinton, often considered the "godfathers" of AI, as well as Stuart Russell, a renowned computer scientist.

The Dangers Of AI At The Forefront Of Open Letter

The letter, hosted on a new site dubbed "Right To Warn," isn't particularly long, but it's well-documented and comes with a list of four primary demands. The authors begin by saying, "we believe in the potential of AI technology to deliver unprecedented benefits to humanity."

But then it gets serious.

The letter details three main major risks of artificial intelligence. These include "further entrenchment of existing inequalities, manipulation and misinformation, [and] the loss of control of autonomous AI systems potentially resulting in human extinction."

We've already seen some of the dangers of AI outlined here playing out in social media and online retailers. However, the most drastic fear listed here — human extinction — likely refers not to where artificial intelligence is, but where several of these companies, including OpenAI, ultimately want to take the technology.

Beyond what we've already seen play out, from questionable ethics to employee exoduses to whirlwind board decisions, perhaps the most concerning thing about artificial intelligence is what companies may do in pursuit of advancing it. The authors of the letters argue that, because all of these companies are motivated by profits and lack government oversight, AI will only become more unsafe without stringent protocols. 

They propose four main agreements that will help company employees, well, protect humanity. 

The Four Agreements

The letter's authors argue that company employees are currently the only safeguards against egregious missteps in AI advancements. As such, they ask that all companies developing technology adopt these four agreements to help employees curb the dangers of AI.

  1. That the company will not enter into or enforce any agreement that prohibits “disparagement” or criticism of the company for risk-related concerns, nor retaliate for risk-related criticism by hindering any vested economic benefit;
  2. That the company will facilitate a verifiably anonymous process for current and former employees to raise risk-related concerns to the company’s board, to regulators, and to an appropriate independent organization with relevant expertise;
  3. That the company will support a culture of open criticism and allow its current and former employees to raise risk-related concerns about its technologies to the public, to the company’s board, to regulators, or to an appropriate independent organization with relevant expertise, so long as trade secrets and other intellectual property interests are appropriately protected;
  4. That the company will not retaliate against current and former employees who publicly share risk-related confidential information after other processes have failed. We accept that any effort to report risk-related concerns should avoid releasing confidential information unnecessarily. Therefore, once an adequate process for anonymously raising concerns to the company’s board, to regulators, and to an appropriate independent organization with relevant expertise exists, we accept that concerns should be raised through such a process initially. However, as long as such a process does not exist, current and former employees should retain their freedom to report their concerns to the public.

Basically, these employees want to make sure that other employees have an opportunity to bring forward serious concerns without threats of retaliation. It sounds a little dystopian, but it's a very real concern, especially at OpenAI

What It Means For The Future Of AI

The letter, dated June 4th, 2024, came less than one week before Apple announced its own form of AI, dubbed "Apple Intelligence." In its announcement, the company notably went in on making sure Apple's version of the technology would be safe and not trained on other peoples' data — only your own. In a way, they basically just want to make personal assistants better. 

But the company also announced that OpenAI's ChatGPT would be integrated into some Siri responses, following similar agreements with companies like Microsoft. If anything, it just means that now the biggest player in the tech hardware space is seemingly "all-in" on AI, the dangers of AI certainly aren't going anywhere. 

But the more noise experts make, the more likely we are to see serious concern and interest from parties that can create regulations and guardrails to make sure AI doesn't become a thing of dystopian action movies. A Pew Research poll revealed that 52 percent of people are more concerned than excited about AI, with only 10 percent of respondents saying they're more excited than concerned. 

A combination of both insider distress and public concern should, in theory, fast-track protections against overly ambitious tech companies with a "shoot first and ask questions later" attitude towards developing the technology. But slow developments so far, at least in the U.S., have made it clear that there's no silver bullet to protect against the potential dangers of AI. 


MORE STORIES FOR YOU

You Should Start A Podcast: Interview With Greg Wasserman

What do you call three men sitting around in a room? A podcast. It’s a cheap joke (heck, we gave it to you for free!), but it says a lot about the perception of podcasting. Like, for instance, how easy it really is to set up. How aimless many podcasts feel. And, yes, how the field is

Read More

RootNote Free Individual Plan Now Available

Content creator data startup RootNote has just launched a free version of the company’s flagship product designed to help content creators, small businesses, influencers, and anybody who relies on digital content as part of their brand. The launch marks a significant step in the company’s mission to make analytics accessible and actionable to everyone. “We believe

Read More

Deepfakes And Protecting Your Likeness

Deepfakes are once again in the headlines, this time thanks to a recent ruling and a generative AI bot with no guardrails. But it’s not just about artificially generated media, either — some creators are finding themselves defending their original content being used by huge companies without credit or compensation. What Are Deepfakes?Deepfakes are artificially generated

Read More

The Artist Corporation Wants To End Hustle Culture

The artist corporation is a new corporate structure that aims to end hustle culture and give creative people the type of security and flexibility enjoyed by other corporations for decades. But can it stick in a creator economy that is perhaps more volatile than ever? The artist corporation, or “A-Corp,” is an experimental new structure being

Read More

Never miss a good story!

 Subscribe to our newsletter to keep up with what's going on in content creation!