Crypto News

AI Risks Spark Concern Among OpenAI, Anthropic, Google DeepMind Staff

Former and current employees from leading AI firms like OpenAI and Google DeepMind raise alarm over AI risks, urging transparency and internal safety protocols amid growing concerns.
AI Risks Spark Concern Among OpenAI, Anthropic, Google DeepMind Staff

Highlights

  • AI industry insiders advocate for the "Right to Warn AI" petition to address risks.
  • Concerns rise over AI's potential to spread misinformation and exacerbate inequalities.
  • Transparency urged as employees call for open dialogue on AI risks within companies.

A group of current and former employees from AI companies, including OpenAI, Google DeepMind, and Anthropic, have expressed concerns about the potential risks associated with AI technologies’ rapid development and deployment.

The problems, outlined in an open letter, range from the spread of misinformation to the possible loss of control over autonomous AI systems and even to the dire possibility of human extinction.

Advertisement

OpenAI, Google DeepMind, Anthropic Staff AI Concerns

13 former and current employees at artificial intelligence (AI) developers OpenAI (ChatGPT), Anthropic (Claude), and DeepMind (Google), along with the “Godfathers of AI” Yoshua Bengio and Geoffrey Hinton and AI scientist Stuart Russell, have initiated a “Right to Warn AI” petition. The petition aims to establish a commitment from frontier AI companies to allow employees to raise risk-related concerns about AI internally and with the public.

In the open letter, the authors explain that due to the financial motives, AI companies focus on product creation rather than its safety. The signatories state that these financial incentives compromise the supervision process and that AI companies have limited legal requirements to disclose information about their systems’ strengths and weaknesses to governments.

The letter also focuses on the current status of AI regulation and argues that the companies cannot be trusted to share essential data.

Subsequently,  they claim that the threats presented by AI without proper regulation, such as dissemination of fake news and the worsening of inequality, call for a more active and responsible approach to AI innovation and application.

Advertisement

Safety Concerns and Calls for Change

The employees have requested changes within the AI industry and have asked companies to implement a system where current and former employees can report their issues concerning risk. They also suggest that AI firms should not impose non-disclosure agreements that prevent criticism, so that people can express concerns about the dangers of AI technologies.

William Saunders, a former OpenAI employee, said,

“Today, those who understand the most about how the cutting-edge AI systems function and the potential dangers associated with their use are not able to share their insights freely because they are afraid of the consequences and non-disclosure agreements are too restrictive.”

The letter is issued at a time when there are concerns within the AI field about the safety of highly sophisticated AI systems. There are already cases when image generators from OpenAI and Microsoft are creating photos with disinformation about voting, although such content is prohibited.

At the same time, there are concerns that AI safety is being ‘de-prioritised,’ especially in the pursuit of AGI that seeks to develop software that can mimic human cognition and learning.

Advertisement

Company Responses and Controversies

OpenAI, Google, and Anthropic still need to address the issues raised by the employees. Nevertheless, OpenAI has stressed the importance of safety and the proper discussion regarding AI technologies. The company has seen internal issues, such as the disbanding of its Superalignment safety team, which has made people doubt the company’s commitment to safety.

Nevertheless, as noted by Coingape earlier, OpenAI created a new Safety and Security Committee to make important decisions and improve the safety of AI as the company advances.

Despite this, some former board members have accused OpenAI management of inefficiency, particularly in terms of the organization’s approach to safety issues. In a podcast, a former board member, Helen Toner, disclosed that OpenAI CEO Sam Altman was allegedly fired for not sharing information with the board.

Read Also: Can GameStop (GME) Price Lead Meme Coin Rally To $1?

Advertisement
Share
Kelvin Munene Murithi

Kelvin Munene is a crypto and finance journalist with over 5 years of experience, offering in-depth market analysis and expert commentary . With a Bachelor's degree in Journalism and Actuarial Science from Mount Kenya University, Kelvin is known for his meticulous research and strong writing skills, particularly in cryptocurrency, blockchain, and financial markets. His work has been featured across top industry publications such as Coingape, Cryptobasic, MetaNews, Cryptotimes, Coinedition, TheCoinrepublic, Cryptotale, and Analytics Insight among others, where he consistently provides timely updates and insightful content. Kelvin’s focus lies in uncovering emerging trends in the crypto space, delivering factual and data-driven analyses that help readers make informed decisions. His expertise extends across market cycles, technological innovations, and regulatory shifts that shape the crypto landscape. Beyond his professional achievements, Kelvin has a passion for chess, traveling, and exploring new adventures.

Why trust CoinGape: CoinGape has covered the cryptocurrency industry since 2017, aiming to provide informative insights to our readers. Our journalists and analysts bring years of experience in market analysis and blockchain technology to ensure factual accuracy and balanced reporting. By following our Editorial Policy, our writers verify every source, fact-check each story, rely on reputable sources, and attribute quotes and media correctly. We also follow a rigorous Review Methodology when evaluating exchanges and tools. From emerging blockchain projects and coin launches to industry events and technical developments, we cover all facets of the digital asset space with unwavering commitment to timely, relevant information.
Investment disclaimer: The content reflects the author’s personal views and current market conditions. Please conduct your own research before investing in cryptocurrencies, as neither the author nor the publication is responsible for any financial losses.
Ad Disclosure: This site may feature sponsored content and affiliate links. All advertisements are clearly labeled, and ad partners have no influence over our editorial content.

Recent Posts

  • Crypto News

Arthur Hayes Predicts Bitcoin Rally To $500K By Next Year Over Fed Easing

BitMEX co-founder Arthur Hayes has predicted that Bitcoin's price could rally to $500,000 by the…

November 29, 2025
  • Crypto News

China Tightens Stance on Stablecoin and Crypto Payments With New Policy Talks

China is tightening its stance on digital assets again with regulators warning of rising risks…

November 29, 2025
  • Crypto News

Hyperliquid Team Moves $90M HYPE as Network Becomes Top Fee Chain

Hyperliquid logged another major on-chain development today after a team-linked wallet shifted $90 million worth…

November 29, 2025
  • Crypto News

Is Kalshi Manipulating Prediction Markets? Platform Hit With Lawsuit Over Violations

Kalshi has been accused of illegal sports gambling and manipulating the market. This comes as…

November 29, 2025
  • Crypto News

Coinbase Submits Recommendations to CFTC on Crypto Market Rules

The policy recommendations are detailed requests that Coinbase filed with the Commodity Futures Trading Commission…

November 29, 2025
  • Crypto News

21Shares XRP ETF To Begin Trading on Monday as Institutional Inflows Hit $666 Million

21Shares’ XRP ETF is set to join the list of funds already trading in the…

November 29, 2025