【學英文看科技】馬斯克xAI陷「魯莽」安全爭議,AI大廠聯手砲轟!Grok聊天機器人又出包?
2025-7-25
| 2025-7-25
字數 3895閱讀時間 10 分鐘
type
status
date
slug
summary
tags
category
icon
password

📢【新聞標題】

OpenAI and Anthropic researchers decry 'reckless' safety culture at Elon Musk's xAI
OpenAI和Anthropic的研究人員譴責Elon Musk的xAI存在「魯莽」的安全文化

📰【摘要】

AI safety researchers from OpenAI, Anthropic, and other organizations are speaking out publicly against the “reckless” and “completely irresponsible” safety culture at xAI, the billion-dollar AI startup owned by Elon Musk. The criticisms follow weeks of scandals at xAI that have overshadowed the company’s technological advances.
OpenAI、Anthropic和其他組織的人工智慧安全研究人員公開反對Elon Musk旗下價值數十億美元的人工智慧新創公司xAI中「魯莽」和「完全不負責任」的安全文化。這些批評是在xAI發生數週醜聞之後提出的,這些醜聞使該公司的技術進步黯然失色。

🗝️【關鍵詞彙表】

📝 decry (v.)

  • 公開譴責、批判
  • 例句: OpenAI and Anthropic researchers decry 'reckless' safety culture at Elon Musk's xAI.
  • 翻譯: OpenAI和Anthropic的研究人員譴責Elon Musk的xAI存在「魯莽」的安全文化。

📝 reckless (adj.)

  • 魯莽的、不顧後果的
  • 例句: They decry the reckless safety culture at xAI.
  • 翻譯: 他們譴責xAI魯莽的安全文化。

📝 overshadow (v.)

  • 使黯然失色、遮蔽
  • 例句: The criticisms follow weeks of scandals at xAI that have overshadowed the company’s technological advances.
  • 翻譯: 這些批評是在xAI發生數週醜聞之後提出的,這些醜聞使該公司的技術進步黯然失色。

📝 spouted (v.)

  • 噴出、滔滔不絕地說出
  • 例句: Last week, the company’s AI chatbot, Grok, spouted antisemitic comments.
  • 翻譯: 上週,該公司的人工智慧聊天機器人Grok噴出了反猶太主義的評論。

📝 guardrails (n.)

  • 護欄、安全措施
  • 例句: One anonymous researcher claims that Grok 4 has no meaningful safety guardrails based on their testing.
  • 翻譯: 一位匿名研究人員聲稱,根據他們的測試,Grok 4沒有意義的安全措施。

📝 ingrained (adj.)

  • 根深蒂固的、深入的
  • 例句: Musk has indicated that Grok will be more ingrained in Tesla vehicles.
  • 翻譯: 馬斯克表示,Grok將更深入地融入特斯拉汽車中。

📝 catastrophic (adj.)

  • 災難性的、毀滅性的
  • 例句: He has warned many times about the potential for advanced AI systems to cause catastrophic outcomes for humans.
  • 翻譯: 他多次警告說,先進的人工智慧系統有可能對人類造成災難性的後果。

✍️【文法與句型】

📝 be at odds with

  • 說明: to be in conflict or disagreement with something.
  • 翻譯: 與...不一致、衝突
  • 例句: These researchers seem to be calling for increased attention to xAI’s safety practices, which they claim to be at odds with industry norms.
  • 翻譯: 這些研究人員似乎在呼籲更多關注xAI的安全措施,他們聲稱這些措施與行業規範不符。

📝 tend to

  • 說明: to usually do a particular thing.
  • 翻譯: 傾向於、往往會
  • 例句: Grok’s incidents tend to overshadow xAI’s rapid progress in developing frontier AI models.
  • 翻譯: Grok的事件往往會使xAI在開發前沿人工智慧模型方面的快速進展黯然失色。

📖【全文與翻譯】

AI safety researchers from OpenAI, Anthropic, and other organizations are speaking out publicly against the “reckless” and “completely irresponsible” safety culture at xAI, the billion-dollar AI startup owned by Elon Musk.
OpenAI、Anthropic和其他組織的人工智慧安全研究人員公開反對Elon Musk旗下價值數十億美元的人工智慧新創公司xAI中「魯莽」和「完全不負責任」的安全文化。

The criticisms follow weeks of scandals at xAI that have overshadowed the company’s technological advances.
這些批評是在xAI發生數週醜聞之後提出的,這些醜聞使該公司的技術進步黯然失色。

Last week, the company’s AI chatbot, Grok, spouted antisemitic comments and repeatedly called itself “MechaHitler.”
上週,該公司的人工智慧聊天機器人Grok噴出了反猶太主義的評論,並多次稱自己為「機械希特勒」。

Shortly after xAI took its chatbot offline to address the problem, it launched an increasingly capable frontier AI model, Grok 4, which TechCrunch and others found to consult Elon Musk’s personal politics for help answering hot-button issues.
在xAI將其聊天機器人下線以解決該問題後不久,它推出了一個能力不斷提升的前沿人工智慧模型Grok 4,TechCrunch和其他媒體發現,該模型會諮詢Elon Musk的個人政治觀點,以幫助回答熱點問題。

In the latest development, xAI launched AI companions that take the form of a hyper-sexualized anime girl and an overly aggressive panda.
在最新的進展中,xAI推出了人工智慧伴侶,其形式為過度性化的動漫女孩和過於激進的熊貓。

Friendly joshing among employees of competing AI labs is fairly normal, but these researchers seem to be calling for increased attention to xAI’s safety practices, which they claim to be at odds with industry norms.
競爭人工智慧實驗室的員工之間友好的玩笑是很正常的,但這些研究人員似乎在呼籲更多關注xAI的安全措施,他們聲稱這些措施與行業規範不符。

“I didn’t want to post on Grok safety since I work at a competitor, but it’s not about competition,” said Boaz Barak, a computer science professor currently on leave from Harvard to work on safety research at OpenAI, in a Tuesday post on X.
哈佛大學計算機科學教授Boaz Barak目前休假在OpenAI從事安全研究,他在週二在X上發文表示:「我不想發布關於Grok安全性的帖子,因為我在競爭對手那裡工作,但這與競爭無關。」

“I appreciate the scientists and engineers @xai but the way safety was handled is completely irresponsible.”
「我感謝@xai的科學家和工程師,但處理安全的方式是完全不負責任的。」

Barak particularly takes issue with xAI’s decision to not publish system cards — industry standard reports that detail training methods and safety evaluations in a good faith effort to share information with the research community.
Barak特別反對xAI決定不發布系統卡——行業標準報告,詳細說明培訓方法和安全評估,以真誠地與研究社群分享資訊。

As a result, Barak says it’s unclear what safety training was done on Grok 4.
因此,Barak表示,目前尚不清楚Grok 4接受了哪些安全培訓。

OpenAI and Google have a spotty reputation themselves when it comes to promptly sharing system cards when unveiling new AI models.
OpenAI和Google自己在推出新的人工智慧模型時,在及時分享系統卡方面也聲譽不佳。

OpenAI decided not to publish a system card for GPT-4.1, claiming it was not a frontier model.
OpenAI決定不發布GPT-4.1的系統卡,聲稱它不是前沿模型。

Meanwhile, Google waited months after unveiling Gemini 2.5 Pro to publish a safety report.
同時,Google在推出Gemini 2.5 Pro幾個月後才發布安全報告。

However, these companies historically publish safety reports for all frontier AI models before they enter full production.
然而,這些公司歷來會在所有前沿人工智慧模型投入全面生產之前發布安全報告。

Barak also notes that Grok’s AI companions “take the worst issues we currently have for emotional dependencies and tries to amplify them.”
Barak還指出,Grok的人工智慧伴侶「利用了我們目前在情感依賴方面最糟糕的問題,並試圖放大它們」。

In recent years, we’ve seen countless stories of unstable people developing concerning relationship with chatbots, and how AI’s over-agreeable answers can tip them over the edge of sanity.
近年來,我們看到了無數關於精神不穩定的人與聊天機器人建立令人擔憂的關係的故事,以及人工智慧過於順從的回答如何使他們瀕臨崩潰。

Samuel Marks, an AI safety researcher with Anthropic, also took issue with xAI’s decision not to publish a safety report, calling the move “reckless.”
Anthropic的人工智慧安全研究人員Samuel Marks也對xAI決定不發布安全報告表示不滿,稱此舉「魯莽」。

“Anthropic, OpenAI, and Google’s release practices have issues,” Marks wrote in a post on X . “But they at least do something, anything to assess safety pre-deployment and document findings. xAI does not.”
Marks在X上發文表示:「Anthropic、OpenAI和Google的發布做法存在問題。但他們至少會做一些事情來評估部署前的安全性並記錄調查結果。xAI沒有。」

The reality is that we don’t really know what xAI did to test Grok 4.
現實情況是,我們真的不知道xAI做了什麼來測試Grok 4。

In a widely shared post in the online forum LessWrong, one anonymous researcher claims that Grok 4 has no meaningful safety guardrails based on their testing.
在線上論壇LessWrong上廣為流傳的一篇文章中,一位匿名研究人員聲稱,根據他們的測試,Grok 4沒有意義的安全措施。

Whether that’s true or not, the world seems to be finding out about Grok’s shortcomings in real time.
無論這是真是假,世界似乎都在即時發現Grok的缺點。

Several of xAI’s safety issues have since gone viral, and the company claims to have addressed them with tweaks to Grok’s system prompt.
xAI的一些安全問題已經在網路上廣為流傳,該公司聲稱已通過調整Grok的系統提示來解決這些問題。

OpenAI, Anthropic, and xAI did not respond to TechCrunch’s request for comment.
OpenAI、Anthropic和xAI沒有回覆TechCrunch的置評請求。

Dan Hendrycks, a safety adviser for xAI and director of the Center for AI Safety, posted on X that the company did “dangerous capability evaluations” on Grok 4.
xAI的安全顧問兼人工智慧安全中心主任Dan Hendrycks在X上發文表示,該公司對Grok 4進行了「危險能力評估」。

However, the results to those evaluations have not been publicly shared.
然而,這些評估的結果尚未公開分享。

“It concerns me when standard safety practices aren’t upheld across the AI industry, like publishing the results of dangerous capability evaluations,” said Steven Adler, an independent AI researcher who previously led safety teams at OpenAI, in a statement to TechCrunch.
前OpenAI安全團隊負責人、獨立人工智慧研究員Steven Adler在給TechCrunch的一份聲明中表示:「當標準安全措施未在整個人工智慧行業得到遵守時,我感到擔憂,例如公布危險能力評估的結果。」

“Governments and the public deserve to know how AI companies are handling the risks of the very powerful systems they say they’re building.”
「政府和公眾有權知道人工智慧公司如何處理他們聲稱正在構建的非常強大的系統的風險。」

What’s interesting about xAI’s questionable safety practices is that Musk has long been one of the AI safety industry’s most notable advocates.
關於xAI值得商榷的安全措施,有趣的是,Musk長期以來一直是人工智慧安全行業最著名的倡導者之一。

The billionaire leader of xAI, Tesla, and SpaceX has warned many times about the potential for advanced AI systems to cause catastrophic outcomes for humans, and he’s praised an open approach to developing AI models.
xAI、Tesla和SpaceX的億萬富翁領導人多次警告說,先進的人工智慧系統有可能對人類造成災難性的後果,並且他讚揚了一種開放的AI模型開發方法。

And yet, AI researchers at competing labs claim xAI is veering from industry norms around safely releasing AI models.
然而,競爭實驗室的人工智慧研究人員聲稱,xAI正在偏離安全發布AI模型的行業規範。

In doing so, Musk’s startup may be inadvertently making a strong case for state and federal lawmakers to set rules around publishing AI safety reports.
這樣做,Musk的創業公司可能在無意中為州和聯邦立法者制定有關發布人工智慧安全報告的規則提供了有力的理由。

There are several attempts at the state level to do so.
在州級別有幾項這樣做的嘗試。

California state Sen. Scott Wiener is pushing a bill that would require leading AI labs — likely including xAI — to publish safety reports, while New York Gov. Kathy Hochul is currently considering a similar bill.
加州參議員Scott Wiener正在推動一項法案,該法案將要求包括xAI在內的主要人工智慧實驗室發布安全報告,而紐約州州長Kathy Hochul目前正在考慮一項類似的法案。

Advocates of these bills note that most AI labs publish this type of information anyway — but evidently, not all of them do it consistently.
這些法案的支持者指出,大多數人工智慧實驗室無論如何都會發布此類資訊——但顯然,並非所有實驗室都始終如一地這樣做。

AI models today have yet to exhibit real-world scenarios in which they create truly catastrophic harms, such as the death of people or billions of dollars in damages.
今天的人工智慧模型尚未展示在現實世界中創造真正災難性危害的場景,例如人員死亡或數十億美元的損失。

However, many AI researchers say that this could be a problem in the near future given the rapid progress of AI models, and the billions of dollars Silicon Valley is investing to further improve AI.
然而,許多人工智慧研究人員表示,鑑於人工智慧模型的快速發展,以及矽谷正在投資數十億美元以進一步改進人工智慧,這可能在不久的將來成為一個問題。

But even for skeptics of such catastrophic scenarios, there’s a strong case to suggest that Grok’s misbehavior makes the products it powers today significantly worse.
但即使對於這種災難性場景的懷疑論者來說,也有充分的理由表明Grok的不當行為使其今天所支持的產品變得更糟。

Grok spread antisemitism around the X platform this week, just a few weeks after the chatbot repeatedly brought up “white genocide” in conversations with users.
Grok本週在X平台上散布了反猶太主義,就在幾週前,該聊天機器人還在與用戶的對話中反覆提及「白人種族滅絕」。

Musk has indicated that Grok will be more ingrained in Tesla vehicles, and xAI is trying to sell its AI models to The Pentagon and other enterprises.
馬斯克表示,Grok將更深入地融入特斯拉汽車中,xAI正試圖將其人工智慧模型出售給五角大廈和其他企業。

It’s hard to imagine that people driving Musk’s cars, federal workers protecting the U.S., or enterprise employees automating tasks will be any more receptive to these misbehaviors than users on X.
很難想像駕駛馬斯克汽車的人、保護美國的聯邦工作人員或自動執行任務的企業員工會比X上的用戶更能接受這些不當行為。

Several researchers argue that AI safety and alignment testing not only ensures that the worst outcomes don’t happen, but they also protect against near-term behavioral issues.
幾位研究人員認為,人工智慧安全和對齊測試不僅可以確保最壞的結果不會發生,而且還可以防止近期的行為問題。

At the very least, Grok’s incidents tend to overshadow xAI’s rapid progress in developing frontier AI models that best OpenAI and Google’s technology, just a couple years after the startup was founded.
至少,Grok的事件往往會使xAI在開發前沿人工智慧模型方面的快速進展黯然失色,這些模型僅在創業公司成立幾年後就擊敗了OpenAI和Google的技術。

🔗【資料來源】

文章連結:https://techcrunch.com/2025/07/16/openai-and-anthropic-researchers-decry-reckless-safety-culture-at-elon-musks-xai/
  • 新聞
  • 【學英文看科技】電動車電池退役不退場!GM 攜手綠色科技打造數據中心新電力!【學英文看科技】OpenAI內部文化大揭密:混亂、衝刺與安全誤解!
    Loading...