AI companies have been widely criticized for potential harm to users, but mass surveillance and weapons development would clearly take that to a new level. Anthropic's potential reply to the Pentagon was seen as a test of its claim to be the most safety-forward AI company, particularly after dropping its flagship safety pledge a few days ago. Now that Amodei has responded, the focus will shift to the Pentagon to see if it follows through on its threats, which could seriously harm Anthropic.
Фонбет Чемпионат КХЛ
。业内人士推荐im钱包官方下载作为进阶阅读
这些公司有一个共同点:它们没有在造更聪明的模型,它们在解决数据从混乱到有序的问题。
OpenAI would retain control over how technical safeguards are implemented and which models are deployed and where, and would limit deployment to cloud environments rather than “edge systems.” (In a military context, edge systems are a category that could include aircraft and drones.) In what would be a major concession, Altman told employees that the government said it is willing to include OpenAI’s named “red lines” in the contract, such as not using AI to power autonomous weapons, conduct domestic mass surveillance, or engage in critical decision-making.
They never did. The flag disappeared eventually, though I received no confirmation.