A group of five U.S. senators are demanding OpenAI submit data showing how it plans to meet safety and security commitments that it has made about its artificial intelligence systems after a growing number of employees and researchers have raised red flags about the technology and the company’s safety protocols.
In a letter to OpenAI CEO Sam Altman, the four Democrats and one Independent lawmaker asked a series of questions on how the company is working to ensure AI cannot be misused to provide potentially harmful information—such as giving instructions on how to build weapons or assisting in the coding of malware—to members of the public. In addition, the group sought assurances that employees who raise potential safety issues would not be silenced or punished.
The concerns voiced by former employees have led to a flurry of media reports—and the senators expressed concerns about how the company is addressing safety concerns.
“We seek additional information from OpenAI about the steps that the company is taking to meet its public commitments on safety, how the company is internally evaluating its progress on those commitments, and on the company’s identification and mitigation of cybersecurity threats,” the five senators wrote in a letter obtained by The Washington Post.
Earlier this month, a whistleblower filed a complaint with the Securities and Exchange Commission (SEC), accusing OpenAI of preventing employees from warning officials about the risks of the company’s products. These allegedly included nondisclosure (and non-disparagement) agreements and the need to obtain prior consent from the company before discussing confidential information with federal regulators.
OpenAI did not reply to Fast Company’s request for comment about the lawmakers’ letter.
The whistleblower letter followed several public warnings from workers who opted to leave OpenAI, the maker of ChatGPT. Jan Leike, a coleader in the company’s superalignment group (dedicated to ensuring that AI stays aligned with the goals of its makers) who left the company in May, was harshly critical of Altman and the company, saying executives made it harder for him and his team to ensure the company’s AI systems aligned with human interests.
“Over the past few months, my team has been sailing against the wind. Sometimes we were struggling for compute [total computational resources] and it was getting harder and harder to get this crucial research done,” he wrote. “Building smarter-than-human machines is an inherently dangerous endeavor. OpenAI is shouldering an enormous responsibility on behalf of all of humanity.”
Months before Leike’s comments, another member of the superalignment team, Leopold Aschenbrenner, was fired for allegedly leaking information to journalists, according to The Information.
OpenAI has previously said it has removed non-disparagement terms from staff employment agreements, but the senators asked company officials to confirm that those will not be enforced for current or former employees and if OpenAI was willing to “commit to removing any other provisions from employment agreements that could be used to penalize employees who publicly raise concerns about company practices.”
The letter also requested additional information on a number of other issues, including:
- Whether OpenAI has procedures in place for when employees raise safety concerns
- Details about its security protocols
- Whether independent experts are allowed to test OpenAI systems before those systems are released
- A detailing of which patterns of misuse and safety risks the team has observed following the release of the most recent large language models
OpenAI was reportedly given a deadline of August 13 to supply the information requested by the senators.