Sam Altman departs OpenAI's safety committee
Sam Altman has resigned from OpenAI's Safety and Security Committee, which will become an independent oversight board. The company faces scrutiny over safety policies and is seeking to raise over $6.5 billion.
Read original articleSam Altman, CEO of OpenAI, has stepped down from the company's Safety and Security Committee, which was established in May to oversee critical safety decisions regarding OpenAI's projects. Following his departure, the committee will transition into an independent oversight board led by Zico Kolter from Carnegie Mellon University, with members including Quora's Adam D'Angelo and retired General Paul Nakasone. The committee will continue to monitor safety assessments and has the authority to delay product releases if safety issues arise. Altman's exit comes amid scrutiny from U.S. senators regarding OpenAI's safety policies and a significant turnover in staff focused on AI's long-term risks. Critics, including former board members, have expressed concerns about OpenAI's ability to self-regulate due to profit motives. OpenAI is reportedly seeking to raise over $6.5 billion, which may lead to a shift away from its hybrid nonprofit structure aimed at limiting investor returns. This move raises questions about the company's commitment to its founding mission of developing AI for the benefit of humanity.
- Sam Altman has left OpenAI's Safety and Security Committee.
- The committee will become an independent oversight board chaired by Zico Kolter.
- OpenAI is under scrutiny for its safety policies and has seen significant staff turnover.
- Critics doubt OpenAI's ability to self-regulate due to profit incentives.
- The company is reportedly raising over $6.5 billion, potentially altering its nonprofit structure.
Related
Former OpenAI employee quit to avoid 'working for the Titanic of AI'
A former OpenAI employee raised concerns about the company's direction, likening it to the Titanic. Departures, lawsuits, and founding rival companies highlight challenges in balancing innovation and safety in AI development.
OpenAI promised to make its AI safe. Employees say it 'failed' its first test
OpenAI faces criticism for failing safety test on GPT-4 Omni model, signaling a shift towards profit over safety. Concerns raised on self-regulation effectiveness and reliance on voluntary commitments for AI risk mitigation. Leadership changes reflect ongoing safety challenges.
US lawmakers send a letter to OpenAI requesting government access
US lawmakers have urged OpenAI to enhance safety standards, allocate resources for AI safety research, and allow pre-deployment testing, following whistleblower allegations and concerns about AI risks and accountability.
Sam Altman is becoming one of the most powerful people on Earth. Be afraid
Sam Altman, CEO of OpenAI, faces scrutiny over his influence in AI, lobbying against regulations, and potential conflicts of interest, raising concerns about transparency and ethical considerations in the industry.
Nearly half of OpenAI's Safety Researchers have left
Nearly half of OpenAI's AGI safety team has left, including key figures, raising concerns about prioritizing product development over safety and the impact of commercial interests on AGI research.
Related
Former OpenAI employee quit to avoid 'working for the Titanic of AI'
A former OpenAI employee raised concerns about the company's direction, likening it to the Titanic. Departures, lawsuits, and founding rival companies highlight challenges in balancing innovation and safety in AI development.
OpenAI promised to make its AI safe. Employees say it 'failed' its first test
OpenAI faces criticism for failing safety test on GPT-4 Omni model, signaling a shift towards profit over safety. Concerns raised on self-regulation effectiveness and reliance on voluntary commitments for AI risk mitigation. Leadership changes reflect ongoing safety challenges.
US lawmakers send a letter to OpenAI requesting government access
US lawmakers have urged OpenAI to enhance safety standards, allocate resources for AI safety research, and allow pre-deployment testing, following whistleblower allegations and concerns about AI risks and accountability.
Sam Altman is becoming one of the most powerful people on Earth. Be afraid
Sam Altman, CEO of OpenAI, faces scrutiny over his influence in AI, lobbying against regulations, and potential conflicts of interest, raising concerns about transparency and ethical considerations in the industry.
Nearly half of OpenAI's Safety Researchers have left
Nearly half of OpenAI's AGI safety team has left, including key figures, raising concerns about prioritizing product development over safety and the impact of commercial interests on AGI research.