OpenAI outlines AI safety plan, allowing board to reverse decisions
DeeperDive is a beta AI feature. Refer to full articles for the facts.
ARTIFICIAL intelligence (AI) company OpenAI laid out a framework to address safety in its most advanced models, including allowing the board to reverse safety decisions, according to a plan published on its website on Monday (Dec 18).
Microsoft-backed OpenAI will only deploy its latest technology if it is deemed safe in specific areas such as cybersecurity and nuclear threats. The company is also creating an advisory group to review safety reports and send them to the company’s executives and board. While executives will make decisions, the board can reverse those decisions.
Since ChatGPT’s launch a year ago, the potential dangers of AI have been top of mind for both AI researchers and the general public. Generative AI technology has dazzled users with its ability to write poetry and essays, but also sparked safety concerns with its potential to spread disinformation and manipulate humans.
In April, a group of AI industry leaders and experts signed an open letter calling for a six-month pause in developing systems more powerful than OpenAI’s GPT-4, citing potential risks to society. A May Reuters/Ipsos poll found that more than two-thirds of Americans are concerned about the possible negative effects of AI and 61 per cent believe it could threaten civilisation. REUTERS
Decoding Asia newsletter: your guide to navigating Asia in a new global order. Sign up here to get Decoding Asia newsletter. Delivered to your inbox. Free.
Share with us your feedback on BT's products and services
TRENDING NOW
‘Boring’ is the new black: The stars are aligning for a Singapore stock market revival
Near sell-out launches in March boost developer sales to 1,300 units after four slow months
China pips the US if Asean is forced to choose, but analysts warn against reading it like a sports result
Genting Singapore’s Lim Kok Thay receives S$7.5 million pay package for FY2025