AI firms mustn’t govern themselves
For humanity’s sake, regulation is needed to tame market forces, argue two former members of OpenAI’s board
CAN private companies pushing forward the frontier of a revolutionary new technology be expected to operate in the interests of both their shareholders and the wider world? When we were recruited to the board of OpenAI – Tasha in 2018 and Helen in 2021 – we were cautiously optimistic that the company’s innovative approach to self-governance could offer a blueprint for responsible artificial intelligence (AI) development.
But based on our experience, we believe that self-governance cannot reliably withstand the pressure of profit incentives. With AI’s enormous potential for both positive and negative impact, it’s not sufficient to assume that such incentives will always be aligned with the public good. For the rise of AI to benefit everyone, governments must begin building effective regulatory frameworks now.
If any company could have successfully governed itself while safely and ethically developing advanced AI systems, it would have been OpenAI. The organisation was originally established as a non-profit with a laudable mission: to ensure that AGI, or artificial general intelligence – AI systems that are generally smarter than humans – would benefit “all of humanity”. Later, a for-profit subsidiary was created to raise the necessary capital, but the non-profit stayed in charge.
Decoding Asia newsletter: your guide to navigating Asia in a new global order. Sign up here to get Decoding Asia newsletter. Delivered to your inbox. Free.
Share with us your feedback on BT's products and services
TRENDING NOW
Middle East-linked energy supply shocks put Asean Power Grid back in focus
From intern to C-suite: JPMorgan’s Teresa Heitsenrether on building a fully AI-powered ‘megabank’
Sunway ‘sniffing around’ for opportunities: Tycoon Jeffrey Cheah wants business to last 10 generations
Keppel begins arbitration against partners in Vietnam JV hit by 6.9 trillion dong bill