this post was submitted on 17 Nov 2023
1 points (100.0% liked)
Machine Learning
1 readers
1 users here now
Community Rules:
- Be nice. No offensive behavior, insults or attacks: we encourage a diverse community in which members feel safe and have a voice.
- Make your post clear and comprehensive: posts that lack insight or effort will be removed. (ex: questions which are easily googled)
- Beginner or career related questions go elsewhere. This community is focused in discussion of research and new projects that advance the state-of-the-art.
- Limit self-promotion. Comments and posts should be first and foremost about topics of interest to ML observers and practitioners. Limited self-promotion is tolerated, but the sub is not here as merely a source for free advertisement. Such posts will be removed at the discretion of the mods.
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Normally in most companies, the responsibility of the board is around representing the interests of the shareholders; for example, ensuring that the financial statements are a genuine representation of the value of the business. And usually, a statement about a lack of candidness from the CEO is corporate speak for "the CEO is using company money for personal gaiin", which again is corporate speak for "we found the CEO is guilty of embezzlement, but we don't want to ruin the prosecution's case and risk defamation by saying that out loud".
But, it could also be anything that materially affects shareholder value that the CEO isn't being honest about. For example, if Sam Altman knew that November's GPT-4 was disastrously worse at programming than previous releases, but told the board that everything was fine, and that it had all checked out as being perfect... that could also count as "lack of candidness". He would have to have done this repeatedly for the board to swoop in; some sort of "this is the final straw" after last week's announcements.
That all said, OpenAI's board does have a wider remit: they are also have a reponsibility around safe-guarding the use of AI and a few other things like that. So if OpenAI had actually built an AGI and Sam Altman was lying about it, they would also have a responsibility to fire him.
In every example I can think of, this announcement means that Sam Altman has been lying about something. In my experience, dishonest executive leadership almost always slows research and development down; so OpenAI has been succeeded despite Sam Altman's leadership, rather than because of it. So I predict that AI research speeds up even more now. (Uh oh.)