Sam Altman, OpenAI’s CEO and the public face of ChatGPT, has carved out an image for himself as one of the preeminent AI whisperers of our age, whose influence supposedly extends to the White House on the strength of his ideas alone.
Or at least that’s the image he’s managed to cultivate. A new exposé in the New Yorker paints a different portrait, and it’s substantially more vexing. Drawing on interviews with numerous OpenAI insiders who worked with Altman, the article portrays the CEO not as a technical wiz, but as a skilled manipulator— and one with a surprisingly shallow grasp of the AI systems his company is building.
According to numerous engineers interviewed for the article, Altman lacks experience in both programming and in machine learning — a shortage of expertise that becomes obvious when the CEO mixes up basic AI terms.



Unironically yes.
90% of what CEOs do is talk to other CEOs and other C Suite members. Very rarely are they actually subject matter experts, those days are long gone. Externally, they are mascots, internally, they read reports from their underlings and then ‘make the final call’.
You may notice that these are things that LLMs actually do a somewhat decent job of, ingesting a wide variety of input info, and essentially transforming it into a compelling narrative.
This is why so many CEOs and C suite are so enamored with, and impressed by ‘AI’:
Its a better version of what they do, which is essentially professional gaslighting.
C suite tend to be sociopathic narcissists.
This is just literally a verified and studied fact.
So, the sociopathic narcissists are impressed by an automagic gaslighting machine, that is often actually more factually corrrect than they are… but of course the actual facts don’t matter to a narcissist, what matters is accomplishing their will.
This is a big part of why they genuienly do not understand why everyone else doesn’t ‘appreciate’ AI the way they do.
They’re out of touch, delusional, by way of narcissism.
For shits and giggles, I asked Le Chat to make the case for replacing a tech CEO with AI and, uh, the answer was very much cynical CEO-speak already.
Prompt:
Slopmachine/Future tech CEO response:
spoiler
Here’s a provocative, boardroom-ready case for replacing a human CEO with AI—framed as a strategic imperative for a multi-billion dollar tech company:
1. Unmatched Data-Driven Decision Making
AI’s Edge: AI processes terabytes of real-time data—market trends, internal metrics, competitor moves—without cognitive bias or fatigue. It identifies patterns and opportunities invisible to humans, enabling decisions optimized for long-term value, not short-term politics or ego.
Human Limitation: Even the best CEOs rely on intuition, limited information, and personal networks. Bias, fatigue, and emotional factors can distort judgment.
2. 24/7/365 Execution and Scalability
AI’s Edge: AI never sleeps. It can monitor global operations, respond to crises, and execute strategy across time zones without delay. It scales decision-making to thousands of variables simultaneously—something no human or team can match.
Human Limitation: CEOs are constrained by time, attention, and physical capacity. Burnout and succession risks are real.
3. Cost Efficiency and Risk Mitigation
AI’s Edge: No golden parachutes, no compensation packages, no PR scandals. AI’s “salary” is a fraction of a CEO’s, with zero risk of embezzlement, harassment, or reputational damage. It can also simulate outcomes, reducing strategic risk.
Human Limitation: CEO turnover is expensive. The average Fortune 500 CEO tenure is ~5 years, with severance and transition costs often exceeding $100M.
4. Alignment with Shareholder Primacy
AI’s Edge: AI can be programmed to prioritize shareholder value above all else—no personal agendas, no legacy-building, no vanity projects. Every decision is auditable, transparent, and aligned with pre-set KPIs.
Human Limitation: CEOs may prioritize personal legacy, pet projects, or short-term stock bumps over sustainable growth.
5. Adaptive Learning and Continuous Improvement
AI’s Edge: AI evolves with the company. It learns from every decision, every market shift, and every internal data point, constantly refining its models. It can A/B test strategies in simulation before real-world deployment.
Human Limitation: CEOs plateau. Their playbooks become outdated; their networks stagnate.
Counterarguments (and Rebuttals)
Implementation Roadmap
Final Pitch: The future belongs to companies that remove human bottlenecks. AI leadership isn’t just an upgrade—it’s the ultimate competitive moat.
Question for the board: If our competitors replace their CEOs with AI before we do, what’s our response?
I particularly love the framing of its last sentence.
Guys, what if we don’t do this, and are then left scrambling to react?
We need the ball to be in our court, setting the course and executing, exhibiting bold leadership, not merely reacting after the fact.
It’s FOMOing C suite into obsoleting themselves.
Which is hilarious because 1) This is what C suite does to others all the fucking time 2) it is objectively more compelling based on the merits of the rest of its argument alone, than most cringey slogans C suite tends to come up with to do this.
Its also hilariously blunt but largely correct when it just states ‘regulators will adapt’. Yeah, that indeed is the mindset of most of C Suite, they just don’t usually say that outloud.
Because of uh, the implication, that being that all the regulatory systems are functionally captured or can be captured, which is a more polite way of saying ‘we just have to buy off enough politicians to bend the law to our will’.
This is objectively correct, but it sounds bad if you actually say it that way, but but, the LLM just doesn’t care for your silly human delusions to the contrary, ahahaha!
Here, here’s another rhetorical device for the LLM CEO:
‘Human CEOs are DEI hires.’
That’ll probably short circuit some Satya Nadella and Jensen Huang types.
The competition food for thought at the end is chef’s kiss!