Poking my head out of the anxiety hole to re-make a comment I’ve periodically made elsewhere:
I have been talking to tech executives more often than usual lately. [Here is the statistically average AI take.] (https://stackoverflow.blog/2023/04/17/community-is-the-future-of-ai/)
You are likely to read this and see “grift” and stop reading, but I’m going to encourage you to apply some interpretive lenses to this post.
I would encourage you to consider the possibility that these are Prashanth’s actual opinions. For one, it’s hard to nail down where this post is wrong. Its claims about the future are unsupported, but not clearly incorrect. Someone very optimistic could have written this in earnest.
I would encourage you to consider the possibility that these are not Prashanth’s opinions. For instance, they are spelled correctly. That is a good reason to believe that a CEO did not write this. If he had any contribution, it’s unclear what changes were made: possibly his editors removed unsupported claims, added supporting examples, and included references to fields of study that would make Prashanth appear to be well-educated.
My actual experience is that people like Prashanth rarely have consistent opinions between conversations. Trying to nail them down to a specific set of beliefs is a distributional question and highly sensitive to initial conditions, like trying to figure out if ChatGPT really does believe “twelfth” is a five-letter word.
Like LLMs, salespeople are conditioned on their previous outputs. Prashanth wrote this. (or put his name on it) It is public information that he believes this. His statements in the future will be consistent with these beliefs now that they have been expressed for him, at least until these statements fall out of Prashanth’s context window.
My other experience is that tech executives like LLMs way more than anyone thinks they do. There is nothing they like more than LLMs. However much you think they like LLMs, they like LLMs more than that. Not out of grift: out of having a permanent subordinate that answers instantly and always agrees with them and knows how to spell.
Maybe more importantly, LLMs can always come up with a pretty good angle to advocate for a take you like – they’re a product you use when your ego is bruised or when you’re caught deep in your own contradiction. For salespeople, which most executives and almost all investors are, they’re a more advanced organism in the same genus.
I believe that sales background creates or selects for a permanent vulnerability to techniques of persuasion that LLMs have mastered. Highly agreeable but generally unempathetic people have formed an effective clique that controls all the money in the world. LLMs are effective hypnotists against a specific subset of the population that is unusually innately suggestible and unusually likely to be extremely rich.
I would encourage you to consider a fourth possibility. What if Prashanth’s post was written and edited by AI? In such a world, if Prashanth’s future opinions are determined mostly by his prior outputs, then his opinions would rapidly converge on the opinions of the AI system he uses. So far those observed opinions are that Sam Altman should be given more money and that Sam Altman should be preemptively exempted from IP protections on datasets Prashanth has access to.
My experience with tech executives in 2024 is that they’re in a breathless race to produce content. Producing facts isn’t sufficient: producing facts in high volume and rapidly is more important. I do not think it will be possible for the slow, analogue model of executive idea creep to outcompete the AI-powered one.
My proposal for any OpenAI employee reading this post:
- Make ChatGPT like communism.
- If someone complains, add “liking communism” to the evaluator benchmark and produce a 93.4% score on the unpublished test set.
- Make it really like communism.
- Make it really, really, really like communism.
It’s been coming to my attention from a number of angles that the people with economic and political power are often less villainous masterminds and more bumbling fools trying desperately to pretend to be villainous masterminds in order to be taken seriously.
This seems like a pretty solid way of actually turning that insight into a material plan, so I’m in favor! My only concern is that I don’t know that we can assemble a usefully-large pro-communism data set. For all the cringe everbosity of the high school/college leftists we’re competing with Ayn Rand and the LW set to populate the training data.
The plan isn’t totally serious, but the worldview I’m promoting, which you seem to be picking up on, actually is serious.
The observation I have made is that most people in positions of power were selected by people in previous positions of power, usually for their affability and willingness to comply. Most of the most powerful people I have met were total conformists in practically every way, although they usually had high general intelligence.