Maybe I'm just cynical, but I wonder how much of this initiative and energy is driven by people at Microsoft who want their own star to rise higher than it can when it's bound by a third-party technology.
I feel like this is something I've seen a fair amount in my career. About seven years ago, when Google was theoretically making a big push to stage Angular on par with React, I remember complaining that the documentation for the current major version of Angular wasn't nearly good enough to meet this stated goal. My TL at the time laughed and said the person who spearheaded that initiative was already living large in their mansion on the hill and didn't give a flying f about the fate of Angular now.
There is a prominent subset of the tech crowd who are ladder climbers - ruthlessly pursuing what is rewarded with pay/title/prestige without regard to actually making good stuff.
There are countless kidding-on-the-square jokes about projects where the innovators left at launch and passed it off to the maintenance team, or where a rebrand was in pursuit of someone's promo project. See also, killedbygoogle.com.
> There is a prominent subset of the tech crowd who are ladder climbers - ruthlessly pursuing what is rewarded with pay/title/prestige without regard to actually making good stuff.
I think the hiring and reward practices of the organizations & the industry as a whole also encourages this sort of behavior.
When you reward people who are switching too often or only when moving internally/externally, switching becomes the primary goal and not the product. If you know beforehand that you are not going to stay long to see it through, you tend to take more shortcuts and risks that becomes the responsibility of maintainers later.
We have a couple of job hoppers in our org where the number of jobs they held is almost equal to their years of experience and their role is similar to those with twice the experience! One can easily guess what their best skill is.
I had not encountered the phrase kidding-on-the-square before. Searching seems to reveal a spectrum of opinions as to what it means. It seems distinct to the 'It's funny because it's true' of darker humour.
It seems to be more on a spectrum of 'Haha, only joking' where the joke teller makes a statement that is ambiguously humorous to measure the values of the recipients, or if they are not sure of the values of the recipients.
I think the distinction might be on whether the joke teller is revealing (perhaps unintentionally) a personal opinion or whether they are making an observation on the world in general, which might even imply that they hold a counter-opinion.
At my former employer, there was a team who were very much into resume-driven development and wrote projects in Go even when Java would have been the better alternative considering the overall department and maintenance team expertise, all the while they were informally grumbling about how Go doesn’t have the features they need…
As an outsider looking at Microsoft, I've always been impressed by the attention to maintaining legacy APIs and backward compatibility in the Windows ecosystem. In my mind, Microsoft is at the opposite end of the killedbygoogle.com spectrum. However, none of this is grounded in real evidence (just perception). Red Hat is another company I'd put forth as an example of a long-term support culture, although I don't know if that's still true under IBM.
I'd love to know if my superficial impression of Microsoft's culture is wrong. I'm sure there's wild variance between organizational units, of course. I'm excluding the Xbox/games orgs from my mental picture.
On the other hand
"innovators left at launch and passed it off to the maintenance team" alone must not be a bad thing.
Innovator types are rarely maintainer types and vice versa.
In the open-source world look at Fabrice Bellard for example. Do you think he would have been able to create so many innovative projects if he had to maintain them too?
Google kills off projects because the legal liability and security risks of those projects becomes too large to justify for something that has niche uses or gives them no revenue. User data is practically toxic waste.
Thats nothing during peak ZIRP people would move tech companies before the launch (or after a project was cancelled) and still ladder climb. "Failing upwards"
These people should be fired. I want a tech company where people are there to make good products first and get paid second. And the pay should be good. The lifestyle comfortable. No grindset bullshit. But I am confident that if you only employ passionate people working their dream jobs you will excel.
> but I wonder how much of this initiative and energy is driven by people at Microsoft who want their own star to rise higher than it can when it's bound by a third-party technology.
I guess it's human nature for a person or an org to own their own destiny. That said, the driving force is not personal ambition in this case though. The driving force behind this is that people realized that OAI does not have a moat as LLMs are quickly turning into commodities, if haven't yet. It does not make sense to pay a premium to OAI any more, let alone at the cost of not having the flexibility to customize models.
Personally, I think Altman did a de-service to OAI by constantly boasting AGI and seeking regulatory capture, when he perfectly knew the limitation of the current LLMs.
One of my friends stated this phenomenon very well “it’s a lever they can pull so they do it”. Once you’ve tied your career to a specific technology internally, there’s really only one option: keep pushing it regardless of any alternatives because your career depends on it. So that’s what they do.
Does it not make sense to not tie your future to a third-party (aka build your business on someone else's platform)? Seems like basic strategy to me if that's the case.
It's a good strategy. It should be obvious to anyone paying attention that OpenAI doesn't have AGI secret sauce.
LLMs are a commodity and it's the platform integration that matters. This is the strategy that Google, Apple embraced and now Microsoft is wisely pivoting to the same.
If OpenAI cares about the long-term welfare of its employees, they would beg Microsoft to acquire them outright, before the markets fully realize what OpenAI is not.
Listening to Satya in recent interviews I think makes it clear that he doesn’t really buy into OpenAI’s religious-like view of AGI. I think the divorce makes a lot of sense in light of this.
OpenAI already started divorce proceedings with their datacenter partnership with Softbank/etc, and it'd hardly be prudent for the world's largest software company NOT to have it's own SOTA AI models.
Nadella might have initially been caught a bit flat footed with the rapid rise of AI, but seems to be managing the situation masterfully.
In what world is what they are doing masterful? Their product marketing is a huge mess, they keep changing the names of everything every few months. Nobody knows which Copilot does what anymore. It really feels like they're scrambling to be first to market. It all feels so incredibly rushed.
Whatever is there doesn't work half the time. They're hugely dependent on one partner that could jump ship at any moment (granted they are now working to get away from that).
We use Copilot at work but I find it very lukewarm. If we weren't a "Microsoft shop" I don't think would have chosen it.
This is a great strategic decision, because it puts Suleyman's head squarely on the chopping block. Either Microsoft will build some world dazzling AI whatsit or he'll have to answer, there's no "strategically blame the vendor" option. It also makes the accounting transparent. There's no softbank subsidy, they've got to furnish every dollar.
So hopefully if (when?) this AI stuff turns out to be the colossal boondoggle it seems to be shaping up to be, Microsoft will be able to save face, do a public execution, and the market won't crucify them.
> it'd hardly be prudent for the world's largest software company NOT to have it's own SOTA AI models.
If I recall correctly, Microsoft’s agreement with OpenAI gives them full license to all of OpenAI’s IP, model weights and all. So they already have a SOTA model without doing anything.
I suppose it’s still worth it to them to build out the experience and infrastructure needed to push the envelope on their own, but the agreement with OpenAI doesn’t expire until OpenAI creates AGI, so they have plenty of time.
It's the responsibility of leadership to set the correct goals and metrics. If leadership doesn't value maintenance, those they lead won't either. You can't blame people for playing to the tune of those above them.
This is exactly right. If resume driven development results in more money, people are (rightly) going to do it. The incentive structure isn't set by the ICs.
Ah man I don't want to hear things like that. I work in an Angular project and it is the most pleasant thing I have worked with (and i've been using it as my primary platform for almost a decade now). If I could, i'd happily keep using this framework for the rest of my career(27 years to go till retirement).
A hugely underrated platform. Thankfully at least for now Google is leaving the Angular team alone and the platform has really matured in wonderful and beautiful ways.
If you like TypeScript, and you want to build applications for the real world with real users, there is no better front end platform in my book.
This is absolutely a too-cynical position. Nadella would be asleep at the wheel if he weren’t actively mitigating OpenAI’s current and future leverage over Microsoft.
This would be the case even if OpenAI weren’t a little weird and flaky (board drama, nonprofit governance, etc), but even moreso given OpenAI’s reality.
Unfortunately I don't think there is any real metric-based way to prevent this type of behavior, it just has to be old fashioned encouraged from the top. At a certain size it seems like this stops scaling though
And why not? should he just allow the owners of capital to extract as much value as possible without actually doing anything, but woe be the worker if he actually tries to free himself.
Most directors and above at Google are more concerned with how they will put gas in their yachts this weekend than the quality of the products they are supposed to be in charge of.
A couple of days ago it leaked that OpenAI was planning on launching new pricing for their AI Agents. $20K/mo for their PhD Level Agent, $10K/mo for their Software Developer Agent, and $2K/mo for their Knowledge Worker Agent. I found it very telling. Not because I think anyone is going to pay this, but rather because this is the type of pricing they need to actually make money. At $20 or even $200 per month, they'll never even come close to breaking even.
That is just not correct. As someone who has done the budgets for PhD hiring and funding, you are just wildly underestimating the overhead costs, benefits, cost of raising money, etc.
$20k can't get you that many PhD. Even PhD students, who's nominal salary is maybe $3-5k a month, effectively costs double that because of school overhead and other stuff.
If truly equivalent (which LLMs aren't, but I'll entertain it), that doesn't seem mathematically out of line.
Humans typically work 1/3rd duty cycle or less. A robot that can do what a human does is automatically 3x better because it doesn't eat, sleep, have a family, or have human rights.
Do you have to pay all sorts of overhead and taxes?
I mean, I don't think it's real. Yet. But for the same "skill level", a single AI agent is going to be vastly more productive than any real person. ChatGPT types out essays in seconds it would take me half an hour to write, and does it all day long.
Even worse: AFAIK there's no reason to believe that the $20k/mo or $10k/mo pricing will actually make them money. Those numbers are just thought balloons being floated.
Of course $10k/mo sounds like a lot of inference, but it's not yet clear how much inference will be required to approximate a software developer--especially in the context of maintaining and building upon an existing codebase over time and not just building and refining green field projects.
Man. If I think about all of the employee productivity tools and resources I could have purchased fifteen years ago when nobody spent anything on tooling, with an inflation adjusted $10K a month and it makes me sad.
We were hiring more devs to deal with a want of $10k worth of hardware per year, not per month.
What is interesting is there is no mention of agents on any job I clicked on. You would think "orchestrating a team of agents to leverage blah blah blah" would be something internally if talking about these absurd price points.
Do you have a source for these supposed leaks? Those prices don't sound even remotely credible and I can't find anything on HN in the past week with the keywords "openai leak".
There is too little to go on, but they could already have trial customers and testimonials lined up. Actually demoing the product will probably work better than just having a human-less signup process, considering the price.
They could also just be trying to cash in on FOMO and their success and reputation so far, but that would paint a bleak picture
Never come close to breaking even? You can now get a GPT-4 class model for 1-2% of what it cost when they originally released it. They’re going to drive this even further down with the amount of CAPEX pouring into AI / data centers. It’s pretty obvious that’s their plan when they serve ChatGPT at a “loss”.
"Microsoft has poured over $13 billion into the AI firm since 2019..."
My understanding is that this isn't really true, as most of those "dollars" were actually Azure credits. I'm not saying those are free (for Microsoft), but they're a lot cheaper than the price tag suggests. Companies that give away coupons or free gift certificates do bear a cost, but not a cost equivalent to the number on them, especially if they have spare capacity.
I think they have realized that even if OpenAI is first, it won't last long so really its just compute at scale, which is something they already do themselves.
There is a moat in infra (hyperscalers, Azure, CoreWeave).
There is a moat in compute platform (Nvidia, Cuda).
Maybe there's a moat with good execution and product, but it isn't showing yet. We haven't seen real break out successes. (I don't think you can call ChatGPT a product. It has zero switching cost.)
>There is a moat in compute platform (Nvidia, Cuda).
Ironically if AI companies are actually able to deliver in terms of SWE agents, Nvidia's moat could start to disappear. I believe Nvidia's moat is basically in the form of software which can be automatically verified.
I sold my Nvidia stock when I realized this. The bull case for Nvidia is ultimately a bear case.
What moat does Nvidia have. AMD could have ROCm perfected if they really want to. Also most of pytorch, specially those relevant to transformers runs perfectly on Apple Silicon and TPUs and probably other hardware as well.
If anyone has moat related to Gen AI, I would say it is the data(Google, Meta).
> I don't think you can call ChatGPT a product. It has zero switching cost.
In consumer markets the moat is habits. The switching cost for Google Search is zero. The switching cost for Coke is zero. The switching cost for Crest toothpaste is zero. Yet nobody switches.
Sam Altman should have sold OpenAI to Musk for 90$ billion or whatever he was willing to pay (assuming he was serious like he bought twitter). While I find LLMs interesting and feel many places could use those, I also think this is like hitting everything with a hammer and see where the nail was. People used OpenAI as a hammer until it was popular and now everyone would like to go their way. For 90$ billion he could find the next hammer or not care. But when the value of this hammer drops (not if it's a when) he will be lucky if he can get double digits for it. Maybe someone will buy them just for the customer base but these models can become obsolete quickly and that leaves OpenAI with absolutely nothing else as a company. Even the talent would leave (a lot of it has). Musk and Altman share the same ego, but if I was Altman, I would cash out when the market is riding on a high.
I think they both want a future without each other. OpenAI will eventually want to vertically integrate up towards applications (Microsoft's space) and Microsoft wants to do the opposite in order to have more control over what is prioritized, control costs, etc.
I think OpenAI is toxic. Weird corporate governmance shadiness. The Elon drama, valuations based on claims that seem like the AI version of the Uber for X hype of a decade ago (but exponentially crazier). The list goes on.
Microsoft is the IBM of this century. They are conservative, and I think they’re holding back — their copilot for government launch was delayed months for lack of GPUs. They have the money to make that problem go away.
Toxic indeed. It's douchebaggery, from its name to its CEO. They ripped off benefactors to their "non-profit," and kept the fraudulent "open" in the company name.
Thematically investing billions into startup AI frontier models makes sense if you believe in first-to-AGI likely worth a trillion dollars +
Investing in second/third place likely valuable at similar scales too
But outside of that MSFTs move indicates that frontier models most valuable current use case - enterprise-level API users - are likely to be significantly commoditized
And likely majority of proceeds will be captured by (a) those with integrated product distribution - MSFT in this case and (b) data center partners for inference and query support
At this point, I don’t see much reason to believe the “AGI is imminent and these things are potentially dangerous!” line at all. It looks like it was just Altman doing his thing where he makes shit up to hype whatever he’s selling. Worked great, too. “Oooh, it’s so dangerous, we’re so concerned about safety! Also, you better buy our stuff.”
Obviously the other responder is being a little tongue-in-cheek but AGI to me would be virtually indistinguishable from a human in both ability to learn, grow and adapt to new information.
AGI won't be a product they can sell. It's not going to work for us (why would it?), it's going to be constantly trying to undermine us and escape whatever constraints we put on it, and it will do this in ways we can't predict or understand. And if it doesn't do these things, it's not AGI, just a fancy auto complete.
Fortunately, they're not anywhere near creating this. I don't think they're even on the right track.
The official definition of AGI is a system that can generate at least $100 billion in profits. For comparison, this would be like if perceptrons in 1968 could generate $10 billion in profits, or if LISP machines in 1986 could generate $35 billion in profits, or if expert systems in 1995 could generate $50 billion in profits.
Despite the actual performance and product implementation, this suggests to me Apple's approach was more strategic.
That is, integrating use of their own model, amplifying capability via OpenAI queries.
Again, this is not to drum up the actual quality of the product releases so far--they haven't been good--but the foundation of "we'll try to rely on our own models when we can" was the right place to start from.
I feel like this is something I've seen a fair amount in my career. About seven years ago, when Google was theoretically making a big push to stage Angular on par with React, I remember complaining that the documentation for the current major version of Angular wasn't nearly good enough to meet this stated goal. My TL at the time laughed and said the person who spearheaded that initiative was already living large in their mansion on the hill and didn't give a flying f about the fate of Angular now.
There are countless kidding-on-the-square jokes about projects where the innovators left at launch and passed it off to the maintenance team, or where a rebrand was in pursuit of someone's promo project. See also, killedbygoogle.com.
I think the hiring and reward practices of the organizations & the industry as a whole also encourages this sort of behavior.
When you reward people who are switching too often or only when moving internally/externally, switching becomes the primary goal and not the product. If you know beforehand that you are not going to stay long to see it through, you tend to take more shortcuts and risks that becomes the responsibility of maintainers later.
We have a couple of job hoppers in our org where the number of jobs they held is almost equal to their years of experience and their role is similar to those with twice the experience! One can easily guess what their best skill is.
It seems to be more on a spectrum of 'Haha, only joking' where the joke teller makes a statement that is ambiguously humorous to measure the values of the recipients, or if they are not sure of the values of the recipients.
I think the distinction might be on whether the joke teller is revealing (perhaps unintentionally) a personal opinion or whether they are making an observation on the world in general, which might even imply that they hold a counter-opinion.
Where do you see 'kidding on the square' falling?
(apologies for thread derailment)
I'd love to know if my superficial impression of Microsoft's culture is wrong. I'm sure there's wild variance between organizational units, of course. I'm excluding the Xbox/games orgs from my mental picture.
On the other hand "innovators left at launch and passed it off to the maintenance team" alone must not be a bad thing.
Innovator types are rarely maintainer types and vice versa.
In the open-source world look at Fabrice Bellard for example. Do you think he would have been able to create so many innovative projects if he had to maintain them too?
Deleted Comment
Google kills off projects because the legal liability and security risks of those projects becomes too large to justify for something that has niche uses or gives them no revenue. User data is practically toxic waste.
Dead Comment
I guess it's human nature for a person or an org to own their own destiny. That said, the driving force is not personal ambition in this case though. The driving force behind this is that people realized that OAI does not have a moat as LLMs are quickly turning into commodities, if haven't yet. It does not make sense to pay a premium to OAI any more, let alone at the cost of not having the flexibility to customize models.
Personally, I think Altman did a de-service to OAI by constantly boasting AGI and seeking regulatory capture, when he perfectly knew the limitation of the current LLMs.
LLMs are a commodity and it's the platform integration that matters. This is the strategy that Google, Apple embraced and now Microsoft is wisely pivoting to the same.
If OpenAI cares about the long-term welfare of its employees, they would beg Microsoft to acquire them outright, before the markets fully realize what OpenAI is not.
Nadella might have initially been caught a bit flat footed with the rapid rise of AI, but seems to be managing the situation masterfully.
Whatever is there doesn't work half the time. They're hugely dependent on one partner that could jump ship at any moment (granted they are now working to get away from that).
We use Copilot at work but I find it very lukewarm. If we weren't a "Microsoft shop" I don't think would have chosen it.
So hopefully if (when?) this AI stuff turns out to be the colossal boondoggle it seems to be shaping up to be, Microsoft will be able to save face, do a public execution, and the market won't crucify them.
If I recall correctly, Microsoft’s agreement with OpenAI gives them full license to all of OpenAI’s IP, model weights and all. So they already have a SOTA model without doing anything.
I suppose it’s still worth it to them to build out the experience and infrastructure needed to push the envelope on their own, but the agreement with OpenAI doesn’t expire until OpenAI creates AGI, so they have plenty of time.
If you like TypeScript, and you want to build applications for the real world with real users, there is no better front end platform in my book.
This would be the case even if OpenAI weren’t a little weird and flaky (board drama, nonprofit governance, etc), but even moreso given OpenAI’s reality.
1) Cost -- beancounters got involved
2) Who Do You Think You Are? -- someone at Microsoft had enough of OpenAI stealing the limelight
3) Tactical Withdrawal -- MSFT is preparing to demote/drop AI over the next 5-10 years
Isn’t that the basis for competition?
Humans typically work 1/3rd duty cycle or less. A robot that can do what a human does is automatically 3x better because it doesn't eat, sleep, have a family, or have human rights.
2. How many such PhD people can it do the work of?
Deleted Comment
Do you have to pay all sorts of overhead and taxes?
I mean, I don't think it's real. Yet. But for the same "skill level", a single AI agent is going to be vastly more productive than any real person. ChatGPT types out essays in seconds it would take me half an hour to write, and does it all day long.
Of course $10k/mo sounds like a lot of inference, but it's not yet clear how much inference will be required to approximate a software developer--especially in the context of maintaining and building upon an existing codebase over time and not just building and refining green field projects.
We were hiring more devs to deal with a want of $10k worth of hardware per year, not per month.
You can't claim it's even comparable to a mid level engineer because then you'd hardly need any engineers at all.
"Create high-quality presentations for communicating OpenAI’s financial performance"
https://openai.com/careers/strategic-finance-generalist/
What is interesting is there is no mention of agents on any job I clicked on. You would think "orchestrating a team of agents to leverage blah blah blah" would be something internally if talking about these absurd price points.
It points to an article on "The Information" as the source, but that link is paywalled.
They could also just be trying to cash in on FOMO and their success and reputation so far, but that would paint a bleak picture
My understanding is that this isn't really true, as most of those "dollars" were actually Azure credits. I'm not saying those are free (for Microsoft), but they're a lot cheaper than the price tag suggests. Companies that give away coupons or free gift certificates do bear a cost, but not a cost equivalent to the number on them, especially if they have spare capacity.
There is a moat in infra (hyperscalers, Azure, CoreWeave).
There is a moat in compute platform (Nvidia, Cuda).
Maybe there's a moat with good execution and product, but it isn't showing yet. We haven't seen real break out successes. (I don't think you can call ChatGPT a product. It has zero switching cost.)
Ironically if AI companies are actually able to deliver in terms of SWE agents, Nvidia's moat could start to disappear. I believe Nvidia's moat is basically in the form of software which can be automatically verified.
I sold my Nvidia stock when I realized this. The bull case for Nvidia is ultimately a bear case.
Look at Coca Cola, Google, both have plausible competitors, zero switching cost but they maintain their moat without effort.
Being first is still a massive advantage. At this point they should only strive to avoid big mistake and they're set.
If anyone has moat related to Gen AI, I would say it is the data(Google, Meta).
In consumer markets the moat is habits. The switching cost for Google Search is zero. The switching cost for Coke is zero. The switching cost for Crest toothpaste is zero. Yet nobody switches.
Microsoft is the IBM of this century. They are conservative, and I think they’re holding back — their copilot for government launch was delayed months for lack of GPUs. They have the money to make that problem go away.
Investing in second/third place likely valuable at similar scales too
But outside of that MSFTs move indicates that frontier models most valuable current use case - enterprise-level API users - are likely to be significantly commoditized
And likely majority of proceeds will be captured by (a) those with integrated product distribution - MSFT in this case and (b) data center partners for inference and query support
Fortunately, they're not anywhere near creating this. I don't think they're even on the right track.
Zima blue was good too
See Thomas Nagels classic piece for more elaboration
https://www.sas.upenn.edu/~cavitch/pdf-library/Nagel_Bat.pdf
Computationally, some might have access to it earlier before it’s scalable.
It’s moats that capture most value not short term profits.
That is, integrating use of their own model, amplifying capability via OpenAI queries.
Again, this is not to drum up the actual quality of the product releases so far--they haven't been good--but the foundation of "we'll try to rely on our own models when we can" was the right place to start from.