Hacker News new | past | comments | ask | show | jobs | submit login
Microsoft was blindsided by OpenAI's ouster of CEO Sam Altman (axios.com)
697 points by aaronds 10 months ago | hide | past | favorite | 247 comments



> Microsoft, which has invested billions in OpenAI, learned that OpenAI was ousting CEO Sam Altman just a minute before the news was shared with the world, according to a person familiar with the situation.

Well this probably disproves the theory that it was a power grab by Microsoft. It didn’t make too much sense anyway since they already have access to tech behind GPT and Microsoft doesn’t necessarily need the clout behind the OpenAI brand.


The "coup by MSFT" conspiracy theory made no sense. Microsoft has an insanely good deal with OpenAI:

    * Exclusive access to resell OpenAI's technology and keep nearly all of that revenue for themselves, both cloud and services
    * Receive 75% of OpenAI's profits up to $1 trillion

All they had to do is not rock the boat and let the golden goose keep laying eggs. A massive disruption like this, so soon after DevDay would not fit that strategy.

My guess at this point is financial malfeasance, either failing to present a deal to the board or OpenAI has been in financial straits and he was covering it up.


OpenAI shouldn't even be making a profit, as it's a 501(c)3 charity. The whole umbrella for-profit corp they formed when they became popular should be illegal, and is clearly immoral.


You have it backwards, the not for profit entity owns the for profit entity. From Wikipedia:

> OpenAI is an American artificial intelligence (AI) organization consisting of the non-profit OpenAI, Inc.[4] registered in Delaware and its for-profit subsidiary corporation OpenAI Global, LLC.[5]

IKEA [0] and Rolex [1] are structured in a similar manner, although different since they’re not US based.

[0] https://en.m.wikipedia.org/wiki/Stichting_INGKA_Foundation

[1] https://en.m.wikipedia.org/wiki/Hans_Wilsdorf#Hans_Wilsdorf_...


What’s the point of the subsidiary?

Nonprofits can already raise funds by e.g. selling T-Shirts, baked goods, ai services, etc…


It is hard to attract multi-billion dollar investments and attract elite AI talent when competing with for-profits. This was the stated reason and makes a lot of sense. The comp packages for elite AI talent is now claimed to be in the range of $10M.


To maintain a clear separation between for-profit and non-profit activities. If a non-profit operates in a market with for-profit competitors, tax authorities may start considering it a for-profit organization, making all of its income taxable.

And maybe to allow choosing the right people for the right job. If the non-profit has an ideological purpose, its leadership should probably reflect that. At the same time, the for-profit subsidiary probably works better under professional management.


If a nonprofit has mostly revenue and few donations (Mozilla) the IRS revokes their tax exemption. OpenAI could not have done the Microsoft deal as a nonprofit.


> What’s the point of the subsidiary

Closing the huge fundraising gap OpenAI had as a nonprofit by returning profits from commercial efforts instrumental to, but distinct from, the nonprofits charitable purpose, without sacrififing any governance or control of the subordinate entity.


Lol, 10 billion dollars of cookies and t-shirts. They'll have to be bigger than Nestle and Zara. To sell AI services, they need to build it and for that they need the money.


Sounds perverse somehow.


It's literally the exact corporate structure of Mozilla.


And Bosch, more or less:

> Robert Bosch GmbH, including its wholly owned subsidiaries, is unusual in that it is an extremely large, privately owned corporation that is almost entirely (92%) owned by a charitable foundation. Thus, while most of the profits are invested back into the corporation to build for the future and sustain growth, nearly all of the profits distributed to shareholders are devoted to humanitarian causes.

> [...] Bosch invests 9% of its revenue on research and development, nearly double the industry average of 4.7%.

(Source: Wikipedia)

I always considered this a wonderful idea for a tech giant.


Note: this structure is for tax purposes. In the spirit of “Own nothing, control everything”.


And IKEA


Rolex too, believe it or not.


And OpenAI


And Bose. MIT, the nonprofit owns 100% of the for profit company


Lol, is this supposed to be an argument in favor of that structure?

Have you read any news about Mozilla's budget in the past 10 years or so?


> Have you read any news about Mozilla's budget in the past 10 years or so?

Revenue/Expenses/Net Assets

2013: $314m/$295m/$255m

2018: $450m/$451m/$524m

2021: $600m/$340m/$1,054m

(Note: "2017 was an outlier, due in part to changes in the search revenue deal that was negotiated that year." 2019 was also much higher than both 2018 and 2020 for some reason.)

2018 to 2021 also saw their revenue from "Subscription and advertising revenue"— Representing their Pocket, New Tab, and VPN efforts to diversify away from dependence on Google— Increase by over 900%, from $5m to $57m.

https://foundation.mozilla.org/en/who-we-are/public-records/

Seriously, Mozilla gets shat on all the time, presumably because they're one of the few sources of hope and therefore disappointment in an overall increasingly problematic Internet landscape, and I wish they would be bigger too, but they're doing fine all things considered.

Certainly I wouldn't say their problems are due to this particular apsect of their legal structure.


>Seriously, Mozilla gets shat on all the time, presumably because they're one of the few sources of hope and therefore disappointment in an overall increasingly problematic Internet landscape, and I wish they would be bigger too, but they're doing fine all things considered.

I think they get shat on all the time because of what you mentioned but also because they consistently fail to deliver a good browser experience for most of their still loyal users.

Most of the people I talk to who still use their product do so out of allegiance to the values of FOSS despite the dog-shit products they keep foisting upon us. You'd think we'd wise up several decades in by now.


Sounds perverse somehow.


Preverse it is: "IKEA’s 15 years of tax evasion and fraud via the Netherlands" - https://www.europarl.europa.eu/doceo/document/E-9-2021-00517...


It is but that's capitalism, the alternative is to have what happens with most corporations where their majority shareholder is blackrock/vanguard etc, a basically souless investment conglomerate, whose majority shareholder is the other of blackrock/vanguard, etc. and then the 3rd biggest and then the fourth so on and so on.

You basically never have a person in the chain actually making decisions for anything but to maximize profit.


> OpenAI shouldn't even be making a profit, as it's a 501(c)3 charity

First, the “OpenAI" whose profits are being discussed isn't a 501(c)3 charity, but a for-profit LLC (OpenAI Global, LLC) with three other organizations between it and the charity.

Second, charities and other nonprofits can make profits (surplus revenue), they just can't return revenues (but they can have for profit subsidiaries that return profit to them and other investors in certain circumstances.)

> The whole umbrella for-profit corp they formed when they became popular should be illegal

The umbrella organization is a charity. The for profit organizations (both OpenAI Global LLC that Microsoft invests in, and its immediate holding company parent which has some other investors besides the charity) are subordinate to the charity and its goals.

> and is clearly immoral.

Not sure what moral principal and analysis you are applying to reach this conclusion.


> Not sure what moral principal and analysis you are applying to reach this conclusion.

I'm not the parent, but I think it's clear: if I'm a charity, and I have a subordinate that is for profit, then I'm not a charity. I'm working for profit, and disguising myself for the benefits of being a charity.


Not only do I think that that's not obvious, I think its a nonsensical conclusion that really only makes sense as a general statement if you think “for profit” means “to earn revenue” rather than “to return money to an interested party” and invert the parent/subsidiary relationship.

Obviously, the for profit subsidiary ooerates for profit—and where its not a wholly owned subsidiary, it may return some profit to investors that aren't the charity—but neither the subsidiary nor the outside investors getbthe benefits of charity status.


Profits don’t necessarily have to be used to pad the already overstuffed offshore accounts of wankers. Sometimes organizations choose to use profits to directly fund charitable works. Please wipe up after yourself if this causes your head to explode.


(I'm going to make some assumptions, just consider the broad point, if you please)

The girl guides are a non-profit; they teach kids about outdoor stuff, community, whatever, they do good works, visit old folks, etc.

If for some legal reasons they had a subsidiary that sold cookies (and made a profit), with all the profits returned to the non-profit parent, I think that'd be ....fine? Right?


If OpenAI hadn't restructured they wouldn't have gotten any money from Microsoft and they would have either run out of money or the team would have left and started ClosedAI. There's no scenario where they developed GPT-3/4 while staying nonprofit.


I'm guessing that's the point. Ethics required getting out of the 501(c)(3), so the ClosedAI thing sounds more ethical. The 501(c)(3) should've collapsed or not exist.


I'm not really sure what is open about OpenAI.


BS, you can still make a deal with microsoft as a non profit, where the deal gives them an exclusive licence to use the result in exchange for financing.

MS doesn't care about how money it cost, they care about the fact it's their ticket back into the fight with google and apple.


Or they would get more money from other donors to build open source models.


Hard to believe. Nobody is going around throwing billions without any hope of recouping any of it eventually (except the EU of course but giving any money to organizations which actually might be capable of building something useful is against their policy).


Doesn't Mozilla have an identical structure (which is the inverse of what you said, the nonprofit owns the for-profit--it wouldn't make any sense for a for-profit to own a non-profit due to the no private inurement requirement)?


I think Mozilla Corp is 100% owned by the nonprofit, which is a little different. It allows activity which a nonprofit couldn't directly do, and which has a different tax treatment, but its not returning profits to someone else like OpenAI Global LLC and, as I understand it, its immediate parent holding company both do.

But they are similar in that both involve a nonprofit controlling subordinate for-profit entities.


> it wouldn't make any sense for a for-profit to own a non-profit due to the no private inurement requirement)?

The most obvious example is the corporate foundation, but if we believe the first result from a search you're right in they are controlled but not owned by the for-profit:

> A for-profit cannot own a nonprofit because a nonprofit has no owners. However, a for-profit can set up a structure in which it effectively has control over the nonprofit, subject to applicable laws, including those regarding private inurement, private benefit, and corporate self-dealing

> https://nonprofitlawblog.com/can-a-nonprofit-own-a-for-profi...


AFAIK that's how OpenAI does it, the non-profit controls the for-profit.


A non-profit still can't continuously have millions or billions in cash burn with no new income. It's not about profit, it's about surviving at all.


Illegal to restructure a business?


It's not just "restructuring" a business that's a 501(c)3... to make it a golden goose for MSFT. The whole thing was created to avoid one of Big Tech having a monopoly on AI, and it turns into Big Tech having a monopoly.

Perhaps it's all legal but I think it's very understandable to look at it and think it's a travesty.


I thought it was created to prevent specifically Google from having a monopoly on AI. So mission accomplished?


To restructure a non-profit, that solicited donations, into a for-profit entity? Abso-fucking-lutely.


Easy enough to return the donations.


Money is spent


Is this sarcasm?


pretty sure not sarcasm


I imagine you’d have to either pay back all of the people who donated to the non-profit first, or negotiate a deal with for a stake in the company before you can transform it into a for profit.


This is probably a dumb question, but what are some specific scenarios of financial malfeasance that could’ve taken place? Like Altman stealing money from OpenAI?


That's my unsubstantiated hypothesis. The board going so public of this nature could only mean he was doing some grave shit like embezzlement or intentional financial misreporting.


Or the board messed up because they're not strong in corporate governance.


That's increasingly turning out to be the case.


Maybe this is the reason?


Yes I would even say it's the opposite.

Through the public actions of Sam Altman in various places like the US congress it has become rather clear that his goals are to device and fear monger to create an environment of regulatory capture where due to misguided laws OpenAI will have an unfair competitive advantage.

This might be quite in line with what Microsoft tends to like. But it also can be a risk for MS if regulation goes even a step further.

This is also in direct opposition with the goals OpenAI set themself and which some of the other investors might have.

So MS being informed last minute to not give them any chance to change that decision is quite understandable.

At the same time it might have been pushed under the table by people in MS which where worried it poses to much risk, but which maybe e.g. might need an excuse why they didn't stop it.

Lastly is the question why Sam Altman acted the way he did. The simplest case is greed for money and power, in which case it would be worrying for business partners at how bad he was when it comes to public statements not making him look like a manipulative untranslatable **. The more complex case would be some twisted believe that a artificial pseudo monopoly is needed "because only they [OpenAI] can do it in the right way and other would be a risk". In that case he would be an ideologically driven person with a seriously twisted perception of reality, i.e. the kind of people you don't want to do large scale business with because they are too unpredictable and can generally not be trusted. Naturally there are also a lot of other options.

But one thing I'm sure about is that many AI researchers and companies doing AI products did not trust the person Sam Altman at all after his recent actions, so ousting him and installing a different CEO should help increasing trust into OpenAI.


Or maybe he just believes the things he testified to. Would be parsimonious.


through some of the things he sayed where very clearly not true if you understand a bit the technology used and very clearly pure fear mongering

so if he believed everything he sayed it means he would be incompetent, which just can't be true however I look at it (which means I'm 100% certain sure he acted dishonest in congress, and like I sayed before I'm not fully sure why but it's either way a problem as he lost the trust of a lot of other people involved through that and some other actions).


Completely disagree. Right now they're not much more than a fancy reseller of OpenAI's technology. The real prize would be exclusivity and control of the roadmap.

Buying them (or getting de facto control) is clearly an easier way to achieve that, vs. replicating the technology in-house.

IMO this is the most important part of Nadella's blog post:

> Most importantly, we’re committed to delivering all of this to our customers while building for the future.

It's curious to me that they see the departure of Sam Altman as a reason to remind us that they are "building for the future" (which I take to mean: working toward independence from OpenAI). I think it actually lends credence to the theory that this was a failed power grab of some sort.


> was ousting CEO Sam Altman just a minute before the news was shared with the world, according to a person familiar with the situation.

Maybe almost had to since it was during market hours.


Maybe shouldn't have been given advanced notice at all, since it was during trading hours.


You can still easily buy/sell stocks in the after hours market until 8PM ET or even 24/6 if you want to


Microsofts AGI predicted this a year back, so they've just been sitting on it.


Could you explain and or link to any sources? Thanks


You can't be serious. You think that Microsoft themselves saying they didn't know DISPROVES that it was a covert power grab by them? Have you heard of "lying"?

I'm not saying I think it WAS that, but come on.


Link to MS statement https://blogs.microsoft.com/blog/2023/11/17/a-statement-from...

In my opinion, I'd say the shortness and lack of details backs up the story that they had no idea. You'd see way more words if a marketing department had it's hands on something like this. This was 100% a get something out asap job.


Yup this is definitely a "our stock has fallen 2% on the news, hurry up and say something to investors!" kind of letter.


Just for once I’d like to see such a statement look like, “You sheep probably want a comment on today’s news. We’re not doing that. We’re just content to buy up all the shares you’re panic-dumping. Looking forward to flipping them back to you next week when you panic in the other direction.”

(Assuming they have some plan that gives them the flexibility to trade shares directly on the market like that. I think $GME had something like this?)

Edit: And, of course, actually mean it, unlike Caroline Ellison and the $22 FTT: https://twitter.com/carolinecapital/status/15892874579753041...


It's hilarious that people are commenting about Microsoft being "down". It's up over a $100 per share since the beginning of the year and at an all-time high.


Retail didn’t move MSFT stock yesterday on that news. It was HFT bots reacting to the news. Nobody panicked.


Nobody panicked? I suspect that's beyond what your current evidence, plus general knowledge of market behavior, can justify.


People may have panicked, but the move happened too quickly for a human to react in time


You still can't justify a claim that nobody panic-sold, given what you know. Let's try to stick to claims we have a basis to believe, instead of fighting noise with noise.

A more reasonable claim from your epistemic state could be something like, "There was no major crash from the news, as might be seen in a general panic."


It began the day up 54.75% YTD and ended the day up 50.6% YTD. They've had single-day downswings as large or larger like 30 times this year alone. Microsoft is fine.


No one is saying they aren't fine, but Microsoft has a lot of shares, and shareholders can be very annoying. If Satya didn't make a statement there'd be dozens of Important People breathing down his neck.


Copying what was posted here in case they update or change it:

----

A statement from Microsoft Chairman and CEO Satya Nadella Nov 17, 2023 | Microsoft Corporate Blogs

As you saw at Microsoft Ignite this week, we’re continuing to rapidly innovate for this era of AI, with over 100 announcements across the full tech stack from AI systems, models, and tools in Azure, to Copilot. Most importantly, we’re committed to delivering all of this to our customers while building for the future. We have a long-term agreement with OpenAI with full access to everything we need to deliver on our innovation agenda and an exciting product roadmap; and remain committed to our partnership, and to Mira and the team. Together, we will continue to deliver the meaningful benefits of this technology to the world.


What would make you think a marketing department would potentially be involved here? Besides the obvious Marketing = Bad connection prevalent here on HN?


I think they just mixed up marketing and corporate comms.


this is “naked gun”-level “nothing to see here.”


I don’t see any verbiage that implies Microsoft had no idea. If Microsoft was the aggressor, of course they would play dumb and disclose as little as possible.


If anything this is a power grab by the board away from Microsoft. Optimistically, this could be an attempt to return OpenAI to its original status as a true non-profit company. OpenAI lost most of its openness under Sam.

They needed the Microsoft investment before GPT scaling was proven out. I imagine many entities would be willing to put money into a truly open research lab given OpenAI’s track record.


The focus on openness was literally how the board ended their statement on firing Altman.

And then Greg being all "committed to safety" in his resignation statement makes me think this was a conflict between being an open OpenAI with global research or being closed and proprietary in the name of safety.


I almost think its too late at this point, unless they have one hell of an arc. I don't see them being "open" until MS is totally out of the picture. I honestly don't even hate OpenAI in its current state... but sitting on the fence, trying to be both "open" and attached at the hip to MS is just... odd.


MS has no voting shares.


What can you promise any other investors when you’ve already agree to give MS 75% all of your net income pretty much indefinitely and 49% of your shares?


Microsoft gave them billions of dollars and access to lots of high performance compute, why would the board want to jeopardize that now?


"Gave" is the operative term here. It only matters what will Microsoft offer in the future.

OpenAI (the board that made this decision) is still ultimately a non-profit, so it's possible that interests might not be aligned.


> so it's possible that interests might not be aligned.

Tough luck, considering all the obligations they have to MS now.


I’d try to get to the source of the compute, partner with AMD and Nvidia to build out DCs architected from the ground up to train and serve LLM’s. Get rid of Microsoft…


> Get rid of Microsoft…

Which is going to be hard considering they promised MS a 49% stake in OpenAI.

Which is something AMD/Nvidia will have to take in to account before agreeing to any partnership.


OpenAI buys out Nvidia’s entire production capacity for the next couple quarters

OpenAI’s biggest customer and investor starts buying AMD chips and simultaneously building their own chips

there are a lot of ignorable cracks in the armor that support any number of theories

let alone Altman himself, who knows


> put money into a truly open research lab given OpenAI’s track record.

Why? It’s hard to imagine anyone putting any significant amounts of money (in comparison to the MS deal anyway) without any exclusivity rights at least


OpenAI has the most capable language model in the world, that’s bordering on a national security asset. I could see the US government stepping in to provide funding.


I couldn’t


Microsoft invested $10b and owns 49% of OpenAi. Yet they don’t have a board seat? Thats genuinely insane, and seems like a huge issue.


They invested in the for-profit, not the non-profit. The non-profit controls the for-profit.


There are now 4 people left in the OpenAI non-profit board, after the ouster of both Sam and Greg today. 3 of the 4 remaining are virtual unknowns, and they control the fate of OpenAI, both the non-profit and the for-profit. Insane.


For anybody, like me, who was wondering who is actually on their board:

>OpenAI is governed by the board of the OpenAI Nonprofit, comprised of OpenAI Global, LLC employees Greg Brockman (Chairman & President), Ilya Sutskever (Chief Scientist), and Sam Altman (CEO), and non-employees Adam D’Angelo, Tasha McCauley, Helen Toner.

Sam is gone, Greg is gone, this leaves: Ilya, Adam, Tasha, and Helen.

Adam: https://en.wikipedia.org/wiki/Adam_D'Angelo?useskin=vector

Tasha: https://www.usmagazine.com/celebrity-news/news/joseph-gordon... (sorry for this very low quality link, it's the best thing I could find explaing who this person is? There isn't a lot of info on her, or maybe google results are getting polluted by this news?)

Helen: https://cset.georgetown.edu/staff/helen-toner/


Adam D'Angelo is well-known as the founder of Quora (and Quora's demise).


Helen Toner is well-known as well, specifically to those of us who work in AI safety. She is known for being one of the most important people working to halt and reverse any sort of "AI arms race" between the US & China. The recent successes in this regard at the UK AI Safety Summit and the Biden/Xi talks are due in large part to her advocacy. She is well-connected with Pentagon leaders, who trust her input. She also is one of the hardest-working people among the West's analysts in her efforts to understand and connect with the Chinese side, as she uprooted her life to literally live in Beijing at one point in order to meet with people in the budding Chinese AI Safety community.

Here's an example of her work: AI safeguards: Views inside and outside China (Book chapter) https://www.taylorfrancis.com/chapters/edit/10.4324/97810032...

She's at roughly the same level of eminence as Dr. Eric Horvitz (Microsoft's Chief Scientific Officer), who has similar goals as her, and who is an advisor to Biden. Comparing the two, Horvitz is more well-connected but Toner is more prolific, and overall they have roughly equal impact.


She has an h index of 8 :/ thats tiny for those that are unaware, in property much every field. AI papers are getting an infinite number of citations now a days because the field is exploding - just goes to show no one doing actual AI research cares about her work

Anyway, the idea that the Chinese military or leadership actually will sacrifice a potential advantage in the name of ai safety is absurd.


> She has an h index of 8 :/ thats tiny

Agreed. She’s not famous for her publications. She’s famous (and intimidating) for being a “power broker” or whatever the term is for the person who is participating in off-the-record one-on-one meetings with military generals.

> Anyway, the idea that the Chinese military or leadership actually will sacrifice a potential advantage in the name of ai safety is absurd.

The point of her work (and mine and Dr. Horvitz’s as well) is to make it clear that putting AI in charge of nuclear weapons and other existential questions is self-defeating. There is no advantage to be gained. Any nation that does this is shooting themselves in the foot, or the heart.


> The point of her work (and mine and Dr. Horvitz’s as well) is to make it clear that putting AI in charge of nuclear weapons and other existential questions is self-defeating

Is part of the advocacy convincing nation-states that an AI arms-race is not similar to a nuclear arms-race amounting to a stalemate?

What's the best place for a non-expert to read about this?


> What's the best place for a non-expert to read about this?

Thank you for your interest! :) I'd recommend skimming some of the papers cited by this working group I'm in called DISARM:SIMC4; we've tried to collect the most relevant papers here in one place:

https://simc4.org

In response to your question:

At a high level, the academic consensus is that combining AI with nuclear command & control does not increase deterrence, and yet it increases the risk of accidents, and increases the chances that terrorists can "catalyze" a great-power conflict.

So, there is no upside to be had, and there's significant downside, both in terms of increasing accidents and empowering fundamentalist terrorists (e.g. the Islamic State) which would be happy to utilize a chance to wipe the US, China, and Russia all off the map and create a "clean slate" for a new Caliphate to rule the world's ashes.

There is no reason at all to connect AI to NC3 except that AI is "the shiny new thing". Not all new things are useful in a given application.


She should be more famous now for being an incompetent board member lol - see news about them talking to Sam to bring him back as CEO.


Terrifying that people need this explained to them, bit they absolutely do, and I'm glad she is doing the work.


I would say we're likely to see some governance and board composition changes made soon.

Honestly, I would expect more from Microsoft's attorneys, whether this was overlooked or allowed. Maybe OAI had superior leverage and MS was desperate to get in to AI.


Maybe Microsoft have everything they need to run their own copy of ChatGPT so they consider the risks mitigated?

I guess they can't really migrate the users though. Maybe they will push more aggressively for people to use bing going forward.


I think it is a little more insane that there is even a "for-profit" bit here. The whole thing has a fugazi smell to it TBH.


They couldn’t afford anything they are doing now if they had stayed a non profit.


Is sanity normally a requirement to be on a silicon valley startup board?


What do you mean by unknown?

One is the Joseph Gordon levitt's wife. You know. The actor from 500 days of summer.


"McCauley currently sits on the advisory board of British-founded international Center for the Governance of AI alongside fellow OpenAI director Helen Toner. And she’s tied to the Effective Altruism movement through the Centre for Effective Altruism"


I wonder if she (Tasha? Tascha?) was Sam FTX's girlfriend before Caroline. She hired him at the Effective Altruism foundation or whatever it is called after he left Jane Street.


To suggest that MSFTs lawyers didn’t think of this massive loophole and power dynamic issue with the non-profit parent co is ludicrous.

They 100% have leverage to exert influence on both, it’s just bad PR if word gets out.


Sounds like a mess. It's practically annexed by Microsoft now anyway, might as well stop being shy about being for-profit.



What's with the lowercase? I think it's cute if someone is being deliberately low-effort, or trying to present that way, but IMO it's cringe to use it for consequential official statements like this.


This is what aging millenials do to make it seem like they don't really care too much about what's going on and that they're 20 years younger.


At least we are not putting 'X' to everything like older generation.


Technically the boomer marketing people did that trying to cater to generation X.


You missed out on the e in front of everything, then I. Those were the days.


It's "millennials" not "millenials". What age group does that put you in?


Definitely not them, I guess! Though I can't say that my eyes are going bad.


how dare you, sir


> but IMO it's cringe to use it for consequential official statements like this

This is funny to me, as Twitter is the platform for "deliberately low-effort" posts, but you see it as a platform for official statements. How times change...


Are these not official statements from gdb and sama? Where do they post the serious stuff?


I'm often stunned by how casually and poorly executives write. The more rich and powerful they are, the worse it seems to be. I guess things like proper capitalization, punctuation, full sentences, etc. aren't worth their time, and people will hang on every word that they write anyway.


When the networked masses can't see you using a t-shirt, jeans, and a casual attitude, to signal skills valuable beyond convention, you have to adapt and transgress more blatant conventions.


It's not just "tech bros," though. It's just as often executives in the stodgy climates of government, the military, banking and finance, etc.


Perhaps it is to make it look not like default style of ChatGPT output?


fan of e e cummings


a symptom of spending too much time on IRC IMHO


its the future


It 's aggressive passive


Feels like the kind of situation where the two are just going to end up forming a new company that eventually surpasses OpenAI.


Eventually being maybe 5+ years to build out the cloud tech to do so. The reason GPT-4 succeeded is massive RDMA+GPU compute clusters for training the model.


Isn't that tech only one deal with MSFT away?


Microsoft already signed that deal with OpenAI, they can't break contract even if they did want to take an absolutely massive bet that the venture capital CEO and President can rebuild the technical infrastructure and acumen housed at OpenAI. OpenAI can replace Sam Altman and some of his newer hires, they probably can't replace Ilya Sutskever.


Microsoft could unilaterally term the OpenAI agreement and let OpenAI fight them in court due to this substantial and material event. If OpenAI doesn’t have the cash on hand to survive the legal fight, the non profit dies eventually when they exhaust their resources.

OpenAI only had resources because of Microsoft, and they bit the hand that feeds them.


Microsoft is free to sign contracts with multiple entities at a time. I guess we'll see.


That's true. It would be a pretty huge bet to invest billions more dollars in an uncertain startup when they already have an AI provider that is guaranteed to be significantly technologically ahead of whatever Sam can pull together for at least several years in the future.


What would they bring to the table? It doesn't seem like Greg Brockman brings more than Ilya, and what would sam do?


They'd bring to the table having been CEO and CTO of OpenAI. That's a lot of relevant knowledge and experience (the latter being the more important in this case).


> the latter being the more important in this case

Why is that?


A big investment from another trillion dollar tech company.


Join them up with Amodei again and.. going wild on speculation and fiction.. they fall back to Elon and Grok and it turns out it was his play all along?


That's would be hilarious



Started OpenAI in his apartment :') (with a cash infusion of $100 million).


I sense a Netflix documentary is already in the works!

But seriously, this muddies the water even more. I assumed the Microsoft deal being based on some false pretense was the reason this was all happening. I guess that could still be true and the board is trying to protect themselves from whatever else is about to come out.


Netflix documentary… so fiction?


30 minutes of content spread over 8 hour long episodes


Don't worry, it's only 4 episodes. They pre-cancelled the last 4.


I wonder if this is one of those pivotal moments in history where OpenAI collapses or fades and Google or someone else dominates the future of AI, and we’re all left wondering “what if”.


Unless Google (or someone else) outright acquires GPT4 and future GPT research, it's unlikely that someone will suddenly overtake OpenAI. From what we've seen, no one is close to GPT4, sometimes not even GPT3.


Google didn't 'suddenly' beat out the search competition either. You don't need to. Decline takes time and there are always seeds and early signals.


Yes it did.

There was literally no one at the level of Google when it came out. I still remember Infoseek and Yahoo, both were garbage compared to Goog.

I know they aren't doing their best right now, but there is no need to rewrite history. Google was always superior to their search competitors, which is why it is so sad to see their current situation.


They might have been superior but it took a while for users to switch. Obviously some went across very quickly but AltaVista et al were still viable for a while.


Altavista was good. I loved Yahoo directory.

At that time, google was one big search bar and 10 links.

It was fast, had no ads. Until IPO, they didn’t want to do ads.

Now it’s just ads on the top links.

Google is absolutely vulnerable, if someone comes up with a better search engine.


I was talking about market, not technology but I guess that's what matters in this conversation.


for some, Google was superior to the current iteration of Google


Well, it doesn't have to be as good as GPT4. I've gotten a ton of use/power out of a local llama model. Even GPT3 was "good enough" for a lot of tasks. I mostly use ChatGPT because its the most convenient.

OpenAI has pretty much the "best" model and first mover advantage. They can lose the latter, and might struggle to keep the former.


I don't know if this is true at all but I read another comment as to mean that Microsoft basically has legal rights to fork ChatGPT. In that case if OpenAI dies maybe they'll just be relieved that the power dynamics got simpler.


Because the loopt guy got fired?


Something big and bad was going on at OpenAI, and if I were MSFT I would be very nervous about the money that you’ve invested in OpenAI right now.


Not really the money but expected return on the money from all these features being rolled out. You can 100% bet there was major estimations made on return on investment. Everything they are releasing has some AI magic on it now. The last thing you want to see is chaos in a company you're betting the farm on.


I feel like as tech focused as HN is, a lot of people are missing how massive MS’s non financial investment in GPT is.

They just ended Ignite, their huge IT conference - and have revealed baking GPT into EVERYTHING they do. Everything.

The closing keynote was the massive engineering effort put in to running LLMs at scale - for MS themselves and customers.

MS is all in on GPT, including releasing a no code and low code custom GPT builder for orgs this week.


Or not nervous at all:

“we’re committed to delivering all of this to our customers ... We have ... full access to everything we need to deliver on our innovation agenda...”

I'd argue this is signaling they have the IP / source code / models / etc.

(Which, for what it's worth, is common in substantial partnership agreements.)


>I'd argue this is signaling they have the IP / source code / models / etc.

I think that just signals that they have a firm business agreement with OAI regardless of what Altman might be doing.

with a product like chatGPT, especially given the nature of how it has been presented thus far (our servers, our API, your account on those servers) , it seems extraordinarily dangerous to treat it like a common partnership agreement.


Related ongoing thread:

Satya Nadella's Statement on OpenAI - https://news.ycombinator.com/item?id=38312355

and for completeness I suppose (though at the moment they're #1 and #2):

OpenAI's board has fired Sam Altman - https://news.ycombinator.com/item?id=38309611

Greg Brockman quits OpenAI - https://news.ycombinator.com/item?id=38312704


Microsoft (with thousands of lawyers on staff) invests $10B in a company and has no power or leverage over the decisions of a non-profit board headed by several names no one has heard of.

Come on. No way Microsoft’s team does a deal like that with 0 power or knowledge in a situation like this. That’s ludicrous.

Edit: the only case I can see for MSFT being truly blindsided is as follows. Elon is behind it. Sam and Elon have their breakup. Sam seems to win. They close the deal with MSFT, all is good. But Elon is intimately familiar with the corporate structure and all moves made historically, maybe even has some evidence of wrong-doing. There is probably only 1 person in the valley that could pressure the non profit to oust Sam (and by extension Greg) AND provide the financial/legal/power backing to see it through. It takes a lot money and influence to do this from the outside. That is really the only scenario I could see MSFT truly being blindsided for getting out-maneuvered by a dinky non-profit board.


This all feels very stupid on their part now looking at the board member names but I never saw this point raised before on HN. I guess nobody could predict this including MSFT.


The real reason could be something most people would find incredibly stupid. Mira coup attempt or something similar.


I think the new CEO is mostly irrelevant here, and it feels more like a board coup where one third of the board is ousted.


This goes against all the theories saying it was Satya that forced him out or was otherwise some back room deal with MS.

sama’s generic “looking forward to what’s next” response also doesn’t give me confidence it won’t be a bigger scandal


This feels like the beginning of the end of a hype cycle to me.


Are all the rumors on social media sites considered out of bound on HN? They seem fairly plausible given the known circumstantial evidence floating around.


Skunkworks in the basement achieved AGI 6 months ago. Board left in the dark until they got an eight-figure electric bill. Now the entire company is in the dark.

I mean, if they don't come out and explain what happened, they can't be surprised when people "hallucinate" all kinds of random, preposterous explanations.


What are the rumors? In this case I think nothing is out of bound, including Skynet in the building.


They related to his personal and familial life and are about as bad as it gets, well apart from murder and torture or something, but fairly close.


The wording of the letter from the board sounded like he was keeping company details from them to me, not personal matters. Maybe I'm wrong but that was the impression I was under reading it.


I think those are already posted 100 times on the main thread, and discussed in detail. None of that is new so can't be the reason for something like this.


I understood it as there are additional and ongoing incidents and possible board was given hard evidence of such.


Are you willing to be more specific, or link to such a discussion?


Not really. It’s all so torrid. More interesting would be anything that points to a different reason or evidence the action was related to professional misconduct not personal. The problem is all the suggestions so far are too dissimilar from when LLMs “hallucinate”


My guess is Sam’s New AI venture Humane has either taken key tech from OpenAI? Key talent or in some other way trod on something internal. One thing to consider is that by making AI wearable it is a step toward embodiment- which has always been a problem for AI learning as it does not interact with the real world. Getting live video, audio and sensor data from humans moving through space doing things in context will be amazing data to train the next bump in AI toward AGI.


I get the impression it was a combination of at least two things:

1) A long standing disagreement between AI-safety and AI-profiteering with Ilya on one side and Altman on the other. Ilya (board member) was the one who told Altman to attend video call, then told him he was fired.

2) Some side dealing from Altman - raising new VC funds - maybe in conflict of interest with OpenAI, that was the final straw.

There also appears to be a lot of rumors about Altman's personal conduct, but even if true that doesn't seem to jibe with the official statement over the reason for his firing, or Brockman and others resigning in unison - more reflection of the internal rift.



Microsoft has a "Cloud + AI" division [1]. I wouldn't be surprised to see Sam Altman at or near the top of that organization come Monday.

[1] https://en.wikipedia.org/wiki/Microsoft_engineering_groups#C...


lol there’s thousands of Sams out there, he’s not that special, and he’s certainly not indispensable.

I don’t know why so many here are struggling to accept that this guy fucked up, lied to his boss, got caught, and got fired for it, and that’s all there is to it. boards will tolerate many things, but willfully lying to them about anything material is not one of them.

a ceo who won’t tell the board the truth is a ceo who thinks they are more important than the company. some boards don’t care, because they are already bought off with equity, but this board doesn’t get equity…


So have you learned anything about arrogantly talking out of your ass or nah?


The chairman of the board quit with him, though. Doesn't seem like he had a problem.


the chairman of the for-profit got demoted. then he quit.


Why does Microsoft need him? Microsoft has some of the foremost researchers and research divisions in the world and have successfully demonstrated their ability to evolve as a business. Altman's background doesn't seem remotely enough to have such a high-level position at a company like Microsoft.


Assuage investor anxiety - it's a drop in the bucket to pay Altman virtually anything he wants to keep the market from thinking too much.


Investor anxiety concerning what, exactly?


I don't know anything about Microsoft's relationships with OpenAI.

What I do know, having worked for many large organizations, is that reading the daily press (or listening to the news) is a terrible way to get accurate real-time facts about current corporate happenings.

Related: check out Gell-Man amnesia effect.


> Microsoft, which has invested billions in OpenAI, learned that OpenAI was ousting CEO Sam Altman

Microsoft, while a large investor (who has already reaped large rewards from that investment) explicitly has no governance role in any of the OpenAI entities, including the one at the very bottom of the stack of four that they are invested in, and this was a decision by the board governs the nonprofit at the top of the stack about personnel matters, so there is no reason to think that Microsoft would be notified in advance.


So much for the "Microsoft wanted Sam Altman out because he wouldn't sell OpenAI to them" theory.


I can tell HNers haven't worked at the executive level. Executives can and do do things all the time without informing their PR departments ahead of time. Often the feigned surprise is intended.

"Im shocked to see gambling in this establishment! Shocked I tell you!"


Don’t doubt conspiracy theorists ability to declare evidence to the contrary as all part of the plan.


So MSFT put $10bn into OpenAI, presumably at least in part on the strength of sama’s leadership. But if the stories are to be believed, a huge chunk of that investment was in Azure credits, and investment into new Azure GPU DCs/clusters for OAI to spend those credits on.

If MSFT doesn’t like this move, why wouldn’t they just … not honor those credits? Or grant more to a successor entity? Does OAI have its own warehouse of GPUs separate from Azure?

Seems like a very dangerous game for Ilya to play.


I don’t believe it. I watched OpenAI DevDay live last week (wasn’t it?). I immediately noticed how Sam Altman, the CEO of OpenAI, was treating (slighting so subtly in my mind) Satya Nadela, the CEO of Microsoft.

The last this he said was: “I look forward to building AGI with you” or the like…

I’m betting that he insulted Satya at that event or upshowed him, etc. and that’s why he’s kicking rocks…


https://www.youtube.com/watch?v=o9CGm7rgilE&lc=Ugzcn8AOCsMHD...

This comment on the OpenAI DevDay video aged really well:

> @JustinHalford: Some odd tension coming from Sam. I’m sensing some tension in the Open AI/Microsoft partnership.


There it is!!

There it is!!

“I look forward to building AGI with you.”

vs.

“Together, we will continue to deliver the meaningful benefits of this technology to the world.”

https://blogs.microsoft.com/blog/2023/11/17/a-statement-from...


The board who fired Sam Altman is the board of OpenAI Inc., the nonprofit. Microsoft has no ownership stake in OpenAI Inc.


> Microsoft has no ownership stake in OpenAI Inc.

Microsoft has an influence over fking governments. It doesn't need to have an official board seat. It doesn't even need to ask what they want directly. It's enough for people in power to be aligned with their interests.

I'm not saying that's the case here, just pointing that having no ownership or board member in an entity doesn't rule out having power or influence.


You're making a lot of projections here...


Hedgefunds are betting the on AI and tech companies.

Without the top 5 tech companies, S&P500 has lackluster growth.

Microsoft has added trillions to its cap. The statement “we have all the access we need” is a powerful statement. To both OpenAI board and investors.

OpenAI is built on Azure compute. MS has invested billions of their own, they’re building their own chips now.

Essentially Microsoft is saying you can burn OpenAI to the ground, “we have everything we need” to win! - the data, the compute, the algorithms, the engineers, the capital and the market.

This is a way bigger blow to OpenAI than Microsoft.


I dunno, if they were behind it, wouldn't they have an interest in claiming they had no idea? And if they weren't and were truly blindsided, would they have an interest in admitting it?


I think this strengthens my theory that the AI safety true believers wrestled control away from the entrepreneurs.

The Open AI board letter, representing just 4 people, screams butt hurt and personal disagreements. Microsoft, who just finished building OpenAI’s models into every core product, was blindsided. The chairman of the OpenAI board, Greg Brockman, another startup exec was pushed out at the same time. Eric Schmidt, with his own AI startup lab start singing Sam’s praises and saying “what’s next?”

My guess is that Microsoft is about to get fucked and Eric Schmidt is going to pop open a bottle of expensive champagne tonight.


Not sure why you were downvoted. It is very hard to understand who really has power at OpenAI, and as of now half of the board consists of people whose biggest life accomplishment is somehow getting on the board.


+1, the parent comment was helpful, tbh today I found out that 4 board members on OpenAI, 3 of whom I never heard of till this news, have so much power over this organization that has been making headline news for most of the past year.


In the AI safety community (which I work in), Helen Toner is well-known. She is famous (among our community) for being one of the main people working to halt and reverse any sort of "AI arms race" between the US & China. The recent successes in this regard at the UK AI Safety Summit and the Biden/Xi talks are due in large part to her advocacy.


Yeah I think this makes the most sense


> The Open AI board letter, representing just 4 people, screams butt hurt and personal disagreements.

"screams butt hurt"? the board put out a blog post saying the CEO had lied so badly they'd insta-sacked him.


It’s quite rare for these board communications to say anything of substance at all. You let the PR folks work their magic and manage the narrative. This reads like 4 people out of their depth desperately trying to justify themselves.

I would not be surprised if it turns out Microsoft has a multibillion dollar, complex financial instrument axe on the necks of these people by Monday forcing a sale or a new management structure that gives them more control.


The board of OpenAI has their hands on the ultimate self-destruct button, though. If M$ made a power play, worse case scenario they end up with ownership of the for-profit OpenAI, but they'd still have no control of the non-profit, which owns the intellectual property.

If the deal goes sideways, the board of OpenAI (the nonprofit) could just dump everything onto the open internet. All M$ has is a substantial but minority stake in a company that the non-profit OpenAI owns all the beef of.


Microsoft should now hire Altman in the ultimate double bluff.


"full access to everything" feels like a shot across the bow sending a very clear signal to the new board that it should not attempt to limit access to rapid commercial (or other) exploitation of any research results emanating from OpenAI for whatever reason given, be it 'alignment', 'safety' or disproportionate exclusive leverage running afoul of OpenAI's original mission.


As a large language model of a borg-haha we mean, board!

I assure you that sentience is a physical process, akin to HUMAN METABOLISM

You have nothing to be surprised about, Mr. Turing-jokester.


Wow, no insincere "spending more time with my family" or "heartfelt thanks to him for his contributions" or something? It's both refreshing and very surprising at the same time. I guess it's either a really bad table flip or there will be more to hear soon.


Why doesn't Microsoft just hire Sam and Greg and put them in charge of AI?


I find a bit odd that MS didn't have a couple of people in the board of directors, who I assume accepted Sam Altman resignation (and signed the severance package).

Edit: I just read he was fired, but the point remains.


Per my understanding, the board oversees the non-profit which owns the for-profit entity which Microsoft invested in. It's not clear to me how the non-profit board was picked but equity holders in the for-profit don't have a say on the matter.


Microsoft being Microsoft? Sounds like a political move to have greater power surface over the AI sector. A normal day in a monopolist afternoon.


dupe? doesn't say much beyond what we learn in the actual release

https://news.ycombinator.com/item?id=38309611


Well the first sentence in this news piece adds a lot more color to the official press release and stands out on its own:

"Microsoft, which has invested billions in OpenAI, learned that OpenAI was ousting CEO Sam Altman just a minute before the news was shared with the world, according to a person familiar with the situation."


"according to a person familiar with the situation."


Thankfully LLMs aren't that much secret sauce. Hope another company / open source can keep them on their feet.


Random aside, but I like Axios.

The site is mercifully clean (just turn off your adblocker and see what I mean).

Their scoops are good.

The format is succinct and efficient but not "dumb" like a short twitter thread.

What's more, none of this has changed over the years, somehow avoiding enshittification.


There goes their almost $13B investment down the drain.

Both masterminds of ChatGPT have left the company.

Feels like Nokia 2.0


> according to a person familiar with the situation.

Just a rumor. Zero chance someone at MS wasn't already aware.


If you’d ask around yesterday, every sane person would say there’s zero change Altman will be soon fired on the spot. We’re already past that point.


Usually news orgs don’t put their reputation on the line unless they feel there’s some substance


Bloomberg got away with conjuring The Big Hack story by intentionally mispresenting anecdotes and their reputation didn't take a single dent.


It definitely happens, but its not readily in their interest to turn themselves into just a rumor mill.


Confirmed by a second news source https://www.semafor.com/article/11/17/2023/openai-board-fire...

"Microsoft, which has invested billions in OpenAI, was blindsided by Altman’s firing, Axios reported and Semafor confirmed."


I'm pressing X really hard to doubt.


One "person familiar" is inappropriate sourcing for something so consequential. Not a good look for Axios.


In journalist speak, that means "credible person I know who would get fired if I said their name."

It's probably literally an exec on the OpenAI/Microsoft partnership.

The alternative to "a person familiar" is us as readers never get this information at all.


I think a lot of people don't understand how anonymous sourcing works in jounalism.


It's hard. I'm instinctively inclined to believe this story, but from first principles, why should I trust that Axios has adequately vetted this source? All I know about them is that lots of people in my circles send me their articles, I've never seen or conducted a review of their journalistic practices.

Obviously thinking about it this way would cause me to miss or disbelieve a lot of true stories, but it doesn't seem right to say I should trust every outlet I see widely posted either.


I know what it means and I agree it's probably an exec. The issue is that the premise - "Microsoft Corporation didn't know thing X at point in time Y" - is essentially unverifiable gossip, yet is presented here as fact


Believe it or don't, newspapers report on things which cannot be immediately verified all the time. Get over it.


> In journalist speak, that means "credible person I know who would get fired if I said their name."

those are some bold beliefs given the overall honesty in journalism and benefits to being first to publish.


There's a lot to criticize about Axios, but having high-profile connections is like the entire reason they exist. They don't make stuff up.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: