Hacker News new | past | comments | ask | show | jobs | submit login
OpenAI – transformer debugger release (github.com/openai)
362 points by nmca 65 days ago | hide | past | favorite | 119 comments



Can’t help but think Elons lawsuit will trigger more releases by OpenAI. His core arguments are BS, but raised legitimate questions about their non-profit status and lack of non-profit related activity.


> raised legitimate questions about their non-profit status and lack of non-profit related activity

What is "non-profit related activity"? IANAL but the only justification needed for non-profit designation is not being for-profit.

And maybe their structure with a for-profit subsidiary is up for debate because of this. But you don't get to be non-profit by doing "good stuff" like open sourcing things. You get it by investing all profits back into the business.

Evil Corp for Doing Evil could be a non-profit if it invested all its profits back into the business.


> What is "non-profit related activity"? What is "non-profit related activity"? IANAL but the only justification needed for non-profit designation is not being for-profit.

Good thing you are NAL, because this is flat out wrong. "Non-profit" is a term used to refer to specific tax-exempt institutions under US tax law 501(c), and has nothing to do with investing profits "back into the business" (otherwise, you could have called Amazon or Uber "non-profits").

OpenAI[1] is specifically a 501(c)(3)[2] which explicitly requires the organization's purpose be one of the following "exempt" purposes:

>charitable, religious, educational, scientific, literary, testing for public safety, fostering national or international amateur sports competition, and preventing cruelty to children or animals. [3]

If it does not continue to fill one of those purposes, it can lose its tax-exempt nonprofit status.

[1] https://projects.propublica.org/nonprofits/organizations/810...

[2] https://www.irs.gov/charities-non-profits/charitable-organiz...

[3] https://www.irs.gov/charities-non-profits/charitable-organiz...


> Good thing you are NAL

I'm offended. Presumably if I was AL I would have gotten it right.


LOL


Lots of Lawyers


My concern is that Elon has shut the door with a sloppy lawsuit.

If some other entity (Government agency? Another company?) Comes around to pull this string, its going to set a bad precedent.


> His core arguments are BS, but raised legitimate questions about their non-profit status and lack of non-profit related activity.

These legitimate questions about their non-profit status ARE his core arguments. His complaints are the same as the popular notion on hackernews that OpenAI has failed to be open.


> These legitimate questions about their non-profit status ARE his core arguments.

Except for the portion where he advocated for a capped-profit arm which he would run under Tesla. You know that part that is shown in the OpenAI emails. It seemed he was ok with it when HE would stand to benefit.

You can't believe that Elon is "a great businessman" while also believing that he is totally altruistic and ignores all of the self serving impulses that have gotten him to where he is.


Elon had always advocated for a for-profit structure so that OpenAI can raise enough money to achieve its vision. In the email they are discussing how the for-profit structure could work. Nothing wrong with that especially when he was the GP.


GP is discussing legal arguments, not motivations.


While I am not a lawyer, I think it’s pretty apparent you can’t sue someone for damages for participating in an action that you conspired to be a part of.


Is that even true? If you are in talks for something but it falls through for whatever reason, if the thing the other people do violates some prior agreement with you then I don't see why you can't sue for damages. You would need to actually alter the agreement, not just discuss potentially altering it, no?


Of course he's not "totally altruistic"; nobody is. Of course he's self-serving; everyone is all the time. Even the most ostensibly selfless acts have a basis in feeling good about selflessness.

But I don't see how your dichotomy makes any sense. There are selfish poor people. There are altruistic poor people. There are selfish wealthy people. And there are altruistic wealthy people (disregarding blanket denials by far left wackos with emotionally derived selective criticisms of capitalism).

Eye-rolling conspiracy theories aside, Bill Gates is absolute proof that someone can be an objectively great businessman and objectively highly altruistic.

Seems to me that the impusles which got Musk to where he is an obsession with extreme risk taking, combined with sufficient education and clarity of vision to sometimes bend extreme odds in his favour. History is littered with examples of wealth incinerated in the pursuit of building cars and rockets and it says a lot that he's been associated with a multitude of ultra-unlikely successes.

Whereas many other billionaires are more directly driven by wealth accumulation, as evidenced by the trappings of extreme wealth which surround them.


> Of course he's not "totally altruistic"; nobody is. Of course he's self-serving; everyone is all the time. Even the most ostensibly selfless acts have a basis in feeling good about selflessness. But I don't see how your dichotomy makes any sense. There are selfish poor people. There are altruistic poor people. There are selfish wealthy people. And there are altruistic wealthy people (disregarding blanket denials by far left wackos with emotionally derived selective criticisms of capitalism).

All of this goes without saying.

> Eye-rolling conspiracy theories aside, Bill Gates is absolute proof that someone can be an objectively great businessman and objectively highly altruistic.

The biggest and most glaring difference here is that Bill Gates is not on record attempting to take board control of a for profit company which is attempting to make malaria vaccines (of the many public health efforts he is pushing).

> Seems to me that the impusles which got Musk to where he is an obsession with extreme risk taking, combined with sufficient education and clarity of vision to sometimes bend extreme odds in his favour. History is littered with examples of wealth incinerated in the pursuit of building cars and rockets and it says a lot that he's been associated with a multitude of ultra-unlikely successes.

> Whereas many other billionaires are more directly driven by wealth accumulation, as evidenced by the trappings of extreme wealth which surround them.

None of this has anything to do with the merits of the case and really just comes of a Musk-coddling, a favorite pastime here it seems.


The analogy would be if Bill Gates put millions of dollars into a new non-profit entity dedicated to developing new royalty-free malaria drugs. Then after using his money to develop a great malaria drug, they transferred the malaria drug assets into a for-profit entity, sold half of that entity to Pfizer, so that Pfizer can manufacture the drug and sell it to malaria patients in central Africa.

If that happened, I'd be surprised if Bill Gates didn't sue them.

(As for my latter paragraphs, I was replying to something you said which didn't have anything to do with the merits of the case either. I'd like to think you could extend me the same courtesy you extended yourself.)


They are part of his core argument, but not the entire argument. The (imo) shakier part of the argument is that he is entitled to damages even though he doesn't own shares in the company. This is atypical, most of the time you lose that right once you get rid of your shares. It seems like he is arguing that he was misled and sold his shares because of that, which will be hard to prove if true, and even if it is true, isn't really illegal.


I believe at the time it was reported that Musk and the OpenAI board chose to part ways due to his for-profit AI research (in Tesla) posing a conflict of interest to the OpenAI mission.

Hence his argument boils down to "You made me sell my shares because my 'closed' AI research conflicted with your 'open' non-profit ideals. However, you have since stoped to adhere to your mission statement, effectively seizing to publish your research findings and pursuing for-profit goals. As a consequence of this I have lost a bunch of money."

And as nuts as Musk is, I kind of see his point here.


That argument doesn't really hold a lot of water with me to be honest. If I sell (forced or otherwise) my shares in company because it isn't working on tech I agree with, and then they pivot and start working on tech I agree with and their shares pop, I am not entitled to sue them for damages.


In the scenario you described above, you would be entitled to sue for damages if you sold shares in a company under false pretenses because the other owners deceived you.

Imagine being the part-owner of a bread focused bakery. You tell the other owners that you think the business should focus on pastries – even if everybody agrees that bread products are more important – as pastries make more money (that money can in turn be invested into the bread business).

The other owners hard disagree and ask you to leave the company, because a) that does not fit into their bread-centric mission and b) you own a pastry-supplier in town and would obviously be profiting from this move. So you sell your shares and move on, no harm no foul.

But what if they then turn around and start producing pastries, claiming it's only doing so to financially support their bread business? What if they start entering lucrative deals with other pastry-suppliers and effectively stop making bread?

In this case I would argue that you would be entitled to some portion of your foregone gains. You sold your stock while being under the impression that the company would never go into the direction which you deem is the *only viable one*, only to find out that they did exactly what you suggested after you left.

> ... because it isn't working on tech I agree with, ...

Please notice that this is not what happened. They all agreed on the tech to be researched. The only disagreement was the business model and company structure required to fund the research.


> under false pretenses because the other owners deceived you.

Even your bread/pastry scenario doesn’t quite make this part clear. And this is what will be difficult to prove - did OpenAI know and discuss a rug-pull of “we willingly plan to say we’re going to be doing non-profit work, but actually our plan is to be a for-profit company”.

If their pivot was pre-meditated then I could see Musk having a case, but if they pivoted purely from market factors and realizing they wouldn’t be able to cut it as a non-profit, I’d think he’s SOL


Yes, I agree with your take.

But please reread the comment which is the root of this discussion: > The (imo) shakier part of the argument is that he is entitled to damages even though he doesn't own shares in the company.

I was never arguing that Musk is entitled to damages. I am merely arguing that it is possible to not own the shares of a company and be entitled to damages. Whether Musk specifically is owed damages is something a judge has to decide.


In the exact case above, you would have to prove intentional deceit (which is quite difficult to do) and even then that isn't actually illegal on its face a lot of the time in non publically traded companies. Further, the timeline isn't as compressed as your comment suggests, if the bread focused bakery pivoted to pastries 4-5 years after you sold your stake in it I do not agree that you would be entitled to damages at all, businesses can and should be allowed to pivot. Selling stock under the impression that a company wouldn't go in a direction you expect, is a normal part of investment and there is nothing wrong with it, people make bad bets all the time.


IANAL but I don't think that intentional deceit is necessary to prove Musk's case, negligent misrepresentation might suffice. And I agree that the judge ultimately has to rule whether or not the time-period between the ousting and pivot is reasonable; this decision would probably be based on precedence, the specifics of the case and the judges opinions.

However, please remember your comment which is the root of this discussion: > The (imo) shakier part of the argument is that he is entitled to damages even though he doesn't own shares in the company.

I was never arguing that Musk is entitled to damages. I am merely arguing that it is possible to not own the shares of a company and be entitled to damages.


Erm, you get it, it's right in your comment: lawsuit about non-profit, HN about open weights.


Remember, open source is just on the beginning and for recruiting purposes


Why do you think they are BS? You say it raised legitimate questions about their non-profit status and lack of non-profit related activity.


I think some believe Elon's argument is BS because Elon claims to be upset about OpenAI charging for their work. However, internal communications/texts show he wanted to do the same thing. So it seems like sour grapes that he isn't the one in charge.


An argument isn’t good or bad because of the motivations of the person behind it.

I get Elon is not popular right now. But jeez this is playground emotional level logic.


I detailed my understanding of this case in a sibling-comment to this one, but I think the fact that he wanted to "do the same thing" (read: use the for-profit engine to fund AI research) only makes his case stronger!

If I understand correctly, the OpenAI board asked Musk to remove himself and sell his stake as his for-profit pursuits conflicted with the mission of open, non-profit research. But after he left, they started a for-profit subsidiary and did exactly what they didn't want him doing. I could see how a judge might side with him.


Let's be real. They think Elon's arguments are BS since he bought Twitter and he destroyed their Leftie safe space.

They are also very upset with Elon's very public shift away from the current Left.

It's Elon derangement syndrome.


Elon’s shift to appeal to a more right-leaning base happened long before any of the OpenAI stuff.


> Why do you think they are BS?

Because me and everyone else here really hates Elon Musk.

> You say it raised legitimate questions about their non-profit status and lack of non-profit related activity.

As soon as OpenAI accepted that investment from Microsoft it has only been used as a for-profit proxy to have their own ‘DeepMind’ to bring down Google which they have tried for years with Bing and the Scroogled campaigns.

OpenAI is now no better than Google DeepMind and both are as closed and secretive and OpenAI ditched from their charter and non-profit status and continues to make excuses for not giving back their research.


> Because me and everyone else here really hates Elon Musk.

Speak for yourself.


How exactly?


When I said this last year, people said

"BUT THEY ARE DEMOCRAITIZING IT"

Its good to see that I was always right, even last year. I just didn't have a dad who owned an emerald mine.


People seem to imagine that Errol Musk is dead and Elon got a big inheritance from his daddy's estate. In fact Errol is still very much alive, became bankrupt sometime in the 1990s, and is now a deadbeat living off financial support from his sons.

Was there ever an emerald mine? Maybe. Though nobody has been able to find a scrap of evidence for it. Literally the only "evidence" is Errol's own boasts, which is that he bought a half-share of a Zambian mine for 100k ZAR ($50k USD). Hardly the stuff generational empires are made of.


The Zambian mine is the emerald mine everyone is talking about. From Walter Isaacson's biography:

> [Errol and a Zambian busienessman] agreed on a price [for Errol's plane], and instead of taking a payment in cash, Errol was given a portion of the emeralds produced at three small mines that the entrepreneur owned in Zambia. [...]

> Errol, who never had an ownership stake in the mine, expanded his trade by importing raw emeralds and having them cut in Johannesburg. "Many people came to me with stolen parcels," he says. "On trips overseas I would sell emeralds to jewelers. It was a cloak-and-dagger thing, because none of it was legal." After producing profits of roughly $210,000, his emerald business collapsed in the 1980s when the Russians created an artificial emerald in the lab. He lost all of his emerald earnings.

$210,000 in 1986 would be about $600,000 today.


Yes, all entirely consistent with what I said. Errol has told this story many times, including to Walter. But that’s all there is — a story. Is it true? Doesn’t seem particularly consequential even if true. $600k is a lot of money, but it’s hardly unimaginable wealth.


What about this emerald diamond that Elon got as a gift for an ex-girlfriend that was on auction (With picture of said girlfriend and Elon)

[1]:https://www.rrauction.com/auctions/lot-detail/34638440643009...

Wonder if it would be possible to somehow trace the origin of that emerald to where it came from. Anyway its probably the Musk Family office that was the winning bidder so we may never find out now.

I don't know why this guy keeps having emeralds come up in his life. This auction is hilarious given that he repeatedly denies having anything to do with Emeralds. In fact the origin of the Emerald story is Elon himself, in a Forbes article that has since been scrubbed from the internet.

The best I can find is this Snopes article that references the Forbes article [2]:https://www.snopes.com/news/2022/11/17/elon-musk-emerald-min...

Here is the quote:

JC: How do you handle fear?

EM: Company death – not succeeding with the company – causes me a lot more stress than physical danger. But I've been in physical danger before. The funny thing is I've not actually been that nervous. In South Africa, my father had a private plane we'd fly in incredibly dangerous weather and barely make it back. This is going to sound slightly crazy, but my father also had a share in an Emerald mine in Zambia. I was 15 and really wanted to go with him but didn't realize how dangerous it was. I couldn't find my passport so I ended up grabbing my brother's – which turned out to be six months overdue! So we had this planeload of contraband and an overdue passport from another person. There were AK-47s all over the place and I'm thinking, "Man, this could really go bad."


I've no doubt Errol told Elon that the mine existed. I've no doubt that Elon believed it when he was 15 years old, and believed it as recently as 2014. And maybe it did exist, or maybe it was a Nigerian Prince scam and Errol, embarrassed to admit he was duped, never recanted that claim. Who knows.

Meanwhile, there's still no trace of this supposed mine. You'd think with all of the obsession people have with this mythical origin story, someone would have found it by now.


if you bought 50k USD worth of $AAPL in 1990, by conservative estimates they will be worth 560 million USD worth today. I'd argue, that's generational stuff right there.


If you bought the winning lottery ticket, you’d be a lottery winner.


If grandmother had balls she would be a grandfather.


Did Errol Musk buy 50k worth of $AAPL in 1990?


Interesting to see the use of ruff and black in the same project. https://github.com/openai/transformer-debugger/blob/main/.pr...


And isort, which Ruff also supports.


I must say, understanding how transformers work is arguably the most important research problem in history, assuming that AGI can be achieved by just scaling up current LLM models on text, video, audio, etc.


That is a very big assumption.


It really depends on the definition.

"Better than the average human at most profitable tasks" is a much lower bar than most people on HN might think.

I have vendors who instead of filling out a web form which remembers their inputs and eventually even fills everything out for them instead print it out and fax it back in.

We're probably only about 2-3 years away from transformers being self-optimizing enough in prompts and evaluations to outpace the average worker in most tasks in most roles. (It won't necessarily be that much cheaper after the multiple passes and context windows required, and crucially probably won't be better at all tasks in most roles.)

If you define AGI as "better than any human at profitable tasks" or "better than average at all tasks" then yes, we're a long ways off and transformers alone probably won't get us there.


> "Better than the average human at most profitable tasks"

I think the HN crowd forgets that what really runs the world are min wage workers running around and doing real world things, not code monkeys and glorified type writers filling excel sheets. So yes, replacing the bullshit jobs we invented to keep people busy will be relatively easy, that's if you don't account for the fact you'll now have to create bullshit+ jobs to keep them busy

And even then we're far away, sure it can shit out code for a todo webapp and create semi realistic images of a monkey eating a burrito but that's about it. More than a year ago someone betted against me here that chatgpt would revolutionise the world in the next year, nothing happened really, geeks are excited, execs are buying the hype, tons of money is transferred, yet there was no 4th industrial revolution.

What happened though is that the web is flooded with absolutely useless content, amazon is full of ai generated books, students rely more and more on chatgpt to generate homeworks, thesis, "find" solutions, &c. it might very well end up being a net negative for the average joe in the long run


>I think the HN crowd forgets that what really runs the world are min wage workers running around and doing real world things, not code monkeys and glorified type writers filling excel sheets.

This is not true at all. How many products do you use that come primarily from minimum wage workers?

If a few people responsible for Google maps running stopped working the GDP loss would be much bigger than if magnitudes more minimum wage workers did the same.


Forget "how many", without those people, you'd end up without food on your table. Then the rest of the products you use wouldn't matter.


Also not true, I know the big farms of my state and they don't depend on minimum wage workers and I live in a third world country.

I doubt agricultural wages or truck drivers are minimum wage jobs where you live. Assuming US https://www.nass.usda.gov/Charts_and_Maps/Farm_Labor/fl_allw...


Farm work, especially work that doesn't require specialization (planting, maintaining, harvesting), is pretty much minimum wage work where I live, in Spain. Minimum wage here is ~1300 EUR / month. But it also differs wildly by region here, as some regions are really poor while others rich (relatively).

Besides the farm work, there is food processing workers (cutting, cleaning, basically assembly lines), packaging, workers at warehouses, people who work at the counters of the store, and all the support roles for those positions. If you go outside and eat, you have all the restaurant personnel to take into account as well.

There is a lot of low skilled labor that goes into making what we eat today. I'm not sure how you could possibly claim that none of those people are on minimum wage.


Not all of the work you cited is essential. Would society crumble without retail?

Minimum wage in Spain is significantly more money than anything I've made in my life. It's a very comfortable position for the vast majority of the world.

>There is a lot of low skilled labor that goes into making what we eat today. I'm not sure how you could possibly claim that none of those people are on minimum wage.

People doing essential work that isn't trivially replaceable have the bargaining power to charge more than the minimum wage in a moderately free market of human work, usually they do.


> Not all of the work you cited is essential. Would society crumble without retail?

Did I miss the part where the other comment mentioned retail, or where you respond to the half dozen other examples of essential work?

> Minimum wage in Spain is significantly more money than anything I've made in my life. It's a very comfortable position for the vast majority of the world.

Instead of moving the bar some more, could you just define what minimum wage would be an acceptable bar for you in this conversation?


Yes you missed retail, read it again.

https://uk.indeed.com/career/warehouse-worker/salaries

Do you really need me to Google every single essential position known before conceding that society is not maintained by minimum wage workers?


No, but I sure would like it if you defined it a bit better, since apparently we're now talking about UK numbers.


France is one of if not the biggest agriculture power in Europe, most farmers can't even generate a 35hr min wage equivalent while working 80+ hours a week.

20% of them live in poverty, half of them make less than 22k euros a year

Truck drivers earn between min wage and 150% of min wage, while being on the road every day and not having a social life, they drive 8 hours per day and sleep in their fucking truck while some code monkey makes 300k+/year coding memeojis at apple. Guess which ones will be automated first by openai lmao


>Truck drivers earn between min wage and 150% of min wage

Where are you getting this information? It's absolutely wrong. Long haul truckers (the one's you're saying don't have social lives because they drive 8 hours per day) make $71,196 on average in the US[1].

[1] https://www.ziprecruiter.com/Salaries/LONG-HAUL-Truck-Driver...


He is talking about France in the sentence before. There are barely any truckers in Germany with a German nationality. They are simply not competitive. Same goes for package delivery.

Just imagine what would happen to a trucker's salary in the US if it were to create a unified market with Mexico and all of Central America.

https://en.wikipedia.org/wiki/2004_enlargement_of_the_Europe...

It's not necessarily a bad thing. Economies of Eastern European countries have been growing after all and Western Europe does not have enough workers because of its demographics anyway. My take is, that everybody is winning, there is less poverty than before, but some sideffects look ugly for a while.


>half of them make less than 22k euros a year

I'd be extremely happy making this amount. Some people are just accustomed to easier lives.

>while some code monkey makes 300k+/year coding memeojis at apple.

Meme position for a privileged caste statistically irrespective of skill in an institution that can piss money on anything and still succeed.


Take a look at what happened to farm employment figures over the last 100 years.

It was a good thing.


The people picking trash in my street stopped working for 2 days and it looks like I live in some third world country now, two fucking days and it looks like I live in the middle of an open air dump

If trucks stopped deliveries every city would die in a week

If construction workers stopped building / maintaining we'd be in deep shit in 6 months or less

If the people in warehouses stopped working for a week the economy would tank like it rarely does

Nurses, doctors, bus/tram/train drivers, police, firefighters, ambulances, janitors, trash pickers, plumbers, sewer workers, electricians, people taking care of water treatment plants, power plants, teachers, social workers, ...

You could delete facebook, openai, instagram, twitter, netflix, tesla and 90% of startups from the face of the earth right now and I'd have the exact same life as yesterday. Remove any of the people I mentioned above and society would crumble in no time

And none of these are even remotely close to being automated at all, nobody cares about most of these jobs. But hey, here is a dancing kangaroo: https://www.youtube.com/watch?v=Zuivg5rz_aA


Are any of the positions you cited minimum wage workers where you live? Again, assuming US:

https://money.cnn.com/2016/02/24/news/economy/trash-workers-...

https://money.usnews.com/careers/best-jobs/garbage-collector...

>You could delete facebook, openai, instagram, twitter, netflix, tesla and 90% of startups from the face of the earth right now and I'd have the exact same life as yesterday. Remove any of the people I mentioned above and society would crumble in no time

Yes because you picked non-essential work. (?)


> Yes because you picked non-essential work. (?)

Then again tell me who we're automating out of the work force right now ? Trash pickers or code monkeys ? Truck drivers or artists ?


Growing up my trash was picked up by a human and the truck crew had two or three people on it jogging house to house to pick up trash as the driver slow rolled through the neighborhood.

Now my trash is serviced by one person who mostly never leaves the cab and who would be better described as a skilled machine operator than as a menial labor role. The work isn't completely automated but technology has reduced the job market for trash truck crews by two-thirds. I'm guessing the barrier is higher now too, requiring training and certifications to run the robotics on the truck instead of physical fitness being the primary prior qualification.


All of them? We've been working on and succeeding at automating physical tasks for decades.


Essential and non-replaceable are different concepts.


Even further, maybe the world would actually be better without these companies.

Now that there are great inventions like TikTok, teenagers are depressed as hell, and they don't go to meet each other to play soccer together, because the "social" networks are giving the illusion of having that connection.


> I think the HN crowd forgets that what really runs the world are min wage workers running around and doing real world things

Is this really true? It's certainly a nice soundbyte when you're making class arguments or trying to dunk on the "HN crowd", but I think it falls apart under any level of scrutiny.

Who keeps your lights on? Who drives your bus? Who manages your sewage? Who teaches your kids? Who builds your roads? None of them make minimum wage and would probably be a little insulted to be characterized as such.

It's pretty reductionist to call anyone outside our realm of tech a "min wage worker", they're just workers like or I. I think it's a pretty stupid and pointless exercise to subdivide people into useful or non-useful workers, serving no purpose but to further pet the smugness of HN AI skeptics.


I think this comment focuses too much on the “minimum wage” aspect - the core of the argument is that those are roles not at risk to AI in its present state, not necessarily the compensation aspect


>there was no 4th industrial revolution.

Yet. The industrial revolution didn't happen in a year either.


"More than a year ago"? Really? What did anyone think was going to happen in a year?

This sort of thing usually takes longer than you expect it to, and then it usually happens faster than you expect it to.


This all reminds me of my asshole great uncle making fun of me as a teenager circa 1997 while I was at the computer and on the internet.

Sarcastically asking me "Can you meet girls on that?" and then laughed.

He wasn't wrong in the short term but laughably wrong in the long term.


> "Better than the average human at most profitable tasks"

This is not the definition of AGI. You can't just make up a random definition to fit your argument, lol.


I'm actually all in on people making up new definitions for vague terms at the start of an argument as long as they're explicit about it.

And I particularly like this one, which is much more clearly measurable. If you feel AGI is taken, maybe we should coin this one as APGI or something like that


I don't think the main intention was to define AGI but to zoom in on an interpretation of AGI that would provide enough value to be revolutionary.


What is the definition? The definition of AGI is one of the central points of contention in the biggest industry legal battle.


It's only in contention because 1 of the sides has a tonne of money, a hurt ego, and is willing to pay lawyers to argue the sky is red in order to get revenge on his former colleagues. I don't think anyone would seriously claim OpenAI has achieved AGI today.


No, what they have is several narrow ASIs.


No that's the economically dominating definition. The philosophical one will happen much later or may never happen, but human society may change beyond recognition with the first one alone.


"The philosophical one" seems to get updated with every new breakthrough. 20 years ago, GPT3 would have been considered AGI (or "strong AI", as we called it back then).

https://en.wikipedia.org/wiki/Artificial_general_intelligenc...


Dennett describes it as real magic. The magic that can be performed is not considered real magic (it's merely a trick of confidence), whereas real magic is that which couldn't possibly be done.


My understanding is that AGI has no formal definition as it means different things to different people.

The poster here created his own definition, but what is wrong with that? He set a very specific bar to achieve, something that most "high-level" thinkers in the space have not really done. Isn't the point of discourse to bring your ideas to the table?


This is the correct way to approach the answer to when and how to achieve AGI. Otherwise please present here your Engineer Specification for defining AGI...

On a timeframe for achieving AGI: https://youtu.be/cEg8cOx7UZk?t=1658


This is literally Sam Altman's definition:

https://www.businessinsider.com/sam-altman-thinks-agi-replac...


There is no single definition of AGI. Performing most intellectual tasks human perform today is both general and a form of intelligence, so I too agree with it.


Robotics is more important to AGI, because the bulk of human intelligence comes from manipulating and navigating the physical world. Which includes a large amount of social interaction. Allan’s are tools to assist humans. They aren’t automating most jobs away anytime soon.


> I have vendors who instead of filling out a web form which remembers their inputs and eventually even fills everything out for them instead print it out and fax it back in.

Somewhere along the way we built computer that are so intuitive people find printing and faxing easier than our web apps. This isn't completely the fault of any single web app, users have a lot of learned avoidance because of bad experiences with many apps.

In the end completely automating the job ended up being easier than building a good interface for a human to do the job.


AI as a reconstruction of UX is one of the more interesting angles in the next few years.


Need to be better than an average expert. Humans are general intelligences since you can train a human to do anything, so a general intelligent machine needs to be able to be trained and become equal to human experts, matching an average untrained human isn't worth much.


Which might turn out to be correct. Might be wrong also. We have no priors to AGI developing. Only NGI, and we know preciously little about how to achieve NGI too, except the bedroom way.


We have a lot of priors - everything we've ever done has not produced AGI.

Maybe scaling transformers is the best way forward. I'm hopeful. But it's a big assumption that it will produce AGI.


In vitro fertilization too!


We already understand how transformers work: their architecture can learn to approximate a very large class of functions-on-sequences (specifically continuous sequence-to-sequence functions with compact support: https://arxiv.org/abs/1912.10077). It can do it more accurately than previous architectures like RNNs because transformers don't "forget" any information from prior items in the sequence. Training transformers to predict the next item in sequences eventually forces them to learn a function that approximates a world model (or at least a model of how the world behaves in the training text/data), and if they're large enough and trained with enough data then this world model is accurate enough for them to be useful for us.

If you're asking for understanding the actual internal world model they develop, it's basically equivalent to trying to understand a human brain's internal world model by analysing its neurons and how they fire.


>Training transformers to predict the next item in sequences eventually forces them to learn a function that approximates a world model

There is absolutely no proof for this statement.


Have you used ChatGPT? It can do (at least) simple reasoning, for example simple spatial reasoning or simple human behavior reasoning.


I suggest you start here:

https://ahtiahde.medium.com/limits-of-turing-machines-and-al...

If you don't have a CS background, I would suggest reading the wikipedia entries referenced in the medium article as well.


> it's basically equivalent to trying to understand a human brain's internal world model by analysing its neurons and how they fire

A major challenge here is that it's very hard/expensive to analyze these neurons, especially at any kind of scale within one human.

Not so with LLMs.


We shouldn't assume that either of those tasks is impossible


>most important research problem in history

That has to be some extremely narrowed version of all research that has happened (or will happen?)


Considering that knowing how knowing works is at the top of the ordo cognoscendi, it s not that narrow.


>assuming that AGI can be achieved by just scaling up current LLM models on text, video, audio, etc.

Is any sane person actually trying this?

I can't imagine an LLM ever going AGI.


> the most important research problem in history

Probably not.

> assuming that AGI can be achieved by just scaling up current LLM models

Lmao.


> Probably not.

Lmao.


We may well look back in future years and view the underlying approach introduced in Reexpress as among the more significant results of the first quarter of the 21st century. With Reexpress, we can generate reliable probability estimates over high-dimensional objects (e.g., LLMs), including in the presence of a non-trivial subset of distribution shifts seen in practice. A non-vacuous argument can be made that this solves the alignment/super-alignment problem (the ultimate goal of the line of work in the post above, and why I mention this here), because we can achieve this behavior via composition with networks of arbitrary size.

Because the parameters of the large neural networks are non-identifiable (in the statistical sense), we operate at the unit of analysis of labeled examples/exemplars (i.e., the observable data), with a direct connection between the Training set and the Calibration set.

This has important practical implications. It works with essentially any generative AI model. For example, we can build an 'uncertainty-aware GPT-4' for use in enterprise and professional settings, such as law: https://github.com/ReexpressAI/Example_Data/blob/main/tutori...

(The need for reliable, controllable estimates is critical regardless of any notion of AGI, since the existing LLMs are already getting baked into higher-risk settings, such as medicine, finance, and law.)


I hope you are not part of the founding team but if you are, you truly are doing your startup a disservice. Sharing your startup/ideas is great but doing it in the form of an advertisement "underlying approach introduced in Reexpress as among the more significant results of the first quarter of the 21st century" is just weird.


Yeah, I am all in for hustling but this post is way over the top, particularly for this forum.


Given his username I'd say.. he is.

Agreed: disclaimer needed.


Re: this is an advertisement for a product by its own employees.


This is pretty cool. I find it intriguing that we're neural surgery on LLMs!


How many transformers are in a typical LLM? Or is the whole thing considered a transformer?


The architecture as a whole is referred to as a transformer (autoregressive decoder-only architecture in the case of gpt style LLMs).

Note that there can be other types of LLMs too that are not necessarily transformer based (SSM, RNN, etc)


Normally, the LLM is composed of multiple transformer blocks, where each block consists of the (mutli-head) attention and fully-connected feedforward components. These are then stacked on top of each other to give the final output of the network.


Hmmm.. but what happens if you allow the LLM to access and query its own debugger? "Why did I answer this to that?" "What happens if I change slightly my assumptions"?


Yearly obligatory open source drop. What was it last time, whisper?


A very meagre attempt to look like they provide open source tools help the world safely make AGI




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: