Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Can someone remind me of anything Gary has contributed to AI?

Last I saw he hasn’t produced anything but general “pop” books on AI and being associated with MIT, which IMO has zero weight on applied or even at this point theoretical AI, as that is primarily coming out of corporate labs.

No new algorithms, frameworks, datasets, products, insights.

Why is this guy relevant enough to keep giving him attention, his entire ouvre is just anti-whatever is getting attention in the “AI” landscape

I don’t see him commenting on any other papers and he has no lab or anything

Someone make it make sense, or is it as simple as “yeah thing makes me feel bad, me repost, me repeat!”



I've come around to the opinion that he's a bad faith actor riding the anti-AI attention train. Everything that he has said has also been said by other, more reasonable people. To give a concrete example: for years Yann LeCun has been banging the drum that LLMs by themselves are insufficient to build general intelligence and that just scaling up will not be enough.

At some point I entertained a few discussions where Gary Marcus was participating but from what I remember, it would never go anywhere other than a focus on playing around with definitions. Even if he's not wrong about some of his claims, I think there are better people worth engaging with. The amount of insight to be gained from listening to Gary Marcus is akin to that of a small puddle.


Regardless of personal opinions about his style, Marcus has been proven correct on several fronts, including the diminishing returns of scaling laws and the lack of true reasoning (out of distribution generalizability) in LLM-type AI.

These are issues that the industry initially denied, only to (years) later acknowledge them as their "own recent discoveries" as soon as they had something new to sell (chain-of-thought approach, RL-based LLM, tbc.).


Care to explain further? He has made far more claims of the limitations of LLMs that have been proven false.

> diminishing returns of scaling laws

This was so obvious it didn't need mentioning. And what Gary really missed is that all you need are more axes to scale over and you can still make significant improvements. Think of where we are now vs 2023.

> lack of true reasoning (out of distribution generalizability) in LLM-type AI

To my understanding, this is one that he has gotten wrong. LLMs do have internal representations, exactly the kind that he predicted they didn't have.

> These are issues that the industry initially denied, only to (years) later acknowledge them

The industry denies all their limitations for hype. The academic literature has all of them listed plain as day. Gary isn't wrong because he's contradicted the hype of the tech labs, he's wrong because his short-term predictions were proven false in the literature he used to publish in. This was all in his efforts to peddle neurosymbolic architectures which were quickly replaced by tool use.


I’m just trying to find where all this hype is

I think the hype is coming from people who have no idea what is going on and just feeding on each other

Much like blockchain, metaverse or whatever was dominated by know nothings who spoke confidently to people even dumber than them

No professionals that have any experience or research credentials have made any crazy claims


The hype is coming from startups, big tech press releases, and grifters who have a vested interest in raising a ton of money from VCs and stakeholders, same as blockchain and metaverse. The difference is that there is a large legitimate body of research underneath deep learning that has been there for many years and remains (somewhat) healthy.

I would argue that the claim of "LLMs will never be able to do this" is crazy without solid mathematical proof, and is risky even with significant empirical evidence. Unfortunately, several professionals have resorted to this language.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: