Hacker News new | past | comments | ask | show | jobs | submit login

The entire point is that LLMs don't think, they're just regressions of information you'd find on the internet. And automation is a good thing. 90% of Americans were farmers in the 1700s; automation completely revolutionized that. Your final point is still a concern in general, specifically what will happen when menial labor no longer exists. At that point, you have to hope that all the wealth generated from automation will be redistributed back in the form of a basic income, otherwise yes, we'll have a lot of people with no means to survive.





Automation is a good thing? It definitely made the rich richer, but I'm not sure it made the average person happier. Depression rates are at an all-time high, after all.

They are not interpolating, which is what I think you meant to say, except for a loose definition humans would meet too.

What do you think of the latest thinking models, and what is your test of thinking?


An LLM is one very big nonlinear regression used to pick a token with a clearly defined input, output, and the corresponding weights. It's still far too straight-forward and non-dynamic (the weights aren't constantly changing even during a single inference) compared to the human brain.

As far as the latest "thinking" techniques, it's all about providing the correct input to get the desired output. If you look at the training data (the internet), the hardest and most ambiguous problems don't have a simple question input and answer response, they instead have a lot of back-and-forth before arriving at the answer, so you need to simulate that same back-and-forth to arrive at the desired answer. Unfortunately model architecture is still too simple to implicitly do this within the model itself, at least reliably.


Learning and thinking are separate things. Today's models think without learning -- they are frozen in time -- but this is a temporary state borne of the cost of training. I actually like it like this because we don't yet have impenetrable guardrails on these things.

> If you look at the training data (the internet), the hardest and most ambiguous problems don't have a simple question input and answer response, they instead have a lot of back-and-forth before arriving at the answer, so you need to simulate that same back-and-forth to arrive at the desired answer. Unfortunately model architecture is still too simple to implicitly do this within the model itself, at least reliably.

Today's thinking models iterate (with function calls and Internet queries) and even backtrack. They are not as reliable as humans but are demonstrating the hallmarks of thinking, I'd say.




Join us for AI Startup School this June 16-17 in San Francisco!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: