Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
andai
2 days ago
|
parent
|
context
|
favorite
| on:
DeepThought-8B: A small, capable reasoning model
Thanks. If I'm reading this right, the limiting factor on the intelligence of current LLMs is not the network size, nor training data (size/quality) but rather the architecture? Do we know of a better one for complex computations / "reasoning"?
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search: