Hacker News new | past | comments | ask | show | jobs | submit | turingbook's comments login

I am curious whether this theory can explain some phenomena of large language models.


I am teaching my daughter computer science. Are the key concepts in this article good starting points for learning?


Open-weights models are not really open source.


I tried this on ChatGLM, a frontier foundation model developed by Zhipu.ai and Tsinghua University, and it gave the correct answer: https://chatglm.cn/share/FoZBJ


They are just trying to see if they can come up with something big.


A comment from Boris Power, an OpenAI guy: The top line number for MMLU is a bit gamed - Gemini is actually worse than GPT-4 when compared on normal few shot or chain of thought https://twitter.com/BorisMPower/status/1732435733045199126


Unfortunately, these are two different processing.


"Applying AI to core search algorithm. We’ve also applied the AI model to our core Bing search ranking engine, which led to the largest jump in relevance in two decades. With this AI model, even basic search queries are more accurate and more relevant." The large foundational model will change lots of things!


I guess GPT-4 needs its own launch event. :)


Peter Principle is not only for individuals, but also for teams, companies, or other orgs.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: