Hacker News new | past | comments | ask | show | jobs | submit login
LTM-1, an LLM with a 5M token context window (magic.dev)
6 points by kristjansson on June 6, 2023 | hide | past | favorite | 2 comments



Absolutely no information about the model or the product? It's not transformer-based, so what is it?


I posted the link to see if there were any plausible theories, or concrete information.

Clearly it’s not an exact-attention transformer - perhaps some sort sparse / approximate attention, or recurrent-transformer-ish-thing like RWKV?

Their twitter announcement[0] does say it’s a novel architecture they’re calling a “Long Term Memory Network”. But who knows what that actually means.

[0] https://twitter.com/magicailabs/status/1666116949560967168




Consider applying for YC's Spring batch! Applications are open till Feb 11.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: