Hacker Newsnew | past | comments | ask | show | jobs | submitlogin
MosaicBERT: Pretraining Bert from Scratch for $20 (mosaicml.com)
4 points by ashvardanian on Jan 2, 2024 | hide | past | favorite | 1 comment


Super cool article - this was a good reminder for me that innovation is still happening in the BERT realm.

Honestly, for task specific tasks methods like this seem like the way to go over the more general LLM.

Does anyone know if there is any benchmarks that show LLM performance on classification tasks? It’d be interesting to have data to back that up.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: