Hacker News new | past | comments | ask | show | jobs | submit login

I’ve trained tokenizers on medium-sized datasets (+5GB of text, although that could be considered small or large depending on who you ask) and have always found training quite fast. As in, it takes a couple minutes.

Maybe if we’re talking terabytes it might not scale as well but so far in my experience training tokenizers has never been an issue. It’s training models that takes ages.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: