Hacker News new | past | comments | ask | show | jobs | submit login

isn't it 2 bytes (fp16) per param. so 7b = 14 GB+some for inference?



This was trained to be run at FP8 with no quality loss.


The model description on huggingface says - Model size - 12.2B params, Tensor type - BF16. Is the Tensor type different from the training param size?


it's very common to run local models in 8 bit int.


Yes, but it's not common for the original model to be 8 bit int. The community can downgrade any model to 8 bit int, but it's always linked to quality loss.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: