Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
hislaziness
6 months ago
|
parent
|
context
|
favorite
| on:
Mistral NeMo
isn't it 2 bytes (fp16) per param. so 7b = 14 GB+some for inference?
ancientworldnow
6 months ago
|
next
[–]
This was trained to be run at FP8 with no quality loss.
hislaziness
6 months ago
|
parent
|
next
[–]
The model description on huggingface says - Model size - 12.2B params, Tensor type - BF16. Is the Tensor type different from the training param size?
fzzzy
6 months ago
|
prev
[–]
it's very common to run local models in 8 bit int.
qwertox
6 months ago
|
parent
[–]
Yes, but it's not common for the original model to be 8 bit int. The community can downgrade any model to 8 bit int, but it's always linked to quality loss.
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search: