Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Reasoning models like qwen3 are even better, and they have more options, for example you can choose the 14B model (at the usual 4KM quantization) instead of the 8B model.


Are they quantized more effectively than the non-reasoning models for some reason?


There is no difference, you can choose a 6 bits quantization if you prefer, at that point it's essentially lossless.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: