Hacker Newsnew | past | comments | ask | show | jobs | submit | chid's commentslogin

interestingly that's what comes up when you search for pi-coding-agent rather than the new domain.


utterly unrelated, the RSP had nothing to do with their usage terms and was entirely about research and release of high-capability models.


Given the high bar of entry 160VRAM GPU - is there anything practical one can use this for?


The model being 32B could run in <20GB VRAM with Q4 quantization (minimal loss of quality), or 80GB unquantized at full fidelity. The quoted 160GB is for their recommended evaluation settings.

There's a few pre-quantized options[0] or you can quantize it yourself with llama.cpp[1]. You can run the resulting gguf with llama.cpp `llama-cli` or `llama-server`, with LM Studio or with Ollama.

0: https://huggingface.co/models?search=cwm%20q4%20gguf

1: https://huggingface.co/spaces/ggml-org/gguf-my-repo


I see, still a fair more VRAM than I have access to. Thanks for sharing that information.


Interesting though one would think this is also an obvious finding.

Quantifying this would be interesting though.


it definitely feels like it.


Took quite a while for it to show up oddly.


I can't think of any other than potentially oil and gas (though they probably use a lot of it in head office type environment).


Has this been implemented anywhere else in the world other than China?


Do they even have advertising normally?


yes i believe so


Anna doesn't have any advertising. Their income is purely driven off donations, most of which are part of subscription packages that offer faster downloads.


yeah meant more libgen, which i believe some instances have ads.

anna's has "donations for speed" and dark pattern hide the links to the fast external websites


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: