Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

This reminds me of bias tuning, a LoRA competitor. One can get decent adapters by only finetuning a vector added to each linear layer activations. I think I saw it first while reading [1] but there are other instances.

[1] https://arxiv.org/pdf/2304.15010.pdf



Please try to share abstract links instead of pdf links, for mobile or low connection readers.


A fine suggestion. For you and others:

https://arxiv.org/abs/2304.15010

also available at:

https://doi.org/10.48550/arXiv.2304.15010




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: