Hacker News new | past | comments | ask | show | jobs | submit login

My Ubuntu desktop has 64 gigs RAM, with a 12G RTX 3060 card. I have 4 bit 13B parameter LLaMA running on it currently, following these instructions - https://github.com/oobabooga/text-generation-webui/wiki/LLaM... . They don't have 30B or 65B ready yet.

Might try other methods to do 30B, or switch to my M1 Macbook if that's useful (as it said here). Don't have an immediate need for it, just futzing with it currently.

I should note that web link is to software for a gradio text generation web UI, reminiscent of Automatic1111.




By 30b and 65b, does it mean model with 30 or 65 million parameters?


Billion




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: