16
Beginner questions thread
(sh.itjust.works)
Community to discuss about LLaMA, the large language model created by Meta AI.
This is intended to be a replacement for r/LocalLLaMA on Reddit.
You shouldn't need nvlink, I'm wondering if it's something to do with AWQ since I know that exllamav2 and llama.cpp both support splitting in oobabooga
I think you're right. Saw a post on Reddit basically mentioning the same things I'm seeing.
It looks like autoawq supports it but it might be an issue with how oobabooga implements it or something...