[-] grahamsz@alien.top 1 points 1 year ago

I can run VMWare's Open LLama 7B v2 Open Instruct on my laptop comfortably (though I have 64GB ram and 16GB VRAM) and my sense is that's it's probably somewhere between GPT2 and GPT3 in inference quality. It is, however, very slow. Even with my comparatively strong hardware, it's slow enough that I wouldn't want to use it in an interactive context (though it may be useful for background processing)

grahamsz

joined 1 year ago