Three cents for every 1k prompt tokens. You pay another six cents per 1k generated tokens in addition to that.
At 8k context size, this adds up quickly. Depending on what you send, you can easily be out ~thirty cents per generation.
Three cents for every 1k prompt tokens. You pay another six cents per 1k generated tokens in addition to that.
At 8k context size, this adds up quickly. Depending on what you send, you can easily be out ~thirty cents per generation.
Claude 2 isn't free though, is it?
Either way, it does depends on what you want to use it for. Claude 2 is very biased towards positivity and it can be like pulling teeth if you're asking it to generate anything it even remotely disapproves of. In that sense, Claude 1 is the superior option.
Presumably you watermark all the training data.
At least, that's my first instinct.
You can make it as complicated as you want, of course.
Out of curiosity, what use-case did you find for it? I'm always interested to see how AI is actually applied in real settings.
Wayne June has excellent readings of Lovecraft's works.
I've used WSL to run deepspeed before because inexplicably microsoft didn't develop it for their own platform...
I read it a long time ago. The format is interesting, novel certainly. I suppose it's the selling point, over the prose.
To me it seemed like there were many competing "ways" to read it as well. Like a maze, you can go different paths. Do you read it front to back? Niggle through the citations? Thread back through the holes? It's not often you get a book that has this much re-read value.
The assertion that they cannot be cheap is funny, when Vicuna 13B was trained on all of $300.
Not $300,000. $300. And that gets you a model that's almost parity with ChatGPT.
It may be an opinion, but pointing it out won't make me like java any more.