You must log in or register to comment.
Llama-2 because you can run it on your own hardware. For the big GPU on a rented instance: Falcon 70b. OpenAI and Google can have turns playing proprietary asshat jack in the box.
How expensive would it be to run on a rented server?
For the bigger ones you could do it under $.50 / hr. I run llama 2 13b-8 bit on my 3090 no problem, which can be rented for $.20/hr.
Some of the lower pricing I’ve seen.