Llama-2 because you can run it on your own hardware. For the big GPU on a rented instance: Falcon 70b. OpenAI and Google can have turns playing proprietary asshat jack in the box.
How expensive would it be to run on a rented server?
For the bigger ones you could do it under $.50 / hr. I run llama 2 13b-8 bit on my 3090 no problem, which can be rented for $.20/hr.
Llama-2 because you can run it on your own hardware. For the big GPU on a rented instance: Falcon 70b. OpenAI and Google can have turns playing proprietary asshat jack in the box.
How expensive would it be to run on a rented server?
For the bigger ones you could do it under $.50 / hr. I run llama 2 13b-8 bit on my 3090 no problem, which can be rented for $.20/hr.
https://vast.ai/#pricing
Some of the lower pricing I've seen.