@ggerganov's LLaMA works on a Pixel 6! LLaMAs been waiting for this, and so have I

Mar 13, 2023 路 7:57 AM UTC

Generations are very slow, seems it's due to load time. Might be cause I'm running it on Termux
Does it uses Pixel 6 TPUs?
No, it's not even optimized for the pixel. I quantized it on my mac and transferred the weights over due to space issues on my phone
Oh man, I need to deploy this on my pixel 6! Is there a repo you have posted anywhere? I can't keep up with all these ai LLaMA advances!
Only 26 seconds / token but it鈥檚 something!
324secs on Pixel 6 is too slow for building an app around, but maybe this model version params cut be cut down somehow?
I wonder how it will do with @kdrag0n NestBox!
Does the pixel have that new tensor chip? Wonder if we could get it to leverage it to speed up generation
Holy shit is termux still a thing?