
Matt “Dev Ops” Trahan
For everyone saying that you can't download a model, it's probably because you don't have enough RAM. On ARM devices the LLM is forced to live in RAM when active. If your device isn't powerful enough it's not going to support a model. just get a PC guys, workstation not gaming, lots of RAM, with a lot of storage however many billion parameters you can just assume that's how many gigabytes you need 7B you're going to need 7-9 gigabytes, depending on quantization. Thanks for attending my TED talk
9 people found this review helpful

Sam M.
- Flag inappropriate
- Show review history
Quite good, its not fast on my device but that is because its an 5 year old phone. But the capability to rum small LLMs from ollama on your phone is cool. Being able to load any model from ollama would we nicer, but maybe there are some adjustments needed to run well on the phone. Oh, and no weird permission, just internet, thats nice!!
11 people found this review helpful

wade heying
It works pretty decent. However, after continuing a conversation for a while, it seems to run out of memory or something on my note 20 ultra. It begins to spew nonsense, getting stuck in what seems to be a loop. It would also be great to have the option to load other models.
3 people found this review helpful