Discussion
Google Gemma 4 Runs Natively on iPhone With Full Offline AI Inference
mistic92: It runs on Android too, with AI Core or even with llama.cpp
andsoitis: is there a comparison of it running on iPhone vs. Android phones?
lrvick: You can run Android on just about anything so it boils down to Linux GPU benchmarks.
bossyTeacher: Is the output coherent though? I am yet to see a local model working on consumer grade hardware being actually useful.
a_paddy: I can try it for you
lrvick: I run qwen3.5 122b on a Framework Desktop at 35/ts as a daily driver doing security and OS systems and software engineering.Never paid an LLM provider and I have no reason to ever start.
camillomiller: Can we please ban content that is CLEARLY written by AI?
fsiefken: Qwen3.5-9b and Qwen3.5-27b are pretty coherent on my 24G android phone
stingraycharles: I find it fascinating that after all this time reporters still don’t even bother to proofread content for obvious AI tells. I guess nobody really cares anymore?
mixermachine: What spec of Framework Desktop do you run this on?
breisa: There is only one and for this model you need the one with 128GiB RAM.
pabs3: > edge AI deploymentIsn't the "edge" meant to be computing near the user, but not on their devices?
stingraycharles: No it does not. This is about as “edge” as AI gets.In a general sense, edge just means moving the computation to the user, rather than in a central cloud (although the two aren’t mutually exclusive, eg Cloudflare Workers)
pgt: Your device is the ultimate edge. The next frontier would be running models on your wetware.
acters: Man can't wait for AI in my brain. And then intelligence will be pay to win.
dax_: That bugged me too, so I started looking at other articles - they all look AI generated to me. Whole website should be banned.
ValleZ: There are many apps to run local LLMs on both iOS & Android
jeroenhd: Google's models work quite well on my Android phone. I haven't found a use case beyond generating shitposts, but the model does its job pretty well. It's not exactly ChatGPT, but minor things like "alter the tone of this email to make it more professional" work like a charm.You need a relatively beefy phone to run this stuff on large amounts of text, though, and you can't have every app run it because your battery wouldn't last more than an hour.I think the real use case for apps is more like going to be something like tiny, purpose-trained models, like the 270M models Google wants people to train and use: https://developers.googleblog.com/on-device-function-calling... With these things, you can set up somewhat intelligent situational automation without having to work out logic trees and edge cases beforehand.