Discussion
Ensu - Ente's Local LLM app
moqster: Heard the first time about them (ente) yesterday in a discussion about "which 2FA are u using?". Directly switched to https://ente.com/auth/ on Android and Linux Desktop and very happy with it.Going to give this a try...
_factor: You presumably had a working 2fa app already, but off the cuff decide to switch to new unvetted variant X; basically unknown auth system after reading a few paragraphs of text in an afternoon?Does this seem sound?
ahofmann: While I would have the same reaction, in this case I think it is a sane decision. Ente is cornering the privacy market and I think they're doing a great job. They have a lot to lose (trust) and it would be stupid if they did something shady with the data entered in the 2FA app.
juliushuijnk: I'm working on a rather simple idea; a Wordpress plugin that allows you to use a local LLM inside your wordpress CMS.It requires a Firefox add-on to act as a bridge: https://addons.mozilla.org/en-US/firefox/addon/ai-s-that-hel...There is honestly not much to test just yet, but feel free to check it out here, provide feedback on the idea: https://codeberg.org/Helpalot/ais-that-helpalotThe essence works, I was able to let it make a simple summary on CMS content. So next is making it do something useful, and making it clear how other plugins could use it.
VladVladikoff: Maybe I’m missing it but the page is really light on technical information. Is this a quantized / distilled model of a larger LLM? Which one? How many parameters? What quantization? What T/s can I expect? What are the VRAM requirements? Etc etc
NitpickLawyer: You can see what it uses here - https://github.com/ente-io/ente/blob/main/web/apps/ensu/src/...Either LFM2.5-1.6B-4bit or Qwen3.5-2B-8bit or Qwen3.5-4B-4bit
sync: Hmm, the Mac app downloaded gemma-3-4b-it-Q4_K_M.gguf for me (on an Apple M4) - maybe the desktop apps download different models?Though, I don't see any references to Gemma at all in the open source code...
HelloUsername: Spam? Ad?Also: "Your AI agent can now create, edit, and manage content on WordPress.com" https://wordpress.com/blog/2026/03/20/ai-agent-manage-conten...
nathan_compton: Please god stop letting LLMs write your copy. My brain just slides right over this slop. Perhaps you have a useful product but christ almighty I cannot countenance this boring machine generated text.
zaphod12: if it helps, I've used ente for a year and I really like it.
mkagenius: Had used cactus before - https://news.ycombinator.com/item?id=44524544Then moved to pocket pal now for local llm.
glitchc: This sounds like an ad.
bilekas: > use a local LLM inside your wordpress CMSFor when wordpress doesn't have enough exploits and bugs as it is. Also why bother with wordpress in the first place if you're already having an LLM spit out content for you ?
juliushuijnk: What's your point? Don't use LLM for CMS content? That my code is buggy? Or that people shouldn't trust the LLM they run on their computer on their own website?You can check the code for exploits yourself. And other than that it's just your LLM talking to your own website.> Also why bother with wordpress in the first placeWeird question, but sure, I use WordPress, because I have a website that I want to run with a simple CMS that can also run my custom Wordpress plugins.
franze: if you are into local LLMs check out apfelhttps://github.com/Arthur-Ficial/apfelApple Ai on the command line
echion: Looks excellent -- thanks; shame older intel macs don't get it (a lot of those still around)
juliushuijnk: Spam for what? This is hackernews, I'm "hacking something" to push more control to users.I'm talking about connecting Ollama to your wordpress.Not via MCP or something that's complicated for a relatively normal user. But thanks for the link.
juliushuijnk: It seems your link about the Wordpress variation validated my idea :)
gwerbret: As do most of the associated comments. I think we're surrounded by bots.
mschulze: I'm not a bot. Check my comment history and account age.
embedding-shape: Huh, 1.6B/2B/4B models, I guess they weren't joking when they said "not as powerful as ChatGPT or Claude Code". Also unsure why they said "Claude Code", it's not an CLI agent AFAIK?
dgb23: This seems to be a general chat app, but otherwise small models can be very effective within the right use cases and orchestration.
embedding-shape: > otherwise small models can be very effective within the right use cases and orchestrationvery limited amount of use cases, perhaps. As a generalized chat assistant? I'm not sure you'd be able to get anything of value out from them, but happy to be proven otherwise. I have all of those locally already, without fine-tuning, what use case could I try right now where any of those are "very effective"?
Mashimo: > Also unsure why they said "Claude Code", it's not an CLI agent AFAIK?Claude Code is a Desktop app as well.
jubilanti: There's dozens of local inference apps that basically wrap llama.cpp and someone else's GGUFs. The decentralized sync history part seems new? Not much else. But the advertisement copy is so insufferably annoying in how it presents this wrapper as a product.Have a comparison chart to Ollama, LMStudio, LocalAI, Exo, Jan.AI, GPT4ALL, PocketPal, etc.
yomismoaqui: The consfusing way AI companies like to name products is something to be studied.
QubridAI: This is the most important part of local AI maturing not just better models, but better productization of on-device inference for normal people.
bee_rider: There are so many wrappers that are obviously wrappers. I wonder if part of the value proposition here is that it is “like a product.” I have no idea if they actually achieve that, though, and doubt it really could be proven on a site.
embedding-shape: Ok, but "Claude Code"/"Claude Desktop" regardless is software, a tool, not a model/LLM. Doesn't make much sense as they've written it.
FusionX: Given how the blog is presented, I assumed this was something novel that solved a unique problem, maybe a local multi-modal assistant for your device.I installed it and it's none of that. It is a mere wrapper around small local LLM models. And, it's not even multi-modal! Anyone could've one-shotted this in Claude in an hour (I'm not exaggerating).What's the target audience here? Your average person doesn't care about the privacy value proposition (at least not by severely sacrificing chat model's quality). And users who do want that control can already install LMStudio/Llama.cpp (which is dead simple to setup).The actual release product should've been what's described in "What's next" section.> Instead of general chat, we shape Ensu to have a more specialized interface, say like a single, never-ending note you keep writing on, while the LLM offers suggestions, critiques, reminders, context, alternatives, viewpoints, quotes. A second brain, if you will.> A more utilitarian take, say like an Android Launcher, where the LLM is an implementation detail behind an existing interaction that people are already used to.> Your agent, running on your phone. No setup, no management, no manual backups. An LLM that grows with you, remembers you, your choices, manages your tasks, and has long-term memory and personality.
jubilanti: > Anyone could've one-shotted this in Claude in an hour (I'm not exaggerating).This probably could have been one-shotted with Sonnet, not even Opus. Given how over indexed they are on LLM coding, Haiku might even be able to do it.This is actually an interesting coding model benchmark task now that I think about it.
BaudouinVH: Installed it on a not-so-young laptop. It crashes immediately after launch. I blame the laptop.If Ente is reading this : please add requirements to make it run (how many RAM, etc.)
post-it: > Anyone could've one-shotted this in Claude in an hourI think they did. If you start the download and then open the sidebar and/or background the app, the download progress bar disappears and is replaced by the download button. If you press the download button again, the progress bar reappears at the correct point.I find that Claude often makes little statefulness mistakes like that. Human developers do too, but the slower and more iterative nature of human development makes it more likely that that would get caught.
xtracto: I would love to see a "distributed LLM" system, where people can easily setup a system to perform a "piece" of a "mega model" inference or training. Kind of like SETI@home but for an open LLM (like https://github.com/evilsocket/cake but massive )Ideally if you "participate" in the network, you would get "credits" to use it proportionally to how much GPU power you have provided to the network. Or if you can't, then buy credits (payment would be distributed as credits to other participants).That way we could build huge LLMs that area really open and are not owned by any network.I would LOVE to participate in building that as well.