TL;DR: I shipped NoAIBills, a Chrome extension that runs Llama, DeepSeek, Qwen, and other LLMs entirely in your browser. No cloud, no Ollama, no subscriptions. It's free—I'm collecting emails and learning what people actually use local AI for.
Try it: noaibills.app
I wanted to use AI for everyday stuff—drafting emails, fixing grammar, summarizing docs, getting unstuck on code. But three things kept bugging me:
1. Privacy concerns. Every time I typed something even slightly sensitive into ChatGPT, I'd hesitate. Work stuff, personal notes, code I didn't want on someone else's server. I know I'm not the only one.
2. The "local AI" setup tax. Ollama is great, but try explaining it to a non-dev. "Open terminal, run this command, download a model..." Eyes glaze over. And if you're on a locked-down work laptop? Forget it.
3. Subscription fatigue. $20/month for something I use a few times a week felt like overkill. I didn't need GPT-4 for grammar checks.
So I thought: what if AI could run inside the browser? No installs, no servers, no monthly fees. Just a Chrome extension.
Turns out WebGPU makes this possible now. So I built it.
NoAIBills is a Chrome extension that runs open-source LLMs locally in your browser using WebGPU, Transformers.js, and Chrome's built-in Prompt API.
It's not trying to replace GPT-4 or Claude. It's for the 80% of tasks where you don't need a massive cloud model—drafts, summaries, quick coding help, grammar fixes.
I went back and forth on pricing. Considered $1, $5, $9. Here's where I landed:
I don't know enough yet.
I don't know who actually wants this, what they use it for, or whether they'd pay. Charging upfront would reduce installs and slow down learning.
So the deal is simple: it's free in exchange for your email and anonymous usage stats (via GA4). Your conversations stay completely private—I never see them.
The goal right now is to build an audience and figure out:
Maybe I'll monetize eventually. Maybe I won't. Right now I'm optimizing for learning, not revenue.
Local AI is good enough for most things. I'm not claiming it replaces GPT-4. But for the 80% of tasks—drafts, summaries, quick coding questions—a 3B parameter model running locally is plenty.
Not positioned as a cloud LLM replacement—it's for local inference on basic text tasks (writing, communication, drafts) with zero internet dependency, no API costs, and complete privacy.
Core fit: organizations with data restrictions that block cloud AI and can't install desktop tools like Ollama/LMStudio. For quick drafts, grammar checks, and basic reasoning without budget or setup barriers.
Need real-time knowledge or complex reasoning? Use cloud models. This serves a different niche—not every problem needs a sledgehammer 😄.
Tech stack for the curious:
The tricky part was making it work as a Chrome extension—CSP restrictions, WASM loading, static export. Wrote a bunch of post-build scripts to handle the edge cases.
One codebase, one build process. Dev mode runs locally, production becomes an extension.
I'm rolling this out gradually:
Spacing it out to avoid spam flags and to actually engage with each community.
1. Distribution is the hard part. Building it was fun. Getting people to find it? That's the real challenge. Chrome Web Store organic discovery is basically zero.
2. "Free" removes friction but adds noise. More installs, but harder to tell who's actually engaged vs. who installed and forgot.
3. Local AI is good enough for most things. A 3B parameter model running on your GPU handles 80% of what I used ChatGPT for. The other 20%? I still use Claude.
4. Be honest about limitations. My landing page has a "When NOT to use NoAIBills" section. I think that builds more trust than pretending it does everything.
If any of this resonates, I'd love for you to give it a spin:
And if you have feedback, questions, or just want to chat about browser-based AI—I'm here. Hit me up in the comments.
What would you do differently? Charge from day one? Different positioning? I'm all ears.