Over the past few weeks, I’ve been building LibreAI — a simple, privacy-focused AI chat app that runs entirely on your own infrastructure using open-source models like Mistral, LLaMA 3, and Phi via Ollama.

LibreAI is built with privacy and simplicity at its core. It streams responses in real-time, stores nothing, and has zero telemetry. No OpenAI, no account wall, no JavaScript bloat.

Try it live

💬 Discuss it on Hacker News

🧠 Why I Built LibreAI

A lot of modern AI tools work well, but they come with trade-offs: vendor lock-in, data collection, or heavy dependencies.

I wanted an AI assistant I could trust and control, powered by models that run locally — without relying on big cloud providers or complex frontend stacks.

So I built LibreAI as a lightweight, clean alternative.

What I focused on:

A UI that loads fast, even on slow networks

Real-time streaming output

Support for multiple open models via Ollama

Full self-hostability — no cloud APIs required

No React, no tracking, no unnecessary complexity

⚙️ The Stack

LibreAI is built using:

Go Fiber – backend web framework

HTMX – frontend interactivity without JS frameworks

TailwindCSS – utility-first styling

Ollama – for serving local LLMs

Plausible – privacy-first analytics (no cookies or tracking)

🧪 What Works (and What’s Coming)

Right now, LibreAI supports:

Real-time streamed responses

Multiple open models via Ollama

Fast performance with a lightweight UI

What I’m improving:

Mobile layout and responsiveness

Model selection UX

Easier deployment options

LibreAI is intentionally minimal.

💬 I'd Love Your Feedback

If you're working on AI tools, local model deployments, or care about privacy-first UX, I’d love to hear your thoughts.

What models are you running locally?
What do you look for in a self-hosted AI tool?
Does lightweight and private beat "smart and centralized" for you?

👉 Try Libre

🗣 Join the Hacker News discussion

Thanks for reading 🙌