A powerful, privacy-first AI chat application that runs entirely on your machine. No cloud dependencies, no data leaves your device — just you and your AI assistant.
Built for developers who want full control over their AI experience without sacrificing convenience.
- Privacy First — Your conversations never leave your machine
- Local & Fast — No API calls or internet required after setup
- Open Source — Full transparency, no vendor lock-in
- Customizable — Tweak system prompts and temperature to your liking
- Real-time AI chat with streaming responses
- Multiple model support (switch between models instantly)
- Chat history persistence (your chats are saved locally)
- Dark mode (easy on the eyes)
- Customizable system prompts
- Temperature control (adjust creativity vs precision)
- Multiple chat threads (organize conversations)
- Framework — Next.js
- Language — TypeScript
- Styling — Tailwind CSS
- AI Engine — Ollama
- Storage — localStorage
Install dependencies:
npm iPull the models you want to use:
ollama pull qwen2.5:3b
ollama pull llama3.2:1bStart the development server:
npm run devOpen http://localhost:5137 and start chatting!
Head to Ollama Library to browse and install models. Popular choices:
- deepseek r1:3b* — General purpose, excellent reasoning
- Qwen 3.5:4b — Strong coding capabilities
See Issues for planned features and progress.
Contributions welcome! Open an issue or PR to get started.
Built with care for developers, by developers.