Vane is an AI-powered answering engine.
-
Updated
Apr 9, 2026 - TypeScript
Vane is an AI-powered answering engine.
On-device audio AI runtime. Local first transcription, speaker diarization, TTS, and voice cloning with an OpenAI compatible API.
High-performance lightweight proxy and load balancer for LLM infrastructure. Intelligent routing, automatic failover and unified model discovery across local and remote inference backends.
Free, open-source alternative to Weavy AI, Krea Nodes, Freepik Spaces & FloraFauna AI — node-based AI workflow builder for generative image & video pipelines
SmarterRouter: An intelligent LLM gateway and VRAM-aware router for Ollama, llama.cpp, and OpenAI. Features semantic caching, model profiling, and automatic failover for local AI labs.
Small Language Model Inference, Fine-Tuning and Observability.
Open-Source Intelligent Command Layer
We gave AI agents a brain. Memory, planning, continuity, and self-repair — the missing cognitive architecture layer. Runs on your Mac.
Recallium is a local, self-hosted universal AI memory system providing a persistent knowledge layer for developer tools (Copilot, Cursor, Claude Desktop). It eliminates "AI amnesia" by automatically capturing, clustering, and surfacing decisions and patterns across all projects. It uses the MCP for universal compatibility and ensures privacy
emotional AI Companions for personal relationships
Your machine's AI brain. One 20MB binary gives every tool, script, and cron job shared AI memory + 114 API routes. Desktop app, CLI, Telegram — all connected. Rust-powered.
Run IBM Granite 4.0 locally on Raspberry Pi 5 with Ollama.This is a privacy-first AI. Your data never leaves your device because it runs 100% locally. There are no cloud uploads and no third-party tracking.
A private, local RAG (Retrieval-Augmented Generation) system using Flowise, Ollama, and open-source LLMs to chat with your documents securely and offline.
AI SMS Auto-Responder for Android. Turn your Android device into an autonomous AI communication hub. A Python-based SMS auto-responder running natively on Termux, powered by LLMs (OpenRouter/Ollama) with a sleek Web & Terminal UI.
Electron + Next.js desktop AI assistant that runs GGUF models locally using llama.cpp. Designed for offline use, portability, and zero-install deployment.
Self-hosted, multi-model AI inference server. Run LLMs, TTS, STT, embeddings, and image generation with an OpenAI-compatible API.
Defense-in-depth platform for running OpenClaw agents on personal hardware
Self-hosted AI chat interface with RAG, long-term memory, and admin controls. Works with TabbyAPI, Ollama, vLLM, and any OpenAI-compatible API.
A growing collection of practical filter, pipeline & tool extensions for Open WebUI — solving real usability gaps in local & cloud LLM workflows.
Production-ready guide for connecting OpenClaw to a Telegram Bot. Build a self-hosted Telegram AI Agent using OpenClaw Gateway, pairing, and streaming responses.
Add a description, image, and links to the self-hosted-ai topic page so that developers can more easily learn about it.
To associate your repository with the self-hosted-ai topic, visit your repo's landing page and select "manage topics."