⚡ Python-free Rust inference server — OpenAI-API compatible. GGUF + SafeTensors, hot model swap, auto-discovery, single binary. FREE now, FREE forever.
+72 stars 24h | +124 7d
0 in 24h | 0 sources
0/5 channels firing
no linked package yet
last commit 1mo ago
Each channel contributes 0-1. Per-channel tiers: GitHub (breakout 1.0 / hot 0.7 / rising 0.4), HN (front-page 1.0 / ≥3 mentions 0.7 / 1-2 mentions 0.4), Bluesky (≥5 mentions 1.0 / 2-4 0.7 / 1 0.4), dev.to (≥3 articles 1.0 / 2 0.7 / 1 0.4), Reddit (corpus-normalized 48h velocity).
* Reddit bar shows a per-repo velocity proxy (raw score / 100); the score formula uses the corpus-normalized version so a single repo's bar may not match its contribution to the corpus-wide ranking.
No mentions on this channel in the last 7 days.
// QUIET HERE DOESN'T MEAN THE REPO IS DEAD — CHECK OTHER TABS
Hundreds of models & providers. One command to find what runs on your hardware.
Auto pilot for Claude Code - connect multiple coding agents to a local LLM brain. 🆕 with a hive mind now
Reliable model swapping for any local OpenAI/Anthropic compatible server - llama.cpp, vllm, etc
AI gateway written in Go. Lightweight unified OpenAI-compatible API for OpenAI, Anthropic, Gemini, Groq, xAI & Ollama. LiteLLM alternative with observability, guardrails, streaming, costs and usage tracking.
From a goal to a task DAG, automatically. TypeScript-native multi-agent orchestration with multi-model teams and parallel execution. Three runtime dependencies.
Translate full-length books and documents with Ollama, OpenAI (comptatible), Gemini, Mistral, Poe or OpenRouter. Preserves formatting. Resumes where you left off. No file size limits.
// KNOWN REPO · PACKAGE · LAUNCH · SITE SURFACES