voocel/litellm is ranked by live GitHub momentum and cross-source evidence. It moved +1 stars in 24h with momentum score 11.5.
0/5 channels firing
no linked package yet
last commit 11h ago
Each channel contributes 0-1. Per-channel tiers: GitHub (breakout 1.0 / hot 0.7 / rising 0.4), HN (front-page 1.0 / ≥3 mentions 0.7 / 1-2 mentions 0.4), Bluesky (≥5 mentions 1.0 / 2-4 0.7 / 1 0.4), dev.to (≥3 articles 1.0 / 2 0.7 / 1 0.4), Reddit (corpus-normalized 48h velocity), X (≥10 mentions 24h 1.0 / 3-9 0.7 / 1-2 0.4).
* Reddit bar shows a per-repo velocity proxy (raw score / 100); the score formula uses the corpus-normalized version so a single repo's bar may not match its contribution to the corpus-wide ranking.
No mentions on this channel in the last 7 days.
// QUIET HERE DOESN'T MEAN THE REPO IS DEAD — CHECK OTHER TABS
Free, local, open-source 24/7 Cowork app and OpenClaw for Gemini CLI, Claude Code, Codex, OpenCode, Qwen Code, Goose CLI, Auggie, and more | 🌟 Star if you like it!
The agent that grows with you
Glamourous agentic coding for all 💘
🪨 why use many token when few token do trick — Claude Code skill that cuts 65% of tokens by talking like caveman
Python SDK, Proxy Server (AI Gateway) to call 100+ LLM APIs in OpenAI (or native) format, with cost tracking, guardrails, loadbalancing and logging. [Bedrock, Azure, OpenAI, VertexAI, Cohere, Anthropic, Sagemaker, HuggingFace, VLLM, NVIDIA NIM]
A Claude skill that writes the accurate prompts for any AI tool. Zero tokens or credits wasted. Full context and memory retention
// KNOWN REPO · PACKAGE · LAUNCH · SITE SURFACES