@bytelyst/llm-router (0.1.1)
Published 2026-03-29 19:38:58 +00:00 by bytelyst-admin
Installation
@bytelyst:registry=npm install @bytelyst/llm-router@0.1.1"@bytelyst/llm-router": "0.1.1"About this package
@bytelyst/llm-router
Pure-code LLM router for free-tier API providers. No LLM-in-the-loop — deterministic routing with automatic fallback, health tracking, and round-robin load distribution.
Features
- 4 free providers out of the box: Groq, OpenRouter, Together AI, Cerebras
- Prompt classification — regex-based detection of code/math/reasoning/creative prompts
- Smart selection — routes to the best model for each prompt category
- Round-robin — distributes load across providers to maximize free-tier usage
- Auto-fallback — retries on 429/5xx with next-best provider
- Health tracking — sliding-window stats (latency, error rate, rate-limit rate)
- Telemetry hook — log every routing decision for analysis
- OpenAI-compatible — same request/response format as OpenAI chat completions
- Zero dependencies — pure TypeScript, uses native
fetch
Quick Start
# Set at least one API key
export GROQ_API_KEY=gsk_...
export OPENROUTER_API_KEY=sk-or-...
export TOGETHER_API_KEY=...
export CEREBRAS_API_KEY=...
import { LlmRouter } from '@bytelyst/llm-router';
const router = new LlmRouter();
// Automatic routing — classifier picks best provider+model
const result = await router.chat({
messages: [{ role: 'user', content: 'Write a quicksort in TypeScript' }],
});
console.log(result.response.choices[0].message.content);
console.log(`Served by: ${result.provider}/${result.model} in ${result.totalLatencyMs}ms`);
Explicit Provider Routing
// Force a specific provider:model
const result = await router.chat({
messages: [{ role: 'user', content: 'Hello' }],
model: 'groq:llama-3.3-70b-versatile',
});
Telemetry
const router = new LlmRouter({
onTelemetry: entry => {
// entry: { event, provider, model, attempt, latencyMs, category, tokens?, error? }
console.log(`[${entry.event}] ${entry.provider}/${entry.model} — ${entry.latencyMs}ms`);
},
});
Health Monitoring
const snapshots = router.getHealth();
// Returns: HealthSnapshot[] with per-provider stats
// { provider, model, totalRequests, successes, rateLimits, errors, avgLatencyMs, p95LatencyMs, healthy }
Configuration
const router = new LlmRouter({
// Override default providers
providers: [...],
// Health window (default: 60s)
healthWindowMs: 120_000,
// Error rate to mark unhealthy (default: 50%)
errorThreshold: 0.4,
// Rate-limit rate to mark unhealthy (default: 30%)
rateLimitThreshold: 0.2,
// Request timeout (default: 30s)
timeoutMs: 15_000,
// Max retry attempts (default: 3)
maxRetries: 4,
});
Provider Selection Logic
- Classify prompt → code, math, reasoning, creative, or general
- Score each available model based on category match, speed tier, context window, and model size
- Filter unhealthy models (based on sliding-window error/rate-limit rates)
- Round-robin across top-scoring providers to spread rate-limit load
- Fallback on 429/5xx → exclude failed model, pick next best
Default Provider Registry
| Provider | Models | Speed | Strengths |
|---|---|---|---|
| Groq | Llama 3.3 70B, Llama 3.1 8B, Gemma 2 9B | ⚡ Fastest | General, reasoning, code |
| OpenRouter | DeepSeek R1, Llama 3.3 70B, Gemma 2 9B | Medium | Reasoning, code, math |
| Together | Llama 3.3 70B Turbo, DeepSeek R1 Distill | Medium | General, reasoning, code |
| Cerebras | Llama 3.3 70B | ⚡ Fastest | General, reasoning, code |
Adding Custom Providers
Any OpenAI-compatible endpoint works:
import { LlmRouter, DEFAULT_PROVIDERS } from '@bytelyst/llm-router';
const router = new LlmRouter({
providers: [
...DEFAULT_PROVIDERS,
{
name: 'my-provider',
baseUrl: 'https://my-api.example.com/v1',
apiKeyEnv: 'MY_PROVIDER_KEY',
rpmLimit: 60,
tpmLimit: 100_000,
models: [
{
id: 'my-model',
label: 'My Model',
contextWindow: 32_000,
strengths: ['general', 'code'],
speedTier: 2,
},
],
},
],
});
Dependencies
Development Dependencies
| ID | Version |
|---|---|
| typescript | ^5.7.0 |
| vitest | ^3.0.0 |
Details
2026-03-29 19:38:58 +00:00
Assets (1)
Versions (2)
View all
npm
6
latest
19 KiB
llm-router-0.1.1.tgz
19 KiB