v6.0 β€” Production Ready

The Universal
LLM Proxy

Self-hosted proxy server that enables Claude Code, Cursor, and AI coding tools to work with any LLM provider β€” with 60-80% cost reduction.

Get Started View on GitHub NPM
12+
Providers
60-80%
Cost Savings
400+
Tests
100%
Local Option

Built for Production

Enterprise-grade features for teams that demand reliability, performance, and cost control.

πŸ”Œ

Multi-Provider Support

Seamlessly switch between Ollama, AWS Bedrock, OpenRouter, Databricks, Azure, and more β€” without changing a single line of code.

πŸ’°

Token Optimization

Smart tool selection, prompt caching, and memory deduplication work together to reduce your LLM costs by 60-80%.

🧠

Long-Term Memory

Titans-inspired memory system persists context across sessions, enabling smarter and more contextual responses.

πŸ”§

Full MCP Integration

Complete tool calling support with Model Context Protocol for advanced agent workflows and custom integrations.

πŸ“Š

Enterprise Ready

Built-in circuit breakers, load shedding, metrics, health checks, and audit logging for production deployments.

🌐

Remote Endpoints

Connect to models on any machine β€” run inference on dedicated GPU servers and share resources across your entire team.

πŸ”€

Hierarchical Routing

Intelligent request routing across model tiers β€” automatically route simple tasks to faster models and complex ones to more capable models.

12+ LLM Providers

From free local models to enterprise cloud solutions β€” use what works best for your use case.

πŸ¦™
Ollama
Local
FREE
⚑
llama.cpp
Local
FREE
🎬
LM Studio
Local
FREE
🍎
MLX Server
Apple Silicon
FREE
☁️
AWS Bedrock
Cloud
πŸ›€οΈ
OpenRouter
Cloud
🧱
Databricks
Enterprise
πŸ”·
Azure OpenAI
Cloud
🟣
Azure Anthropic
Cloud
πŸ€–
OpenAI
Cloud
Z
z.ai
Cloud
πŸ’Ž
Vertex AI
Cloud

Works With Your Stack

Drop-in replacement for Anthropic API β€” no code changes required in any of your tools.

⚑
Claude Code CLI
Official Anthropic terminal assistant
πŸ“
Cursor IDE
AI-first code editor
πŸ’»
Codex CLI
OpenAI's terminal assistant
πŸ€–
ClawdBot
GitHub Copilot alternative
πŸ”—
Cline / Continue.dev
OpenAI-compatible VS Code extensions

Up and Running in Minutes

Three simple steps to unlock any LLM provider with your favorite coding tools.

1
Install Lynkr
Install globally via npm or Homebrew β€” your choice.
Terminal
# Using npm
npm install -g lynkr

# Or using Homebrew
brew tap vishalveerareddy123/lynkr && brew install lynkr
2
Configure Your Provider
Set your preferred LLM provider using environment variables.
Local with Ollama (FREE)
export MODEL_PROVIDER=ollama
export OLLAMA_MODEL=qwen2.5-coder:latest
Cloud with AWS Bedrock
export MODEL_PROVIDER=bedrock
export AWS_BEDROCK_API_KEY=your-api-key
export AWS_BEDROCK_MODEL_ID=anthropic.claude-3-5-sonnet-20241022-v2:0
3
Connect Your Tools
Point your AI coding tools to Lynkr and you're done.
Start Lynkr & Connect Claude Code
# Start the proxy server
npm start  # β†’ Running at http://localhost:8081

# Connect Claude Code CLI
export ANTHROPIC_BASE_URL=http://localhost:8081
export ANTHROPIC_API_KEY=dummy
claude

Architecture

Lynkr sits between your tools and any LLM provider, handling format conversion, caching, and optimization transparently.

       Claude Code  /  Cursor  /  Codex  /  ClawdBot
                          β”‚
                          β–Ό
                β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
                β”‚    Lynkr Proxy      β”‚
                β”‚   localhost:8081    β”‚
                β”‚                     β”‚
                β”‚  β€’ Format Conversionβ”‚
                β”‚  β€’ Prompt Caching   β”‚
                β”‚  β€’ Token Optimization
                β”‚  β€’ Tool Calling     β”‚
                β”‚  β€’ Memory System    β”‚
                β”‚  β€’ Metrics & Logs   β”‚
                β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
                           β”‚
           β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”Όβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
           β–Ό               β–Ό               β–Ό
      β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”    β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”    β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”
      β”‚  Local  β”‚    β”‚  Cloud  β”‚    β”‚ Remote  β”‚
      β”œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€    β”œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€    β”œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€
      β”‚ Ollama  β”‚    β”‚ Bedrock β”‚    β”‚GPU Serverβ”‚
      β”‚llama.cppβ”‚    β”‚OpenRouter    β”‚ (any IP)β”‚
      β”‚LM Studioβ”‚    β”‚Databricks    β”‚         β”‚
      β”‚MLX Server    β”‚Azure/OpenAI  β”‚         β”‚
      β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜    β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜    β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜

Ready to Get Started?

Join developers saving 60-80% on LLM costs while using their favorite coding tools with any provider.