Self-hosted proxy server that enables Claude Code, Cursor, and AI coding tools to work with any LLM provider β with 60-80% cost reduction.
Enterprise-grade features for teams that demand reliability, performance, and cost control.
Seamlessly switch between Ollama, AWS Bedrock, OpenRouter, Databricks, Azure, and more β without changing a single line of code.
Smart tool selection, prompt caching, and memory deduplication work together to reduce your LLM costs by 60-80%.
Titans-inspired memory system persists context across sessions, enabling smarter and more contextual responses.
Complete tool calling support with Model Context Protocol for advanced agent workflows and custom integrations.
Built-in circuit breakers, load shedding, metrics, health checks, and audit logging for production deployments.
Connect to models on any machine β run inference on dedicated GPU servers and share resources across your entire team.
Intelligent request routing across model tiers β automatically route simple tasks to faster models and complex ones to more capable models.
From free local models to enterprise cloud solutions β use what works best for your use case.
Drop-in replacement for Anthropic API β no code changes required in any of your tools.
Three simple steps to unlock any LLM provider with your favorite coding tools.
# Using npm
npm install -g lynkr
# Or using Homebrew
brew tap vishalveerareddy123/lynkr && brew install lynkr
export MODEL_PROVIDER=ollama
export OLLAMA_MODEL=qwen2.5-coder:latest
export MODEL_PROVIDER=bedrock
export AWS_BEDROCK_API_KEY=your-api-key
export AWS_BEDROCK_MODEL_ID=anthropic.claude-3-5-sonnet-20241022-v2:0
# Start the proxy server
npm start # β Running at http://localhost:8081
# Connect Claude Code CLI
export ANTHROPIC_BASE_URL=http://localhost:8081
export ANTHROPIC_API_KEY=dummy
claude
Lynkr sits between your tools and any LLM provider, handling format conversion, caching, and optimization transparently.
Claude Code / Cursor / Codex / ClawdBot
β
βΌ
βββββββββββββββββββββββ
β Lynkr Proxy β
β localhost:8081 β
β β
β β’ Format Conversionβ
β β’ Prompt Caching β
β β’ Token Optimization
β β’ Tool Calling β
β β’ Memory System β
β β’ Metrics & Logs β
ββββββββββββ¬βββββββββββ
β
βββββββββββββββββΌββββββββββββββββ
βΌ βΌ βΌ
βββββββββββ βββββββββββ βββββββββββ
β Local β β Cloud β β Remote β
βββββββββββ€ βββββββββββ€ βββββββββββ€
β Ollama β β Bedrock β βGPU Serverβ
βllama.cppβ βOpenRouter β (any IP)β
βLM Studioβ βDatabricks β β
βMLX Server βAzure/OpenAI β β
βββββββββββ βββββββββββ βββββββββββ
Join developers saving 60-80% on LLM costs while using their favorite coding tools with any provider.