CC-Relay is a fast LLM API gateway written in Go. It routes Claude Code requests to multiple Anthropic-compatible providers. Operations teams use it to optimize LLM API calls, reduce costs, and improve reliability. It connects to various LLM providers like Anthropic, OpenAI, and Mistral.
git clone https://github.com/omarluq/cc-relay.gitCC-Relay is a fast LLM API gateway written in Go. It routes Claude Code requests to multiple Anthropic-compatible providers. Operations teams use it to optimize LLM API calls, reduce costs, and improve reliability. It connects to various LLM providers like Anthropic, OpenAI, and Mistral.
No install command available. Check the GitHub repository for manual installation instructions.
git clone https://github.com/omarluq/cc-relayCopy the install command above and run it in your terminal.
Launch Claude Code, Cursor, or your preferred AI coding agent.
Use the prompt template or examples below to test the skill.
Adapt the skill to your specific use case and workflow.
I'm setting up CC-Relay to optimize my Claude Code requests. I need to configure it to route requests to [PROVIDER_1], [PROVIDER_2], and [PROVIDER_3]. Can you provide a step-by-step guide for setting up CC-Relay with these providers and explain how to monitor the performance and cost savings?
## CC-Relay Configuration Guide
### Step 1: Install CC-Relay
1. Clone the CC-Relay repository from GitHub.
2. Navigate to the project directory.
3. Run `go build` to compile the application.
4. Run `./cc-relay` to start the service.
### Step 2: Configure Providers
1. Open the `config.yaml` file.
2. Add the following configuration for your providers:
```yaml
providers:
- name: Anthropic
api_key: your_anthropic_api_key
base_url: https://api.anthropic.com
- name: OpenAI
api_key: your_openai_api_key
base_url: https://api.openai.com
- name: Mistral
api_key: your_mistral_api_key
base_url: https://api.mistral.ai
```
3. Save the file.
### Step 3: Monitor Performance and Cost Savings
1. Use the built-in metrics endpoint to monitor request latency and success rates.
2. Analyze the cost savings by comparing the usage of each provider.
3. Adjust the provider configuration as needed to optimize performance and cost.
AI for humanity, built with safety first
Open-weight LLMs for enterprise AI deployment
Advanced foundation models via API and ChatGPT
Run large language models locally on your machine
Access business banking with customizable permissions, real-time notifications, and expense tracking.
AI-powered equity research for informed decisions
Take a free 3-minute scan and get personalized AI skill recommendations.
Take free scan