From c2e0414d567c87f124c514647a49acb51b7331a4 Mon Sep 17 00:00:00 2001 From: Aljaz Ceru Date: Mon, 23 Jun 2025 17:27:37 +0200 Subject: [PATCH] readme --- README.md | 30 +++++++++--------------------- 1 file changed, 9 insertions(+), 21 deletions(-) diff --git a/README.md b/README.md index 648d74c..7318093 100644 --- a/README.md +++ b/README.md @@ -1,10 +1,7 @@ -# Enchanted Proxy for OpenRouter -This repository is specifically made for use with the [Enchanted project](https://github.com/gluonfield/enchanted/tree/main). -The original author of this proxy is [marknefedov](https://github.com/marknefedov/ollama-openrouter-proxy). +## Free Model Proxy for OpenRouter +This proxy automatically selects and uses free models from OpenRouter, exposing them through Ollama and OpenAI-compatible APIs. It is a work in progress, designed to utilize free resources available on OpenRouter automatically, without requiring manual model selection. -## Description -This repository provides a proxy server that emulates [Ollama's REST API](https://github.com/ollama/ollama) but forwards requests to [OpenRouter](https://openrouter.ai/). It uses the [sashabaranov/go-openai](https://github.com/sashabaranov/go-openai) library under the hood, with minimal code changes to keep the Ollama API calls the same. This allows you to use Ollama-compatible tooling and clients, but run your requests on OpenRouter-managed models. -Currently, it is enough for usage with [Jetbrains AI assistant](https://blog.jetbrains.com/ai/2024/11/jetbrains-ai-assistant-2024-3/#more-control-over-your-chat-experience-choose-between-gemini,-openai,-and-local-models). +It hasn't been extensively tested with paid models, but it should work with any OpenRouter model that is compatible with the OpenAI API. ## Features - **Free Mode (Default)**: Automatically selects and uses free models from OpenRouter with intelligent fallback. Enabled by default unless `FREE_MODE=false` is set. @@ -116,23 +113,10 @@ curl -X POST http://localhost:11434/v1/chat/completions \ }' ``` -## Installation -1. **Clone the Repository**: - - git clone https://github.com/your-username/ollama-openrouter-proxy.git - cd ollama-openrouter-proxy - -2. **Install Dependencies**: - - go mod tidy - -3. **Build**: - - go build -o ollama-proxy ## Docker Usage -### Using Docker Compose (Recommended) +### Using Docker Compose 1. **Clone the repository and create environment file**: ```bash @@ -149,7 +133,7 @@ curl -X POST http://localhost:11434/v1/chat/completions \ 3. **Run with Docker Compose**: ```bash - docker-compose up -d + docker compose up -d ``` The service will be available at `http://localhost:11434`. @@ -160,3 +144,7 @@ The service will be available at `http://localhost:11434`. docker build -t ollama-proxy . docker run -p 11434:11434 -e OPENAI_API_KEY="your-openrouter-api-key" ollama-proxy ``` + + +## Acknowledgements +Inspiration for this project was [xsharov/enchanted-ollama-openrouter-proxy](https://github.com/xsharov/enchanted-ollama-openrouter-proxy) who took inspiration from [marknefedov](https://github.com/marknefedov/ollama-openrouter-proxy). \ No newline at end of file