Files
goose/documentation/docs/configuration/providers.md
Bradley Axen 1c9a7c0b05 feat: V1.0 (#734)
Co-authored-by: Michael Neale <michael.neale@gmail.com>
Co-authored-by: Wendy Tang <wendytang@squareup.com>
Co-authored-by: Jarrod Sibbison <72240382+jsibbison-square@users.noreply.github.com>
Co-authored-by: Alex Hancock <alex.hancock@example.com>
Co-authored-by: Alex Hancock <alexhancock@block.xyz>
Co-authored-by: Lifei Zhou <lifei@squareup.com>
Co-authored-by: Wes <141185334+wesrblock@users.noreply.github.com>
Co-authored-by: Max Novich <maksymstepanenko1990@gmail.com>
Co-authored-by: Zaki Ali <zaki@squareup.com>
Co-authored-by: Salman Mohammed <smohammed@squareup.com>
Co-authored-by: Kalvin C <kalvinnchau@users.noreply.github.com>
Co-authored-by: Alec Thomas <alec@swapoff.org>
Co-authored-by: lily-de <119957291+lily-de@users.noreply.github.com>
Co-authored-by: kalvinnchau <kalvin@block.xyz>
Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com>
Co-authored-by: Rizel Scarlett <rizel@squareup.com>
Co-authored-by: bwrage <bwrage@squareup.com>
Co-authored-by: Kalvin Chau <kalvin@squareup.com>
Co-authored-by: Alice Hau <110418948+ahau-square@users.noreply.github.com>
Co-authored-by: Alistair Gray <ajgray@stripe.com>
Co-authored-by: Nahiyan Khan <nahiyan.khan@gmail.com>
Co-authored-by: Alex Hancock <alexhancock@squareup.com>
Co-authored-by: Nahiyan Khan <nahiyan@squareup.com>
Co-authored-by: marcelle <1852848+laanak08@users.noreply.github.com>
Co-authored-by: Yingjie He <yingjiehe@block.xyz>
Co-authored-by: Yingjie He <yingjiehe@squareup.com>
Co-authored-by: Lily Delalande <ldelalande@block.xyz>
Co-authored-by: Adewale Abati <acekyd01@gmail.com>
Co-authored-by: Ebony Louis <ebony774@gmail.com>
Co-authored-by: Angie Jones <jones.angie@gmail.com>
Co-authored-by: Ebony Louis <55366651+EbonyLouis@users.noreply.github.com>
2025-01-24 13:04:43 -08:00

3.9 KiB

sidebar_position
sidebar_position
1

import Tabs from '@theme/Tabs'; import TabItem from '@theme/TabItem';

Supported LLM Providers

You can use Goose with your preferred LLM. Goose supports a variety of LLM providers. To configure your chosen provider or see available options, run goose configure in the CLI or visit the Provider Settings page in the Goose UI.

1. Run the following command: ```sh goose configure ``` 2. Select `Configure Providers` from the menu.
![Provider Config](../assets/guides/goose-providers-cli.png)
**To update your LLM provider and API key:**
1. Click on the three dots in the top-right corner.
2. Select `Provider Settings` from the menu.
3. Click Edit, enter your API key, and click `Set as Active`.

Available Providers

  • OpenAI
  • Databricks
  • Ollama
  • Anthropic
  • Google Gemini
  • Groq
  • OpenRouter

OpenAI

OpenAI offers powerful language models that include GPT-4, GPT-3.5-turbo, and more.

  1. Run the following command and choose "Configure Providers":
    goose configure
    
  2. Select OpenAI from the list of available providers.
  3. Enter your OPENAI_API_KEY when prompted, which you can obtain by registering at OpenAI's platform.

Databricks

Databricks is a data analytics and AI platform that provides access to various AI models and tools. They offer integration with popular models and custom model deployment.

  1. Run the following command and choose "Configure Providers":
    goose configure
    
  2. Select Databricks as your provider.
  3. Enter your DATABRICKS_HOST and DATABRICKS_TOKEN, which can be generated in your Databricks Account Settings.

Ollama

Ollama is an open-source project that allows running large language models locally. It supports various open-source models and provides an API for integration.

  1. Run the following command and choose "Configure Providers":
    goose configure
    
  2. Select Ollama and follow the steps to download and set up your models as detailed on Ollama's site. Requires OLLAMA_HOST.

Anthropic

Anthropic is an AI research company that offers advanced language models through its API. Their primary model is Claude, which comes in various versions.

  1. Run the following command and choose "Configure Providers":
    goose configure
    
  2. Choose Anthropic and provide the ANTHROPIC_API_KEY, obtainable via Anthropic's platform.

Google Gemini

Google Gemini is a suite of large language models developed by Google. It offers multimodal capabilities and can be accessed through the Google AI Studio.

  1. Run the following command and choose "Configure Providers":
    goose configure
    
  2. Pick Google Gemini from the list of providers and input your GOOGLE_API_KEY. .

Groq

Groq is an AI company that offers high-performance inference for large language models. They provide access to various models through their API.

  1. Run the following command and choose "Configure Providers":
    goose configure
    
  2. Select Groq from the list of providers and input your GROQ_API_KEY, set up via the Groq Console.

OpenRouter

OpenRouter is a platform that provides access to multiple AI models from various providers through a single API. It simplifies the process of using different AI models in applications.

  1. Run the following command and choose "Configure Providers":

    goose configure
    
  2. Select OpenRouter from the list of providers and input your OPENROUTER_API_KEY, set up via the OpenRouter Console.