Files
goose/documentation/docs/troubleshooting.md
2025-01-25 11:11:41 -06:00

3.2 KiB

title
title
Troubleshooting

Troubleshooting

Goose, like any system, may run into occasional issues. This guide provides solutions for common problems.

Goose Edits Files

Goose can and will edit files as part of its workflow. To avoid losing personal changes, use version control to stage your personal edits. Leave Goose edits unstaged until reviewed. Consider separate commits for Goose's edits so you can easily revert them if needed.


Interrupting Goose

If Goose is heading in the wrong direction or gets stuck, you can interrupt it by pressing CTRL+C. This will stop Goose and give you the opportunity to correct its actions or provide additional information.


Stuck in a Loop or Unresponsive

In rare cases, Goose may enter a "death loop" or become unresponsive during a long session. This is often resolved by ending the current session, and starting a new session.

  1. Hold down Ctrl + C to cancel
  2. Start a new session:
goose session

:::tip For particularly large or complex tasks, consider breaking them into smaller sessions. :::


Context Length Exceeded Error

This error occurs when the input provided to Goose exceeds the maximum token limit of the LLM being used. To resolve this try breaking down your input into smaller parts. You can also use .goosehints as a way to provide goose with detailed context. Refer to the Using Goosehints Guide for more information.


Handling Rate Limit Errors

Goose may encounter a 429 error (rate limit exceeded) when interacting with LLM providers. The recommended solution is to use OpenRouter. See Handling LLM Rate Limits for more info.


API Errors

Users may run into an error like the one below when there are issues with their LLM API tokens, such as running out of credits or incorrect configuration:

Traceback (most recent call last):
  File "/Users/admin/.local/pipx/venvs/goose-ai/lib/python3.13/site-packages/exchange/providers/utils.py",
line 30, in raise_for_status
    response.raise_for_status()
    ~~~~~~~~~~~~~~~~~~~~~~~~~^^
  File "/Users/admin/.local/pipx/venvs/goose-ai/lib/python3.13/site-packages/httpx/_models.py",
line 829, in raise_for_status
    raise HTTPStatusError(message, request=request, response=self)
httpx.HTTPStatusError: Client error '404 Not Found' for url
'https://api.openai.com/v1/chat/completions'

...

This error typically occurs when LLM API credits are exhausted or your API key is invalid. To resolve this issue:

  1. Check Your API Credits:
    • Log into your LLM provider's dashboard
    • Verify that you have enough credits. If not, refill them
  2. Verify API Key:
    • Run the following command to reconfigure your API key:
    goose configure
    

For detailed steps on updating your LLM provider, refer to the Installation Guide.


Need Further Help?

If you have questions, run into issues, or just need to brainstorm ideas join the Discord Community!