mirror of
https://github.com/aljazceru/chatgpt-telegram-bot.git
synced 2025-12-20 14:14:52 +01:00
increase default max tokens for gpt-4 models
This commit is contained in:
@@ -3,7 +3,7 @@ import os
|
|||||||
|
|
||||||
from dotenv import load_dotenv
|
from dotenv import load_dotenv
|
||||||
|
|
||||||
from openai_helper import OpenAIHelper
|
from openai_helper import OpenAIHelper, default_max_tokens
|
||||||
from telegram_bot import ChatGPTTelegramBot
|
from telegram_bot import ChatGPTTelegramBot
|
||||||
|
|
||||||
|
|
||||||
@@ -25,6 +25,8 @@ def main():
|
|||||||
exit(1)
|
exit(1)
|
||||||
|
|
||||||
# Setup configurations
|
# Setup configurations
|
||||||
|
model = os.environ.get('OPENAI_MODEL', 'gpt-3.5-turbo')
|
||||||
|
max_tokens_default = default_max_tokens(model=model)
|
||||||
openai_config = {
|
openai_config = {
|
||||||
'api_key': os.environ['OPENAI_API_KEY'],
|
'api_key': os.environ['OPENAI_API_KEY'],
|
||||||
'show_usage': os.environ.get('SHOW_USAGE', 'false').lower() == 'true',
|
'show_usage': os.environ.get('SHOW_USAGE', 'false').lower() == 'true',
|
||||||
@@ -33,11 +35,11 @@ def main():
|
|||||||
'max_history_size': int(os.environ.get('MAX_HISTORY_SIZE', 15)),
|
'max_history_size': int(os.environ.get('MAX_HISTORY_SIZE', 15)),
|
||||||
'max_conversation_age_minutes': int(os.environ.get('MAX_CONVERSATION_AGE_MINUTES', 180)),
|
'max_conversation_age_minutes': int(os.environ.get('MAX_CONVERSATION_AGE_MINUTES', 180)),
|
||||||
'assistant_prompt': os.environ.get('ASSISTANT_PROMPT', 'You are a helpful assistant.'),
|
'assistant_prompt': os.environ.get('ASSISTANT_PROMPT', 'You are a helpful assistant.'),
|
||||||
'max_tokens': int(os.environ.get('MAX_TOKENS', 1200)),
|
'max_tokens': int(os.environ.get('MAX_TOKENS', max_tokens_default)),
|
||||||
'n_choices': int(os.environ.get('N_CHOICES', 1)),
|
'n_choices': int(os.environ.get('N_CHOICES', 1)),
|
||||||
'temperature': float(os.environ.get('TEMPERATURE', 1.0)),
|
'temperature': float(os.environ.get('TEMPERATURE', 1.0)),
|
||||||
'image_size': os.environ.get('IMAGE_SIZE', '512x512'),
|
'image_size': os.environ.get('IMAGE_SIZE', '512x512'),
|
||||||
'model': os.environ.get('OPENAI_MODEL', 'gpt-3.5-turbo'),
|
'model': model,
|
||||||
'presence_penalty': int(os.environ.get('PRESENCE_PENALTY', 0)),
|
'presence_penalty': int(os.environ.get('PRESENCE_PENALTY', 0)),
|
||||||
'frequency_penalty': int(os.environ.get('FREQUENCY_PENALTY', 0)),
|
'frequency_penalty': int(os.environ.get('FREQUENCY_PENALTY', 0)),
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -16,6 +16,16 @@ GPT_4_MODELS = ("gpt-4", "gpt-4-0314")
|
|||||||
GPT_4_32K_MODELS = ("gpt-4-32k", "gpt-4-32k-0314")
|
GPT_4_32K_MODELS = ("gpt-4-32k", "gpt-4-32k-0314")
|
||||||
GPT_ALL_MODELS = GPT_3_MODELS + GPT_4_MODELS + GPT_4_32K_MODELS
|
GPT_ALL_MODELS = GPT_3_MODELS + GPT_4_MODELS + GPT_4_32K_MODELS
|
||||||
|
|
||||||
|
|
||||||
|
def default_max_tokens(model: str) -> int:
|
||||||
|
"""
|
||||||
|
Gets the default number of max tokens for the given model.
|
||||||
|
:param model: The model name
|
||||||
|
:return: The default number of max tokens
|
||||||
|
"""
|
||||||
|
return 1200 if model in GPT_3_MODELS else 2400
|
||||||
|
|
||||||
|
|
||||||
class OpenAIHelper:
|
class OpenAIHelper:
|
||||||
"""
|
"""
|
||||||
ChatGPT helper class.
|
ChatGPT helper class.
|
||||||
|
|||||||
Reference in New Issue
Block a user