feat: Update default AI model to meta-llama/llama-4-maverick:free in Teto and Roleplay cogs
This commit is contained in:
parent
4c377a47be
commit
835b8d37c9
@ -10,7 +10,7 @@ import aiohttp
|
||||
CONVERSATION_HISTORY_FILE = 'data/roleplay_conversations.json'
|
||||
|
||||
# Default AI model
|
||||
DEFAULT_AI_MODEL = "deepseek/deepseek-chat-v3-0324"
|
||||
DEFAULT_AI_MODEL = "meta-llama/llama-4-maverick:free"
|
||||
|
||||
def strip_think_blocks(text):
|
||||
# Removes all <think>...</think> blocks, including multiline
|
||||
|
@ -21,7 +21,7 @@ import aiohttp
|
||||
class TetoCog(commands.Cog):
|
||||
def __init__(self, bot: commands.Bot):
|
||||
self.bot = bot
|
||||
self._ai_model = "google/gemini-2.5-flash-preview" # Default model
|
||||
self._ai_model = "meta-llama/llama-4-maverick:free" # Default model
|
||||
self._api_endpoint = "https://openrouter.ai/api/v1/chat/completions" # Default endpoint
|
||||
|
||||
async def _teto_reply_ai_with_messages(self, messages, system_mode="reply"):
|
||||
|
Loading…
x
Reference in New Issue
Block a user