From 01b7a6b240921b7177a1165ccf2df6349f2730bb Mon Sep 17 00:00:00 2001 From: AJ Isaacs Date: Tue, 24 Feb 2026 17:08:32 -0500 Subject: [PATCH] Bump health check max_completion_tokens to 16 gpt-5-nano can't produce output with max_completion_tokens=1. Co-Authored-By: Claude Opus 4.6 --- bot.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/bot.py b/bot.py index bcac6cf..e92bf21 100644 --- a/bot.py +++ b/bot.py @@ -146,7 +146,7 @@ class BCSBot(commands.Bot): await self.llm._client.chat.completions.create( model=self.llm.model, messages=[{"role": "user", "content": "hi"}], - max_completion_tokens=1, + max_completion_tokens=16, ) logger.info("LLM connectivity check passed.") except Exception as e: