This commit is contained in:
Matt Eng 2025-03-14 21:41:43 -07:00
parent b17c9d31a9
commit 49ccd72815

View File

@ -151,7 +151,9 @@ class LLM:
params["max_completion_tokens"] = self.max_tokens
else:
params["max_tokens"] = self.max_tokens
params["temperature"] = temperature if temperature is not None else self.temperature
params["temperature"] = (
temperature if temperature is not None else self.temperature
)
if not stream:
# Non-streaming request
@ -255,7 +257,9 @@ class LLM:
params["max_completion_tokens"] = self.max_tokens
else:
params["max_tokens"] = self.max_tokens
params["temperature"] = temperature if temperature is not None else self.temperature
params["temperature"] = (
temperature if temperature is not None else self.temperature
)
response = await self.client.chat.completions.create(**params)