Fix temperature using default if 0
This commit is contained in:
parent
3671e1d866
commit
b17c9d31a9
@ -151,7 +151,7 @@ class LLM:
|
|||||||
params["max_completion_tokens"] = self.max_tokens
|
params["max_completion_tokens"] = self.max_tokens
|
||||||
else:
|
else:
|
||||||
params["max_tokens"] = self.max_tokens
|
params["max_tokens"] = self.max_tokens
|
||||||
params["temperature"] = temperature or self.temperature
|
params["temperature"] = temperature if temperature is not None else self.temperature
|
||||||
|
|
||||||
if not stream:
|
if not stream:
|
||||||
# Non-streaming request
|
# Non-streaming request
|
||||||
@ -255,7 +255,7 @@ class LLM:
|
|||||||
params["max_completion_tokens"] = self.max_tokens
|
params["max_completion_tokens"] = self.max_tokens
|
||||||
else:
|
else:
|
||||||
params["max_tokens"] = self.max_tokens
|
params["max_tokens"] = self.max_tokens
|
||||||
params["temperature"] = temperature or self.temperature
|
params["temperature"] = temperature if temperature is not None else self.temperature
|
||||||
|
|
||||||
response = await self.client.chat.completions.create(**params)
|
response = await self.client.chat.completions.create(**params)
|
||||||
|
|
||||||
|
Loading…
x
Reference in New Issue
Block a user