Reformat
This commit is contained in:
parent
b17c9d31a9
commit
49ccd72815
@ -151,7 +151,9 @@ class LLM:
|
||||
params["max_completion_tokens"] = self.max_tokens
|
||||
else:
|
||||
params["max_tokens"] = self.max_tokens
|
||||
params["temperature"] = temperature if temperature is not None else self.temperature
|
||||
params["temperature"] = (
|
||||
temperature if temperature is not None else self.temperature
|
||||
)
|
||||
|
||||
if not stream:
|
||||
# Non-streaming request
|
||||
@ -255,7 +257,9 @@ class LLM:
|
||||
params["max_completion_tokens"] = self.max_tokens
|
||||
else:
|
||||
params["max_tokens"] = self.max_tokens
|
||||
params["temperature"] = temperature if temperature is not None else self.temperature
|
||||
params["temperature"] = (
|
||||
temperature if temperature is not None else self.temperature
|
||||
)
|
||||
|
||||
response = await self.client.chat.completions.create(**params)
|
||||
|
||||
|
Loading…
x
Reference in New Issue
Block a user