chore: ensure TOML configuration files are formatted well
This commit is contained in:
parent
3671e1d866
commit
b6f8f825e0
8
.vscode/extensions.json
vendored
Normal file
8
.vscode/extensions.json
vendored
Normal file
@ -0,0 +1,8 @@
|
|||||||
|
{
|
||||||
|
"recommendations": [
|
||||||
|
"tamasfe.even-better-toml",
|
||||||
|
"ms-python.black-formatter",
|
||||||
|
"ms-python.isort"
|
||||||
|
],
|
||||||
|
"unwantedRecommendations": []
|
||||||
|
}
|
17
.vscode/settings.json
vendored
Normal file
17
.vscode/settings.json
vendored
Normal file
@ -0,0 +1,17 @@
|
|||||||
|
{
|
||||||
|
"[python]": {
|
||||||
|
"editor.defaultFormatter": "ms-python.black-formatter",
|
||||||
|
"editor.codeActionsOnSave": {
|
||||||
|
"source.organizeImports": "always"
|
||||||
|
}
|
||||||
|
},
|
||||||
|
"[toml]": {
|
||||||
|
"editor.defaultFormatter": "tamasfe.even-better-toml",
|
||||||
|
},
|
||||||
|
"pre-commit-helper.runOnSave": "none",
|
||||||
|
"pre-commit-helper.config": ".pre-commit-config.yaml",
|
||||||
|
"evenBetterToml.schema.enabled": true,
|
||||||
|
"evenBetterToml.schema.associations": {
|
||||||
|
"^.+config[/\\\\].+\\.toml$": "../config/schema.config.json"
|
||||||
|
}
|
||||||
|
}
|
2
config/.gitignore
vendored
Normal file
2
config/.gitignore
vendored
Normal file
@ -0,0 +1,2 @@
|
|||||||
|
# prevent the local config file from being uploaded to the remote repository
|
||||||
|
config.toml
|
@ -1,10 +1,10 @@
|
|||||||
# Global LLM configuration
|
# Global LLM configuration
|
||||||
[llm]
|
[llm]
|
||||||
model = "claude-3-7-sonnet" # The LLM model to use
|
model = "claude-3-7-sonnet" # The LLM model to use
|
||||||
base_url = "https://api.openai.com/v1" # API endpoint URL
|
base_url = "https://api.openai.com/v1" # API endpoint URL
|
||||||
api_key = "sk-..." # Your API key
|
api_key = "sk-..." # Your API key
|
||||||
max_tokens = 8192 # Maximum number of tokens in the response
|
max_tokens = 8192 # Maximum number of tokens in the response
|
||||||
temperature = 0.0 # Controls randomness
|
temperature = 0.0 # Controls randomness
|
||||||
|
|
||||||
# [llm] #AZURE OPENAI:
|
# [llm] #AZURE OPENAI:
|
||||||
# api_type= 'azure'
|
# api_type= 'azure'
|
||||||
@ -25,11 +25,11 @@ temperature = 0.0 # Controls randomness
|
|||||||
|
|
||||||
# Optional configuration for specific LLM models
|
# Optional configuration for specific LLM models
|
||||||
[llm.vision]
|
[llm.vision]
|
||||||
model = "claude-3-7-sonnet" # The vision model to use
|
model = "claude-3-7-sonnet" # The vision model to use
|
||||||
base_url = "https://api.openai.com/v1" # API endpoint URL for vision model
|
base_url = "https://api.openai.com/v1" # API endpoint URL for vision model
|
||||||
api_key = "sk-..." # Your API key for vision model
|
api_key = "sk-..." # Your API key for vision model
|
||||||
max_tokens = 8192 # Maximum number of tokens in the response
|
max_tokens = 8192 # Maximum number of tokens in the response
|
||||||
temperature = 0.0 # Controls randomness for vision model
|
temperature = 0.0 # Controls randomness for vision model
|
||||||
|
|
||||||
# [llm.vision] #OLLAMA VISION:
|
# [llm.vision] #OLLAMA VISION:
|
||||||
# api_type = 'ollama'
|
# api_type = 'ollama'
|
||||||
|
Loading…
x
Reference in New Issue
Block a user