Merge branch 'worktree-agent-a090b6ec'
This commit is contained in:
16
pkg/providers/definitions/gpt4all.yaml
Normal file
16
pkg/providers/definitions/gpt4all.yaml
Normal file
@@ -0,0 +1,16 @@
|
||||
format_version: 1
|
||||
name: gpt4all
|
||||
display_name: GPT4All
|
||||
tier: 8
|
||||
last_verified: "2026-04-05"
|
||||
keywords:
|
||||
- "gpt4all"
|
||||
- "nomic-ai"
|
||||
- "GPT4ALL_API_KEY"
|
||||
- "gpt4all.io"
|
||||
verify:
|
||||
method: GET
|
||||
url: ""
|
||||
headers: {}
|
||||
valid_status: []
|
||||
invalid_status: []
|
||||
17
pkg/providers/definitions/jan.yaml
Normal file
17
pkg/providers/definitions/jan.yaml
Normal file
@@ -0,0 +1,17 @@
|
||||
format_version: 1
|
||||
name: jan
|
||||
display_name: Jan AI
|
||||
tier: 8
|
||||
last_verified: "2026-04-05"
|
||||
keywords:
|
||||
- "jan-ai"
|
||||
- "janhq"
|
||||
- "JAN_API_KEY"
|
||||
- "jan.ai"
|
||||
- "cortex-cpp"
|
||||
verify:
|
||||
method: GET
|
||||
url: ""
|
||||
headers: {}
|
||||
valid_status: []
|
||||
invalid_status: []
|
||||
18
pkg/providers/definitions/llamacpp.yaml
Normal file
18
pkg/providers/definitions/llamacpp.yaml
Normal file
@@ -0,0 +1,18 @@
|
||||
format_version: 1
|
||||
name: llamacpp
|
||||
display_name: llama.cpp server
|
||||
tier: 8
|
||||
last_verified: "2026-04-05"
|
||||
keywords:
|
||||
- "llama.cpp"
|
||||
- "llama-cpp"
|
||||
- "llama_cpp"
|
||||
- "LLAMA_API_KEY"
|
||||
- "ggml"
|
||||
- "gguf"
|
||||
verify:
|
||||
method: GET
|
||||
url: ""
|
||||
headers: {}
|
||||
valid_status: []
|
||||
invalid_status: []
|
||||
17
pkg/providers/definitions/lmstudio.yaml
Normal file
17
pkg/providers/definitions/lmstudio.yaml
Normal file
@@ -0,0 +1,17 @@
|
||||
format_version: 1
|
||||
name: lmstudio
|
||||
display_name: LM Studio
|
||||
tier: 8
|
||||
last_verified: "2026-04-05"
|
||||
keywords:
|
||||
- "lmstudio"
|
||||
- "lm-studio"
|
||||
- "LMSTUDIO_API_KEY"
|
||||
- "localhost:1234"
|
||||
- "lmstudio.ai"
|
||||
verify:
|
||||
method: GET
|
||||
url: ""
|
||||
headers: {}
|
||||
valid_status: []
|
||||
invalid_status: []
|
||||
17
pkg/providers/definitions/localai.yaml
Normal file
17
pkg/providers/definitions/localai.yaml
Normal file
@@ -0,0 +1,17 @@
|
||||
format_version: 1
|
||||
name: localai
|
||||
display_name: LocalAI
|
||||
tier: 8
|
||||
last_verified: "2026-04-05"
|
||||
keywords:
|
||||
- "localai"
|
||||
- "LOCALAI_API_KEY"
|
||||
- "go-skynet"
|
||||
- "localai.io"
|
||||
- "localhost:8080"
|
||||
verify:
|
||||
method: GET
|
||||
url: ""
|
||||
headers: {}
|
||||
valid_status: []
|
||||
invalid_status: []
|
||||
19
pkg/providers/definitions/ollama.yaml
Normal file
19
pkg/providers/definitions/ollama.yaml
Normal file
@@ -0,0 +1,19 @@
|
||||
format_version: 1
|
||||
name: ollama
|
||||
display_name: Ollama
|
||||
tier: 8
|
||||
last_verified: "2026-04-05"
|
||||
keywords:
|
||||
- "ollama"
|
||||
- "OLLAMA_HOST"
|
||||
- "OLLAMA_API_KEY"
|
||||
- "OLLAMA_MODELS"
|
||||
- "localhost:11434"
|
||||
- "127.0.0.1:11434"
|
||||
- "api/generate"
|
||||
verify:
|
||||
method: GET
|
||||
url: ""
|
||||
headers: {}
|
||||
valid_status: []
|
||||
invalid_status: []
|
||||
17
pkg/providers/definitions/tensorrt-llm.yaml
Normal file
17
pkg/providers/definitions/tensorrt-llm.yaml
Normal file
@@ -0,0 +1,17 @@
|
||||
format_version: 1
|
||||
name: tensorrt-llm
|
||||
display_name: NVIDIA TensorRT-LLM
|
||||
tier: 8
|
||||
last_verified: "2026-04-05"
|
||||
keywords:
|
||||
- "tensorrt-llm"
|
||||
- "trtllm"
|
||||
- "TRTLLM_API_KEY"
|
||||
- "tensorrt_llm"
|
||||
- "nvidia-nim"
|
||||
verify:
|
||||
method: GET
|
||||
url: ""
|
||||
headers: {}
|
||||
valid_status: []
|
||||
invalid_status: []
|
||||
17
pkg/providers/definitions/text-gen-webui.yaml
Normal file
17
pkg/providers/definitions/text-gen-webui.yaml
Normal file
@@ -0,0 +1,17 @@
|
||||
format_version: 1
|
||||
name: text-gen-webui
|
||||
display_name: text-generation-webui (oobabooga)
|
||||
tier: 8
|
||||
last_verified: "2026-04-05"
|
||||
keywords:
|
||||
- "text-generation-webui"
|
||||
- "oobabooga"
|
||||
- "TEXTGEN_API_KEY"
|
||||
- "text-gen-webui"
|
||||
- "localhost:5000"
|
||||
verify:
|
||||
method: GET
|
||||
url: ""
|
||||
headers: {}
|
||||
valid_status: []
|
||||
invalid_status: []
|
||||
17
pkg/providers/definitions/triton.yaml
Normal file
17
pkg/providers/definitions/triton.yaml
Normal file
@@ -0,0 +1,17 @@
|
||||
format_version: 1
|
||||
name: triton
|
||||
display_name: NVIDIA Triton Inference Server
|
||||
tier: 8
|
||||
last_verified: "2026-04-05"
|
||||
keywords:
|
||||
- "triton-inference-server"
|
||||
- "tritonserver"
|
||||
- "TRITON_API_KEY"
|
||||
- "triton_grpc"
|
||||
- "v2/models"
|
||||
verify:
|
||||
method: GET
|
||||
url: ""
|
||||
headers: {}
|
||||
valid_status: []
|
||||
invalid_status: []
|
||||
18
pkg/providers/definitions/vllm.yaml
Normal file
18
pkg/providers/definitions/vllm.yaml
Normal file
@@ -0,0 +1,18 @@
|
||||
format_version: 1
|
||||
name: vllm
|
||||
display_name: vLLM
|
||||
tier: 8
|
||||
last_verified: "2026-04-05"
|
||||
keywords:
|
||||
- "vllm"
|
||||
- "VLLM_API_KEY"
|
||||
- "vllm-openai"
|
||||
- "--api-key"
|
||||
- "openai.api_server"
|
||||
- "vllm.entrypoints"
|
||||
verify:
|
||||
method: GET
|
||||
url: ""
|
||||
headers: {}
|
||||
valid_status: []
|
||||
invalid_status: []
|
||||
Reference in New Issue
Block a user