diff --git a/pkg/providers/definitions/llamacpp.yaml b/pkg/providers/definitions/llamacpp.yaml new file mode 100644 index 0000000..89ba2e8 --- /dev/null +++ b/pkg/providers/definitions/llamacpp.yaml @@ -0,0 +1,18 @@ +format_version: 1 +name: llamacpp +display_name: llama.cpp server +tier: 8 +last_verified: "2026-04-05" +keywords: + - "llama.cpp" + - "llama-cpp" + - "llama_cpp" + - "LLAMA_API_KEY" + - "ggml" + - "gguf" +verify: + method: GET + url: "" + headers: {} + valid_status: [] + invalid_status: [] diff --git a/pkg/providers/definitions/lmstudio.yaml b/pkg/providers/definitions/lmstudio.yaml new file mode 100644 index 0000000..57dc9c0 --- /dev/null +++ b/pkg/providers/definitions/lmstudio.yaml @@ -0,0 +1,17 @@ +format_version: 1 +name: lmstudio +display_name: LM Studio +tier: 8 +last_verified: "2026-04-05" +keywords: + - "lmstudio" + - "lm-studio" + - "LMSTUDIO_API_KEY" + - "localhost:1234" + - "lmstudio.ai" +verify: + method: GET + url: "" + headers: {} + valid_status: [] + invalid_status: [] diff --git a/pkg/providers/definitions/localai.yaml b/pkg/providers/definitions/localai.yaml new file mode 100644 index 0000000..b0f6896 --- /dev/null +++ b/pkg/providers/definitions/localai.yaml @@ -0,0 +1,17 @@ +format_version: 1 +name: localai +display_name: LocalAI +tier: 8 +last_verified: "2026-04-05" +keywords: + - "localai" + - "LOCALAI_API_KEY" + - "go-skynet" + - "localai.io" + - "localhost:8080" +verify: + method: GET + url: "" + headers: {} + valid_status: [] + invalid_status: [] diff --git a/pkg/providers/definitions/ollama.yaml b/pkg/providers/definitions/ollama.yaml new file mode 100644 index 0000000..ed8735f --- /dev/null +++ b/pkg/providers/definitions/ollama.yaml @@ -0,0 +1,19 @@ +format_version: 1 +name: ollama +display_name: Ollama +tier: 8 +last_verified: "2026-04-05" +keywords: + - "ollama" + - "OLLAMA_HOST" + - "OLLAMA_API_KEY" + - "OLLAMA_MODELS" + - "localhost:11434" + - "127.0.0.1:11434" + - "api/generate" +verify: + method: GET + url: "" + headers: {} + valid_status: [] + invalid_status: [] diff --git a/pkg/providers/definitions/vllm.yaml b/pkg/providers/definitions/vllm.yaml new file mode 100644 index 0000000..b360e05 --- /dev/null +++ b/pkg/providers/definitions/vllm.yaml @@ -0,0 +1,18 @@ +format_version: 1 +name: vllm +display_name: vLLM +tier: 8 +last_verified: "2026-04-05" +keywords: + - "vllm" + - "VLLM_API_KEY" + - "vllm-openai" + - "--api-key" + - "openai.api_server" + - "vllm.entrypoints" +verify: + method: GET + url: "" + headers: {} + valid_status: [] + invalid_status: [] diff --git a/providers/llamacpp.yaml b/providers/llamacpp.yaml new file mode 100644 index 0000000..89ba2e8 --- /dev/null +++ b/providers/llamacpp.yaml @@ -0,0 +1,18 @@ +format_version: 1 +name: llamacpp +display_name: llama.cpp server +tier: 8 +last_verified: "2026-04-05" +keywords: + - "llama.cpp" + - "llama-cpp" + - "llama_cpp" + - "LLAMA_API_KEY" + - "ggml" + - "gguf" +verify: + method: GET + url: "" + headers: {} + valid_status: [] + invalid_status: [] diff --git a/providers/lmstudio.yaml b/providers/lmstudio.yaml new file mode 100644 index 0000000..57dc9c0 --- /dev/null +++ b/providers/lmstudio.yaml @@ -0,0 +1,17 @@ +format_version: 1 +name: lmstudio +display_name: LM Studio +tier: 8 +last_verified: "2026-04-05" +keywords: + - "lmstudio" + - "lm-studio" + - "LMSTUDIO_API_KEY" + - "localhost:1234" + - "lmstudio.ai" +verify: + method: GET + url: "" + headers: {} + valid_status: [] + invalid_status: [] diff --git a/providers/localai.yaml b/providers/localai.yaml new file mode 100644 index 0000000..b0f6896 --- /dev/null +++ b/providers/localai.yaml @@ -0,0 +1,17 @@ +format_version: 1 +name: localai +display_name: LocalAI +tier: 8 +last_verified: "2026-04-05" +keywords: + - "localai" + - "LOCALAI_API_KEY" + - "go-skynet" + - "localai.io" + - "localhost:8080" +verify: + method: GET + url: "" + headers: {} + valid_status: [] + invalid_status: [] diff --git a/providers/ollama.yaml b/providers/ollama.yaml new file mode 100644 index 0000000..ed8735f --- /dev/null +++ b/providers/ollama.yaml @@ -0,0 +1,19 @@ +format_version: 1 +name: ollama +display_name: Ollama +tier: 8 +last_verified: "2026-04-05" +keywords: + - "ollama" + - "OLLAMA_HOST" + - "OLLAMA_API_KEY" + - "OLLAMA_MODELS" + - "localhost:11434" + - "127.0.0.1:11434" + - "api/generate" +verify: + method: GET + url: "" + headers: {} + valid_status: [] + invalid_status: [] diff --git a/providers/vllm.yaml b/providers/vllm.yaml new file mode 100644 index 0000000..b360e05 --- /dev/null +++ b/providers/vllm.yaml @@ -0,0 +1,18 @@ +format_version: 1 +name: vllm +display_name: vLLM +tier: 8 +last_verified: "2026-04-05" +keywords: + - "vllm" + - "VLLM_API_KEY" + - "vllm-openai" + - "--api-key" + - "openai.api_server" + - "vllm.entrypoints" +verify: + method: GET + url: "" + headers: {} + valid_status: [] + invalid_status: []