Fix workflow continuation and provider setup gaps

This commit is contained in:
Advait Paliwal
2026-04-17 09:47:38 -07:00
parent d30506c82a
commit 9841342866
28 changed files with 359 additions and 36 deletions

View File

@@ -261,7 +261,7 @@ This usually means the release exists, but not all platform bundles were uploade
Workarounds:
- try again after the release finishes publishing
- pass the latest published version explicitly, e.g.:
curl -fsSL https://feynman.is/install | bash -s -- 0.2.24
curl -fsSL https://feynman.is/install | bash -s -- 0.2.25
EOF
exit 1
fi

View File

@@ -110,7 +110,7 @@ This usually means the release exists, but not all platform bundles were uploade
Workarounds:
- try again after the release finishes publishing
- pass the latest published version explicitly, e.g.:
& ([scriptblock]::Create((irm https://feynman.is/install.ps1))) -Version 0.2.24
& ([scriptblock]::Create((irm https://feynman.is/install.ps1))) -Version 0.2.25
"@
}

View File

@@ -117,13 +117,13 @@ These installers download the bundled `skills/` and `prompts/` trees plus the re
The one-line installer already targets the latest tagged release. To pin an exact version, pass it explicitly:
```bash
curl -fsSL https://feynman.is/install | bash -s -- 0.2.24
curl -fsSL https://feynman.is/install | bash -s -- 0.2.25
```
On Windows:
```powershell
& ([scriptblock]::Create((irm https://feynman.is/install.ps1))) -Version 0.2.24
& ([scriptblock]::Create((irm https://feynman.is/install.ps1))) -Version 0.2.25
```
## Post-install setup

View File

@@ -52,7 +52,7 @@ Amazon Bedrock (AWS credential chain)
Feynman verifies the same AWS credential chain Pi uses at runtime, including `AWS_PROFILE`, `~/.aws` credentials/config, SSO, ECS/IRSA, and EC2 instance roles. Once that check passes, Bedrock models become available in `feynman model list` without needing a traditional API key.
### Local models: LM Studio, Ollama, vLLM
### Local models: LM Studio, LiteLLM, Ollama, vLLM
If you want to use LM Studio, start the LM Studio local server, load a model, choose the API-key flow, and then select:
@@ -70,6 +70,22 @@ API key: lm-studio
Feynman attempts to read LM Studio's `/models` endpoint and prefill the loaded model id.
For LiteLLM, start the proxy, choose the API-key flow, and then select:
```text
LiteLLM Proxy (OpenAI-compatible gateway)
```
The default settings are:
```text
Base URL: http://localhost:4000/v1
API mode: openai-completions
Master key: optional, read from LITELLM_MASTER_KEY
```
Feynman attempts to read LiteLLM's `/models` endpoint and prefill model ids from the proxy config.
For Ollama, vLLM, or another OpenAI-compatible local server, choose:
```text