- Remove all manually registered Pi tools (alpha_search, alpha_get_paper, alpha_ask_paper, alpha_annotate_paper, alpha_list_annotations, alpha_read_code, session_search, preview_file) and their wrappers (alpha.ts, preview.ts, session-search.ts, alpha-tools.test.ts) - Add Pi skill files for alpha-research, session-search, preview, modal-compute, and runpod-compute in skills/ - Sync skills to ~/.feynman/agent/skills/ on startup via syncBundledAssets - Add node_modules/.bin to Pi subprocess PATH so alpha CLI is accessible - Add /outputs extension command to browse research artifacts via dialog - Add Modal and RunPod as execution environments in /replicate and /autoresearch prompts - Remove redundant /alpha-login /alpha-logout /alpha-status REPL commands (feynman alpha CLI still works) - Update README, researcher agent, metadata, and website docs Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
1.5 KiB
1.5 KiB
name, description
| name | description |
|---|---|
| runpod-compute | Provision and manage GPU pods on RunPod for long-running experiments. Use when the user needs persistent GPU compute with SSH access, large datasets, or multi-step experiments. |
RunPod Compute
Use runpodctl CLI for persistent GPU pods with SSH access.
Setup
brew install runpod/runpodctl/runpodctl # macOS
runpodctl config --apiKey=YOUR_KEY
Commands
| Command | Description |
|---|---|
runpodctl create pod --gpuType "NVIDIA A100 80GB PCIe" --imageName "runpod/pytorch:2.4.0-py3.11-cuda12.4.1-devel-ubuntu22.04" --name experiment |
Create a pod |
runpodctl get pod |
List all pods |
runpodctl stop pod <id> |
Stop (preserves volume) |
runpodctl start pod <id> |
Resume a stopped pod |
runpodctl remove pod <id> |
Terminate and delete |
runpodctl gpu list |
List available GPU types and prices |
runpodctl send <file> |
Transfer files to/from pods |
runpodctl receive <code> |
Receive transferred files |
SSH access
ssh root@<IP> -p <PORT> -i ~/.ssh/id_ed25519
Get connection details from runpodctl get pod <id>. Pods must expose port 22/tcp.
GPU types
NVIDIA GeForce RTX 4090, NVIDIA RTX A6000, NVIDIA A40, NVIDIA A100 80GB PCIe, NVIDIA H100 80GB HBM3
When to use
- Long-running experiments needing persistent state
- Large dataset processing
- Multi-step work with SSH access between iterations
- Always stop or remove pods after experiments
- Check availability:
command -v runpodctl