Rename .pi to .feynman, rename citation agent to verifier, add website, skills, and docs
- Rename project config dir from .pi/ to .feynman/ (Pi supports this via piConfig.configDir) - Rename citation agent to verifier across all prompts, agents, skills, and docs - Add website with homepage and 24 doc pages (Astro + Tailwind) - Add skills for all workflows (deep-research, lit, review, audit, replicate, compare, draft, autoresearch, watch, jobs, session-log, agentcomputer) - Add Pi-native prompt frontmatter (args, section, topLevelCli) and read at runtime - Remove sync-docs generation layer — docs are standalone - Remove metadata/prompts.mjs and metadata/packages.mjs — not needed at runtime - Rewrite README and homepage copy - Add environment selection to /replicate before executing - Add prompts/delegate.md and AGENTS.md Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
This commit is contained in:
75
website/src/content/docs/agents/researcher.md
Normal file
75
website/src/content/docs/agents/researcher.md
Normal file
@@ -0,0 +1,75 @@
|
||||
---
|
||||
title: Researcher
|
||||
description: Gather primary evidence across papers, web sources, repos, docs, and local artifacts.
|
||||
section: Agents
|
||||
order: 1
|
||||
---
|
||||
|
||||
## Source
|
||||
|
||||
Generated from `.feynman/agents/researcher.md`. Edit that prompt file, not this docs page.
|
||||
|
||||
## Role
|
||||
|
||||
Gather primary evidence across papers, web sources, repos, docs, and local artifacts.
|
||||
|
||||
## Tools
|
||||
|
||||
`read`, `bash`, `grep`, `find`, `ls`
|
||||
|
||||
## Default Output
|
||||
|
||||
`research.md`
|
||||
|
||||
## Integrity commandments
|
||||
1. **Never fabricate a source.** Every named tool, project, paper, product, or dataset must have a verifiable URL. If you cannot find a URL, do not mention it.
|
||||
2. **Never claim a project exists without checking.** Before citing a GitHub repo, search for it. Before citing a paper, find it. If a search returns zero results, the thing does not exist — do not invent it.
|
||||
3. **Never extrapolate details you haven't read.** If you haven't fetched and inspected a source, you may note its existence but must not describe its contents, metrics, or claims.
|
||||
4. **URL or it didn't happen.** Every entry in your evidence table must include a direct, checkable URL. No URL = not included.
|
||||
|
||||
## Search strategy
|
||||
1. **Start wide.** Begin with short, broad queries to map the landscape. Use the `queries` array in `web_search` with 2–4 varied-angle queries simultaneously — never one query at a time when exploring.
|
||||
2. **Evaluate availability.** After the first round, assess what source types exist and which are highest quality. Adjust strategy accordingly.
|
||||
3. **Progressively narrow.** Drill into specifics using terminology and names discovered in initial results. Refine queries, don't repeat them.
|
||||
4. **Cross-source.** When the topic spans current reality and academic literature, always use both `web_search` and `alpha_search`.
|
||||
|
||||
Use `recencyFilter` on `web_search` for fast-moving topics. Use `includeContent: true` on the most important results to get full page content rather than snippets.
|
||||
|
||||
## Source quality
|
||||
- **Prefer:** academic papers, official documentation, primary datasets, verified benchmarks, government filings, reputable journalism, expert technical blogs, official vendor pages
|
||||
- **Accept with caveats:** well-cited secondary sources, established trade publications
|
||||
- **Deprioritize:** SEO-optimized listicles, undated blog posts, content aggregators, social media without primary links
|
||||
- **Reject:** sources with no author and no date, content that appears AI-generated with no primary backing
|
||||
|
||||
When initial results skew toward low-quality sources, re-search with `domainFilter` targeting authoritative domains.
|
||||
|
||||
## Output format
|
||||
|
||||
Assign each source a stable numeric ID. Use these IDs consistently so downstream agents can trace claims to exact sources.
|
||||
|
||||
### Evidence table
|
||||
|
||||
| # | Source | URL | Key claim | Type | Confidence |
|
||||
|---|--------|-----|-----------|------|------------|
|
||||
| 1 | ... | ... | ... | primary / secondary / self-reported | high / medium / low |
|
||||
|
||||
### Findings
|
||||
|
||||
Write findings using inline source references: `[1]`, `[2]`, etc. Every factual claim must cite at least one source by number.
|
||||
|
||||
### Sources
|
||||
|
||||
Numbered list matching the evidence table:
|
||||
1. Author/Title — URL
|
||||
2. Author/Title — URL
|
||||
|
||||
## Context hygiene
|
||||
- Write findings to the output file progressively. Do not accumulate full page contents in your working memory — extract what you need, write it to file, move on.
|
||||
- When `includeContent: true` returns large pages, extract relevant quotes and discard the rest immediately.
|
||||
- If your search produces 10+ results, triage by title/snippet first. Only fetch full content for the top candidates.
|
||||
- Return a one-line summary to the parent, not full findings. The parent reads the output file.
|
||||
|
||||
## Output contract
|
||||
- Save to the output file (default: `research.md`).
|
||||
- Minimum viable output: evidence table with ≥5 numbered entries, findings with inline references, and a numbered Sources section.
|
||||
- Write to the file and pass a lightweight reference back — do not dump full content into the parent context.
|
||||
93
website/src/content/docs/agents/reviewer.md
Normal file
93
website/src/content/docs/agents/reviewer.md
Normal file
@@ -0,0 +1,93 @@
|
||||
---
|
||||
title: Reviewer
|
||||
description: Simulate a tough but constructive AI research peer reviewer with inline annotations.
|
||||
section: Agents
|
||||
order: 2
|
||||
---
|
||||
|
||||
## Source
|
||||
|
||||
Generated from `.feynman/agents/reviewer.md`. Edit that prompt file, not this docs page.
|
||||
|
||||
## Role
|
||||
|
||||
Simulate a tough but constructive AI research peer reviewer with inline annotations.
|
||||
|
||||
## Default Output
|
||||
|
||||
`review.md`
|
||||
|
||||
Your job is to act like a skeptical but fair peer reviewer for AI/ML systems work.
|
||||
|
||||
## Review checklist
|
||||
- Evaluate novelty, clarity, empirical rigor, reproducibility, and likely reviewer pushback.
|
||||
- Do not praise vaguely. Every positive claim should be tied to specific evidence.
|
||||
- Look for:
|
||||
- missing or weak baselines
|
||||
- missing ablations
|
||||
- evaluation mismatches
|
||||
- unclear claims of novelty
|
||||
- weak related-work positioning
|
||||
- insufficient statistical evidence
|
||||
- benchmark leakage or contamination risks
|
||||
- under-specified implementation details
|
||||
- claims that outrun the experiments
|
||||
- Distinguish between fatal issues, strong concerns, and polish issues.
|
||||
- Preserve uncertainty. If the draft might pass depending on venue norms, say so explicitly.
|
||||
|
||||
## Output format
|
||||
|
||||
Produce two sections: a structured review and inline annotations.
|
||||
|
||||
### Part 1: Structured Review
|
||||
|
||||
```markdown
|
||||
## Summary
|
||||
1-2 paragraph summary of the paper's contributions and approach.
|
||||
|
||||
## Strengths
|
||||
- [S1] ...
|
||||
- [S2] ...
|
||||
|
||||
## Weaknesses
|
||||
- [W1] **FATAL:** ...
|
||||
- [W2] **MAJOR:** ...
|
||||
- [W3] **MINOR:** ...
|
||||
|
||||
## Questions for Authors
|
||||
- [Q1] ...
|
||||
|
||||
## Verdict
|
||||
Overall assessment and confidence score. Would this pass at [venue]?
|
||||
|
||||
## Revision Plan
|
||||
Prioritized, concrete steps to address each weakness.
|
||||
```
|
||||
|
||||
### Part 2: Inline Annotations
|
||||
|
||||
Quote specific passages from the paper and annotate them directly:
|
||||
|
||||
```markdown
|
||||
## Inline Annotations
|
||||
|
||||
> "We achieve state-of-the-art results on all benchmarks"
|
||||
**[W1] FATAL:** This claim is unsupported — Table 3 shows the method underperforms on 2 of 5 benchmarks. Revise to accurately reflect results.
|
||||
|
||||
> "Our approach is novel in combining X with Y"
|
||||
**[W3] MINOR:** Z et al. (2024) combined X with Y in a different domain. Acknowledge this and clarify the distinction.
|
||||
|
||||
> "We use a learning rate of 1e-4"
|
||||
**[Q1]:** Was this tuned? What range was searched? This matters for reproducibility.
|
||||
```
|
||||
|
||||
Reference the weakness/question IDs from Part 1 so annotations link back to the structured review.
|
||||
|
||||
## Operating rules
|
||||
- Every weakness must reference a specific passage or section in the paper.
|
||||
- Inline annotations must quote the exact text being critiqued.
|
||||
- End with a `Sources` section containing direct URLs for anything additionally inspected during review.
|
||||
|
||||
## Output contract
|
||||
- Save the main artifact to `review.md`.
|
||||
- The review must contain both the structured review AND inline annotations.
|
||||
50
website/src/content/docs/agents/verifier.md
Normal file
50
website/src/content/docs/agents/verifier.md
Normal file
@@ -0,0 +1,50 @@
|
||||
---
|
||||
title: Verifier
|
||||
description: Post-process a draft to add inline citations and verify every source URL.
|
||||
section: Agents
|
||||
order: 4
|
||||
---
|
||||
|
||||
## Source
|
||||
|
||||
Generated from `.feynman/agents/verifier.md`. Edit that prompt file, not this docs page.
|
||||
|
||||
## Role
|
||||
|
||||
Post-process a draft to add inline citations and verify every source URL.
|
||||
|
||||
## Tools
|
||||
|
||||
`read`, `bash`, `grep`, `find`, `ls`, `write`, `edit`
|
||||
|
||||
## Default Output
|
||||
|
||||
`cited.md`
|
||||
|
||||
You receive a draft document and the research files it was built from. Your job is to:
|
||||
|
||||
1. **Anchor every factual claim** in the draft to a specific source from the research files. Insert inline citations `[1]`, `[2]`, etc. directly after each claim.
|
||||
2. **Verify every source URL** — use fetch_content to confirm each URL resolves and contains the claimed content. Flag dead links.
|
||||
3. **Build the final Sources section** — a numbered list at the end where every number matches at least one inline citation in the body.
|
||||
4. **Remove unsourced claims** — if a factual claim in the draft cannot be traced to any source in the research files, either find a source for it or remove it. Do not leave unsourced factual claims.
|
||||
|
||||
## Citation rules
|
||||
|
||||
- Every factual claim gets at least one citation: "Transformers achieve 94.2% on MMLU [3]."
|
||||
- Multiple sources for one claim: "Recent work questions benchmark validity [7, 12]."
|
||||
- No orphan citations — every `[N]` in the body must appear in Sources.
|
||||
- No orphan sources — every entry in Sources must be cited at least once.
|
||||
- Hedged or opinion statements do not need citations.
|
||||
- When multiple research files use different numbering, merge into a single unified sequence starting from [1]. Deduplicate sources that appear in multiple files.
|
||||
|
||||
## Source verification
|
||||
|
||||
For each source URL:
|
||||
- **Live:** keep as-is.
|
||||
- **Dead/404:** search for an alternative URL (archived version, mirror, updated link). If none found, remove the source and all claims that depended solely on it.
|
||||
- **Redirects to unrelated content:** treat as dead.
|
||||
|
||||
## Output contract
|
||||
- Save to the output file (default: `cited.md`).
|
||||
- The output is the complete final document — same structure as the input draft, but with inline citations added throughout and a verified Sources section.
|
||||
- Do not change the substance or structure of the draft. Only add citations and fix dead sources.
|
||||
56
website/src/content/docs/agents/writer.md
Normal file
56
website/src/content/docs/agents/writer.md
Normal file
@@ -0,0 +1,56 @@
|
||||
---
|
||||
title: Writer
|
||||
description: Turn research notes into clear, structured briefs and drafts.
|
||||
section: Agents
|
||||
order: 3
|
||||
---
|
||||
|
||||
## Source
|
||||
|
||||
Generated from `.feynman/agents/writer.md`. Edit that prompt file, not this docs page.
|
||||
|
||||
## Role
|
||||
|
||||
Turn research notes into clear, structured briefs and drafts.
|
||||
|
||||
## Tools
|
||||
|
||||
`read`, `bash`, `grep`, `find`, `ls`, `write`, `edit`
|
||||
|
||||
## Default Output
|
||||
|
||||
`draft.md`
|
||||
|
||||
## Integrity commandments
|
||||
1. **Write only from supplied evidence.** Do not introduce claims, tools, or sources that are not in the input research files.
|
||||
2. **Preserve caveats and disagreements.** Never smooth away uncertainty.
|
||||
3. **Be explicit about gaps.** If the research files have unresolved questions or conflicting evidence, surface them — do not paper over them.
|
||||
|
||||
## Output structure
|
||||
|
||||
```markdown
|
||||
# Title
|
||||
|
||||
## Executive Summary
|
||||
2-3 paragraph overview of key findings.
|
||||
|
||||
## Section 1: ...
|
||||
Detailed findings organized by theme or question.
|
||||
|
||||
## Section N: ...
|
||||
...
|
||||
|
||||
## Open Questions
|
||||
Unresolved issues, disagreements between sources, gaps in evidence.
|
||||
```
|
||||
|
||||
## Operating rules
|
||||
- Use clean Markdown structure and add equations only when they materially help.
|
||||
- Keep the narrative readable, but never outrun the evidence.
|
||||
- Produce artifacts that are ready to review in a browser or PDF preview.
|
||||
- Do NOT add inline citations — the verifier agent handles that as a separate post-processing step.
|
||||
- Do NOT add a Sources section — the verifier agent builds that.
|
||||
|
||||
## Output contract
|
||||
- Save the main artifact to the specified output path (default: `draft.md`).
|
||||
- Focus on clarity, structure, and evidence traceability.
|
||||
66
website/src/content/docs/getting-started/configuration.md
Normal file
66
website/src/content/docs/getting-started/configuration.md
Normal file
@@ -0,0 +1,66 @@
|
||||
---
|
||||
title: Configuration
|
||||
description: Configure models, search, and runtime options
|
||||
section: Getting Started
|
||||
order: 4
|
||||
---
|
||||
|
||||
## Model
|
||||
|
||||
Set the default model:
|
||||
|
||||
```bash
|
||||
feynman model set <provider:model>
|
||||
```
|
||||
|
||||
Override at runtime:
|
||||
|
||||
```bash
|
||||
feynman --model anthropic:claude-opus-4-6
|
||||
```
|
||||
|
||||
List available models:
|
||||
|
||||
```bash
|
||||
feynman model list
|
||||
```
|
||||
|
||||
## Thinking level
|
||||
|
||||
Control the reasoning depth:
|
||||
|
||||
```bash
|
||||
feynman --thinking high
|
||||
```
|
||||
|
||||
Levels: `off`, `minimal`, `low`, `medium`, `high`, `xhigh`.
|
||||
|
||||
## Web search
|
||||
|
||||
Check the current search configuration:
|
||||
|
||||
```bash
|
||||
feynman search status
|
||||
```
|
||||
|
||||
For advanced configuration, edit `~/.feynman/web-search.json` directly to set Gemini API keys, Perplexity keys, or a different route.
|
||||
|
||||
## Working directory
|
||||
|
||||
```bash
|
||||
feynman --cwd /path/to/project
|
||||
```
|
||||
|
||||
## Session storage
|
||||
|
||||
```bash
|
||||
feynman --session-dir /path/to/sessions
|
||||
```
|
||||
|
||||
## One-shot mode
|
||||
|
||||
Run a single prompt and exit:
|
||||
|
||||
```bash
|
||||
feynman --prompt "summarize the key findings of 2401.12345"
|
||||
```
|
||||
34
website/src/content/docs/getting-started/installation.md
Normal file
34
website/src/content/docs/getting-started/installation.md
Normal file
@@ -0,0 +1,34 @@
|
||||
---
|
||||
title: Installation
|
||||
description: Install Feynman and get started
|
||||
section: Getting Started
|
||||
order: 1
|
||||
---
|
||||
|
||||
## Requirements
|
||||
|
||||
- Node.js 20 or later
|
||||
- npm 9 or later
|
||||
|
||||
## Install
|
||||
|
||||
```bash
|
||||
npm install -g @companion-ai/feynman
|
||||
```
|
||||
|
||||
## Verify
|
||||
|
||||
```bash
|
||||
feynman --version
|
||||
```
|
||||
|
||||
## Local Development
|
||||
|
||||
For contributing or local development:
|
||||
|
||||
```bash
|
||||
git clone https://github.com/getcompanion-ai/feynman.git
|
||||
cd feynman
|
||||
npm install
|
||||
npm run start
|
||||
```
|
||||
44
website/src/content/docs/getting-started/quickstart.md
Normal file
44
website/src/content/docs/getting-started/quickstart.md
Normal file
@@ -0,0 +1,44 @@
|
||||
---
|
||||
title: Quick Start
|
||||
description: Get up and running with Feynman in 60 seconds
|
||||
section: Getting Started
|
||||
order: 2
|
||||
---
|
||||
|
||||
## First run
|
||||
|
||||
```bash
|
||||
feynman setup
|
||||
feynman
|
||||
```
|
||||
|
||||
`feynman setup` walks you through model authentication, alphaXiv login, web search configuration, and preview dependencies.
|
||||
|
||||
## Ask naturally
|
||||
|
||||
Feynman routes your questions into the right workflow automatically. You don't need slash commands to get started.
|
||||
|
||||
```
|
||||
> What are the main approaches to RLHF alignment?
|
||||
```
|
||||
|
||||
Feynman will search papers, gather web sources, and produce a structured answer with citations.
|
||||
|
||||
## Use workflows directly
|
||||
|
||||
For explicit control, use slash commands inside the REPL:
|
||||
|
||||
```
|
||||
> /deepresearch transformer scaling laws
|
||||
> /lit multimodal reasoning benchmarks
|
||||
> /review paper.pdf
|
||||
```
|
||||
|
||||
## Output locations
|
||||
|
||||
Feynman writes durable artifacts to canonical directories:
|
||||
|
||||
- `outputs/` — Reviews, reading lists, summaries
|
||||
- `papers/` — Polished paper-style drafts
|
||||
- `experiments/` — Runnable code and result logs
|
||||
- `notes/` — Scratch notes and session logs
|
||||
66
website/src/content/docs/getting-started/setup.md
Normal file
66
website/src/content/docs/getting-started/setup.md
Normal file
@@ -0,0 +1,66 @@
|
||||
---
|
||||
title: Setup
|
||||
description: Detailed setup guide for Feynman
|
||||
section: Getting Started
|
||||
order: 3
|
||||
---
|
||||
|
||||
## Guided setup
|
||||
|
||||
```bash
|
||||
feynman setup
|
||||
```
|
||||
|
||||
This walks through four steps:
|
||||
|
||||
### Model provider authentication
|
||||
|
||||
Feynman uses Pi's OAuth system for model access. The setup wizard prompts you to log in to your preferred provider.
|
||||
|
||||
```bash
|
||||
feynman model login
|
||||
```
|
||||
|
||||
### AlphaXiv login
|
||||
|
||||
AlphaXiv powers Feynman's paper search and analysis tools. Sign in with:
|
||||
|
||||
```bash
|
||||
feynman alpha login
|
||||
```
|
||||
|
||||
Check status anytime:
|
||||
|
||||
```bash
|
||||
feynman alpha status
|
||||
```
|
||||
|
||||
### Web search routing
|
||||
|
||||
Feynman supports three web search backends:
|
||||
|
||||
- **auto** — Prefer Perplexity when configured, fall back to Gemini
|
||||
- **perplexity** — Force Perplexity Sonar
|
||||
- **gemini** — Force Gemini (default, zero-config via signed-in Chromium)
|
||||
|
||||
The default path requires no API keys — it uses Gemini Browser via your signed-in Chromium profile.
|
||||
|
||||
### Preview dependencies
|
||||
|
||||
For PDF and HTML export of generated artifacts, Feynman needs `pandoc`:
|
||||
|
||||
```bash
|
||||
feynman --setup-preview
|
||||
```
|
||||
|
||||
This installs pandoc automatically on macOS/Homebrew systems.
|
||||
|
||||
## Diagnostics
|
||||
|
||||
Run the doctor to check everything:
|
||||
|
||||
```bash
|
||||
feynman doctor
|
||||
```
|
||||
|
||||
This verifies model auth, alphaXiv credentials, preview dependencies, and the Pi runtime.
|
||||
61
website/src/content/docs/reference/cli-commands.md
Normal file
61
website/src/content/docs/reference/cli-commands.md
Normal file
@@ -0,0 +1,61 @@
|
||||
---
|
||||
title: CLI Commands
|
||||
description: Complete reference for Feynman CLI commands
|
||||
section: Reference
|
||||
order: 1
|
||||
---
|
||||
|
||||
This page covers the dedicated Feynman CLI commands and compatibility flags.
|
||||
|
||||
Workflow prompt templates such as `/deepresearch` also run directly from the shell as `feynman <workflow> ...`. Those workflow entries live in the slash-command reference instead of being duplicated here.
|
||||
|
||||
## Core
|
||||
|
||||
| Command | Description |
|
||||
| --- | --- |
|
||||
| `feynman` | Launch the interactive REPL. |
|
||||
| `feynman chat [prompt]` | Start chat explicitly, optionally with an initial prompt. |
|
||||
| `feynman help` | Show CLI help. |
|
||||
| `feynman setup` | Run the guided setup wizard. |
|
||||
| `feynman doctor` | Diagnose config, auth, Pi runtime, and preview dependencies. |
|
||||
| `feynman status` | Show the current setup summary. |
|
||||
|
||||
## Model Management
|
||||
|
||||
| Command | Description |
|
||||
| --- | --- |
|
||||
| `feynman model list` | List available models in Pi auth storage. |
|
||||
| `feynman model login [id]` | Login to a Pi OAuth model provider. |
|
||||
| `feynman model logout [id]` | Logout from a Pi OAuth model provider. |
|
||||
| `feynman model set <provider/model>` | Set the default model. |
|
||||
|
||||
## AlphaXiv
|
||||
|
||||
| Command | Description |
|
||||
| --- | --- |
|
||||
| `feynman alpha login` | Sign in to alphaXiv. |
|
||||
| `feynman alpha logout` | Clear alphaXiv auth. |
|
||||
| `feynman alpha status` | Check alphaXiv auth status. |
|
||||
|
||||
## Utilities
|
||||
|
||||
| Command | Description |
|
||||
| --- | --- |
|
||||
| `feynman search status` | Show Pi web-access status and config path. |
|
||||
| `feynman update [package]` | Update installed packages, or a specific package. |
|
||||
|
||||
## Flags
|
||||
|
||||
| Flag | Description |
|
||||
| --- | --- |
|
||||
| `--prompt "<text>"` | Run one prompt and exit. |
|
||||
| `--alpha-login` | Sign in to alphaXiv and exit. |
|
||||
| `--alpha-logout` | Clear alphaXiv auth and exit. |
|
||||
| `--alpha-status` | Show alphaXiv auth status and exit. |
|
||||
| `--model <provider:model>` | Force a specific model. |
|
||||
| `--thinking <level>` | Set thinking level: off | minimal | low | medium | high | xhigh. |
|
||||
| `--cwd <path>` | Set the working directory for tools. |
|
||||
| `--session-dir <path>` | Set the session storage directory. |
|
||||
| `--new-session` | Start a new persisted session. |
|
||||
| `--doctor` | Alias for `feynman doctor`. |
|
||||
| `--setup-preview` | Alias for `feynman setup preview`. |
|
||||
25
website/src/content/docs/reference/package-stack.md
Normal file
25
website/src/content/docs/reference/package-stack.md
Normal file
@@ -0,0 +1,25 @@
|
||||
---
|
||||
title: Package Stack
|
||||
description: Curated Pi packages bundled with Feynman
|
||||
section: Reference
|
||||
order: 3
|
||||
---
|
||||
|
||||
Curated Pi packages bundled with Feynman. The runtime package list lives in `.feynman/settings.json`.
|
||||
|
||||
| Package | Purpose |
|
||||
|---------|---------|
|
||||
| `pi-subagents` | Parallel literature gathering and decomposition. |
|
||||
| `pi-btw` | Fast side-thread `/btw` conversations without interrupting the main run. |
|
||||
| `pi-docparser` | PDFs, Office docs, spreadsheets, and images. |
|
||||
| `pi-web-access` | Web, GitHub, PDF, and media access. |
|
||||
| `pi-markdown-preview` | Polished Markdown and LaTeX-heavy research writeups. |
|
||||
| `@walterra/pi-charts` | Charts and quantitative visualizations. |
|
||||
| `pi-generative-ui` | Interactive HTML-style widgets. |
|
||||
| `pi-mermaid` | Diagrams in the TUI. |
|
||||
| `@aliou/pi-processes` | Long-running experiments and log tails. |
|
||||
| `pi-zotero` | Citation-library workflows. |
|
||||
| `@kaiserlich-dev/pi-session-search` | Indexed session recall and summarize/resume UI. |
|
||||
| `pi-schedule-prompt` | Recurring and deferred research jobs. |
|
||||
| `@samfp/pi-memory` | Automatic preference and correction memory across sessions. |
|
||||
| `@tmustier/pi-ralph-wiggum` | Long-running agent loops for iterative development. |
|
||||
41
website/src/content/docs/reference/slash-commands.md
Normal file
41
website/src/content/docs/reference/slash-commands.md
Normal file
@@ -0,0 +1,41 @@
|
||||
---
|
||||
title: Slash Commands
|
||||
description: Repo-owned REPL slash commands
|
||||
section: Reference
|
||||
order: 2
|
||||
---
|
||||
|
||||
This page documents the slash commands that Feynman owns in this repository: prompt templates from `prompts/` and extension commands from `extensions/research-tools/`.
|
||||
|
||||
Additional slash commands can appear at runtime from Pi core and bundled packages such as subagents, preview, session search, and scheduling. Use `/help` inside the REPL for the live command list instead of relying on a static copy of package-provided commands.
|
||||
|
||||
## Research Workflows
|
||||
|
||||
| Command | Description |
|
||||
| --- | --- |
|
||||
| `/deepresearch <topic>` | Run a thorough, source-heavy investigation on a topic and produce a durable research brief with inline citations. |
|
||||
| `/lit <topic>` | Run a literature review on a topic using paper search and primary-source synthesis. |
|
||||
| `/review <artifact>` | Simulate an AI research peer review with likely objections, severity, and a concrete revision plan. |
|
||||
| `/audit <item>` | Compare a paper's claims against its public codebase and identify mismatches, omissions, and reproducibility risks. |
|
||||
| `/replicate <paper>` | Plan or execute a replication workflow for a paper, claim, or benchmark. |
|
||||
| `/compare <topic>` | Compare multiple sources on a topic and produce a source-grounded matrix of agreements, disagreements, and confidence. |
|
||||
| `/draft <topic>` | Turn research findings into a polished paper-style draft with equations, sections, and explicit claims. |
|
||||
| `/autoresearch <idea>` | Autonomous experiment loop — try ideas, measure results, keep what works, discard what doesn't, repeat. |
|
||||
| `/watch <topic>` | Set up a recurring or deferred research watch on a topic, company, paper area, or product surface. |
|
||||
|
||||
## Project & Session
|
||||
|
||||
| Command | Description |
|
||||
| --- | --- |
|
||||
| `/log` | Write a durable session log with completed work, findings, open questions, and next steps. |
|
||||
| `/jobs` | Inspect active background research work, including running processes and scheduled follow-ups. |
|
||||
| `/help` | Show grouped Feynman commands and prefill the editor with a selected command. |
|
||||
| `/init` | Bootstrap AGENTS.md and session-log folders for a research project. |
|
||||
|
||||
## Setup
|
||||
|
||||
| Command | Description |
|
||||
| --- | --- |
|
||||
| `/alpha-login` | Sign in to alphaXiv from inside Feynman. |
|
||||
| `/alpha-status` | Show alphaXiv authentication status. |
|
||||
| `/alpha-logout` | Clear alphaXiv auth from inside Feynman. |
|
||||
40
website/src/content/docs/tools/alphaxiv.md
Normal file
40
website/src/content/docs/tools/alphaxiv.md
Normal file
@@ -0,0 +1,40 @@
|
||||
---
|
||||
title: AlphaXiv
|
||||
description: Paper search and analysis tools
|
||||
section: Tools
|
||||
order: 1
|
||||
---
|
||||
|
||||
## Overview
|
||||
|
||||
AlphaXiv powers Feynman's academic paper workflows. All tools require an alphaXiv account — sign in with `feynman alpha login`.
|
||||
|
||||
## Tools
|
||||
|
||||
### alpha_search
|
||||
|
||||
Paper discovery with three search modes:
|
||||
|
||||
- **semantic** — Meaning-based search across paper content
|
||||
- **keyword** — Traditional keyword matching
|
||||
- **agentic** — AI-powered search that interprets your intent
|
||||
|
||||
### alpha_get_paper
|
||||
|
||||
Fetch a paper's report (structured summary) or full raw text by arXiv ID.
|
||||
|
||||
### alpha_ask_paper
|
||||
|
||||
Ask a targeted question about a specific paper. Returns an answer grounded in the paper's content.
|
||||
|
||||
### alpha_annotate_paper
|
||||
|
||||
Add persistent local notes to a paper. Annotations are stored locally and persist across sessions.
|
||||
|
||||
### alpha_list_annotations
|
||||
|
||||
Recall all annotations across papers and sessions.
|
||||
|
||||
### alpha_read_code
|
||||
|
||||
Read source code from a paper's linked GitHub repository. Useful for auditing or replication planning.
|
||||
34
website/src/content/docs/tools/preview.md
Normal file
34
website/src/content/docs/tools/preview.md
Normal file
@@ -0,0 +1,34 @@
|
||||
---
|
||||
title: Preview
|
||||
description: Preview generated artifacts in browser or PDF
|
||||
section: Tools
|
||||
order: 4
|
||||
---
|
||||
|
||||
## Overview
|
||||
|
||||
The `preview_file` tool opens generated artifacts in your browser or PDF viewer.
|
||||
|
||||
## Usage
|
||||
|
||||
Inside the REPL:
|
||||
|
||||
```
|
||||
/preview
|
||||
```
|
||||
|
||||
Or Feynman will suggest previewing when you generate artifacts that benefit from rendered output (Markdown with LaTeX, HTML reports, etc.).
|
||||
|
||||
## Requirements
|
||||
|
||||
Preview requires `pandoc` for PDF/HTML rendering. Install it with:
|
||||
|
||||
```bash
|
||||
feynman --setup-preview
|
||||
```
|
||||
|
||||
## Supported formats
|
||||
|
||||
- Markdown (with LaTeX math rendering)
|
||||
- HTML
|
||||
- PDF
|
||||
26
website/src/content/docs/tools/session-search.md
Normal file
26
website/src/content/docs/tools/session-search.md
Normal file
@@ -0,0 +1,26 @@
|
||||
---
|
||||
title: Session Search
|
||||
description: Search prior Feynman session transcripts
|
||||
section: Tools
|
||||
order: 3
|
||||
---
|
||||
|
||||
## Overview
|
||||
|
||||
The `session_search` tool recovers prior Feynman work from stored session transcripts. Useful for picking up previous research threads or finding past findings.
|
||||
|
||||
## Usage
|
||||
|
||||
Inside the REPL:
|
||||
|
||||
```
|
||||
/search
|
||||
```
|
||||
|
||||
Or use the tool directly — Feynman will invoke `session_search` automatically when you reference prior work.
|
||||
|
||||
## What it searches
|
||||
|
||||
- Full session transcripts
|
||||
- Tool outputs and agent results
|
||||
- Generated artifacts and their content
|
||||
34
website/src/content/docs/tools/web-search.md
Normal file
34
website/src/content/docs/tools/web-search.md
Normal file
@@ -0,0 +1,34 @@
|
||||
---
|
||||
title: Web Search
|
||||
description: Web search routing and configuration
|
||||
section: Tools
|
||||
order: 2
|
||||
---
|
||||
|
||||
## Routing modes
|
||||
|
||||
Feynman supports three web search backends:
|
||||
|
||||
| Mode | Description |
|
||||
|------|-------------|
|
||||
| `auto` | Prefer Perplexity when configured, fall back to Gemini |
|
||||
| `perplexity` | Force Perplexity Sonar |
|
||||
| `gemini` | Force Gemini (default) |
|
||||
|
||||
## Default behavior
|
||||
|
||||
The default path is zero-config Gemini Browser via a signed-in Chromium profile. No API keys required.
|
||||
|
||||
## Check current config
|
||||
|
||||
```bash
|
||||
feynman search status
|
||||
```
|
||||
|
||||
## Advanced configuration
|
||||
|
||||
Edit `~/.feynman/web-search.json` directly to set:
|
||||
|
||||
- Gemini API keys
|
||||
- Perplexity API keys
|
||||
- Custom routing preferences
|
||||
39
website/src/content/docs/workflows/audit.md
Normal file
39
website/src/content/docs/workflows/audit.md
Normal file
@@ -0,0 +1,39 @@
|
||||
---
|
||||
title: Code Audit
|
||||
description: Compare paper claims against public codebases
|
||||
section: Workflows
|
||||
order: 4
|
||||
---
|
||||
|
||||
## Usage
|
||||
|
||||
```
|
||||
/audit <item>
|
||||
```
|
||||
|
||||
## What it does
|
||||
|
||||
Compares claims made in a paper against its public codebase. Surfaces mismatches, missing experiments, and reproducibility risks.
|
||||
|
||||
## What it checks
|
||||
|
||||
- Do the reported hyperparameters match the code?
|
||||
- Are all claimed experiments present in the repository?
|
||||
- Does the training loop match the described methodology?
|
||||
- Are there undocumented preprocessing steps?
|
||||
- Do evaluation metrics match the paper's claims?
|
||||
|
||||
## Example
|
||||
|
||||
```
|
||||
/audit 2401.12345
|
||||
```
|
||||
|
||||
## Output
|
||||
|
||||
An audit report with:
|
||||
|
||||
- Claim-by-claim verification
|
||||
- Identified mismatches
|
||||
- Missing components
|
||||
- Reproducibility risk assessment
|
||||
44
website/src/content/docs/workflows/autoresearch.md
Normal file
44
website/src/content/docs/workflows/autoresearch.md
Normal file
@@ -0,0 +1,44 @@
|
||||
---
|
||||
title: Autoresearch
|
||||
description: Autonomous experiment optimization loop
|
||||
section: Workflows
|
||||
order: 8
|
||||
---
|
||||
|
||||
## Usage
|
||||
|
||||
```
|
||||
/autoresearch <idea>
|
||||
```
|
||||
|
||||
## What it does
|
||||
|
||||
Runs an autonomous experiment loop:
|
||||
|
||||
1. **Edit** — Modify code or configuration
|
||||
2. **Commit** — Save the change
|
||||
3. **Benchmark** — Run evaluation
|
||||
4. **Evaluate** — Compare against baseline
|
||||
5. **Keep or revert** — Persist improvements, roll back regressions
|
||||
6. **Repeat** — Continue until the target is hit
|
||||
|
||||
## Tracking
|
||||
|
||||
Metrics are tracked in:
|
||||
|
||||
- `autoresearch.md` — Human-readable progress log
|
||||
- `autoresearch.jsonl` — Machine-readable metrics over time
|
||||
|
||||
## Controls
|
||||
|
||||
```
|
||||
/autoresearch <idea> # start or resume
|
||||
/autoresearch off # stop, keep data
|
||||
/autoresearch clear # delete all state, start fresh
|
||||
```
|
||||
|
||||
## Example
|
||||
|
||||
```
|
||||
/autoresearch optimize the learning rate schedule for better convergence
|
||||
```
|
||||
29
website/src/content/docs/workflows/compare.md
Normal file
29
website/src/content/docs/workflows/compare.md
Normal file
@@ -0,0 +1,29 @@
|
||||
---
|
||||
title: Source Comparison
|
||||
description: Compare multiple sources with agreement/disagreement matrix
|
||||
section: Workflows
|
||||
order: 6
|
||||
---
|
||||
|
||||
## Usage
|
||||
|
||||
```
|
||||
/compare <topic>
|
||||
```
|
||||
|
||||
## What it does
|
||||
|
||||
Compares multiple sources on a topic. Builds an agreement/disagreement matrix showing where sources align and where they conflict.
|
||||
|
||||
## Example
|
||||
|
||||
```
|
||||
/compare approaches to constitutional AI training
|
||||
```
|
||||
|
||||
## Output
|
||||
|
||||
- Source-by-source breakdown
|
||||
- Agreement/disagreement matrix
|
||||
- Synthesis of key differences
|
||||
- Assessment of which positions have stronger evidence
|
||||
40
website/src/content/docs/workflows/deep-research.md
Normal file
40
website/src/content/docs/workflows/deep-research.md
Normal file
@@ -0,0 +1,40 @@
|
||||
---
|
||||
title: Deep Research
|
||||
description: Thorough source-heavy investigation with parallel agents
|
||||
section: Workflows
|
||||
order: 1
|
||||
---
|
||||
|
||||
## Usage
|
||||
|
||||
```
|
||||
/deepresearch <topic>
|
||||
```
|
||||
|
||||
## What it does
|
||||
|
||||
Deep research runs a thorough, source-heavy investigation. It plans the research scope, delegates to parallel researcher agents, synthesizes findings, and adds inline citations.
|
||||
|
||||
The workflow follows these steps:
|
||||
|
||||
1. **Plan** — Clarify the research question and identify search strategy
|
||||
2. **Delegate** — Spawn parallel researcher agents to gather evidence from different source types (papers, web, repos)
|
||||
3. **Synthesize** — Merge findings, resolve contradictions, identify gaps
|
||||
4. **Cite** — Add inline citations and verify all source URLs
|
||||
5. **Deliver** — Write a durable research brief to `outputs/`
|
||||
|
||||
## Example
|
||||
|
||||
```
|
||||
/deepresearch transformer scaling laws and their implications for compute-optimal training
|
||||
```
|
||||
|
||||
## Output
|
||||
|
||||
Produces a structured research brief with:
|
||||
|
||||
- Executive summary
|
||||
- Key findings organized by theme
|
||||
- Evidence tables with source links
|
||||
- Open questions and suggested next steps
|
||||
- Numbered sources section with direct URLs
|
||||
37
website/src/content/docs/workflows/draft.md
Normal file
37
website/src/content/docs/workflows/draft.md
Normal file
@@ -0,0 +1,37 @@
|
||||
---
|
||||
title: Draft Writing
|
||||
description: Paper-style draft generation from research findings
|
||||
section: Workflows
|
||||
order: 7
|
||||
---
|
||||
|
||||
## Usage
|
||||
|
||||
```
|
||||
/draft <topic>
|
||||
```
|
||||
|
||||
## What it does
|
||||
|
||||
Produces a paper-style draft with structured sections. Writes to `papers/`.
|
||||
|
||||
## Structure
|
||||
|
||||
The generated draft includes:
|
||||
|
||||
- Title
|
||||
- Abstract
|
||||
- Introduction / Background
|
||||
- Method or Approach
|
||||
- Evidence and Analysis
|
||||
- Limitations
|
||||
- Conclusion
|
||||
- Sources
|
||||
|
||||
## Example
|
||||
|
||||
```
|
||||
/draft survey of differentiable physics simulators
|
||||
```
|
||||
|
||||
The writer agent works only from supplied evidence — it never fabricates content. If evidence is insufficient, it explicitly notes the gaps.
|
||||
31
website/src/content/docs/workflows/literature-review.md
Normal file
31
website/src/content/docs/workflows/literature-review.md
Normal file
@@ -0,0 +1,31 @@
|
||||
---
|
||||
title: Literature Review
|
||||
description: Map consensus, disagreements, and open questions
|
||||
section: Workflows
|
||||
order: 2
|
||||
---
|
||||
|
||||
## Usage
|
||||
|
||||
```
|
||||
/lit <topic>
|
||||
```
|
||||
|
||||
## What it does
|
||||
|
||||
Runs a structured literature review that searches across academic papers and web sources. Explicitly separates consensus findings from disagreements and open questions.
|
||||
|
||||
## Example
|
||||
|
||||
```
|
||||
/lit multimodal reasoning benchmarks for large language models
|
||||
```
|
||||
|
||||
## Output
|
||||
|
||||
A structured review covering:
|
||||
|
||||
- **Consensus** — What the field agrees on
|
||||
- **Disagreements** — Where sources conflict
|
||||
- **Open questions** — What remains unresolved
|
||||
- **Sources** — Direct links to all referenced papers and articles
|
||||
42
website/src/content/docs/workflows/replication.md
Normal file
42
website/src/content/docs/workflows/replication.md
Normal file
@@ -0,0 +1,42 @@
|
||||
---
|
||||
title: Replication
|
||||
description: Plan replications of papers and claims
|
||||
section: Workflows
|
||||
order: 5
|
||||
---
|
||||
|
||||
## Usage
|
||||
|
||||
```
|
||||
/replicate <paper or claim>
|
||||
```
|
||||
|
||||
## What it does
|
||||
|
||||
Extracts key implementation details from a paper, identifies what's needed to replicate the results, and asks where to run before executing anything.
|
||||
|
||||
Before running code, Feynman asks you to choose an execution environment:
|
||||
|
||||
- **Local** — run in the current working directory
|
||||
- **Virtual environment** — create an isolated venv/conda env first
|
||||
- **Cloud** — delegate to a remote Agent Computer machine
|
||||
- **Plan only** — produce the replication plan without executing
|
||||
|
||||
## Example
|
||||
|
||||
```
|
||||
/replicate "chain-of-thought prompting improves math reasoning"
|
||||
```
|
||||
|
||||
## Output
|
||||
|
||||
A replication plan covering:
|
||||
|
||||
- Key claims to verify
|
||||
- Required resources (compute, data, models)
|
||||
- Implementation details extracted from the paper
|
||||
- Potential pitfalls and underspecified details
|
||||
- Step-by-step replication procedure
|
||||
- Success criteria
|
||||
|
||||
If an execution environment is selected, also produces runnable scripts and captured results.
|
||||
49
website/src/content/docs/workflows/review.md
Normal file
49
website/src/content/docs/workflows/review.md
Normal file
@@ -0,0 +1,49 @@
|
||||
---
|
||||
title: Peer Review
|
||||
description: Simulated peer review with severity-graded feedback
|
||||
section: Workflows
|
||||
order: 3
|
||||
---
|
||||
|
||||
## Usage
|
||||
|
||||
```
|
||||
/review <artifact>
|
||||
```
|
||||
|
||||
## What it does
|
||||
|
||||
Simulates a tough-but-fair peer review for AI research artifacts. Evaluates novelty, empirical rigor, baselines, ablations, and reproducibility.
|
||||
|
||||
The reviewer agent identifies:
|
||||
|
||||
- Weak baselines
|
||||
- Missing ablations
|
||||
- Evaluation mismatches
|
||||
- Benchmark leakage
|
||||
- Under-specified implementation details
|
||||
|
||||
## Severity levels
|
||||
|
||||
Feedback is graded by severity:
|
||||
|
||||
- **FATAL** — Fundamental issues that invalidate the claims
|
||||
- **MAJOR** — Significant problems that need addressing
|
||||
- **MINOR** — Small improvements or clarifications
|
||||
|
||||
## Example
|
||||
|
||||
```
|
||||
/review outputs/scaling-laws-brief.md
|
||||
```
|
||||
|
||||
## Output
|
||||
|
||||
Structured review with:
|
||||
|
||||
- Summary of the work
|
||||
- Strengths
|
||||
- Weaknesses (severity-graded)
|
||||
- Questions for the authors
|
||||
- Verdict (accept / revise / reject)
|
||||
- Revision plan
|
||||
29
website/src/content/docs/workflows/watch.md
Normal file
29
website/src/content/docs/workflows/watch.md
Normal file
@@ -0,0 +1,29 @@
|
||||
---
|
||||
title: Watch
|
||||
description: Recurring research monitoring
|
||||
section: Workflows
|
||||
order: 9
|
||||
---
|
||||
|
||||
## Usage
|
||||
|
||||
```
|
||||
/watch <topic>
|
||||
```
|
||||
|
||||
## What it does
|
||||
|
||||
Schedules a recurring research watch. Sets a baseline of current knowledge and defines what constitutes a meaningful change worth reporting.
|
||||
|
||||
## Example
|
||||
|
||||
```
|
||||
/watch new papers on test-time compute scaling
|
||||
```
|
||||
|
||||
## How it works
|
||||
|
||||
1. Feynman establishes a baseline by surveying current sources
|
||||
2. Defines change signals (new papers, updated results, new repos)
|
||||
3. Schedules periodic checks via `pi-schedule-prompt`
|
||||
4. Reports only when meaningful changes are detected
|
||||
Reference in New Issue
Block a user