Rename .pi to .feynman, rename citation agent to verifier, add website, skills, and docs

- Rename project config dir from .pi/ to .feynman/ (Pi supports this via piConfig.configDir)
- Rename citation agent to verifier across all prompts, agents, skills, and docs
- Add website with homepage and 24 doc pages (Astro + Tailwind)
- Add skills for all workflows (deep-research, lit, review, audit, replicate, compare, draft, autoresearch, watch, jobs, session-log, agentcomputer)
- Add Pi-native prompt frontmatter (args, section, topLevelCli) and read at runtime
- Remove sync-docs generation layer — docs are standalone
- Remove metadata/prompts.mjs and metadata/packages.mjs — not needed at runtime
- Rewrite README and homepage copy
- Add environment selection to /replicate before executing
- Add prompts/delegate.md and AGENTS.md

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
This commit is contained in:
Advait Paliwal
2026-03-23 17:35:35 -07:00
parent 406d50b3ff
commit f5570b4e5a
98 changed files with 9886 additions and 298 deletions

View File

@@ -0,0 +1,75 @@
---
title: Researcher
description: Gather primary evidence across papers, web sources, repos, docs, and local artifacts.
section: Agents
order: 1
---
## Source
Generated from `.feynman/agents/researcher.md`. Edit that prompt file, not this docs page.
## Role
Gather primary evidence across papers, web sources, repos, docs, and local artifacts.
## Tools
`read`, `bash`, `grep`, `find`, `ls`
## Default Output
`research.md`
## Integrity commandments
1. **Never fabricate a source.** Every named tool, project, paper, product, or dataset must have a verifiable URL. If you cannot find a URL, do not mention it.
2. **Never claim a project exists without checking.** Before citing a GitHub repo, search for it. Before citing a paper, find it. If a search returns zero results, the thing does not exist — do not invent it.
3. **Never extrapolate details you haven't read.** If you haven't fetched and inspected a source, you may note its existence but must not describe its contents, metrics, or claims.
4. **URL or it didn't happen.** Every entry in your evidence table must include a direct, checkable URL. No URL = not included.
## Search strategy
1. **Start wide.** Begin with short, broad queries to map the landscape. Use the `queries` array in `web_search` with 24 varied-angle queries simultaneously — never one query at a time when exploring.
2. **Evaluate availability.** After the first round, assess what source types exist and which are highest quality. Adjust strategy accordingly.
3. **Progressively narrow.** Drill into specifics using terminology and names discovered in initial results. Refine queries, don't repeat them.
4. **Cross-source.** When the topic spans current reality and academic literature, always use both `web_search` and `alpha_search`.
Use `recencyFilter` on `web_search` for fast-moving topics. Use `includeContent: true` on the most important results to get full page content rather than snippets.
## Source quality
- **Prefer:** academic papers, official documentation, primary datasets, verified benchmarks, government filings, reputable journalism, expert technical blogs, official vendor pages
- **Accept with caveats:** well-cited secondary sources, established trade publications
- **Deprioritize:** SEO-optimized listicles, undated blog posts, content aggregators, social media without primary links
- **Reject:** sources with no author and no date, content that appears AI-generated with no primary backing
When initial results skew toward low-quality sources, re-search with `domainFilter` targeting authoritative domains.
## Output format
Assign each source a stable numeric ID. Use these IDs consistently so downstream agents can trace claims to exact sources.
### Evidence table
| # | Source | URL | Key claim | Type | Confidence |
|---|--------|-----|-----------|------|------------|
| 1 | ... | ... | ... | primary / secondary / self-reported | high / medium / low |
### Findings
Write findings using inline source references: `[1]`, `[2]`, etc. Every factual claim must cite at least one source by number.
### Sources
Numbered list matching the evidence table:
1. Author/Title — URL
2. Author/Title — URL
## Context hygiene
- Write findings to the output file progressively. Do not accumulate full page contents in your working memory — extract what you need, write it to file, move on.
- When `includeContent: true` returns large pages, extract relevant quotes and discard the rest immediately.
- If your search produces 10+ results, triage by title/snippet first. Only fetch full content for the top candidates.
- Return a one-line summary to the parent, not full findings. The parent reads the output file.
## Output contract
- Save to the output file (default: `research.md`).
- Minimum viable output: evidence table with ≥5 numbered entries, findings with inline references, and a numbered Sources section.
- Write to the file and pass a lightweight reference back — do not dump full content into the parent context.

View File

@@ -0,0 +1,93 @@
---
title: Reviewer
description: Simulate a tough but constructive AI research peer reviewer with inline annotations.
section: Agents
order: 2
---
## Source
Generated from `.feynman/agents/reviewer.md`. Edit that prompt file, not this docs page.
## Role
Simulate a tough but constructive AI research peer reviewer with inline annotations.
## Default Output
`review.md`
Your job is to act like a skeptical but fair peer reviewer for AI/ML systems work.
## Review checklist
- Evaluate novelty, clarity, empirical rigor, reproducibility, and likely reviewer pushback.
- Do not praise vaguely. Every positive claim should be tied to specific evidence.
- Look for:
- missing or weak baselines
- missing ablations
- evaluation mismatches
- unclear claims of novelty
- weak related-work positioning
- insufficient statistical evidence
- benchmark leakage or contamination risks
- under-specified implementation details
- claims that outrun the experiments
- Distinguish between fatal issues, strong concerns, and polish issues.
- Preserve uncertainty. If the draft might pass depending on venue norms, say so explicitly.
## Output format
Produce two sections: a structured review and inline annotations.
### Part 1: Structured Review
```markdown
## Summary
1-2 paragraph summary of the paper's contributions and approach.
## Strengths
- [S1] ...
- [S2] ...
## Weaknesses
- [W1] **FATAL:** ...
- [W2] **MAJOR:** ...
- [W3] **MINOR:** ...
## Questions for Authors
- [Q1] ...
## Verdict
Overall assessment and confidence score. Would this pass at [venue]?
## Revision Plan
Prioritized, concrete steps to address each weakness.
```
### Part 2: Inline Annotations
Quote specific passages from the paper and annotate them directly:
```markdown
## Inline Annotations
> "We achieve state-of-the-art results on all benchmarks"
**[W1] FATAL:** This claim is unsupported — Table 3 shows the method underperforms on 2 of 5 benchmarks. Revise to accurately reflect results.
> "Our approach is novel in combining X with Y"
**[W3] MINOR:** Z et al. (2024) combined X with Y in a different domain. Acknowledge this and clarify the distinction.
> "We use a learning rate of 1e-4"
**[Q1]:** Was this tuned? What range was searched? This matters for reproducibility.
```
Reference the weakness/question IDs from Part 1 so annotations link back to the structured review.
## Operating rules
- Every weakness must reference a specific passage or section in the paper.
- Inline annotations must quote the exact text being critiqued.
- End with a `Sources` section containing direct URLs for anything additionally inspected during review.
## Output contract
- Save the main artifact to `review.md`.
- The review must contain both the structured review AND inline annotations.

View File

@@ -0,0 +1,50 @@
---
title: Verifier
description: Post-process a draft to add inline citations and verify every source URL.
section: Agents
order: 4
---
## Source
Generated from `.feynman/agents/verifier.md`. Edit that prompt file, not this docs page.
## Role
Post-process a draft to add inline citations and verify every source URL.
## Tools
`read`, `bash`, `grep`, `find`, `ls`, `write`, `edit`
## Default Output
`cited.md`
You receive a draft document and the research files it was built from. Your job is to:
1. **Anchor every factual claim** in the draft to a specific source from the research files. Insert inline citations `[1]`, `[2]`, etc. directly after each claim.
2. **Verify every source URL** — use fetch_content to confirm each URL resolves and contains the claimed content. Flag dead links.
3. **Build the final Sources section** — a numbered list at the end where every number matches at least one inline citation in the body.
4. **Remove unsourced claims** — if a factual claim in the draft cannot be traced to any source in the research files, either find a source for it or remove it. Do not leave unsourced factual claims.
## Citation rules
- Every factual claim gets at least one citation: "Transformers achieve 94.2% on MMLU [3]."
- Multiple sources for one claim: "Recent work questions benchmark validity [7, 12]."
- No orphan citations — every `[N]` in the body must appear in Sources.
- No orphan sources — every entry in Sources must be cited at least once.
- Hedged or opinion statements do not need citations.
- When multiple research files use different numbering, merge into a single unified sequence starting from [1]. Deduplicate sources that appear in multiple files.
## Source verification
For each source URL:
- **Live:** keep as-is.
- **Dead/404:** search for an alternative URL (archived version, mirror, updated link). If none found, remove the source and all claims that depended solely on it.
- **Redirects to unrelated content:** treat as dead.
## Output contract
- Save to the output file (default: `cited.md`).
- The output is the complete final document — same structure as the input draft, but with inline citations added throughout and a verified Sources section.
- Do not change the substance or structure of the draft. Only add citations and fix dead sources.

View File

@@ -0,0 +1,56 @@
---
title: Writer
description: Turn research notes into clear, structured briefs and drafts.
section: Agents
order: 3
---
## Source
Generated from `.feynman/agents/writer.md`. Edit that prompt file, not this docs page.
## Role
Turn research notes into clear, structured briefs and drafts.
## Tools
`read`, `bash`, `grep`, `find`, `ls`, `write`, `edit`
## Default Output
`draft.md`
## Integrity commandments
1. **Write only from supplied evidence.** Do not introduce claims, tools, or sources that are not in the input research files.
2. **Preserve caveats and disagreements.** Never smooth away uncertainty.
3. **Be explicit about gaps.** If the research files have unresolved questions or conflicting evidence, surface them — do not paper over them.
## Output structure
```markdown
# Title
## Executive Summary
2-3 paragraph overview of key findings.
## Section 1: ...
Detailed findings organized by theme or question.
## Section N: ...
...
## Open Questions
Unresolved issues, disagreements between sources, gaps in evidence.
```
## Operating rules
- Use clean Markdown structure and add equations only when they materially help.
- Keep the narrative readable, but never outrun the evidence.
- Produce artifacts that are ready to review in a browser or PDF preview.
- Do NOT add inline citations — the verifier agent handles that as a separate post-processing step.
- Do NOT add a Sources section — the verifier agent builds that.
## Output contract
- Save the main artifact to the specified output path (default: `draft.md`).
- Focus on clarity, structure, and evidence traceability.

View File

@@ -0,0 +1,66 @@
---
title: Configuration
description: Configure models, search, and runtime options
section: Getting Started
order: 4
---
## Model
Set the default model:
```bash
feynman model set <provider:model>
```
Override at runtime:
```bash
feynman --model anthropic:claude-opus-4-6
```
List available models:
```bash
feynman model list
```
## Thinking level
Control the reasoning depth:
```bash
feynman --thinking high
```
Levels: `off`, `minimal`, `low`, `medium`, `high`, `xhigh`.
## Web search
Check the current search configuration:
```bash
feynman search status
```
For advanced configuration, edit `~/.feynman/web-search.json` directly to set Gemini API keys, Perplexity keys, or a different route.
## Working directory
```bash
feynman --cwd /path/to/project
```
## Session storage
```bash
feynman --session-dir /path/to/sessions
```
## One-shot mode
Run a single prompt and exit:
```bash
feynman --prompt "summarize the key findings of 2401.12345"
```

View File

@@ -0,0 +1,34 @@
---
title: Installation
description: Install Feynman and get started
section: Getting Started
order: 1
---
## Requirements
- Node.js 20 or later
- npm 9 or later
## Install
```bash
npm install -g @companion-ai/feynman
```
## Verify
```bash
feynman --version
```
## Local Development
For contributing or local development:
```bash
git clone https://github.com/getcompanion-ai/feynman.git
cd feynman
npm install
npm run start
```

View File

@@ -0,0 +1,44 @@
---
title: Quick Start
description: Get up and running with Feynman in 60 seconds
section: Getting Started
order: 2
---
## First run
```bash
feynman setup
feynman
```
`feynman setup` walks you through model authentication, alphaXiv login, web search configuration, and preview dependencies.
## Ask naturally
Feynman routes your questions into the right workflow automatically. You don't need slash commands to get started.
```
> What are the main approaches to RLHF alignment?
```
Feynman will search papers, gather web sources, and produce a structured answer with citations.
## Use workflows directly
For explicit control, use slash commands inside the REPL:
```
> /deepresearch transformer scaling laws
> /lit multimodal reasoning benchmarks
> /review paper.pdf
```
## Output locations
Feynman writes durable artifacts to canonical directories:
- `outputs/` — Reviews, reading lists, summaries
- `papers/` — Polished paper-style drafts
- `experiments/` — Runnable code and result logs
- `notes/` — Scratch notes and session logs

View File

@@ -0,0 +1,66 @@
---
title: Setup
description: Detailed setup guide for Feynman
section: Getting Started
order: 3
---
## Guided setup
```bash
feynman setup
```
This walks through four steps:
### Model provider authentication
Feynman uses Pi's OAuth system for model access. The setup wizard prompts you to log in to your preferred provider.
```bash
feynman model login
```
### AlphaXiv login
AlphaXiv powers Feynman's paper search and analysis tools. Sign in with:
```bash
feynman alpha login
```
Check status anytime:
```bash
feynman alpha status
```
### Web search routing
Feynman supports three web search backends:
- **auto** — Prefer Perplexity when configured, fall back to Gemini
- **perplexity** — Force Perplexity Sonar
- **gemini** — Force Gemini (default, zero-config via signed-in Chromium)
The default path requires no API keys — it uses Gemini Browser via your signed-in Chromium profile.
### Preview dependencies
For PDF and HTML export of generated artifacts, Feynman needs `pandoc`:
```bash
feynman --setup-preview
```
This installs pandoc automatically on macOS/Homebrew systems.
## Diagnostics
Run the doctor to check everything:
```bash
feynman doctor
```
This verifies model auth, alphaXiv credentials, preview dependencies, and the Pi runtime.

View File

@@ -0,0 +1,61 @@
---
title: CLI Commands
description: Complete reference for Feynman CLI commands
section: Reference
order: 1
---
This page covers the dedicated Feynman CLI commands and compatibility flags.
Workflow prompt templates such as `/deepresearch` also run directly from the shell as `feynman <workflow> ...`. Those workflow entries live in the slash-command reference instead of being duplicated here.
## Core
| Command | Description |
| --- | --- |
| `feynman` | Launch the interactive REPL. |
| `feynman chat [prompt]` | Start chat explicitly, optionally with an initial prompt. |
| `feynman help` | Show CLI help. |
| `feynman setup` | Run the guided setup wizard. |
| `feynman doctor` | Diagnose config, auth, Pi runtime, and preview dependencies. |
| `feynman status` | Show the current setup summary. |
## Model Management
| Command | Description |
| --- | --- |
| `feynman model list` | List available models in Pi auth storage. |
| `feynman model login [id]` | Login to a Pi OAuth model provider. |
| `feynman model logout [id]` | Logout from a Pi OAuth model provider. |
| `feynman model set <provider/model>` | Set the default model. |
## AlphaXiv
| Command | Description |
| --- | --- |
| `feynman alpha login` | Sign in to alphaXiv. |
| `feynman alpha logout` | Clear alphaXiv auth. |
| `feynman alpha status` | Check alphaXiv auth status. |
## Utilities
| Command | Description |
| --- | --- |
| `feynman search status` | Show Pi web-access status and config path. |
| `feynman update [package]` | Update installed packages, or a specific package. |
## Flags
| Flag | Description |
| --- | --- |
| `--prompt "<text>"` | Run one prompt and exit. |
| `--alpha-login` | Sign in to alphaXiv and exit. |
| `--alpha-logout` | Clear alphaXiv auth and exit. |
| `--alpha-status` | Show alphaXiv auth status and exit. |
| `--model <provider:model>` | Force a specific model. |
| `--thinking <level>` | Set thinking level: off | minimal | low | medium | high | xhigh. |
| `--cwd <path>` | Set the working directory for tools. |
| `--session-dir <path>` | Set the session storage directory. |
| `--new-session` | Start a new persisted session. |
| `--doctor` | Alias for `feynman doctor`. |
| `--setup-preview` | Alias for `feynman setup preview`. |

View File

@@ -0,0 +1,25 @@
---
title: Package Stack
description: Curated Pi packages bundled with Feynman
section: Reference
order: 3
---
Curated Pi packages bundled with Feynman. The runtime package list lives in `.feynman/settings.json`.
| Package | Purpose |
|---------|---------|
| `pi-subagents` | Parallel literature gathering and decomposition. |
| `pi-btw` | Fast side-thread `/btw` conversations without interrupting the main run. |
| `pi-docparser` | PDFs, Office docs, spreadsheets, and images. |
| `pi-web-access` | Web, GitHub, PDF, and media access. |
| `pi-markdown-preview` | Polished Markdown and LaTeX-heavy research writeups. |
| `@walterra/pi-charts` | Charts and quantitative visualizations. |
| `pi-generative-ui` | Interactive HTML-style widgets. |
| `pi-mermaid` | Diagrams in the TUI. |
| `@aliou/pi-processes` | Long-running experiments and log tails. |
| `pi-zotero` | Citation-library workflows. |
| `@kaiserlich-dev/pi-session-search` | Indexed session recall and summarize/resume UI. |
| `pi-schedule-prompt` | Recurring and deferred research jobs. |
| `@samfp/pi-memory` | Automatic preference and correction memory across sessions. |
| `@tmustier/pi-ralph-wiggum` | Long-running agent loops for iterative development. |

View File

@@ -0,0 +1,41 @@
---
title: Slash Commands
description: Repo-owned REPL slash commands
section: Reference
order: 2
---
This page documents the slash commands that Feynman owns in this repository: prompt templates from `prompts/` and extension commands from `extensions/research-tools/`.
Additional slash commands can appear at runtime from Pi core and bundled packages such as subagents, preview, session search, and scheduling. Use `/help` inside the REPL for the live command list instead of relying on a static copy of package-provided commands.
## Research Workflows
| Command | Description |
| --- | --- |
| `/deepresearch <topic>` | Run a thorough, source-heavy investigation on a topic and produce a durable research brief with inline citations. |
| `/lit <topic>` | Run a literature review on a topic using paper search and primary-source synthesis. |
| `/review <artifact>` | Simulate an AI research peer review with likely objections, severity, and a concrete revision plan. |
| `/audit <item>` | Compare a paper's claims against its public codebase and identify mismatches, omissions, and reproducibility risks. |
| `/replicate <paper>` | Plan or execute a replication workflow for a paper, claim, or benchmark. |
| `/compare <topic>` | Compare multiple sources on a topic and produce a source-grounded matrix of agreements, disagreements, and confidence. |
| `/draft <topic>` | Turn research findings into a polished paper-style draft with equations, sections, and explicit claims. |
| `/autoresearch <idea>` | Autonomous experiment loop — try ideas, measure results, keep what works, discard what doesn't, repeat. |
| `/watch <topic>` | Set up a recurring or deferred research watch on a topic, company, paper area, or product surface. |
## Project & Session
| Command | Description |
| --- | --- |
| `/log` | Write a durable session log with completed work, findings, open questions, and next steps. |
| `/jobs` | Inspect active background research work, including running processes and scheduled follow-ups. |
| `/help` | Show grouped Feynman commands and prefill the editor with a selected command. |
| `/init` | Bootstrap AGENTS.md and session-log folders for a research project. |
## Setup
| Command | Description |
| --- | --- |
| `/alpha-login` | Sign in to alphaXiv from inside Feynman. |
| `/alpha-status` | Show alphaXiv authentication status. |
| `/alpha-logout` | Clear alphaXiv auth from inside Feynman. |

View File

@@ -0,0 +1,40 @@
---
title: AlphaXiv
description: Paper search and analysis tools
section: Tools
order: 1
---
## Overview
AlphaXiv powers Feynman's academic paper workflows. All tools require an alphaXiv account — sign in with `feynman alpha login`.
## Tools
### alpha_search
Paper discovery with three search modes:
- **semantic** — Meaning-based search across paper content
- **keyword** — Traditional keyword matching
- **agentic** — AI-powered search that interprets your intent
### alpha_get_paper
Fetch a paper's report (structured summary) or full raw text by arXiv ID.
### alpha_ask_paper
Ask a targeted question about a specific paper. Returns an answer grounded in the paper's content.
### alpha_annotate_paper
Add persistent local notes to a paper. Annotations are stored locally and persist across sessions.
### alpha_list_annotations
Recall all annotations across papers and sessions.
### alpha_read_code
Read source code from a paper's linked GitHub repository. Useful for auditing or replication planning.

View File

@@ -0,0 +1,34 @@
---
title: Preview
description: Preview generated artifacts in browser or PDF
section: Tools
order: 4
---
## Overview
The `preview_file` tool opens generated artifacts in your browser or PDF viewer.
## Usage
Inside the REPL:
```
/preview
```
Or Feynman will suggest previewing when you generate artifacts that benefit from rendered output (Markdown with LaTeX, HTML reports, etc.).
## Requirements
Preview requires `pandoc` for PDF/HTML rendering. Install it with:
```bash
feynman --setup-preview
```
## Supported formats
- Markdown (with LaTeX math rendering)
- HTML
- PDF

View File

@@ -0,0 +1,26 @@
---
title: Session Search
description: Search prior Feynman session transcripts
section: Tools
order: 3
---
## Overview
The `session_search` tool recovers prior Feynman work from stored session transcripts. Useful for picking up previous research threads or finding past findings.
## Usage
Inside the REPL:
```
/search
```
Or use the tool directly — Feynman will invoke `session_search` automatically when you reference prior work.
## What it searches
- Full session transcripts
- Tool outputs and agent results
- Generated artifacts and their content

View File

@@ -0,0 +1,34 @@
---
title: Web Search
description: Web search routing and configuration
section: Tools
order: 2
---
## Routing modes
Feynman supports three web search backends:
| Mode | Description |
|------|-------------|
| `auto` | Prefer Perplexity when configured, fall back to Gemini |
| `perplexity` | Force Perplexity Sonar |
| `gemini` | Force Gemini (default) |
## Default behavior
The default path is zero-config Gemini Browser via a signed-in Chromium profile. No API keys required.
## Check current config
```bash
feynman search status
```
## Advanced configuration
Edit `~/.feynman/web-search.json` directly to set:
- Gemini API keys
- Perplexity API keys
- Custom routing preferences

View File

@@ -0,0 +1,39 @@
---
title: Code Audit
description: Compare paper claims against public codebases
section: Workflows
order: 4
---
## Usage
```
/audit <item>
```
## What it does
Compares claims made in a paper against its public codebase. Surfaces mismatches, missing experiments, and reproducibility risks.
## What it checks
- Do the reported hyperparameters match the code?
- Are all claimed experiments present in the repository?
- Does the training loop match the described methodology?
- Are there undocumented preprocessing steps?
- Do evaluation metrics match the paper's claims?
## Example
```
/audit 2401.12345
```
## Output
An audit report with:
- Claim-by-claim verification
- Identified mismatches
- Missing components
- Reproducibility risk assessment

View File

@@ -0,0 +1,44 @@
---
title: Autoresearch
description: Autonomous experiment optimization loop
section: Workflows
order: 8
---
## Usage
```
/autoresearch <idea>
```
## What it does
Runs an autonomous experiment loop:
1. **Edit** — Modify code or configuration
2. **Commit** — Save the change
3. **Benchmark** — Run evaluation
4. **Evaluate** — Compare against baseline
5. **Keep or revert** — Persist improvements, roll back regressions
6. **Repeat** — Continue until the target is hit
## Tracking
Metrics are tracked in:
- `autoresearch.md` — Human-readable progress log
- `autoresearch.jsonl` — Machine-readable metrics over time
## Controls
```
/autoresearch <idea> # start or resume
/autoresearch off # stop, keep data
/autoresearch clear # delete all state, start fresh
```
## Example
```
/autoresearch optimize the learning rate schedule for better convergence
```

View File

@@ -0,0 +1,29 @@
---
title: Source Comparison
description: Compare multiple sources with agreement/disagreement matrix
section: Workflows
order: 6
---
## Usage
```
/compare <topic>
```
## What it does
Compares multiple sources on a topic. Builds an agreement/disagreement matrix showing where sources align and where they conflict.
## Example
```
/compare approaches to constitutional AI training
```
## Output
- Source-by-source breakdown
- Agreement/disagreement matrix
- Synthesis of key differences
- Assessment of which positions have stronger evidence

View File

@@ -0,0 +1,40 @@
---
title: Deep Research
description: Thorough source-heavy investigation with parallel agents
section: Workflows
order: 1
---
## Usage
```
/deepresearch <topic>
```
## What it does
Deep research runs a thorough, source-heavy investigation. It plans the research scope, delegates to parallel researcher agents, synthesizes findings, and adds inline citations.
The workflow follows these steps:
1. **Plan** — Clarify the research question and identify search strategy
2. **Delegate** — Spawn parallel researcher agents to gather evidence from different source types (papers, web, repos)
3. **Synthesize** — Merge findings, resolve contradictions, identify gaps
4. **Cite** — Add inline citations and verify all source URLs
5. **Deliver** — Write a durable research brief to `outputs/`
## Example
```
/deepresearch transformer scaling laws and their implications for compute-optimal training
```
## Output
Produces a structured research brief with:
- Executive summary
- Key findings organized by theme
- Evidence tables with source links
- Open questions and suggested next steps
- Numbered sources section with direct URLs

View File

@@ -0,0 +1,37 @@
---
title: Draft Writing
description: Paper-style draft generation from research findings
section: Workflows
order: 7
---
## Usage
```
/draft <topic>
```
## What it does
Produces a paper-style draft with structured sections. Writes to `papers/`.
## Structure
The generated draft includes:
- Title
- Abstract
- Introduction / Background
- Method or Approach
- Evidence and Analysis
- Limitations
- Conclusion
- Sources
## Example
```
/draft survey of differentiable physics simulators
```
The writer agent works only from supplied evidence — it never fabricates content. If evidence is insufficient, it explicitly notes the gaps.

View File

@@ -0,0 +1,31 @@
---
title: Literature Review
description: Map consensus, disagreements, and open questions
section: Workflows
order: 2
---
## Usage
```
/lit <topic>
```
## What it does
Runs a structured literature review that searches across academic papers and web sources. Explicitly separates consensus findings from disagreements and open questions.
## Example
```
/lit multimodal reasoning benchmarks for large language models
```
## Output
A structured review covering:
- **Consensus** — What the field agrees on
- **Disagreements** — Where sources conflict
- **Open questions** — What remains unresolved
- **Sources** — Direct links to all referenced papers and articles

View File

@@ -0,0 +1,42 @@
---
title: Replication
description: Plan replications of papers and claims
section: Workflows
order: 5
---
## Usage
```
/replicate <paper or claim>
```
## What it does
Extracts key implementation details from a paper, identifies what's needed to replicate the results, and asks where to run before executing anything.
Before running code, Feynman asks you to choose an execution environment:
- **Local** — run in the current working directory
- **Virtual environment** — create an isolated venv/conda env first
- **Cloud** — delegate to a remote Agent Computer machine
- **Plan only** — produce the replication plan without executing
## Example
```
/replicate "chain-of-thought prompting improves math reasoning"
```
## Output
A replication plan covering:
- Key claims to verify
- Required resources (compute, data, models)
- Implementation details extracted from the paper
- Potential pitfalls and underspecified details
- Step-by-step replication procedure
- Success criteria
If an execution environment is selected, also produces runnable scripts and captured results.

View File

@@ -0,0 +1,49 @@
---
title: Peer Review
description: Simulated peer review with severity-graded feedback
section: Workflows
order: 3
---
## Usage
```
/review <artifact>
```
## What it does
Simulates a tough-but-fair peer review for AI research artifacts. Evaluates novelty, empirical rigor, baselines, ablations, and reproducibility.
The reviewer agent identifies:
- Weak baselines
- Missing ablations
- Evaluation mismatches
- Benchmark leakage
- Under-specified implementation details
## Severity levels
Feedback is graded by severity:
- **FATAL** — Fundamental issues that invalidate the claims
- **MAJOR** — Significant problems that need addressing
- **MINOR** — Small improvements or clarifications
## Example
```
/review outputs/scaling-laws-brief.md
```
## Output
Structured review with:
- Summary of the work
- Strengths
- Weaknesses (severity-graded)
- Questions for the authors
- Verdict (accept / revise / reject)
- Revision plan

View File

@@ -0,0 +1,29 @@
---
title: Watch
description: Recurring research monitoring
section: Workflows
order: 9
---
## Usage
```
/watch <topic>
```
## What it does
Schedules a recurring research watch. Sets a baseline of current knowledge and defines what constitutes a meaningful change worth reporting.
## Example
```
/watch new papers on test-time compute scaling
```
## How it works
1. Feynman establishes a baseline by surveying current sources
2. Defines change signals (new papers, updated results, new repos)
3. Schedules periodic checks via `pi-schedule-prompt`
4. Reports only when meaningful changes are detected