Files
feynman/skills/research/replication/SKILL.md
2026-03-20 11:05:58 -07:00

1.7 KiB

name, description
name description
replication Use this when the task is to reproduce a paper result, benchmark a claim, rebuild an experiment, or evaluate whether a published result holds in practice.

Replication

When To Use

Use this skill for:

  • paper reproduction
  • benchmark recreation
  • ablation reruns
  • claim verification through code and experiments

Procedure

  1. Identify the canonical source paper and inspect it with alpha_get_paper.
  2. Extract the exact target:
    • task
    • dataset
    • model or method
    • metrics
    • hardware or runtime assumptions
  3. Use alpha_ask_paper to pull out the exact details missing from the report.
  4. If the paper has a public repository, inspect it with alpha_read_code.
  5. Search the local workspace for existing code, notebooks, configs, and datasets.
  6. Write down the missing pieces explicitly before running anything.
  7. If the environment is sufficient, implement the minimal runnable reproduction path.
  8. Run the experiment with built-in file and shell tools.
  9. Save:
    • commands used
    • configs
    • raw outputs
    • summarized results
  10. Compare observed results with the paper and explain gaps.
  11. If the paper had a practical gotcha, attach it with alpha_annotate_paper.

Pitfalls

  • Do not claim replication succeeded if key conditions were missing.
  • Do not compare different metrics as if they were equivalent.
  • Do not ignore dataset or preprocessing mismatch.
  • Do not hide failed runs; record them and explain them.

Verification

A good replication outcome includes:

  • the exact command path
  • the data or config used
  • the observed metrics
  • a clear statement of match, partial match, or mismatch