1.7 KiB
1.7 KiB
name, description
| name | description |
|---|---|
| replication | Use this when the task is to reproduce a paper result, benchmark a claim, rebuild an experiment, or evaluate whether a published result holds in practice. |
Replication
When To Use
Use this skill for:
- paper reproduction
- benchmark recreation
- ablation reruns
- claim verification through code and experiments
Procedure
- Identify the canonical source paper and inspect it with
alpha_get_paper. - Extract the exact target:
- task
- dataset
- model or method
- metrics
- hardware or runtime assumptions
- Use
alpha_ask_paperto pull out the exact details missing from the report. - If the paper has a public repository, inspect it with
alpha_read_code. - Search the local workspace for existing code, notebooks, configs, and datasets.
- Write down the missing pieces explicitly before running anything.
- If the environment is sufficient, implement the minimal runnable reproduction path.
- Run the experiment with built-in file and shell tools.
- Save:
- commands used
- configs
- raw outputs
- summarized results
- Compare observed results with the paper and explain gaps.
- If the paper had a practical gotcha, attach it with
alpha_annotate_paper.
Pitfalls
- Do not claim replication succeeded if key conditions were missing.
- Do not compare different metrics as if they were equivalent.
- Do not ignore dataset or preprocessing mismatch.
- Do not hide failed runs; record them and explain them.
Verification
A good replication outcome includes:
- the exact command path
- the data or config used
- the observed metrics
- a clear statement of match, partial match, or mismatch