setup

alirezarezvani/claude-skills · updated Apr 8, 2026

$npx skills add https://github.com/alirezarezvani/claude-skills --skill setup
0 commentsdiscussion
summary

Set up a new autoresearch experiment with all required configuration.

skill.md

/ar:setup — Create New Experiment

Set up a new autoresearch experiment with all required configuration.

Usage

/ar:setup                                    # Interactive mode
/ar:setup engineering api-speed src/api.py "pytest bench.py" p50_ms lower
/ar:setup --list                             # Show existing experiments
/ar:setup --list-evaluators                  # Show available evaluators

What It Does

If arguments provided

Pass them directly to the setup script:

python {skill_path}/scripts/setup_experiment.py \
  --domain {domain} --name {name} \
  --target {target} --eval "{eval_cmd}" \
  --metric {metric} --direction {direction} \
  [--evaluator {evaluator}] [--scope {scope}]

If no arguments (interactive mode)

Collect each parameter one at a time:

  1. Domain — Ask: "What domain? (engineering, marketing, content, prompts, custom)"
  2. Name — Ask: "Experiment name? (e.g., api-speed, blog-titles)"
  3. Target file — Ask: "Which file to optimize?" Verify it exists.
  4. Eval command — Ask: "How to measure it? (e.g., pytest bench.py, python evaluate.py)"
  5. Metric — Ask: "What metric does the eval output? (e.g., p50_ms, ctr_score)"
  6. Direction — Ask: "Is lower or higher better?"
  7. Evaluator (optional) — Show built-in evaluators. Ask: "Use a built-in evaluator, or your own?"
  8. Scope — Ask: "Store in project (.autoresearch/) or user (~/.autoresearch/)?"

Then run setup_experiment.py with the collected parameters.

Listing

# Show existing experiments
python {skill_path}/scripts/setup_experiment.py --list

# Show available evaluators
python {skill_path}/scripts/setup_experiment.py --list-evaluators

Built-in Evaluators

Name Metric Use Case
benchmark_speed p50_ms (lower) Function/API execution time
benchmark_size size_bytes (lower) File, bundle, Docker image size
test_pass_rate pass_rate (higher) Test suite pass percentage
build_speed build_seconds (lower) Build/compile/Docker build time
memory_usage peak_mb (lower) Peak memory during execution
llm_judge_content ctr_score (higher) Headlines, titles, descriptions
llm_judge_prompt quality_score (higher) System prompts, agent instructions
llm_judge_copy engagement_score (higher) Social posts, ad copy, emails

After Setup

Report to the user:

  • Experiment path and branch name
  • Whether the eval command worked and the baseline metric
  • Suggest: "Run /ar:run {domain}/{name} to start iterating, or /ar:loop {domain}/{name} for autonomous mode."

Discussion

Product Hunt–style comments (not star reviews)
  • No comments yet — start the thread.
general reviews

Ratings

4.664 reviews
  • Chaitanya Patil· Dec 28, 2024

    setup fits our agent workflows well — practical, well scoped, and easy to wire into existing repos.

  • James Chawla· Dec 28, 2024

    Keeps context tight: setup is the kind of skill you can hand to a new teammate without a long onboarding doc.

  • Min Sharma· Dec 24, 2024

    setup fits our agent workflows well — practical, well scoped, and easy to wire into existing repos.

  • Noor Rao· Dec 24, 2024

    Registry listing for setup matched our evaluation — installs cleanly and behaves as described in the markdown.

  • Min Lopez· Dec 20, 2024

    Solid pick for teams standardizing on skills: setup is focused, and the summary matches what you get after install.

  • Anika Kapoor· Dec 12, 2024

    setup is among the better-maintained entries we tried; worth keeping pinned for repeat workflows.

  • Rahul Santra· Nov 27, 2024

    setup has been reliable in day-to-day use. Documentation quality is above average for community skills.

  • Anika Desai· Nov 27, 2024

    Solid pick for teams standardizing on skills: setup is focused, and the summary matches what you get after install.

  • Piyush G· Nov 19, 2024

    setup is among the better-maintained entries we tried; worth keeping pinned for repeat workflows.

  • Anika Jain· Nov 19, 2024

    We added setup from the explainx registry; install was straightforward and the SKILL.md answered most questions upfront.

showing 1-10 of 64

1 / 7