parallel-deep-research▌
parallel-web/parallel-agent-skills · updated Apr 8, 2026
Exhaustive research with configurable depth, latency, and cost trade-offs for complex topics.
- ›Three processor tiers (pro-fast, ultra-fast, ultra) ranging from 30 seconds to 25 minutes, with cost scaling from 1x to 3x baseline
- ›Asynchronous execution with polling: kick off research instantly, monitor progress via URL, retrieve results when ready without blocking
- ›Outputs formatted markdown report and JSON metadata; executive summary printed to stdout for quick overview
- ›Designed for e
Deep Research
Research topic: $ARGUMENTS
When to use (vs parallel-web-search)
ONLY use this skill when the user explicitly requests deep/exhaustive research. Deep research is 10-100x slower and more expensive than parallel-web-search. For normal "research X" requests, quick lookups, or fact-checking, use parallel-web-search instead.
Step 1: Start the research
parallel-cli research run "$ARGUMENTS" --processor pro-fast --no-wait --json
If this is a follow-up to a previous research or enrichment task where you know the interaction_id, add context chaining:
parallel-cli research run "$ARGUMENTS" --processor lite --no-wait --json --previous-interaction-id "$INTERACTION_ID"
By chaining interaction_id values across requests, each follow-up question automatically has the full context of prior turns — so you can drill deeper into a topic without restating what was already researched. Use --processor lite for follow-ups since the heavy research was already done in the initial turn and the follow-up just needs to build on that context.
This returns instantly. Do NOT omit --no-wait — without it the command blocks for minutes and will time out.
Processor options (choose based on user request):
| Processor | Expected latency | Use when |
|---|---|---|
pro-fast |
30s – 5 min | Default — good balance of depth and speed |
ultra-fast |
1 – 10 min | Deeper analysis, more sources (~2x cost) |
ultra |
5 – 25 min | Maximum depth, only when explicitly requested (~3x cost) |
Parse the JSON output to extract the run_id, interaction_id, and monitoring URL. Immediately tell the user:
- Deep research has been kicked off
- The expected latency for the processor tier chosen (from the table above)
- The monitoring URL where they can track progress
Tell them they can background the polling step to continue working while it runs.
Step 2: Poll for results
Choose a descriptive filename based on the topic (e.g., ai-chip-market-2026, react-vs-vue-comparison). Use lowercase with hyphens, no spaces.
parallel-cli research poll "$RUN_ID" -o "$FILENAME" --timeout 540
Important:
- Use
--timeout 540(9 minutes) to stay within tool execution limits - Do NOT pass
--json— the full output is large and will flood context. The-oflag writes results to files instead. - The
-oflag generates two output files:$FILENAME.json— metadata and basis$FILENAME.md— formatted markdown report
- The poll command prints an executive summary to stdout when the research completes. Share this executive summary with the user — it gives them a quick overview without having to open the files.
If the poll times out
Higher processor tiers can take longer than 9 minutes. If the poll exits without completing:
- Tell the user the research is still running server-side
- Re-run the same
parallel-cli research pollcommand to continue waiting
Response format
After step 1: Share the monitoring URL (for tracking progress only — it is not the final report).
After step 2:
- Share the executive summary that the poll command printed to stdout
- Tell the user the two generated file paths:
$FILENAME.md— formatted markdown report$FILENAME.json— metadata and basis
- Share the
interaction_idand tell the user they can ask follow-up questions that build on this research (e.g., "drill deeper into X" or "compare that to Y")
Do NOT re-share the monitoring URL after completion — the results are in the files, not at that link.
Ask the user if they would like to read through the files for more detail. Do NOT read the file contents into context unless the user asks.
Remember the interaction_id — if the user asks a follow-up question that relates to this research, use it as --previous-interaction-id in the next research or enrichment command.
Setup
If parallel-cli is not found, install and authenticate:
curl -fsSL https://parallel.ai/install.sh | bash
If unable to install that way, install via pipx instead:
pipx install "parallel-web-tools[cli]"
pipx ensurepath
Then authenticate:
parallel-cli login
Or set an API key: export PARALLEL_API_KEY="your-key"