productivity

TaskQueue

by chriscarrollsmith

TaskQueue is a project tracking software for managing complex projects, featuring progress tracking and approval checkpo

Structured task management system that breaks down complex projects into manageable tasks with progress tracking, user approval checkpoints, and support for multiple LLM providers.

github stars

67

Multiple LLM provider supportBuilt-in CLI for task inspectionUser approval workflow system

best for

  • / Managing multi-step software development projects
  • / Breaking down complex workflows into manageable chunks
  • / Teams needing structured task approval processes
  • / AI assistants handling long-running projects

capabilities

  • / Create projects with structured task breakdowns
  • / Track task progress and completion status
  • / Generate project plans using AI/LLMs
  • / Set user approval checkpoints for tasks
  • / Manage multiple projects simultaneously
  • / Export task details and project status

what it does

Manages complex projects by breaking them into structured tasks with progress tracking and user approval checkpoints. Includes AI-powered project planning using multiple LLM providers.

about

TaskQueue is a community-built MCP server published by chriscarrollsmith that provides AI assistants with tools and capabilities via the Model Context Protocol. TaskQueue is a project tracking software for managing complex projects, featuring progress tracking and approval checkpo It is categorized under productivity. This server exposes 14 tools that AI clients can invoke during conversations and coding sessions.

how to install

You can install TaskQueue in your AI client of choice. Use the install panel on this page to get one-click setup for Cursor, Claude Desktop, VS Code, and other MCP-compatible clients. This server runs locally on your machine via the stdio transport.

license

MIT

TaskQueue is released under the MIT license. This is a permissive open-source license, meaning you can freely use, modify, and distribute the software.

readme

MCP Task Manager

smithery badge

MCP Task Manager (npm package: taskqueue-mcp) is a Model Context Protocol (MCP) server for AI task management. This tool helps AI assistants handle multi-step tasks in a structured way, with optional user approval checkpoints.

Features

  • Task planning with multiple steps
  • Progress tracking
  • User approval of completed tasks
  • Project completion approval
  • Task details visualization
  • Task status state management
  • Enhanced CLI for task inspection and management

Basic Setup

Usually you will set the tool configuration in Claude Desktop, Cursor, or another MCP client as follows:

{
  "tools": {
    "taskqueue": {
      "command": "npx",
      "args": ["-y", "taskqueue-mcp"]
    }
  }
}

To use the CLI utility, you can install the package globally and then use the following command:

npx taskqueue --help

This will show the available commands and options.

Advanced Configuration

The task manager supports multiple LLM providers for generating project plans. You can configure one or more of the following environment variables depending on which providers you want to use:

  • OPENAI_API_KEY: Required for using OpenAI models (e.g., GPT-4)
  • GOOGLE_GENERATIVE_AI_API_KEY: Required for using Google's Gemini models
  • DEEPSEEK_API_KEY: Required for using Deepseek models

To generate project plans using the CLI, set these environment variables in your shell:

export OPENAI_API_KEY="your-api-key"
export GOOGLE_GENERATIVE_AI_API_KEY="your-api-key"
export DEEPSEEK_API_KEY="your-api-key"

Or you can include them in your MCP client configuration to generate project plans with MCP tool calls:

{
  "tools": {
    "taskqueue": {
      "command": "npx",
      "args": ["-y", "taskqueue-mcp"],
      "env": {
        "OPENAI_API_KEY": "your-api-key",
        "GOOGLE_GENERATIVE_AI_API_KEY": "your-api-key",
        "DEEPSEEK_API_KEY": "your-api-key"
      }
    }
  }
}

Available MCP Tools

The TaskManager now uses a direct tools interface with specific, purpose-built tools for each operation:

Project Management Tools

  • list_projects: Lists all projects in the system
  • read_project: Gets details about a specific project
  • create_project: Creates a new project with initial tasks
  • delete_project: Removes a project
  • add_tasks_to_project: Adds new tasks to an existing project
  • finalize_project: Finalizes a project after all tasks are done

Task Management Tools

  • list_tasks: Lists all tasks for a specific project
  • read_task: Gets details of a specific task
  • create_task: Creates a new task in a project
  • update_task: Modifies a task's properties (title, description, status)
  • delete_task: Removes a task from a project
  • approve_task: Approves a completed task
  • get_next_task: Gets the next pending task in a project
  • mark_task_done: Marks a task as completed with details

Task Status and Workflows

Tasks have a status field that can be one of:

  • not started: Task has not been started yet
  • in progress: Task is currently being worked on
  • done: Task has been completed (requires completedDetails)

Status Transition Rules

The system enforces the following rules for task status transitions:

  • Tasks follow a specific workflow with defined valid transitions:
    • From not started: Can only move to in progress
    • From in progress: Can move to either done or back to not started
    • From done: Can move back to in progress if additional work is needed
  • When a task is marked as "done", the completedDetails field must be provided to document what was completed
  • Approved tasks cannot be modified
  • A project can only be approved when all tasks are both done and approved

These rules help maintain the integrity of task progress and ensure proper documentation of completed work.

Usage Workflow

A typical workflow for an LLM using this task manager would be:

  1. create_project: Start a project with initial tasks
  2. get_next_task: Get the first pending task
  3. Work on the task
  4. mark_task_done: Mark the task as complete with details
  5. Wait for approval (user must call approve_task through the CLI)
  6. get_next_task: Get the next pending task
  7. Repeat steps 3-6 until all tasks are complete
  8. finalize_project: Complete the project (requires user approval)

CLI Commands

To use the CLI, you will need to install the package globally:

npm install -g taskqueue-mcp

Alternatively, you can run the CLI with npx using the --package=taskqueue-mcp flag to tell npx what package it's from.

npx --package=taskqueue-mcp taskqueue --help

Task Approval

By default, all tasks and projects will be auto-approved when marked "done" by the AI agent. To require manual human task approval, set autoApprove to false when creating a project.

Task approval is controlled exclusively by the human user through the CLI:

npx taskqueue approve-task -- <projectId> <taskId>

Options:

  • -f, --force: Force approval even if the task is not marked as done

Note: Tasks must be marked as "done" with completed details by the AI agent before they can be approved (unless using --force).

Listing Tasks and Projects

The CLI provides a command to list all projects and tasks:

npx taskqueue list-tasks

To view details of a specific project:

npx taskqueue list-tasks -- -p <projectId>

This command displays information about all projects in the system or a specific project, including:

  • Project ID and initial prompt
  • Completion status
  • Task details (title, description, status, approval)
  • Progress metrics (approved/completed/total tasks)

Data Schema and Storage

File Location

The task manager stores data in a JSON file that must be accessible to both the server and CLI.

The default platform-specific location is:

  • Linux: ~/.local/share/taskqueue-mcp/tasks.json
  • macOS: ~/Library/Application Support/taskqueue-mcp/tasks.json
  • Windows: %APPDATA% askqueue-mcp asks.json

Using a custom file path for storing task data is not recommended, because you have to remember to set the same path for both the MCP server and the CLI, or they won't be able to coordinate with each other. But if you do want to use a custom path, you can set the TASK_MANAGER_FILE_PATH environment variable in your MCP client configuration:

{
  "tools": {
    "taskqueue": {
      "command": "npx",
      "args": ["-y", "taskqueue-mcp"],
      "env": {
        "TASK_MANAGER_FILE_PATH": "/path/to/tasks.json"
      }
    }
  }
}

Then, before running the CLI, you should export the same path in your shell:

export TASK_MANAGER_FILE_PATH="/path/to/tasks.json"

Data Schema

The JSON file uses the following structure:

TaskManagerFile
├── projects: Project[]
    ├── projectId: string            # Format: "proj-{number}"
    ├── initialPrompt: string        # Original user request text
    ├── projectPlan: string          # Additional project details
    ├── completed: boolean           # Project completion status
    ├── autoApprove: boolean         # Set `false` to require manual user approval
    └── tasks: Task[]                # Array of tasks
        ├── id: string               # Format: "task-{number}"
        ├── title: string            # Short task title
        ├── description: string      # Detailed task description
        ├── status: string           # Task status: "not started", "in progress", or "done"
        ├── approved: boolean        # Task approval status
        ├── completedDetails: string # Completion information (required when status is "done")
        ├── toolRecommendations: string # Suggested tools that might be helpful for this task
        └── ruleRecommendations: string # Suggested rules/guidelines to follow for this task

License

MIT

FAQ

What is the TaskQueue MCP server?
TaskQueue is a Model Context Protocol (MCP) server profile on explainx.ai. MCP lets AI hosts (e.g. Claude Desktop, Cursor) call tools and resources through a standard interface; this page summarizes categories, install hints, and community ratings.
How do MCP servers relate to agent skills?
Skills are reusable instruction packages (often SKILL.md); MCP servers expose live capabilities. Teams frequently combine both—skills for workflows, MCP for APIs and data. See explainx.ai/skills and explainx.ai/mcp-servers for parallel directories.
How are reviews shown for TaskQueue?
This profile displays 10 aggregated ratings (sample rows for discoverability plus signed-in user reviews). Average score is about 4.5 out of 5—verify behavior in your own environment before production use.
MCP server reviews

Ratings

4.510 reviews
  • Shikha Mishra· Oct 10, 2024

    TaskQueue is among the better-indexed MCP projects we tried; the explainx.ai summary tracks the official description.

  • Piyush G· Sep 9, 2024

    We evaluated TaskQueue against two servers with overlapping tools; this profile had the clearer scope statement.

  • Chaitanya Patil· Aug 8, 2024

    Useful MCP listing: TaskQueue is the kind of server we cite when onboarding engineers to host + tool permissions.

  • Sakshi Patil· Jul 7, 2024

    TaskQueue reduced integration guesswork — categories and install configs on the listing matched the upstream repo.

  • Ganesh Mohane· Jun 6, 2024

    I recommend TaskQueue for teams standardizing on MCP; the explainx.ai page compares cleanly with sibling servers.

  • Oshnikdeep· May 5, 2024

    Strong directory entry: TaskQueue surfaces stars and publisher context so we could sanity-check maintenance before adopting.

  • Dhruvi Jain· Apr 4, 2024

    TaskQueue has been reliable for tool-calling workflows; the MCP profile page is a good permalink for internal docs.

  • Rahul Santra· Mar 3, 2024

    According to our notes, TaskQueue benefits from clear Model Context Protocol framing — fewer ambiguous “AI plugin” claims.

  • Pratham Ware· Feb 2, 2024

    We wired TaskQueue into a staging workspace; the listing’s GitHub and npm pointers saved time versus hunting across READMEs.

  • Yash Thakker· Jan 1, 2024

    TaskQueue is a well-scoped MCP server in the explainx.ai directory — install snippets and categories matched our Claude Code setup.