{
  "version": "1.0",
  "name": "GPT Researcher",
  "description": "Skills published by GPT Researcher for AI coding agents and autonomous agent platforms.",
  "homepage": "https://gptr.dev",
  "skills": [
    {
      "name": "deep-research",
      "description": "Conduct autonomous, multi-source deep research on a topic and produce a long-form report with inline citations. Use when the user asks for a research brief, market analysis, due-diligence write-up, or literature review that requires synthesizing 10-30 sources.",
      "url": "https://docs.gptr.dev/docs/gpt-researcher/getting-started/introduction",
      "tags": ["research", "report", "citations", "autonomous", "long-form"],
      "inputs": ["query: string", "report_type?: string"],
      "outputs": ["report: markdown", "sources: string[]"]
    },
    {
      "name": "quick-search",
      "description": "Run a fast web search across a configured retriever (Tavily, Bing, Google, DuckDuckGo, SearXNG) and return snippet-level results. Use when an agent needs grounded search snippets without a full research run.",
      "url": "https://docs.gptr.dev/docs/gpt-researcher/search-engines",
      "tags": ["search", "snippets", "retriever"],
      "inputs": ["query: string", "retriever?: string"],
      "outputs": ["results: { title, url, content }[]"]
    },
    {
      "name": "mcp-server",
      "description": "Run gptr-mcp, the official Model Context Protocol server that exposes deep_research, quick_search, write_report, get_research_sources, and get_research_context as MCP tools. Compatible with Claude Desktop, Cursor, ChatGPT custom GPTs, n8n, and any MCP-compatible client.",
      "url": "https://github.com/assafelovic/gptr-mcp",
      "tags": ["mcp", "model-context-protocol", "claude", "cursor", "agent-integration"],
      "inputs": [],
      "outputs": ["tools: deep_research, quick_search, write_report, get_research_sources, get_research_context"]
    },
    {
      "name": "self-host",
      "description": "Self-host GPT Researcher as a FastAPI service or a Docker container. Configure your own LLM provider (OpenAI, Anthropic, Google, Mistral, DeepSeek, local via Ollama) and retriever.",
      "url": "https://docs.gptr.dev/docs/gpt-researcher/getting-started/getting-started-docker",
      "tags": ["self-host", "docker", "fastapi", "deployment"],
      "inputs": [],
      "outputs": ["http server on configurable port"]
    }
  ]
}
