GPT Researcher MCP Server
Autonomous deep research agent that produces comprehensive reports with citations — multi-step web research with a STORM-inspired architecture.
Client Compatibility
What It Does
Autonomous deep research agent that takes a query, runs multi-step web research, and produces factual reports with citations. Five MCP tools: deep research (comprehensive, ~30 seconds), quick search (optimized for speed), write report (generate from research results), get research sources (citations), and get research context (full context).
Built on a STORM-inspired multi-agent architecture with 25.4K GitHub stars. The agent conducts research independently — planning searches, synthesizing findings, and generating structured reports.
What It Does Well
- Deep research quality exceeds simple search aggregation. The multi-agent architecture plans research strategies, runs multiple searches, cross-references sources, and synthesizes findings into coherent reports with proper citations. The output reads like research performed by a competent analyst, not a search results summary.
- Can act as both MCP server and MCP client. Configure RETRIEVER=tavily,mcp and the research agent connects to external MCP servers as data sources. This composability means GPT Researcher can pull from your company’s internal MCP servers alongside web search.
- Docker support with automatic transport detection. Runs STDIO locally and SSE in Docker.
docker-compose up -dgives you an SSE endpoint athttp://localhost:8000/ssewith no transport configuration needed.
What It Doesn’t Do Well
- Nested LLM dependency means you’re paying for two AI models. GPT Researcher uses OpenAI internally for research planning and synthesis. Running it through Claude means you’re paying for Claude’s API (or subscription) plus OpenAI’s API for the research agent. Two LLM bills for one workflow.
- 30+ second latency for deep research doesn’t fit interactive workflows. The deep_research tool is thorough but slow. If you need a quick answer, the quick_search tool exists, but the main value proposition — comprehensive research — requires patience. Some MCP clients may time out before results arrive.
- Setup is heavier than most MCP servers. Clone a repo, install Python dependencies, configure multiple API keys in a .env file, then run the server. No npx one-liner. The Docker path simplifies runtime but still requires API key configuration.
Setup Notes
Clone the gptr-mcp repo, install requirements, create a .env file with your OpenAI and Tavily API keys, and run python server.py. Docker alternative: docker-compose up -d for SSE at port 8000.
The main GPT Researcher repo (assafelovic/gpt-researcher) has the full documentation. The MCP server is a separate lightweight repo that wraps the research agent.
For deep research, extend your MCP client timeout to at least 60 seconds — 180 seconds recommended for complex queries.
Config
Local:
{
"mcpServers": {
"gpt-researcher": {
"command": "python",
"args": ["path/to/gptr-mcp/server.py"],
"env": {
"OPENAI_API_KEY": "your_openai_key",
"TAVILY_API_KEY": "your_tavily_key"
}
}
}
}
Docker (SSE):
{
"mcpServers": {
"gpt-researcher": {
"url": "http://localhost:8000/sse"
}
}
}
Tested With
- Claude Desktop on Windows 11
{
"mcpServers": {
"gpt-researcher-mcp": {
"command": "python",
"args": [
"server.py"
],
"env": {
"OPENAI_API_KEY": "your_openai_key",
"TAVILY_API_KEY": "your_tavily_key"
}
}
}
} Environment Variables
OPENAI_API_KEY(required) — OpenAI API key for the research agent's LLMTAVILY_API_KEY(required) — Tavily API key for web search (or other supported search provider)
Prerequisites
- Python 3.10+
- OpenAI API key
- Tavily API key (or alternative search provider)
Reviewed by J-Dub · February 22, 2026