Repo Crawler for AI agents.
Turn any GitHub repo into structured intelligence. Metadata, issues, security alerts, SBOMs — all through one tool call.
Quick scan
get_repo_summary({ owner: "facebook", repo: "react" })
Deep crawl
crawl_repo({ owner: "myorg", repo: "api", tier: "3" })
Compare
compare_repos({ repos: [{ owner: "vitejs", repo: "vite" }, ...] })
Features
Everything an AI agent needs to understand a repository.
5 MCP Tools
Crawl repos, crawl orgs, summarize, compare, and export — each purpose-built for a different workflow.
3-Tier Data Model
Start light with metadata, go deeper with issues and PRs, or get the full security picture with Dependabot, SBOM, and secret scanning.
Section-Selective
Only calls the APIs you ask for. Request specific sections to save quota and shrink response size.
Graceful Degradation
A 403 on Dependabot won't kill the crawl. Each section fetches independently with permission tracking.
Rate-Limit Safe
Built-in Octokit throttling with automatic retry on 429s. No more blowing through your API quota.
Safe Exports
CSV with formula injection prevention, Markdown with pipe escaping. Export crawl data without security risks.
Quick Start
Add to your MCP client config and start crawling.
Claude Code / Claude Desktop
{
"mcpServers": {
"repo-crawler": {
"command": "npx",
"args": ["-y", "@mcptoolshop/repo-crawler-mcp"],
"env": {
"GITHUB_TOKEN": "ghp_your_token"
}
}
}
} Try it
// Quick repo triage (4 API calls)
get_repo_summary({ owner: "expressjs", repo: "express" })
// Full security audit
crawl_repo({ owner: "myorg", repo: "api", tier: "3" })
// Compare frameworks side-by-side
compare_repos({ repos: [
{ owner: "sveltejs", repo: "svelte" },
{ owner: "vuejs", repo: "core" },
{ owner: "facebook", repo: "react" }
]}) Data Tiers
Choose the depth of data you need. Each tier includes everything below it.
Tools
Five MCP tools covering the full GitHub data surface.
crawl_repo({ owner, repo, tier, sections? })
Crawl a single repository at any data tier. Supports section-selective fetching and configurable limits for commits, issues, PRs, and alerts.
crawl_org({ org, tier, min_stars?, language? })
Crawl every repo in an organization with filters for stars, language, forks, and archived status.
get_repo_summary({ owner, repo })
Quick human-readable summary using only 4 API calls. Ideal for triage and initial assessment.
compare_repos({ repos, aspects? })
Side-by-side comparison of 2–5 repos. Stars, languages, activity, community health, and size.
export_data({ data, format, sections? })
Export crawl results as JSON, CSV, or Markdown. CSV includes formula injection prevention.