by bneil
MCP Go Colly is a sophisticated web crawling framework that integrates the Model Context Protocol (MCP) with the powerful Colly web scraping library. This project aims to provide a flexible and extensible solution for extracting web content for large language model (LLM) applications.
git clone https://github.com/yourusername/mcp-go-colly.git cd mcp-go-colly
make deps
The project includes a Makefile with several useful commands:
# Build the binary (outputs to bin/mcp-go-colly) make build # Build for all platforms (Linux, Windows, macOS) make build-all # Run tests make test # Clean build artifacts make clean # Format code make fmt # Run linter make lint
All binaries will be generated in the bin/
directory.
Then you need to add the following configuration to the claude_desktop_config.json
file:
*Configuration content*
The crawler is implemented as an MCP tool that can be called with the following parameters:
*Configuration content*
result, err := crawlerTool.Call(ctx, mcp.CallToolRequest{ Params: struct{ Arguments map[string]interface{} }{ Arguments: map[string]interface{}{ "urls": []string{"https://example.com"}, "max_depth": 2, }, }, })
max_depth
: Set maximum crawl depth (default: 2)urls
: Single URL string or array of URLs to crawlMIT
No version information available