UseScraper网页提取
STDIO基于UseScraper的网页内容抓取服务
基于UseScraper的网页内容抓取服务
This is a TypeScript-based MCP server that provides web scraping capabilities using the UseScraper API. It exposes a single tool 'scrape' that can extract content from web pages in various formats.
scrape
- Extract content from a webpage
url
(required): The URL of the webpage to scrapeformat
(optional): The format to save the content (text, html, markdown). Default: markdownadvanced_proxy
(optional): Use advanced proxy to circumvent bot detection. Default: falseextract_object
(optional): Object specifying data to extractTo install UseScraper for Claude Desktop automatically via Smithery:
npx -y @smithery/cli install usescraper-server --client claude
Clone the repository:
git clone https://github.com/your-repo/usescraper-server.git cd usescraper-server
Install dependencies:
npm install
Build the server:
npm run build
To use with Claude Desktop, add the server config:
On MacOS: ~/Library/Application Support/Claude/claude_desktop_config.json
On Windows: %APPDATA%/Claude/claude_desktop_config.json
{ "mcpServers": { "usescraper-server": { "command": "node", "args": ["/path/to/usescraper-server/build/index.js"], "env": { "USESCRAPER_API_KEY": "your-api-key-here" } } } }
Replace /path/to/usescraper-server
with the actual path to the server and your-api-key-here
with your UseScraper API key.
Once configured, you can use the 'scrape' tool through the MCP interface. Example usage:
{ "name": "scrape", "arguments": { "url": "https://example.com", "format": "markdown" } }
For development with auto-rebuild:
npm run watch
Since MCP servers communicate over stdio, debugging can be challenging. We recommend using the MCP Inspector, which is available as a package script:
npm run inspector
The Inspector will provide a URL to access debugging tools in your browser.