CrawlForge
HomeUse CasesIntegrationsPricingDocumentationBlog
How to Scrape Websites in Cursor IDE with CrawlForge MCP
Tutorials
Back to Blog
Tutorials

How to Scrape Websites in Cursor IDE with CrawlForge MCP

C
CrawlForge Team
Engineering Team
April 14, 2026
9 min read

On this page

Cursor IDE is great at reasoning about your code but cannot see the live web. Add CrawlForge via Cursor's Model Context Protocol integration and the Composer gains 20 scraping tools -- no Python script, no curl, no leaving the editor.

Json

This guide walks through setting up web scraping in Cursor IDE step by step, with runnable examples for research, structured extraction, and competitor monitoring.

Table of Contents

  • Why Scrape from Inside Cursor?
  • Prerequisites
  • Step 1: Install the MCP Server
  • Step 2: Configure Cursor's MCP Settings
  • Step 3: Restart and Verify
  • Step 4: Your First Scrape in Composer
  • Full Example: Build a Competitor Price Tracker
  • Workflow: Use Scrapes to Write Code
  • Troubleshooting
  • FAQ

Why Scrape from Inside Cursor?

Cursor Composer treats MCP tools as first-class actions: it picks the right tool for a task, passes typed arguments, and feeds results back into the conversation. When you scrape websites in Cursor IDE through CrawlForge, the extracted data is immediately available for Cursor to generate tests, write TypeScript interfaces, or update dashboards. No copy-paste, no context switch.

If you already use Cursor rules to shape Composer's behavior, MCP tools slot right in -- rules describe how to code, tools expose what Composer can do.

Prerequisites

  • Cursor IDE 0.42+ -- download from cursor.com
  • Node.js 18+ -- node --version to check
  • CrawlForge account -- free at crawlforge.dev/signup

Step 1: Install the MCP Server

Bash

Confirm it is on your PATH:

Bash

Step 2: Configure Cursor's MCP Settings

Cursor reads MCP servers from ~/.cursor/mcp.json. Create it if it does not exist:

Bash

Paste this config (replace the key):

Json

On Windows the file lives at %USERPROFILE%\.cursor\mcp.json and the command should be crawlforge-mcp-server.cmd.

Step 3: Restart and Verify

  1. Quit Cursor completely (Cmd+Q on macOS).
  2. Reopen the project.
  3. Go to Settings -> Features -> MCP. You should see crawlforge with a green dot and 20 tools listed.

If the server is red or the tools list is empty, skip to Troubleshooting.

Step 4: Your First Scrape in Composer

Open Composer (Cmd+I) and paste:

Use CrawlForge to fetch https://news.ycombinator.com and list the top 5 story titles.

Cursor will call fetch_url (1 credit) and display the response. When you approve the tool call, Composer parses the HTML and returns a clean list.

Full Example: Build a Competitor Price Tracker

Say you want to track pricing changes on a competitor SaaS. Open Composer and paste:

Use scrape_structured to pull pricing from https://competitor.example.com/pricing. Fields: plan (h3), price (.price), features (ul li). Then generate a TypeScript type for the response.

Cursor issues a scrape_structured call with your selectors, returns JSON, and emits this TypeScript in the next editor chunk:

Typescript

Total cost: 2 credits per run. Schedule it via Vercel Cron or GitHub Actions and you have a free-tier price tracker.

Workflow: Use Scrapes to Write Code

The real unlock is feeding scraped data into Cursor's code generation. Proven patterns:

  • Type generation from live APIs: "Fetch https://api.example.com/users, then generate a Zod schema matching the response."
  • Test fixtures from real pages: "Scrape the top 3 articles from Hacker News and save them as JSON fixtures in tests/fixtures/."
  • Documentation extraction: "Use extract_content on the React docs for useState, then write an idiomatic example that matches."
  • Competitor feature parity: "Use map_site on competitor.com and flag any URL patterns we do not have in our own sitemap."

Each pattern is 1-5 credits per run and keeps you inside Cursor.

Credit Costs Summary

OperationToolCredits
Fetch HTMLfetch_url1
Clean textextract_text1
Readable articleextract_content2
CSS-selector extractscrape_structured2
Sitemap discoverymap_site2
Web searchsearch_web5
SPA with clicksscrape_with_actions5
Anti-bot bypassstealth_mode5

Troubleshooting

Tools list empty in Cursor Settings -- Cursor caches MCP config. Fully quit (Cmd+Q), then reopen. Check ~/.cursor/logs/ for parse errors.

"Command not found: crawlforge-mcp-server" -- npm's global bin is not on Cursor's PATH. Fix by setting an absolute path in mcp.json: "command": "/usr/local/bin/crawlforge-mcp-server".

Every call returns 401 -- API key missing or still the placeholder. Verify with: curl -H "Authorization: Bearer $CRAWLFORGE_API_KEY" https://crawlforge.dev/api/v1/credits/balance.

Cursor asks for approval on every tool call -- That is expected default behavior. Enable "Auto-approve for trusted servers" in MCP settings if you want Composer to run scrapes silently.

Composer ignores the MCP tool -- Explicitly prompt: "Use CrawlForge's scrape_structured tool to...". Cursor sometimes defaults to its built-in web fetch, which is less capable.

Next Steps

  • Read the Cursor rules guide to optimize Composer behavior for scraping
  • Browse the 20-tools overview to see what else you can automate
  • Check the getting started docs for API reference and credit pricing
  • Compare vendors at Firecrawl alternative

Start free with 1,000 credits at crawlforge.dev/signup.

Tags

Cursor-IDEweb-scrapingMCPtutorialComposerdeveloper-toolsAI-agents

About the Author

C

CrawlForge Team

Engineering Team

Building the most comprehensive web scraping MCP server. We create tools that help developers extract, analyze, and transform web data for AI applications.

On this page

Frequently Asked Questions

Where does Cursor IDE look for MCP server configuration?+

Cursor reads MCP configs from ~/.cursor/mcp.json on macOS and Linux, and %USERPROFILE%\\.cursor\\mcp.json on Windows. Changes require a full restart (Cmd+Q then reopen) for Cursor to pick them up.

Can Cursor Composer run CrawlForge tools automatically?+

Yes. Enable auto-approve for trusted servers in Settings > Features > MCP. Composer will then call fetch_url, scrape_structured, and other CrawlForge tools without a per-call confirmation dialog, which is ideal for iterative workflows.

How is this different from the Cursor Rules guide?+

Cursor Rules control how Composer writes code (style, frameworks, file conventions). MCP tools like CrawlForge add new capabilities -- web scraping, structured extraction, deep research. You usually want both: rules to shape output and MCP tools to expand what Composer can do.

Does Cursor support MCP tools on the free plan?+

MCP integration works on all Cursor plans, including free. You only pay for Cursor Pro if you need the advanced models. CrawlForge also has a free tier (1,000 credits/month) so the full stack can run at zero cost for light usage.

Can I use Cursor MCP tools in a monorepo with multiple projects?+

Yes -- MCP config is global to your user profile, not per-project. All projects opened in Cursor have access to the same CrawlForge tools. For project-specific overrides, add a .cursor/mcp.json inside the repo root and Cursor merges it with the global config.

Related Articles

How to Scrape Websites with Claude Code (2026 Guide)
Tutorials

How to Scrape Websites with Claude Code (2026 Guide)

Scrape any website from your terminal with Claude Code and CrawlForge MCP. Fetch pages, extract data, bypass anti-bot -- in under 2 minutes.

C
CrawlForge Team
|
Apr 14
|
10m
How to Scrape Websites in Zed AI with CrawlForge MCP
Tutorials

How to Scrape Websites in Zed AI with CrawlForge MCP

Add web scraping to Zed AI in 3 minutes. Configure CrawlForge MCP in Zed so your editor can fetch, extract, and research live web data on demand.

C
CrawlForge Team
|
Apr 14
|
9m
How to Scrape Websites with GitHub Copilot Agents in VS Code
Tutorials

How to Scrape Websites with GitHub Copilot Agents in VS Code

Add web scraping to GitHub Copilot agents in VS Code. Configure CrawlForge MCP so Copilot can fetch, extract, and research live web data on demand.

C
CrawlForge Team
|
Apr 14
|
10m

Footer

CrawlForge

Enterprise web scraping for AI Agents. 20 specialized MCP tools designed for modern developers building intelligent systems.

Product

  • Features
  • Pricing
  • Use Cases
  • Integrations
  • Changelog

Resources

  • Getting Started
  • API Reference
  • Templates
  • Guides
  • Blog
  • FAQ

Developers

  • MCP Protocol
  • Claude Desktop
  • Cursor IDE
  • LangChain
  • LlamaIndex

Company

  • About
  • Contact
  • Privacy
  • Terms

Stay updated

Get the latest updates on new tools and features.

Built with Next.js and MCP protocol

© 2025-2026 CrawlForge. All rights reserved.