Content Migration
Extract and restructure content from legacy sites for migration to modern platforms.
The Problem
Migrating content from legacy CMS platforms means extracting thousands of pages while preserving structure, metadata, and formatting. Manual copy-paste is error-prone and painfully slow.
The Solution
CrawlForge crawl_deep traverses entire sites following internal links, while extract_text pulls clean content from each page. Migrate hundreds of pages in a single automated run.
Code Example
// Crawl legacy site and extract all content for migration
const crawl = await mcp.crawl_deep({
url: "https://legacy-site.com/blog",
max_depth: 3,
follow_links: true,
include_patterns: ["/blog/*"],
});
// Extract clean text from each discovered page
const pages = await Promise.all(
crawl.urls.map(url =>
mcp.extract_text({ url, preserve_structure: true })
)
);
console.log(`Migrated ${pages.length} pages`);Tools Used
crawl_deep5 credits
extract_text1 credit
Estimated cost: ~6 credits per page
Ready to Get Started?
Every new account gets 1,000 free credits. No credit card required.
Start Free with 1,000 CreditsRelated Use Cases
SEO Analysis & Auditing
Audit your site and competitors for metadata, broken links, content gaps, and ranking opportunities.
extract_metadata (2 cr)map_site (3 cr)analyze_content (3 cr)
AI Training Data Collection
Collect and structure large-scale web datasets for fine-tuning and training AI models.
batch_scrape (5 cr)extract_content (2 cr)