Sitemap
Web ScrapingDefinition
A sitemap is an XML file that lists all the URLs on a website, along with metadata like last modification date and priority. It helps search engines and crawlers discover and index all pages efficiently.
How It Relates to CrawlForge
Sitemaps provide a complete inventory of a website's pages without needing to discover them by following links. This makes them invaluable for comprehensive scraping, SEO audits, and content migration where you need to process every page.
CrawlForge map_site generates sitemaps for any domain, discovering URLs through both link-following and existing sitemap files. This gives you a reliable starting point for batch operations with batch_scrape.
Related CrawlForge Tools
Related Terms
Web Crawler
A web crawler is a program that systematically browses the web by following links from page to page. Crawlers discover and index content across entire websites or domains.
Robots.txt
Robots.txt is a standard text file placed at the root of a website that tells web crawlers which pages they are allowed or disallowed from accessing. It is part of the Robots Exclusion Protocol.
SEO Audit
An SEO audit is a comprehensive analysis of a website's search engine optimization performance. It evaluates technical SEO, on-page content, metadata, site structure, and identifies opportunities for improvement.
Schema Markup
Schema markup is a vocabulary of tags (from schema.org) that you add to HTML to improve how search engines read and represent your page. It defines types like Product, Article, Organization, and their properties.
Start Scraping with 1,000 Free Credits
Get started with CrawlForge today. No credit card required.
Start scraping with 1,000 free credits