π The most advanced free sitemap generator - Now with async crawling, XML export, and professional SEO features!
- Async Concurrent Crawling - Crawl up to 50 pages simultaneously
- Connection Pooling - Reuse connections for faster crawling
- Smart Rate Limiting - Respects server resources while maximizing speed
- XML Sitemap - Full sitemaps.org compliance
- XML Sitemap Index - For large sites (50,000+ URLs)
- GZip Compression - Reduce file sizes by 80%+
- Text, CSV, JSON - Multiple export options
- Automatic Priority Calculation - Based on page depth and importance
- Change Frequency Detection - Smart changefreq assignment
- Last Modified Extraction - From HTTP headers
- Image Sitemap Support - Include page images
- Robots.txt Compliance - Optional respect for crawl rules
The Free Sitemap Generator is a professional-grade desktop application for creating XML sitemaps. It crawls your website, discovers all pages, and generates standards-compliant sitemaps for search engine submission.
- π Website Owners - Improve SEO and search indexing
- π’ SEO Professionals - Client site audits and optimization
- π¨βπ» Developers - Automated sitemap generation
- π¨ Agencies - Batch site processing
# Clone the repository
git clone https://github.com/jtgsystems/free-sitemap-generator.git
cd free-sitemap-generator
# Install dependencies
pip install -r requirements.txt
# Run the application
python main.py- Enter your website URL (e.g.,
https://example.com) - Configure crawl settings (optional)
- Click "Start Crawl"
- Export your sitemap in your preferred format
| Feature | Description |
|---|---|
| Async Concurrent | Crawl up to 50 pages simultaneously |
| Depth Control | Limit crawl depth (1-10 levels) |
| URL Limits | Configurable max URLs (100-50,000) |
| Smart Filtering | Excludes non-HTML, external links |
| Robots.txt | Optional compliance with crawl rules |
| Rate Limiting | Configurable delay between requests |
| Format | Description | Best For |
|---|---|---|
| XML Sitemap | Standard sitemaps.org format | Google, Bing submission |
| XML + GZip | Compressed XML | Large sites, bandwidth saving |
| Sitemap Index | Multiple sitemap files | 50,000+ URLs |
| Text | One URL per line | Quick reviews |
| CSV | Spreadsheet format | Analysis in Excel |
| JSON | Structured data | API integration |
- Max Depth - How many link levels to follow
- Max URLs - Limit total discovered URLs
- Concurrency - Number of simultaneous requests
- Crawl Delay - Seconds between requests
- Respect robots.txt - Follow crawl rules
- Include Images - Add image sitemap entries
<url>
<loc>https://example.com/</loc>
<lastmod>2026-02-05T10:30:00+00:00</lastmod>
<changefreq>daily</changefreq>
<priority>1.0</priority>
</url>- Priority Calculation - Homepage gets 1.0, deeper pages get lower priority
- Change Frequency - Detected from URL patterns (blog=weekly, about=monthly)
- Last Modified - Extracted from HTTP Last-Modified headers
<url>
<loc>https://example.com/gallery</loc>
<image:image>
<image:loc>https://example.com/photo1.jpg</image:loc>
</image:image>
</url># Install PyInstaller
pip install pyinstaller
# Build
python setup.py
# Find executable in dist/ folderfree-sitemap-generator/
βββ sitemap_generator/ # Main package
β βββ __init__.py
β βββ crawler.py # Async web crawler
β βββ exporter.py # XML/export formats
β βββ gui.py # PyQt6 interface
βββ main.py # Entry point
βββ requirements.txt # Dependencies
βββ setup.py # Build script
βββ README.md # This file
βββ LICENSE # MIT License
βββ .github/ # GitHub templates
- β sitemaps.org protocol
- β Google Search Console compatible
- β Bing Webmaster Tools compatible
- β 50,000 URL limit per file
- β 50MB file size limit
- β UTF-8 encoding
- β Proper XML escaping
- Crawling: Up to 500 pages/minute (depending on site)
- Memory: Efficient streaming for large sites
- CPU: Multi-threaded processing
- Network: Connection pooling, keep-alive
"Timeout fetching URL"
- Check your internet connection
- Increase timeout in settings
- Website may be blocking crawlers
"No URLs found"
- Ensure URL includes http:// or https://
- Check that website has crawlable links
- Verify robots.txt allows crawling
"Application freezes"
- Reduce concurrency setting
- Increase crawl delay
- Check system resources
| Code | Meaning | Solution |
|---|---|---|
| 403 | Access Forbidden | Check robots.txt, add User-Agent |
| 404 | Page Not Found | Check URL spelling |
| 500 | Server Error | Try again later |
| Timeout | Request Timeout | Increase timeout setting |
Contributions welcome! Please read CONTRIBUTING.md for guidelines.
# Clone repo
git clone https://github.com/jtgsystems/free-sitemap-generator.git
# Create virtual environment
python -m venv venv
source venv/bin/activate # Windows: venv\Scripts\activate
# Install dev dependencies
pip install -r requirements.txt
pip install pytest black flake8
# Run tests
pytest
# Format code
black sitemap_generator/This project is licensed under the MIT License - see LICENSE file.
- Built with PyQt6 for the GUI
- Async crawling powered by aiohttp
- HTML parsing by BeautifulSoup
- π Website: https://jtgsystems.com
- π» GitHub: https://github.com/jtgsystems/free-sitemap-generator
- π Issues: https://github.com/jtgsystems/free-sitemap-generator/issues
- π Wiki: https://github.com/jtgsystems/free-sitemap-generator/wiki
Made with β€οΈ by JTGSYSTEMS
sitemap generator crawler python pyqt6 gui web seo indexing discovery urls links analytics architecture structure navigation automation crawling spider performance lxml requests beautifulsoup multithreading progress reporting visualization export xml website audit diagnostics optimization accessibility compliance marketing content developers agencies freelancers startup enterprise batch queue resilience reliability uptime monitoring insights roadmap async concurrent gzip sitemap-index image-sitemap google bing search-console
