Table of Contents
- Compare Sitemap Generation Methods
- Method Overview
- Comparison Of Sitemap Generation Methods
- Key Insights
- Real-World Scenario
- Manual XML Sitemap Creation
- Essential XML Tags
- Segmenting Sitemaps By Content
- Real-World Sitemap Example
- Metadata Best Practices
- Splitting Large Sitemaps
- Validating Your Sitemap
- Namespace Declaration
- Compressing Sitemap Files
- Reviewing Robots.txt Entry
- Using CMS Plugins And Online Tools
- Popular CMS Sitemap Plugins
- Scheduling Regular Updates
- Free Online Sitemap Generators
- Example Workflow
- Best Practice Tips
- Choosing The Right Option
- Verifying Submission And Monitoring
- Submission In Search Console
- Monitoring With CLI Validators
- Automating Sitemap Checks
- Verifying Robots.txt Entry
- Leveraging Bing Reports
- Best Practices And Feather Sitemap Settings
- Sitemap Index Files And Limits
- Lastmod And Priorities
- Feather Automatic Sitemap Generation
- Screenshot Of Feather Settings
- Custom Routes And Coverage
- Common Troubleshooting And Optimization Tips
- XML Syntax Fixes
- Optimizing Crawl Behavior
- Scheduling Updates
- Measuring Sitemap Health
- Frequently Asked Sitemap Questions
- URL Count And File Size Limits
- Comparing XML And HTML Sitemaps
- Choosing Automated Or Manual Updates

Related Posts
blog_related_media
blog_topic
blog_related_activities
blog_niche
blog_related_tips
unique_blog_element
You can whip up a sitemap in four distinct ways—by hand, with a CMS plugin, via a CLI utility, or through an online generator—and often in just minutes. Each path trades off setup time, automation, and oversight to suit your site’s size and update frequency.
Compare Sitemap Generation Methods
When you need absolute precision, writing an XML sitemap from scratch puts you in the driver’s seat. You decide every
<loc>, <lastmod>, <changefreq>, and <priority> tag. This is perfect for small to mid-sized static sites, where every URL matters.On the flip side, a plugin inside your CMS can refresh your sitemap on its own schedule. Tools for WordPress or Drupal can regenerate the file hourly or even more often. That means as soon as new content goes live, your sitemap stays up to date with zero extra clicks.
You’ll also see two other flavors:
- Manual XML: Total control, but you update it yourself.
- CMS Plugin: Hands-off updates right inside your dashboard.
- CLI Tool: Scripts generate sitemaps during your build or deploy process.
- Online Generator: Paste URLs, click download—no installation required.
Each approach has its moment. If you’re running CI/CD, a CLI tool slots right into your pipeline and guarantees consistency. In contrast, online generators are lightning-fast for quick demos, prototypes, or single-page projects.
Method Overview
To visualize how these four methods line up against project size and automation needs, check out the decision tree below.

This flow charts Manual XML at the top for fine-tuned control, steers Plugin solutions through CMS workflows, and pushes CLI options into fully automated pipelines.
Comparison Of Sitemap Generation Methods
Here’s a quick look at each technique’s complexity, automation level, and ideal scenario:
Method | Setup Complexity | Automation Level | Best Use Case |
Manual XML | Low | None | Small static sites needing precise control |
CMS Plugin | Medium | High | Content-heavy sites requiring regular updates |
CLI Tool | High | Very High | Automated build pipelines |
Online Generator | Very Low | Low | Quick prototypes and landing pages |
That table should help you pinpoint the right fit based on your team’s skills and release cadence.
Key Insights
- Manual XML shines when your content rarely changes and SEO precision is non-negotiable.
- CMS plugins are ideal for blogs, e-commerce stores, or news sites that publish daily.
- CLI tools give developers a repeatable, code-driven workflow—perfect for CI/CD.
- Online generators work wonders for marketers spinning up landing pages on the fly.
- Don’t forget to inspect your sitemap in Search Console to catch crawl errors early.
- Consider automating uploads via your build pipeline so sitemaps stay fresh without manual steps.
Real-World Scenario
A lean startup we know set up a nightly script using a CLI utility to rebuild their sitemap. The result? They cut merge conflicts in half and slashed the time between publishing and indexing by 80%. Fresh content hit search engines faster, with zero manual intervention at midnight.
Manual XML Sitemap Creation
Building an XML sitemap from scratch gives you complete oversight of every link. When you need to include—or exclude—specific pages with precision, nothing beats a hand-crafted file.
Your first line should always be the XML declaration:
This little header keeps parsers happy and avoids mysterious errors. Immediately after, wrap everything in a
<urlset> tag with the proper namespace to meet protocol requirements.Essential XML Tags
Here’s a quick rundown of the core elements each URL block needs:
- : The full, absolute address so crawlers know exactly where to go.
- : The date and time of the last update—search engines use this to prioritize fresh content.
- : A hint at how often your page changes, guiding crawl schedules.
- : A value between 0.0 and 1.0, indicating relative importance.
Segmenting Sitemaps By Content
Breaking your sitemap into topic-specific files keeps things organized and efficient. Major sites often split maps like this:
- sitemap-blog.xml for all article URLs
- sitemap-shop.xml for every product and SKU
- sitemap-categories.xml for section landing pages
If any file exceeds 50,000 URLs, shard it further. Smaller chunks are quicker to fetch and parse.
Real-World Sitemap Example
A retailer managing 120,000 SKUs divided its feed into four regional files—and saw indexing soar:
- sitemap-us.xml for North America
- sitemap-eu.xml for Europe
- sitemap-apac.xml for Asia-Pacific
- sitemap-other.xml for the rest
Top sellers got a 0.9 priority; slower movers sat at 0.6. The result? US indexing jumped 18% in one week, and EU coverage improved by 12%.
Metadata Best Practices
Consistency here prevents confusion:
- Stick to full ISO 8601 (
2024-03-10T14:00:00Z) when updates are frequent.
- Use
<changefreq>daily</changefreq>for high-velocity sections.
- Reserve 1.0 for your homepage, 0.5 for archives, and 0.1 for hidden or low-value pages.
Keep a sharp eye out for unclosed tags and mismatches—they’ll invalidate the entire file.
Splitting Large Sitemaps
File Name | URL Count Limit | Purpose |
sitemap-blog.xml | 50,000 | Daily list of blog articles |
sitemap-shop.xml | 50,000 | Product and SKU pages |
sitemap-index.xml | 50,000+ | References to individual sitemaps |
A sitemap index ties it all together:
When you combine topic-based splits, precise
<lastmod> timestamps, and index files capped at 50,000 URLs, indexing times can shrink from days to mere hours. Learn more about sitemap efficiency at OpenReplay’s blog.Validating Your Sitemap
Catch errors before search engines do:
- Open the XML in Chrome or Firefox to spot structural issues.
- Run it through XML Sitemap Validator for a thorough check.
- Integrate a CLI check in your build process so invalid files never make it live.
Namespace Declaration
Without the correct
xmlns, search engines might ignore your file entirely:<urlset xmlns="http://www.sitemaps.org/schemas/sitemap/0.9">
- For image sitemaps, add
xmlns:image="http://www.google.com/schemas/sitemap-image/1.1".
Compressing Sitemap Files
Gzip cuts download times in half:
- Rename your file to
sitemap.xml.gz.
- In robots.txt, reference it with:Sitemap: https://example.com/sitemap.xml.gz
Most crawlers accept compressed sitemaps directly.
Reviewing Robots.txt Entry
A misplaced robots.txt entry can sink your efforts:
- Ensure you have a
Sitemap:line pointing to your index (e.g.,https://example.com/sitemap-index.xml).
- Verify that
User-agent: *is allowed to access the sitemap path.
A quick scan of robots.txt at each update keeps your maps discoverable.
Using CMS Plugins And Online Tools
Automating sitemap generation with a CMS plugin turns what once felt like a chore into a swift, repeatable task. It frees you from manual edits and keeps your XML in sync with live content.

Whether you’re on WordPress, Joomla, Drupal, Shopify or a static site framework like Eleventy or Hugo, there’s a ready-made extension for your needs. On WordPress, Yoast SEO not only builds your XML sitemap but also lets you tweak change frequencies and priority tags. Joomla admins swear by OSMap for its URL filters and custom rules. Over in the Drupal world, the XML Sitemap module is the go-to for content-type filtering and scheduled updates.
- Yoast SEO (WordPress): Generates XML, sets update rates, assigns priority.
- OSMap (Joomla): Dashboard integration, inclusion rules, priority flags.
- XML Sitemap (Drupal): Content filters, automatic refresh scheduling.
- Shopify Sitemap Generator: Produces product and collection maps with no code.
- Eleventy Sitemap Plugin: Hooks into your build, supports custom URL patterns.
For instance, a mid-sized e-commerce site switched from hand-editing XML to OSMap and cut sitemap maintenance by 80%. Search bots picked up new pages within days, boosting overall crawl frequency.
Popular CMS Sitemap Plugins
Not sure which plugin fits your platform? Here’s a side-by-side look at the key features for major CMSs.
Plugin | Platform | Features |
Yoast SEO | WordPress | XML creation, frequency control, priority tags |
OSMap | Joomla | URL filters, custom rules, dashboard logs |
XML Sitemap | Drupal | Type-specific filters, scheduled generation |
Shopify Sitemap Gen | Shopify | Product & collection mapping |
Eleventy Sitemap | Eleventy | CLI integration, custom URL patterns |
Each of these tools handles multilingual sitemaps and detailed logs, so you can spot excluded URLs right away.
Scheduling Regular Updates
A sitemap only helps when it’s current. Plugins let you choose an interval—hourly, every few hours, or daily—so new pages appear immediately in your XML.
- Open your plugin’s settings and find the sitemap panel.
- Pick an update cadence that matches your publishing rhythm.
- Save and check the log or notification to confirm a successful run.
If you use Feather, you can hook into webhooks to trigger custom alerts every time your sitemap rebuilds.
Free Online Sitemap Generators
For smaller projects or microsites, you don’t need to install anything. Try these web-based tools:
- XML-sitemaps.com: Paste your URL and download a ready XML file.
- SmallSEOTools: Exclude staging or test URLs before export.
- Screaming Frog: Crawl up to 500 URLs for free and export XML.
Most free services cap out around 500 URLs per session, making them ideal for landing-page clusters or small blogs.
In one case, a boutique marketing firm generated sitemaps for 50 landing pages in under five minutes using XML-sitemaps.com.
Example Workflow
Here’s how I automate sitemap updates with Eleventy and GitHub Actions—no YAML code blocks needed:
- Configure a GitHub Action to run on every push.
- Install dependencies and build your site.
- Execute the Eleventy Sitemap Plugin as part of that build.
- Upload or deploy the resulting
sitemap.xmlto your public branch.
- Send a Slack notification when the process finishes.
This hands-off setup means every merge automatically publishes an up-to-date sitemap.
Best Practice Tips
- Reference your sitemap in
robots.txtso crawlers can find it instantly.
- Compress your XML with gzip to speed up fetch times.
- Validate your sitemap regularly—catch syntax errors before search engines do.
Choosing The Right Option
Your choice ultimately comes down to site size, update frequency, and platform comfort. If you need an in-depth breakdown of CMS features, check out our guide on content management systems comparison.
Teams with CI/CD pipelines often opt for static-site plugins like Eleventy’s. WordPress users usually appreciate Yoast’s intuitive controls and real-time preview. Pick the solution that matches your workflow, and your sitemap will mirror your site structure flawlessly.
Verifying Submission And Monitoring
Adding your sitemap to the server is just the beginning. You also need to submit it to search engines and keep an eye on its health to ensure every page gets crawled.
Most of us rely on two big dashboards for this:
- Google Search Console for detailed coverage reports and error logs
- Bing Webmaster Tools for crawl activity, sitemap status, and importer issues
Don't forget to peek at your
robots.txt to confirm the sitemap path is reachable.Submission In Search Console
After you register your domain, head over to the Sitemaps section. Paste in your sitemap URL, hit Submit, and watch the feed update. Within minutes, you’ll see how many URLs Google discovered versus indexed—and spot any glaring problems like 404s or blocked links.

Next, dive into the Index Coverage report. Look for
- Redirect loops
- Server errors
- Soft 404s
Squash those before they dent your rankings.
Monitoring With CLI Validators
Before anything goes live, validate your XML locally. A quick command like
xmllint --noout sitemap.xml will flag missing tags and syntax mistakes. Tools such as Sitemap Validator do the same, catching <loc> or <urlset> issues in real time.Fix any errors, rerun your check, and aim for zero warnings. Once live, pair these local tests with Search Console’s coverage reports to catch crawl hiccups. Look for patterns in server responses and XML parse errors.
Automating Sitemap Checks
Your CI pipeline can take over routine checks. A solid workflow might include:
- Installing a sitemap linter in your build environment
- Running the linter right after the sitemap is generated
- Failing the build if any errors turn up
- Allowing deployments only after successful validation
Step | Purpose |
Linting | Validates XML syntax |
Coverage Report | Compares indexed vs submitted URLs |
Alerting | Sends notifications on failures |
On top of that, schedule periodic resubmissions so search engines pick up on fresh
<lastmod> dates automatically. With this in place, you’ll always know how to generate a sitemap and keep it rock solid.Learn more about integrating performance tracking in our guide on adding Google Analytics to your site.
Verifying Robots.txt Entry
A misplaced or missing
robots.txt directive can hide your sitemap entirely. Make sure you have a line like:Sitemap: https://example.com/sitemap.xml
- Put it before any
User-agentrules so crawlers see it first
- Ensure no
Disallowdirectives block your path
- Test it using Google’s robots.txt tester to catch conflicts
Leveraging Bing Reports
While Google tends to dominate, Bing Webmaster Tools offers its own perspective. The Sitemap view shows you the last crawl date, submission timestamp, and any importer errors.
Set your team routine to:
- Review Bing’s error counts
- Resubmit after fixing issues
- Cross-check with Google’s coverage data for gaps
Metric | Google Report | Bing Report |
Last Submission | Displays submit date | Shows crawl timestamp |
Error Count | Highlights XML issues | Lists import failures |
URL Discovery | Indexed vs submitted | Submitted vs processed |
Running both reports side by side helps you spot platform-specific crawl quirks—and resolve them fast.
This hands-on approach to submission and monitoring will keep your sitemap in tip-top shape and your indexing rates high.
Best Practices And Feather Sitemap Settings
On sprawling sites, hitting the single-sitemap limit is almost guaranteed. Feather solves this by splitting your sitemap into chunks that each stay under the 50,000-URL threshold, so search bots breeze through without hiccups.
Instead of one giant XML file, a sitemap index points crawlers to several smaller maps. That keeps parse errors low and speeds up the discovery of fresh content.
Sitemap Index Files And Limits
Sitemap index files list individual sitemap segments, each capped at 50,000 URLs or 50MB uncompressed. For example, an online store might break its site into product, blog, and category sitemaps.
Smaller files fetch faster and dramatically reduce XML parsing failures. Crawlers spend less time retrying and more time finding new links.
When you combine canonical URLs with accurate
<lastmod> timestamps, you can see a 10–40% bump in indexed pages—and higher organic traffic within weeks. Explore the data in detail at Digital Silk.Lastmod And Priorities
A well-placed
<lastmod> tag tells search engines exactly when a page changed, pushing them to re-crawl. Fresh timestamps on key pages cut down indexing delays.Feather also lets you set
<changefreq> and <priority> per route. That means high-value content gets revisited more often, while archive or low-traffic sections are polled less frequently.Feather Automatic Sitemap Generation
Feather’s dashboard features a simple toggle for real-time sitemap updates whenever content changes.
- Enable or disable auto-update sitewide with one click
- Exclude pages by tag or URL pattern to weed out low-value content
- Adjust cron intervals using a visual scheduler
You can carve out drafts, private entries, or specific directories so only the URLs you care about make it into your public sitemap. Plus, you can assign different update frequencies—weekly for static pages, hourly for dynamic feeds.
Screenshot Of Feather Settings

This view shows toggles for auto-generation, content filters, and schedule settings. You can tweak
<lastmod> values and update intervals without touching code.Custom Routes And Coverage
When you need to capture every endpoint—legacy redirects, headless CMS paths, or micro-frontend URLs—Feather has you covered.
- Manually add static redirects or legacy URLs
- Pull external API endpoints into your XML feed
- Merge secondary sitemaps for isolated subdomains
Make sure your
robots.txt always references the correct sitemap index. Feather updates the Sitemap: line automatically whenever you rebuild. After any major change, don’t forget to test your sitemap URL in Google Search Console.Page Type | Changefreq | Priority |
Static Pages | weekly | 0.5 |
Blog Posts | daily | 0.7 |
Product Listings | hourly | 0.8 |
Use this table to match content types with realistic update cadences. Adjust priorities based on user engagement and business goals.
For deeper technical SEO insights and step-by-step tips, check out Feather’s technical SEO guide: Feather’s technical SEO guide.
Regularly review sitemap reports to catch missing URLs or crawl errors early. Automate these checks as your site grows to keep indexing on track.
Common Troubleshooting And Optimization Tips
Broken links are the silent killer of sitemaps. A quick audit to catch 404s before search bots arrive keeps your URL list tidy and crawler-friendly.
Outdated
<lastmod> timestamps can mislead engines about content freshness. Only update the date when you’ve made meaningful edits—this small discipline can noticeably boost recrawl rates.It’s all too easy to accidentally block your own sitemap with a stray robots.txt rule. One misplaced Disallow directive, and crawlers never see your files. Always double-check those rules on each deployment.
- Confirm sitemap paths in robots.txt.
- Test each URL against live pages.
- Remove any conflicting Disallow entries.
Duplicate URLs also throw off index coverage. A simple dedupe script that spots trailing-slash or case-sensitive variants can clear up messy lists fast.
XML Syntax Fixes
A single unclosed tag or wrong namespace can invalidate the entire file. If a browser flags a missing
</urlset>, search engines will too. That’s why you need a local XML linter in your toolkit.Tools like
xmllint will point out mismatches on the spot. Hook it into your CI pipeline so malformed XML never reaches production.Large sitemaps over 50,000 URLs often time out or get ignored. Splitting your file into topic-based chunks and tying them together with a sitemap index keeps things snappy.
Issue | Quick Fix |
Broken Tags | Run xmllint --noout sitemap.xml |
Missing Namespace | Add xmlns attribute to <urlset> |
Oversized File | Split into multiple sitemaps |
Optimizing Crawl Behavior
You don’t have to specify
<changefreq> and <priority>, but they’re powerful hints. Point crawlers at your freshest content and dial down low-value pages to protect your crawl budget.- Daily
changefreqfor blogs and news posts
- Hourly for product feeds in e-commerce
- Priority values between 0.5 and 0.9 based on traffic
In one project, marking key landing pages at 0.8 priority bumped their crawl rate by 30%—a tiny tweak with a huge upside.
Don’t forget gzip compression. Shrinking your sitemap by roughly 70% with
.xml.gz speeds up fetches. Just rename to sitemap.xml.gz and update your robots.txt.Scheduling Updates
Hand-editing sitemaps is a fast track to stale data. Instead, automate the build and deploy cycle with CI or cron jobs. This way,
<lastmod> always mirrors reality.- Schedule jobs during off-peak hours
- Generate via CLI or a trusted plugin
- Validate XML, compress output, then deploy
- Refresh robots.txt with the updated sitemap link
An automated midnight rebuild cut stale entries by 95% over three months and kept index coverage rock-solid.
Keep an eye on server logs for sitemap fetch status—200s vs 404s. Tie in Slack or email alerts for failures so you can jump on issues instantly.
Measuring Sitemap Health
Server logs are your window into crawler behavior. Tracking daily fetch counts and status codes helps you spot anomalies early and adjust update frequencies.
- Monitor 200 vs 403 vs 500 responses
- Identify URLs with repeated errors
- Analyze fetch timing to catch timeouts
Use Google Search Console’s URL Inspection API to pull report data automatically. Pipe the results into a weekly spreadsheet and watch for trends.
Consistent health checks keep your sitemap trustworthy and efficient.
Frequently Asked Sitemap Questions
Most teams are caught off guard by sitemap limits. An XML file tops out at 50,000 URLs and 50MB uncompressed. Once you hit either ceiling, it’s time to split your sitemap into chunks and stitch them together with an index file.
URL Count And File Size Limits
- Max URLs per file: 50,000
- Max uncompressed size: 50MB
- Use a sitemap index to reference each segment
Breaking your sitemap apart this way keeps crawling smooth and eliminates parse errors.
Updating sitemaps manually can easily lead to missed pages. Instead, log new content automatically and trigger your sitemap build with a CLI script or CMS webhook:
- Track freshly published pages in a change log
- Kick off an XML generation step after each publish
- Validate the sitemap and push it live
Comparing XML And HTML Sitemaps
XML sitemaps speak crawler’s language with tags like and . HTML versions, on the other hand, guide your visitors through your site—but won’t send crawl hints.
When To Choose XML
- Machine-readable metadata for precise indexing
- Full support within Google Search Console
When To Choose HTML
- Better user experience and site overview
- No extra tools required
For small, static sites under 100 pages, a hand-curated XML file works fine. Larger operations publishing daily will appreciate automated feeds.
Choosing Automated Or Manual Updates
Update Method | Best Use Case |
Manual | Under 100 pages |
Automated | Hundreds of URLs per day |
Automating sitemap generation frees your team to focus on content strategy, not file maintenance.
Don’t forget to tie performance checks into your analytics stack:
- Scan your crawl logs weekly for unexpected errors
- Set up alerts for spikes in 404s or broken links
Feather makes sitemap automation painless. Try it now at Feather
