SEO

Sitemap

What is a Sitemap? XML file to communicate website structure to search engine crawlers.

Sitemap is a structured file (usually in XML format) that tells search engines all the URLs of a website so they can crawl and index them. A sitemap is not for users, but for search engine crawlers. It functions like a map: "Here are all the important places on my website, please visit these." In technical SEO, a well-maintained sitemap is essential for large websites and signals to Google which URLs are important.

Most small to medium websites don't strictly need a sitemap - Google finds URLs through internal links. But for complex websites, large catalogs (e-commerce), new websites, or websites with many JavaScript-based pages, a sitemap is critical.

What is a Sitemap?

A sitemap is an XML file, typically at /sitemap.xml, that lists all URLs of a website with additional metadata:

Example Sitemap Struktur:

"<?xml version="1.0" encoding="UTF-8"?> <urlset xmlns="http://www.sitemaps.org/schemas/sitemap/0.9"> <url> <loc>https://leadanic.com/</loc> <lastmod>2026-03-04</lastmod> <changefreq>weekly</changefreq> <priority>1.0</priority> </url> <url> <loc>https://leadanic.com/lexikon/marketing-automation/</loc> <lastmod>2026-02-28</lastmod> <changefreq>monthly</changefreq> <priority>0.8</priority> </url> </urlset>"

Sitemap attributes:

  • loc: The complete URL of the page
  • lastmod: When was this page last updated? (Optional, but recommended)
  • changefreq: How often does this page change? ("always", "hourly", "daily", "weekly", "monthly", "yearly", "never")
  • priority: Relative priority (0.0 to 1.0). Default is 0.5

Types of sitemaps:

  • XML sitemap: For search engines, machine-readable
  • HTML sitemap: For users, a page index as HTML
  • Video sitemap: Specialized sitemap for videos
  • News sitemap: Specialized sitemap for news sites
  • Image sitemap: Specialized sitemap for images

Most focus is on XML sitemaps.

Sitemap in a B2B Context

B2B websites need special sitemap considerations:

Dynamic URLs and parameters

Many B2B pages generate URLs dynamically (e.g., filters in product catalogs):

/produkte?kategorie=marketing-automation&budgetrang=10k-50k&industrie=technologie

Sitemaps should not contain every URL variant (would become too large), but rather the canonical, most important version:

/produkte/marketing-automation

Gated content and authentication**

URLs behind login (e.g., "/my-account/", "/dashboard/") should typically not be in the sitemap, as Google cannot crawl these.

Large catalogs (> 50,000 URLs)**

Google accepts sitemaps up to 50,000 URLs and 50 MB. For larger catalogs, you need a sitemap index:

"<sitemapindex xmlns="http://www.sitemaps.org/schemas/sitemap/0.9"> <sitemap> <loc>https://leadanic.com/sitemap-products-1.xml</loc> </sitemap> <sitemap> <loc>https://leadanic.com/sitemap-products-2.xml</loc> </sitemap> </sitemapindex>"

This creates multiple sitemaps (e.g., "sitemap-products-1.xml", "sitemap-products-2.xml") and an index.

Creating a sitemap**

Option 1: Automatically via CMS**

Most modern CMS platforms (WordPress, Webflow, HubSpot) generate sitemaps automatically:

  • WordPress: Yoast SEO Plugin or ein anderes generiert automatisch
  • Webflow: Automatically generated on each publish
  • HubSpot: CMS generates automatically

This is often the best option, as it is updated, when new pages are added.

Option 2: XML sitemap generator (online tools)**

Tools like:

  • XML-Sitemaps.com (free, up to 500 URLs)
  • Screaming Frog SEO Spider (Paid, unlimited)
  • Sitemap Generator Tools

These crawl the website and generate a sitemap XML file.

Option 3: Manual or programmatic**

For larger catalogs: generate die Sitemap programmatically from a database:

  • Database query: SELECT url, lastupdate FROM pages
  • Generate XML file based on data
  • Upload file to /sitemap.xml
  • Reference in robots.txt

Sitemap best practices**

1. Reference in robots.txt**

In your /robots.txt, the sitemap should the sitemap should be referenced:

"Sitemap: https://leadanic.com/sitemap.xml"

This signals to Google, where the sitemap is.

2. Include only crawlable URLs**

Don't include URLs, that are behind login, are 404s, or block user experience:

  • URLs that are not publicly accessible
  • Pages with noindex tag
  • Duplicate/canonical URLs (only the canonical)

3. Update lastmod**

The lastmod date should be updated, when a page is actually updated. Google uses this, to decide how often to re-crawl:

  • Page was updated: update lastmod
  • Only typo fixes: don't update lastmod (not necessary)

4. Use changefreq thoughtfully**

changefreq is a hint, not a guarantee. Google often ignores it and uses historical data instead.

  • Homepage: "weekly" or "daily" (frequent updates)
  • Blog articles: "monthly" or "yearly" (not often updated after publishing)
  • Old archives: "never" (not updated)

5. Set priority realistically**

Priority tells Google, which pages are relatively more important:

  • Homepage: 1.0 (most important)
  • Top-level service pages: 0.8 - 0.9
  • Blog articles: 0.6 - 0.7
  • Archive/old content: 0.3 - 0.5

Don't set all pages to 1.0 - that doesn't mean anything.

6. Test sitemap in Google Search Console**

After adding a new sitemap:

  1. Go to Google Search Console
  2. Go to "Sitemaps"
  3. Enter your sitemap URL
  4. Google validates and shows status

Search Console will show:

  • How many URLs in the sitemap
  • How many of them are indexed
  • Any errors

Sitemap errors and problems**

URLs in sitemap but not indexed**

If Google sees the sitemap, but pages are not indexed, this could be because:

  • robots.txt blockiert die URL
  • noindex Tag on der Seite
  • Seite ist canonical to einer anderen URL
  • Page Speed sehr langsam
  • Seite hat wenig Backlinks / Authority

Zu groe Sitemap (> 50MB or > 50.000 URLs)**

Google akzeptiert nicht mehr. Lsung:

  • Mehrere kleinere Sitemaps erstellen (Sitemap Index)
  • Nicht-wichtige URLs entfernen (e.g., Archive)
  • Paginated URLs zusammenfassen (nur erste Seite)

Veraltete Sitemap**

Eine Sitemap, die nicht aktualisiert wird, wird to einer SEO-Liability:

  • Tote Links in the Sitemap
  • Neue Seiten nicht eingefgt
  • Inaccurate lastmod

Lsung: Automatisieren (CMS or Programmatic), nicht manuell pflegen.

Sitemap Gre and Limits**

Limit Wert Was to tun, wenn berschritten
Max URLs pro Sitemap 50.000 Sitemap Index + mehrere Sitemaps
Max Dateisize 50 MB Sitemap Index + mehrere Sitemaps
Max Sitemaps in Index Keine harte Limit, aber 1.000+ ist nicht ideal Seiten konsolidieren, alte Archive entfernen

Sitemap vs. Interne Verlinkung**

Eine hufige Frage: "Brauche ich eine Sitemap, wenn ich gute interne Links habe?"

Antwort: Fr kleine Websites (< 1.000 URLs) with guter interner Verlinkung ist eine Sitemap optional. Fr grere or komplexe Websites ist Sitemap + gute interne Links die beste combination.

Sitemap and interne Links sind complementary, nicht exclusive.

Eine gut gepflegte Sitemap ist kostenlos and hilft Google, Ihre Website to crawlen and to indexieren. Besonders for grere Websites or solche with vielen neuen Inhalten sollte eine Sitemap Standard sein. Fr kleine Websites ist es weniger kritisch, aber es schadet nicht, eine to haben.

Sounds like a topic for you?

We analyze your situation and show concrete improvement potential. The consultation is free and non-binding.

Book Free Consultation