A well-optimized website is crucial for SEO success, and technical issues can severely impact search engine rankings if left unchecked. Semrush Site Audit is a powerful tool that helps website owners identify critical errors affecting site performance, crawlability, and search visibility. These errors—from broken links and missing metadata to server and security issues—are the most severe problems detected in a site audit.
Unlike minor technical improvements, errors indicate fundamental flaws that can prevent search engines from properly indexing and ranking a website. If these issues persist, they can lead to lower search rankings, decreased organic traffic, and even a loss of credibility with search engines.
Fixing errors promptly is essential for maintaining a healthy, high-performing website. Search engines prioritise sites that are accessible, well-structured, and free from critical issues. Pages with broken internal links, incorrect URL structures, or missing security protocols can hinder user experience and send negative signals to search engine crawlers.
Common Errors Identified in Semrush Site Audit (With Fixes)
Technical Errors Affecting Website Accessibility
Ensuring that a website is accessible to both users and search engine crawlers is fundamental for SEO success. When critical technical errors arise, they can block pages from being indexed, disrupt user experience, and weaken overall site authority. The Semrush Site Audit detects various technical errors affecting a website’s accessibility.
Pages Returning 5XX Status Codes (Server Errors)
Definition and Impact on Website Availability
5XX status codes indicate server-side errors that prevent pages from being accessed by users and search engines. These errors occur when a server fails to fulfill a request, causing temporary or permanent downtime. If search engine bots frequently encounter 5XX errors, they may reduce the crawl frequency of your site, leading to indexing issues and a drop in rankings.
Common Causes
- Server overload: High traffic spikes or insufficient server resources
- Misconfiguration: Issues with server settings, database connections, or incorrect .htaccess rules
- Hosting problems: Poor hosting infrastructure, expired hosting plans, or maintenance downtime
Solutions
- Optimise server performance: Upgrade hosting plans if traffic exceeds current capacity.
- Check server logs: Identify patterns of failure and resolve configuration issues.
- Use a CDN (Content Delivery Network): Distribute server load for better uptime.
- Monitor uptime with tools: Services like UptimeRobot or Pingdom can track downtime and alert issues in real-time.
Pages Returning 4XX Status Codes (Client Errors)
Explanation of 404, 403, and 410 Errors
4XX errors indicate client-side issues where a webpage cannot be accessed due to incorrect requests or missing content. The most common status code types include:
- 404 Not Found: The page no longer exists or has been moved without proper redirection.
- 403 Forbidden: The server refuses access due to permissions or security settings.
- 410 Gone: The page has been permanently deleted with no forwarding address.
Causes
- Deleted or moved pages without redirects
- Incorrect URL entry or broken internal links
- Restricted access to pages due to incorrect permissions
Fixes
- Set up 301 redirects for permanently moved pages to direct users and crawlers to the correct URLs
- Update internal links to avoid broken navigation paths
- Check file permissions and ensure proper access settings for necessary pages
Broken Internal Links
How They Negatively Impact SEO and User Experience
Internal links are essential for website navigation and search engine crawling. Broken internal links lead to 404 errors, disrupt user journeys, and reduce the authority flow between pages, ultimately hurting rankings.
Tools to Detect Broken Links
- Semrush Site Audit (Detects broken links and provides status codes)
- Screaming Frog SEO Spider (Crawls websites to identify link errors)
- Ahrefs Site Audit (Highlights broken links and their referring pages)
Fixing Broken Links
- Use proper redirects (301) when content moves to a new location
- Regularly audit internal links and update outdated URLs
- Ensure proper linking structure to avoid orphaned pages (pages with no internal links pointing to them)
Broken Internal Images
Causes of Broken Images
- Deleted or moved images without updating the file path
- Incorrect image URLs in HTML or CSS
- CDN issues or external image hosting downtime
Solutions
- Check image URLs to ensure they point to the correct location
- Re-upload missing images and update file paths where necessary
- Use WebP format for optimised, faster-loading images
- Implement lazy loading to improve performance and prevent unnecessary image loading delays
Pages Not Crawled Due to DNS Issues
What DNS Resolution Errors Mean
Domain Name System (DNS) issues prevent search engines from resolving your domain, making your website inaccessible to users and crawlers. This can significantly impact indexing and rankings.
How to Diagnose Issues
- Use Google Search Console: Check for crawl errors related to DNS resolution.
- Run a DNS lookup tool like MXToolBox to verify domain settings
- Check Semrush Site Audit for DNS-related issues
Fixes
- Contact your hosting provider to resolve misconfigured DNS settings
- Verify DNS records (A Record, CNAME, MX, and TXT settings)
- Ensure your domain isn’t expired or misconfigured at the registrar
Pages with Incorrect URL Formats
Examples of Incorrect URL Structures
- Dynamic URLs with excessive parameters (e.g., example.com/page?id=123&ref=source)
- Mixed uppercase and lowercase characters (Causing duplicate content issues)
- Long, unreadable URLs with unnecessary words (e.g., example.com/buy-best-products-category-online-2025)
Impact on Search Engine Indexing
- Confuses search engines and may result in duplicate content indexing
- Reduces user experience and trust in the link
- Can cause keyword cannibalisation if multiple URLs serve similar content
Best Practices for SEO-Friendly URLs
- Use clean, short URLs (e.g., example.com/best-products)
- Keep URLs lowercase to avoid case-sensitive duplicate content issues
- Avoid unnecessary parameters and session IDs in URLs
- Use hyphens (–) instead of underscores (_) for better readability
On-Page SEO Errors
On-page SEO elements play a crucial role in helping search engines understand a webpage’s content and relevance. Errors in these areas, such as duplicate or missing metadata, improper content structuring, and indexing issues, can negatively impact search rankings and user experience.
Duplicate Title Tags
Why Duplicate Titles Confuse Search Engines
Title tags are one of the most important on-page SEO elements, as they help search engines understand a page’s topic. When multiple pages have the same title tag, it creates confusion for search engines, making it harder for them to determine which page is the most relevant for a search query. This can lead to:
- Keyword cannibalisation, where search engines struggle to prioritise the correct page
- Decreased rankings due to diluted SEO value across multiple pages
- Poor click-through rates (CTR), as identical titles in search results reduce user engagement
Fix: Implementing Unique, Keyword-Optimized Title Tags
- Ensure each page has a unique title that accurately represents its content
- Use primary keywords naturally while maintaining readability (e.g., “Best Laptops for Students | TechReviews 2025”)
- Keep title length within 50-60 characters to avoid truncation in search results
- Use dynamic title tags for category and product pages to differentiate them (e.g., “Running Shoes – Nike vs. Adidas | Sports Gear Review”)
Missing Title Tags
How Missing Titles Reduce Search Visibility
Title tags are a ranking factor influencing a webpage’s CTR in search engine results. Search engines often generate a default title without a title tag, which may not be relevant or optimised for search intent. This can lead to:
- Lower search visibility, as missing titles weaken the page’s ranking potential.
- Reduced user engagement, since users may see incomplete or auto-generated titles.
- Poor site structure, making it harder for crawlers to understand the purpose of a page.
Best Practices for Writing Compelling SEO Titles
- Include target keywords naturally at the beginning of the title
- Use action words or engaging phrases to improve CTR (e.g., “Learn SEO in 30 Days – A Complete Guide”)
- Align with user intent, ensuring titles match what people are searching for
- Avoid keyword stuffing, as unnatural titles can get penalised by search engines
Duplicate Content Issues
Causes of Duplicate Content
Duplicate content occurs when the same or very similar content appears on multiple URLs, either within the same website or across different sites. This can cause indexing confusion, where search engines struggle to determine the source of the content. Common causes include:
- Content scraping: Unauthorized duplication of content across multiple domains
- Poor site architecture: Multiple URLs serving identical content (e.g., /product and /product?id=123)
- Missing canonical tags: Search engines treating variations of the same content as separate pages
Fixes: Using Canonical Tags and Rewriting Duplicate Content
- Implement canonical tags (rel=”canonical”) to tell search engines which version of a page to index
- Rewrite similar content to ensure uniqueness and provide additional value to users
- Use 301 redirects to consolidate duplicate URLs into a single authoritative version
- Prevent duplicate product pages on e-commerce sites by ensuring proper URL structures and filtering parameters
Missing Meta Descriptions
Why Meta Descriptions Impact Click-Through Rate (CTR)
A meta description summarises a webpage that appears in search engine results. While it is not a direct ranking factor, it plays a crucial role in influencing CTR. When a page is missing a meta description:
- Search engines generate one automatically, which may not be compelling or relevant
- Users are less likely to click, reducing organic traffic
- It can weaken brand messaging, as the page preview lacks control over what users see
Best Practices for Writing Compelling Meta Descriptions
- Keep meta descriptions between 150-160 characters to prevent truncation
- Incorporate primary keywords naturally without stuffing
- Use a call-to-action (CTA) to encourage clicks (e.g., “Discover the top 10 laptops for 2025 – Read our expert review now!”)
- Ensure descriptions are unique across all pages to avoid duplicate content issues
Duplicate Meta Descriptions
The Risk of Multiple Pages Having Identical Descriptions
Like duplicate title tags, duplicate meta descriptions can confuse search engines and negatively impact CTR. If multiple pages share the same description, search engines may:
- Randomly select a description or generate one from page content, which may not be optimised.
- Reduce user engagement, as repetitive descriptions across different pages fail to highlight unique content.
- Cause keyword dilution, making it harder for individual pages to stand out in search results.
How to Create Unique Descriptions for Better Search Visibility
- Write distinct descriptions that summarise the content of each page.
- Avoid generic phrases, instead using specific value-driven statements for each page.
- Use dynamic descriptions for large websites by incorporating unique page elements (e.g., “Shop Nike Running Shoes – Lightweight, Durable, and Built for Speed”).
- Regularly audit and update descriptions to align with search trends and user intent.
Sitemap and Indexing Issues
A well-structured sitemap and a properly configured robots.txt file ensure that search engines efficiently crawl and index a website. Errors in these files can prevent important pages from appearing in search results, leading to lower rankings and reduced organic traffic.
Incorrect Pages in Sitemap.xml
How Incorrect URLs in Sitemap Affect Indexing
The sitemap.xml file is a roadmap for search engine crawlers, helping them discover and index important pages. However, if incorrect URLs are listed, it can lead to:
- Wasted crawl budget, as search engines attempt to index non-existent or low-priority pages
- Indexing of unnecessary pages, such as test pages, outdated content, or parameterised URLs
- 404 errors in search console, if URLs in the sitemap no longer exist
Fixing Sitemap Errors Using XML Validation Tools
- Use Semrush Site Audit to detect incorrect or outdated URLs in the sitemap
- Validate the sitemap with Google Search Console (under the Sitemaps section) to identify submission errors
- Ensure only indexable URLs are included, avoiding noindex, redirected, or broken pages
- Regularly update the sitemap to reflect the current site structure and content
Sitemap.xml Formatting Errors
Common Sitemap Issues That Block Crawling
A poorly formatted sitemap can prevent search engines from properly crawling a website, leading to indexation issues. Some common problems include:
- Incorrect XML syntax, causing parsing errors
- Non-canonical URLs, which confuse search engines about the preferred version of a page
- Too many URLs in a single sitemap file (Google recommends limiting it to 50,000 URLs or 50MB)
Fix: Correcting Structure and Updating Robots.txt
- Validate the sitemap using XML validation tools like XML Sitemap Validator
- Ensure correct formatting by following Google’s sitemap guidelines
- Break large sitemaps into multiple smaller files, and use a sitemap index file if needed
- Update the robots.txt file to reference the correct sitemap location using:
plaintext
Robots.txt Formatting Errors
Mistakes in Robots.txt That Block Important Pages
The robots.txt file controls which parts of a website search engine crawlers can access. Misconfigured rules can block essential pages from indexing, leading to severe SEO issues. Common mistakes include:
- Blocking the entire website (Disallow: / prevents search engines from crawling any page)
- Restricting important sections like category pages or blogs, preventing them from appearing in search results
- Conflicts with meta robots tags, where a page is allowed in robots.txt but set to noindex in the page’s metadata
Best Practices for Robots.txt to Control Crawler Behavior
- Check for accidental disallow rules using the robots.txt tester in Google Search Console
- Ensure critical pages are not blocked, such as the homepage, category pages, and service pages
- Allow important bots like Googlebot while restricting unnecessary crawlers to save server resources
- Use precise blocking rules, avoiding broad disallow commands like
Instead, specify directories that should not be crawled, such as:
- Ensure sitemap.xml is referenced correctly in the robots.txt file.
Mobile and Performance Errors
In today’s mobile-first world, Google prioritises websites that offer a fast, responsive, and mobile-friendly experience. Errors related to mobile responsiveness and performance can lead to higher bounce rates, slower load times, and lower search rankings.
Pages Without Viewport Tags
Why Viewport Tags Are Crucial for Mobile SEO
The meta viewport tag tells browsers how to scale and display a webpage on different screen sizes. Without it, pages may appear zoomed out, improperly formatted, or difficult to navigate on mobile devices. Issues caused by missing viewport tags include:
- Poor mobile user experience, as text and images do not resize correctly
- Google’s mobile-first indexing penalties, reducing rankings for mobile searches
- Higher bounce rates, since users struggle to interact with non-responsive pages
Fix: Adding Correct Meta Viewport Tags for Responsive Design
- Ensure all pages include a viewport meta tag in the <head> section:
- Use responsive design (CSS media queries) to adapt layouts for different screen sizes
- Test mobile-friendliness using Google’s Mobile-Friendly Test
- Avoid fixed-width elements that prevent proper scaling on smaller screens
Excessive HTML Size on Pages
How Bloated HTML Slows Down Websites
A large HTML file size increases page load time, hurting user experience and SEO rankings. Google favors fast-loading pages, and excessive HTML size can lead to:
- Longer server response times, delaying page rendering
- Lower Core Web Vitals scores, affecting user engagement metrics
- Higher bounce rates, as users abandon slow-loading sites
Fix: Minifying HTML, CSS, and JavaScript
- Use HTML minification tools to remove unnecessary spaces and comments:
- HTMLMinifier
- Compress CSS and JavaScript using tools like:
- CSS Minifier (cssnano)
- JS Minifier (UglifyJS)
- Enable Gzip or Brotli compression to reduce page file size
- Use asynchronous loading (async) for JavaScript files to prevent rendering delays
AMP Pages Missing Canonical Tags
Why AMP Pages Need Canonical Tags
Accelerated Mobile Pages (AMP) are lightweight webpages designed for faster loading on mobile devices. However, if AMP pages lack a canonical tag, search engines may:
- Treat AMP and non-AMP pages as duplicate content, leading to indexing issues
- Fail to link AMP pages to their original counterparts, reducing ranking authority
- Display incorrect page versions in search results, affecting user navigation
Fix: Proper Canonicalization to Avoid Duplicate Content
- Ensure each AMP page includes a canonical tag pointing to the original version.
- Verify AMP implementation using Google Search Console under the AMP Reports section.
- Test pages with Google’s AMP Validator (AMP Test)
Hreflang Conflicts in Source Code
The Role of Hreflang in Multilingual SEO
Hreflang is an HTML attribute used to indicate a webpage’s language and regional targeting. It ensures that:
- The correct language version of a page is shown to users in different regions
- Search engines avoid duplicate content penalties by recognising different language versions as intentional
- User experience improves, as visitors land on the most relevant content for their location
When hreflang conflicts occur, search engines may struggle to determine the correct page version, leading to incorrect indexing and ranking issues.
Fix: Ensuring Correct Language and Regional Targeting
- Each page should have a self-referencing hreflang tag to confirm its language version:
- Hreflang annotations must be bidirectional—if example.com/es links to example.com/en, then example.com/en must link back to example.com/es.
- Use correct language and country codes based on ISO 639-1 for languages and ISO 3166-1 for regions. Example:
- Avoid mixing incorrect language-country codes (e.g., using es-UK instead of es-ES for Spanish in Spain).
- Check for missing return tags, as incomplete hreflang implementations can cause errors.
Invalid Hreflang Links
Common Hreflang Mistakes (Pointing to Incorrect URLs)
Incorrect hreflang links can cause search engines to ignore language targeting altogether. Some common mistakes include:
- Pointing hreflang tags to non-existent pages (404 errors)
- Using relative URLs instead of absolute URLs
- Mismatched hreflang implementation between pages
- Referencing pages blocked by robots.txt or marked as noindex
Fix: Using Hreflang Validation Tools to Correct Issues
- Use Google Search Console’s International Targeting tool to detect hreflang errors
- Validate hreflang tags with Semrush Site Audit to identify incorrect or missing annotations
- Ensure all hreflang URLs are absolute, not relative (e.g., use https://example.com/es, not /es)
- Check if referenced pages are indexable and not blocked by robots.txt or noindex tags
- Use hreflang testing tools, such as:
Security and HTTPS Issues
Website security is critical for SEO and user trust. Google has explicitly stated that HTTPS is a ranking factor, meaning that sites without SSL certificates are disadvantaged in search results. Moreover, modern browsers warn users about insecure websites, which can significantly impact traffic and conversions.
Pages Not Using HTTPS
Why HTTPS Is a Ranking Factor
HTTPS (Hypertext Transfer Protocol Secure) encrypts data exchanged between users and the website, preventing security threats such as:
- Data interception (man-in-the-middle attacks) that compromise sensitive user information
- Phishing attacks, where malicious actors create fake versions of unsecured websites
- SEO penalties, as Google favors secure sites and may flag HTTP pages as “Not Secure”
Using HTTPS improves SEO and user trust by:
- Providing a ranking boost, as Google prioritises HTTPS pages over HTTP
- Ensuring secure transactions, especially for e-commerce and login-based sites
- Preventing browser security warnings, which can deter visitors from staying on a site
Fix: Installing SSL Certificates and Updating Mixed Content
- Install an SSL certificate from a trusted provider (e.g., Let’s Encrypt, Cloudflare, GoDaddy)
- Redirect all HTTP URLs to HTTPS using 301 redirects in the .htaccess file: