What are some common errors reported in Google Search Console under the "Coverage" report, and how do you fix them?
In Google Search Console, the Coverage report helps website owners and SEO professionals monitor how well their website is indexed by Google and identify issues that may prevent proper indexing. Here are some of the most common errors reported in the Coverage report, along with suggestions on how to fix them:
1. Submitted URL is not valid
-
Description: This error indicates that a URL was submitted to Google (typically through a sitemap), but Google couldn't process it properly.
-
Cause: The URL might be malformed, contain unsupported characters, or be blocked by robots.txt.
-
Fix:
-
Check the URL for typos or issues in formatting.
-
Ensure the URL is not blocked by your robots.txt file or set to "noindex."
-
Confirm the page is accessible and returns a 200 OK status code.
-
2. Submitted URL has crawl issue
-
Description: Google attempted to crawl the URL, but encountered an error, such as a timeout or server issue.
-
Cause: This could be caused by server problems, slow response times, or temporary unavailability.
-
Fix:
-
Check your server's performance to ensure it's not slow or experiencing downtime.
-
Review server logs to diagnose any timeouts or errors.
-
Ensure the page loads correctly for both users and bots.
-
3. 404 Not Found (Page Not Found)
-
Description: The page has been deleted, moved, or is unavailable.
-
Cause: The URL no longer exists or was never created in the first place.
-
Fix:
-
Ensure that the page still exists, or if it was moved, set up a 301 redirect to the correct page.
-
For pages that are intentionally removed, update your sitemap to remove the invalid URL.
-
4. Soft 404 (Page Not Found)
-
Description: The page appears to return a "200 OK" status but has content that indicates it's a 404 (e.g., a page with a "not found" message).
-
Cause: The server might return a 200 status, but the content is not meaningful or relevant (like a blank page or "not found" text).
-
Fix:
-
Ensure the page returns a correct 404 or 410 status code if the content is truly missing or deleted.
-
If the page should be kept, update the content to make it useful and relevant.
-
5. Redirect Error
-
Description: The URL redirects to another URL, but the redirection is broken or leads to a non-canonical destination.
-
Cause: Misconfigured redirects or incorrect redirect loops.
-
Fix:
-
Check and update redirects to ensure they are working properly (using 301 redirects).
-
Ensure that there are no redirect loops or chains.
-
6. Blocked by robots.txt
-
Description: The page is being blocked from crawling by the robots.txt file, which is preventing Googlebot from accessing it.
-
Cause: The robots.txt file may have been configured to disallow Googlebot from crawling certain parts of your site.
-
Fix:
-
Review your robots.txt file and ensure that it doesn't block important pages.
-
If necessary, update the file to allow access to the relevant URLs.
-
7. Crawl anomaly
-
Description: Googlebot encountered an unusual situation or error when trying to crawl the URL, but the exact cause isn’t clear.
-
Cause: This could be due to intermittent server issues, access problems, or other unexpected behavior during crawling.
-
Fix:
-
Check for server issues like slow response times, timeouts, or other anomalies.
-
If it’s a temporary issue, it might resolve itself over time, but keep monitoring the URL in Google Search Console.
-
8. Noindex (Page blocked by noindex directive)
-
Description: The page has a noindex directive, which tells Google not to index the page.
-
Cause: The page may have a meta tag or HTTP header that instructs Google not to index it.
-
Fix:
-
If the page should be indexed, remove the noindex directive from the page's HTML or HTTP header.
-
If the page should not be indexed, no action is needed, but ensure you’re not accidentally blocking important pages.
-
9. Server errors (5xx)
-
Description: Googlebot encounters a server-side error when attempting to access the page, such as a 500 Internal Server Error or 503 Service Unavailable.
-
Cause: These errors are usually due to server misconfigurations or temporary downtime.
-
Fix:
-
Check server logs to identify the specific error and resolve it.
-
Ensure the server is online and accessible for Googlebot.
-
If the error was temporary, Google will attempt to recrawl the page after a while.
-
10. Duplicate without canonical
-
Description: Google has detected duplicate content but has not identified the preferred (canonical) version.
-
Cause: Two or more pages have similar content, but one is not marked as the canonical version.
-
Fix:
-
Add a rel=canonical link tag to the duplicate page that points to the preferred version.
-
Ensure there are no other issues with duplicate content (e.g., the same content appearing under different URLs).
-
11. Invalid URL
-
Description: Google detected an invalid URL in your sitemap or internal links, such as URLs with unsupported characters.
-
Cause: URLs with spaces, special characters, or incorrect formats that Google cannot process.
-
Fix:
-
Ensure all URLs follow proper syntax and avoid using unsupported characters.
-
Update your sitemap and internal links accordingly.
-
General Fixes and Best Practices:
-
Regularly Check Your Coverage Report: The Coverage report is an ongoing source of insights into how Google sees your site. Regularly review and fix issues to ensure proper indexing.
-
Fix Errors in Sitemap: Ensure your sitemap is up-to-date and contains only valid URLs that are accessible and properly configured.
-
Monitor Performance: Use Google Search Console's performance metrics to assess the impact of any errors and monitor how your site is being indexed over time.
By addressing these errors, you can improve your site's visibility and ensure Google indexes your pages correctly.