Page 1 of 1

Most common problems found by web crawlers:

Posted: Sun Dec 22, 2024 5:51 am
by Md5656se
14.- Nofollow attributes in outgoing internal links
Internal links that contain a nofollow attribute block any potential passing through your site.

In relation to this, Ana Belén Leíño Head of SEO at Valor de Ley thinks that:

When implementing a nofollow link on an internal link, our goal is to ensure that the page from which it comes does not pass its link juice to the destination page, causing the search engine crawler to not follow it. The paradox is that you do this action within your own “house” so you have to be, in my opinion, extremely cautious.
By implementing it you are telling the search engine robots not to follow the link even though the user’s view is not affected. Well, if several people enter your house and you show them a room with another door inside, and you let some of your guests (users) access to see what is inside and others (Google) you only show them the door but you block their way and do not let them in, don’t you think that you will raise some suspicion among those guests?
A difference between what is shown to the user and what is shown to the Google robot is one of the biggest problems with nofollow, especially if it is abused in a misunderstood “PageRank distribution” between the URLs of the website.
It was once a habit to put nofollow on multiple internal links, for example those included in the footer. When the same links were repeated on all the pages of the site, they were blocked in order not to let the coveted link juice escape, but that is not how it works. Nowadays, even nofollow links transmit authority. Any link always consumes PageRank, whether internal or external. I am in favor of not using them except in a few exceptions such as links to URLs with logins or private parts of the site. In this case, it is better to be a good host and let your guests flow freely through your house.

15.- Incorrect pages found in the sitemap.xml file
Your sitemap.xml file should not contain broken pages.

Check any redirect chains and non-canonical pages and make sure they return a 200 code.

16.- Sitemap.xml not found
Lack of sitemaps makes it harder for search engines to crawl, crawl and index your website pages.

17.- Sitemap.xml not specified in robots.txt
Without a link to your sitemap.xml in your robots.txt file, search engines will not be able to understand the structure of your website.

Other common traceability errors i philippines number list nclude:
Pages not crawled.

Broken internal images.

Broken internal links.

URLs containing underscores.

4xx errors.

Resources formatted as page links.

External resources blocked in robots.txt.

Nofollow attributes on outgoing external links.

Tracking blocking.

Pages with only one internal link.


Image



Orphaned pages in the sitemap.

Crawling pages with a depth of more than 3 links.

Temporary redirects.