Google Search Console – Using Relative URLs in

I've just noticed that Google Support "Consolidating duplicate URLs" says the following:

Use absolute paths instead of relative paths with rel = "canonical"
Link element.

Use this structure: https://www.example.com/dresses/green/greendresss.html
Not This structure: /dresses/green/greendress.html

I always just used the file name without extension, as in:

public/Recipes/Mushy_Peas.xhtml:   

The following was reported on the Google Webmaster reporting pages without obvious issues:

User declared canonical
https://rbutterworth.nfshost.com/Recipes/Mushy_Peas

Using the detached filename is much easier than typing the entire URL.

(Note that the page contains multiple URLs only with / without the ".html" suffix.)

Is there a reason to change this (and a few hundred others) to show the full URL?

SEO – Google can not retrieve a large sitemap with 50,000 URLs and is not rendered by browsers

My sitemap contains 50,000 URLs / 7.8 MB and the following URL syntax:




 https://www.ninjogos.com.br/resultados?pesquisa=vestido, maquiagem,   2019-10-03T17:12:01-03:00 
 1.00 


The problems are:

• The search console reports that the sitemap could not be read.

• Sitemap loading takes 1 hour and Chrome stops working.

Enter image description here

• In Firefox, the sitemap was downloaded in 1483 ms and fully loaded after 5 minutes.

Things that I have done without success:

• Disable GZip compression.

• Delete my .htaccess file;

• Create a test sitemap with 1 KB URLs and the same syntax, and send it to Search Console. However, the sitemap with 50 KB URLs still displays "" that no sitemap can be retrieved.

Enter image description here

• An attempt was made to directly check the URL, but an error occurred and you are asked to try again later while the 1KB URLs worked.

• An attempt was made to validate the sitemap on five different websites (YANDEX, ETC), and all worked without error / warning

Any light?

SEO – Problems with Biggers Sitemaps with 50K URLs

My sitemap contains 50,000 URLs / 7.8 MB and the following URL syntax:




 https://www.ninjogos.com.br/resultados?pesquisa=vestido, maquiagem,   2019-10-03T17:12:01-03:00 
 1.00 


The problems are:

• The search console reports that the sitemap could not be read.

• Sitemap loading takes 1 hour and Chrome stops working.

Enter image description here

• In Firefox, the sitemap was downloaded in 1483 ms and fully loaded after 5 minutes.

Things that I have done without success:

• Disable GZip compression.

• Delete my .htaccess file;

• Create a test sitemap with 1 KB URLs and the same syntax, and send it to Search Console. However, the sitemap with 50 KB URLs still displays "" that no sitemap can be retrieved.

Enter image description here

• An attempt was made to directly check the URL, but an error occurred and you are asked to try again later while the 1KB URLs worked.

• An attempt was made to validate the sitemap on five different websites (YANDEX, ETC), and all worked without error / warning

Any light?

How do I filter out URLs that already have a website URL?

Here is the situation. I created a comment poster on blogging platforms and received 30,000 successful posts before Windows restarted and ended the campaign. Only 20% were done. So my question is, how can I avoid posting again to the same 30,000 URLs when I'm running another campaign? I could not export the successfully posted URLs. Is there a feature that scans the newly scrapped URLs and removes those that my sites are already linked to? I want to avoid double bookings.