seo – Can I delete sitemap.xml from my website? Does that have a negative impact?

I'll post a controversial response that many probably will not agree with. Is sitemap.xml still necessary? It is no longer decisive, no. Is it good to have it and does it help? For sure. Should you delete it? I can not say, because I do not know what your above reasons are.

View this article written by my former colleague with years of SEO, UX, and Web Dev experience. I agree with many things:

https://www.imarc.com/blog/do-sites-really-need-a-sitemap-for-seo

If you do not actively update your Sitemap and do not submit to GSC or Bing Webmasters, you probably do not need them. A sitemap is a suggestion for search engines how to crawl your site. So think of websites like news sites with hundreds of thousands of pages: Sitemaps have size limits. For news sites, they do not contain every published article. However, these articles are found, indexed and classified.

If your website is updated regularly; has a clear information architecture and category hierarchy; and contains a well-planned internal link structure. The search engines index and arrange the content independently of the existence of the sitemap.

Note, however, that you will lose the ability to manually submit the Sitemap to GSC, which will trigger a crawl again. You also lose the ability to suggest search engines which parts of your website are most important or which ones are current.

Many SEOs still say that Sitemaps are critical, but that's traditional SEO wisdom. What matters is not how I would define it; it is more helpful, as I have observed in recent years. In your case, it depends on your reasons for deleting and how much you currently manage.

seo – Temporarily hosting a sitemap.xml file to retrieve URLs

I have a GZ sitemap file that contains 4 different sitemaps, each 50 MB in size. I want to get all the URLs in these 4 sitemap files. I would like to use https://robhammond.co/tools/xml-extract for this, but this tool requires inserting the URL of the sitemap. However, with sitemap.xml.gz, you can only download the XML files so that a locally saved file and no URL is displayed at the end (for example, example.com/sitemap1.xml)

How can I do that?
1) Host the 4 sitemap.xml files for the above tool
or
2) Extract the URLs from the downloaded sitemap files

robots.txt and sitemap.xml site.webmanifest

Hello im having touble to load files robots.txt and sitemap.xml site.webmanifest form https://example.com/robots.txt for example, since magento root directory is set under /pub through nginx.conf.sample (from Magento 2 installation). I can load files from http://example.com/pub/robots.txt but its not what i want. I want to be able to load files like https://example.com/robots.txt https://example.com/sitemap.xml etc. Any ide how to solve this?

sitemap.xml with spaces in line 1

I have a plugin that generates a Sitemap and creates a redirect for a multi-site installation. When the sitemap is created, it is stored in the upload directory of each blog: /wp-content/blogs.dir/1/sitemap.xml, but is available through /sitemap.xml

If you access the Sitemap from the actual location, it works as expected without adding an extra row. However, if you access it through the redirect, it adds an empty space to the first line of the sitemap, which causes an error.

I've looked into this problem in several support forums and they all pointed out that an extra line at the end of the functions.php (after?>) Theme could cause this problem. Alternatively, an extra line in wp-config.php, index.php or another plugin. I tested all these options without success.

A recommended solution was to use the Health Check plugin, which disables all plugins and changes the theme to the default. Then you can activate plugins individually to find the problem. This also offered no solution.

If all plugins are disabled and the default theme is installed, the sitemap redirect will still exist, and the extra row has been added.

So I started wondering if this was a core issue of WordPress and reinstalled all the core scripts, but to no avail.

I looked through the error logs and nothing.

Any consideration of where to go next to try to solve this problem would be greatly appreciated.

Many Thanks,
Beat

Where will robot.txt and sitemap.xml for WordPress be installed in a separate directory?

My site is set up to have WordPress installed in its own subdirectory, but the content is deployed as if it were in the domain root.
(This WordPress Codex page describes the configuration.)

WordPress URL: https://www.example.com/wordpress
Page URL: https://www.example.com

In this case URLs are NOT inserted into include / WordPress in the URL it serves pages of (though image URLs do lock in: https://www.example.com/wordpress/wp-content/uploads/ ...).

URL for a page is: www.example.com/mypage/
NOT: www.example.com/wordpress/mypage/
but image link is: www.example.com/wordpress/wp-content/uploads/2018/12/25/image1.jpg

Question:
Should I bet? robots.txt and sitemap.xml in the directory "webiste root" on my server (example.com)
or his WordPress Directory (example.com/wordpress)

I can not find this particular case in any search.
Even trial and error do not help, because I get conflicting results from various sitemap and robot validators.

In addition, it seems that a plugin or part of WordPress is creating a virtual sitemap (sitemap.xml.gz) and robots (robot.txt) Files, which further annoys my efforts definitely determine The place to put these files for this general, but non-standard setup.

magento 1.8 – Exclude disabled products from generating sitemap.xml

Some of your previous answers have not been well received and you run the risk of being blocked from answering.

Please note the following notes exactly:

  • Please be sure answer the question, Enter details and share your research!

But avoid

  • Ask for help, clarification or answering other answers.
  • Make statements based on opinions; secure them with references or personal experiences.

For more information, see our tips for writing great answers.

how to maintain the sitemap.xml automatically well?

Hello,

if my domain is domain.com,
I put my main page HTML pages under / public_html /,
and I install a blog at / public_html / blog /.

for my main page sometimes
I can update manually or upload new pages.

and my blog,
I can write a new post every day.

I think to mix them into a sitemap.xml and send them to the search engine.
I find the article https://stackoverflow.com/questions/…e-sitemap-file,
I may need a sitemap.xml to merge the two sitemap files as follows:


http://domain.com/sitemap1.xml
2012


http://domain.com/blog/sitemap2.xml
2016-10-11

but it takes the day ,
I would like to ask if a script or method can automatically update them well and update them into the main sitemap.
or is not important ?