indexing – How to test multiple Sitemap.xml files?

Is there a way to test multiple sitemap.xml files?
The validation works well, Google accepts all subfiles, but "Verification of the server response" in Yandex returns "The document does not contain any text".

By the rate of analysis and the overall progress of the indexing, I have the impression that both search engines fail to read the contents of the sitemap files. Because there is a large amount of "Discovered – Currently Unindexed" = 2/3 of all content, they have never been explored and due to the low indexation of ration in Yandex.

This website contains about 750,000 links in site map files. When I generate 50,000 links per file (about 11 MB), the crawl graph goes up, then falls. When there are 10,000 links per file, the graph drops much faster and stays at about the same level.

We have done various checks and, technically, everything seems to be going well, but if you look at the performance, it's pretty dubious.
Robots.txt gives full access. robots meta tags too.

  • Can anyone suggest a way to check why "Checking the server response" returns an error when the file exists?
  • Is there a way to check if the whole sitemap file system really works – which really means being read correctly by the search engines?
  • Can this problem be related to the settings defined in the .htaccess file?

Please see the screen shots below.
Location of the Sitemap file: https://www.rusinfo.eu/sitemap.xml
Link Yandex Server Check: https://webmaster.yandex.ru/tools/server-response/

Thanks in advance
enter the description of the image here
enter the description of the image here
enter the description of the image here
enter the description of the image here
enter the description of the image here
enter the description of the image here

seo – Can I remove sitemap.xml from my website, will the impact be negative?

I will post a controversial answer with which many will probably disagree. Is Sitemap.xml still necessary? This is no more crucial, no. Is it good to have and help? Sure. Do you have to delete it? I can not say, because I do not know what are the reasons above.

Take a look at this article, written by my former colleague with several years of experience in the areas of SEO, user experience and web development; I agree with a lot of this:

https://www.imarc.com/blog/do-sites-really-need-a-sitemap-for-seo

If you do not actively update your sitemap and do not send it to GSC or Bing Webmaster, you probably do not need it. A sitemap is a suggestion to search engines on how your site should be explored. Think of sites like news sites with hundreds of thousands of pages: sitemaps have size limits. They do not contain all published articles, but these articles are found, indexed and classified.

If your site is updated regularly; has a clear information architecture and a hierarchy of categories; and includes a well-planned internal link structure, the search engines will eventually index and classify the content, regardless of the existence of the sitemap.

However, keep in mind that you will lose the ability to submit the Sitemap manually in GSC, which triggers an analysis again. You will also lose the ability to suggest to search engines which parts of your site are the most important or what are the latest articles.

Many SEOs still say that XML sitemaps are crucial, but that is the conventional wisdom of SEO. Crucial is not how I would define it; Helpful is more like that, from what I've observed in recent years. Regarding your case, it will depend on the reasons for which you delete it and the duration of your maintenance.

Get the status' Unable to recover & # 39; when submitting sitemap.xml for our site [on hold]

I get the status' Unable to recover & # 39; when submitting sitemap.xml for our site. I've validated the site map with the help of some online tools and have the impression that everything is fine. Not sure what could be the reason

seo – How to temporarily host a sitemap.xml file for URLs

I have a GZ sitemap file that contains 4 different sitemaps of 50 MB each. What I want to do is I want to get all the URLs in these 4 sitemap files. I want to use https://robhammond.co/tools/xml-extract for this, but this tool requires to paste the sitemap URL. However, sitemap.xml.gz only allows you to download the XML files. So I end up with a file stored locally and not with a URL (for example, as example.com/sitemap1.xml).

How can I be
1) Host the 4 sitemap.xml files for the above tool
or
2) extract URLs from downloaded sitemap files

robots.txt and sitemap.xml site.webmanifest

Hello, I am able to load the robots.txt and sitemap.xml site.webmanifest files in the form https://example.com/robots.txt, for example, the root directory of magento being defined under / pub via nginx. conf.sample (from Magento 2 installation). I can load files from http://example.com/pub/robots.txt but that's not what I want. I want to be able to upload files like https://example.com/robots.txt https://example.com/sitemap.xml etc. You know how to solve this problem?

sitemap.xml with blank space on line 1

I have a plugin that generates a sitemap and creates a redirect for a multisite installation. When the sitemap is generated, it is stored in the download directory of individual blogs: /wp-content/blogs.dir/1/sitemap.xml but is available via /sitemap.xml

When you navigate to the site map through its current location, it works as expected, with no additional line. However, when you access it through the redirect, an empty space is added to the first line of the site map, causing an error.

I've investigated this problem on different support forums and they all indicated that an extra line at the end of the themes functions.php (after?>) Could create this problem. Alternatively, an additional line in wp-config.php, index.php or another plugin. I've reviewed all these options to no avail.

One suggested solution was to use the Health Check plug-in, which disables all plug-ins and modifies the default theme, and then allows you to activate the plug-ins one after the other. to try to find the problem. Again, this has not brought any solution.

With all plugins disabled and the default theme installed, site plan redirection still existed and resulted in the addition of the extra line.

As such, I started asking myself if it was a fundamental WordPress problem and I reinstalled all the Core scripts, but again in vain.

I have gone through the logs of errors and nothing.

Any idea of ​​where I should turn to try to solve this problem would be very welcome.

thank you,
Tap

Where to place robot.txt and sitemap.xml for wordpress installed in its own directory

My website is set up for wordpress to be installed in its own subdirectory, but the content is broadcast as it was at the root of the domain.
(This wordpress codex page details the configuration)

WordPress URL: https://www.example.com/wordpress
Website URL: https://www.example.com

So in this case the URLs are rewritten to NOT include / wordpress in the URL, it serves pages from (although the URLs of the image make include it: https://www.example.com/wordpress/wp-content/uploads / ...).

The URL of a page is: www.example.com/mypage/
DO NOT: www.example.com/wordpress/mypage/
but the link of the image is: www.example.com/wordpress/wp-content/uploads/2018/12/25/image1.jpg

Question:
Do I put robots.txt and sitemap.xml in the "root of the website" directory on my server (exemple.com)
or his wordpress directory (example.com/wordpress)?

I do not find this particular case in research.
Moreover, testing and errors do not help, because I get conflicting results from various sitemap validators and robots.

In addition, it seems that a plugin or part of WordPress insists on creating a virtual sitemap (sitemap.xml.gz) and robot (robots.txt), which further exacerbates my efforts to definitively determine the appropriate location to place these files for this common but non-standard installation.

magento 1.8 – Exclude deactivated products from the sitemap.xml build

Some of your past responses have not been well received and you may be stuck.

Please pay close attention to the following tips:

  • Please make sure to respond to the question. Provide details and share your research!

But to avoid

  • Ask for help, clarification, or answer other answers.
  • Make statements based on opinions; save them with references or personal experience.

To learn more, read our tips for writing good answers.

how to properly maintain the sitemap.xml automatically?

Hi,

if my domain is domain.com,
I put my main site html pages under / public_html /,
and I'm setting up a blog at / public_html / blog /.

for my main site, sometimes,
I can manually update or download new pages.

and my blog,
I can write a new message every day.

I think to merge them into a sitemap.xml and send them to the search engine.
I find the article https://stackoverflow.com/questions/…e-sitemap-file,
I may need a sitemap.xml to merge the two sitemap files as follows:


http://domain.com/sitemap1.xml
2012


http://domain.com/blog/sitemap2.xml
2016-10-11

but you have to tag it ,
I want to ask if a script or method can automatically update and update them in the main site map?
or is not important?