One can sort
robots.txt this way:
User-agent: DESIRED_INPUT Sitemap: https://example.com/sitemap-index.xml Disallow: /
User-agent: DESIRED_INPUT Disallow: / Sitemap: https://example.com/sitemap-index.xml
I assume both are okay because it’s likely the file is compiled in correct order by generally all crawlers.
Is it a best practice to put
Sitemap: to prevent an extremely unlikely bug of a crawler’s bad compilation of crawling before ignoring