Robots Governance
Each robots will contain this core set of URLs:
Disallow: /stage.digital2.aveva.com
Disallow: /stage.digital3.aveva.com
Disallow: /search-results
Disallow: /search-results*
Disallow: /404
Disallow: /500
Disallow: /search
Disallow: /search/*
Disallow: Search.aspx
Disallow: Search?search*
Disallow: search?q=*
Disallow: admin
Disallow: admin/*
Disallow: assets/*
Disallow: /styles
Disallow: /styles/*
Only en will include these:
Sitemap: https://www.aveva.com/sitemap.xml
Sitemap: https://www.aveva.com/sitemap_es.xml
Sitemap: https://www.aveva.com/sitemap_de.xml
Sitemap: https://www.aveva.com/sitemap_pt.xml
Sitemap: https://www.aveva.com/sitemap_ko.xml
Sitemap: https://www.aveva.com/sitemap_ja.xml
Sitemap: https://www.aveva.com/sitemap_cn.xml
Sitemap: https://www.aveva.com/sitemap_fr.xml
Only cn will include this:
Sitemap: https://www.aveva.cn/sitemap.xml
Live Site
The correct version of the robots.txt file must show on these URLs.
If updates are reflecting at the PROD URLs, but they are not showing on these live site URLs, then ask Adobe to cache the URLs.
- https://www.aveva.com/robots.txt (used for en site, as well as all international sites, except for China site)
- https://www.aveva.cn/robots.txt (used for cn site only)
PROD
- https://www.aveva.com/en/robots/ (used for en site, as well as all international sites, except for China site)
- https://www.aveva.cn/robots/ (used for cn site only)
STAGE, QA, DEV
A recurring problem that AVEVA experiences is the URLs from the lower environments (STAGE, QA, DEV) getting into external search results.
To minimize or prevent this issue from occuring, the robots file on the lower environments should contain the following additional rule at the top of the file:
- Disallow: /