Bot releases are visible (Hide)
Published by cicirello about 4 years ago
This action generates a sitemap for a website hosted on GitHub Pages. It supports both xml and txt sitemaps. When generating an xml sitemap, it uses the last commit date of each file to generate the <lastmod>
tag in the sitemap entry. It can include html as well as pdf files in the sitemap, and has inputs to control the included file types (defaults include both html and pdf files in the sitemap). It skips over html files that contain <meta name="robots" content="noindex">
. It otherwise does not currently attempt to respect a robots.txt file. The sitemap entries are sorted in a consistent order (primary sort is by depth of page in site, and URLs at same depth are then sorted alphabetically).
Published by cicirello about 4 years ago
This action generates a sitemap for a website hosted on GitHub Pages. It supports both xml and txt sitemaps. When generating an xml sitemap, it uses the last commit date of each file to generate the <lastmod>
tag in the sitemap entry. It can include html as well as pdf files in the sitemap, and has inputs to control the included file types (defaults include both html and pdf files in the sitemap). It skips over html files that contain <meta name="robots" content="noindex">
. It otherwise does not currently attempt to respect a robots.txt file. The sitemap entries are sorted in a consistent order (html pages before pdf files, then by depth of page in site, and finally alphabetically).
Published by cicirello about 4 years ago
This action generates a sitemap for a website hosted on GitHub Pages. It supports both xml and txt sitemaps. When generating an xml sitemap, it uses the last commit date of each file to generate the <lastmod>
tag in the sitemap entry. It can include html as well as pdf files in the sitemap, and has inputs to control the included file types (defaults include both html and pdf files in the sitemap). It skips over html files that contain <meta name="robots" content="noindex">
. It otherwise does not currently attempt to respect a robots.txt file. The sitemap entries are sorted in a consistent order (html pages before pdf files, then by depth of page in site, and finally alphabetically).
Published by cicirello about 4 years ago
This action generates a sitemap for a website hosted on GitHub Pages. It supports both xml and txt sitemaps. When generating an xml sitemap, it uses the last commit date of each file to generate the <lastmod>
tag in the sitemap entry. It can include html as well as pdf files in the sitemap, and has inputs to control the included file types (defaults include both html and pdf files in the sitemap). It skips over html files that contain <meta name="robots" content="noindex">
. It otherwise does not currently attempt to respect a robots.txt file.