Bot releases are visible (Hide)
Published by cicirello 11 months ago
exclude-paths
.Published by cicirello about 1 year ago
Published by cicirello almost 2 years ago
<meta name="robots" content="noindex">
in head of html files.<meta content="noindex" name="robots">
(i.e., content before name).Published by cicirello almost 2 years ago
Published by cicirello almost 2 years ago
GITHUB_OUTPUT
environment file. This patch adds backwards compatibility for thoseset-output
if GITHUB_OUTPUT
doesn't exist).Published by cicirello almost 2 years ago
set-output
with the new $GITHUB_OUTPUT
env file.Published by cicirello about 2 years ago
Published by cicirello over 2 years ago
Published by cicirello over 2 years ago
Published by cicirello about 3 years ago
Published by cicirello over 3 years ago
.html
from URLs listed in the sitemapgenerate-sitemap
action enables your sitemap todrop-html-extension
, to control this behavior.Published by cicirello over 3 years ago
Published by cicirello over 3 years ago
Published by cicirello over 3 years ago
additional-extensions
, that enables addingPublished by cicirello over 3 years ago
Published by cicirello about 4 years ago
Published by cicirello about 4 years ago
<meta name="robots" content="noindex">
directives, the generate-sitemap GitHub action now parses a robots.txt file, if present at the root of the website, excluding any URLs from the sitemap that match Disallow:
rules for User-agent: *
.Published by cicirello about 4 years ago
This action generates a sitemap for a website hosted on GitHub Pages. It supports both xml and txt sitemaps. When generating an xml sitemap, it uses the last commit date of each file to generate the <lastmod>
tag in the sitemap entry. It can include html as well as pdf files in the sitemap, and has inputs to control the included file types (defaults include both html and pdf files in the sitemap). It skips html files that contain <meta name="robots" content="noindex">
. It otherwise does not currently attempt to respect a robots.txt file. The sitemap entries are sorted in a consistent order (primary sort is by depth of page in site, and URLs at same depth are then sorted alphabetically).
Published by cicirello about 4 years ago
This action generates a sitemap for a website hosted on GitHub Pages. It supports both xml and txt sitemaps. When generating an xml sitemap, it uses the last commit date of each file to generate the <lastmod>
tag in the sitemap entry. It can include html as well as pdf files in the sitemap, and has inputs to control the included file types (defaults include both html and pdf files in the sitemap). It skips over html files that contain <meta name="robots" content="noindex">
. It otherwise does not currently attempt to respect a robots.txt file. The sitemap entries are sorted in a consistent order (primary sort is by depth of page in site, and URLs at same depth are then sorted alphabetically).