# Copyright (c) 2016-2023 Martin Donath # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to # deal in the Software without restriction, including without limitation the # rights to use, copy, modify, merge, publish, distribute, sublicense, and/or # sell copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # The above copyright notice and this permission notice shall be included in # all copies or substantial portions of the Software. # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING # FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS # IN THE SOFTWARE. import logging import os import paginate import posixpath import re import readtime import sys from babel.dates import format_date from copy import copy from datetime import date, datetime, time from hashlib import sha1 from lxml.html import fragment_fromstring, tostring from mkdocs import utils from mkdocs.utils.meta import get_data from mkdocs.commands.build import _populate_page from mkdocs.contrib.search import SearchIndex from mkdocs.plugins import BasePlugin from mkdocs.structure.files import File, Files from mkdocs.structure.nav import Link, Section from mkdocs.structure.pages import Page from tempfile import gettempdir from yaml import SafeLoader, load from material.plugins.blog.config import BlogConfig # ----------------------------------------------------------------------------- # Class # ----------------------------------------------------------------------------- # Blog plugin class BlogPlugin(BasePlugin[BlogConfig]): supports_multiple_instances = True # Determine whether we're running under dirty reload def on_startup(self, *, command, dirty): self.is_serve = (command == "serve") self.is_dirtyreload = False self.is_dirty = dirty # Initialize plugin def on_config(self, config): if not self.config.enabled: return # Resolve source directory for posts and generated files self.post_dir = self._resolve("posts") self.temp_dir = gettempdir() # Initialize posts self.post_map = dict() self.post_meta_map = dict() self.post_pages = [] self.post_pager_pages = [] # Initialize archive if self.config.archive: self.archive_map = dict() self.archive_post_map = dict() # Initialize categories if self.config.categories: self.category_map = dict() self.category_name_map = dict() self.category_post_map = dict() # Initialize authors if self.config.authors: self.authors_map = dict() # Resolve authors file path = os.path.normpath(os.path.join( config.docs_dir, self.config.authors_file.format( blog = self.config.blog_dir ) )) # Load authors map, if it exists if os.path.isfile(path): with open(path, encoding = "utf-8") as f: self.authors_map = load(f, SafeLoader) or {} # Ensure that format strings have no trailing slashes for option in [ "post_url_format", "archive_url_format", "categories_url_format", "pagination_url_format" ]: if self.config[option].endswith("/"): log.error(f"Option '{option}' must not contain trailing slash.") sys.exit(1) # Inherit global table of contents setting if not isinstance(self.config.archive_toc, bool): self.config.archive_toc = self.config.blog_toc if not isinstance(self.config.categories_toc, bool): self.config.categories_toc = self.config.blog_toc # If pagination should not be used, set to large value if not self.config.pagination: self.config.pagination_per_page = 1e7 # By default, drafts are rendered when the documentation is served, # but not when it is built. This should nicely align with the expected # user experience when authoring documentation. if self.is_serve and self.config.draft_on_serve: self.config.draft = True # Adjust paths to assets in the posts directory and preprocess posts def on_files(self, files, *, config): if not self.config.enabled: return # Adjust destination paths for assets path = self._resolve("assets") for file in files.media_files(): if self.post_dir not in file.src_uri: continue # Compute destination URL file.url = file.url.replace(self.post_dir, path) # Compute destination file system path file.dest_uri = file.dest_uri.replace(self.post_dir, path) file.abs_dest_path = os.path.join(config.site_dir, file.dest_path) # Hack: as post URLs are dynamically computed and can be configured by # the author, we need to compute them before we process the contents of # any other page or post. If we wouldn't do that, URLs would be invalid # and we would need to patch them afterwards. The only way to do this # correctly is to first extract the metadata of all posts. Additionally, # while we're at it, generate all archive and category pages as we have # the post metadata on our hands. This ensures that we can safely link # from anywhere to all pages that are generated as part of the blog. for file in files.documentation_pages(): if self.post_dir not in file.src_uri: continue # Read and preprocess post with open(file.abs_src_path, encoding = "utf-8") as f: markdown, meta = get_data(f.read()) # Ensure post has a date set if not meta.get("date"): log.error(f"Blog post '{file.src_uri}' has no date set.") sys.exit(1) # Compute slug from metadata, content or file name headline = utils.get_markdown_title(markdown) slug = meta.get("title", headline or file.name) # Front matter can be defind in YAML, guarded by two lines with # `---` markers, or MultiMarkdown, separated by an empty line. # If the author chooses to use MultiMarkdown syntax, date is # returned as a string, which is different from YAML behavior, # which returns a date. Thus, we must check for its type, and # parse the date for normalization purposes. if isinstance(meta["date"], str): meta["date"] = date.fromisoformat(meta["date"]) # Normalize date to datetime for proper sorting if not isinstance(meta["date"], datetime): meta["date"] = datetime.combine(meta["date"], time()) # Compute category slugs categories = [] for name in meta.get("categories", []): categories.append(self.config.categories_slugify( name, self.config.categories_slugify_separator )) # Check if maximum number of categories is reached max_categories = self.config.post_url_max_categories if len(categories) == max_categories: break # Compute path from format string date_format = self.config.post_url_date_format path = self.config.post_url_format.format( categories = "/".join(categories), date = self._format_date(meta["date"], date_format, config), file = file.name, slug = meta.get("slug", self.config.post_slugify( slug, self.config.post_slugify_separator )) ) # Normalize path, as it may begin with a slash path = posixpath.normpath("/".join([".", path])) # Compute destination URL according to settings file.url = self._resolve(path) if not config.use_directory_urls: file.url += ".html" else: file.url += "/" # Compute destination file system path file.dest_uri = re.sub(r"(?<=\/)$", "index.html", file.url) file.abs_dest_path = os.path.join( config.site_dir, file.dest_path ) # Add post metadata self.post_meta_map[file.src_uri] = meta # Sort post metadata by date (descending) self.post_meta_map = dict(sorted( self.post_meta_map.items(), key = lambda item: item[1]["date"], reverse = True )) # Find and extract the section hosting the blog path = self._resolve("index.md") root = _host(config.nav, path) # Ensure blog root exists file = files.get_file_from_path(path) if not file: log.error(f"Blog root '{path}' does not exist.") sys.exit(1) # Ensure blog root is part of navigation if not root: log.error(f"Blog root '{path}' not in navigation.") sys.exit(1) # Generate and register files for archive if self.config.archive: name = self._translate(config, self.config.archive_name) data = self._generate_files_for_archive(config, files) if data: root.append({ name: data }) # Generate and register files for categories if self.config.categories: name = self._translate(config, self.config.categories_name) data = self._generate_files_for_categories(config, files) if data: root.append({ name: data }) # Hack: add posts temporarily, so MkDocs doesn't complain name = sha1(path.encode("utf-8")).hexdigest() root.append({ f"__posts_${name}": list(self.post_meta_map.keys()) }) # Cleanup navigation before proceeding def on_nav(self, nav, *, config, files): if not self.config.enabled: return # Find and resolve index for cleanup path = self._resolve("index.md") file = files.get_file_from_path(path) # Determine blog root section self.main = file.page if self.main.parent: root = self.main.parent.children else: root = nav.items # Hack: remove temporarily added posts from the navigation name = sha1(path.encode("utf-8")).hexdigest() for item in root: if not item.is_section or item.title != f"__posts_${name}": continue # Detach previous and next links of posts if item.children: head = item.children[+0] tail = item.children[-1] # Link page prior to posts to page after posts if head.previous_page: head.previous_page.next_page = tail.next_page # Link page after posts to page prior to posts if tail.next_page: tail.next_page.previous_page = head.previous_page # Contain previous and next links inside posts head.previous_page = None tail.next_page = None # Set blog as parent page for page in item.children: page.parent = self.main next = page.next_page # Switch previous and next links page.next_page = page.previous_page page.previous_page = next # Remove posts from navigation root.remove(item) break # Prepare post for rendering def on_page_markdown(self, markdown, *, page, config, files): if not self.config.enabled: return # Only process posts if self.post_dir not in page.file.src_uri: return # Skip processing of drafts if self._is_draft(page.file.src_uri): return # Ensure template is set or use default if "template" not in page.meta: page.meta["template"] = "blog-post.html" # Use previously normalized date page.meta["date"] = self.post_meta_map[page.file.src_uri]["date"] # Ensure navigation is hidden page.meta["hide"] = page.meta.get("hide", []) if "navigation" not in page.meta["hide"]: page.meta["hide"].append("navigation") # Format date for rendering date_format = self.config.post_date_format page.meta["date_format"] = self._format_date( page.meta["date"], date_format, config ) # Compute readtime if desired and not explicitly set if self.config.post_readtime: # There's a bug in the readtime library, which causes it to fail # when the input string contains emojis (reported in #5555) encoded = markdown.encode("unicode_escape") if "readtime" not in page.meta: rate = self.config.post_readtime_words_per_minute read = readtime.of_markdown(encoded, rate) page.meta["readtime"] = read.minutes # Compute post categories page.categories = [] if self.config.categories: for name in page.meta.get("categories", []): file = files.get_file_from_path(self.category_name_map[name]) page.categories.append(file.page) # Compute post authors page.authors = [] if self.config.authors: for name in page.meta.get("authors", []): if name not in self.authors_map: log.error( f"Blog post '{page.file.src_uri}' author '{name}' " f"unknown, not listed in .authors.yml" ) sys.exit(1) # Add author to page page.authors.append(self.authors_map[name]) # Fix stale link if previous post is a draft prev = page.previous_page while prev and self._is_draft(prev.file.src_uri): page.previous_page = prev.previous_page prev = prev.previous_page # Fix stale link if next post is a draft next = page.next_page while next and self._is_draft(next.file.src_uri): page.next_page = next.next_page next = next.next_page # Filter posts and generate excerpts for generated pages def on_env(self, env, *, config, files): if not self.config.enabled: return # Skip post excerpts on dirty reload to save time if self.is_dirtyreload: return # Copy configuration and enable 'toc' extension config = copy(config) config.mdx_configs["toc"] = copy(config.mdx_configs.get("toc", {})) # Ensure that post titles are links config.mdx_configs["toc"]["anchorlink"] = True config.mdx_configs["toc"]["permalink"] = False # Filter posts that should not be published for file in files.documentation_pages(): if self.post_dir in file.src_uri: if self._is_draft(file.src_uri): files.remove(file) # Ensure template is set if "template" not in self.main.meta: self.main.meta["template"] = "blog.html" # Populate archive if self.config.archive: for path in self.archive_map: self.archive_post_map[path] = [] # Generate post excerpts for archive base = files.get_file_from_path(path) for file in self.archive_map[path]: self.archive_post_map[path].append( self._generate_excerpt(file, base, config, files) ) # Ensure template is set page = base.page if "template" not in page.meta: page.meta["template"] = "blog-archive.html" # Populate categories if self.config.categories: for path in self.category_map: self.category_post_map[path] = [] # Generate post excerpts for categories base = files.get_file_from_path(path) for file in self.category_map[path]: self.category_post_map[path].append( self._generate_excerpt(file, base, config, files) ) # Ensure template is set page = base.page if "template" not in page.meta: page.meta["template"] = "blog-category.html" # Resolve path of initial index curr = self._resolve("index.md") base = self.main.file # Initialize index self.post_map[curr] = [] self.post_pager_pages.append(self.main) # Generate indexes by paginating through posts for path in self.post_meta_map.keys(): file = files.get_file_from_path(path) if not self._is_draft(path): self.post_pages.append(file.page) else: continue # Generate new index when the current is full per_page = self.config.pagination_per_page if len(self.post_map[curr]) == per_page: offset = 1 + len(self.post_map) # Resolve path of new index curr = self.config.pagination_url_format.format(page = offset) curr = self._resolve(curr + ".md") # Generate file self._generate_file(curr, f"# {self.main.title}") # Register file and page base = self._register_file(curr, config, files) page = self._register_page(base, config, files) # Inherit page metadata, title and position page.meta = self.main.meta page.title = self.main.title page.parent = self.main page.previous_page = self.main.previous_page page.next_page = self.main.next_page # Initialize next index self.post_map[curr] = [] self.post_pager_pages.append(page) # Assign post excerpt to current index self.post_map[curr].append( self._generate_excerpt(file, base, config, files) ) # Populate generated pages def on_page_context(self, context, *, page, config, nav): if not self.config.enabled: return # Provide post excerpts for index path = page.file.src_uri if path in self.post_map: context["posts"] = self.post_map[path] if self.config.blog_toc: self._populate_toc(page, context["posts"]) # Create pagination pagination = paginate.Page( self.post_pages, page = list(self.post_map.keys()).index(path) + 1, items_per_page = self.config.pagination_per_page, url_maker = lambda n: utils.get_relative_url( self.post_pager_pages[n - 1].url, page.url ) ) # Create pagination pager context["pagination"] = lambda args: pagination.pager( format = self.config.pagination_template, show_if_single_page = False, **args ) # Provide post excerpts for archive if self.config.archive: if path in self.archive_post_map: context["posts"] = self.archive_post_map[path] if self.config.archive_toc: self._populate_toc(page, context["posts"]) # Provide post excerpts for categories if self.config.categories: if path in self.category_post_map: context["posts"] = self.category_post_map[path] if self.config.categories_toc: self._populate_toc(page, context["posts"]) # Determine whether we're running under dirty reload def on_serve(self, server, *, config, builder): self.is_dirtyreload = self.is_dirty # ------------------------------------------------------------------------- # Generate and register files for archive def _generate_files_for_archive(self, config, files): for path, meta in self.post_meta_map.items(): file = files.get_file_from_path(path) if self._is_draft(path): continue # Compute name from format string date_format = self.config.archive_date_format name = self._format_date(meta["date"], date_format, config) # Compute path from format string date_format = self.config.archive_url_date_format path = self.config.archive_url_format.format( date = self._format_date(meta["date"], date_format, config) ) # Create file for archive if it doesn't exist path = self._resolve(path + ".md") if path not in self.archive_map: self.archive_map[path] = [] # Generate and register file for archive self._generate_file(path, f"# {name}") self._register_file(path, config, files) # Assign current post to archive self.archive_map[path].append(file) # Return generated archive files return list(self.archive_map.keys()) # Generate and register files for categories def _generate_files_for_categories(self, config, files): allowed = set(self.config.categories_allowed) for path, meta in self.post_meta_map.items(): file = files.get_file_from_path(path) if self._is_draft(path): continue # Ensure category is in (non-empty) allow list categories = set(meta.get("categories", [])) if allowed: for name in categories - allowed: log.error( f"Blog post '{file.src_uri}' uses a category " f"which is not in allow list: {name}" ) sys.exit(1) # Traverse all categories of the post for name in categories: path = self.config.categories_url_format.format( slug = self.config.categories_slugify( name, self.config.categories_slugify_separator ) ) # Create file for category if it doesn't exist path = self._resolve(path + ".md") if path not in self.category_map: self.category_map[path] = [] # Generate and register file for category self._generate_file(path, f"# {name}") self._register_file(path, config, files) # Link category path to name self.category_name_map[name] = path # Assign current post to category self.category_map[path].append(file) # Sort categories alphabetically (ascending) self.category_map = dict(sorted(self.category_map.items())) # Return generated category files return list(self.category_map.keys()) # ------------------------------------------------------------------------- # Check if a post is a draft def _is_draft(self, path): meta = self.post_meta_map[path] if not self.config.draft: # Check if post date is in the future future = False if self.config.draft_if_future_date: future = meta["date"] > datetime.now() # Check if post is marked as draft return meta.get("draft", future) # Post is not a draft return False # Generate a post excerpt relative to base def _generate_excerpt(self, file, base, config, files): page = file.page # Generate temporary file and page for post excerpt temp = self._register_file(file.src_uri, config) excerpt = Page(page.title, temp, config) # Check for separator, if post excerpt is required separator = self.config.post_excerpt_separator if self.config.post_excerpt == "required": if separator not in page.markdown: log.error(f"Blog post '{temp.src_uri}' has no excerpt.") sys.exit(1) # Ensure separator at the end to strip footnotes and patch h1-h5 markdown = "\n\n".join([page.markdown, separator]) markdown = re.sub(r"(^#{1,5})", "#\\1", markdown, flags = re.MULTILINE) # Extract content and metadata from original post excerpt.file.url = base.url excerpt.markdown = markdown excerpt.meta = page.meta # Render post and revert page URL excerpt.render(config, files) excerpt.file.url = page.url # Find all anchor links expr = re.compile( r"]+href=['\"]?#[^>]+>", re.IGNORECASE | re.MULTILINE ) # Replace callback first = True def replace(match): value = match.group() # Handle anchor link el = fragment_fromstring(value.encode("utf-8")) if el.tag == "a": nonlocal first # Fix up each anchor link of the excerpt with a link to the # anchor of the actual post, except for the first one – that # one needs to go to the top of the post. A better way might # be a Markdown extension, but for now this should be fine. url = utils.get_relative_url(excerpt.file.url, base.url) if first: el.set("href", url) else: el.set("href", url + el.get("href")) # From now on reference anchors first = False # Replace link opening tag (without closing tag) return tostring(el, encoding = "unicode")[:-4] # Extract excerpt from post and replace anchor links excerpt.content = expr.sub( replace, excerpt.content.split(separator)[0] ) # Determine maximum number of authors and categories max_authors = self.config.post_excerpt_max_authors max_categories = self.config.post_excerpt_max_categories # Obtain computed metadata from original post excerpt.authors = page.authors[:max_authors] excerpt.categories = page.categories[:max_categories] # Return post excerpt return excerpt # Generate a file with the given template and content def _generate_file(self, path, content): content = f"---\nsearch:\n exclude: true\n---\n\n{content}" utils.write_file( bytes(content, "utf-8"), os.path.join(self.temp_dir, path) ) # Register a file def _register_file(self, path, config, files = Files([])): file = files.get_file_from_path(path) if not file: urls = config.use_directory_urls file = File(path, self.temp_dir, config.site_dir, urls) files.append(file) # Mark file as generated, so other plugins don't think it's part # of the file system. This is more or less a new quasi-standard # for plugins that generate files which was introduced by the # git-revision-date-localized-plugin - see https://bit.ly/3ZUmdBx file.generated_by = "material/blog" # Return file return file # Register and populate a page def _register_page(self, file, config, files): page = Page(None, file, config) _populate_page(page, config, files) return page # Populate table of contents of given page def _populate_toc(self, page, posts): toc = page.toc.items[0] for post in posts: toc.children.append(post.toc.items[0]) # Remove anchors below the second level post.toc.items[0].children = [] # Translate the given placeholder value def _translate(self, config, value): env = config.theme.get_env() # Load language template and return translation for placeholder language = "partials/language.html" template = env.get_template(language, None, { "config": config }) return template.module.t(value) # Resolve path relative to blog root def _resolve(self, *args): path = posixpath.join(self.config.blog_dir, *args) return posixpath.normpath(path) # Format date according to locale def _format_date(self, date, format, config): return format_date( date, format = format, locale = config.theme["language"] ) # ----------------------------------------------------------------------------- # Helper functions # ----------------------------------------------------------------------------- # Search the given navigation section (from the configuration) recursively to # find the section to host all generated pages (archive, categories, etc.) def _host(nav, path): # Search navigation dictionary if isinstance(nav, dict): for _, item in nav.items(): result = _host(item, path) if result: return result # Search navigation list elif isinstance(nav, list): if path in nav: return nav # Search each list item for item in nav: if isinstance(item, dict) and path in item.values(): if path in item.values(): return nav else: result = _host(item, path) if result: return result # Copied and adapted from MkDocs, because we need to return existing pages and # support anchor names as subtitles, which is pretty fucking cool. def _data_to_navigation(nav, config, files): # Search navigation dictionary if isinstance(nav, dict): return [ _data_to_navigation((key, value), config, files) if isinstance(value, str) else Section( title = key, children = _data_to_navigation(value, config, files) ) for key, value in nav.items() ] # Search navigation list elif isinstance(nav, list): return [ _data_to_navigation(item, config, files)[0] if isinstance(item, dict) and len(item) == 1 else _data_to_navigation(item, config, files) for item in nav ] # Extract navigation title and path and split anchors title, path = nav if isinstance(nav, tuple) else (None, nav) path, _, anchor = path.partition("#") # Try to retrieve existing file file = files.get_file_from_path(path) if not file: return Link(title, path) # Use resolved assets destination path if not path.endswith(".md"): return Link(title or os.path.basename(path), file.url) # Generate temporary file as for post excerpts else: urls = config.use_directory_urls link = File(path, config.docs_dir, config.site_dir, urls) page = Page(title or file.page.title, link, config) # Set destination file system path and URL from original file link.dest_uri = file.dest_uri link.abs_dest_path = file.abs_dest_path link.url = file.url # Retrieve name of anchor by misusing the search index if anchor: item = SearchIndex()._find_toc_by_id(file.page.toc, anchor) # Set anchor name as subtitle page.meta["subtitle"] = item.title link.url += f"#{anchor}" # Return navigation item return page # ----------------------------------------------------------------------------- # Data # ----------------------------------------------------------------------------- # Set up logging log = logging.getLogger("mkdocs.material.blog")