Compare commits
38 commits
Author | SHA1 | Date | |
---|---|---|---|
|
45a5371074 | ||
|
79555fb14f | ||
|
620086c17a | ||
|
2458d2cc0c | ||
ba801995a8 | |||
7bea868899 | |||
6de5db8485 | |||
aec603f2e8 | |||
cece2bee68 | |||
e11c3b2688 | |||
50207d7e25 | |||
8d74cb0c26 | |||
90d85fedb4 | |||
279d885478 | |||
56417c6763 | |||
d78393f1b3 | |||
bf80f73950 | |||
38b079d42d | |||
43897cc961 | |||
bf714a30bc | |||
|
89ea18a805 | ||
e3b30e6a98 | |||
|
817b23ac69 | ||
d517416077 | |||
|
390ca758b9 | ||
609dd14d90 | |||
958f4106cc | |||
34b187d0d4 | |||
7530ecf7ba | |||
0eb29e8c5a | |||
150af7f868 | |||
afc4e09784 | |||
e06a2145c2 | |||
261e5bfc0e | |||
cfd929f36f | |||
07b257e152 | |||
d993d1c29b | |||
3ff839e5d3 |
12 changed files with 234 additions and 153 deletions
15
CHANGELOG.md
15
CHANGELOG.md
|
@ -1,5 +1,20 @@
|
||||||
# Changelog
|
# Changelog
|
||||||
|
|
||||||
|
## 0.2.24
|
||||||
|
|
||||||
|
- Better default fonts - `sans-serif` style for the index and the titles,
|
||||||
|
`serif` for the articles' body.
|
||||||
|
|
||||||
|
## 0.2.19
|
||||||
|
|
||||||
|
- Added `short_feed` configuration flag to permanently disable returning the
|
||||||
|
full content of the articles in the RSS feed.
|
||||||
|
|
||||||
|
## 0.2.16
|
||||||
|
|
||||||
|
- Removed `alt` attribute from LaTeX rendered `<img>` tags. It may generate
|
||||||
|
non-standard Unicode characters that break the RSS feed.
|
||||||
|
|
||||||
## 0.2.14
|
## 0.2.14
|
||||||
|
|
||||||
- Better support for PWA tags and added a default config-generated `/manifest.json`.
|
- Better support for PWA tags and added a default config-generated `/manifest.json`.
|
||||||
|
|
|
@ -72,6 +72,8 @@ logo: /path/or/url/here
|
||||||
language: en-US
|
language: en-US
|
||||||
# Show/hide the header (default: true)
|
# Show/hide the header (default: true)
|
||||||
header: true
|
header: true
|
||||||
|
# Enable/disable the short RSS feed (default: false)
|
||||||
|
short_feed: false
|
||||||
|
|
||||||
categories:
|
categories:
|
||||||
- category1
|
- category1
|
||||||
|
@ -133,4 +135,9 @@ $$
|
||||||
RSS feeds for the blog are provided under the `/rss` URL.
|
RSS feeds for the blog are provided under the `/rss` URL.
|
||||||
|
|
||||||
By default, the whole HTML-rendered content of an article is returned under `rss.channel.item.description`.
|
By default, the whole HTML-rendered content of an article is returned under `rss.channel.item.description`.
|
||||||
|
|
||||||
If you only want to include the short description of an article in the feed, use `/rss?short` instead.
|
If you only want to include the short description of an article in the feed, use `/rss?short` instead.
|
||||||
|
|
||||||
|
If you want the short feed (i.e. without the fully rendered article as a
|
||||||
|
description) to be always returned, then you can specify `short_feed=true` in
|
||||||
|
your configuration.
|
||||||
|
|
|
@ -1 +1 @@
|
||||||
__version__ = '0.2.14'
|
__version__ = "0.2.31"
|
||||||
|
|
114
madblog/app.py
114
madblog/app.py
|
@ -12,11 +12,11 @@ from ._sorters import PagesSorter, PagesSortByTime
|
||||||
|
|
||||||
|
|
||||||
class BlogApp(Flask):
|
class BlogApp(Flask):
|
||||||
_title_header_regex = re.compile(r'^#\s*((\[(.*)\])|(.*))')
|
_title_header_regex = re.compile(r"^#\s*((\[(.*)\])|(.*))")
|
||||||
|
|
||||||
def __init__(self, *args, **kwargs):
|
def __init__(self, *args, **kwargs):
|
||||||
super().__init__(*args, template_folder=config.templates_dir, **kwargs)
|
super().__init__(*args, template_folder=config.templates_dir, **kwargs)
|
||||||
self.pages_dir = os.path.join(config.content_dir, 'markdown')
|
self.pages_dir = os.path.join(config.content_dir, "markdown")
|
||||||
self.img_dir = config.default_img_dir
|
self.img_dir = config.default_img_dir
|
||||||
self.css_dir = config.default_css_dir
|
self.css_dir = config.default_css_dir
|
||||||
self.js_dir = config.default_js_dir
|
self.js_dir = config.default_js_dir
|
||||||
|
@ -27,73 +27,77 @@ class BlogApp(Flask):
|
||||||
# `config.content_dir` is treated as the root for markdown files.
|
# `config.content_dir` is treated as the root for markdown files.
|
||||||
self.pages_dir = config.content_dir
|
self.pages_dir = config.content_dir
|
||||||
|
|
||||||
img_dir = os.path.join(config.content_dir, 'img')
|
img_dir = os.path.join(config.content_dir, "img")
|
||||||
if os.path.isdir(img_dir):
|
if os.path.isdir(img_dir):
|
||||||
self.img_dir = os.path.abspath(img_dir)
|
self.img_dir = os.path.abspath(img_dir)
|
||||||
else:
|
else:
|
||||||
self.img_dir = config.content_dir
|
self.img_dir = config.content_dir
|
||||||
|
|
||||||
css_dir = os.path.join(config.content_dir, 'css')
|
css_dir = os.path.join(config.content_dir, "css")
|
||||||
if os.path.isdir(css_dir):
|
if os.path.isdir(css_dir):
|
||||||
self.css_dir = os.path.abspath(css_dir)
|
self.css_dir = os.path.abspath(css_dir)
|
||||||
|
|
||||||
js_dir = os.path.join(config.content_dir, 'js')
|
js_dir = os.path.join(config.content_dir, "js")
|
||||||
if os.path.isdir(js_dir):
|
if os.path.isdir(js_dir):
|
||||||
self.js_dir = os.path.abspath(js_dir)
|
self.js_dir = os.path.abspath(js_dir)
|
||||||
|
|
||||||
fonts_dir = os.path.join(config.content_dir, 'fonts')
|
fonts_dir = os.path.join(config.content_dir, "fonts")
|
||||||
if os.path.isdir(fonts_dir):
|
if os.path.isdir(fonts_dir):
|
||||||
self.fonts_dir = os.path.abspath(fonts_dir)
|
self.fonts_dir = os.path.abspath(fonts_dir)
|
||||||
|
|
||||||
templates_dir = os.path.join(config.content_dir, 'templates')
|
templates_dir = os.path.join(config.content_dir, "templates")
|
||||||
if os.path.isdir(templates_dir):
|
if os.path.isdir(templates_dir):
|
||||||
self.template_folder = os.path.abspath(templates_dir)
|
self.template_folder = os.path.abspath(templates_dir)
|
||||||
|
|
||||||
def get_page_metadata(self, page: str) -> dict:
|
def get_page_metadata(self, page: str) -> dict:
|
||||||
if not page.endswith('.md'):
|
if not page.endswith(".md"):
|
||||||
page = page + '.md'
|
page = page + ".md"
|
||||||
|
|
||||||
md_file = os.path.join(self.pages_dir, page)
|
md_file = os.path.join(self.pages_dir, page)
|
||||||
if not os.path.isfile(md_file):
|
if not os.path.isfile(md_file):
|
||||||
abort(404)
|
abort(404)
|
||||||
|
|
||||||
metadata = {}
|
metadata = {}
|
||||||
with open(md_file, 'r') as f:
|
with open(md_file, "r") as f:
|
||||||
metadata['uri'] = '/article/' + page[:-3]
|
metadata["uri"] = "/article/" + page[:-3]
|
||||||
|
|
||||||
for line in f.readlines():
|
for line in f:
|
||||||
if not line:
|
if not line:
|
||||||
continue
|
continue
|
||||||
|
|
||||||
if not (m := re.match(r'^\[//]: # \(([^:]+):\s*([^)]+)\)\s*$', line)):
|
if not (m := re.match(r"^\[//]: # \(([^:]+):\s*(.*)\)\s*$", line)):
|
||||||
break
|
break
|
||||||
|
|
||||||
if m.group(1) == 'published':
|
if m.group(1) == "published":
|
||||||
metadata[m.group(1)] = datetime.date.fromisoformat(m.group(2))
|
metadata[m.group(1)] = datetime.datetime.fromisoformat(
|
||||||
|
m.group(2)
|
||||||
|
).date()
|
||||||
else:
|
else:
|
||||||
metadata[m.group(1)] = m.group(2)
|
metadata[m.group(1)] = m.group(2)
|
||||||
|
|
||||||
if not metadata.get('title'):
|
if not metadata.get("title"):
|
||||||
# If the `title` header isn't available in the file,
|
# If the `title` header isn't available in the file,
|
||||||
# infer it from the first line of the file
|
# infer it from the first line of the file
|
||||||
with open(md_file, 'r') as f:
|
with open(md_file, "r") as f:
|
||||||
header = ''
|
header = ""
|
||||||
for line in f.readlines():
|
for line in f.readlines():
|
||||||
header = line
|
header = line
|
||||||
break
|
break
|
||||||
|
|
||||||
metadata['title_inferred'] = True
|
metadata["title_inferred"] = True
|
||||||
m = self._title_header_regex.search(header)
|
m = self._title_header_regex.search(header)
|
||||||
if m:
|
if m:
|
||||||
metadata['title'] = m.group(3) or m.group(1)
|
metadata["title"] = m.group(3) or m.group(1)
|
||||||
else:
|
else:
|
||||||
metadata['title'] = os.path.basename(md_file)
|
metadata["title"] = os.path.basename(md_file)
|
||||||
|
|
||||||
if not metadata.get('published'):
|
if not metadata.get("published"):
|
||||||
# If the `published` header isn't available in the file,
|
# If the `published` header isn't available in the file,
|
||||||
# infer it from the file's creation date
|
# infer it from the file's creation date
|
||||||
metadata['published'] = datetime.date.fromtimestamp(os.stat(md_file).st_ctime)
|
metadata["published"] = datetime.date.fromtimestamp(
|
||||||
metadata['published_inferred'] = True
|
os.stat(md_file).st_ctime
|
||||||
|
)
|
||||||
|
metadata["published_inferred"] = True
|
||||||
|
|
||||||
return metadata
|
return metadata
|
||||||
|
|
||||||
|
@ -102,37 +106,42 @@ class BlogApp(Flask):
|
||||||
page: str,
|
page: str,
|
||||||
title: Optional[str] = None,
|
title: Optional[str] = None,
|
||||||
skip_header: bool = False,
|
skip_header: bool = False,
|
||||||
skip_html_head: bool = False
|
skip_html_head: bool = False,
|
||||||
):
|
):
|
||||||
if not page.endswith('.md'):
|
if not page.endswith(".md"):
|
||||||
page = page + '.md'
|
page = page + ".md"
|
||||||
|
|
||||||
metadata = self.get_page_metadata(page)
|
metadata = self.get_page_metadata(page)
|
||||||
# Don't duplicate the page title if it's been inferred
|
# Don't duplicate the page title if it's been inferred
|
||||||
if not (title or metadata.get('title_inferred')):
|
if not (title or metadata.get("title_inferred")):
|
||||||
title = metadata.get('title', config.title)
|
title = metadata.get("title", config.title)
|
||||||
|
|
||||||
with open(os.path.join(self.pages_dir, page), 'r') as f:
|
with open(os.path.join(self.pages_dir, page), "r") as f:
|
||||||
return render_template(
|
return render_template(
|
||||||
'article.html',
|
"article.html",
|
||||||
config=config,
|
config=config,
|
||||||
title=title,
|
title=title,
|
||||||
image=metadata.get('image'),
|
image=metadata.get("image"),
|
||||||
description=metadata.get('description'),
|
description=metadata.get("description"),
|
||||||
author=(
|
author=(
|
||||||
re.match(r'(.+?)\s+<([^>]+>)', metadata['author'])[1]
|
re.match(r"(.+?)\s+<([^>]+>)", metadata["author"])[1]
|
||||||
if 'author' in metadata else None
|
if "author" in metadata
|
||||||
),
|
|
||||||
author_email=(
|
|
||||||
re.match(r'(.+?)\s+<([^>]+)>', metadata['author'])[2]
|
|
||||||
if 'author' in metadata else None
|
|
||||||
),
|
|
||||||
published=(
|
|
||||||
metadata['published'].strftime('%b %d, %Y')
|
|
||||||
if metadata.get('published') and not metadata.get('published_inferred')
|
|
||||||
else None
|
else None
|
||||||
),
|
),
|
||||||
content=markdown(f.read(), extensions=['fenced_code', 'codehilite', MarkdownLatex()]),
|
author_email=(
|
||||||
|
re.match(r"(.+?)\s+<([^>]+)>", metadata["author"])[2]
|
||||||
|
if "author" in metadata
|
||||||
|
else None
|
||||||
|
),
|
||||||
|
published=(
|
||||||
|
metadata["published"].strftime("%b %d, %Y")
|
||||||
|
if metadata.get("published")
|
||||||
|
and not metadata.get("published_inferred")
|
||||||
|
else None
|
||||||
|
),
|
||||||
|
content=markdown(
|
||||||
|
f.read(), extensions=["fenced_code", "codehilite", "tables", MarkdownLatex()]
|
||||||
|
),
|
||||||
skip_header=skip_header,
|
skip_header=skip_header,
|
||||||
skip_html_head=skip_html_head,
|
skip_html_head=skip_html_head,
|
||||||
)
|
)
|
||||||
|
@ -145,26 +154,25 @@ class BlogApp(Flask):
|
||||||
sorter: Type[PagesSorter] = PagesSortByTime,
|
sorter: Type[PagesSorter] = PagesSortByTime,
|
||||||
reverse: bool = True,
|
reverse: bool = True,
|
||||||
) -> List[Tuple[int, dict]]:
|
) -> List[Tuple[int, dict]]:
|
||||||
pages_dir = app.pages_dir.rstrip('/')
|
pages_dir = app.pages_dir.rstrip("/")
|
||||||
pages = [
|
pages = [
|
||||||
{
|
{
|
||||||
'path': os.path.join(root[len(pages_dir)+1:], f),
|
"path": os.path.join(root[len(pages_dir) + 1 :], f),
|
||||||
'folder': root[len(pages_dir)+1:],
|
"folder": root[len(pages_dir) + 1 :],
|
||||||
'content': (
|
"content": (
|
||||||
self.get_page(
|
self.get_page(
|
||||||
os.path.join(root, f),
|
os.path.join(root, f),
|
||||||
skip_header=skip_header,
|
skip_header=skip_header,
|
||||||
skip_html_head=skip_html_head,
|
skip_html_head=skip_html_head,
|
||||||
)
|
)
|
||||||
if with_content else ''
|
if with_content
|
||||||
),
|
else ""
|
||||||
**self.get_page_metadata(
|
|
||||||
os.path.join(root[len(pages_dir)+1:], f)
|
|
||||||
),
|
),
|
||||||
|
**self.get_page_metadata(os.path.join(root[len(pages_dir) + 1 :], f)),
|
||||||
}
|
}
|
||||||
for root, _, files in os.walk(pages_dir, followlinks=True)
|
for root, _, files in os.walk(pages_dir, followlinks=True)
|
||||||
for f in files
|
for f in files
|
||||||
if f.endswith('.md')
|
if f.endswith(".md")
|
||||||
]
|
]
|
||||||
|
|
||||||
sorter_func = sorter(pages)
|
sorter_func = sorter(pages)
|
||||||
|
|
|
@ -1,57 +1,61 @@
|
||||||
import os
|
import os
|
||||||
|
from typing import List
|
||||||
import yaml
|
import yaml
|
||||||
|
|
||||||
from dataclasses import dataclass
|
from dataclasses import dataclass, field
|
||||||
|
|
||||||
|
|
||||||
@dataclass
|
@dataclass
|
||||||
class Config:
|
class Config:
|
||||||
title = 'Blog'
|
title = "Blog"
|
||||||
description = ''
|
description = ""
|
||||||
link = '/'
|
link = "/"
|
||||||
home_link = '/'
|
home_link = "/"
|
||||||
language = 'en-US'
|
language = "en-US"
|
||||||
logo = '/img/icon.png'
|
logo = "/img/icon.png"
|
||||||
header = True
|
header = True
|
||||||
content_dir = '.'
|
content_dir = "."
|
||||||
categories = None
|
categories: List[str] = field(default_factory=list)
|
||||||
|
short_feed = False
|
||||||
|
|
||||||
basedir = os.path.abspath(os.path.dirname(__file__))
|
basedir = os.path.abspath(os.path.dirname(__file__))
|
||||||
templates_dir = os.path.join(basedir, 'templates')
|
templates_dir = os.path.join(basedir, "templates")
|
||||||
static_dir = os.path.join(basedir, 'static')
|
static_dir = os.path.join(basedir, "static")
|
||||||
default_css_dir = os.path.join(static_dir, 'css')
|
default_css_dir = os.path.join(static_dir, "css")
|
||||||
default_js_dir = os.path.join(static_dir, 'js')
|
default_js_dir = os.path.join(static_dir, "js")
|
||||||
default_fonts_dir = os.path.join(static_dir, 'fonts')
|
default_fonts_dir = os.path.join(static_dir, "fonts")
|
||||||
default_img_dir = os.path.join(static_dir, 'img')
|
default_img_dir = os.path.join(static_dir, "img")
|
||||||
|
|
||||||
|
|
||||||
config = Config()
|
config = Config()
|
||||||
|
|
||||||
|
|
||||||
def init_config(content_dir='.', config_file='config.yaml'):
|
def init_config(content_dir=".", config_file="config.yaml"):
|
||||||
cfg = {}
|
cfg = {}
|
||||||
config.content_dir = content_dir
|
config.content_dir = content_dir
|
||||||
|
|
||||||
if os.path.isfile(config_file):
|
if os.path.isfile(config_file):
|
||||||
with open(config_file, 'r') as f:
|
with open(config_file, "r") as f:
|
||||||
cfg = yaml.safe_load(f)
|
cfg = yaml.safe_load(f)
|
||||||
|
|
||||||
if cfg.get('title'):
|
if cfg.get("title"):
|
||||||
config.title = cfg['title']
|
config.title = cfg["title"]
|
||||||
if cfg.get('description'):
|
if cfg.get("description"):
|
||||||
config.description = cfg['description']
|
config.description = cfg["description"]
|
||||||
if cfg.get('link'):
|
if cfg.get("link"):
|
||||||
config.link = cfg['link']
|
config.link = cfg["link"]
|
||||||
if cfg.get('home_link'):
|
if cfg.get("home_link"):
|
||||||
config.home_link = cfg['home_link']
|
config.home_link = cfg["home_link"]
|
||||||
if cfg.get('logo') is not None:
|
if cfg.get("logo") is not None:
|
||||||
config.logo = cfg['logo']
|
config.logo = cfg["logo"]
|
||||||
if cfg.get('language'):
|
if cfg.get("language"):
|
||||||
config.language = cfg['language']
|
config.language = cfg["language"]
|
||||||
if cfg.get('header') is False:
|
if cfg.get("header") is False:
|
||||||
config.header = False
|
config.header = False
|
||||||
|
if cfg.get("short_feed"):
|
||||||
|
config.short_feed = True
|
||||||
|
|
||||||
config.categories = cfg.get('categories', [])
|
config.categories = cfg.get("categories", [])
|
||||||
|
|
||||||
|
|
||||||
# vim:sw=4:ts=4:et:
|
# vim:sw=4:ts=4:et:
|
||||||
|
|
|
@ -31,11 +31,11 @@ def call(*args, **kwargs):
|
||||||
|
|
||||||
|
|
||||||
# Defines our basic inline image
|
# Defines our basic inline image
|
||||||
img_expr = '<img class="latex inline math-%s" alt="%s" id="%s" src="data:image/png;base64,%s">'
|
img_expr = '<img class="latex inline math-%s" id="%s" src="data:image/png;base64,%s">'
|
||||||
|
|
||||||
# Defines multiline expression image
|
# Defines multiline expression image
|
||||||
multiline_img_expr = '''<div class="multiline-wrapper">
|
multiline_img_expr = """<div class="multiline-wrapper">
|
||||||
<img class="latex multiline math-%s" alt="%s" id="%s" src="data:image/png;base64,%s"></div>'''
|
<img class="latex multiline math-%s" id="%s" src="data:image/png;base64,%s"></div>"""
|
||||||
|
|
||||||
# Base CSS template
|
# Base CSS template
|
||||||
img_css = """<style scoped>
|
img_css = """<style scoped>
|
||||||
|
@ -55,8 +55,8 @@ img.latex.inline {
|
||||||
</style>"""
|
</style>"""
|
||||||
|
|
||||||
# Cache and temp file paths
|
# Cache and temp file paths
|
||||||
tmpdir = tempfile.gettempdir() + '/markdown-latex'
|
tmpdir = tempfile.gettempdir() + "/markdown-latex"
|
||||||
cache_file = tmpdir + '/latex.cache'
|
cache_file = tmpdir + "/latex.cache"
|
||||||
|
|
||||||
|
|
||||||
class LaTeXPreprocessor(markdown.preprocessors.Preprocessor):
|
class LaTeXPreprocessor(markdown.preprocessors.Preprocessor):
|
||||||
|
@ -75,16 +75,20 @@ class LaTeXPreprocessor(markdown.preprocessors.Preprocessor):
|
||||||
"""
|
"""
|
||||||
|
|
||||||
# Math TeX extraction regex
|
# Math TeX extraction regex
|
||||||
math_extract_regex = re.compile(r'(.+?)((\\\(.+?\\\))|(\$\$\n.+?\n\$\$\n))(.*)', re.MULTILINE | re.DOTALL)
|
math_extract_regex = re.compile(
|
||||||
|
r"(.+?)((\\\(.+?\\\))|(\$\$\n.+?\n\$\$\n))(.*)", re.MULTILINE | re.DOTALL
|
||||||
|
)
|
||||||
|
|
||||||
# Math TeX matching regex
|
# Math TeX matching regex
|
||||||
math_match_regex = re.compile(r'\s*(\\\(.+?\\\))|(\$\$\n.+?\n\$\$\n)\s*', re.MULTILINE | re.DOTALL)
|
math_match_regex = re.compile(
|
||||||
|
r"\s*(\\\(.+?\\\))|(\$\$\n.+?\n\$\$\n)\s*", re.MULTILINE | re.DOTALL
|
||||||
|
)
|
||||||
|
|
||||||
def __init__(self, *_, **__):
|
def __init__(self, *_, **__):
|
||||||
if not os.path.isdir(tmpdir):
|
if not os.path.isdir(tmpdir):
|
||||||
os.makedirs(tmpdir)
|
os.makedirs(tmpdir)
|
||||||
try:
|
try:
|
||||||
with open(cache_file, 'r') as f:
|
with open(cache_file, "r") as f:
|
||||||
self.cached = json.load(f)
|
self.cached = json.load(f)
|
||||||
except (IOError, json.JSONDecodeError):
|
except (IOError, json.JSONDecodeError):
|
||||||
self.cached = {}
|
self.cached = {}
|
||||||
|
@ -94,7 +98,8 @@ class LaTeXPreprocessor(markdown.preprocessors.Preprocessor):
|
||||||
("dvipng", "args"): "-q -T tight -bg Transparent -z 9 -D 200",
|
("dvipng", "args"): "-q -T tight -bg Transparent -z 9 -D 200",
|
||||||
("delimiters", "text"): "%",
|
("delimiters", "text"): "%",
|
||||||
("delimiters", "math"): "$",
|
("delimiters", "math"): "$",
|
||||||
("delimiters", "preamble"): "%%"}
|
("delimiters", "preamble"): "%%",
|
||||||
|
}
|
||||||
|
|
||||||
def _latex_to_base64(self, tex):
|
def _latex_to_base64(self, tex):
|
||||||
"""Generates a base64 representation of TeX string"""
|
"""Generates a base64 representation of TeX string"""
|
||||||
|
@ -104,18 +109,24 @@ class LaTeXPreprocessor(markdown.preprocessors.Preprocessor):
|
||||||
with os.fdopen(tmp_file_fd, "w") as tmp_file:
|
with os.fdopen(tmp_file_fd, "w") as tmp_file:
|
||||||
tmp_file.write(self.tex_preamble)
|
tmp_file.write(self.tex_preamble)
|
||||||
tmp_file.write(tex)
|
tmp_file.write(tex)
|
||||||
tmp_file.write('\n\\end{document}')
|
tmp_file.write("\n\\end{document}")
|
||||||
|
|
||||||
# compile LaTeX document. A DVI file is created
|
# compile LaTeX document. A DVI file is created
|
||||||
status = call(('latex -halt-on-error -output-directory={:s} {:s}'
|
status = call(
|
||||||
.format(tmpdir, path)).split(),
|
(
|
||||||
stdout=PIPE, timeout=10)
|
"latex -halt-on-error -output-directory={:s} {:s}".format(tmpdir, path)
|
||||||
|
).split(),
|
||||||
|
stdout=PIPE,
|
||||||
|
timeout=10,
|
||||||
|
)
|
||||||
|
|
||||||
# clean up if the above failed
|
# clean up if the above failed
|
||||||
if status:
|
if status:
|
||||||
self._cleanup(path, err=True)
|
self._cleanup(path, err=True)
|
||||||
raise Exception("Couldn't compile LaTeX document." +
|
raise Exception(
|
||||||
"Please read '%s.log' for more detail." % path)
|
"Couldn't compile LaTeX document."
|
||||||
|
+ "Please read '%s.log' for more detail." % path
|
||||||
|
)
|
||||||
|
|
||||||
# Run dvipng on the generated DVI file. Use tight bounding box.
|
# Run dvipng on the generated DVI file. Use tight bounding box.
|
||||||
# Magnification is set to 1200
|
# Magnification is set to 1200
|
||||||
|
@ -129,8 +140,10 @@ class LaTeXPreprocessor(markdown.preprocessors.Preprocessor):
|
||||||
# clean up if we couldn't make the above work
|
# clean up if we couldn't make the above work
|
||||||
if status:
|
if status:
|
||||||
self._cleanup(path, err=True)
|
self._cleanup(path, err=True)
|
||||||
raise Exception("Couldn't convert LaTeX to image." +
|
raise Exception(
|
||||||
"Please read '%s.log' for more detail." % path)
|
"Couldn't convert LaTeX to image."
|
||||||
|
+ "Please read '%s.log' for more detail." % path
|
||||||
|
)
|
||||||
|
|
||||||
# Read the png and encode the data
|
# Read the png and encode the data
|
||||||
try:
|
try:
|
||||||
|
@ -157,7 +170,7 @@ class LaTeXPreprocessor(markdown.preprocessors.Preprocessor):
|
||||||
def run(self, lines):
|
def run(self, lines):
|
||||||
"""Parses the actual page"""
|
"""Parses the actual page"""
|
||||||
# Checks for the LaTeX header
|
# Checks for the LaTeX header
|
||||||
use_latex = any(line == '[//]: # (latex: 1)' for line in lines)
|
use_latex = any(line == "[//]: # (latex: 1)" for line in lines)
|
||||||
if not use_latex:
|
if not use_latex:
|
||||||
return lines
|
return lines
|
||||||
|
|
||||||
|
@ -165,7 +178,9 @@ class LaTeXPreprocessor(markdown.preprocessors.Preprocessor):
|
||||||
page = "\n".join(lines)
|
page = "\n".join(lines)
|
||||||
|
|
||||||
# Adds a preamble mode
|
# Adds a preamble mode
|
||||||
self.tex_preamble += self.config[("general", "preamble")] + "\n\\begin{document}\n"
|
self.tex_preamble += (
|
||||||
|
self.config[("general", "preamble")] + "\n\\begin{document}\n"
|
||||||
|
)
|
||||||
|
|
||||||
# Figure out our text strings and math-mode strings
|
# Figure out our text strings and math-mode strings
|
||||||
tex_expr = self.math_extract_regex.findall(page)
|
tex_expr = self.math_extract_regex.findall(page)
|
||||||
|
@ -176,7 +191,7 @@ class LaTeXPreprocessor(markdown.preprocessors.Preprocessor):
|
||||||
|
|
||||||
# Parse the expressions
|
# Parse the expressions
|
||||||
new_cache = {}
|
new_cache = {}
|
||||||
new_page = ''
|
new_page = ""
|
||||||
n_multiline_expressions = 0
|
n_multiline_expressions = 0
|
||||||
|
|
||||||
while page:
|
while page:
|
||||||
|
@ -200,21 +215,25 @@ class LaTeXPreprocessor(markdown.preprocessors.Preprocessor):
|
||||||
new_cache[tex_hash] = data
|
new_cache[tex_hash] = data
|
||||||
|
|
||||||
if is_multiline and n_multiline_expressions > 0:
|
if is_multiline and n_multiline_expressions > 0:
|
||||||
new_page += '</p>'
|
new_page += "</p>"
|
||||||
new_page += (multiline_img_expr if is_multiline else img_expr) % ('true', expr, tex_hash, data)
|
new_page += (multiline_img_expr if is_multiline else img_expr) % (
|
||||||
|
"true",
|
||||||
|
tex_hash,
|
||||||
|
data,
|
||||||
|
)
|
||||||
|
|
||||||
if is_multiline:
|
if is_multiline:
|
||||||
new_page += '<p>'
|
new_page += "<p>"
|
||||||
n_multiline_expressions += 1
|
n_multiline_expressions += 1
|
||||||
|
|
||||||
page = m.group(5)
|
page = m.group(5)
|
||||||
|
|
||||||
if n_multiline_expressions > 0:
|
if n_multiline_expressions > 0:
|
||||||
new_page += '</p>'
|
new_page += "</p>"
|
||||||
|
|
||||||
# Cache our data
|
# Cache our data
|
||||||
self.cached.update(new_cache)
|
self.cached.update(new_cache)
|
||||||
with open(cache_file, 'w') as f:
|
with open(cache_file, "w") as f:
|
||||||
json.dump(self.cached, f)
|
json.dump(self.cached, f)
|
||||||
|
|
||||||
# Make sure to re-split the lines
|
# Make sure to re-split the lines
|
||||||
|
@ -231,6 +250,6 @@ class MarkdownLatex(markdown.Extension):
|
||||||
def extendMarkdown(self, md):
|
def extendMarkdown(self, md):
|
||||||
md.preprocessors.register(
|
md.preprocessors.register(
|
||||||
LaTeXPreprocessor(self),
|
LaTeXPreprocessor(self),
|
||||||
'latex',
|
"latex",
|
||||||
1,
|
1,
|
||||||
)
|
)
|
||||||
|
|
|
@ -1,5 +1,7 @@
|
||||||
import os
|
import os
|
||||||
|
import re
|
||||||
from typing import Optional
|
from typing import Optional
|
||||||
|
from urllib.parse import urljoin
|
||||||
|
|
||||||
from flask import (
|
from flask import (
|
||||||
jsonify,
|
jsonify,
|
||||||
|
@ -97,8 +99,8 @@ def manifest_route():
|
||||||
"scope": "",
|
"scope": "",
|
||||||
"orientation": "portrait",
|
"orientation": "portrait",
|
||||||
"display": "standalone",
|
"display": "standalone",
|
||||||
"theme_color": "#ffffff",
|
"theme_color": "#000000",
|
||||||
"background_color": "#000000",
|
"background_color": "#ffffff",
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -115,8 +117,12 @@ def article_route(article: str):
|
||||||
|
|
||||||
@app.route("/rss", methods=["GET"])
|
@app.route("/rss", methods=["GET"])
|
||||||
def rss_route():
|
def rss_route():
|
||||||
pages = app.get_pages(with_content=True, skip_header=True, skip_html_head=True)
|
short_description = "short" in request.args or config.short_feed
|
||||||
short_description = "short" in request.args
|
pages = app.get_pages(
|
||||||
|
with_content=not short_description,
|
||||||
|
skip_header=True,
|
||||||
|
skip_html_head=True,
|
||||||
|
)
|
||||||
|
|
||||||
return Response(
|
return Response(
|
||||||
"""<?xml version="1.0" encoding="UTF-8" ?>
|
"""<?xml version="1.0" encoding="UTF-8" ?>
|
||||||
|
@ -149,27 +155,36 @@ def rss_route():
|
||||||
),
|
),
|
||||||
items="\n\n".join(
|
items="\n\n".join(
|
||||||
[
|
[
|
||||||
|
(
|
||||||
"""
|
"""
|
||||||
<item>
|
<item>
|
||||||
<title>{title}</title>
|
<title>{title}</title>
|
||||||
<link>{base_link}{link}</link>
|
<link>{base_link}{link}</link>
|
||||||
<pubDate>{published}</pubDate>
|
<pubDate>{published}</pubDate>
|
||||||
<description><![CDATA[{content}]]></description>
|
<description><![CDATA[{content}]]></description>
|
||||||
<media:content medium="image" url="{base_link}{image}" width="200" height="150" />
|
<media:content medium="image" url="{image}" width="200" height="150" />
|
||||||
</item>
|
</item>
|
||||||
""".format(
|
"""
|
||||||
|
).format(
|
||||||
base_link=config.link,
|
base_link=config.link,
|
||||||
title=page.get("title", "[No Title]"),
|
title=page.get("title", "[No Title]"),
|
||||||
link=page.get("uri", ""),
|
link=page.get("uri", ""),
|
||||||
published=page["published"].strftime(
|
published=(
|
||||||
"%a, %d %b %Y %H:%M:%S GMT"
|
page["published"].strftime("%a, %d %b %Y %H:%M:%S GMT")
|
||||||
)
|
|
||||||
if "published" in page
|
if "published" in page
|
||||||
else "",
|
else ""
|
||||||
content=page.get("description", "")
|
),
|
||||||
|
content=(
|
||||||
|
page.get("description", "")
|
||||||
if short_description
|
if short_description
|
||||||
else page.get("content", ""),
|
else page.get("content", "")
|
||||||
image=page.get("image", ""),
|
),
|
||||||
|
image=(
|
||||||
|
urljoin(config.link, page["image"])
|
||||||
|
if page.get("image")
|
||||||
|
and not re.search(r"^https?://", page["image"])
|
||||||
|
else page.get("image", "")
|
||||||
|
),
|
||||||
)
|
)
|
||||||
for _, page in pages
|
for _, page in pages
|
||||||
]
|
]
|
||||||
|
|
|
@ -6,19 +6,26 @@ main .content {
|
||||||
|
|
||||||
main .content p,
|
main .content p,
|
||||||
main .content ul {
|
main .content ul {
|
||||||
|
font-family: Lora,'Palatino Linotype','Book Antiqua','New York','DejaVu serif',serif;
|
||||||
text-align: justify;
|
text-align: justify;
|
||||||
overflow-wrap: break-word;
|
overflow-wrap: break-word;
|
||||||
word-break: break-word;
|
word-break: break-word;
|
||||||
|
font-size: 0.95em;
|
||||||
}
|
}
|
||||||
|
|
||||||
main .content code, .codehilite {
|
main .content code, .codehilite {
|
||||||
font-size: .85em;
|
font-size: 0.9em;
|
||||||
}
|
}
|
||||||
|
|
||||||
@media screen and (max-width: 767px) {
|
@media screen and (max-width: 767px) {
|
||||||
main {
|
main {
|
||||||
font-size: 0.9em;
|
font-size: 0.9em;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
main p {
|
||||||
|
font-size: 0.9em;
|
||||||
|
padding: 0 1em;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
a:hover {
|
a:hover {
|
||||||
|
|
|
@ -2,7 +2,8 @@ html {
|
||||||
height: -webkit-fill-available;
|
height: -webkit-fill-available;
|
||||||
height: -moz-available;
|
height: -moz-available;
|
||||||
font-size: 20px;
|
font-size: 20px;
|
||||||
font-family: -apple-system, "Segoe UI", Roboto, Oxygen, Ubuntu, Cantarell, "Fira Sans", "Open Sans", "Droid Sans", "Helvetica Neue", Helvetica, Arial, sans-serif;
|
/* font-family: Lora, "Palatino Linotype", "Book Antiqua", "New York", "DejaVu serif", serif; */
|
||||||
|
font-family: system-ui, -apple-system, "Segoe UI", Roboto, Helvetica, Arial, sans-serif, "Apple Color Emoji", "Segoe UI Emoji";
|
||||||
font-weight: 400;
|
font-weight: 400;
|
||||||
text-rendering: optimizeLegibility;
|
text-rendering: optimizeLegibility;
|
||||||
}
|
}
|
||||||
|
@ -119,6 +120,10 @@ h2 {
|
||||||
line-height: 1.1em;
|
line-height: 1.1em;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
h1, h2, h3, h4, h5, h6 {
|
||||||
|
font-family: -apple-system, "Segoe UI", Roboto, Oxygen, Ubuntu, Cantarell, "Fira Sans", "Open Sans", "Droid Sans", "Helvetica Neue", Helvetica, Arial, sans-serif;
|
||||||
|
}
|
||||||
|
|
||||||
footer {
|
footer {
|
||||||
width: 100%;
|
width: 100%;
|
||||||
font-size: .65em;
|
font-size: .65em;
|
||||||
|
@ -129,3 +134,7 @@ footer {
|
||||||
text-align: center;
|
text-align: center;
|
||||||
box-shadow: 1px -2px 2px 0 #bbb;
|
box-shadow: 1px -2px 2px 0 #bbb;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
.index {
|
||||||
|
font-family: -apple-system, "Segoe UI", Roboto, Oxygen, Ubuntu, Cantarell, "Fira Sans", "Open Sans", "Droid Sans", "Helvetica Neue", Helvetica, Arial, sans-serif;
|
||||||
|
}
|
||||||
|
|
|
@ -6,7 +6,6 @@
|
||||||
|
|
||||||
<!-- PWA & Viewport -->
|
<!-- PWA & Viewport -->
|
||||||
<meta name="viewport" content="width=device-width, initial-scale=1, maximum-scale=1" />
|
<meta name="viewport" content="width=device-width, initial-scale=1, maximum-scale=1" />
|
||||||
<meta name="viewport" content="uc-fitscreen=yes"/>
|
|
||||||
<meta name="description" content="{{ config.description }}">
|
<meta name="description" content="{{ config.description }}">
|
||||||
<link rel="manifest" href="/manifest.json">
|
<link rel="manifest" href="/manifest.json">
|
||||||
<!-- Android PWA -->
|
<!-- Android PWA -->
|
||||||
|
@ -17,8 +16,6 @@
|
||||||
<meta name="apple-mobile-web-app-title" content="{{ config.title }}">
|
<meta name="apple-mobile-web-app-title" content="{{ config.title }}">
|
||||||
<meta name="apple-mobile-web-app-capable" content="yes">
|
<meta name="apple-mobile-web-app-capable" content="yes">
|
||||||
<meta name="apple-mobile-web-app-status-bar-style" content="default">
|
<meta name="apple-mobile-web-app-status-bar-style" content="default">
|
||||||
<!-- Layout mode -->
|
|
||||||
<meta name="layoutmode" content="fitscreen/standard">
|
|
||||||
<!-- Orientation -->
|
<!-- Orientation -->
|
||||||
<meta name="screen-orientation" content="portrait">
|
<meta name="screen-orientation" content="portrait">
|
||||||
<!-- RSS feed -->
|
<!-- RSS feed -->
|
||||||
|
|
|
@ -1,5 +1,5 @@
|
||||||
[bumpversion]
|
[bumpversion]
|
||||||
current_version = 0.2.14
|
current_version = 0.2.31
|
||||||
commit = True
|
commit = True
|
||||||
tag = True
|
tag = True
|
||||||
|
|
||||||
|
|
38
setup.py
38
setup.py
|
@ -5,38 +5,38 @@ from setuptools import setup, find_packages
|
||||||
|
|
||||||
|
|
||||||
def readfile(file):
|
def readfile(file):
|
||||||
with open(file, 'r') as f:
|
with open(file, "r") as f:
|
||||||
return f.read()
|
return f.read()
|
||||||
|
|
||||||
|
|
||||||
setup(
|
setup(
|
||||||
name='madblog',
|
name="madblog",
|
||||||
version='0.2.14',
|
version="0.2.31",
|
||||||
author='Fabio Manganiello',
|
author="Fabio Manganiello",
|
||||||
author_email='info@fabiomanganiello.com',
|
author_email="info@fabiomanganiello.com",
|
||||||
description='A minimal platform for Markdown-based blogs',
|
description="A minimal platform for Markdown-based blogs",
|
||||||
license='MIT',
|
license="MIT",
|
||||||
python_requires='>= 3.8',
|
python_requires=">= 3.8",
|
||||||
keywords='blog markdown',
|
keywords="blog markdown",
|
||||||
url='https://git.platypush.tech/blacklight/madblog',
|
url="https://git.platypush.tech/blacklight/madblog",
|
||||||
packages=find_packages(include=['madblog']),
|
packages=find_packages(include=["madblog"]),
|
||||||
include_package_data=True,
|
include_package_data=True,
|
||||||
entry_points={
|
entry_points={
|
||||||
'console_scripts': [
|
"console_scripts": [
|
||||||
'madblog=madblog.cli:run',
|
"madblog=madblog.cli:run",
|
||||||
],
|
],
|
||||||
},
|
},
|
||||||
long_description=readfile('README.md'),
|
long_description=readfile("README.md"),
|
||||||
long_description_content_type='text/markdown',
|
long_description_content_type="text/markdown",
|
||||||
classifiers=[
|
classifiers=[
|
||||||
"Topic :: Utilities",
|
"Topic :: Utilities",
|
||||||
"License :: OSI Approved :: MIT License",
|
"License :: OSI Approved :: MIT License",
|
||||||
"Development Status :: 4 - Beta",
|
"Development Status :: 4 - Beta",
|
||||||
],
|
],
|
||||||
install_requires=[
|
install_requires=[
|
||||||
'flask',
|
"flask",
|
||||||
'markdown',
|
"markdown",
|
||||||
'pygments',
|
"pygments",
|
||||||
'pyyaml',
|
"pyyaml",
|
||||||
],
|
],
|
||||||
)
|
)
|
||||||
|
|
Loading…
Reference in a new issue