* feat: extractor for pholder.com
Closes #2568
* feat[pholder]: support gallery_id properly and tags
* doc[text.nameext_from_name]: minor typo in docstring
* remove '__init__' & 'request' methods and 'json' import
* use 'text.nameext_from_url' to ensure a 'filename' value
* fix 'imgur' links by disabling auto-Referer
* fix 'data["id"].partition()' call
'partition' returns 3 elements
* use 'item["_source"]' data directly
* remove unused supportedsites overwrite
* catch all exceptions in '_thumb_resolution'
fixes "KeyError: 'width'"
* use 'author' name for user folders
---------
Co-authored-by: Mike Fährmann <mike_faehrmann@web.de>
350 lines
8.8 KiB
Python
350 lines
8.8 KiB
Python
# -*- coding: utf-8 -*-
|
|
|
|
# Copyright 2015-2026 Mike Fährmann
|
|
#
|
|
# This program is free software; you can redistribute it and/or modify
|
|
# it under the terms of the GNU General Public License version 2 as
|
|
# published by the Free Software Foundation.
|
|
|
|
"""Collection of functions that work on strings/text"""
|
|
|
|
import html
|
|
import urllib.parse
|
|
import re as re_module
|
|
|
|
try:
|
|
re_compile = re_module._compiler.compile
|
|
except AttributeError:
|
|
re_compile = re_module.sre_compile.compile
|
|
|
|
HTML_RE = re_compile(r"<[^>]+>")
|
|
PATTERN_CACHE = {}
|
|
|
|
|
|
def re(pattern):
|
|
"""Compile a regular expression pattern"""
|
|
try:
|
|
return PATTERN_CACHE[pattern]
|
|
except KeyError:
|
|
p = PATTERN_CACHE[pattern] = re_compile(pattern)
|
|
return p
|
|
|
|
|
|
def remove_html(txt, repl=" ", sep=" "):
|
|
"""Remove html-tags from a string"""
|
|
try:
|
|
txt = HTML_RE.sub(repl, txt)
|
|
except TypeError:
|
|
return ""
|
|
if sep:
|
|
return sep.join(txt.split())
|
|
return txt.strip()
|
|
|
|
|
|
def split_html(txt):
|
|
"""Split input string by HTML tags"""
|
|
try:
|
|
return [
|
|
unescape(x).strip()
|
|
for x in HTML_RE.split(txt)
|
|
if x and not x.isspace()
|
|
]
|
|
except TypeError:
|
|
return []
|
|
|
|
|
|
def slugify(value):
|
|
"""Convert a string to a URL slug
|
|
|
|
Adapted from:
|
|
https://github.com/django/django/blob/master/django/utils/text.py
|
|
"""
|
|
value = re(r"[^\w\s-]").sub("", str(value).lower())
|
|
return re(r"[-\s]+").sub("-", value).strip("-_")
|
|
|
|
|
|
def sanitize_whitespace(value):
|
|
"""Replace all whitespace characters with a single space"""
|
|
return re(r"\s+").sub(" ", value.strip())
|
|
|
|
|
|
def ensure_http_scheme(url, scheme="https://"):
|
|
"""Prepend 'scheme' to 'url' if it doesn't have one"""
|
|
if url and not url.startswith(("https://", "http://")):
|
|
return scheme + url.lstrip("/:")
|
|
return url
|
|
|
|
|
|
def root_from_url(url, scheme="https://"):
|
|
"""Extract scheme and domain from a URL"""
|
|
if not url.startswith(("https://", "http://")):
|
|
try:
|
|
return scheme + url[:url.index("/")]
|
|
except ValueError:
|
|
return scheme + url
|
|
try:
|
|
return url[:url.index("/", 8)]
|
|
except ValueError:
|
|
return url
|
|
|
|
|
|
def filename_from_url(url):
|
|
"""Extract the last part of an URL to use as a filename"""
|
|
try:
|
|
return url.partition("?")[0].rpartition("/")[2]
|
|
except Exception:
|
|
return ""
|
|
|
|
|
|
def ext_from_url(url):
|
|
"""Extract the filename extension of an URL"""
|
|
name, _, ext = filename_from_url(url).rpartition(".")
|
|
return ext.lower() if name else ""
|
|
|
|
|
|
def nameext_from_url(url, data=None):
|
|
"""Extract the last part of an URL and fill 'data' accordingly"""
|
|
if data is None:
|
|
data = {}
|
|
|
|
filename = unquote(filename_from_url(url))
|
|
name, _, ext = filename.rpartition(".")
|
|
if name and len(ext) <= 16:
|
|
data["filename"] = name
|
|
data["extension"] = ext.lower()
|
|
else:
|
|
data["filename"] = filename
|
|
data["extension"] = ""
|
|
|
|
return data
|
|
|
|
|
|
def nameext_from_name(filename, data=None):
|
|
"""Extract the last part of a file name and fill 'data' accordingly"""
|
|
if data is None:
|
|
data = {}
|
|
|
|
name, _, ext = filename.rpartition(".")
|
|
if name and len(ext) <= 16:
|
|
data["filename"] = name
|
|
data["extension"] = ext.lower()
|
|
else:
|
|
data["filename"] = filename
|
|
data["extension"] = ""
|
|
|
|
return data
|
|
|
|
|
|
def extract(txt, begin, end, pos=None):
|
|
"""Extract the text between 'begin' and 'end' from 'txt'
|
|
|
|
Args:
|
|
txt: String to search in
|
|
begin: First string to be searched for
|
|
end: Second string to be searched for after 'begin'
|
|
pos: Starting position for searches in 'txt'
|
|
|
|
Returns:
|
|
The string between the two search-strings 'begin' and 'end' beginning
|
|
with position 'pos' in 'txt' as well as the position after 'end'.
|
|
|
|
If at least one of 'begin' or 'end' is not found, None and the original
|
|
value of 'pos' is returned
|
|
|
|
Examples:
|
|
extract("abcde", "b", "d") -> "c" , 4
|
|
extract("abcde", "b", "d", 3) -> None, 3
|
|
"""
|
|
try:
|
|
first = txt.index(begin, pos) + len(begin)
|
|
last = txt.index(end, first)
|
|
return txt[first:last], last+len(end)
|
|
except Exception:
|
|
return None, 0 if pos is None else pos
|
|
|
|
|
|
def extr(txt, begin, end, default=""):
|
|
"""Stripped-down version of 'extract()'"""
|
|
try:
|
|
first = txt.index(begin) + len(begin)
|
|
return txt[first:txt.index(end, first)]
|
|
except Exception:
|
|
return default
|
|
|
|
|
|
def rextract(txt, begin, end, pos=None):
|
|
try:
|
|
lbeg = len(begin)
|
|
first = txt.rindex(begin, None, pos)
|
|
last = txt.index(end, first + lbeg)
|
|
return txt[first + lbeg:last], first
|
|
except Exception:
|
|
return None, -1 if pos is None else pos
|
|
|
|
|
|
def rextr(txt, begin, end, pos=None, default=""):
|
|
"""Stripped-down version of 'rextract()'"""
|
|
try:
|
|
first = txt.rindex(begin, None, pos) + len(begin)
|
|
return txt[first:txt.index(end, first)]
|
|
except Exception:
|
|
return default
|
|
|
|
|
|
def extract_all(txt, rules, pos=None, values=None):
|
|
"""Calls extract for each rule and returns the result in a dict"""
|
|
if values is None:
|
|
values = {}
|
|
for key, begin, end in rules:
|
|
result, pos = extract(txt, begin, end, pos)
|
|
if key:
|
|
values[key] = result
|
|
return values, 0 if pos is None else pos
|
|
|
|
|
|
def extract_iter(txt, begin, end, pos=None):
|
|
"""Yield values that would be returned by repeated calls of extract()"""
|
|
try:
|
|
index = txt.index
|
|
lbeg = len(begin)
|
|
lend = len(end)
|
|
while True:
|
|
first = index(begin, pos) + lbeg
|
|
last = index(end, first)
|
|
pos = last + lend
|
|
yield txt[first:last]
|
|
except Exception:
|
|
return
|
|
|
|
|
|
def extract_from(txt, pos=None, default=""):
|
|
"""Returns a function object that extracts from 'txt'"""
|
|
def extr(begin, end, index=txt.index, txt=txt):
|
|
nonlocal pos
|
|
try:
|
|
first = index(begin, pos) + len(begin)
|
|
last = index(end, first)
|
|
pos = last + len(end)
|
|
return txt[first:last]
|
|
except Exception:
|
|
return default
|
|
return extr
|
|
|
|
|
|
extract_urls = re(r"https?://[^\s\"'<>\\]+").findall
|
|
|
|
|
|
def parse_unicode_escapes(txt):
|
|
"""Convert JSON Unicode escapes in 'txt' into actual characters"""
|
|
if "\\u" in txt:
|
|
return re(r"\\u([0-9a-fA-F]{4})").sub(_hex_to_char, txt)
|
|
return txt
|
|
|
|
|
|
def _hex_to_char(match):
|
|
return chr(int(match[1], 16))
|
|
|
|
|
|
def parse_bytes(value, default=0, suffixes="bkmgtp"):
|
|
"""Convert a bytes-amount ("500k", "2.5M", ...) to int"""
|
|
if not value:
|
|
return default
|
|
|
|
value = str(value).strip()
|
|
last = value[-1].lower()
|
|
|
|
if last in suffixes:
|
|
mul = 1024 ** suffixes.index(last)
|
|
value = value[:-1]
|
|
else:
|
|
mul = 1
|
|
|
|
try:
|
|
return round(float(value) * mul)
|
|
except ValueError:
|
|
return default
|
|
|
|
|
|
def parse_int(value, default=0):
|
|
"""Convert 'value' to int"""
|
|
if not value:
|
|
return default
|
|
try:
|
|
return int(value)
|
|
except Exception:
|
|
return default
|
|
|
|
|
|
def parse_float(value, default=0.0):
|
|
"""Convert 'value' to float"""
|
|
if not value:
|
|
return default
|
|
try:
|
|
return float(value)
|
|
except Exception:
|
|
return default
|
|
|
|
|
|
def parse_query(qs, empty=False):
|
|
"""Parse a query string into name-value pairs
|
|
|
|
Ignore values whose name has been seen before
|
|
"""
|
|
if not qs:
|
|
return {}
|
|
|
|
result = {}
|
|
try:
|
|
for name_value in qs.split("&"):
|
|
name, eq, value = name_value.partition("=")
|
|
if eq or empty:
|
|
name = unquote(name.replace("+", " "))
|
|
if name not in result:
|
|
result[name] = unquote(value.replace("+", " "))
|
|
except Exception:
|
|
pass
|
|
return result
|
|
|
|
|
|
def parse_query_list(qs, as_list=()):
|
|
"""Parse a query string into name-value pairs
|
|
|
|
Combine values of names in 'as_list' into lists
|
|
"""
|
|
if not qs:
|
|
return {}
|
|
|
|
result = {}
|
|
try:
|
|
for name_value in qs.split("&"):
|
|
name, eq, value = name_value.partition("=")
|
|
if eq:
|
|
name = unquote(name.replace("+", " "))
|
|
value = unquote(value.replace("+", " "))
|
|
if name in as_list:
|
|
if name in result:
|
|
result[name].append(value)
|
|
else:
|
|
result[name] = [value]
|
|
elif name not in result:
|
|
result[name] = value
|
|
except Exception:
|
|
pass
|
|
return result
|
|
|
|
|
|
def build_query(params):
|
|
return "&".join([
|
|
f"{quote(name)}={quote(value)}"
|
|
for name, value in params.items()
|
|
])
|
|
|
|
|
|
urljoin = urllib.parse.urljoin
|
|
|
|
quote = urllib.parse.quote
|
|
unquote = urllib.parse.unquote
|
|
|
|
escape = html.escape
|
|
unescape = html.unescape
|