[py] Fix styling

This commit is contained in:
Nemo
2022-10-16 13:59:39 +05:30
committed by Nemo
parent 9a35436353
commit fac9620241
9 changed files with 283 additions and 249 deletions

View File

@@ -6,85 +6,92 @@ import re
from html.parser import HTMLParser from html.parser import HTMLParser
URLS = [ URLS = [
"https://support.apple.com/en-us/HT201222", # latest "https://support.apple.com/en-us/HT201222", # latest
"https://support.apple.com/kb/HT213078", # 2018-2019 "https://support.apple.com/kb/HT213078", # 2018-2019
"https://support.apple.com/kb/HT213077", # 2016-2017 "https://support.apple.com/kb/HT213077", # 2016-2017
"https://support.apple.com/kb/HT209441", # 2015 "https://support.apple.com/kb/HT209441", # 2015
"https://support.apple.com/kb/HT205762", # 2014 "https://support.apple.com/kb/HT205762", # 2014
"https://support.apple.com/kb/HT205759", # 2013 "https://support.apple.com/kb/HT205759", # 2013
"https://support.apple.com/kb/HT204611", # 2011 to 2012 "https://support.apple.com/kb/HT204611", # 2011 to 2012
"https://support.apple.com/kb/HT5165", # 2010 "https://support.apple.com/kb/HT5165", # 2010
"https://support.apple.com/kb/HT4218", # 2008-2009 "https://support.apple.com/kb/HT4218", # 2008-2009
"https://support.apple.com/kb/HT1263", # 2005-2007 "https://support.apple.com/kb/HT1263", # 2005-2007
] ]
# If you are changing these, please # If you are changing these, please
# use https://www.toptal.com/developers/hastebin/mikahukube.txt as your corpus # use https://www.toptal.com/developers/hastebin/mikahukube.txt as your corpus
# to validate your changes # to validate your changes
CONFIG = { CONFIG = {
"macos": [ "macos": [
# This covers Sierra and beyond # This covers Sierra and beyond
r"macOS [\w ]+ (?P<version>\d{2}(?:\.\d+)+)", r"macOS [\w ]+ (?P<version>\d{2}(?:\.\d+)+)",
# This covers Mavericks - El Capitan # This covers Mavericks - El Capitan
r"OS X [\w ]+ v?(?P<version>\d{2}(?:\.\d+)+)", r"OS X [\w ]+ v?(?P<version>\d{2}(?:\.\d+)+)",
# This covers even older versions (OS X) # This covers even older versions (OS X)
r"^Mac OS X [\w ]+ v?(?P<version>\d{2}(?:\.\d+)+)", r"^Mac OS X [\w ]+ v?(?P<version>\d{2}(?:\.\d+)+)",
], ],
"ios": [ "ios": [
r"iOS (?P<version>\d+(?:)(?:\.\d+)+)", r"iOS (?P<version>\d+(?:)(?:\.\d+)+)",
r"iPhone v?(?P<version>\d+(?:)(?:\.\d+)+)" r"iPhone v?(?P<version>\d+(?:)(?:\.\d+)+)",
], ],
"ipados": [ "ipados": [r"iPadOS (?P<version>\d+(?:)(?:\.\d+)+)"],
r"iPadOS (?P<version>\d+(?:)(?:\.\d+)+)" "watchos": [r"watchOS (?P<version>\d+(?:)(?:\.\d+)+)"],
],
"watchos": [
r"watchOS (?P<version>\d+(?:)(?:\.\d+)+)"
]
} }
release_lists = {k: {} for k in CONFIG.keys()} release_lists = {k: {} for k in CONFIG.keys()}
print("::group::apple") print("::group::apple")
def parse_date(input): def parse_date(input):
d,m,y = input.strip().split(" ") d, m, y = input.strip().split(" ")
m=m[0:3].lower() m = m[0:3].lower()
return datetime.datetime.strptime("%s %s %s" % (d,m,y), "%d %b %Y") return datetime.datetime.strptime("%s %s %s" % (d, m, y), "%d %b %Y")
for url in URLS: for url in URLS:
with urllib.request.urlopen(url, data=None, timeout=5) as response: with urllib.request.urlopen(url, data=None, timeout=5) as response:
soup = BeautifulSoup(response, features="html5lib") soup = BeautifulSoup(response, features="html5lib")
table = soup.find(id='tableWraper') table = soup.find(id="tableWraper")
for tr in reversed(table.findAll('tr')[1:]): for tr in reversed(table.findAll("tr")[1:]):
td_list = tr.findAll('td') td_list = tr.findAll("td")
version_text = td_list[0].get_text() version_text = td_list[0].get_text()
for key,regexes in CONFIG.items(): for key, regexes in CONFIG.items():
for regex in regexes: for regex in regexes:
matches = re.findall(regex, version_text, re.MULTILINE) matches = re.findall(regex, version_text, re.MULTILINE)
if matches: if matches:
for version in matches: for version in matches:
abs_date = None abs_date = None
try: try:
abs_date = parse_date(td_list[2].get_text()) abs_date = parse_date(td_list[2].get_text())
print_date = abs_date.strftime("%Y-%m-%d") print_date = abs_date.strftime("%Y-%m-%d")
# Only update the date if we are adding first time # Only update the date if we are adding first time
# or if the date is lower # or if the date is lower
if version not in release_lists[key]: if version not in release_lists[key]:
release_lists[key][version] = abs_date release_lists[key][version] = abs_date
print("%s-%s: %s" % (key, version, print_date)) print("%s-%s: %s" % (key, version, print_date))
elif release_lists[key][version] < abs_date: elif release_lists[key][version] < abs_date:
print("%s-%s: %s [IGNORED]" % (key, version, print_date)) print(
elif release_lists[key][version] > abs_date: "%s-%s: %s [IGNORED]"
# This is a lower date, so we mark it with a bang % (key, version, print_date)
print("%s-%s: %s [UPDATED]" % (key, version, print_date)) )
release_lists[key][version] = abs_date elif release_lists[key][version] > abs_date:
except ValueError as e: # This is a lower date, so we mark it with a bang
print("%s-%s Failed to parse Date (%s)" % (key, version, td_list[2].get_text())) print(
next "%s-%s: %s [UPDATED]"
% (key, version, print_date)
)
release_lists[key][version] = abs_date
except ValueError as e:
print(
"%s-%s Failed to parse Date (%s)"
% (key, version, td_list[2].get_text())
)
next
for k in CONFIG.keys(): for k in CONFIG.keys():
with open("releases/%s.json" % k, 'w') as f: with open("releases/%s.json" % k, "w") as f:
data = {v: d.strftime("%Y-%m-%d") for v,d in release_lists[k].items()} data = {v: d.strftime("%Y-%m-%d") for v, d in release_lists[k].items()}
f.write(json.dumps(data, indent=2)) f.write(json.dumps(data, indent=2))
print("::endgroup::") print("::endgroup::")

View File

@@ -10,19 +10,23 @@ from html.parser import HTMLParser
from liquid import Template from liquid import Template
# Same as used in Ruby (update.rb) # Same as used in Ruby (update.rb)
DEFAULT_TAG_TEMPLATE = "{{major}}{% if minor %}.{{minor}}{% if patch %}.{{patch}}{%endif%}{%endif%}" DEFAULT_TAG_TEMPLATE = (
"{{major}}{% if minor %}.{{minor}}{% if patch %}.{{patch}}{%endif%}{%endif%}"
)
def get_versions_from_headline(regex, headline, template): def get_versions_from_headline(regex, headline, template):
if not isinstance(regex, list): if not isinstance(regex, list):
regex = [regex] regex = [regex]
for r in regex: for r in regex:
matches = re.match(r.strip(), headline) matches = re.match(r.strip(), headline)
if matches: if matches:
match_data = matches.groupdict() match_data = matches.groupdict()
version_string = template.render(**match_data) version_string = template.render(**match_data)
return version_string.split("\n") return version_string.split("\n")
return {}
return {}
def fetch_releases(distrowatch_id, regex, template): def fetch_releases(distrowatch_id, regex, template):
releases = {} releases = {}
@@ -34,34 +38,37 @@ def fetch_releases(distrowatch_id, regex, template):
headline = table.select_one("td.NewsHeadline a[href]").get_text().strip() headline = table.select_one("td.NewsHeadline a[href]").get_text().strip()
date = table.select_one("td.NewsDate").get_text() date = table.select_one("td.NewsDate").get_text()
for v in get_versions_from_headline(regex, headline, l_template): for v in get_versions_from_headline(regex, headline, l_template):
print("%s: %s" % (v, date)) print("%s: %s" % (v, date))
releases[v] = date releases[v] = date
return releases return releases
def update_releases(product_filter=None): def update_releases(product_filter=None):
for product_file in glob("website/products/*.md"): for product_file in glob("website/products/*.md"):
product_name = os.path.splitext(os.path.basename(product_file))[0] product_name = os.path.splitext(os.path.basename(product_file))[0]
if product_filter and product_name != product_filter: if product_filter and product_name != product_filter:
continue continue
with open(product_file, "r") as f: with open(product_file, "r") as f:
data = frontmatter.load(f) data = frontmatter.load(f)
if "auto" in data: if "auto" in data:
for config in data["auto"]: for config in data["auto"]:
for key, d_id in config.items(): for key, d_id in config.items():
if key == "distrowatch": if key == "distrowatch":
update_product(product_name, config) update_product(product_name, config)
def update_product(product_name, config): def update_product(product_name, config):
t = config.get("template", DEFAULT_TAG_TEMPLATE) t = config.get("template", DEFAULT_TAG_TEMPLATE)
if "regex" in config: if "regex" in config:
print("::group::%s" % product_name) print("::group::%s" % product_name)
r = fetch_releases(config['distrowatch'], config["regex"], t) r = fetch_releases(config["distrowatch"], config["regex"], t)
with open("releases/%s.json" % product_name, "w") as f: with open("releases/%s.json" % product_name, "w") as f:
f.write(json.dumps(r, indent=2)) f.write(json.dumps(r, indent=2))
print("::endgroup::") print("::endgroup::")
if __name__ == "__main__": if __name__ == "__main__":
if len(sys.argv) > 1: if len(sys.argv) > 1:
update_releases(sys.argv[1]) update_releases(sys.argv[1])
else: else:
update_releases() update_releases()

View File

@@ -3,19 +3,19 @@ import re
import urllib.request import urllib.request
# https://regex101.com/r/1JCnFC/1 # https://regex101.com/r/1JCnFC/1
REGEX = r'^(\d{4})\/(\d{2})\/(\d{2})\s+:\s+(\d+\.\d+\.\d.?)$' REGEX = r"^(\d{4})\/(\d{2})\/(\d{2})\s+:\s+(\d+\.\d+\.\d.?)$"
list = {} list = {}
for i in range(17, 27): for i in range(17, 27):
url = "https://www.haproxy.org/download/%s/src/CHANGELOG" % (i/10) url = "https://www.haproxy.org/download/%s/src/CHANGELOG" % (i / 10)
with urllib.request.urlopen(url) as response: with urllib.request.urlopen(url) as response:
for line in response: for line in response:
m = re.match(REGEX, line.decode('utf-8')) m = re.match(REGEX, line.decode("utf-8"))
if m: if m:
year,month,date,version = m.groups() year, month, date, version = m.groups()
abs_date = "%s-%s-%s" % (year, month, date) abs_date = "%s-%s-%s" % (year, month, date)
list[version] = abs_date list[version] = abs_date
with open('releases/haproxy.json', 'w') as f: with open("releases/haproxy.json", "w") as f:
f.write(json.dumps(list, indent=2)) f.write(json.dumps(list, indent=2))

View File

@@ -6,48 +6,58 @@ import frontmatter
import urllib.request import urllib.request
import datetime import datetime
def fetch_releases(package_identifier): def fetch_releases(package_identifier):
groupId, artifactId = package_identifier.split("/") groupId, artifactId = package_identifier.split("/")
releases = {} releases = {}
start = 0 start = 0
while True: while True:
url = "https://search.maven.org/solrsearch/select?q=g:%s+AND+a:%s&core=gav&rows=100&wt=json&start=%s" % (groupId, artifactId, start) url = (
with urllib.request.urlopen(url, data=None, timeout=5) as response: "https://search.maven.org/solrsearch/select?q=g:%s+AND+a:%s&core=gav&rows=100&wt=json&start=%s"
data = json.load(response) % (groupId, artifactId, start)
for row in data['response']['docs']: )
date = datetime.datetime.utcfromtimestamp(row['timestamp'] / 1000) with urllib.request.urlopen(url, data=None, timeout=5) as response:
version = row['v'] data = json.load(response)
if not any(exception in version for exception in ['alpha', 'beta', 'nightly', 'rc', '-M']): for row in data["response"]["docs"]:
abs_date = date.strftime("%Y-%m-%d") date = datetime.datetime.utcfromtimestamp(row["timestamp"] / 1000)
releases[version] = abs_date version = row["v"]
print("%s: %s" % (version, abs_date)) if not any(
start+=100 exception in version
if data['response']['numFound'] <= start: for exception in ["alpha", "beta", "nightly", "rc", "-M"]
break ):
abs_date = date.strftime("%Y-%m-%d")
releases[version] = abs_date
print("%s: %s" % (version, abs_date))
start += 100
if data["response"]["numFound"] <= start:
break
return releases return releases
def update_releases(product_filter=None): def update_releases(product_filter=None):
for product_file in glob("website/products/*.md"): for product_file in glob("website/products/*.md"):
product_name = os.path.splitext(os.path.basename(product_file))[0] product_name = os.path.splitext(os.path.basename(product_file))[0]
if product_filter and product_name != product_filter: if product_filter and product_name != product_filter:
continue continue
with open(product_file, "r") as f: with open(product_file, "r") as f:
data = frontmatter.load(f) data = frontmatter.load(f)
if "auto" in data: if "auto" in data:
for config in data["auto"]: for config in data["auto"]:
for key, _ in config.items(): for key, _ in config.items():
if key == "maven": if key == "maven":
update_product(product_name, config) update_product(product_name, config)
def update_product(product_name, config): def update_product(product_name, config):
print("::group::%s" % product_name) print("::group::%s" % product_name)
r = fetch_releases(config['maven']) r = fetch_releases(config["maven"])
with open("releases/%s.json" % product_name, "w") as f: with open("releases/%s.json" % product_name, "w") as f:
f.write(json.dumps(r, indent=2)) f.write(json.dumps(r, indent=2))
print("::endgroup::") print("::endgroup::")
if __name__ == "__main__": if __name__ == "__main__":
if len(sys.argv) > 1: if len(sys.argv) > 1:
update_releases(sys.argv[1]) update_releases(sys.argv[1])
else: else:
update_releases() update_releases()

View File

@@ -6,36 +6,40 @@ from bs4 import BeautifulSoup
URL = "https://www.paloaltonetworks.com/services/support/end-of-life-announcements/end-of-life-summary" URL = "https://www.paloaltonetworks.com/services/support/end-of-life-announcements/end-of-life-summary"
ID_MAPPING = { ID_MAPPING = {
"pan-os-panorama": "pan-os", "pan-os-panorama": "pan-os",
"globalprotect": "pan-gp", "globalprotect": "pan-gp",
"traps-esm-and-cortex": "pan-xdr" "traps-esm-and-cortex": "pan-xdr",
} }
def update_releases(html_id, file): def update_releases(html_id, file):
list = {} list = {}
with urllib.request.urlopen(URL, data=None, timeout=5) as response: with urllib.request.urlopen(URL, data=None, timeout=5) as response:
soup = BeautifulSoup(response, features="html5lib") soup = BeautifulSoup(response, features="html5lib")
table = soup.find(id=html_id) table = soup.find(id=html_id)
for tr in table.findAll('tr')[3:]: for tr in table.findAll("tr")[3:]:
td_list = tr.findAll('td') td_list = tr.findAll("td")
version = td_list[0].get_text().strip().lower().replace(" ", "-").replace("*","") version = (
if file == 'pan-xdr': td_list[0].get_text().strip().lower().replace(" ", "-").replace("*", "")
if "xdr" not in version: )
continue if file == "pan-xdr":
version = version.removesuffix("-(cortex-xdr-agent)") if "xdr" not in version:
version = version.removesuffix("-(vm-series-only)") continue
version = version.removesuffix("-(panorama-only)") version = version.removesuffix("-(cortex-xdr-agent)")
try: version = version.removesuffix("-(vm-series-only)")
month,date,year = td_list[1].get_text().split('/') version = version.removesuffix("-(panorama-only)")
abs_date = f"{year}-{month:0>2}-{date:0>2}" try:
except Exception as e: month, date, year = td_list[1].get_text().split("/")
date = datetime.datetime.strptime(td_list[1].get_text(), "%B %d, %Y") abs_date = f"{year}-{month:0>2}-{date:0>2}"
abs_date = date.strftime("%Y-%m-%d") except Exception as e:
date = datetime.datetime.strptime(td_list[1].get_text(), "%B %d, %Y")
abs_date = date.strftime("%Y-%m-%d")
list[version] = abs_date list[version] = abs_date
with open("releases/%s.json" % file, "w") as f:
f.write(json.dumps(list, indent=2))
with open('releases/%s.json' % file, 'w') as f:
f.write(json.dumps(list, indent=2))
for html_id in ID_MAPPING: for html_id in ID_MAPPING:
update_releases(html_id, ID_MAPPING[html_id]) update_releases(html_id, ID_MAPPING[html_id])

View File

@@ -11,55 +11,61 @@ from html.parser import HTMLParser
from liquid import Template from liquid import Template
# Same as used in Ruby (update.rb) # Same as used in Ruby (update.rb)
DEFAULT_TAG_TEMPLATE = "{{major}}{% if minor %}.{{minor}}{% if patch %}.{{patch}}{%endif%}{%endif%}" DEFAULT_TAG_TEMPLATE = (
REGEX = r'^(?:(\d+\.(?:\d+\.)*\d+))$' "{{major}}{% if minor %}.{{minor}}{% if patch %}.{{patch}}{%endif%}{%endif%}"
)
REGEX = r"^(?:(\d+\.(?:\d+\.)*\d+))$"
def fetch_releases(pypi_id, regex): def fetch_releases(pypi_id, regex):
releases = {} releases = {}
if not isinstance(regex, list): if not isinstance(regex, list):
regex = [regex] regex = [regex]
url = "https://pypi.org/pypi/%s/json" % pypi_id url = "https://pypi.org/pypi/%s/json" % pypi_id
with urllib.request.urlopen(url, data=None, timeout=5) as response: with urllib.request.urlopen(url, data=None, timeout=5) as response:
data = json.loads(response.read().decode("utf-8")) data = json.loads(response.read().decode("utf-8"))
for version in data['releases']: for version in data["releases"]:
R = data['releases'][version] R = data["releases"][version]
matches = False matches = False
for r in regex: for r in regex:
if re.match(r, version): if re.match(r, version):
matches = True matches = True
if matches: if matches:
d = datetime.fromisoformat(R[0]['upload_time']).strftime("%Y-%m-%d") d = datetime.fromisoformat(R[0]["upload_time"]).strftime("%Y-%m-%d")
releases[version] = d releases[version] = d
print("%s: %s" % (version, d)) print("%s: %s" % (version, d))
return releases return releases
def update_releases(product_filter=None): def update_releases(product_filter=None):
for product_file in glob("website/products/*.md"): for product_file in glob("website/products/*.md"):
product_name = os.path.splitext(os.path.basename(product_file))[0] product_name = os.path.splitext(os.path.basename(product_file))[0]
if product_filter and product_name != product_filter: if product_filter and product_name != product_filter:
continue continue
with open(product_file, "r") as f: with open(product_file, "r") as f:
data = frontmatter.load(f) data = frontmatter.load(f)
if "auto" in data: if "auto" in data:
for config in data["auto"]: for config in data["auto"]:
for key, d_id in config.items(): for key, d_id in config.items():
if key == "pypi": if key == "pypi":
update_product(product_name, config) update_product(product_name, config)
def update_product(product_name, config): def update_product(product_name, config):
if "pypi" in config: if "pypi" in config:
print("::group::%s" % product_name) print("::group::%s" % product_name)
config = config | {"regex": REGEX} config = config | {"regex": REGEX}
r = fetch_releases(config['pypi'], config["regex"]) r = fetch_releases(config["pypi"], config["regex"])
with open("releases/%s.json" % product_name, "w") as f: with open("releases/%s.json" % product_name, "w") as f:
f.write(json.dumps(r, indent=2)) f.write(json.dumps(r, indent=2))
print("::endgroup::") print("::endgroup::")
if __name__ == "__main__": if __name__ == "__main__":
if len(sys.argv) > 1: if len(sys.argv) > 1:
update_releases(sys.argv[1]) update_releases(sys.argv[1])
else: else:
update_releases() update_releases()

View File

@@ -8,24 +8,22 @@ URL = "https://access.redhat.com/articles/3078"
regex = r"RHEL (?P<major>\d)(\. ?(?P<minor>\d+))?(( Update (?P<minor2>\d))| GA)?" regex = r"RHEL (?P<major>\d)(\. ?(?P<minor>\d+))?(( Update (?P<minor2>\d))| GA)?"
list = {} list = {}
headers = { headers = {"user-agent": "mozilla"}
'user-agent': 'mozilla' req = urllib.request.Request(URL, headers=headers)
}
req = urllib.request.Request(URL,headers=headers)
with urllib.request.urlopen(req, timeout=5) as response: with urllib.request.urlopen(req, timeout=5) as response:
soup = BeautifulSoup(response, features="html5lib") soup = BeautifulSoup(response, features="html5lib")
for tr in soup.findAll('tr'): for tr in soup.findAll("tr"):
td_list = tr.findAll('td') td_list = tr.findAll("td")
if len(td_list)>0: if len(td_list) > 0:
version = td_list[0].get_text() version = td_list[0].get_text()
m = re.match(regex, version.strip()).groupdict() m = re.match(regex, version.strip()).groupdict()
version = m['major'] version = m["major"]
if m['minor']: if m["minor"]:
version += ".%s" % m['minor'] version += ".%s" % m["minor"]
if m['minor2']: if m["minor2"]:
version += ".%s" % m['minor2'] version += ".%s" % m["minor2"]
list[version] = td_list[1].get_text() list[version] = td_list[1].get_text()
with open('releases/redhat.json', 'w') as f: with open("releases/redhat.json", "w") as f:
f.write(json.dumps(list, indent=2)) f.write(json.dumps(list, indent=2))

View File

@@ -11,24 +11,26 @@ regex = r"^ROS (?P<name>(\w| )+)"
list = {} list = {}
with urllib.request.urlopen(URL, timeout=5) as response: with urllib.request.urlopen(URL, timeout=5) as response:
soup = BeautifulSoup(response, features="html5lib") soup = BeautifulSoup(response, features="html5lib")
for tr in soup.findAll('tr'): for tr in soup.findAll("tr"):
td_list = tr.findAll('td') td_list = tr.findAll("td")
if len(td_list)>0: if len(td_list) > 0:
version = td_list[0].get_text() version = td_list[0].get_text()
m = re.match(regex, version.strip()) m = re.match(regex, version.strip())
if m: if m:
version = td_list[0].findAll('a')[0]['href'][1:] version = td_list[0].findAll("a")[0]["href"][1:]
try: try:
date = datetime.datetime.strptime(td_list[1].get_text().strip(), "%B %d, %Y") date = datetime.datetime.strptime(
# The date is a suffix (May 23rd, 2020) td_list[1].get_text().strip(), "%B %d, %Y"
except Exception as e: )
x = td_list[1].get_text().split(',') # The date is a suffix (May 23rd, 2020)
date = datetime.datetime.strptime(x[0][:-2] + x[1], "%B %d %Y") except Exception as e:
abs_date = date.strftime("%Y-%m-%d") x = td_list[1].get_text().split(",")
list[version] = abs_date date = datetime.datetime.strptime(x[0][:-2] + x[1], "%B %d %Y")
print("%s: %s" % (version, abs_date)) abs_date = date.strftime("%Y-%m-%d")
list[version] = abs_date
print("%s: %s" % (version, abs_date))
with open('releases/ros.json', 'w') as f: with open("releases/ros.json", "w") as f:
f.write(json.dumps(list, indent=2)) f.write(json.dumps(list, indent=2))

View File

@@ -4,21 +4,21 @@ import re
import urllib.request import urllib.request
URL = "https://www.unrealircd.org/docwiki/index.php?title=History_of_UnrealIRCd_releases&action=raw" URL = "https://www.unrealircd.org/docwiki/index.php?title=History_of_UnrealIRCd_releases&action=raw"
REGEX = r'^(?:(\d+\.(?:\d+\.)*\d+))$' REGEX = r"^(?:(\d+\.(?:\d+\.)*\d+))$"
list = {} list = {}
with urllib.request.urlopen(URL) as response: with urllib.request.urlopen(URL) as response:
text = response.read() text = response.read()
wikicode = mwparserfromhell.parse(text) wikicode = mwparserfromhell.parse(text)
for tr in wikicode.ifilter_tags(matches=lambda node: node.tag == "tr"): for tr in wikicode.ifilter_tags(matches=lambda node: node.tag == "tr"):
items = tr.contents.filter_tags(matches=lambda node: node.tag == "td") items = tr.contents.filter_tags(matches=lambda node: node.tag == "td")
if len(items) >=2: if len(items) >= 2:
maybe_version = items[0].__strip__() maybe_version = items[0].__strip__()
if re.match(REGEX, maybe_version): if re.match(REGEX, maybe_version):
maybe_date = items[1].__strip__() maybe_date = items[1].__strip__()
if re.match(r'\d{4}-\d{2}-\d{2}', maybe_date): if re.match(r"\d{4}-\d{2}-\d{2}", maybe_date):
list[maybe_version] = maybe_date list[maybe_version] = maybe_date
with open('releases/unrealircd.json', 'w') as f: with open("releases/unrealircd.json", "w") as f:
f.write(json.dumps(list, indent=2)) f.write(json.dumps(list, indent=2))