[manganelo] update and fix metadata extraction

This commit is contained in:
Mike Fährmann
2023-02-16 22:31:18 +01:00
parent 363bb76dff
commit 7bdc1d6d3d

View File

@@ -13,22 +13,52 @@ import re
BASE_PATTERN = r"(?:https?://)?((?:chap|read|www\.|m\.)?mangan(?:at|el)o\.com)" BASE_PATTERN = r"(?:https?://)?((?:chap|read|www\.|m\.)?mangan(?:at|el)o\.com)"
class ManganeloChapterExtractor(ChapterExtractor): class ManganeloBase():
"""Extractor for manga chapters from manganelo.com"""
category = "manganelo" category = "manganelo"
root = "https://chapmanganato.com" root = "https://chapmanganato.com"
def __init__(self, match):
domain, path = match.groups()
super().__init__(match, "https://" + domain + path)
self.session.headers['Referer'] = self.root
self._match_chapter = re.compile(
r"(?:[Vv]ol\.?\s*(\d+)\s?)?"
r"[Cc]hapter\s*([^:]+)"
r"(?::\s*(.+))?").match
def _parse_chapter(self, info, manga, author, date=None):
match = self._match_chapter(info)
volume, chapter, title = match.groups() if match else ("", "", info)
chapter, sep, minor = chapter.partition(".")
return {
"manga" : manga,
"author" : author,
"date" : date,
"title" : text.unescape(title) if title else "",
"volume" : text.parse_int(volume),
"chapter" : text.parse_int(chapter),
"chapter_minor": sep + minor,
"lang" : "en",
"language" : "English",
}
class ManganeloChapterExtractor(ManganeloBase, ChapterExtractor):
"""Extractor for manga chapters from manganelo.com"""
pattern = BASE_PATTERN + r"(/(?:manga-\w+|chapter/\w+)/chapter[-_][^/?#]+)" pattern = BASE_PATTERN + r"(/(?:manga-\w+|chapter/\w+)/chapter[-_][^/?#]+)"
test = ( test = (
("https://chapmanganato.com/manga-gn983696/chapter-23", { ("https://chapmanganato.com/manga-gn983696/chapter-23", {
"pattern": r"https://v\d+\.mkklcdnv6tempv5\.com/img/tab_17/03/23" "pattern": r"https://v\d+\.mkklcdnv6tempv5\.com/img/tab_17/03/23"
r"/39/gn983696/vol_3_chapter_23_24_yen/\d+-[no]\.jpg", r"/39/gn983696/vol_3_chapter_23_24_yen/\d+-[no]\.jpg",
"keyword": "2c5cd59342f149375df9bcb50aa416b4d04a43cf", "keyword": "17faaea7f0fb8c2675a327bf3aa0bcd7a6311d68",
"count": 25, "count": 25,
}), }),
("https://chapmanganelo.com/manga-ti107776/chapter-4", { ("https://chapmanganelo.com/manga-ti107776/chapter-4", {
"pattern": r"https://v\d+\.mkklcdnv6tempv5\.com/img/tab_17/01/92" "pattern": r"https://v\d+\.mkklcdnv6tempv5\.com/img/tab_17/01/92"
r"/08/ti970565/chapter_4_caster/\d+-o\.jpg", r"/08/ti970565/chapter_4_caster/\d+-o\.jpg",
"keyword": "16ac03503c920fb98490ad2c3621c27fddcb6c4b", "keyword": "06e01fa9b3fc9b5b954c0d4a98f0153b40922ded",
"count": 45, "count": 45,
}), }),
("https://readmanganato.com/manga-gn983696/chapter-23"), ("https://readmanganato.com/manga-gn983696/chapter-23"),
@@ -36,34 +66,15 @@ class ManganeloChapterExtractor(ChapterExtractor):
("https://manganelo.com/chapter/gq921227/chapter_23"), ("https://manganelo.com/chapter/gq921227/chapter_23"),
) )
def __init__(self, match):
domain, path = match.groups()
ChapterExtractor.__init__(self, match, "https://" + domain + path)
self.session.headers['Referer'] = self.root
def metadata(self, page): def metadata(self, page):
_ , pos = text.extract(page, 'class="a-h"', ">") extr = text.extract_from(page)
manga , pos = text.extract(page, 'title="', '"', pos) extr('class="a-h"', ">")
info , pos = text.extract(page, 'title="', '"', pos) manga = extr('title="', '"')
author, pos = text.extract(page, '- Author(s) : ', '</p>', pos) info = extr('title="', '"')
author = extr("- Author(s) : ", "</p>")
match = re.match( return self._parse_chapter(
r"(?:[Vv]ol\. *(\d+) )?" info, text.unescape(manga), text.unescape(author))
r"[Cc]hapter *([^:]*)"
r"(?:: *(.+))?", info)
volume, chapter, title = match.groups() if match else ("", "", info)
chapter, sep, minor = chapter.partition(".")
return {
"manga" : text.unescape(manga),
"title" : text.unescape(title) if title else "",
"author" : text.unescape(author) if author else "",
"volume" : text.parse_int(volume),
"chapter" : text.parse_int(chapter),
"chapter_minor": sep + minor,
"lang" : "en",
"language" : "English",
}
def images(self, page): def images(self, page):
page = text.extr( page = text.extr(
@@ -78,10 +89,8 @@ class ManganeloChapterExtractor(ChapterExtractor):
] ]
class ManganeloMangaExtractor(MangaExtractor): class ManganeloMangaExtractor(ManganeloBase, MangaExtractor):
"""Extractor for manga from manganelo.com""" """Extractor for manga from manganelo.com"""
category = "manganelo"
root = "https://chapmanganato.com"
chapterclass = ManganeloChapterExtractor chapterclass = ManganeloChapterExtractor
pattern = BASE_PATTERN + r"(/(?:manga[-/]|read_)\w+)/?$" pattern = BASE_PATTERN + r"(/(?:manga[-/]|read_)\w+)/?$"
test = ( test = (
@@ -98,35 +107,19 @@ class ManganeloMangaExtractor(MangaExtractor):
("https://manganelo.com/manga/ol921234/"), ("https://manganelo.com/manga/ol921234/"),
) )
def __init__(self, match):
domain, path = match.groups()
MangaExtractor.__init__(self, match, "https://" + domain + path)
self.session.headers['Referer'] = self.root
def chapters(self, page): def chapters(self, page):
results = [] results = []
data = self.parse_page(page, {"lang": "en", "language": "English"}) append = results.append
needle = 'class="chapter-name text-nowrap" href="' extr = text.extract_from(page)
pos = page.index('class="row-content-chapter') manga = text.unescape(extr("<h1>", "<"))
author = text.remove_html(extr("</i>Author(s) :</td>", "</tr>"))
extr('class="row-content-chapter', '')
while True: while True:
url, pos = text.extract(page, needle, '"', pos) url = extr('class="chapter-name text-nowrap" href="', '"')
if not url: if not url:
return results return results
data["title"], pos = text.extract(page, '>', '</a>', pos) info = extr(">", "<")
data["date"] , pos = text.extract( date = extr('class="chapter-time text-nowrap" title="', '"')
page, 'class="chapter-time text-nowrap" title="', '">', pos) append((url, self._parse_chapter(info, manga, author, date)))
chapter, sep, minor = url.rpartition("/chapter_")[2].partition(".")
data["chapter"] = text.parse_int(chapter)
data["chapter_minor"] = sep + minor
results.append((url, data.copy()))
@staticmethod
def parse_page(page, data):
"""Parse metadata on 'page' and add it to 'data'"""
text.extract_all(page, (
("manga" , '<h1>', '</h1>'),
('author' , '</i>Author(s) :</td>', '</tr>'),
), values=data)
data["author"] = text.remove_html(data["author"])
return data