From 1c6b9ba322eda47bae5e16f3b816dbb0b7f71301 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Mike=20F=C3=A4hrmann?= Date: Sun, 9 Dec 2018 14:54:55 +0100 Subject: [PATCH] [readcomiconline] use HTTPS --- gallery_dl/extractor/mangapark.py | 2 +- gallery_dl/extractor/piczel.py | 2 +- gallery_dl/extractor/readcomiconline.py | 25 +++++++++++++------------ 3 files changed, 15 insertions(+), 14 deletions(-) diff --git a/gallery_dl/extractor/mangapark.py b/gallery_dl/extractor/mangapark.py index 517adfdb..705cf9e5 100644 --- a/gallery_dl/extractor/mangapark.py +++ b/gallery_dl/extractor/mangapark.py @@ -83,7 +83,7 @@ class MangaparkChapterExtractor(MangaparkExtractor, ChapterExtractor): test = [ ("https://mangapark.me/manga/gosu/s2/c55", { "count": 50, - "keyword": "5454f37e6ca4113c5639c9bdad5185ccdce4bb10", + "keyword": "72ac1714b492b021a1fe26d9271ed132d51a930e", }), (("https://mangapark.me/manga/" "ad-astra-per-aspera-hata-kenjirou/s5/c1.2"), { diff --git a/gallery_dl/extractor/piczel.py b/gallery_dl/extractor/piczel.py index 2bc5d030..b46fb6ef 100644 --- a/gallery_dl/extractor/piczel.py +++ b/gallery_dl/extractor/piczel.py @@ -90,7 +90,7 @@ class PiczelImageExtractor(PiczelExtractor): subcategory = "image" pattern = [r"(?:https?://)?(?:www\.)?piczel\.tv/gallery/image/(\d+)"] test = [("https://piczel.tv/gallery/image/7807", { - "url": "c8caccac9fa798dc4fd4b920890e4d8b42cb44e5", + "url": "9b9e416b6ab7e58676fab84453d5028f306ece34", "content": "df9a053a24234474a19bce2b7e27e0dec23bff87", "keyword": { "created_at": "2018-07-22T05:13:58.000Z", diff --git a/gallery_dl/extractor/readcomiconline.py b/gallery_dl/extractor/readcomiconline.py index a2ad68eb..0e3511e7 100644 --- a/gallery_dl/extractor/readcomiconline.py +++ b/gallery_dl/extractor/readcomiconline.py @@ -19,7 +19,7 @@ class ReadcomiconlineBase(): directory_fmt = ["{category}", "{comic}", "{issue:>03}"] filename_fmt = "{comic}_{issue:>03}_{page:>03}.{extension}" archive_fmt = "{issue_id}_{page}" - root = "http://readcomiconline.to" + root = "https://readcomiconline.to" useragent = "Wget/1.19.2 (linux-gnu)" request = cloudflare.request_func @@ -28,21 +28,22 @@ class ReadcomiconlineBase(): class ReadcomiconlineComicExtractor(ReadcomiconlineBase, MangaExtractor): """Extractor for comics from readcomiconline.to""" subcategory = "comic" - pattern = [r"(?i)(?:https?://)?(?:www\.)?(readcomiconline\.to" - r"/Comic/[^/?&#]+/?)$"] + pattern = [r"(?i)(?:https?://)?(?:www\.)?readcomiconline\.to" + r"(/Comic/[^/?&#]+/?)$"] test = [ - ("http://readcomiconline.to/Comic/W-i-t-c-h", { - "url": "c5a530538a30b176916e30cbe223a93d83cb2691", + ("https://readcomiconline.to/Comic/W-i-t-c-h", { + "url": "e231bc2a293edb465133c37a8e36a7e7d94cab14", "keyword": "3986248e4458fa44a201ec073c3684917f48ee0c", }), - ("http://readcomiconline.to/Comic/Bazooka-Jules", { - "url": "e517dca61dff489f18ca781084f59a9eeb60a6b6", + ("https://readcomiconline.to/Comic/Bazooka-Jules", { + "url": "711674cb78ed10bd2557315f7a67552d01b33985", "keyword": "f5ba5246cd787bb750924d9690cb1549199bd516", }), ] + scheme = "https" def __init__(self, match): - MangaExtractor.__init__(self, match) + MangaExtractor.__init__(self, match, self.root + match.group(1)) self.session.headers["User-Agent"] = self.useragent def chapters(self, page): @@ -66,15 +67,15 @@ class ReadcomiconlineIssueExtractor(ReadcomiconlineBase, ChapterExtractor): """Extractor for comic-issues from readcomiconline.to""" subcategory = "issue" pattern = [r"(?i)(?:https?://)?(?:www\.)?readcomiconline\.to" - r"/Comic/[^/?&#]+/[^/?&#]+\?id=(\d+)"] - test = [("http://readcomiconline.to/Comic/W-i-t-c-h/Issue-130?id=22289", { + r"(/Comic/[^/?&#]+/[^/?&#]+\?id=(\d+))"] + test = [("https://readcomiconline.to/Comic/W-i-t-c-h/Issue-130?id=22289", { "url": "2bbab6ec4fbc05d269cca420a82a9b5acda28682", "keyword": "c6de1c9c8a307dc4be56783c4ac6f1338ffac6fc", })] def __init__(self, match): - ChapterExtractor.__init__(self, match.group(0)) - self.issue_id = match.group(1) + ChapterExtractor.__init__(self, self.root + match.group(1)) + self.issue_id = match.group(2) self.session.headers["User-Agent"] = self.useragent def get_metadata(self, page):