diff --git a/docs/configuration.rst b/docs/configuration.rst
index ddb028ef..819dc959 100644
--- a/docs/configuration.rst
+++ b/docs/configuration.rst
@@ -390,7 +390,6 @@ Description
* ``e621`` (*)
* ``e926`` (*)
* ``exhentai``
- * ``gfycat``
* ``idolcomplex``
* ``imgbb``
* ``inkbunny``
@@ -714,7 +713,7 @@ Type
Default
``["oauth", "recursive", "test"]`` + current extractor category
Example
- ``["imgur", "gfycat:user", "*:image"]``
+ ``["imgur", "redgifs:user", "*:image"]``
Description
A list of extractor identifiers to ignore (or allow)
when spawning child extractors for unknown URLs,
@@ -723,7 +722,7 @@ Description
Each identifier can be
* A category or basecategory name (``"imgur"``, ``"mastodon"``)
- * | A (base)category-subcategory pair, where both names are separated by a colon (``"gfycat:user"``).
+ * | A (base)category-subcategory pair, where both names are separated by a colon (``"redgifs:user"``).
| Both names can be a `*` or left empty, matching all possible names (``"*:image"``, ``":user"``).
Note: Any ``blacklist`` setting will automatically include
@@ -1706,29 +1705,6 @@ Description
even ones without a ``generic:`` prefix.
-extractor.gfycat.format
------------------------
-Type
- * ``string``
- * ``list`` of ``strings``
-Default
- ``["mp4", "webm", "mobile", "gif"]``
-Description
- List of names of the preferred animation format, which can be
- ``"mp4"``,
- ``"webm"``,
- ``"mobile"``,
- ``"gif"``, or
- ``"webp"``.
-
- If a selected format is not available, the next one in the list will be
- tried until an available format is found.
-
- If the format is given as ``string``, it will be extended with
- ``["mp4", "webm", "mobile", "gif"]``. Use a list with one element to
- restrict it to only one possible format.
-
-
extractor.gofile.api-token
--------------------------
Type
diff --git a/docs/gallery-dl-example.conf b/docs/gallery-dl-example.conf
index da386dd6..c3f80493 100644
--- a/docs/gallery-dl-example.conf
+++ b/docs/gallery-dl-example.conf
@@ -165,7 +165,7 @@
"reddit":
{
"#": "only spawn child extractors for links to specific sites",
- "whitelist": ["imgur", "redgifs", "gfycat"],
+ "whitelist": ["imgur", "redgifs"],
"#": "put files from child extractors into the reddit directory",
"parent-directory": true,
diff --git a/docs/gallery-dl.conf b/docs/gallery-dl.conf
index 6f397e68..e0594cad 100644
--- a/docs/gallery-dl.conf
+++ b/docs/gallery-dl.conf
@@ -125,10 +125,6 @@
"api-key": null,
"user-id": null
},
- "gfycat":
- {
- "format": ["mp4", "webm", "mobile", "gif"]
- },
"gofile": {
"api-token": null,
"website-token": null
diff --git a/docs/supportedsites.md b/docs/supportedsites.md
index 93eb3e36..111cef65 100644
--- a/docs/supportedsites.md
+++ b/docs/supportedsites.md
@@ -247,12 +247,6 @@ Consider all sites to be NSFW unless otherwise known.
Favorites, Pools, Posts, Redirects, Tag Searches |
|
-
- | Gfycat |
- https://gfycat.com/ |
- Collections, individual Images, Search Results, User Profiles |
- Supported |
-
| Gofile |
https://gofile.io/ |
diff --git a/gallery_dl/extractor/__init__.py b/gallery_dl/extractor/__init__.py
index fa56bfb4..3abe74b6 100644
--- a/gallery_dl/extractor/__init__.py
+++ b/gallery_dl/extractor/__init__.py
@@ -49,7 +49,6 @@ modules = [
"gelbooru",
"gelbooru_v01",
"gelbooru_v02",
- "gfycat",
"gofile",
"hbrowse",
"hentai2read",
diff --git a/gallery_dl/extractor/gfycat.py b/gallery_dl/extractor/gfycat.py
deleted file mode 100644
index a95dab85..00000000
--- a/gallery_dl/extractor/gfycat.py
+++ /dev/null
@@ -1,312 +0,0 @@
-# -*- coding: utf-8 -*-
-
-# Copyright 2017-2023 Mike Fährmann
-#
-# This program is free software; you can redistribute it and/or modify
-# it under the terms of the GNU General Public License version 2 as
-# published by the Free Software Foundation.
-
-"""Extractors for https://gfycat.com/"""
-
-from .common import Extractor, Message
-from .. import text, exception
-from ..cache import cache
-
-
-class GfycatExtractor(Extractor):
- """Base class for gfycat extractors"""
- category = "gfycat"
- filename_fmt = "{category}_{gfyName}{title:?_//}.{extension}"
- archive_fmt = "{gfyName}"
- root = "https://gfycat.com"
-
- def __init__(self, match):
- Extractor.__init__(self, match)
- self.key = match.group(1).lower()
-
- def _init(self):
- formats = self.config("format")
- if formats is None:
- formats = ("mp4", "webm", "mobile", "gif")
- elif isinstance(formats, str):
- formats = (formats, "mp4", "webm", "mobile", "gif")
- self.formats = formats
-
- def items(self):
- metadata = self.metadata()
- for gfycat in self.gfycats():
- if "gfyName" not in gfycat:
- self.log.warning("Skipping '%s' (malformed)", gfycat["gfyId"])
- continue
-
- url = self._process(gfycat)
- if not url:
- self.log.warning("Skipping '%s' (format not available)",
- gfycat["gfyId"])
- continue
-
- gfycat.update(metadata)
- yield Message.Directory, gfycat
- yield Message.Url, url, gfycat
-
- def _process(self, gfycat):
- gfycat["_fallback"] = formats = self._formats(gfycat)
- gfycat["date"] = text.parse_timestamp(gfycat.get("createDate"))
- return next(formats, None)
-
- def _formats(self, gfycat):
- for fmt in self.formats:
- key = fmt + "Url"
- if key in gfycat:
- url = gfycat[key]
- if url.startswith("http:"):
- url = "https" + url[4:]
- gfycat["extension"] = url.rpartition(".")[2]
- yield url
-
- def metadata(self):
- return {}
-
- def gfycats(self):
- return ()
-
-
-class GfycatUserExtractor(GfycatExtractor):
- """Extractor for gfycat user profiles"""
- subcategory = "user"
- directory_fmt = ("{category}", "{username}")
- pattern = r"(?:https?://)?gfycat\.com/@([^/?#]+)/?(?:$|\?|#)"
- test = ("https://gfycat.com/@gretta", {
- "pattern": r"https://giant\.gfycat\.com/[A-Za-z]+\.mp4",
- "count": ">= 100",
- })
-
- def gfycats(self):
- if self.key == "me":
- return GfycatAPI(self).me()
- return GfycatAPI(self).user(self.key)
-
-
-class GfycatCollectionExtractor(GfycatExtractor):
- """Extractor for a gfycat collection"""
- subcategory = "collection"
- directory_fmt = ("{category}", "{collection_owner}",
- "{collection_name|collection_id}")
- pattern = (r"(?:https?://)?gfycat\.com/@([^/?#]+)/collections"
- r"/(\w+)(?:/([^/?#]+))?")
- test = ("https://gfycat.com/@reactions/collections/nHgy2DtE/no-text", {
- "pattern": r"https://\w+\.gfycat\.com/[A-Za-z]+\.mp4",
- "count": ">= 100",
- })
-
- def __init__(self, match):
- GfycatExtractor.__init__(self, match)
- self.collection_id = match.group(2)
- self.collection_name = match.group(3)
-
- def metadata(self):
- return {
- "collection_owner": self.key,
- "collection_name" : self.collection_name,
- "collection_id" : self.collection_id,
- }
-
- def gfycats(self):
- return GfycatAPI(self).collection(self.key, self.collection_id)
-
-
-class GfycatCollectionsExtractor(GfycatExtractor):
- """Extractor for a gfycat user's collections"""
- subcategory = "collections"
- pattern = r"(?:https?://)?gfycat\.com/@([^/?#]+)/collections/?(?:$|\?|#)"
- test = ("https://gfycat.com/@sannahparker/collections", {
- "pattern": GfycatCollectionExtractor.pattern,
- "count": ">= 20",
- })
-
- def items(self):
- for col in GfycatAPI(self).collections(self.key):
- url = "https://gfycat.com/@{}/collections/{}/{}".format(
- col["userId"], col["folderId"], col["linkText"])
- col["_extractor"] = GfycatCollectionExtractor
- yield Message.Queue, url, col
-
-
-class GfycatSearchExtractor(GfycatExtractor):
- """Extractor for gfycat search results"""
- subcategory = "search"
- directory_fmt = ("{category}", "Search", "{search}")
- pattern = r"(?:https?://)?gfycat\.com/gifs/search/([^/?#]+)"
- test = ("https://gfycat.com/gifs/search/funny+animals", {
- "pattern": r"https://\w+\.gfycat\.com/[A-Za-z]+\.mp4",
- "archive": False,
- "range": "100-300",
- "count": "> 200",
- })
-
- def metadata(self):
- self.key = text.unquote(self.key).replace("+", " ")
- return {"search": self.key}
-
- def gfycats(self):
- return GfycatAPI(self).search(self.key)
-
-
-class GfycatImageExtractor(GfycatExtractor):
- """Extractor for individual images from gfycat.com"""
- subcategory = "image"
- pattern = (r"(?:https?://)?(?:\w+\.)?gfycat\.com"
- r"/(?:gifs/detail/|\w+/)?([A-Za-z]{8,})")
- test = (
- ("https://gfycat.com/GrayGenerousCowrie", {
- "url": "e0b5e1d7223108249b15c3c7898dd358dbfae045",
- "content": "5786028e04b155baa20b87c5f4f77453cd5edc37",
- "keyword": {
- "gfyId": "graygenerouscowrie",
- "gfyName": "GrayGenerousCowrie",
- "gfyNumber": 755075459,
- "title": "Bottom's up",
- "username": "jackson3oh3",
- "createDate": 1495884169,
- "date": "dt:2017-05-27 11:22:49",
- "md5": "a4796e05b0db9ba9ce5140145cd318aa",
- "width": 400,
- "height": 224,
- "frameRate": 23.0,
- "numFrames": 158.0,
- "views": int,
- },
- }),
- (("https://thumbs.gfycat.com/SillyLameIsabellinewheatear"
- "-size_restricted.gif"), {
- "url": "13b32e6cc169d086577d7dd3fd36ee6cdbc02726",
- }),
- ("https://gfycat.com/detail/UnequaledHastyAnkole?tagname=aww", {
- "url": "e24c9f69897fd223343782425a429c5cab6a768e",
- }),
- # retry 404'ed videos on redgifs (#874)
- ("https://www.gfycat.com/foolishforkedabyssiniancat", {
- "pattern": "https://redgifs.com/watch/foolishforkedabyssiniancat",
- }),
- # malformed API response (#902)
- ("https://gfycat.com/illexcitablehairstreak", {
- "count": 0,
- }),
- ("https://gfycat.com/gifs/detail/UnequaledHastyAnkole"),
- ("https://gfycat.com/ifr/UnequaledHastyAnkole"),
- ("https://gfycat.com/ru/UnequaledHastyAnkole"),
- )
-
- def items(self):
- try:
- gfycat = GfycatAPI(self).gfycat(self.key)
- except exception.HttpError:
- from .redgifs import RedgifsImageExtractor
- url = "https://redgifs.com/watch/" + self.key
- data = {"_extractor": RedgifsImageExtractor}
- yield Message.Queue, url, data
- else:
- if "gfyName" not in gfycat:
- self.log.warning("Skipping '%s' (malformed)", gfycat["gfyId"])
- return
- url = self._process(gfycat)
- if not url:
- self.log.warning("Skipping '%s' (format not available)",
- gfycat["gfyId"])
- return
- yield Message.Directory, gfycat
- yield Message.Url, url, gfycat
-
-
-class GfycatAPI():
- API_ROOT = "https://api.gfycat.com"
-
- def __init__(self, extractor):
- self.extractor = extractor
- self.headers = {}
- self.username, self.password = extractor._get_auth_info()
-
- def collection(self, user, collection):
- endpoint = "/v1/users/{}/collections/{}/gfycats".format(
- user, collection)
- params = {"count": 100}
- return self._pagination(endpoint, params)
-
- def collections(self, user):
- endpoint = "/v1/users/{}/collections".format(user)
- params = {"count": 100}
- return self._pagination(endpoint, params, "gfyCollections")
-
- def gfycat(self, gfycat_id):
- endpoint = "/v1/gfycats/" + gfycat_id
- return self._call(endpoint)["gfyItem"]
-
- def me(self):
- endpoint = "/v1/me/gfycats"
- params = {"count": 100}
- return self._pagination(endpoint, params)
-
- def search(self, query):
- endpoint = "/v1/gfycats/search"
- params = {"search_text": query, "count": 100}
- return self._pagination(endpoint, params)
-
- def user(self, user):
- endpoint = "/v1/users/{}/gfycats".format(user.lower())
- params = {"count": 100}
- return self._pagination(endpoint, params)
-
- def authenticate(self):
- self.headers["Authorization"] = \
- self._authenticate_impl(self.username, self.password)
-
- @cache(maxage=3600, keyarg=1)
- def _authenticate_impl(self, username, password):
- self.extractor.log.info("Logging in as %s", username)
-
- url = "https://weblogin.gfycat.com/oauth/webtoken"
- headers = {"Origin": "https://gfycat.com"}
- data = {
- "access_key": "Anr96uuqt9EdamSCwK4txKPjMsf2"
- "M95Rfa5FLLhPFucu8H5HTzeutyAa",
- }
- response = self.extractor.request(
- url, method="POST", headers=headers, json=data).json()
-
- url = "https://weblogin.gfycat.com/oauth/weblogin"
- headers["authorization"] = "Bearer " + response["access_token"]
- data = {
- "grant_type": "password",
- "username" : username,
- "password" : password,
- }
- response = self.extractor.request(
- url, method="POST", headers=headers, json=data, fatal=None).json()
-
- if "errorMessage" in response:
- raise exception.AuthenticationError(
- response["errorMessage"]["description"])
- return "Bearer " + response["access_token"]
-
- def _call(self, endpoint, params=None):
- if self.username:
- self.authenticate()
-
- url = self.API_ROOT + endpoint
- return self.extractor.request(
- url, params=params, headers=self.headers).json()
-
- def _pagination(self, endpoint, params, key="gfycats"):
- while True:
- data = self._call(endpoint, params)
- yield from data[key]
-
- if data["cursor"]:
- params["cursor"] = data["cursor"]
- elif "cursor" in params:
- return
- else:
- start = params.get("start", 0)
- if start + len(data[key]) >= data.get("found", 0):
- return
- params["start"] = start + params["count"]
diff --git a/scripts/supportedsites.py b/scripts/supportedsites.py
index 9077a036..b8052164 100755
--- a/scripts/supportedsites.py
+++ b/scripts/supportedsites.py
@@ -178,9 +178,6 @@ SUBCATEGORY_MAP = {
"fapello": {
"path": "Videos, Trending Posts, Popular Videos, Top Models",
},
- "gfycat": {
- "collections": "",
- },
"hentaifoundry": {
"story": "",
},
@@ -314,7 +311,6 @@ AUTH_MAP = {
"fanbox" : _COOKIES,
"fantia" : _COOKIES,
"flickr" : _OAUTH,
- "gfycat" : "Supported",
"furaffinity" : _COOKIES,
"horne" : "Required",
"idolcomplex" : "Supported",
diff --git a/test/test_results.py b/test/test_results.py
index f434b2fc..a555b721 100644
--- a/test/test_results.py
+++ b/test/test_results.py
@@ -326,7 +326,7 @@ def setup_test_config():
"e621", "e926", "e6ai",
"instagram", "twitter", "subscribestar", "deviantart",
"inkbunny", "tapas", "pillowfort", "mangadex",
- "vipergirls", "gfycat"):
+ "vipergirls"):
config.set(("extractor", category), "username", None)
config.set(("extractor", "mastodon.social"), "access-token",