mirror of
https://github.com/karl0ss/GoGoDownloader.git
synced 2025-04-26 19:49:23 +01:00
Removed the 1 to 99 episodes limit
This commit is contained in:
parent
1626601b56
commit
89301e6f19
BIN
images/ba-screenshot-updated.png
Normal file
BIN
images/ba-screenshot-updated.png
Normal file
Binary file not shown.
After Width: | Height: | Size: 24 KiB |
BIN
images/downloaded.PNG
Normal file
BIN
images/downloaded.PNG
Normal file
Binary file not shown.
After Width: | Height: | Size: 652 KiB |
@ -23,7 +23,7 @@ def get_links(name, episode_number, source=None):
|
|||||||
|
|
||||||
|
|
||||||
def get_download_links(episode_link):
|
def get_download_links(episode_link):
|
||||||
episode_link_resp = requests.get(episode_link, stream=True)
|
episode_link_resp = requests.get(episode_link)
|
||||||
soup = BeautifulSoup(episode_link_resp.content, "html.parser")
|
soup = BeautifulSoup(episode_link_resp.content, "html.parser")
|
||||||
exist = soup.find("h1", {"class": "entry-title"})
|
exist = soup.find("h1", {"class": "entry-title"})
|
||||||
if exist is None:
|
if exist is None:
|
||||||
@ -33,7 +33,7 @@ def get_download_links(episode_link):
|
|||||||
else:
|
else:
|
||||||
# 404
|
# 404
|
||||||
episode_link = f"{episode_link}-"
|
episode_link = f"{episode_link}-"
|
||||||
episode_link_resp = requests.get(episode_link, stream=True)
|
episode_link_resp = requests.get(episode_link)
|
||||||
soup = BeautifulSoup(episode_link_resp.content, "html.parser")
|
soup = BeautifulSoup(episode_link_resp.content, "html.parser")
|
||||||
exist = soup.find("h1", {"class": "entry-title"})
|
exist = soup.find("h1", {"class": "entry-title"})
|
||||||
if exist is None:
|
if exist is None:
|
||||||
@ -41,10 +41,10 @@ def get_download_links(episode_link):
|
|||||||
return link.a.get("href")
|
return link.a.get("href")
|
||||||
else:
|
else:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
def get_download_urls(download_link):
|
def get_download_urls(download_link):
|
||||||
link = requests.get(download_link, stream=True)
|
link = requests.get(download_link)
|
||||||
soup = BeautifulSoup(link.content, "html.parser")
|
soup = BeautifulSoup(link.content, "html.parser")
|
||||||
link = soup.find_all("div", {"class": "dowload"})
|
link = soup.find_all("div", {"class": "dowload"})
|
||||||
return link[0].a.get("href")
|
return link[0].a.get("href")
|
||||||
@ -62,4 +62,4 @@ def download_episodes(url):
|
|||||||
episode_name = f"{url.split('-')[-2]}.mp4"
|
episode_name = f"{url.split('-')[-2]}.mp4"
|
||||||
file_name = os.path.join(folder_path, episode_name)
|
file_name = os.path.join(folder_path, episode_name)
|
||||||
with open(file_name, "wb") as file:
|
with open(file_name, "wb") as file:
|
||||||
shutil.copyfileobj(url_resp.raw, file)
|
shutil.copyfileobj(url_resp.raw, file)
|
||||||
|
Loading…
x
Reference in New Issue
Block a user