mirror of
https://github.com/RicterZ/nhentai.git
synced 2025-07-01 16:09:28 +02:00
Compare commits
16 Commits
0.6.0-beta
...
dependabot
Author | SHA1 | Date | |
---|---|---|---|
ee3de81931 | |||
3be7c02458 | |||
fd1a40867e | |||
6752edfc9d | |||
9a5fcd7d23 | |||
b4cc498a5f | |||
a4eb7f3b5f | |||
36aa321ade | |||
aa84b57a43 | |||
a3c70a0c30 | |||
86060ae0a6 | |||
9648c21b32 | |||
625feb5d21 | |||
6efbc73c10 | |||
34c1ea8952 | |||
2e895d8d0f |
@ -6,10 +6,10 @@ import json
|
|||||||
import nhentai.constant as constant
|
import nhentai.constant as constant
|
||||||
|
|
||||||
from urllib.parse import urlparse
|
from urllib.parse import urlparse
|
||||||
from optparse import OptionParser
|
from argparse import ArgumentParser
|
||||||
|
|
||||||
from nhentai import __version__
|
from nhentai import __version__
|
||||||
from nhentai.utils import generate_html, generate_main_html, DB
|
from nhentai.utils import generate_html, generate_main_html, DB, EXTENSIONS
|
||||||
from nhentai.logger import logger
|
from nhentai.logger import logger
|
||||||
from nhentai.constant import PATH_SEPARATOR
|
from nhentai.constant import PATH_SEPARATOR
|
||||||
|
|
||||||
@ -57,109 +57,133 @@ def callback(option, _opt_str, _value, parser):
|
|||||||
def cmd_parser():
|
def cmd_parser():
|
||||||
load_config()
|
load_config()
|
||||||
|
|
||||||
parser = OptionParser('\n nhentai --search [keyword] --download'
|
parser = ArgumentParser(
|
||||||
'\n NHENTAI=https://nhentai-mirror-url/ nhentai --id [ID ...]'
|
description='\n nhentai --search [keyword] --download'
|
||||||
'\n nhentai --file [filename]'
|
'\n NHENTAI=https://nhentai-mirror-url/ nhentai --id [ID ...]'
|
||||||
'\n\nEnvironment Variable:\n'
|
'\n nhentai --file [filename]'
|
||||||
' NHENTAI nhentai mirror url')
|
'\n\nEnvironment Variable:\n'
|
||||||
|
' NHENTAI nhentai mirror url'
|
||||||
|
)
|
||||||
|
|
||||||
# operation options
|
# operation options
|
||||||
parser.add_option('--download', '-D', dest='is_download', action='store_true',
|
parser.add_argument('--download', '-D', dest='is_download', action='store_true',
|
||||||
help='download doujinshi (for search results)')
|
help='download doujinshi (for search results)')
|
||||||
parser.add_option('--no-download', dest='no_download', action='store_true', default=False,
|
parser.add_argument('--no-download', dest='no_download', action='store_true', default=False,
|
||||||
help='download doujinshi (for search results)')
|
help='download doujinshi (for search results)')
|
||||||
parser.add_option('--show', '-S', dest='is_show', action='store_true',
|
parser.add_argument('--show', '-S', dest='is_show', action='store_true',
|
||||||
help='just show the doujinshi information')
|
help='just show the doujinshi information')
|
||||||
|
|
||||||
# doujinshi options
|
# doujinshi options
|
||||||
parser.add_option('--id', dest='id', action='callback', callback=callback,
|
parser.add_argument('--id', dest='id', nargs='+', type=int,
|
||||||
help='doujinshi ids set, e.g. 167680 167681 167682')
|
help='doujinshi ids set, e.g. 167680 167681 167682')
|
||||||
parser.add_option('--search', '-s', type='string', dest='keyword', action='store',
|
parser.add_argument('--search', '-s', type=str, dest='keyword',
|
||||||
help='search doujinshi by keyword')
|
help='search doujinshi by keyword')
|
||||||
parser.add_option('--favorites', '-F', action='store_true', dest='favorites',
|
parser.add_argument('--favorites', '-F', action='store_true', dest='favorites',
|
||||||
help='list or download your favorites')
|
help='list or download your favorites')
|
||||||
parser.add_option('--artist', '-a', action='store', dest='artist',
|
parser.add_argument('--artist', '-a', type=str, dest='artist',
|
||||||
help='list doujinshi by artist name')
|
help='list doujinshi by artist name')
|
||||||
|
|
||||||
# page options
|
# page options
|
||||||
parser.add_option('--page-all', dest='page_all', action='store_true', default=False,
|
parser.add_argument('--page-all', dest='page_all', action='store_true', default=False,
|
||||||
help='all search results')
|
help='all search results')
|
||||||
parser.add_option('--page', '--page-range', type='string', dest='page', action='store',
|
parser.add_argument('--page', '--page-range', type=str, dest='page',
|
||||||
help='page number of search results. e.g. 1,2-5,14')
|
help='page number of search results. e.g. 1,2-5,14')
|
||||||
parser.add_option('--sorting', '--sort', dest='sorting', action='store', default='popular',
|
parser.add_argument('--sorting', '--sort', dest='sorting', type=str, default='popular',
|
||||||
help='sorting of doujinshi (recent / popular / popular-[today|week])',
|
help='sorting of doujinshi (recent / popular / popular-[today|week])',
|
||||||
choices=['recent', 'popular', 'popular-today', 'popular-week', 'date'])
|
choices=['recent', 'popular', 'popular-today', 'popular-week', 'date'])
|
||||||
|
|
||||||
# download options
|
# download options
|
||||||
parser.add_option('--output', '-o', type='string', dest='output_dir', action='store',
|
parser.add_argument('--output', '-o', type=str, dest='output_dir', default='.',
|
||||||
default=f'.{PATH_SEPARATOR}',
|
help='output dir')
|
||||||
help='output dir')
|
parser.add_argument('--threads', '-t', type=int, dest='threads', default=5,
|
||||||
parser.add_option('--threads', '-t', type='int', dest='threads', action='store', default=5,
|
help='thread count for downloading doujinshi')
|
||||||
help='thread count for downloading doujinshi')
|
parser.add_argument('--timeout', '-T', type=int, dest='timeout', default=30,
|
||||||
parser.add_option('--timeout', '-T', type='int', dest='timeout', action='store', default=30,
|
help='timeout for downloading doujinshi')
|
||||||
help='timeout for downloading doujinshi')
|
parser.add_argument('--delay', '-d', type=int, dest='delay', default=0,
|
||||||
parser.add_option('--delay', '-d', type='int', dest='delay', action='store', default=0,
|
help='slow down between downloading every doujinshi')
|
||||||
help='slow down between downloading every doujinshi')
|
parser.add_argument('--retry', type=int, dest='retry', default=3,
|
||||||
parser.add_option('--retry', type='int', dest='retry', action='store', default=3,
|
help='retry times when downloading failed')
|
||||||
help='retry times when downloading failed')
|
parser.add_argument('--exit-on-fail', dest='exit_on_fail', action='store_true', default=False,
|
||||||
parser.add_option('--exit-on-fail', dest='exit_on_fail', action='store_true', default=False,
|
help='exit on fail to prevent generating incomplete files')
|
||||||
help='exit on fail to prevent generating incomplete files')
|
parser.add_argument('--proxy', type=str, dest='proxy',
|
||||||
parser.add_option('--proxy', type='string', dest='proxy', action='store',
|
help='store a proxy, for example: -p "http://127.0.0.1:1080"')
|
||||||
help='store a proxy, for example: -p "http://127.0.0.1:1080"')
|
parser.add_argument('--file', '-f', type=str, dest='file',
|
||||||
parser.add_option('--file', '-f', type='string', dest='file', action='store',
|
help='read gallery IDs from file.')
|
||||||
help='read gallery IDs from file.')
|
parser.add_argument('--format', type=str, dest='name_format', default='[%i][%a][%t]',
|
||||||
parser.add_option('--format', type='string', dest='name_format', action='store',
|
help='format the saved folder name')
|
||||||
help='format the saved folder name', default='[%i][%a][%t]')
|
|
||||||
|
|
||||||
parser.add_option('--no-filename-padding', action='store_true', dest='no_filename_padding',
|
parser.add_argument('--no-filename-padding', action='store_true', dest='no_filename_padding',
|
||||||
default=False, help='no padding in the images filename, such as \'001.jpg\'')
|
default=False, help='no padding in the images filename, such as \'001.jpg\'')
|
||||||
|
|
||||||
# generate options
|
# generate options
|
||||||
parser.add_option('--html', dest='html_viewer', action='store_true',
|
parser.add_argument('--html', dest='html_viewer', type=str, nargs='?', const='.',
|
||||||
help='generate a html viewer at current directory')
|
help='generate an HTML viewer in the specified directory, or scan all subfolders '
|
||||||
parser.add_option('--no-html', dest='is_nohtml', action='store_true',
|
'within the entire directory to generate the HTML viewer. By default, current '
|
||||||
help='don\'t generate HTML after downloading')
|
'working directory is used.')
|
||||||
parser.add_option('--gen-main', dest='main_viewer', action='store_true',
|
parser.add_argument('--no-html', dest='is_nohtml', action='store_true',
|
||||||
help='generate a main viewer contain all the doujin in the folder')
|
help='don\'t generate HTML after downloading')
|
||||||
parser.add_option('--cbz', '-C', dest='is_cbz', action='store_true',
|
parser.add_argument('--gen-main', dest='main_viewer', action='store_true',
|
||||||
help='generate Comic Book CBZ File')
|
help='generate a main viewer contain all the doujin in the folder')
|
||||||
parser.add_option('--pdf', '-P', dest='is_pdf', action='store_true',
|
parser.add_argument('--cbz', '-C', dest='is_cbz', action='store_true',
|
||||||
help='generate PDF file')
|
help='generate Comic Book CBZ File')
|
||||||
|
parser.add_argument('--pdf', '-P', dest='is_pdf', action='store_true',
|
||||||
|
help='generate PDF file')
|
||||||
|
|
||||||
parser.add_option('--meta', dest='generate_metadata', action='store_true', default=False,
|
parser.add_argument('--meta', dest='generate_metadata', action='store_true', default=False,
|
||||||
help='generate a metadata file in doujinshi format')
|
help='generate a metadata file in doujinshi format')
|
||||||
parser.add_option('--update-meta', dest='update_metadata', action='store_true', default=False,
|
parser.add_argument('--update-meta', dest='update_metadata', action='store_true', default=False,
|
||||||
help='update the metadata file of a doujinshi, update CBZ metadata if exists')
|
help='update the metadata file of a doujinshi, update CBZ metadata if exists')
|
||||||
|
|
||||||
parser.add_option('--rm-origin-dir', dest='rm_origin_dir', action='store_true', default=False,
|
parser.add_argument('--rm-origin-dir', dest='rm_origin_dir', action='store_true', default=False,
|
||||||
help='remove downloaded doujinshi dir when generated CBZ or PDF file')
|
help='remove downloaded doujinshi dir when generated CBZ or PDF file')
|
||||||
parser.add_option('--move-to-folder', dest='move_to_folder', action='store_true', default=False,
|
parser.add_argument('--move-to-folder', dest='move_to_folder', action='store_true', default=False,
|
||||||
help='remove files in doujinshi dir then move new file to folder when generated CBZ or PDF file')
|
help='remove files in doujinshi dir then move new file to folder when generated CBZ or PDF file')
|
||||||
|
|
||||||
parser.add_option('--regenerate', dest='regenerate', action='store_true', default=False,
|
parser.add_argument('--regenerate', dest='regenerate', action='store_true', default=False,
|
||||||
help='regenerate the cbz or pdf file if exists')
|
help='regenerate the cbz or pdf file if exists')
|
||||||
|
parser.add_argument('--zip', action='store_true', help='Package into a single zip file')
|
||||||
|
|
||||||
# nhentai options
|
# nhentai options
|
||||||
parser.add_option('--cookie', type='str', dest='cookie', action='store',
|
parser.add_argument('--cookie', type=str, dest='cookie',
|
||||||
help='set cookie of nhentai to bypass Cloudflare captcha')
|
help='set cookie of nhentai to bypass Cloudflare captcha')
|
||||||
parser.add_option('--useragent', '--user-agent', type='str', dest='useragent', action='store',
|
parser.add_argument('--useragent', '--user-agent', type=str, dest='useragent',
|
||||||
help='set useragent to bypass Cloudflare captcha')
|
help='set useragent to bypass Cloudflare captcha')
|
||||||
parser.add_option('--language', type='str', dest='language', action='store',
|
parser.add_argument('--language', type=str, dest='language',
|
||||||
help='set default language to parse doujinshis')
|
help='set default language to parse doujinshis')
|
||||||
parser.add_option('--clean-language', dest='clean_language', action='store_true', default=False,
|
parser.add_argument('--clean-language', dest='clean_language', action='store_true', default=False,
|
||||||
help='set DEFAULT as language to parse doujinshis')
|
help='set DEFAULT as language to parse doujinshis')
|
||||||
parser.add_option('--save-download-history', dest='is_save_download_history', action='store_true',
|
parser.add_argument('--save-download-history', dest='is_save_download_history', action='store_true',
|
||||||
default=False, help='save downloaded doujinshis, whose will be skipped if you re-download them')
|
default=False, help='save downloaded doujinshis, whose will be skipped if you re-download them')
|
||||||
parser.add_option('--clean-download-history', action='store_true', default=False, dest='clean_download_history',
|
parser.add_argument('--clean-download-history', action='store_true', default=False, dest='clean_download_history',
|
||||||
help='clean download history')
|
help='clean download history')
|
||||||
parser.add_option('--template', dest='viewer_template', action='store',
|
parser.add_argument('--template', dest='viewer_template', type=str, default='',
|
||||||
help='set viewer template', default='')
|
help='set viewer template')
|
||||||
parser.add_option('--legacy', dest='legacy', action='store_true', default=False,
|
parser.add_argument('--legacy', dest='legacy', action='store_true', default=False,
|
||||||
help='use legacy searching method')
|
help='use legacy searching method')
|
||||||
|
|
||||||
args, _ = parser.parse_args(sys.argv[1:])
|
args = parser.parse_args()
|
||||||
|
|
||||||
if args.html_viewer:
|
if args.html_viewer:
|
||||||
generate_html(template=constant.CONFIG['template'])
|
if not os.path.exists(args.html_viewer):
|
||||||
|
logger.error(f'Path \'{args.html_viewer}\' not exists')
|
||||||
|
sys.exit(1)
|
||||||
|
|
||||||
|
for root, dirs, files in os.walk(args.html_viewer):
|
||||||
|
if not dirs:
|
||||||
|
generate_html(output_dir=args.html_viewer, template=constant.CONFIG['template'])
|
||||||
|
sys.exit(0)
|
||||||
|
|
||||||
|
for dir_name in dirs:
|
||||||
|
# it will scan the entire subdirectories
|
||||||
|
doujinshi_dir = os.path.join(root, dir_name)
|
||||||
|
items = set(map(lambda s: os.path.splitext(s)[1], os.listdir(doujinshi_dir)))
|
||||||
|
|
||||||
|
# skip directory without any images
|
||||||
|
if items & set(EXTENSIONS):
|
||||||
|
generate_html(output_dir=doujinshi_dir, template=constant.CONFIG['template'])
|
||||||
|
|
||||||
|
sys.exit(0)
|
||||||
|
|
||||||
sys.exit(0)
|
sys.exit(0)
|
||||||
|
|
||||||
if args.main_viewer and not args.id and not args.keyword and not args.favorites:
|
if args.main_viewer and not args.id and not args.keyword and not args.favorites:
|
||||||
|
@ -10,7 +10,7 @@ from nhentai import constant
|
|||||||
from nhentai.cmdline import cmd_parser, banner, write_config
|
from nhentai.cmdline import cmd_parser, banner, write_config
|
||||||
from nhentai.parser import doujinshi_parser, search_parser, legacy_search_parser, print_doujinshi, favorites_parser
|
from nhentai.parser import doujinshi_parser, search_parser, legacy_search_parser, print_doujinshi, favorites_parser
|
||||||
from nhentai.doujinshi import Doujinshi
|
from nhentai.doujinshi import Doujinshi
|
||||||
from nhentai.downloader import Downloader
|
from nhentai.downloader import Downloader, CompressedDownloader
|
||||||
from nhentai.logger import logger
|
from nhentai.logger import logger
|
||||||
from nhentai.constant import BASE_URL
|
from nhentai.constant import BASE_URL
|
||||||
from nhentai.utils import generate_html, generate_doc, generate_main_html, generate_metadata, \
|
from nhentai.utils import generate_html, generate_doc, generate_main_html, generate_metadata, \
|
||||||
@ -80,12 +80,16 @@ def main():
|
|||||||
|
|
||||||
if options.is_save_download_history:
|
if options.is_save_download_history:
|
||||||
with DB() as db:
|
with DB() as db:
|
||||||
data = map(int, db.get_all())
|
data = set(map(int, db.get_all()))
|
||||||
|
|
||||||
doujinshi_ids = list(set(map(int, doujinshi_ids)) - set(data))
|
doujinshi_ids = list(set(map(int, doujinshi_ids)) - set(data))
|
||||||
|
logger.info(f'New doujinshis account: {len(doujinshi_ids)}')
|
||||||
|
|
||||||
|
if options.zip:
|
||||||
|
options.is_nohtml = True
|
||||||
|
|
||||||
if not options.is_show:
|
if not options.is_show:
|
||||||
downloader = Downloader(path=options.output_dir, threads=options.threads,
|
downloader = (CompressedDownloader if options.zip else Downloader)(path=options.output_dir, threads=options.threads,
|
||||||
timeout=options.timeout, delay=options.delay,
|
timeout=options.timeout, delay=options.delay,
|
||||||
exit_on_fail=options.exit_on_fail,
|
exit_on_fail=options.exit_on_fail,
|
||||||
no_filename_padding=options.no_filename_padding)
|
no_filename_padding=options.no_filename_padding)
|
||||||
|
@ -4,6 +4,8 @@ import os
|
|||||||
import asyncio
|
import asyncio
|
||||||
import httpx
|
import httpx
|
||||||
import urllib3.exceptions
|
import urllib3.exceptions
|
||||||
|
import zipfile
|
||||||
|
import io
|
||||||
|
|
||||||
from urllib.parse import urlparse
|
from urllib.parse import urlparse
|
||||||
from nhentai import constant
|
from nhentai import constant
|
||||||
@ -13,11 +15,6 @@ from nhentai.utils import Singleton, async_request
|
|||||||
|
|
||||||
urllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning)
|
urllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning)
|
||||||
|
|
||||||
|
|
||||||
class NHentaiImageNotExistException(Exception):
|
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
def download_callback(result):
|
def download_callback(result):
|
||||||
result, data = result
|
result, data = result
|
||||||
if result == 0:
|
if result == 0:
|
||||||
@ -77,13 +74,7 @@ class Downloader(Singleton):
|
|||||||
else:
|
else:
|
||||||
filename = base_filename + extension
|
filename = base_filename + extension
|
||||||
|
|
||||||
save_file_path = os.path.join(self.folder, filename)
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
if os.path.exists(save_file_path):
|
|
||||||
logger.warning(f'Skipped download: {save_file_path} already exists')
|
|
||||||
return 1, url
|
|
||||||
|
|
||||||
response = await async_request('GET', url, timeout=self.timeout, proxy=proxy)
|
response = await async_request('GET', url, timeout=self.timeout, proxy=proxy)
|
||||||
|
|
||||||
if response.status_code != 200:
|
if response.status_code != 200:
|
||||||
@ -113,10 +104,6 @@ class Downloader(Singleton):
|
|||||||
logger.warning(f'Download {filename} failed with {constant.RETRY_TIMES} times retried, skipped')
|
logger.warning(f'Download {filename} failed with {constant.RETRY_TIMES} times retried, skipped')
|
||||||
return -2, url
|
return -2, url
|
||||||
|
|
||||||
except NHentaiImageNotExistException as e:
|
|
||||||
os.remove(save_file_path)
|
|
||||||
return -3, url
|
|
||||||
|
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
import traceback
|
import traceback
|
||||||
|
|
||||||
@ -130,11 +117,11 @@ class Downloader(Singleton):
|
|||||||
|
|
||||||
return 1, url
|
return 1, url
|
||||||
|
|
||||||
async def save(self, save_file_path, response) -> bool:
|
async def save(self, filename, response) -> bool:
|
||||||
if response is None:
|
if response is None:
|
||||||
logger.error('Error: Response is None')
|
logger.error('Error: Response is None')
|
||||||
return False
|
return False
|
||||||
save_file_path = os.path.join(self.folder, save_file_path)
|
save_file_path = os.path.join(self.folder, filename)
|
||||||
with open(save_file_path, 'wb') as f:
|
with open(save_file_path, 'wb') as f:
|
||||||
if response is not None:
|
if response is not None:
|
||||||
length = response.headers.get('content-length')
|
length = response.headers.get('content-length')
|
||||||
@ -145,6 +132,15 @@ class Downloader(Singleton):
|
|||||||
f.write(chunk)
|
f.write(chunk)
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
def create_storage_object(self, folder:str):
|
||||||
|
if not os.path.exists(folder):
|
||||||
|
try:
|
||||||
|
os.makedirs(folder)
|
||||||
|
except EnvironmentError as e:
|
||||||
|
logger.critical(str(e))
|
||||||
|
self.folder:str = folder
|
||||||
|
self.close = lambda: None # Only available in class CompressedDownloader
|
||||||
|
|
||||||
def start_download(self, queue, folder='') -> bool:
|
def start_download(self, queue, folder='') -> bool:
|
||||||
if not isinstance(folder, (str,)):
|
if not isinstance(folder, (str,)):
|
||||||
folder = str(folder)
|
folder = str(folder)
|
||||||
@ -153,12 +149,7 @@ class Downloader(Singleton):
|
|||||||
folder = os.path.join(self.path, folder)
|
folder = os.path.join(self.path, folder)
|
||||||
|
|
||||||
logger.info(f'Doujinshi will be saved at "{folder}"')
|
logger.info(f'Doujinshi will be saved at "{folder}"')
|
||||||
if not os.path.exists(folder):
|
self.create_storage_object(folder)
|
||||||
try:
|
|
||||||
os.makedirs(folder)
|
|
||||||
except EnvironmentError as e:
|
|
||||||
logger.critical(str(e))
|
|
||||||
self.folder = folder
|
|
||||||
|
|
||||||
if os.getenv('DEBUG', None) == 'NODOWNLOAD':
|
if os.getenv('DEBUG', None) == 'NODOWNLOAD':
|
||||||
# Assuming we want to continue with rest of process.
|
# Assuming we want to continue with rest of process.
|
||||||
@ -174,4 +165,31 @@ class Downloader(Singleton):
|
|||||||
# Prevent coroutines infection
|
# Prevent coroutines infection
|
||||||
asyncio.run(self.fiber(coroutines))
|
asyncio.run(self.fiber(coroutines))
|
||||||
|
|
||||||
|
self.close()
|
||||||
|
|
||||||
|
return True
|
||||||
|
|
||||||
|
class CompressedDownloader(Downloader):
|
||||||
|
def create_storage_object(self, folder):
|
||||||
|
filename = f'{folder}.zip'
|
||||||
|
print(filename)
|
||||||
|
self.zipfile = zipfile.ZipFile(filename,'w')
|
||||||
|
self.close = lambda: self.zipfile.close()
|
||||||
|
|
||||||
|
async def save(self, filename, response) -> bool:
|
||||||
|
if response is None:
|
||||||
|
logger.error('Error: Response is None')
|
||||||
|
return False
|
||||||
|
|
||||||
|
image_data = io.BytesIO()
|
||||||
|
length = response.headers.get('content-length')
|
||||||
|
if length is None:
|
||||||
|
content = await response.read()
|
||||||
|
image_data.write(content)
|
||||||
|
else:
|
||||||
|
async for chunk in response.aiter_bytes(2048):
|
||||||
|
image_data.write(chunk)
|
||||||
|
|
||||||
|
image_data.seek(0)
|
||||||
|
self.zipfile.writestr(filename, image_data.read())
|
||||||
return True
|
return True
|
||||||
|
@ -61,6 +61,8 @@ def serialize_comic_xml(doujinshi, output_dir):
|
|||||||
xml_write_simple_tag(f, 'Day', dt.day)
|
xml_write_simple_tag(f, 'Day', dt.day)
|
||||||
if doujinshi.info.parodies:
|
if doujinshi.info.parodies:
|
||||||
xml_write_simple_tag(f, 'Series', doujinshi.info.parodies)
|
xml_write_simple_tag(f, 'Series', doujinshi.info.parodies)
|
||||||
|
if doujinshi.info.groups:
|
||||||
|
xml_write_simple_tag(f, 'Groups', doujinshi.info.groups)
|
||||||
if doujinshi.info.characters:
|
if doujinshi.info.characters:
|
||||||
xml_write_simple_tag(f, 'Characters', doujinshi.info.characters)
|
xml_write_simple_tag(f, 'Characters', doujinshi.info.characters)
|
||||||
if doujinshi.info.tags:
|
if doujinshi.info.tags:
|
||||||
|
@ -1,5 +1,5 @@
|
|||||||
# coding: utf-8
|
# coding: utf-8
|
||||||
|
import json
|
||||||
import sys
|
import sys
|
||||||
import re
|
import re
|
||||||
import os
|
import os
|
||||||
@ -20,14 +20,24 @@ from nhentai.serializer import serialize_comic_xml, serialize_json, serialize_in
|
|||||||
MAX_FIELD_LENGTH = 100
|
MAX_FIELD_LENGTH = 100
|
||||||
EXTENSIONS = ('.png', '.jpg', '.jpeg', '.gif', '.webp')
|
EXTENSIONS = ('.png', '.jpg', '.jpeg', '.gif', '.webp')
|
||||||
|
|
||||||
|
def get_headers():
|
||||||
|
headers = {
|
||||||
|
'Referer': constant.LOGIN_URL
|
||||||
|
}
|
||||||
|
|
||||||
|
user_agent = constant.CONFIG.get('useragent')
|
||||||
|
if user_agent and user_agent.strip():
|
||||||
|
headers['User-Agent'] = user_agent
|
||||||
|
|
||||||
|
cookie = constant.CONFIG.get('cookie')
|
||||||
|
if cookie and cookie.strip():
|
||||||
|
headers['Cookie'] = cookie
|
||||||
|
|
||||||
|
return headers
|
||||||
|
|
||||||
def request(method, url, **kwargs):
|
def request(method, url, **kwargs):
|
||||||
session = requests.Session()
|
session = requests.Session()
|
||||||
session.headers.update({
|
session.headers.update(get_headers())
|
||||||
'Referer': constant.LOGIN_URL,
|
|
||||||
'User-Agent': constant.CONFIG['useragent'],
|
|
||||||
'Cookie': constant.CONFIG['cookie']
|
|
||||||
})
|
|
||||||
|
|
||||||
if not kwargs.get('proxies', None):
|
if not kwargs.get('proxies', None):
|
||||||
kwargs['proxies'] = {
|
kwargs['proxies'] = {
|
||||||
@ -39,11 +49,7 @@ def request(method, url, **kwargs):
|
|||||||
|
|
||||||
|
|
||||||
async def async_request(method, url, proxy = None, **kwargs):
|
async def async_request(method, url, proxy = None, **kwargs):
|
||||||
headers = {
|
headers=get_headers()
|
||||||
'Referer': constant.LOGIN_URL,
|
|
||||||
'User-Agent': constant.CONFIG['useragent'],
|
|
||||||
'Cookie': constant.CONFIG['cookie'],
|
|
||||||
}
|
|
||||||
|
|
||||||
if proxy is None:
|
if proxy is None:
|
||||||
proxy = constant.CONFIG['proxy']
|
proxy = constant.CONFIG['proxy']
|
||||||
@ -109,6 +115,9 @@ def parse_doujinshi_obj(
|
|||||||
|
|
||||||
filename = os.path.join(output_dir, _filename)
|
filename = os.path.join(output_dir, _filename)
|
||||||
else:
|
else:
|
||||||
|
if file_type == 'html':
|
||||||
|
return output_dir, 'index.html'
|
||||||
|
|
||||||
doujinshi_dir = f'.{PATH_SEPARATOR}'
|
doujinshi_dir = f'.{PATH_SEPARATOR}'
|
||||||
|
|
||||||
if not os.path.exists(doujinshi_dir):
|
if not os.path.exists(doujinshi_dir):
|
||||||
@ -118,7 +127,7 @@ def parse_doujinshi_obj(
|
|||||||
|
|
||||||
|
|
||||||
def generate_html(output_dir='.', doujinshi_obj=None, template='default'):
|
def generate_html(output_dir='.', doujinshi_obj=None, template='default'):
|
||||||
doujinshi_dir, filename = parse_doujinshi_obj(output_dir, doujinshi_obj, '.html')
|
doujinshi_dir, filename = parse_doujinshi_obj(output_dir, doujinshi_obj, 'html')
|
||||||
image_html = ''
|
image_html = ''
|
||||||
|
|
||||||
if not os.path.exists(doujinshi_dir):
|
if not os.path.exists(doujinshi_dir):
|
||||||
@ -144,7 +153,13 @@ def generate_html(output_dir='.', doujinshi_obj=None, template='default'):
|
|||||||
# serialize_json(doujinshi_obj, doujinshi_dir)
|
# serialize_json(doujinshi_obj, doujinshi_dir)
|
||||||
name = doujinshi_obj.name
|
name = doujinshi_obj.name
|
||||||
else:
|
else:
|
||||||
name = {'title': 'nHentai HTML Viewer'}
|
metadata_path = os.path.join(doujinshi_dir, "metadata.json")
|
||||||
|
if os.path.exists(metadata_path):
|
||||||
|
with open(metadata_path, 'r') as file:
|
||||||
|
doujinshi_info = json.loads(file.read())
|
||||||
|
name = doujinshi_info.get("title")
|
||||||
|
else:
|
||||||
|
name = 'nHentai HTML Viewer'
|
||||||
|
|
||||||
data = html.format(TITLE=name, IMAGES=image_html, SCRIPTS=js, STYLES=css)
|
data = html.format(TITLE=name, IMAGES=image_html, SCRIPTS=js, STYLES=css)
|
||||||
try:
|
try:
|
||||||
|
39
poetry.lock
generated
39
poetry.lock
generated
@ -1,4 +1,4 @@
|
|||||||
# This file is automatically @generated by Poetry 1.8.3 and should not be changed by hand.
|
# This file is automatically @generated by Poetry 2.1.1 and should not be changed by hand.
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "anyio"
|
name = "anyio"
|
||||||
@ -6,6 +6,7 @@ version = "4.5.2"
|
|||||||
description = "High level compatibility layer for multiple asynchronous event loop implementations"
|
description = "High level compatibility layer for multiple asynchronous event loop implementations"
|
||||||
optional = false
|
optional = false
|
||||||
python-versions = ">=3.8"
|
python-versions = ">=3.8"
|
||||||
|
groups = ["main"]
|
||||||
files = [
|
files = [
|
||||||
{file = "anyio-4.5.2-py3-none-any.whl", hash = "sha256:c011ee36bc1e8ba40e5a81cb9df91925c218fe9b778554e0b56a21e1b5d4716f"},
|
{file = "anyio-4.5.2-py3-none-any.whl", hash = "sha256:c011ee36bc1e8ba40e5a81cb9df91925c218fe9b778554e0b56a21e1b5d4716f"},
|
||||||
{file = "anyio-4.5.2.tar.gz", hash = "sha256:23009af4ed04ce05991845451e11ef02fc7c5ed29179ac9a420e5ad0ac7ddc5b"},
|
{file = "anyio-4.5.2.tar.gz", hash = "sha256:23009af4ed04ce05991845451e11ef02fc7c5ed29179ac9a420e5ad0ac7ddc5b"},
|
||||||
@ -19,7 +20,7 @@ typing-extensions = {version = ">=4.1", markers = "python_version < \"3.11\""}
|
|||||||
|
|
||||||
[package.extras]
|
[package.extras]
|
||||||
doc = ["Sphinx (>=7.4,<8.0)", "packaging", "sphinx-autodoc-typehints (>=1.2.0)", "sphinx-rtd-theme"]
|
doc = ["Sphinx (>=7.4,<8.0)", "packaging", "sphinx-autodoc-typehints (>=1.2.0)", "sphinx-rtd-theme"]
|
||||||
test = ["anyio[trio]", "coverage[toml] (>=7)", "exceptiongroup (>=1.2.0)", "hypothesis (>=4.0)", "psutil (>=5.9)", "pytest (>=7.0)", "pytest-mock (>=3.6.1)", "trustme", "truststore (>=0.9.1)", "uvloop (>=0.21.0b1)"]
|
test = ["anyio[trio]", "coverage[toml] (>=7)", "exceptiongroup (>=1.2.0)", "hypothesis (>=4.0)", "psutil (>=5.9)", "pytest (>=7.0)", "pytest-mock (>=3.6.1)", "trustme", "truststore (>=0.9.1) ; python_version >= \"3.10\"", "uvloop (>=0.21.0b1) ; platform_python_implementation == \"CPython\" and platform_system != \"Windows\""]
|
||||||
trio = ["trio (>=0.26.1)"]
|
trio = ["trio (>=0.26.1)"]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
@ -28,6 +29,7 @@ version = "4.12.3"
|
|||||||
description = "Screen-scraping library"
|
description = "Screen-scraping library"
|
||||||
optional = false
|
optional = false
|
||||||
python-versions = ">=3.6.0"
|
python-versions = ">=3.6.0"
|
||||||
|
groups = ["main"]
|
||||||
files = [
|
files = [
|
||||||
{file = "beautifulsoup4-4.12.3-py3-none-any.whl", hash = "sha256:b80878c9f40111313e55da8ba20bdba06d8fa3969fc68304167741bbf9e082ed"},
|
{file = "beautifulsoup4-4.12.3-py3-none-any.whl", hash = "sha256:b80878c9f40111313e55da8ba20bdba06d8fa3969fc68304167741bbf9e082ed"},
|
||||||
{file = "beautifulsoup4-4.12.3.tar.gz", hash = "sha256:74e3d1928edc070d21748185c46e3fb33490f22f52a3addee9aee0f4f7781051"},
|
{file = "beautifulsoup4-4.12.3.tar.gz", hash = "sha256:74e3d1928edc070d21748185c46e3fb33490f22f52a3addee9aee0f4f7781051"},
|
||||||
@ -49,6 +51,7 @@ version = "2024.12.14"
|
|||||||
description = "Python package for providing Mozilla's CA Bundle."
|
description = "Python package for providing Mozilla's CA Bundle."
|
||||||
optional = false
|
optional = false
|
||||||
python-versions = ">=3.6"
|
python-versions = ">=3.6"
|
||||||
|
groups = ["main"]
|
||||||
files = [
|
files = [
|
||||||
{file = "certifi-2024.12.14-py3-none-any.whl", hash = "sha256:1275f7a45be9464efc1173084eaa30f866fe2e47d389406136d332ed4967ec56"},
|
{file = "certifi-2024.12.14-py3-none-any.whl", hash = "sha256:1275f7a45be9464efc1173084eaa30f866fe2e47d389406136d332ed4967ec56"},
|
||||||
{file = "certifi-2024.12.14.tar.gz", hash = "sha256:b650d30f370c2b724812bee08008be0c4163b163ddaec3f2546c1caf65f191db"},
|
{file = "certifi-2024.12.14.tar.gz", hash = "sha256:b650d30f370c2b724812bee08008be0c4163b163ddaec3f2546c1caf65f191db"},
|
||||||
@ -60,6 +63,7 @@ version = "5.2.0"
|
|||||||
description = "Universal encoding detector for Python 3"
|
description = "Universal encoding detector for Python 3"
|
||||||
optional = false
|
optional = false
|
||||||
python-versions = ">=3.7"
|
python-versions = ">=3.7"
|
||||||
|
groups = ["main"]
|
||||||
files = [
|
files = [
|
||||||
{file = "chardet-5.2.0-py3-none-any.whl", hash = "sha256:e1cf59446890a00105fe7b7912492ea04b6e6f06d4b742b2c788469e34c82970"},
|
{file = "chardet-5.2.0-py3-none-any.whl", hash = "sha256:e1cf59446890a00105fe7b7912492ea04b6e6f06d4b742b2c788469e34c82970"},
|
||||||
{file = "chardet-5.2.0.tar.gz", hash = "sha256:1b3b6ff479a8c414bc3fa2c0852995695c4a026dcd6d0633b2dd092ca39c1cf7"},
|
{file = "chardet-5.2.0.tar.gz", hash = "sha256:1b3b6ff479a8c414bc3fa2c0852995695c4a026dcd6d0633b2dd092ca39c1cf7"},
|
||||||
@ -71,6 +75,7 @@ version = "3.4.1"
|
|||||||
description = "The Real First Universal Charset Detector. Open, modern and actively maintained alternative to Chardet."
|
description = "The Real First Universal Charset Detector. Open, modern and actively maintained alternative to Chardet."
|
||||||
optional = false
|
optional = false
|
||||||
python-versions = ">=3.7"
|
python-versions = ">=3.7"
|
||||||
|
groups = ["main"]
|
||||||
files = [
|
files = [
|
||||||
{file = "charset_normalizer-3.4.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:91b36a978b5ae0ee86c394f5a54d6ef44db1de0815eb43de826d41d21e4af3de"},
|
{file = "charset_normalizer-3.4.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:91b36a978b5ae0ee86c394f5a54d6ef44db1de0815eb43de826d41d21e4af3de"},
|
||||||
{file = "charset_normalizer-3.4.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7461baadb4dc00fd9e0acbe254e3d7d2112e7f92ced2adc96e54ef6501c5f176"},
|
{file = "charset_normalizer-3.4.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7461baadb4dc00fd9e0acbe254e3d7d2112e7f92ced2adc96e54ef6501c5f176"},
|
||||||
@ -172,6 +177,8 @@ version = "1.2.2"
|
|||||||
description = "Backport of PEP 654 (exception groups)"
|
description = "Backport of PEP 654 (exception groups)"
|
||||||
optional = false
|
optional = false
|
||||||
python-versions = ">=3.7"
|
python-versions = ">=3.7"
|
||||||
|
groups = ["main"]
|
||||||
|
markers = "python_version < \"3.11\""
|
||||||
files = [
|
files = [
|
||||||
{file = "exceptiongroup-1.2.2-py3-none-any.whl", hash = "sha256:3111b9d131c238bec2f8f516e123e14ba243563fb135d3fe885990585aa7795b"},
|
{file = "exceptiongroup-1.2.2-py3-none-any.whl", hash = "sha256:3111b9d131c238bec2f8f516e123e14ba243563fb135d3fe885990585aa7795b"},
|
||||||
{file = "exceptiongroup-1.2.2.tar.gz", hash = "sha256:47c2edf7c6738fafb49fd34290706d1a1a2f4d1c6df275526b62cbb4aa5393cc"},
|
{file = "exceptiongroup-1.2.2.tar.gz", hash = "sha256:47c2edf7c6738fafb49fd34290706d1a1a2f4d1c6df275526b62cbb4aa5393cc"},
|
||||||
@ -186,6 +193,7 @@ version = "0.14.0"
|
|||||||
description = "A pure-Python, bring-your-own-I/O implementation of HTTP/1.1"
|
description = "A pure-Python, bring-your-own-I/O implementation of HTTP/1.1"
|
||||||
optional = false
|
optional = false
|
||||||
python-versions = ">=3.7"
|
python-versions = ">=3.7"
|
||||||
|
groups = ["main"]
|
||||||
files = [
|
files = [
|
||||||
{file = "h11-0.14.0-py3-none-any.whl", hash = "sha256:e3fe4ac4b851c468cc8363d500db52c2ead036020723024a109d37346efaa761"},
|
{file = "h11-0.14.0-py3-none-any.whl", hash = "sha256:e3fe4ac4b851c468cc8363d500db52c2ead036020723024a109d37346efaa761"},
|
||||||
{file = "h11-0.14.0.tar.gz", hash = "sha256:8f19fbbe99e72420ff35c00b27a34cb9937e902a8b810e2c88300c6f0a3b699d"},
|
{file = "h11-0.14.0.tar.gz", hash = "sha256:8f19fbbe99e72420ff35c00b27a34cb9937e902a8b810e2c88300c6f0a3b699d"},
|
||||||
@ -197,6 +205,7 @@ version = "1.0.7"
|
|||||||
description = "A minimal low-level HTTP client."
|
description = "A minimal low-level HTTP client."
|
||||||
optional = false
|
optional = false
|
||||||
python-versions = ">=3.8"
|
python-versions = ">=3.8"
|
||||||
|
groups = ["main"]
|
||||||
files = [
|
files = [
|
||||||
{file = "httpcore-1.0.7-py3-none-any.whl", hash = "sha256:a3fff8f43dc260d5bd363d9f9cf1830fa3a458b332856f34282de498ed420edd"},
|
{file = "httpcore-1.0.7-py3-none-any.whl", hash = "sha256:a3fff8f43dc260d5bd363d9f9cf1830fa3a458b332856f34282de498ed420edd"},
|
||||||
{file = "httpcore-1.0.7.tar.gz", hash = "sha256:8551cb62a169ec7162ac7be8d4817d561f60e08eaa485234898414bb5a8a0b4c"},
|
{file = "httpcore-1.0.7.tar.gz", hash = "sha256:8551cb62a169ec7162ac7be8d4817d561f60e08eaa485234898414bb5a8a0b4c"},
|
||||||
@ -218,6 +227,7 @@ version = "0.28.1"
|
|||||||
description = "The next generation HTTP client."
|
description = "The next generation HTTP client."
|
||||||
optional = false
|
optional = false
|
||||||
python-versions = ">=3.8"
|
python-versions = ">=3.8"
|
||||||
|
groups = ["main"]
|
||||||
files = [
|
files = [
|
||||||
{file = "httpx-0.28.1-py3-none-any.whl", hash = "sha256:d909fcccc110f8c7faf814ca82a9a4d816bc5a6dbfea25d6591d6985b8ba59ad"},
|
{file = "httpx-0.28.1-py3-none-any.whl", hash = "sha256:d909fcccc110f8c7faf814ca82a9a4d816bc5a6dbfea25d6591d6985b8ba59ad"},
|
||||||
{file = "httpx-0.28.1.tar.gz", hash = "sha256:75e98c5f16b0f35b567856f597f06ff2270a374470a5c2392242528e3e3e42fc"},
|
{file = "httpx-0.28.1.tar.gz", hash = "sha256:75e98c5f16b0f35b567856f597f06ff2270a374470a5c2392242528e3e3e42fc"},
|
||||||
@ -230,7 +240,7 @@ httpcore = "==1.*"
|
|||||||
idna = "*"
|
idna = "*"
|
||||||
|
|
||||||
[package.extras]
|
[package.extras]
|
||||||
brotli = ["brotli", "brotlicffi"]
|
brotli = ["brotli ; platform_python_implementation == \"CPython\"", "brotlicffi ; platform_python_implementation != \"CPython\""]
|
||||||
cli = ["click (==8.*)", "pygments (==2.*)", "rich (>=10,<14)"]
|
cli = ["click (==8.*)", "pygments (==2.*)", "rich (>=10,<14)"]
|
||||||
http2 = ["h2 (>=3,<5)"]
|
http2 = ["h2 (>=3,<5)"]
|
||||||
socks = ["socksio (==1.*)"]
|
socks = ["socksio (==1.*)"]
|
||||||
@ -242,6 +252,7 @@ version = "3.10"
|
|||||||
description = "Internationalized Domain Names in Applications (IDNA)"
|
description = "Internationalized Domain Names in Applications (IDNA)"
|
||||||
optional = false
|
optional = false
|
||||||
python-versions = ">=3.6"
|
python-versions = ">=3.6"
|
||||||
|
groups = ["main"]
|
||||||
files = [
|
files = [
|
||||||
{file = "idna-3.10-py3-none-any.whl", hash = "sha256:946d195a0d259cbba61165e88e65941f16e9b36ea6ddb97f00452bae8b1287d3"},
|
{file = "idna-3.10-py3-none-any.whl", hash = "sha256:946d195a0d259cbba61165e88e65941f16e9b36ea6ddb97f00452bae8b1287d3"},
|
||||||
{file = "idna-3.10.tar.gz", hash = "sha256:12f65c9b470abda6dc35cf8e63cc574b1c52b11df2c86030af0ac09b01b13ea9"},
|
{file = "idna-3.10.tar.gz", hash = "sha256:12f65c9b470abda6dc35cf8e63cc574b1c52b11df2c86030af0ac09b01b13ea9"},
|
||||||
@ -256,6 +267,7 @@ version = "1.1.0"
|
|||||||
description = "Simple module to parse ISO 8601 dates"
|
description = "Simple module to parse ISO 8601 dates"
|
||||||
optional = false
|
optional = false
|
||||||
python-versions = ">=3.6.2,<4.0"
|
python-versions = ">=3.6.2,<4.0"
|
||||||
|
groups = ["main"]
|
||||||
files = [
|
files = [
|
||||||
{file = "iso8601-1.1.0-py3-none-any.whl", hash = "sha256:8400e90141bf792bce2634df533dc57e3bee19ea120a87bebcd3da89a58ad73f"},
|
{file = "iso8601-1.1.0-py3-none-any.whl", hash = "sha256:8400e90141bf792bce2634df533dc57e3bee19ea120a87bebcd3da89a58ad73f"},
|
||||||
{file = "iso8601-1.1.0.tar.gz", hash = "sha256:32811e7b81deee2063ea6d2e94f8819a86d1f3811e49d23623a41fa832bef03f"},
|
{file = "iso8601-1.1.0.tar.gz", hash = "sha256:32811e7b81deee2063ea6d2e94f8819a86d1f3811e49d23623a41fa832bef03f"},
|
||||||
@ -263,18 +275,19 @@ files = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "requests"
|
name = "requests"
|
||||||
version = "2.32.3"
|
version = "2.32.4"
|
||||||
description = "Python HTTP for Humans."
|
description = "Python HTTP for Humans."
|
||||||
optional = false
|
optional = false
|
||||||
python-versions = ">=3.8"
|
python-versions = ">=3.8"
|
||||||
|
groups = ["main"]
|
||||||
files = [
|
files = [
|
||||||
{file = "requests-2.32.3-py3-none-any.whl", hash = "sha256:70761cfe03c773ceb22aa2f671b4757976145175cdfca038c02654d061d6dcc6"},
|
{file = "requests-2.32.4-py3-none-any.whl", hash = "sha256:27babd3cda2a6d50b30443204ee89830707d396671944c998b5975b031ac2b2c"},
|
||||||
{file = "requests-2.32.3.tar.gz", hash = "sha256:55365417734eb18255590a9ff9eb97e9e1da868d4ccd6402399eaf68af20a760"},
|
{file = "requests-2.32.4.tar.gz", hash = "sha256:27d0316682c8a29834d3264820024b62a36942083d52caf2f14c0591336d3422"},
|
||||||
]
|
]
|
||||||
|
|
||||||
[package.dependencies]
|
[package.dependencies]
|
||||||
certifi = ">=2017.4.17"
|
certifi = ">=2017.4.17"
|
||||||
charset-normalizer = ">=2,<4"
|
charset_normalizer = ">=2,<4"
|
||||||
idna = ">=2.5,<4"
|
idna = ">=2.5,<4"
|
||||||
urllib3 = ">=1.21.1,<3"
|
urllib3 = ">=1.21.1,<3"
|
||||||
|
|
||||||
@ -288,6 +301,7 @@ version = "1.3.1"
|
|||||||
description = "Sniff out which async library your code is running under"
|
description = "Sniff out which async library your code is running under"
|
||||||
optional = false
|
optional = false
|
||||||
python-versions = ">=3.7"
|
python-versions = ">=3.7"
|
||||||
|
groups = ["main"]
|
||||||
files = [
|
files = [
|
||||||
{file = "sniffio-1.3.1-py3-none-any.whl", hash = "sha256:2f6da418d1f1e0fddd844478f41680e794e6051915791a034ff65e5f100525a2"},
|
{file = "sniffio-1.3.1-py3-none-any.whl", hash = "sha256:2f6da418d1f1e0fddd844478f41680e794e6051915791a034ff65e5f100525a2"},
|
||||||
{file = "sniffio-1.3.1.tar.gz", hash = "sha256:f4324edc670a0f49750a81b895f35c3adb843cca46f0530f79fc1babb23789dc"},
|
{file = "sniffio-1.3.1.tar.gz", hash = "sha256:f4324edc670a0f49750a81b895f35c3adb843cca46f0530f79fc1babb23789dc"},
|
||||||
@ -299,6 +313,7 @@ version = "2.6"
|
|||||||
description = "A modern CSS selector implementation for Beautiful Soup."
|
description = "A modern CSS selector implementation for Beautiful Soup."
|
||||||
optional = false
|
optional = false
|
||||||
python-versions = ">=3.8"
|
python-versions = ">=3.8"
|
||||||
|
groups = ["main"]
|
||||||
files = [
|
files = [
|
||||||
{file = "soupsieve-2.6-py3-none-any.whl", hash = "sha256:e72c4ff06e4fb6e4b5a9f0f55fe6e81514581fca1515028625d0f299c602ccc9"},
|
{file = "soupsieve-2.6-py3-none-any.whl", hash = "sha256:e72c4ff06e4fb6e4b5a9f0f55fe6e81514581fca1515028625d0f299c602ccc9"},
|
||||||
{file = "soupsieve-2.6.tar.gz", hash = "sha256:e2e68417777af359ec65daac1057404a3c8a5455bb8abc36f1a9866ab1a51abb"},
|
{file = "soupsieve-2.6.tar.gz", hash = "sha256:e2e68417777af359ec65daac1057404a3c8a5455bb8abc36f1a9866ab1a51abb"},
|
||||||
@ -310,6 +325,7 @@ version = "0.9.0"
|
|||||||
description = "Pretty-print tabular data"
|
description = "Pretty-print tabular data"
|
||||||
optional = false
|
optional = false
|
||||||
python-versions = ">=3.7"
|
python-versions = ">=3.7"
|
||||||
|
groups = ["main"]
|
||||||
files = [
|
files = [
|
||||||
{file = "tabulate-0.9.0-py3-none-any.whl", hash = "sha256:024ca478df22e9340661486f85298cff5f6dcdba14f3813e8830015b9ed1948f"},
|
{file = "tabulate-0.9.0-py3-none-any.whl", hash = "sha256:024ca478df22e9340661486f85298cff5f6dcdba14f3813e8830015b9ed1948f"},
|
||||||
{file = "tabulate-0.9.0.tar.gz", hash = "sha256:0095b12bf5966de529c0feb1fa08671671b3368eec77d7ef7ab114be2c068b3c"},
|
{file = "tabulate-0.9.0.tar.gz", hash = "sha256:0095b12bf5966de529c0feb1fa08671671b3368eec77d7ef7ab114be2c068b3c"},
|
||||||
@ -324,6 +340,8 @@ version = "4.12.2"
|
|||||||
description = "Backported and Experimental Type Hints for Python 3.8+"
|
description = "Backported and Experimental Type Hints for Python 3.8+"
|
||||||
optional = false
|
optional = false
|
||||||
python-versions = ">=3.8"
|
python-versions = ">=3.8"
|
||||||
|
groups = ["main"]
|
||||||
|
markers = "python_version < \"3.11\""
|
||||||
files = [
|
files = [
|
||||||
{file = "typing_extensions-4.12.2-py3-none-any.whl", hash = "sha256:04e5ca0351e0f3f85c6853954072df659d0d13fac324d0072316b67d7794700d"},
|
{file = "typing_extensions-4.12.2-py3-none-any.whl", hash = "sha256:04e5ca0351e0f3f85c6853954072df659d0d13fac324d0072316b67d7794700d"},
|
||||||
{file = "typing_extensions-4.12.2.tar.gz", hash = "sha256:1a7ead55c7e559dd4dee8856e3a88b41225abfe1ce8df57b7c13915fe121ffb8"},
|
{file = "typing_extensions-4.12.2.tar.gz", hash = "sha256:1a7ead55c7e559dd4dee8856e3a88b41225abfe1ce8df57b7c13915fe121ffb8"},
|
||||||
@ -335,17 +353,18 @@ version = "1.26.20"
|
|||||||
description = "HTTP library with thread-safe connection pooling, file post, and more."
|
description = "HTTP library with thread-safe connection pooling, file post, and more."
|
||||||
optional = false
|
optional = false
|
||||||
python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,>=2.7"
|
python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,>=2.7"
|
||||||
|
groups = ["main"]
|
||||||
files = [
|
files = [
|
||||||
{file = "urllib3-1.26.20-py2.py3-none-any.whl", hash = "sha256:0ed14ccfbf1c30a9072c7ca157e4319b70d65f623e91e7b32fadb2853431016e"},
|
{file = "urllib3-1.26.20-py2.py3-none-any.whl", hash = "sha256:0ed14ccfbf1c30a9072c7ca157e4319b70d65f623e91e7b32fadb2853431016e"},
|
||||||
{file = "urllib3-1.26.20.tar.gz", hash = "sha256:40c2dc0c681e47eb8f90e7e27bf6ff7df2e677421fd46756da1161c39ca70d32"},
|
{file = "urllib3-1.26.20.tar.gz", hash = "sha256:40c2dc0c681e47eb8f90e7e27bf6ff7df2e677421fd46756da1161c39ca70d32"},
|
||||||
]
|
]
|
||||||
|
|
||||||
[package.extras]
|
[package.extras]
|
||||||
brotli = ["brotli (==1.0.9)", "brotli (>=1.0.9)", "brotlicffi (>=0.8.0)", "brotlipy (>=0.6.0)"]
|
brotli = ["brotli (==1.0.9) ; os_name != \"nt\" and python_version < \"3\" and platform_python_implementation == \"CPython\"", "brotli (>=1.0.9) ; python_version >= \"3\" and platform_python_implementation == \"CPython\"", "brotlicffi (>=0.8.0) ; (os_name != \"nt\" or python_version >= \"3\") and platform_python_implementation != \"CPython\"", "brotlipy (>=0.6.0) ; os_name == \"nt\" and python_version < \"3\""]
|
||||||
secure = ["certifi", "cryptography (>=1.3.4)", "idna (>=2.0.0)", "ipaddress", "pyOpenSSL (>=0.14)", "urllib3-secure-extra"]
|
secure = ["certifi", "cryptography (>=1.3.4)", "idna (>=2.0.0)", "ipaddress ; python_version == \"2.7\"", "pyOpenSSL (>=0.14)", "urllib3-secure-extra"]
|
||||||
socks = ["PySocks (>=1.5.6,!=1.5.7,<2.0)"]
|
socks = ["PySocks (>=1.5.6,!=1.5.7,<2.0)"]
|
||||||
|
|
||||||
[metadata]
|
[metadata]
|
||||||
lock-version = "2.0"
|
lock-version = "2.1"
|
||||||
python-versions = "^3.8"
|
python-versions = "^3.8"
|
||||||
content-hash = "b17c2cdd4b140f2ab8081bca7d94630e821fa2e882ac768b1bd8cf3ec58726ce"
|
content-hash = "b17c2cdd4b140f2ab8081bca7d94630e821fa2e882ac768b1bd8cf3ec58726ce"
|
||||||
|
@ -1,14 +1,27 @@
|
|||||||
import unittest
|
import unittest
|
||||||
import os
|
import os
|
||||||
|
import zipfile
|
||||||
import urllib3.exceptions
|
import urllib3.exceptions
|
||||||
|
|
||||||
from nhentai import constant
|
from nhentai import constant
|
||||||
from nhentai.cmdline import load_config
|
from nhentai.cmdline import load_config
|
||||||
from nhentai.downloader import Downloader
|
from nhentai.downloader import Downloader, CompressedDownloader
|
||||||
from nhentai.parser import doujinshi_parser
|
from nhentai.parser import doujinshi_parser
|
||||||
from nhentai.doujinshi import Doujinshi
|
from nhentai.doujinshi import Doujinshi
|
||||||
from nhentai.utils import generate_html
|
from nhentai.utils import generate_html
|
||||||
|
|
||||||
|
did = 440546
|
||||||
|
|
||||||
|
def has_jepg_file(path):
|
||||||
|
with zipfile.ZipFile(path, 'r') as zf:
|
||||||
|
return '01.jpg' in zf.namelist()
|
||||||
|
|
||||||
|
def is_zip_file(path):
|
||||||
|
try:
|
||||||
|
with zipfile.ZipFile(path, 'r') as _:
|
||||||
|
return True
|
||||||
|
except (zipfile.BadZipFile, FileNotFoundError):
|
||||||
|
return False
|
||||||
|
|
||||||
class TestDownload(unittest.TestCase):
|
class TestDownload(unittest.TestCase):
|
||||||
def setUp(self) -> None:
|
def setUp(self) -> None:
|
||||||
@ -17,17 +30,27 @@ class TestDownload(unittest.TestCase):
|
|||||||
constant.CONFIG['cookie'] = os.getenv('NHENTAI_COOKIE')
|
constant.CONFIG['cookie'] = os.getenv('NHENTAI_COOKIE')
|
||||||
constant.CONFIG['useragent'] = os.getenv('NHENTAI_UA')
|
constant.CONFIG['useragent'] = os.getenv('NHENTAI_UA')
|
||||||
|
|
||||||
|
self.info = Doujinshi(**doujinshi_parser(did), name_format='%i')
|
||||||
|
|
||||||
def test_download(self):
|
def test_download(self):
|
||||||
did = 440546
|
info = self.info
|
||||||
info = Doujinshi(**doujinshi_parser(did), name_format='%i')
|
|
||||||
info.downloader = Downloader(path='/tmp', threads=5)
|
info.downloader = Downloader(path='/tmp', threads=5)
|
||||||
info.download()
|
info.download()
|
||||||
|
|
||||||
self.assertTrue(os.path.exists(f'/tmp/{did}/001.jpg'))
|
self.assertTrue(os.path.exists(f'/tmp/{did}/01.jpg'))
|
||||||
|
|
||||||
generate_html('/tmp', info)
|
generate_html('/tmp', info)
|
||||||
self.assertTrue(os.path.exists(f'/tmp/{did}/index.html'))
|
self.assertTrue(os.path.exists(f'/tmp/{did}/index.html'))
|
||||||
|
|
||||||
|
def test_zipfile_download(self):
|
||||||
|
info = self.info
|
||||||
|
info.downloader = CompressedDownloader(path='/tmp', threads=5)
|
||||||
|
info.download()
|
||||||
|
|
||||||
|
zipfile_path = f'/tmp/{did}.zip'
|
||||||
|
self.assertTrue(os.path.exists(zipfile_path))
|
||||||
|
self.assertTrue(is_zip_file(zipfile_path))
|
||||||
|
self.assertTrue(has_jepg_file(zipfile_path))
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
unittest.main()
|
unittest.main()
|
||||||
|
Reference in New Issue
Block a user