]> git.lizzy.rs Git - nhentai.git/blobdiff - nhentai/downloader.py
Fix
[nhentai.git] / nhentai / downloader.py
index 3d2db81d13b8e113f5541584fc85f80b848a27c9..f11bb60b1682c1c046032864a99d86155703f7ef 100644 (file)
@@ -1,14 +1,14 @@
 # coding: utf-
 from __future__ import unicode_literals, print_function
 
+import multiprocessing
 import signal
 
 from future.builtins import str as text
+import sys
 import os
 import requests
-import threadpool
 import time
-import multiprocessing as mp
 
 try:
     from urllib.parse import urlparse
@@ -17,36 +17,21 @@ except ImportError:
 
 from nhentai.logger import logger
 from nhentai.parser import request
-from nhentai.utils import Singleton, signal_handler
+from nhentai.utils import Singleton
 
 requests.packages.urllib3.disable_warnings()
-semaphore = mp.Semaphore()
+semaphore = multiprocessing.Semaphore(1)
 
 
 class NHentaiImageNotExistException(Exception):
     pass
 
 
-class Pool(Singleton):
-    pool = None
-
-    def __init__(self, size, init):
-        if self.pool is None:
-            if os.getenv('DEBUG'):
-                logger.info('Process pool created')
-
-            self.pool = mp.Pool(size, initializer=init)
-
-
 class Downloader(Singleton):
 
-    def __init__(self, path='', thread=1, timeout=30, delay=0):
-        if not isinstance(thread, (int, )) or thread < 1 or thread > 15:
-            raise ValueError('Invalid threads count')
-
+    def __init__(self, path='', size=5, timeout=30, delay=0):
+        self.size = size
         self.path = str(path)
-        self.thread_count = thread
-        self.threads = []
         self.timeout = timeout
         self.delay = delay
 
@@ -140,22 +125,21 @@ class Downloader(Singleton):
                 os.makedirs(folder)
             except EnvironmentError as e:
                 logger.critical('{0}'.format(str(e)))
-                exit(1)
+
         else:
             logger.warn('Path \'{0}\' already exist.'.format(folder))
 
         queue = [(self, url, folder) for url in queue]
 
-        pool = Pool(self.thread_count, init_worker).pool
-        for item in queue:
-            pool.apply_async(download_wrapper, args=item, callback=self._download_callback)
+        pool = multiprocessing.Pool(self.size, init_worker)
+        [pool.apply_async(download_wrapper, args=item) for item in queue]
 
         pool.close()
         pool.join()
 
 
 def download_wrapper(obj, url, folder=''):
-    if semaphore.get_value():
+    if sys.platform == 'darwin' or semaphore.get_value():
         return Downloader.download_(obj, url=url, folder=folder)
     else:
         return -3, None