#VERSION: 1.23 #AUTHORS: Douman (custparasite@gmx.se) #CONTRIBUTORS: Diego de las Heras (ngosang@hotmail.es) # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # # * Redistributions of source code must retain the above copyright notice, # this list of conditions and the following disclaimer. # * Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # * Neither the name of the author nor the names of its contributors may be # used to endorse or promote products derived from this software without # specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" # AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE # IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE # ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE # LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR # CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF # SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS # INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN # CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) # ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE # POSSIBILITY OF SUCH DAMAGE. from html.parser import HTMLParser from re import compile as re_compile from re import DOTALL from itertools import islice #qBt from novaprinter import prettyPrinter from helpers import download_file, retrieve_url class demonoid(object): """ Search engine class """ url = "https://www.demonoid.pw" name = "Demonoid" supported_categories = {'all': '0', 'music': '2', 'movies': '1', 'games': '4', 'software': '5', 'books': '11', 'anime': '9', 'tv': '3'} def download_torrent(self, info): """ Downloader """ print(download_file(info)) class MyHtmlParseWithBlackJack(HTMLParser): """ Parser class """ def __init__(self, url): HTMLParser.__init__(self) self.url = url self.current_item = None self.save_data = None self.seeds_leech = False self.size_repl = re_compile(",") def handle_starttag(self, tag, attrs): """ Parser's start tag handler """ if tag == "a": params = dict(attrs) if "href" in params: link = params["href"] if link.startswith("/files/details"): self.current_item = dict() self.current_item["desc_link"] = "".join((self.url, link)) self.current_item["engine_url"] = self.url self.save_data = "name" elif link.startswith("/files/download"): self.current_item["link"] = "".join((self.url, link)) elif self.current_item: if tag == "td": params = dict(attrs) if "class" in params and "align" in params: if params["class"].startswith("tone"): if params["align"] == "right": self.save_data = "size" elif params["align"] == "center": self.seeds_leech = True elif self.seeds_leech and tag == "font": for attr in attrs: if "class" in attr: if attr[1] == "green": self.save_data = "seeds" elif attr[1] == "red": self.save_data = "leech" self.seeds_leech = False def handle_data(self, data): """ Parser's data handler """ if self.save_data: if self.save_data == "name": # names with special characters like '&' are splitted in several pieces if 'name' not in self.current_item: self.current_item['name'] = '' self.current_item['name'] += data else: self.current_item[self.save_data] = data self.save_data = None if self.current_item.__len__() == 7: self.current_item["size"] = self.size_repl.sub("", self.current_item["size"]) prettyPrinter(self.current_item) self.current_item = None def handle_endtag(self, tag): """ Parser's end tag handler """ if self.save_data == "name": self.save_data = None def search(self, what, cat='all'): """ Performs search """ #prepare query cat = self.supported_categories[cat.lower()] query = "".join((self.url, "/files/?category=", cat, "&subcategory=All&quality=All&seeded=2&external=2&query=", what, "&uid=0&sort=S")) data = retrieve_url(query) add_res_list = re_compile("/files.*page=[0-9]+") torrent_list = re_compile("start torrent list -->(.*)