def get_movie_links(self, url): print 'Get movie links: {url}'.format(url=url) data = util.get_remote_data(url) soup = BeautifulSoup(data, 'html.parser', parse_only=self.SoupStrainer_Movie_Link ) items = [] for item in soup.findAll('a', href=True): lower = item.text.lower() if 'in' in lower and 'now' in lower: label = util.encode(item.text) link = util.encode(item['href']) pk = label items.append({ 'label': label, 'url': link, 'pk': pk, 'is_playable': True }) return items
def get_movie_links(self, url): print 'Get movie links: {url}'.format(url=url) data = util.get_remote_data(url) soup = BeautifulSoup(data, 'html.parser', parse_only=self.SoupStrainer_Movie_Link) items = [] for item in soup.findAll('p'): if item.strong and item.a: href = item.a.get('href', None) if href: label = util.encode(item.strong.text) link = util.encode(href) pk = label items.append({ 'label': label, 'url': link, 'pk': pk, 'is_playable': True }) return items
def get_movie_links(self, url): print 'Get movie links: {url}'.format(url=url) data = util.get_remote_data(url) soup = BeautifulSoup(data, 'html.parser', parse_only=self.SoupStrainer_Movie_Link) items = [] for item in soup.findAll('a', href=True): lower = item.text.lower() if 'in' in lower and 'now' in lower: label = util.encode(item.text) link = util.encode(item['href']) pk = label items.append({ 'label': label, 'url': link, 'pk': pk, 'is_playable': True }) return items
def get_movie_links(self, url): print 'Get movie links: {url}'.format(url=url) data = util.get_remote_data(url) soup = BeautifulSoup(data, 'html.parser', parse_only=self.SoupStrainer_Movie_Link ) items = [] for item in soup.findAll('p'): if item.strong and item.a: href = item.a.get('href', None) if href: label = util.encode(item.strong.text) link = util.encode(href) pk = label items.append({ 'label': label, 'url': link, 'pk': pk, 'is_playable': True }) return items
def get_menu_category(self, api): ''' Get main list of categories ''' print 'Get list categories' url = self.BASE_URL data = util.get_remote_data(url) soup = BeautifulSoup(data, 'html.parser', parse_only=self.SoupStrainer_Category) items = [] for item in soup.findAll('li'): if item.a.has_attr('href'): link = util.encode(item.a['href']) pk = item['id'] # ignore invalid links if 'category/' not in link: continue items.append({ 'label': item.a.text, 'url': link, 'pk': pk, }) return items
def get_movie_links(self, url): print 'Get movie links: {url}'.format(url=url) data = util.get_remote_data(url) soup = BeautifulSoup(data, 'html.parser', parse_only=self.SoupStrainer_Movie_Link) items = [] pk_regex = re.compile('http://([\w\.]+)\/(?:([\w-]+)\/|)') for a in soup: if ('Full' in a.text or \ 'Play' in a.text) and \ 'Online' in a.text: link = util.encode(a['href']) match = pk_regex.search(link) if match: group1 = match.group(1) group2 = match.group(2) label = group2 if group2 else group1 pk = label items.append({ 'label': label, 'url': link, 'pk': pk, 'is_playable': True }) return items
def get_menu_category(self, api): ''' Get main list of categories ''' print 'Get list categories' url = self.BASE_URL data = util.get_remote_data(url) soup = BeautifulSoup(data, 'html.parser', parse_only=self.SoupStrainer_Category ) items = [] pk_regex = re.compile('\/([\w\-]+)\/') for item in soup.findAll('li'): if item.a.has_attr('href'): link = util.encode(item.a['href']) # ignore invalid links if 'category/' not in link: continue pk = pk_regex.search(item.a['href']).group(1) items.append({ 'label': item.a.text, 'url': link, 'pk': pk, }) return items
def get_movie_links(self, url): print 'Get movie links: {url}'.format(url=url) data = util.get_remote_data(url) soup = BeautifulSoup(data, 'html.parser', parse_only=self.SoupStrainer_Movie_Link ) items = [] pk_regex = re.compile('http://([\w\.]+)\/(?:([\w-]+)\/|)') for a in soup: if ('Full' in a.text or \ 'Play' in a.text) and \ 'Online' in a.text: link = util.encode(a['href']) match = pk_regex.search(link) if match: group1 = match.group(1) group2 = match.group(2) label = group2 if group2 else group1 pk = label items.append({ 'label': label, 'url': link, 'pk': pk, 'is_playable': True }) return items
def get_menu_movies(self, url): ''' Get movie titles for category ''' print 'Get list movies: {url}'.format(url=url) data = util.get_remote_data(url) # Get list of movie titles soup = BeautifulSoup(data, 'html.parser', parse_only=self.SoupStrainer_Movies) items = [] pk_regex = re.compile('\/([\w\-]+)\/') for item in soup: img = item.a.img thumb = util.encode(img['src']) if img else '' link = util.encode(item.a['href']) txt = item.text.strip() or item.a.get('title', None) info = util.encode(txt.strip()) label = info pk = pk_regex.search(item.a['href']).group(1) items.append({ 'label': label, 'url': link, 'thumb': thumb, 'info': info, 'pk': pk, 'is_playable': False }) return items
def get_menu_movies(self, url): ''' Get movie titles for category ''' print 'Get list movies: {url}'.format(url=url) data = util.get_remote_data(url, False) # Get list of movie titles soup = BeautifulSoup(data, 'html.parser', parse_only=self.SoupStrainer_Movies ) items = [] pk_regex = re.compile('\/([\w\-]+)\/') for item in soup: img = item.a.img thumb = util.encode(img['src']) if img else '' link = util.encode(item.a['href']) txt = item.text.strip() or item.a.get('title', None) info = util.encode(txt.strip()) label = info pk = pk_regex.search(item.a['href']).group(1) items.append({ 'label': label, 'url': link, 'thumb': thumb, 'info': info, 'pk': pk, 'is_playable': False }) return items