def newMp4star(url): global qualityVideo nurl = url.split('/')[2] print(nurl) if not nurl.startswith('embed'): content = ccurlNew(url+'#'+'-I') if "Location:" in content: m = re.findall('Location: [^\n]*',content) url = re.sub('Location: |\r','',m[-1]) content = ccurlNew(url) soup = BeautifulSoup(content,'lxml') link = soup.find('video') link1 = link.findAll('source') final = '' n = [] for i in link1: n.append(i['src']) if n: if len(n) == 1: src = n[0] else: if qualityVideo == 'hd': src = n[0] else: src = n[-1] content = ccurlNew(src+'#'+'-I') if "Location:" in content: m = re.findall('Location: [^\n]*',content) final = re.sub('Location: |\r','',m[-1]) else: final = src return final
def uploadcrazy(url): content = ccurlNew(url) """ m = re.findall('file: "http[^"]*uploadcrazy.net[^"]*mp4[^"]*',content) if m: url = re.sub('file: "','',m[0]) else: url = "" """ url = mp4starUrl(content,'uploadcrazy') return url
def uploadcrazy(url): content = ccurlNew(url) """ m = re.findall('file: "http[^"]*uploadcrazy.net[^"]*mp4[^"]*',content) if m: url = re.sub('file: "','',m[0]) else: url = "" """ url = mp4starUrl(content, 'uploadcrazy') return url
def mp4star(url): hdr = "Mozilla/5.0 (X11; Ubuntu; Linux i686; rv:35.0) Gecko/20100101 Firefox/35.0" global qualityVideo m = [] #content = (subprocess.check_output(["curl","-L","-I","-A",hdr,url])) content = ccurlNew(url+'#'+'-I') found = "" if isinstance(content,bytes): print("I'm byte") content = str((content).decode('utf-8')) else: print(type(content)) content = str(content) print("I'm unicode") if "Location:" in content: m = re.findall('Location: [^\n]*',content) found = re.sub('Location: |\r','',m[-1]) print(found) if found: #content = ccurl(found,'') #content = (subprocess.check_output(["curl","-A",hdr,found])) content = ccurlNew(found) print(content) url1 = mp4starUrl(content,'mp4star') print(url1,'**********') content = ccurlNew(url1+'#'+'-I') if "Location:" in content: m = re.findall('Location: [^\n]*',content) found = re.sub('Location: |\r','',m[-1]) print(found) else: found = url1 url = str(urllib.parse.unquote(found)) return url
def simplyfind(i): content = ccurlNew(i) #replc = {' ':'%20', '[':'%5B', ']':'%5D','!':'%21'} m = re.findall('["]http://[^"]*.mp4[^"]*|["]http://[^"]*.flv[^"]*|["]https://redirector[^"]*|["]https://[^"]*.mp4', content) m1 = re.findall("[']http://[^']*.mp4[^']*|[']http://[^']*.flv[^']*|[']https://redirector[^']*|[']https://[^']*.mp4", content) print(m) if m: found = m[0] found = found[1:] found = str(urllib.parse.unquote(found)) #found = replace_all(found, replc) elif m1: found = m1[0] found = found[1:] found = str(urllib.parse.unquote(found)) #found = replace_all(found, replc) else: found = "" return found
def get_correct_mirror(self,m,mirrorNo): length = len(m) j = 1 final = '' while (j <= length): mirrorNo = mirrorNo - 1 msg = "Total " + str(len(m)) + " Mirrors, Selecting Mirror "+str(mirrorNo + 1) #subprocess.Popen(["notify-send",msg]) send_notification(msg) if mirrorNo < len(m): src = m[mirrorNo] else: src = m[0] print(src) if 'vidcrazy' in src or 'uploadcrazy' in src: final = uploadcrazy(src) elif 'vidkai' in src: final = findurl(src) elif 'mp4star' in src or 'justmp4' in src: try: content = ccurlNew(src) content = content.replace('\\\\','') print('????????????????????????????') print(content) print('>>>>>>>>>>>>>>>>>>>>>>>>>>>>>') final = mp4starUrl(content,'mp4star') except Exception as e: print(e,'getting next link') elif 'mp4buddy' in src or 'mp4upload' in src: final = findurl(src) print(final,'--final--') if final and final.startswith('http'): break j = j + 1 mirrorNo = j return final
def getEpnList(self,name,opt,depth_list,extra_info,siteName,category): m = [] if siteName == "Cartoon-World" or siteName == "Cartoon-World-Cartoon" or siteName == "Cartoon-World-Movies": base = "http://allmyanime.io/" url = base+ "watch/" + name+"/" elif siteName == "Dubcrazy": #base = "http://www.dubbedanimeonline.us/" base = "http://www.dubbedanimeon.me/" url = base+ "view/" + name+"/" elif siteName == "Animetycoon": base = "http://www.animetycoon.org/" url = base+ "watch/" + name+"/" elif siteName == "AniDub": base = "https://www.watchcartoononline.io/" if category == "Movie": url = "https://www.watchcartoononline.io/" + name else: url = "https://www.watchcartoononline.io/anime/" + name elif siteName == "AnimeStatic": base = "http://www.animestatic.co/" if category == "Movies": url = "http://www.animestatic.co/" + name + '/' else: url = "http://www.animestatic.co/anime/" + name + '/' elif siteName == "CartoonMax": url = "http://gogocartoon.us/category/" + name base = "http://gogocartoon.us/" print(url) #if base_url == 0: #content = subprocess.check_output(['curl','-A',hdr,url]) #else: # content = ccurl(url,"no_redir") if siteName == "Cartoon-World" or siteName == "Cartoon-World-Cartoon" or siteName == "Cartoon-World-Movies": content = ccurlNew(url+'#'+'-L') else: #content = ccurlNew(url+'#'+'-L') if siteName == 'Dubcrazy': content = ccurlRequest(url,method='GET',curl_opt='-L') else: content = ccurlNew(url+'#'+'-L') soup = BeautifulSoup(content,'lxml') print(soup.prettify()) if siteName == "Cartoon-World" or siteName == "Cartoon-World-Cartoon" or siteName == "Cartoon-World-Movies": """ link1 = soup.findAll('div',{'class':'ani-row'}) print(link1) try: img1 = link1[0].find('img',{'class':'anime'}) print(img1) img = img1['src'] if not "http://" in img: img2 = re.findall('/images/[^"]*',img) img = "http://allmyanime.io"+img2[0] print(img) #picn = "/tmp/AnimeWatch/"+name+'.jpg' picn = os.path.join(self.tmp_dir,name+'.jpg') if not os.path.isfile(picn) and img: #subprocess.call(["curl","-A",self.hdr,"-L","-o",picn,img]) ccurlNew(img+'#'+'-o'+'#'+picn) except: picn = "No.jpg" img = "" try: summary=str(link1[1]) summary = re.sub('</table>','</table><div class="desc">',summary) summary = re.sub('</div>','</div></div>',summary) print(summary) soup = BeautifulSoup(summary,'lxml') info = soup.findAll('td',{'class':'ani-table-ans'}) summary = info[0].text+'\nType: '+ info[1].text+ '\nAired: ' + info[2].text + '\nGenre: ' + info[3].text+soup.find('div',{'class':'desc'}).text except: summary = "No Summary Available" """ arr = [] try: summary = "" link = soup.find('p',{'class':'margin-top-10'}) #print(link summary = link.text link_img = soup.findAll('img',{'class':'img-responsive margin-top-10'}) #img = "http://www.dubbedanimeon.me/images/" + name+".jpg" img = link_img[-1]['src'] print(img) #picn = "/tmp/AnimeWatch/" + name + ".jpg" picn = os.path.join(self.tmp_dir,name+'.jpg') if not os.path.isfile(picn): #subprocess.call(["curl","-A",self.hdr,"-L","-o",picn,img]) ccurlNew(img+'#'+'-o'+'#'+picn) except: summary = "No Summary Available" picn = "No" m = soup.findAll('a',{'class':'subbed'}) n = soup.findAll('a',{'class':'dubbed'}) arr = [] for i in m: if 'href' in str(i): arr.append(i['href'].split('/')[-1]+'-subbed'+' '+i['href']) for i in n: if 'href' in str(i): arr.append(i['href'].split('/')[-1]+'-dubbed'+' '+i['href']) elif siteName == "AniDub" or siteName == "AnimeStatic": m = [] summary = '' if category == "Movies": m.append(name) else: if siteName == "AniDub": link = soup.findAll('div',{'id':'catlist-listview'}) else: link = soup.findAll('ul',{ 'class':'eps eps-list'}) for i in link: a = i.findAll('a') for j in a: k=(j['href']).split('/') if siteName == "AniDub": m.append(k[-1]) else: m.append(k[-2]) if siteName == "AniDub": img = [] link = soup.findAll('div',{'class':'iltext'}) for i in link: summary = re.sub('\n','',i.text) img = re.findall('http[^"]*.jpg',content) elif siteName == "AnimeStatic": link = soup.find("div",{ "class":"deskripsi"}) summary = "" img = [] if link: sumr= link.find('p') summary = sumr.text #summary = re.sub('Genres[^\n]*\n','Genres : ',summary) #summary = re.sub('Title[^\n]*\n','Title : ',summary) #summary = re.sub('Rating[^\n]*\n','Rating : ',summary) #summary = re.sub('[)]','',summary) #summary = re.sub('[,][^"]\n','\n',summary) link = soup.find('div',{'class':'imganime'}) if link: img1 = link.find('img') if img1: img.append(img1['src']) #picn = "/tmp/AnimeWatch/" + name + ".jpg" picn = os.path.join(self.tmp_dir,name+'.jpg') if not os.path.isfile(picn) and img: #subprocess.call(["curl","-A",self.hdr,"-L","-o",picn,img[0]]) ccurlNew(img[0]+'#'+'-o'+'#'+picn) elif siteName == "Animetycoon": img =[] #text = str(text) #print(text try: text = soup.find('article') text1 = text.find('p') summary = text1.text try: img1 = text.find('img')['src'] if 'http' not in img1: img1 = 'http:' + img1 img.append(img1) except: img = re.findall('//[^"]*posters/[^"]*.jpg',content) img[0] = "http:" + img[0] #picn = "/tmp/AnimeWatch/" + name + ".jpg" picn = os.path.join(self.tmp_dir,name+'.jpg') if not os.path.isfile(picn): #subprocess.call(["curl","-L","-o",picn,img[0]]) ccurlNew(img[0]+'#'+'-o'+'#'+picn) except: summary = "No Summary Available" picn = "No" elif siteName == "CartoonMax": m = [] link = soup.find('div',{'class':'list-chapter mCustomScrollbar'}) if link: j = link.findAll('a') for k in j: tmp = k['href'].split('/')[-1] m.append(tmp) else: link = soup.find('div',{'class':'anime_info_episodes'}) link1 = link.findAll('a') for i in link1: k = i['href'].split('/')[-1] m.append(k) summary = "" link = soup.find('div',{ 'class':'description'}) img = [] summary = link.text link = soup.find('div',{ 'class':'box-content'}) img1_src = link.find('div',{ 'class':'img'}) img_src = link.find('img')['src'] if ' ' in img_src: img_src = re.sub(" ","%20",img_src) print(img_src) if img_src: img.append(img_src) print(img) #picn = "/tmp/AnimeWatch/" + name + ".jpg" picn = os.path.join(self.tmp_dir,name+'.jpg') try: if not os.path.isfile(picn): ccurlNew(img[0]+'#'+'-o'+'#'+picn) except: pass elif siteName == "Dubcrazy": try: summary = "" link = soup.find('div',{'class':'well well-sm'}) #print(link summary = link.text link_img = soup.findAll('img',{'class':'img-responsive'}) #img = "http://www.dubbedanimeon.me/images/" + name+".jpg" img = link_img[-1]['src'] print(img) #picn = "/tmp/AnimeWatch/" + name + ".jpg" picn = os.path.join(self.tmp_dir,name+'.jpg') if not os.path.isfile(picn): #subprocess.call(["curl","-A",self.hdr,"-L","-o",picn,img]) ccurlNew(img+'#'+'-o'+'#'+picn) except: summary = "No Summary Available" picn = "No" #print(img sort_arr = True if siteName != "AniDub" and siteName != "CartoonMax": if (siteName == "Cartoon-World" or siteName == "Cartoon-World-Cartoon" or siteName == "Cartoon-World-Movies"): m = arr sort_arr = False else: fi = base + name+ '[^"]*/' m = re.findall(fi, content) found = False if not m: m = re.findall(name+'-dubbed-[^"]*', content) found = True j=0 if not found: for i in m: i = re.sub(base,"",i) m[j] = i[:-1] j = j + 1 if sort_arr: m=naturallysorted(m) #m.append(picn) #m.append(summary) record_history = True display_list = True return (m,summary,picn,record_history,depth_list)
def getCompleteList(self,siteName,category,opt): content = "" if siteName == "Cartoon-World": url = "http://allmyanime.io/anime-list/" elif siteName == "Cartoon-World-Movies": url = "http://allmyanime.io/movie-list/" elif siteName == "Cartoon-World-Cartoon": url = "http://allmyanime.io/cartoon-list/" elif siteName == "Dubcrazy": if category == "Movies": url = "http://www.dubbedanimeonline.us/dubbed-movies-list/" else: url = "http://www.dubbedanimeonline.us/dubbed-anime-list/" elif siteName == "Animetycoon": url = "http://www.animetycoon.org/full-index/" elif siteName == "AniDub": if category == "Movies": url = "https://www.watchcartoononline.io/movie-list" else: url = "https://www.watchcartoononline.io/dubbed-anime-list" urlc = "https://www.watchcartoononline.io/cartoon-list" elif siteName == "AnimeStatic": if category == "Movies": url = "http://www.animestatic.co/anime-movies/" else: url = "http://www.animestatic.co/anime-list/" elif siteName == "CartoonMax": url = "http://gogocartoon.us/cartoon-list.html" print(url) if siteName == "Animetycoon" or siteName == "AnimeStatic": #hdrs = {'user-agent':self.hdr} #req = requests.get(url,headers=hdrs) #content = str(req.text) content = ccurlNew(url) else: content = ccurlNew(url) soup = BeautifulSoup(content,'lxml') if siteName == "Cartoon-World" or siteName == "Cartoon-World-Movies" or siteName == "Cartoon-World-Cartoon" or siteName == "Dubcrazy" or siteName == "Animetycoon": if siteName == "Cartoon-World" or siteName == "Cartoon-World-Cartoon" or siteName == "Animetycoon" or siteName == "Cartoon-World-Movies": m = re.findall('watch/[^"]*', content) else: m = re.findall('view/[^"]*', content) #m = list(set(m)) #m.sort() j=0 for i in m: if siteName == "Cartoon-World" or siteName == "Cartoon-World-Cartoon" or siteName == "Animetycoon" or siteName == "Cartoon-World-Movies": i = re.sub("watch/","",i) else: i = re.sub("view/","",i) i = re.sub("/","",i) m[j] = i j = j+1 if opt == 7: strnamr = str(name) s = [] for i in m: m = re.search('[^"]*'+strname+'[^"]*',i) if m: found = m.group(0) s.append(found) m = s elif siteName == "AniDub" or siteName == "AnimeStatic": m = [] if siteName == "AniDub" and category != "Movies": content1 = ccurlNew(urlc) soup1 = BeautifulSoup(content1,'lxml') link1 = soup1.findAll('div',{'id':'ddmcc_container'}) link2 = soup.findAll('div',{'id':'ddmcc_container'}) link = link1 + link2 else: link = soup.findAll('div',{'id':'ddmcc_container'}) for i in link: a = i.findAll('a') for j in a: if 'href' in str(j): k=(j['href']).split('/') if siteName == "AniDub": if k[-1]: m.append(k[-1]) else: if k[-2]: m.append(k[-2]) elif siteName == "CartoonMax": m = [] soup = BeautifulSoup(content,'lxml') #link = soup.findAll('div',{'class':'anime_list_body'}) link = soup.findAll('div',{'class':'box-content list'}) #print(link for i in link: j = i.findAll('a') for k in j: tmp = k['href'].split('/')[-1] if tmp : m.append(tmp) if opt == "Random": m = random.sample(m, len(m)) return m
def getFinalUrl(self,siteName,name,epn,mirrorNo,quality): global qualityVideo,mirror_number mirror_number = mirrorNo qualityVideo = quality final = '' if siteName == "Cartoon-World" or siteName == "Cartoon-World-Cartoon" or siteName == "Cartoon-World-Movies": if ' ' in epn: epn = epn.split(' ')[1] url = "http://allmyanime.io" + epn content = ccurlNew(url) print(url) m = [] soup = BeautifulSoup(content,'lxml') link = soup.findAll('iframe') print(link) for i in link: if 'vidcrazy' in i['src'] or 'uploadcrazy' in i['src'] or 'mp4star' in i['src'] or 'justmp4' in i['src'] or 'gotvb8' in i['src'] or 'vidkai' in i['src'] or 'mp4buddy' in i['src'] or 'mp4upload' in i['src']: m.append(i['src']) print(m) final = self.get_correct_mirror(m,mirrorNo) elif siteName == "Animetycoon": url = "http://www.animetycoon.org/" + epn + "/" content = ccurlNew(url) print(url) final = "" m = re.findall('http://[^"]*uploadcrazy[^"]*|http://[^"]*vidkai[^"]*|http://[^"]*justmp4[^"]*|http://[^"]*mp4star[^"]*',content) print(m) arr = [] for i in m: if i.endswith('.jpg') or i.endswith('.jpeg'): pass else: arr.append(i) final = self.get_correct_mirror(arr,mirrorNo) elif siteName == "Dubcrazy": url = "http://www.dubbedanimeon.me/" + epn + "/" content = ccurlNew(url) print(url) m = [] n =[] soup = BeautifulSoup(content,'lxml') m = re.findall('http://[^"]*embed[^"]*',content) final = self.get_correct_mirror(m,mirrorNo) elif siteName == "AniDub": url = "https://www.watchcartoononline.io/" + epn print(url) content = ccurlNew(url) m = re.findall('["]https://[^"]*embed[^"]*',content) print(m) n = [] for i in m: j= i[1:] print(j) replc = {' ':'%20', '[':'%5B', ']':'%5D','!':'%21'} #j = str(urllib.parse.unquote(j)) n.append(j) print(n) post = 'confirm="Click+Here+to+Watch+Free!!"' for i in n: content = ccurlNew(i+'#'+'-d'+'#'+post) #print(content) m = re.findall('file: "http[^"]*',content) print(m) if m: final1 = re.findall('http://[^"]*',m[0]) if final1: print(final1[0],'++++++++++++++++++++++') k = final1[0] replc = {' ':'%20', '[':'%5B', ']':'%5D','!':'%21'} k = re.sub('\n','',k) content = ccurlNew(k+'#'+'-I') print(content,'-------------ccurlNew--------') n = re.findall('Location: [^\n]*',content) if n: final = re.sub('Location: |\r','',n[-1]) else: final = k print(final) final = re.sub(' ','%20',final) if final: break elif siteName == "AnimeStatic": url = "http://www.animestatic.co/" + epn + '/' print(url) content = ccurlNew(url) m = re.findall('["]http://[^"]*embed[^"]*',content) print(m) n = [] for i in m: j= i[1:] n.append(j) print(n) mirrorNo = mirrorNo -1 for i in n: url = n[mirrorNo] final = findurl(url) if final: break mirrorNo = mirrorNo + 1 elif siteName == "CartoonMax": final = '' url = "http://gogocartoon.us/" + epn print(url,'----------') content = ccurlNew(url) soup = BeautifulSoup(content,'lxml') #link = soup.find('div',{'class':'anime_video_body_watch'}) #if not link: #link = soup.find('div',{'class':'anime_video_body_watch_items_2'}) link = soup.find('div',{'class':'main-video'}) #print(link) sd = '' hd = '' sd480 = '' if link: link2 = link.find('iframe') print(link2) if link2: if 'src' in str(link2): link1 = link2['src'] print(link1,'---') if link1: if ' ' in str(link1): link1 = re.sub(' ','%20',str(link1)) print(link1) content1 = ccurlNew(link1) soup = BeautifulSoup(content1,'lxml') links = soup.findAll('source') for i in links: if 'src' in str(i): j = i['src'] if 'itag=22' in j: hd = j elif 'itag=18' in j: sd = j elif 'itag=59' in j: sd480 = j elif 'itag=43' in j: sd = j print (sd) print(sd480) print(hd) if not sd and not hd and not sd480: soup = BeautifulSoup(content,'lxml') link = soup.find('select',{'id':'selectQuality'}) if link: link1 = link.findAll('option') for i in link1: j = i['value'] if 'itag=18' in j: sd = j elif 'itag=22' in j: hd = j elif 'itag=37' in j: full_hd = j elif '=m18' in j: sd = j elif '=m22' in j: hd = j final_cnt = 0 final_quality = '' if sd: final_cnt = final_cnt+1 final_quality = final_quality + 'SD ' if sd480: final_cnt = final_cnt+1 final_quality = final_quality + '480P ' if hd: final_cnt = final_cnt+1 final_quality = final_quality + 'HD ' msg = "Total " + str(final_cnt) + " Quality Video Available "+final_quality+" Selecting "+str(quality) + " Quality" #subprocess.Popen(["notify-send",msg]) send_notification(msg) if quality == "sd": final_q = sd elif quality == 'sd480p': final_q = sd480 elif quality == 'hd': final_q = hd if not final_q and sd: final_q = sd print(final_q) if final_q: #content = (subprocess.check_output(['curl','-L','-I','-A',self.hdr,final_q])) #content = self.getContent(content) content = ccurlNew(final_q+'#'+'-I') print(content) m = re.findall('Location: https[^\n]*', content) #print(m if m: #print(m final = m[0] final = re.sub('Location: |\r', '', final) else: final = '' return final
def findurl(url): final = "" hdr = 'Mozilla/5.0 (X11; Ubuntu; Linux i686; rv:45.0) Gecko/20100101 Firefox/45.0' print('--------findurl-----',url) if "myvidstream" in str(url): #print("myvidstream="+url packed = '' content = ccurlNew(url) final = mp4starUrl(content,'myvidstream') elif 'mp4star' in str(url) or 'justmp4' in str(url): final = mp4star(url) elif "vidup" in str(url): m = re.findall('http://[^"]*&bg',url) m1 = re.sub('&bg*','',m[0]) final = simplyfind(m1) elif 'uploadcrazy' in str(url): print('------Inside Uploadcrazy----') content = ccurlNew(url) #print(content+'\n') print('-------------uploadcrazy---------') m = re.findall('file: "http[^"]*uploadcrazy.net[^"]*mp4[^"]*',content) if m: final = re.sub('file: "','',m[0]) else: final = "" elif "yourupload" in str(url): i = url.replace(r'#038;','') #content = subprocess.check_output(["curl","-L","-A",hdr,i]) content = ccurlNew(i) m = re.findall("file: 'http://[^']*video.mp4",content) print(m) if m: url = re.sub("file: '","",m[0]) else: url = "" print("File Does Not exist") print(url) #content = (subprocess.check_output(["curl","-L","-I","-A",hdr,"-e",i,url])) content = ccurlNew(url+'#'+'-Ie'+'#'+i) if "Location:" in content: m = re.findall('Location: [^\n]*',content) found = re.sub('Location: |\r','',m[-1]) print(found) url = found return url elif "vidkai" in str(url): content = ccurlNew(url) #print(content soup = BeautifulSoup(content,'lxml') src = soup.find('source')['src'] if src: #content = (subprocess.check_output(['curl','-I','-L','-A',hdr,src])) #print(content content = ccurlNew(src+'#'+'-I') if "Location:" in content: m = re.findall('Location: [^\n]*',content) final = re.sub('Location: |\r','',m[-1]) print(final) else: final = "" elif "uploadc" in str(url): content = ccurlNew(url) replc = {' ':'%20', '[':'%5B', ']':'%5D','!':'%21'} m = re.findall("[']http://[^']*.mp4",content) print(m) if m: #final = replace_all(m[0], replc) final = str(urllib.parse.unquote(m[0])) final = re.sub("[']",'',final) final = final + "?start=0" print(final) elif 'mp4buddy' in str(url): global mirror_number content = ccurlNew(url) soup = BeautifulSoup(content,'lxml') links = soup.findAll('div',{'class':'pestana'}) arr = [] content_new = '' for i in links: l = i['data-src'] k = str(base64.b64decode(l).decode('utf-8')) arr.append(k) print(arr) if arr: if mirror_number < len(arr): url = arr[mirror_number-1] else: url = arr[0] content = ccurlNew(url) soup = BeautifulSoup(content,'lxml') m = soup.findAll('script,{"type":"text/javascript"}') if not m: m = soup.findAll('script') print(m) for i in m: if 'eval(' in i.text and ('https' in i.text or 'http' in i.text): print(i.text) content_new = i.text break if content_new: content_arr = content_new.split('|') content_arr = content_arr[10:] print(content_arr) for p in range(0,5): max_ = 0 max_len = '' j = 0 mark_val = 0 for i in content_arr: if len(i) >= max_: max_len = i max_ = len(i) mark_val = j j = j+1 print(max_len,max_) del content_arr[mark_val] final = 'http://mp4buddy.com/api/'+max_len print(final) content = ccurlNew(final) print(content) js = json.loads(content) final = js['playlist'][0]['sources'][0]['file'] try: #info = ccurlRequest(final,method='HEAD',curl_opt='-I') info = ccurlNew(final+'#'+'-I') print(info,'----------------') if 'HTTP/1.1 503' in info or 'HTTP/1.1 200' in info: print(final) break #if info['Content-Type'] == 'application/octet-stream' or info['Content-Type'] == 'video/mp4': # print(final) # break #if info == 200 or info == 503: # print(final) # break except Exception as e: print(e) print(final) elif "mp4upload" in str(url): content = ccurlRequest(url,method='GET') m = re.findall("'file': 'https://[^']*mp4upload.com[^']*video.mp4",content) print(m) if m: url = re.sub("'file': '","",m[0]) else: m = re.findall('"file": "https://[^"]*mp4upload.com[^"]*video.mp4',content) if m: url = re.sub('"file": "',"",m[0]) else: url = "" print("File Does Not exist") print(url) final = url return final
def uploadcrazy(url): content = ccurlNew(url) url = mp4starUrl(content,'uploadcrazy') return url
def findurl(url): final = "" hdr = 'Mozilla/5.0 (X11; Ubuntu; Linux i686; rv:45.0) Gecko/20100101 Firefox/45.0' print('--------findurl-----',url) if "myvidstream" in str(url): #print("myvidstream="+url packed = '' content = ccurlNew(url) final = mp4starUrl(content,'myvidstream') """ soup = BeautifulSoup(content,'lxml') link = soup.findAll('script',{'type':'text/javascript'}) for i in link: if 'myvidstream' in i.text: #print(i.text) packed = i.text break if packed: val = packer.unpack(packed) print(val) soup = BeautifulSoup(val,'lxml') m = (re.search('file[^)]*',val)).group() print(m) if m: n = re.search("http[^']*",m).group() if n: print(n) final = n final = re.sub(' |"','',final) final = re.sub("'",'',final) fi = final.split('\\') if fi: final = fi[0] """ elif 'mp4star' in str(url) or 'justmp4' in str(url): final = mp4star(url) elif "vidup" in str(url): m = re.findall('http://[^"]*&bg',url) m1 = re.sub('&bg*','',m[0]) final = simplyfind(m1) elif 'uploadcrazy' in str(url): print('------Inside Uploadcrazy----') content = ccurlNew(url) #print(content+'\n') print('-------------uploadcrazy---------') m = re.findall('file: "http[^"]*uploadcrazy.net[^"]*mp4[^"]*',content) if m: final = re.sub('file: "','',m[0]) else: final = "" elif "yourupload" in str(url): i = url.replace(r'#038;','') #content = subprocess.check_output(["curl","-L","-A",hdr,i]) content = ccurlNew(i) m = re.findall("file: 'http://[^']*video.mp4",content) print(m) if m: url = re.sub("file: '","",m[0]) else: url = "" print("File Does Not exist") print(url) #content = (subprocess.check_output(["curl","-L","-I","-A",hdr,"-e",i,url])) content = ccurlNew(url+'#'+'-Ie'+'#'+i) if "Location:" in content: m = re.findall('Location: [^\n]*',content) found = re.sub('Location: |\r','',m[-1]) print(found) url = found return url elif "vidkai" in str(url): content = ccurlNew(url) #print(content soup = BeautifulSoup(content,'lxml') src = soup.find('source')['src'] if src: #content = (subprocess.check_output(['curl','-I','-L','-A',hdr,src])) #print(content content = ccurlNew(src+'#'+'-I') if "Location:" in content: m = re.findall('Location: [^\n]*',content) final = re.sub('Location: |\r','',m[-1]) print(final) else: final = "" elif "uploadc" in str(url): content = ccurlNew(url) replc = {' ':'%20', '[':'%5B', ']':'%5D','!':'%21'} m = re.findall("[']http://[^']*.mp4",content) print(m) if m: #final = replace_all(m[0], replc) final = str(urllib.parse.unquote(m[0])) final = re.sub("[']",'',final) final = final + "?start=0" print(final) return final
def mp4star(url): hdr = "Mozilla/5.0 (X11; Ubuntu; Linux i686; rv:35.0) Gecko/20100101 Firefox/35.0" global qualityVideo m = [] #content = (subprocess.check_output(["curl","-L","-I","-A",hdr,url])) content = ccurlNew(url+'#'+'-I') found = "" if isinstance(content,bytes): print("I'm byte") content = str((content).decode('utf-8')) else: print(type(content)) content = str(content) print("I'm unicode") if "Location:" in content: m = re.findall('Location: [^\n]*',content) found = re.sub('Location: |\r','',m[-1]) print(found) if found: #content = ccurl(found,'') #content = (subprocess.check_output(["curl","-A",hdr,found])) content = ccurlNew(found) print(content) url1 = mp4starUrl(content,'mp4star') print(url1,'**********') content = ccurlNew(url1+'#'+'-I') if "Location:" in content: m = re.findall('Location: [^\n]*',content) found = re.sub('Location: |\r','',m[-1]) print(found) else: found = url1 """ m = re.findall('value="[^"]*',content) value = re.sub('value="',"",m[0]) #print(value #content = (subprocess.check_output(["curl","-A",hdr,'-d','id='+'"'+value+'"',found])) content = ccurl(found,value) if isinstance(content,bytes): print("I'm byte") try: content = str((content).decode('utf-8')) except: content = str(content) else: print(type(content)) content = str(content) print("I'm unicode") #content = ccurl(found,value) """ """ else: content = ccurl(url,'') m = re.findall('value="[^"]*',content) value = re.sub('value="',"",m[0]) #print(value content = ccurl(url,value) #print(content #m = re.findall('https[^"]*googleusercontent[^"]*',content) if 'mp4star' in url: m = re.findall('https[^"]*redirector[^"]*',content) else: m = re.findall("file: 'http[^']*",content) n = [] for i in m: i = i.replace("file: '",'') n.append(i) m[:]=[] m = n print(m) if m: #content = ccurl(m[0],"") if qualityVideo == "sd": #content = (subprocess.check_output(["curl","-L","-I","-A",hdr,m[0]])) content = ccurlNew(m[0]+'#'+'-I') else: #content = (subprocess.check_output(["curl","-L","-I","-A",hdr,m[-1]])) content = ccurlNew(m[-1]+'#'+'-I') if "Location:" in content: m = re.findall('Location: [^\n]*',content) found = re.sub('Location: |\r','',m[-1]) print(found) else: url1 = mp4starUrl(content,'mp4star') print(url1,'**********') content = ccurlNew(url1+'#'+'-I') if "Location:" in content: m = re.findall('Location: [^\n]*',content) found = re.sub('Location: |\r','',m[-1]) print(found) """ url = str(urllib.parse.unquote(found)) return url
def findurl(url): final = "" hdr = 'Mozilla/5.0 (X11; Ubuntu; Linux i686; rv:45.0) Gecko/20100101 Firefox/45.0' print('--------findurl-----', url) if "myvidstream" in str(url): #print("myvidstream="+url packed = '' content = ccurlNew(url) final = mp4starUrl(content, 'myvidstream') """ soup = BeautifulSoup(content,'lxml') link = soup.findAll('script',{'type':'text/javascript'}) for i in link: if 'myvidstream' in i.text: #print(i.text) packed = i.text break if packed: val = packer.unpack(packed) print(val) soup = BeautifulSoup(val,'lxml') m = (re.search('file[^)]*',val)).group() print(m) if m: n = re.search("http[^']*",m).group() if n: print(n) final = n final = re.sub(' |"','',final) final = re.sub("'",'',final) fi = final.split('\\') if fi: final = fi[0] """ elif 'mp4star' in str(url) or 'justmp4' in str(url): final = mp4star(url) elif "vidup" in str(url): m = re.findall('http://[^"]*&bg', url) m1 = re.sub('&bg*', '', m[0]) final = simplyfind(m1) elif 'uploadcrazy' in str(url): print('------Inside Uploadcrazy----') content = ccurlNew(url) #print(content+'\n') print('-------------uploadcrazy---------') m = re.findall('file: "http[^"]*uploadcrazy.net[^"]*mp4[^"]*', content) if m: final = re.sub('file: "', '', m[0]) else: final = "" elif "yourupload" in str(url): i = url.replace(r'#038;', '') #content = subprocess.check_output(["curl","-L","-A",hdr,i]) content = ccurlNew(i) m = re.findall("file: 'http://[^']*video.mp4", content) print(m) if m: url = re.sub("file: '", "", m[0]) else: url = "" print("File Does Not exist") print(url) #content = (subprocess.check_output(["curl","-L","-I","-A",hdr,"-e",i,url])) content = ccurlNew(url + '#' + '-Ie' + '#' + i) if "Location:" in content: m = re.findall('Location: [^\n]*', content) found = re.sub('Location: |\r', '', m[-1]) print(found) url = found return url elif "vidkai" in str(url): content = ccurlNew(url) #print(content soup = BeautifulSoup(content, 'lxml') src = soup.find('source')['src'] if src: #content = (subprocess.check_output(['curl','-I','-L','-A',hdr,src])) #print(content content = ccurlNew(src + '#' + '-I') if "Location:" in content: m = re.findall('Location: [^\n]*', content) final = re.sub('Location: |\r', '', m[-1]) print(final) else: final = "" elif "uploadc" in str(url): content = ccurlNew(url) replc = {' ': '%20', '[': '%5B', ']': '%5D', '!': '%21'} m = re.findall("[']http://[^']*.mp4", content) print(m) if m: #final = replace_all(m[0], replc) final = str(urllib.parse.unquote(m[0])) final = re.sub("[']", '', final) final = final + "?start=0" print(final) return final
def mp4star(url): hdr = "Mozilla/5.0 (X11; Ubuntu; Linux i686; rv:35.0) Gecko/20100101 Firefox/35.0" global qualityVideo m = [] #content = (subprocess.check_output(["curl","-L","-I","-A",hdr,url])) content = ccurlNew(url + '#' + '-I') found = "" if isinstance(content, bytes): print("I'm byte") content = str((content).decode('utf-8')) else: print(type(content)) content = str(content) print("I'm unicode") if "Location:" in content: m = re.findall('Location: [^\n]*', content) found = re.sub('Location: |\r', '', m[-1]) print(found) if found: #content = ccurl(found,'') #content = (subprocess.check_output(["curl","-A",hdr,found])) content = ccurlNew(found) print(content) url1 = mp4starUrl(content, 'mp4star') print(url1, '**********') content = ccurlNew(url1 + '#' + '-I') if "Location:" in content: m = re.findall('Location: [^\n]*', content) found = re.sub('Location: |\r', '', m[-1]) print(found) else: found = url1 """ m = re.findall('value="[^"]*',content) value = re.sub('value="',"",m[0]) #print(value #content = (subprocess.check_output(["curl","-A",hdr,'-d','id='+'"'+value+'"',found])) content = ccurl(found,value) if isinstance(content,bytes): print("I'm byte") try: content = str((content).decode('utf-8')) except: content = str(content) else: print(type(content)) content = str(content) print("I'm unicode") #content = ccurl(found,value) """ """ else: content = ccurl(url,'') m = re.findall('value="[^"]*',content) value = re.sub('value="',"",m[0]) #print(value content = ccurl(url,value) #print(content #m = re.findall('https[^"]*googleusercontent[^"]*',content) if 'mp4star' in url: m = re.findall('https[^"]*redirector[^"]*',content) else: m = re.findall("file: 'http[^']*",content) n = [] for i in m: i = i.replace("file: '",'') n.append(i) m[:]=[] m = n print(m) if m: #content = ccurl(m[0],"") if qualityVideo == "sd": #content = (subprocess.check_output(["curl","-L","-I","-A",hdr,m[0]])) content = ccurlNew(m[0]+'#'+'-I') else: #content = (subprocess.check_output(["curl","-L","-I","-A",hdr,m[-1]])) content = ccurlNew(m[-1]+'#'+'-I') if "Location:" in content: m = re.findall('Location: [^\n]*',content) found = re.sub('Location: |\r','',m[-1]) print(found) else: url1 = mp4starUrl(content,'mp4star') print(url1,'**********') content = ccurlNew(url1+'#'+'-I') if "Location:" in content: m = re.findall('Location: [^\n]*',content) found = re.sub('Location: |\r','',m[-1]) print(found) """ url = str(urllib.parse.unquote(found)) return url