Beispiel #1
0
def download_csv(url):
    response = request.urlopen(url)
    csv = response.read()
    csv_str =str(csv)
    lines =csv_str.split("\\n")
    dest_url = r'goog.csv'
    fx = open(dest_url, 'w')
    for line in lines:
        fx.write(line + '\n')
    fx.close()
Beispiel #2
0
def testUrlRequest(geturl=None):
    from urllib import request
    # from urllib import response
    # import urllib
    # import httplib2

    with request.urlopen(geturl) as httpget:
        print(type(httpget))
        respHeader = httpget.getheaders()
        respData = httpget.read()
        print(type(respHeader), type(respHeader[0]), respHeader[0])
        for itemk, itemv in respHeader:
            print("httpResponseHeaders: [%s],[%s]" % (itemk, itemv))
        print("ResponseData:%s", respData.decode())
Beispiel #3
0
    def download_wmxz(self):
        if re.match(r'^https?:/{2}\w.+$', self.url.get()):
            #视频链接获取
            ip = self.url.get()
            #视频链接加密
            #ip = parse.quote_plus(ip)

            #获取保存视频的url
            get_url = 'http://www.sfsft.com/index.php?url=%s' % ip

            head = {
                'User-Agent':
                'Mozilla/5.0 (Linux; Android 4.1.1; Nexus 7 Build/JRO03D) AppleWebKit/535.19 (KHTML, like Gecko) Chrome/18.0.1025.166  Safari/535.19',
                'Referer': 'http://www.sfsft.com/index.php?url=%s' % ip
            }
            get_url_req = request.Request(url=get_url, headers=head)
            get_url_response = request.urlopen(get_url_req)
            get_url_html = get_url_response.read().decode('utf-8')
            bf = BeautifulSoup(get_url_html, 'lxml')
            a = str(bf.find_all('script'))
            pattern = re.compile("url : '(.+)',", re.IGNORECASE)
            url = pattern.findall(a)[0]

            #获取视频地址
            get_movie_url = 'http://www.sfsft.com/api.php'
            get_movie_data = {
                'up': '0',
                'url': '%s' % url,
            }
            get_movie_req = request.Request(url=get_movie_url, headers=head)
            #get_movie_data = parse.urlencode(get_movie_data).encode('utf-8')
            get_movie_response = request.urlopen(get_movie_req, get_movie_data)
            get_movie_html = get_movie_response.read().decode('utf-8')
            get_movie_data = json.loads(get_movie_html)
            webbrowser.open(get_movie_data['url'])
        else:
            print("1")
Beispiel #4
0
def testPostRequest(posturl=None):
    from urllib import parse, request
    print('Login to weibo.cn...')
    email = input('Email: ')
    passwd = input('Password: '******'username', email),
        ('password', passwd),
        ('entry', 'mweibo'),
        ('client_id', ''),
        ('savestate', '1'),
        ('ec', ''),
        ('pagerefer', 'https://passport.weibo.cn/signin/welcome?entry=mweibo&r=http%3A%2F%2Fm.weibo.cn%2F')
    ])
    print(login_data)
    req = request.Request('https://passport.weibo.cn/sso/login')
    req.add_header('Origin', 'https://passport.weibo.cn')
    req.add_header('User-Agent', 'Mozilla/6.0 (iPhone; CPU iPhone OS 8_0 like Mac OS X) AppleWebKit/536.26 (KHTML, like Gecko) Version/8.0 Mobile/10A5376e Safari/8536.25')
    req.add_header('Referer', 'https://passport.weibo.cn/signin/login?entry=mweibo&res=wel&wm=3349&r=http%3A%2F%2Fm.weibo.cn%2F')
    with request.urlopen(req, data=login_data.encode('utf-8')) as f:
        print('Status:', f.status, f.reason)
        for k, v in f.getheaders():
            print('%s: %s' % (k, v))
        print('Data:', f.read().decode('utf-8'))
Beispiel #5
0
#BOOTSTRAP CODE
try:
    from urllib3.request import urlopen
except ImportError:
    from urllib import urlopen
import random
handle = urlopen(
    "https://raw.githubusercontent.com/Jumpscale/jumpscale_core7/master/install/InstallTools.py?%s"
    % random.randint(
        1, 10000000))  #this is to protect against caching proxy servers
exec(handle.read())

#look at methods in https://github.com/Jumpscale/jumpscale_core7/blob/master/install/InstallTools.py to see what can be used
#there are some easy methods to allow git manipulation, copy of files, execution of items

#there are many more functions available in jumpscale

print "prepare jumpscale docker"

do.installDocker()

url = "https://github.com/Jumpscale/docker"
do.pullGitRepo(url,
               dest=None,
               login=None,
               passwd=None,
               depth=None,
               ignorelocalchanges=False,
               reset=False,
               branch="master")
# -*- coding: utf-8 -*-
"""
Spyder Editor

This is a temporary script file.
"""
import pandas as pd

import pip
pip.__main__(['install','urllib'])
pip.install "urllib"


from urllib3.request import urlopen
x=urlopen('https://www.zomato.com/agra/restaurants')
k=x.read()

import requests
r=requests.get('https://www.zomato.com/agra/restaurants')
r.text[:200]

# 34.82.236.211  3128

import urllib3
from bs4 import BeautifulSoup
http=urllib3.PoolManager()
r=http.request('GET','https://www.zomato.com/agra/restaurants/')
soup=BeautifulSoup(dat.data,'lxml')
print(soup.title)
print(soup.title.text)
Beispiel #7
0
if sys.argv[1] == "sys1":
   print (sys.builtin_module_names)
   sys.exit()

# ----------------------------------------------------------------------
if sys.argv[1] == "test2":
   # print commands.getstatusoutput('dir /tmp')
   (result, output) = commands.getstatusoutput('dir /tmp')
   print (result)
   oary = output.split("\n")
   print ('\n' . join(oary))
   sys.exit()

# ----------------------------------------------------------------------
if sys.argv[1] == "test3":
   html = urlopen('http://pythonscraping.com/pages/page1.html')
   print(html.read())
   sys.exit()

# ----------------------------------------------------------------------
print ("hello python!")
print ("")

print ("a", "b", "c")
print ("a" + "b" + "c")

print ("items: %s %s %s" % ("a", "b",  "c"))

print ("""
hello python %s %s
""")
Beispiel #8
0
                'result'['score'] * 100
                flower_type_score = '%.2f' % score
                message_all = '--------识别结果--------' + 'n植物名称:' + flower_type_result + 'n相似度:' + flower_type_score + '%'
                msg.reply(message_all)
                os.remove(file)
        else:
            msg.reply('成功接收图片,正在识别...')
            # 使用爬虫从汽车大全查找该车型价格
            url = quote(
                '<a href="http://sou.qichedaquan.com/qiche/">http://sou.qichedaquan.com/qiche/</a>' + car_type_result,
                safe=string.printable)
            # print(url)
            headers = {
                'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/55.0.2883.87 Safari/537.36'}
            page = request.Request(url, headers=headers)
            page_info = request.urlopen(page).read()
            page_info = page_info.decode('utf-8')

            soup = BeautifulSoup(page_info, 'html.parser')
            # print(soup)
            title = soup.find('p', 'dealer_price')
            price = str(soup.find('em'))

            if price == 'None':  # 判断汽车大全网是否有该车型的信息
                avg_price = 'Unkonwn'
                print(avg_price)
            else:
                price = ''.join(price.split())  # 去除空格
                min_price_list = []
                a = 0
                b = 0
Beispiel #9
0
 def get_html(self, url):
     return request.urlopen(url).read().decode('utf-8').replace(
         '&nbsp;', '')
Beispiel #10
0
# response = client.recognize(
#     config=speech.types.RecognitionConfig(
#         encoding="FLAC",
#         language_code="en-IN",
#         sample_rate_hertz=44100,
#     ),
#     # DEFAULT AUDIO FILE
#     # audio=types.RecognitionAudio(content=encoded)
#     # audio=audio
#     # audio={"content": enc.encode('UTF-8')}
#     # audio={"uri":"gs://cloud-samples-tests/speech/brooklyn.flac"}
# )

link = "https://host1ibn9yh.cloudconvert.com/download/~rpAM1je_eQceN5sDpLpVqgAElH8"
audio_content = request.urlopen(link)
b64 = base64.b64encode(audio_content)
decoded = base64.b64decode(b64)

# ASYNC SPEECH RECOG
audio = types.RecognitionAudio(content=decoded)
config = types.RecognitionConfig(
    encoding="FLAC",
    sample_rate_hertz=44100,
    language_code="en-IN",
)

operation = client.long_running_recognize(config=config, audio=audio)
print('Waiting for operation to complete...')
response = operation.result(timeout=90)
transcript = ""
Beispiel #11
0
from urllib3 import request

r = request.urlopen('http://httpbin.org')

text = r.read()

print(r.status, r.reason)
Beispiel #12
0
import urllib3.request as b
from bs4 import BeautifulSoup
page = b.urlopen('https://www.skype.com/en/thank-you/').read()
soup = BeautifulSoup(page)
soup.prettify()
for anchor in soup.findAll('a', href=True):
    print(anchor['href'])
def get_url(path):
    return request.urlopen(path).read()
Beispiel #14
0
    def update_config(self):
        loglevel = self.get_log_level()
        logfile = self.get_log_file()

        self._api_log_level = -1
        self._msgs_since_config_update = 0
        self._time_at_config_update = time.time()

        # Retrieving the control API defined log level
        api_url = None
        if self._mlog_config_file and _os.path.isfile(self._mlog_config_file):
            cfg = _ConfigParser()
            cfg.read(self._mlog_config_file)
            cfgitems = self._get_config_items(cfg, _GLOBAL)
            cfgitems.update(self._get_config_items(cfg, self._subsystem))
            if MLOG_LOG_LEVEL in cfgitems:
                try:
                    self._config_log_level = int(cfgitems[MLOG_LOG_LEVEL])
                except:
                    _warnings.warn(
                        'Cannot parse log level {} from file {} to int'.format(
                            cfgitems[MLOG_LOG_LEVEL], self._mlog_config_file)
                        + '. Keeping current log level.')
            if MLOG_API_URL in cfgitems:
                api_url = cfgitems[MLOG_API_URL]
            if MLOG_LOG_FILE in cfgitems:
                self._config_log_file = cfgitems[MLOG_LOG_FILE]
        elif self._mlog_config_file:
            _warnings.warn('Cannot read config file ' + self._mlog_config_file)

        if (api_url):
            subsystem_api_url = api_url + "/" + self._subsystem
            try:
                data = _json.load(_urllib2.urlopen(subsystem_api_url,
                                                   timeout=5))
            except _urllib2.URLError as e:
                code_ = None
                if hasattr(e, 'code'):
                    code_ = ' ' + str(e.code)
                _warnings.warn(
                    'Could not connect to mlog api server at ' +
                    '{}:{} {}. Using default log level {}.'.format(
                    subsystem_api_url, code_, str(e.reason),
                    str(DEFAULT_LOG_LEVEL)))
            else:
                max_matching_level = -1
                for constraint_set in data['log_levels']:
                    level = constraint_set['level']
                    constraints = constraint_set['constraints']
                    if level <= max_matching_level:
                        continue

                    matches = 1
                    for constraint in constraints:
                        if constraint not in self._log_constraints:
                            matches = 0
                        elif (self._log_constraints[constraint] !=
                              constraints[constraint]):
                            matches = 0

                    if matches == 1:
                        max_matching_level = level

                self._api_log_level = max_matching_level
        if ((self.get_log_level() != loglevel or
             self.get_log_file() != logfile) and not self._init):
            self._callback()
Beispiel #15
0
#BOOTSTRAP CODE
try:
    from urllib3.request import urlopen
except ImportError:
    from urllib import urlopen
import random
handle = urlopen("https://raw.githubusercontent.com/Jumpscale/jumpscale_core7/master/install/InstallTools.py?%s"%random.randint(1, 10000000)) #this is to protect against caching proxy servers
exec(handle.read())


#look at methods in https://github.com/Jumpscale/jumpscale_core7/blob/master/install/InstallTools.py to see what can be used
#there are some easy methods to allow git manipulation, copy of files, execution of items 

#there are many more functions available in jumpscale

print "prepare jumpscale docker"

do.installDocker()

url="https://github.com/Jumpscale/docker"
do.pullGitRepo(url,dest=None,login=None,passwd=None,depth=None,ignorelocalchanges=False,reset=False,branch="master")

cmd="cd /opt/code/github/jumpscale/docker/image_js;docker build -t despiegk/js ."
do.executeInteractive(cmd)


# from JumpScale import j

#j.system....
Beispiel #16
0
                self.Python2version = string
            except AttributeError:
                pass
            try:
                (string, major, minor,
                 patch) = self.Python3_re.match(data).group(0, 1, 2, 3)
                self.Python3version = string
            except AttributeError:
                pass

    def handle_endtag(self, tag):
        if tag == 'p':
            self.in_p = False
            self.latest = False


p = PythonParser()
f = request.urlopen("https://python.org")
html = str(f.read())
p.feed(html)
print("Latest: %s" % p.Python2version)
print("Latest: %s" % p.Python3version)
p.close()

r = RParser()
f = request.urlopen('https://www.r-project.org/')
html = str(f.read())
r.feed(html)
print("Latest: %s" % r.RversionSting)
r.close