コード例 #1
0
 def fetchMedia(self):
     f = Fetch(self.mediaUrl)
     if f.connected:
         self.media = f.site
         self.mediaData = self.media.read()
         self.urlinfo = self.media.info()
         self.fetched = True
コード例 #2
0
 def load(self, loadUrl):
     o = urlparse(loadUrl)
     print o.scheme
     print o.port
     print o
     o = urlsplit(loadUrl)
     print o.geturl()
     reader = Fetch()
     htmlFile = reader.get(loadUrl)
     soup = BeautifulSoup(htmlFile)
     form = soup.find('form')
     print form
     print form['action']
     print form['method']
     #print form['onsubmit']
     if 'action' in form:
         print form['action']
     if 'method' in form:
         print form['method']
     if 'onsubmit' in form:
         print form['onsubmit']
     data = {"username": "******", "password": "******"}
     response = reader.post('https://210.154.183.61:4443/mantis/login.php',
                            data)
     print response
コード例 #3
0
 def __init__(self, auth:str, logger=logging.getLogger()):
     self.auth = auth
     self.logger = logger
     self.fetch = Fetch(auth, logger)
     self.user = self.fetch.fetch("/v1/me")
     self.userId = self.user["id"]
     self.imageUrl = self.user["images"][0]["url"]
     self.market = self.user["country"]
     logger.debug("Initialized Spotify: {0} @ {1}".format(self.userId, self.market))
     self.fetch.fetch("/v1/me/tracks", limit=1, market=self.market)
コード例 #4
0
    def __init__(self):

        self.config = {
            "max_result": 1000,
            "topic_score": 10.0,
            "threshold": 1.0,
            "base_url": "http://export.arxiv.org/api/query?",
        }
        self.config = Bunch(self.config)
        self.fetch = Fetch(self.config)
コード例 #5
0
    def subscribe_to_feed(self, feed_url):
        feed_id = db.get_feed_id(feed_url)

        if feed_id < 0:
            feed_fetch = Fetch(feed_url)
            feed_id = db.create_new_feed(feed_fetch.get_feed_details())
            db.add_new_item(feed_fetch.get_entries())

        user_feed_id = db.add_user_to_feed(self.uid, feed_id)

        return feed_id
コード例 #6
0
ファイル: execute.py プロジェクト: ddusek/everyday_sheets
def execute():
    """Run script every X minutes.
    """
    # Get google api credentials.
    my_creds = Creds()

    # Construct a service for interacting with sheets api.
    service = build('sheets', 'v4', credentials=my_creds.creds)

    print(CONSOLE_COLOR_START_DONE + f'{datetime.now()} running script' +
          CONSOLE_ENDC)
    main_start = timer()

    # Create new spreadsheet.
    start = timer()
    print('creating spreadsheet...')
    sheet = Spreadsheet(service, f'everyday_sheet{datetime.now()}')
    print(CONSOLE_COLOR +
          f'new sheet created in {round(timer() - start, 3)}s' + CONSOLE_ENDC)

    # Fetch data from all data sources defined in json file.
    fetcher = Fetch('reddit_inputs.json')
    start = timer()
    print('fetching data...')
    data = fetcher.fetch_all()
    print(CONSOLE_COLOR + f'data fetched in {round(timer() - start, 3)}s' +
          CONSOLE_ENDC)

    # Convert data into format needed for spreadsheet.
    sheet.convert_data(data)

    # Insert data into spreadsheet.
    start = timer()
    print('inserting data into spreadsheet...')
    sheet.insert_data()
    print(CONSOLE_COLOR +
          f'data inserted into a sheet in {round(timer() - start, 3)}s' +
          CONSOLE_ENDC)

    # Adjust columns size.
    sheet.set_col_size()

    print(CONSOLE_COLOR +
          f'elapsed time since start {round(timer() - main_start, 3)}s' +
          CONSOLE_ENDC)

    print('added ', datetime.now())
    print(CONSOLE_COLOR_START_DONE + f'{datetime.now()} script finished' +
          CONSOLE_ENDC)

    return f'https://docs.google.com/spreadsheets/d/{sheet.sheet_id}/edit#gid=0'
コード例 #7
0
    def set_wallpaper(self, img_path):
        if system == "Linux":
            if "~" in img_path:
                img_path = img_path.replace("~", HOME)
            if img_path == ".":
                img_path = start_path
            if os.path.isfile(img_path):
                self.linux_wallpaper(img_path)
            elif os.path.isfile(join(start_path, img_path)):
                self.linux_wallpaper(join(start_path, img_path))
            elif os.path.isdir(img_path):
                Fetch().custom_folder([img_path])
            elif os.path.isdir(join(start_path, img_path)):
                Fetch().custom_folder([join(start_path, img_path)])
            else:
                print(
                    f"{colors['red']}Error, file path not recognized{colors['normal']}"
                )

        else:
            print(
                f"{colors['red']}Sorry, your system is not supported yet.{colors['normal']}"
            )
コード例 #8
0
 def main_menu(self, message=message):
     refresh(message)
     choice = input(f"""{colors['green']}
             Welcome to Redpaper. This is a TUI used to
             control the underlying Redpaper program.
             Select an option:\n{colors['normal']}
         {colors['red']} 1 {colors['normal']}: {colors['blue']} Download wallpapers {colors['normal']} \n
         {colors['red']} 2 {colors['normal']}: {colors['blue']} Next wallpaper{colors['normal']}\n
         {colors['red']} 3 {colors['normal']}: {colors['blue']} Previous wallpaper{colors['normal']}\n
         {colors['red']} 4 {colors['normal']}: {colors['blue']} Settings{colors['normal']}\n
         {colors['red']} 5 {colors['normal']}: {colors['blue']} Help {colors['normal']}\n
         {colors['red']} x {colors['normal']}: {colors['blue']} exit {colors['normal']}\n
             >>>  """)
     if choice == "1":
         refresh(message)
         Fetch().wall_dl()
     elif choice == "2":
         message = f"{colors['green']} Changed wallpaper {colors['normal']}\n"
         refresh(message)
         img_path = WallSet().sequetial(0)
         WallSet().set_wallpaper(img_path)
         self.main_menu()
     elif choice == "3":
         message = f"{colors['green']} Changed wallpaper {colors['normal']}\n"
         refresh(message)
         img_path = WallSet().sequetial(1)
         WallSet().set_wallpaper(img_path)
         self.main_menu()
     elif choice == "4":
         message = ""
         Settings().main_settings()
     elif choice == "5":
         # TODO: create a help page
         message = "HELP\n"
         refresh(message)
         print(f"""
             {colors['green']}You can check the wiki for help:
             https://github.com/keystroke3/redpaper/wiki{colors['normal']}"""
               )
         self.main_menu()
     elif choice == "x" or choice == "X":
         clear()
     else:
         Home().main_menu()
コード例 #9
0
    def sequetial(self, go_back):
        try:
            with open(wall_data_file, encoding="utf-8") as data:
                saved_walls = json.load(data)
        except (FileNotFoundError, ValueError):
            Fetch().wall_dl()
        """chooses the wallpaper in the order in which they were downloaded"""
        with open(wall_data_file, "r") as data:
            saved_walls = json.load(data)

        with open("point.pickle", "rb+") as wall_point:
            # selection_point stores the value of the current wallpaper
            # it is necessary so that wallpapers don't repeat
            selection_point = pickle.load(wall_point)

            if selection_point > len(saved_walls):
                selection_point = 1
            elif selection_point == len(saved_walls) and go_back == 1:
                selection_point -= 1
            elif selection_point == len(saved_walls) and go_back == 0:
                selection_point = 1
            elif (selection_point < len(saved_walls) and selection_point != 1
                  and go_back == 1):
                selection_point -= 1
            elif selection_point < len(saved_walls) and go_back == 0:
                selection_point += 1
            elif (selection_point < len(saved_walls) and selection_point == 1
                  and go_back == 0):
                selection_point += 1
            elif (selection_point < len(saved_walls) and selection_point == 1
                  and go_back == 1):
                selection_point = len(saved_walls)
            elif (selection_point < len(saved_walls) and selection_point == 0
                  and go_back == 0):
                selection_point = 1
            elif (selection_point < len(saved_walls) and selection_point == 0
                  and go_back == 1):
                selection_point = len(saved_walls)
            img_name = str(saved_walls.get(str(selection_point)))
        # the new value of selection point is stored for the next run
        print(f"selection point is {selection_point}")
        with open("point.pickle", "wb") as point:
            pickle.dump(selection_point, point)
        return join(pictures, str(img_name))
コード例 #10
0
    def parseFeed(self):
        print "Start to fetch and parse Feed list"
        seed = self.seed
        f = Fetch(seed.prefixurl, seed.charset, self.seed.timeout)
        if f.isReady():
            feed = feedparser.parse(f.read())
            items = feed["entries"]
            if len(items) > 0:
                for item in items:
                    _item = Item({"url": item["link"], "type": self.seed_type})

                    if self.guid_rule is None:
                        self.guid_rule = "url"

                    guid = self.getItemGUID(item)
                    self.items[guid] = _item

        print "List has finished parsing. It has %s docs." % ansicolor.red(
            self.__len__())
コード例 #11
0
    def fetchListPages(self, listtype="html"):
        print "Start to fetch and parse List"
        urls = self.listRule.getListUrls()
        for url in urls:
            print "Fetching list page: ", url, "charset:", safestr(
                self.seed["charset"]), "timeout:", safestr(
                    self.seed["timeout"])
            f = Fetch(url,
                      charset=self.seed["charset"],
                      timeout=self.seed["timeout"])
            if f.isReady():
                doc = f.read()

                if listtype == "html":
                    self.parseListPage(f, doc, url)
                elif listtype == "json":
                    self.parseJsonPage(f, doc, url)

        print "List has finished parsing. It has %s docs." % ansicolor.red(
            self.__len__())
コード例 #12
0
 async def on_message(self, message):
     channel = message.channel
     channel.id = 842475416160698379
     if message.content.startswith('tag everyone'):
         self.members_to_ping = Fetch().return_data()
         print(self.members_to_ping)
         self.ids, self.members_to_tag = [], []
         for member in self.members:
             if member[2] in self.members_to_ping or member[
                     0] in self.members_to_ping:
                 self.ids.append(member[1])
                 self.members_to_tag.append(member[2] if member[2] in self.
                                            members_to_ping else member[0])
         print(self.members_to_tag)
         s = ''
         for members in self.ids:
             s += '<@{}>'.format(
                 members) if members == self.ids[-1] else '<@{}>,'.format(
                     members)
         await channel.send("{} Please report for your shift".format(s))
コード例 #13
0
    def __init__(self, item, seed):
        '''
	document base url
	'''
        self.url = item["url"]

        self.data = item

        self.seed = seed

        item["tags"] = ",".join(self.seed.tags)

        #文章采集规则
        self.articleRule = seed.getRule().getArticleRule()

        print "Document %s is fetcing" % ansicolor.green(self.url)
        firstContent = Fetch(self.url,
                             charset=seed["charset"],
                             timeout=seed["timeout"]).read()
        if firstContent:
            self.parseDocument(firstContent)
コード例 #14
0
ファイル: dollybs.py プロジェクト: gauryn/sandbox
 def load(self, loadUrl):
     o = urlparse(loadUrl)
     print o.scheme
     print o.port
     print o
     o = urlsplit(loadUrl)
     print o.geturl()
     reader = Fetch()
     htmlFile = reader.get(loadUrl)
     soup = BeautifulSoup(htmlFile)
     form = soup.find('form')
     print form
     print form['action']
     print form['method']
     #print form['onsubmit']
     if 'action' in form:
         print form['action']
     if 'method' in form:
         print form['method']
     if 'onsubmit' in form:
         print form['onsubmit']
コード例 #15
0
 async def on_ready(self):
     self.members = []
     self.required_members = []
     # for guild in self.guilds:
     #     for channel in guild.channels:
     #         print(channel.id,channel.name)
     for guild in self.guilds:
         data = guild.members
         for x in data:
             if x.name == 'REAPER':
                 print(x.name, x.id, x.nick)
             self.members.append([x.name, x.id, x.nick])
     print('Logged on as {0}!'.format(self.user))
     # await asyncio.sleep(60 * 50 * 2)
     self.members_to_ping = Fetch().return_data()
     print(self.members_to_ping)
     self.ids, self.members_to_tag = [], []
     for member in self.members:
         if member[2] in self.members_to_ping or member[
                 1] in self.members_to_ping:
             self.ids.append(member[1])
             self.members_to_tag.append(member[2])
コード例 #16
0
ファイル: main.py プロジェクト: AugustineLe0501/Fetch_Demo
import rospy
import time
import math
import cv2
import tf

from sensor_msgs.msg import Image
from ar_track_alvar_msgs.msg import AlvarMarkers
from geometry_msgs.msg import PoseStamped
from tf.transformations import euler_from_quaternion, quaternion_from_euler
from fetch import Fetch

if ((__name__ == "__main__") and (not rospy.is_shutdown())):
    rospy.init_node("fetch_builder",anonymous=True)

    Fetch_Robot = Fetch()
    rospy.loginfo("Initialization")

    Fetch_Robot.Head.look_at(0.7,0,0.5,"base_link")
    rospy.loginfo("Till head")

    Fetch_Robot.Gripper.Open()
    rospy.loginfo("Gripper Open")

    Fetch_Robot.Arm.Tuck()
    rospy.loginfo("Tuck Arm")

    rospy.sleep(rospy.Duration(2))
    #Take images
    RGB_image = Fetch_Robot.GetRGBImage()
    rospy.loginfo("Get RGB image")
コード例 #17
0
def main(args=None):
	parser = argparse.ArgumentParser(description='Simple training script.')
	parser.add_argument('--cls_id', help='class id', type=int)
	parser.add_argument('--version', help='model version', type=float)
	parser.add_argument('--gamma', help='gamma for the SoftL1Loss', type=float, default=9.0)
	parser.add_argument('--lr', help='lr for optimization', type=float, default=1e-4)
	parser.add_argument('--epoches', help='num of epoches for optimization', type=int, default=4)
	parser.add_argument('--resume_epoch', help='trained model for resume', type=int, default=0)
	parser.add_argument('--batch_size', help='batch size for optimization', type=int, default=10)
	parser.add_argument('--checkpoints', help='checkpoints path', type=str, default='voc_checkpoints')
	parser = parser.parse_args(args)

	cls_name = classes[parser.cls_id]
	parser.checkpoints = '_'.join([parser.checkpoints,cls_name])
	if not os.path.isdir(parser.checkpoints):
		os.mkdir(parser.checkpoints)
	print('will save checkpoints in '+parser.checkpoints)
	cls_dir = "../context_profile/voc_detection_{:s}_p10/"\
		.format(cls_name)
	batch_size = parser.batch_size
	print('[data prepare]....')
	dataloader_train = DataLoader(Fetch('train_benign', root_dir=cls_dir), batch_size=batch_size, num_workers=2, shuffle=True)

	print('[model prepare]....')
	use_gpu = torch.cuda.device_count()>0

	model = AutoEncoder(parser.gamma)
	if use_gpu:
		model = torch.nn.DataParallel(model).cuda()
	optimizer = torch.optim.Adam(model.parameters(), lr=parser.lr)
	scheduler = torch.optim.lr_scheduler.ReduceLROnPlateau(optimizer, patience=2, verbose=True)
	if parser.resume_epoch > 0 :
		checkpoint_name = os.path.join(parser.checkpoints, 'model_{:1.1f}_epoch{:d}.pt'.format(parser.version, parser.resume_epoch))
		if not os.path.isfile(checkpoint_name):
			raise ValueError('No checkpoint file {:s}'.format(checkpoint_name))
		model.load_state_dict(torch.load(checkpoint_name))
		print('model loaded from {:s}'.format(checkpoint_name))

	print('[model training]...')
	loss_hist = []
	epoch_loss = []
	num_iter = len(dataloader_train)
	for epoch_num in range(parser.resume_epoch, parser.epoches):
		model.train()
		for iter_num, sample in enumerate(dataloader_train):
			if True:#try:
				optimizer.zero_grad()
				if use_gpu:
					data = sample['data'].cuda().float()
				else:
					data = sample['data'].float()
					
				loss = model(data).mean()
				if bool(loss==0):
					continue 
				loss.backward()
				torch.nn.utils.clip_grad_norm_(model.parameters(), 0.1)
				optimizer.step()
				epoch_loss.append(float(loss))
				loss_hist.append(float(loss))
				if iter_num % 30 == 0:
					print('Epoch {:d}/{:d} | Iteration: {:d}/{:d} | loss: {:1.5f}'.format(
						epoch_num+1, parser.epoches, iter_num+1, num_iter, float(loss)))
				if iter_num % 3000 == 0:
					scheduler.step(np.mean(epoch_loss))
					epoch_loss = []
		if epoch_num < 1:
			continue
		checkpoint_name = os.path.join(parser.checkpoints, 'model_{:1.1f}_epoch{:d}.pt'.format(parser.version, epoch_num+1))
		torch.save(model.state_dict(), checkpoint_name)
		print('Model saved as {:s}'.format(checkpoint_name))

	np.save('loss_hist.npy', loss_hist)
コード例 #18
0
def coinpaprika():
    """Return coinpaprika json data.
    """
    fetcher = Fetch()
    data = fetcher.fetch_coinpaprika()
    return jsonify(data)
コード例 #19
0
def newsapi(topic):
    """Return newsapi json data according to topic.
    """
    fetcher = Fetch()
    data = fetcher.fetch_newsapi(topic)
    return jsonify(data)
コード例 #20
0
ファイル: test_AutoEncoders.py プロジェクト: sli057/SCEME
def get_preds(args=None):
    parser = argparse.ArgumentParser(description='Simple testing script.')
    parser.add_argument('--cls_id', help='class id', type=int)
    parser.add_argument('--version', help='model version', type=float)
    parser.add_argument('--resume_epoch',
                        help='trained model for resume',
                        type=int)
    parser.add_argument('--set_name',
                        help='imply attack goal',
                        type=str,
                        default='test_digi_ifgsm_hiding')
    parser.add_argument('--gamma',
                        help='gamma for the SoftL1Loss',
                        type=float,
                        default=9.0)
    parser.add_argument('--checkpoints',
                        help='checkpoints path',
                        type=str,
                        default='voc_checkpoints')
    parser.add_argument('--saves_dir',
                        help='the save path for tested reconstruction error',
                        type=str,
                        default='voc_reconstruction_error')
    parser.add_argument('--batch_size',
                        help='batch size for optimization',
                        type=int,
                        default=1)
    parser = parser.parse_args(args)

    batch_size = parser.batch_size
    if not os.path.isdir(parser.saves_dir):
        os.mkdir(parser.saves_dir)

    cls_name = classes[parser.cls_id]
    parser.checkpoints = '_'.join([parser.checkpoints, cls_name])

    checkpoint_name = os.path.join(
        parser.checkpoints,
        'model_{:1.1f}_epoch{:d}.pt'.format(parser.version,
                                            parser.resume_epoch))
    if not os.path.isfile(checkpoint_name):
        raise ValueError('No checkpoint file {:s}'.format(checkpoint_name))
    assert batch_size == 1

    print('[data prepare]....')
    cls_dir = "../context_profile/voc_detection_{:s}_p10/"\
     .format(cls_name)
    dataloader_test = DataLoader(Fetch(parser.set_name, root_dir=cls_dir),
                                 batch_size=batch_size,
                                 num_workers=1,
                                 shuffle=False)

    print('[model prepare]....')
    use_gpu = torch.cuda.device_count() > 0
    model = AutoEncoder(parser.gamma)
    if use_gpu:
        model = torch.nn.DataParallel(model).cuda()
    model.load_state_dict(torch.load(checkpoint_name))
    print('model loaded from {:s}'.format(checkpoint_name))

    print('[model testing]...')
    model.eval()
    preds = []
    with torch.no_grad():
        for sample in iter(dataloader_test):
            if use_gpu:
                data = sample['data'].cuda().float()

            else:
                data = sample['data'].float()
            loss = model(data)
            preds.append(float(loss))
    preds_name = '_model{:1.1f}_' + parser.set_name
    save_name = os.path.join(parser.saves_dir,
                             cls_name + preds_name.format(parser.version))
    np.save(save_name, preds)
    print('save preds in {:s}'.format(save_name))
コード例 #21
0
def reddit():
    """Return reddit json data.
    """
    fetcher = Fetch('reddit_inputs.json')
    data = fetcher.fetch_reddit()
    return jsonify(data)
コード例 #22
0
def get_data(url):
    req_obj = Fetch(url)
    ret_data = req_obj.get_data()
    if ret_data is not None:
        return json.loads(ret_data)
コード例 #23
0
    def parseDocument(self, doc):
        doc = pq(doc)

        wrapparent = self.articleRule.wrapparent
        pageparent = self.articleRule.pageparent
        content_re = ""
        #子页面url
        urls = []

        #文本数据内容
        content = ""

        article = doc.find(wrapparent)
        #pages
        if pageparent:
            urls = self.parsePage(article, pageparent)
#need title, tags
        extrarules = self.articleRule.extrarules

        #只有文章是有content
        #TODO: 这里目前缺失一些特性
        if len(extrarules):
            for key, rule, fetch_all, page_type in extrarules:
                field = Field(name=key, rule=rule)
                value = getElementData(doc, rule, self.data["images"],
                                       fetch_all)

                self.data[field.get('name')] = field

                if self.is_article_content(field):
                    content_re = field.get("rule")
                    content = value
                elif self.is_gallery_content(field):
                    content_re = field.get("rule")
                    content = []
                    if (isinstance(value, list)):
                        content += value
                else:
                    field.value = value

#采集分页内容
        if len(urls) > 0 and content_re:
            for next_url in urls:
                next_page = Fetch(next_url,
                                  charset=self.seed["charset"],
                                  timeout=self.seed["timeout"]).read()
                if next_page is not None:
                    next_page = self._getContent(next_page, wrapparent,
                                                 content_re)
                    if next_page:
                        if isinstance(content, list):
                            content.append(next_page)
                        else:
                            content += next_page

        if content and content_re:
            if isinstance(content, list):
                self.data['content'].value = content
                self.data['images'] += content
            else:
                content = Readability(content, self.url,
                                      self.articleRule.filters)
                images = content.getImages()

                self.data['content'].value = content.getContent()
                self.data['images'] += images
コード例 #24
0
ファイル: kernel.py プロジェクト: veghen/Y86
    lst.write('E', cur.regE)
    lst.write('M', cur.regM)
    lst.write('W', cur, regW)


mem = Memory()
InsCode = {}
Init(InsCode, mem)
reg = Register()
pipereg = PipeRegister()
tmp_pipereg = PipeRegister()
CC = ConditionCode()
Stat = Status()

PC = 0
while Stat.stat == 'AOK':
    print 'Current Time:', PC
    tmp_pipereg = PipeRegister()
    Fetch(tmp_pipereg, InsCode[hex(PC)], PC)
    Decode(pipereg, tmp_pipereg, reg)
    Execute(pipereg, tmp_pipereg)
    Memory(pipereg, tmp_pipereg)
    WriteBack(pipereg, tmp_pipereg)
    PC = pipereg.regF['predPC']
    Update(cur=tmp_pipereg, lst=pipireg)
    print 'RegF:', reg.regF
    print 'RegD:', reg.regD
    print 'RegE:', reg.regE
    print 'RegM:', reg.regM
    print 'RegW:', reg.regW
コード例 #25
0
ファイル: server.py プロジェクト: Jexulie/Python-RSS-Server
from fetch import Fetch
from parsing import Parser, idGenerator
from flask import Flask, Response, json, request
from database import addDB, updateDB, removeDBbyCID, removeDBbyURL, getDBbyCID, getDBbyURL, getAllDB
import json
import re

# Todo User System.
# Todo 2 Database Implentation.
# Todo 3 on Wrong RSS remove added URL and send back response.
# Todo 4 If URL already exists reject
# Todo 5 Proper Err msging and on resp

newFetcher = Fetch()

app = Flask(__name__)


def checkURL(url):
    match = re.match(r'http[s]?://.*', url)
    if match:
        return True
    else:
        return False


@app.route("/", methods=['GET'])
def index():
    r = {'success': True, 'msg': "Server Index Point ..."}
    resp = Response(json.dumps(r))
    resp.headers['Content-Type'] = 'application/json'
コード例 #26
0
import datetime as dt
from fetch import Fetch
from display import Display
import pprint

if __name__ == '__main__':

    # Performs queries:
    fetchy = Fetch()
    #fetchy.fetchTweets("prayforsyria", 1000, begindate=dt.date(2011,1,21), enddate=dt.date.today())

    # Defines words that we want to use in the visualization. All other words will be ignored:
    chosenWords = [
        "Syria", "people", "you", "all", "no", "we", "world", "will", "about",
        "innocent", "Allah", "paz", "just", "pray", "heart", "Syrian", "mundo",
        "children", "who", "one", "don't", "please", "help", "Pray", "more",
        "like", "apoyo", "oración", "#Aleppo", "going", "war", "God", "uno",
        "need", "prayers", "ayuno", "happening", "killed", "being", "peace",
        "sad", "NO", "know", "now", "because", "stop", "many", "everyone",
        "live", "unimos", "really", "lives", "orar", "much", "love",
        "suffering", "protect", "guerra", "morning", "even", "todos", "hope",
        "country", "#LLAP", "forget", "never", "over", "every", "still",
        "brothers", "think", "llamado", "ayunar", "human", "time", "feel",
        "praying", "keep", "bomb", "bombing", "kids", "Virgen", "today",
        "Assad", "better", "diligencia", "killing", "breaks", "trabajo",
        "video", "life", "firmemente", "rezar", "where", "lost"
        "kill", "pidiendo", "humanity", "deserve", "always", "gente", "little",
        "take", "safe", "end", "say", "sisters", "doesn't", "any", "dying",
        "dead", "after", "things", "good", "Trump", "some", "child", "needs",
        "heartbreaking", "other", "Dios", "care", "Syrians", "US", "personas",
        "news", "civilians", "não", "against", "inocentes", "unidos", "media",
コード例 #27
0
from network import Network
from fetch import Fetch

co = Network()
co.send("Test\n")
num = co.receive()
dataFetching = Fetch(num)

data = co.receive()
print(dataFetching.fetchMessage(data))
コード例 #28
0
# App Modules
from fetch import Fetch
from processor import Processor

## Load ENV
load_dotenv()

isDev = os.getenv('FLASK_ENV') == 'development'
accepted = '*' if isDev else 'https://wilsonj806.github.io/*'

app = Flask(__name__)
app.config['CORS_HEADERS'] = 'Content-Type'

CORS(app, origins=accepted)

fetcher = Fetch()


@app.route('/')
def initial_ping():
    return """NYC Tree Data Fetcher; Version: 0.0.0"""


@app.route('/data')
def fetch_all():
    json = fetcher.check_cache()
    return {'_data_length': len(json), 'data': json}


@app.route('/data/count')
def count_per_boro():
コード例 #29
0
#import sys
#sys.path.append('/home/ubuntu/projects/financial')
from fetch import Fetch

fetch = Fetch()
params = [('sc', 431)]

stocklist = fetch.fetch_stocks(params)
print stocklist
コード例 #30
0
def main():
    parser = argparse.ArgumentParser(
        description="""This is a simple program that allows you to change
    you desktop wallpaper. It fetches the best wallpapers from Reddit
    and sets one as the wallpaper.""")
    parser.add_argument("-d",
                        "--download",
                        action="store_true",
                        help="Downloads new wallpapers")
    parser.add_argument(
        "-c",
        "--change",
        action="store_true",
        help="sets a wallpaper without downloading new ones",
    )
    parser.add_argument(
        "-a",
        "--all",
        action="store_true",
        help="Download new wallpapers and set one of them",
    )
    parser.add_argument("-l",
                        "--limit",
                        help="Number of wallpapers to look for. Default = 1")
    parser.add_argument(
        "-p",
        "--path",
        metavar="PATH",
        help="Sets the download location for new wallpapers\n"
        "The img_path has to be in quotes",
    )
    parser.add_argument("-i",
                        "--image",
                        help="Sets a user specified image as wallpaper.\n")
    parser.add_argument("-r",
                        "--sub",
                        help="Sets a user specified subreddit(s) as source.\n")
    parser.add_argument(
        "-f",
        "--folder",
        help="Uses images stored in the specified folders\n"
        "Path has to be in quotes",
    )
    parser.add_argument("-s",
                        "--settings",
                        action="store_true",
                        help="change settings permanently")
    parser.add_argument("-b",
                        "--back",
                        action="store_true",
                        help="Sets the previous image as wallpaper")

    # args = parser.parse_args()
    args, unknown = parser.parse_known_args()
    if not len(sys.argv) > 1:
        Home().main_menu()
    if args.settings:
        if args.path:
            Settings().change_dl_path(args.path, True)
            return
        elif args.limit:
            Settings().max_dl_choice(args.limit, True)
            return
        elif args.sub:
            Settings().change_subs(args.sub, True)
            return
        else:
            print("No option selected or selection not understood")
            return
    if args.download:
        sub_list = ""
        if args.sub:
            sub_list = args.sub
        if args.limit:
            Fetch().d_limit = int(args.limit)
        Fetch(sub_list).wall_dl()
    elif args.change:
        if args.back:
            img_path = WallSet().sequetial(1)
            WallSet().set_wallpaper(img_path)
        else:
            img_path = WallSet().sequetial(0)
            WallSet().set_wallpaper(img_path)
    elif args.image:
        WallSet().set_wallpaper(args.image)
    elif args.folder:
        Fetch().custom_folder(args.folder[0])
        args.change
    elif args.all:
        img_path = WallSet().sequetial(0)
        if args.limit:
            Fetch().d_limit = int(args.limit)
            Fetch().wall_dl()
        WallSet().set_wallpaper(img_path)
    elif args.back:
        img_path = WallSet().sequetial(1)
        WallSet().set_wallpaper(img_path)
    elif unknown:
        if len(unknown) == 1:
            WallSet().set_wallpaper(unknown[0])
        else:
            Fetch().custom_folder(unknown)
            args.change