import hashlib import os.path import sys import datetime from logLoader import loadLogger from database.dbExecutor import dbExecutor SOURCE_ID = "CPI" # source identifier MAX_HTTP_RETRIES = 10 # set max number of http request retries if a page load fails BASE_URL = "http://www.cpi.si" URLS_TO_CHECK = [ BASE_URL + "/mednarodno-sodelovanje/" + i for i in ["projekti-eu-komisije.aspx", "erasmus-.aspx"] ] logger = loadLogger(SOURCE_ID) # makes a sha1 hash out of title and date strings # returns string hash def makeHash(articleTitle): hash_object = hashlib.sha1((articleTitle).encode("utf-8")) return hash_object.hexdigest() # navigates to the given link and extracts the article description def getArticleDescr(session, link): resp = session.get(link) soup = bs.BeautifulSoup(resp.text, "html.parser") # print (soup.encode("utf-8")) description = soup.find("div", class_="freetext")
# -*- coding: utf-8 -*- import requests import bs4 as bs import sys import datetime import logging from dbExecutor import dbExecutor from logLoader import loadLogger from config import AUTH, LOCAL_IP, SOURCE logger = loadLogger(SOURCE) logger.setLevel(logging.INFO) def getAuth(username, password): import base64 return base64.b64encode(username + ":" + password) def getConnectedDevicesList(): urlDevices = "http://" + LOCAL_IP + "/userRpm/AssignedIpAddrListRpm.htm?Refresh=Refresh" todayDateStr = datetime.datetime.now().strftime( "%Y-%m-%dT%H:%M:%S") # today date in the uniform format if len(AUTH) < 1: logger.error( "Unset AUTH key. Set it in the config file or use the getAuth() function in place of \"AUTH\" fields in the main program." ) sys.exit(1)