from flathunter.crawl_immobilienscout import CrawlImmobilienscout from flathunter.crawl_wggesucht import CrawlWgGesucht from flathunter.crawl_immowelt import CrawlImmowelt from flathunter.idmaintainer import IdMaintainer from flathunter.googlecloud_idmaintainer import GoogleCloudIdMaintainer from flathunter.hunter import Hunter from flathunter.config import Config from flathunter.web import app searchers = [ CrawlImmobilienscout(), CrawlWgGesucht(), CrawlEbayKleinanzeigen(), CrawlImmowelt() ] if __name__ == '__main__': # Use the SQLite DB file if we are running locally id_watch = IdMaintainer('%s/processed_ids.db' % os.path.dirname(os.path.abspath(__file__))) else: # Use Google Cloud DB if we run on the cloud id_watch = GoogleCloudIdMaintainer() hunter = Hunter(Config(), searchers, id_watch) app.config["HUNTER"] = hunter if __name__ == '__main__': app.run(host='127.0.0.1', port=8080, debug=True)
from flathunter.idmaintainer import IdMaintainer from flathunter.googlecloud_idmaintainer import GoogleCloudIdMaintainer from flathunter.web_hunter import WebHunter from flathunter.config import Config from flathunter.web import app config = Config() if __name__ == '__main__': # Use the SQLite DB file if we are running locally id_watch = IdMaintainer('%s/processed_ids.db' % config.database_location()) else: # Use Google Cloud DB if we run on the cloud id_watch = GoogleCloudIdMaintainer() hunter = WebHunter(config, id_watch) app.config["HUNTER"] = hunter if 'website' in config: app.secret_key = config['website']['session_key'] app.config["DOMAIN"] = config['website']['domain'] app.config["BOT_NAME"] = config['website']['bot_name'] else: app.secret_key = b'Not a secret' app.config["BOT_TOKEN"] = config['telegram']['bot_token'] if __name__ == '__main__': listen = config['website']['listen'] app.run(host=listen['host'], port=listen['port'], debug=True)