def main(*args): site = pywikibot.Site() fetcher = CategoryFetcher(site) checker = TemplateChecker() checker.load_config("config/templates.json") checker_bot = CheckerBot(checker, site) all_categories = fetcher.get_categories() callbacks = ArticleIteratorCallbacks( category_callback=checker_bot.cb_store_category_result, article_callback=checker_bot.cb_check_article, logging_callback=pywikibot.log, ) article_iterator = ArticleIterator(callbacks, categories=all_categories) parser = ArticleIteratorArgumentParser(article_iterator, fetcher) for argument in pywikibot.handle_args(list(args)): if parser.check_argument(argument): continue elif argument.find("-outputpage:") == 0: checker_bot.outputpage = argument[12:] elif argument.find("-exclude-articles:") == 0: page = pywikibot.Page(site, argument[18:]) article_iterator.excluded_articles = load_excluded_articles_from_wiki(page) article_iterator.iterate_categories() if article_iterator.categories != all_categories: # Don't update summary page if only single categories were crawled return summary = checker_bot.generate_summary_page() if checker_bot.outputpage: checker_bot.save_wikipage(summary, checker_bot.outputpage + u"/Zusammenfassung") else: pywikibot.output(u"Zusammenfassung") pywikibot.output(u"===============") pywikibot.output(summary) pywikibot.output(checker_bot.generate_config_table())
def main(*args): wikipedia_site = pywikibot.Site() # Use the site configured in params/user-config commons_site = pywikibot.Site("commons", "commons") checker = TemplateChecker() checker.load_config("config/templates.json") commons_bot = CommonsBot(wikipedia_site, checker) callbacks = ArticleIteratorCallbacks( logging_callback=pywikibot.log, article_callback=commons_bot.cb_check_article ) article_iterator = ArticleIterator(callbacks) article_iterator.log_every_n = 1 parser = ArticleIteratorArgumentParser(article_iterator, None) run_cmd = commons_bot.run_continuously category_name = u"Images from Wiki Loves Monuments 2015 in Germany" start_time = first_day_of_month() for argument in pywikibot.handle_args(args): if argument.find("-category:") == 0: category_name = argument[10:] continue elif parser.check_argument(argument): continue elif argument.find("-start-at:") == 0: start_time_iso = argument[10:] + "T0:00:00Z" start_time = pywikibot.Timestamp.fromISOformat(start_time_iso) elif argument.find("-sleep-seconds:") == 0 and int(argument[15:]) > 0: commons_bot.sleep_seconds = int(argument[15:]) elif argument == "-once": run_cmd = commons_bot.run_once elif argument == "-local-media": commons_site = wikipedia_site category = pywikibot.Category(commons_site, category_name) run_cmd(article_iterator, start_time, category)
def main(): parser = argparse.ArgumentParser(description='Generate JSON info about monument data in wiki text.') parser.add_argument('--monument_id', '-i', help='Unique ID of the monument. Validity will be checked.', default='', metavar='ID') parser.add_argument('infile', nargs='?', type=argparse.FileType('r'), default=sys.stdin) args = parser.parse_args() checker = TemplateChecker() checker.load_config("config/templates.json") mapper = CommonscatMapper() info = get_template_info(checker, mapper, args.infile.read(), args.monument_id) utf8_writer = codecs.getwriter('utf8') json.dump(info, utf8_writer(sys.stdout))
def setup_instances(): g.site_commons = mwclient.Site("commons.wikimedia.org") g.site_wikipedia = mwclient.Site("de.wikipedia.org") g.campaign_validator = CampaignValidator(g.site_commons) checker = TemplateChecker() checker.load_config("config/templates.json") mapper = CommonscatMapper() mapper.load_mapping("config/commonscat_mapping.json") g.page_information_collector = PageInformationCollector(checker, mapper) if app.config["REDIS_CACHE_PREFIX"]: g.campaign_cache = RedisCache(host=app.config["REDIS_HOST"], key_prefix=app.config["REDIS_CACHE_PREFIX"]) else: g.campaign_cache = SimpleCache()