def main(): # pragma: no cover parser = argparse.ArgumentParser() parser.add_argument("--platform", "-p", help="scrape platform. (eg: doctolib,keldoc or all)") parser.add_argument("--url", "-u", action="append", help="scrape one url, can be repeated") parser.add_argument("--merge", "-m", help="merge platform results", action="store_true") args = parser.parse_args() if args.merge: merge_platforms() return if args.url: scrape_debug(args.url) return platforms = [] if args.platform and args.platform != 'all': platforms = args.platform.split(',') scrape(platforms=platforms)
def main(): # pragma: no cover parser = argparse.ArgumentParser() parser.add_argument("--platform", "-p", help="scrape platform. (eg: doctolib,keldoc or all)") parser.add_argument("--url", "-u", action="append", help="scrape one url, can be repeated") parser.add_argument("--url-file", type=argparse.FileType("r"), help="scrape urls listed in file (one per line)") parser.add_argument("--merge", "-m", help="merge platform results", action="store_true") args = parser.parse_args() if args.merge: merge_platforms() return if args.url_file: args.url = [line.rstrip() for line in args.url_file] if args.url: scrape_debug(args.url) return platforms = [] if args.platform and args.platform != "all": platforms = args.platform.split(",") scrape(platforms=platforms)