def load_links_index(out_dir=OUTPUT_DIR, import_path=None): """parse and load existing index with any new links from import_path merged in""" existing_links = [] if out_dir: existing_links = parse_json_links_index(out_dir) check_links_structure(existing_links) new_links = [] if import_path: # parse and validate the import file log_parsing_started(import_path) raw_links, parser_name = parse_links(import_path) new_links = validate_links(raw_links) check_links_structure(new_links) # merge existing links in out_dir and new links all_links = validate_links(existing_links + new_links) check_links_structure(all_links) num_new_links = len(all_links) - len(existing_links) if import_path and parser_name: log_parsing_finished(num_new_links, parser_name) return all_links, new_links
def load_links_index(out_dir=OUTPUT_DIR, import_path=None): """parse and load existing index with any new links form import_path merged in""" existing_links = [] if out_dir: existing_links = parse_json_links_index(out_dir) check_links_structure(existing_link) new_links = [] if import_path: # parse and validate import file # this serves as a logging function log_parsing_started(import_path) raw_links, parser_name = parse_links(import_path) new_links = validate_links(raw_links) check_links_structure(new_links) # merge existing links in out_dir and new links all_links = validate_links(existing_links + new_links) return all_links, new_links