def test(): base_dir = os.path.join(os.getcwd(), "data") print "base_dir: " + base_dir openurl = "url_ver=Z39.88-2004&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.atitle=Isolation of a common receptor for coxsackie B&rft.jtitle=Science&rft.aulast=Bergelson&rft.auinit=J&rft.date=1997&rft.volume=275&rft.spage=1320&rft.epage=1323" metajson_list = crossref.query_openurl_and_retrieve_metadata(openurl, True) if metajson_list: output_path = os.path.join(base_dir, "result", "result_crossref_metajon.json") export_service.export_metajson(metajson_list, output_path) print json.dumps(metajson_list, indent=4, ensure_ascii=False, encoding="utf-8", sort_keys=True) else: assert False
def test(): base_dir = os.path.join(os.getcwd(), "data") print "base_dir: " + base_dir input_dir = os.path.join(base_dir, "unixref") metajson_list = [] for file_name in os.listdir(input_dir): if file_name.endswith(".xml"): print file_name metajson_list.extend(crossref_unixref_to_metajson.convert_crossref_unixref_file_to_metajson_document_list(os.path.join(input_dir, file_name), file_name, False)) if metajson_list: output_path = os.path.join(base_dir, "result", "result_unixref_metajon.json") export_service.export_metajson(metajson_list, output_path) print json.dumps(metajson_list, indent=4, ensure_ascii=False, encoding="utf-8", sort_keys=True) else: assert False
def import_references(): print("import_references") #base_dir = os.getcwd() base_dir = os.path.dirname(__file__) print "base_dir: " + base_dir base_dir = os.path.join(base_dir, os.pardir, "data") print "base_dir: " + base_dir #filenames = ["endnote-aime.xml", "endnote-ref.xml", "endnote-bib.xml"] filenames = ["endnote-aime.xml"] errors_file = os.path.join(base_dir, "result", "result_validation_errors.txt") result_mla = os.path.join(base_dir, "result", "result_mla.html") result_metajson = os.path.join(base_dir, "result", "result_metajson.json") # import date_start = datetime.datetime.now() input_files = [] for filename in filenames: input_file = os.path.join(base_dir, "endnotexml", filename) input_files.append(input_file) import_service.import_metadata_files(input_files, "endnotexml", errors_file, True) date_import = datetime.datetime.now() chrono.chrono_trace("import", date_start, date_import, None) # fetch metajson_list = repository_service.get_references() date_fetch = datetime.datetime.now() chrono.chrono_trace("fetch", date_import, date_fetch, len(metajson_list)) # export citations export_service.export_html_webpage(metajson_list, result_mla) date_citations = datetime.datetime.now() chrono.chrono_trace("citations", date_fetch, date_citations, len(metajson_list)) # export json export_service.export_metajson(metajson_list, result_metajson) date_json = datetime.datetime.now() chrono.chrono_trace("json", date_citations, date_json, len(metajson_list))