def test():
    base_dir = os.path.join(os.getcwd(), "data")
    print "base_dir: " + base_dir

    openurl = "url_ver=Z39.88-2004&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.atitle=Isolation of a common receptor for coxsackie B&rft.jtitle=Science&rft.aulast=Bergelson&rft.auinit=J&rft.date=1997&rft.volume=275&rft.spage=1320&rft.epage=1323"
    metajson_list = crossref.query_openurl_and_retrieve_metadata(openurl, True)
    if metajson_list:
        output_path = os.path.join(base_dir, "result", "result_crossref_metajon.json")
        export_service.export_metajson(metajson_list, output_path)
        print json.dumps(metajson_list, indent=4, ensure_ascii=False, encoding="utf-8", sort_keys=True)
    else:
        assert False
def test():
    base_dir = os.path.join(os.getcwd(), "data")
    print "base_dir: " + base_dir

    input_dir = os.path.join(base_dir, "unixref")
    metajson_list = []
    for file_name in os.listdir(input_dir):
        if file_name.endswith(".xml"):
            print file_name
            metajson_list.extend(crossref_unixref_to_metajson.convert_crossref_unixref_file_to_metajson_document_list(os.path.join(input_dir, file_name), file_name, False))

    if metajson_list:
        output_path = os.path.join(base_dir, "result", "result_unixref_metajon.json")
        export_service.export_metajson(metajson_list, output_path)
        print json.dumps(metajson_list, indent=4, ensure_ascii=False, encoding="utf-8", sort_keys=True)
    else:
        assert False
Beispiel #3
0
def import_references():
    print("import_references")
    #base_dir = os.getcwd()
    base_dir = os.path.dirname(__file__)
    print "base_dir: " + base_dir
    base_dir = os.path.join(base_dir, os.pardir, "data")
    print "base_dir: " + base_dir

    #filenames = ["endnote-aime.xml", "endnote-ref.xml", "endnote-bib.xml"]
    filenames = ["endnote-aime.xml"]
    errors_file = os.path.join(base_dir, "result", "result_validation_errors.txt")
    result_mla = os.path.join(base_dir, "result", "result_mla.html")
    result_metajson = os.path.join(base_dir, "result", "result_metajson.json")

    # import
    date_start = datetime.datetime.now()
    input_files = []
    for filename in filenames:
        input_file = os.path.join(base_dir, "endnotexml", filename)
        input_files.append(input_file)

    import_service.import_metadata_files(input_files, "endnotexml", errors_file, True)

    date_import = datetime.datetime.now()
    chrono.chrono_trace("import", date_start, date_import, None)

    # fetch
    metajson_list = repository_service.get_references()

    date_fetch = datetime.datetime.now()
    chrono.chrono_trace("fetch", date_import, date_fetch, len(metajson_list))

    # export citations
    export_service.export_html_webpage(metajson_list, result_mla)

    date_citations = datetime.datetime.now()
    chrono.chrono_trace("citations", date_fetch, date_citations, len(metajson_list))

    # export json
    export_service.export_metajson(metajson_list, result_metajson)

    date_json = datetime.datetime.now()
    chrono.chrono_trace("json", date_citations, date_json, len(metajson_list))