# logging import logging MIN_LEVEL = logging.DEBUG dtlutil.setup_log(MIN_LEVEL) ############################################################## #%% read in rdf graph import rdflib from rdflib.graph import Graph, Store, URIRef, Literal, BNode from rdflib.namespace import RDF, FOAF, DC MO, TL, EVENT, OLO, DTL, initNs = dtlutil.init_namespaces() DBP, REL, LJ = dtlutil.LOD_namespaces() g = dtlutil.create_graph() dtlutil.read_in_rdf(g, RDFfile_merged) gDTL1000 = dtlutil.create_graph() #%% gDTL1000 = copy.deepcopy(g) #%% read in fingerprints dtl1000_fprints = [] logging.info("\nReading fingerprints from %s", FPRINTfile) with open(FPRINTfile, 'r') as csvfile: csvreader = csv.reader(csvfile, delimiter=',') count = 0 for row in csvreader:
logging.info("\nAppending RDF to: %s ...", rdffilename) self.append_rdf_to_file(rdffilename, rdffilename_tmp) logging.info("done") # clear temporary graph self.g_temp.remove((None, None, None)) logging.debug("\ng_temp cleared: \n%s", str(self.g_temp.serialize(format="turtle"))) else: # track already in graph logging.info("skipping") # clean up at the end # correct olo:indexx to olo:index logging.info("\nCorrecting olo:indexx to olo:index ...") self.indexx2index() logging.info("Writing RDF to: %s ...", rdffilename) self.write_rdf_to_file(rdffilename, rdffilename_tmp) logging.info("done") #logging.debug(str(g.serialize(format='turtle'))) # create graphs g = dtlutil.create_graph('JEGraphID') g_temp = dtlutil.create_graph('JE_temp_GraphID') dtlutil.read_in_rdf(g, rdffilename) # create RDF RDFcreator = JEtoRDF(g, g_temp) RDFcreator.create_rdf()
dtlutil.setup_log(MIN_LEVEL) ############################################################## # create rdf graph import rdflib from rdflib.graph import Graph, Store, URIRef, Literal, BNode from rdflib.namespace import Namespace, RDFS from rdflib import plugin from rdflib.plugins import sparql from rdflib import Namespace from rdflib.namespace import RDF, FOAF, RDFS, DC, XSD MO, TL, EVENT, OLO, DTL, initNs = dtlutil.init_namespaces() g = dtlutil.create_graph('Ill1000GraphID') ######################################################### # read in sqlite tables logging.info("Reading tables from %s", DBfile) import sqlite3 import pandas as pd pd.set_option("display.max_rows", 5) # Create your connection. cnx = sqlite3.connect(DBfile) sessions_table = pd.read_sql_query("SELECT * FROM sessions", cnx) #sessions_table.head()
dtlutil.setup_log(MIN_LEVEL) ############################################################## # read in rdf graph import rdflib from rdflib.graph import Graph, Store, URIRef, Literal, BNode from rdflib.namespace import Namespace, RDFS from rdflib import plugin from rdflib.plugins import sparql from rdflib import Namespace from rdflib.namespace import RDF, FOAF, RDFS, DC, XSD MO, TL, EVENT, OLO, DTL, initNs = dtlutil.init_namespaces() g = dtlutil.create_graph() dtlutil.read_in_rdf(g, RDFfile) logging.debug("\ngraph has %i triples", len(g)) ############################################################## logging.info("\nReading styles from %s", STYLESfile) with open(STYLESfile, 'r') as csvfile: csvreader = csv.reader(csvfile, delimiter=',') count = 0 for row in csvreader: if len(row) > 0: fprint = row[0] style = row[1] signalURI = g.value(subject=None, predicate=DTL.fingerprint_short, \
############################################################## # read in rdf graph import rdflib from rdflib.graph import Graph, Store, URIRef, Literal, BNode from rdflib.namespace import Namespace, RDFS from rdflib import plugin from rdflib.plugins import sparql from rdflib import Namespace from rdflib.namespace import RDF, FOAF, RDFS, DC, XSD, OWL MO, TL, EVENT, OLO, DTL, initNs = dtlutil.init_namespaces() DBP, REL, LJ = dtlutil.LOD_namespaces() g = dtlutil.create_graph() dtlutil.read_in_rdf(g, RDFfile) ljg = dtlutil.create_graph("LJgraph") dtlutil.read_in_rdf(ljg, NTfile, myformat = "nt") ############################################################## mcount = 0 lodcount = 0 for musician in g.subjects(RDF.type, MO.MusicArtist): name = g.value(musician, FOAF.name) mcount +=1 for LODmusician in ljg.subjects(predicate=FOAF.name, object=Literal(name, lang = "en")): logging.info("found match for musician %s in DB: %s", name, LODmusician) g.add( (musician, OWL.sameAs, LODmusician) )
############################################################## # read in rdf graph import rdflib from rdflib.graph import Graph, Store, URIRef, Literal, BNode from rdflib.namespace import Namespace, RDFS from rdflib import plugin from rdflib.plugins import sparql from rdflib import Namespace from rdflib.namespace import RDF, FOAF, RDFS, DC, XSD MO, TL, EVENT, OLO, DTL, initNs = dtlutil.init_namespaces() DBP, REL, LJ = dtlutil.LOD_namespaces() gJE = dtlutil.create_graph() dtlutil.read_in_rdf(gJE, RDFfile_JE) gILL = dtlutil.create_graph() dtlutil.read_in_rdf(gILL, RDFfile_ILL) ############################################################## def merge(uri_JE, uri_ILL): for s, p, o in gILL.triples((uri_ILL, None, None)): #logging.debug("\nremoving: (%s, %s, %s)", str(s), str(p), str(o)) g_merge.remove((s, p, o)) if not (uri_JE, p, o) in g_merge: #logging.debug("adding: (%s, %s, %s)", str(uri_JE), str(p), str(o))
dtlutil.setup_log(MIN_LEVEL) ############################################################## # create rdf graph import rdflib from rdflib.graph import Graph, Store, URIRef, Literal, BNode from rdflib.namespace import Namespace, RDFS from rdflib import plugin from rdflib.plugins import sparql from rdflib import Namespace from rdflib.namespace import RDF, FOAF, RDFS, DC, XSD MO, TL, EVENT, OLO, DTL, initNs = dtlutil.init_namespaces() g = dtlutil.create_graph('Ill1000GraphID') g_temp = dtlutil.create_graph('Ill1000GraphID_temp') dtlutil.read_in_rdf(g, RDFfile) ######################################################### # read in sqlite tables logging.info("Reading tables from %s", DBfile) import sqlite3 import pandas as pd pd.set_option("display.max_rows", 5) # Create your connection. cnx = sqlite3.connect(DBfile) sessions_table = pd.read_sql_query("SELECT * FROM sessions", cnx)