Exemple #1
0
import logging

if __name__ == '__main__':
    # Load the configuration file
    config = Configuration('config-cedar.ini')

    # Configure the logger
    root_logger = logging.getLogger('')
    root_logger.setLevel(logging.DEBUG if config.verbose() else logging.INFO)
    logFormat = '%(asctime)s %(name)-18s %(levelname)-8s %(message)s'
    ch = logging.StreamHandler()
    ch.setFormatter(logging.Formatter(logFormat))
    root_logger.addHandler(ch)
    fh = logging.FileHandler('rules.log')
    fh.setFormatter(logging.Formatter(logFormat))
    root_logger.addHandler(fh)

    #dataset = config.getURI('cedar',"BRT_1889_02_T1-S0")
    #dataset = config.getURI('cedar',"VT_1869_01_H1-S0")
    #dataset = config.getURI('cedar','VT_1879_01_H1-S0')
    #dataset = config.getURI('cedar','VT_1859_01_H1-S6')
    dataset = 'VT_1899_07_H1-S0'

    # Test
    rulesMaker = RuleMaker(config.get_SPARQL(), dataset, "/tmp/test.ttl")
    rulesMaker.loadMappings(
        config.get_path('mappings'))  #, ['Sex','MaritalStatus']
    rulesMaker.loadHeaders(config.get_graph_name('raw-data'))
    rulesMaker.process()  # ['Sex','MaritalStatus']
Exemple #2
0
def generate_harmonization_rules_thread(parameters):
    '''
    Worker thread for generate_harmonization_rules
    '''
    dataset = parameters['dataset']
    output = parameters['output']
    log.info("[{}] Calling RulesMaker".format(dataset))
    try:
        rulesMaker = RuleMaker(parameters['endpoint'], dataset, output)
        rulesMaker.set_target_namespace(parameters['target'])
        rulesMaker.set_compress(parameters['compress'])
        rulesMaker.loadMappings(parameters['mappings'])
        rulesMaker.loadHeaders(parameters['raw-data'])
        rulesMaker.process()
    except Exception as e:
        log.error("[{}] Error in RulesMaker: {}".format(dataset.n3(), e))
Exemple #3
0
def generate_harmonization_rules_thread(parameters):
    '''
    Worker thread for generate_harmonization_rules
    '''
    dataset = parameters['dataset']
    output = parameters['output']
    log.info("[{}] Calling RulesMaker".format(dataset))
    try:
        rulesMaker = RuleMaker(parameters['endpoint'], dataset, output)
        rulesMaker.set_target_namespace(parameters['target'])
        rulesMaker.set_compress(parameters['compress'])
        rulesMaker.loadMappings(parameters['mappings']) 
        rulesMaker.loadHeaders(parameters['raw-data'])
        rulesMaker.process()
    except Exception as e:
        log.error("[{}] Error in RulesMaker: {}".format(dataset.n3(), e))
from util.configuration import Configuration

import logging

if __name__ == '__main__':
    # Load the configuration file
    config = Configuration('config-cedar.ini')

    # Configure the logger
    root_logger = logging.getLogger('')
    root_logger.setLevel(logging.DEBUG if config.verbose() else logging.INFO)
    logFormat = '%(asctime)s %(name)-18s %(levelname)-8s %(message)s'    
    ch = logging.StreamHandler()
    ch.setFormatter(logging.Formatter(logFormat))
    root_logger.addHandler(ch)
    fh = logging.FileHandler('rules.log')
    fh.setFormatter(logging.Formatter(logFormat))
    root_logger.addHandler(fh)
    
    #dataset = config.getURI('cedar',"BRT_1889_02_T1-S0")
    #dataset = config.getURI('cedar',"VT_1869_01_H1-S0")
    #dataset = config.getURI('cedar','VT_1879_01_H1-S0')
    #dataset = config.getURI('cedar','VT_1859_01_H1-S6')
    dataset = 'VT_1899_07_H1-S0'
    
    # Test
    rulesMaker = RuleMaker(config.get_SPARQL(), dataset, "/tmp/test.ttl")
    rulesMaker.loadMappings(config.get_path('mappings')) #, ['Sex','MaritalStatus']
    rulesMaker.loadHeaders(config.get_graph_name('raw-data'))
    rulesMaker.process() # ['Sex','MaritalStatus']