def fetchAndConvertToDCAT(self,
                              graph,
                              portal_ref,
                              portal_api,
                              snapshot,
                              activity,
                              format="ttl"):

        logger.debug('Fetching CKAN portal via RDF endpoint: ' + portal_api)

        with no_ssl_verification():
            graph.parse(portal_api, format=format)
            cur = graph.value(predicate=RDF.type,
                              object=namespaces['hydra'].PagedCollection)
            next_page = graph.value(subject=cur,
                                    predicate=namespaces['hydra'].nextPage)
            page = 0
            while next_page:
                page += 1
                if page % 10 == 0:
                    logger.debug('Processed pages:' + str(page))

                p = str(next_page)
                g = rdflib.Graph()
                g.parse(p, format=format)
                next_page = g.value(subject=URIRef(next_page),
                                    predicate=namespaces['hydra'].nextPage)
                graph.parse(p, format=format)

            logger.debug('Total pages:' + str(page))
            logger.info('Fetching finished')

            for d in graph.subjects(RDF.type, DCAT.Dataset):
                quality.add_quality_measures(d, graph, activity)
    def fetchAndConvertToDCAT(self, graph, portal_ref, portal_api, snapshot,
                              activity):

        with no_ssl_verification():
            graph = rdflib.Graph()
            graph.parse(portal_api, format="xml")

            for d in graph.subjects(RDF.type, DCAT.Dataset):
                quality.add_quality_measures(d, graph, activity)
    def fetchAndConvertToDCAT(self, graph, portal_ref, portal_api, snapshot,
                              activity):

        url = portal_api + "?format=text/turtle&query="
        query = """
        construct {?dataset a dcat:Dataset}  {
            ?dataset a dcat:Dataset.
        }
        """

        limit = 10000
        offset = 0
        download_url = url + urllib.parse.quote(query + " OFFSET " +
                                                str(offset) + " LIMIT " +
                                                str(limit))
        tmpgraph = rdflib.Graph()
        tmpgraph.parse(download_url, format='ttl')
        datasets = [d for d in tmpgraph.subjects(RDF.type, DCAT.Dataset)]

        with no_ssl_verification():
            while len(datasets) > 0:
                for dataset_uri in tmpgraph.subjects(RDF.type, DCAT.Dataset):
                    construct_query = """
                    CONSTRUCT {{ <{0}> ?p ?o. ?o ?q ?r}}
                    WHERE {{
                    <{0}> a dcat:Dataset.
                    <{0}> ?p ?o
                    OPTIONAL {{?o ?q ?r}}
                    }}
                    """.format(str(dataset_uri))

                    ds_url = url + urllib.parse.quote(construct_query)
                    graph.parse(ds_url, format='ttl')
                    graph.add((portal_ref, DCAT.dataset, dataset_uri))
                    graph.add((dataset_uri, RDF.type, DCAT.Dataset))
                    quality.add_quality_measures(dataset_uri, graph, activity)

                offset += limit
                download_url = url + urllib.parse.quote(query + " OFFSET " +
                                                        str(offset) +
                                                        " LIMIT " + str(limit))
                tmpgraph = rdflib.Graph()
                tmpgraph.parse(download_url, format='ttl')
                datasets = [
                    d for d in tmpgraph.subjects(RDF.type, DCAT.Dataset)
                ]
    def fetchAndConvertToDCAT(self,
                              graph,
                              portal_ref,
                              portal_api,
                              snapshot,
                              activity,
                              dcat=True):

        api = urllib.parse.urljoin(portal_api,
                                   '/api/1/datasets/?page_size=100')
        processed = set([])

        with no_ssl_verification():
            while True:
                resp = requests.get(api, verify=False)
                if resp.status_code != requests.codes.ok:
                    # TODO wait? appropriate message
                    pass

                res = resp.json()
                # returns a list of datasets
                if not res or 'data' not in res:
                    break
                for datasetJSON in res['data']:
                    if 'id' not in datasetJSON:
                        continue

                    datasetID = datasetJSON['id']
                    if datasetID not in processed:
                        processed.add(datasetID)
                        dataset_ref = graph_from_data_gouv_fr(
                            graph, datasetJSON, portal_api)
                        graph.add((portal_ref, DCAT.dataset, dataset_ref))
                        quality.add_quality_measures(dataset_ref, graph,
                                                     activity)

                        if len(processed) % 1000 == 0:
                            logger.info("ProgressDSFetch: " + portal_api +
                                        ", processed= " + str(len(processed)))
                if 'next_page' in res and res['next_page']:
                    api = res['next_page']
                else:
                    break
    def fetchAndConvertToDCAT(self, graph, portal_ref, portal_api, snapshot,
                              activity):

        start = 0
        rows = 10000
        processed = set([])

        with no_ssl_verification():
            while True:
                query = '/api/datasets/1.0/search?rows=' + str(
                    rows) + '&start=' + str(start)
                resp = requests.get(urllib.parse.urljoin(portal_api, query),
                                    verify=False)
                res = resp.json()
                datasets = res['datasets']
                if datasets:
                    rows = len(datasets) if start == 0 else rows
                    start += rows
                    for datasetJSON in datasets:
                        if 'datasetid' not in datasetJSON:
                            continue
                        datasetID = datasetJSON['datasetid']

                        if datasetID not in processed:
                            dataset_ref = graph_from_opendatasoft(
                                graph, datasetJSON, portal_api)
                            graph.add((portal_ref, DCAT.dataset, dataset_ref))
                            quality.add_quality_measures(
                                dataset_ref, graph, activity)
                            processed.add(datasetID)

                            if len(processed) % 1000 == 0:
                                logger.info("ProgressDSFetch: " + portal_api +
                                            ", processed= " +
                                            str(len(processed)))
                else:
                    break
    def fetchAndConvertToDCAT(self, graph, portal_ref, portal_api, snapshot,
                              activity):

        api = urllib.parse.urljoin(portal_api, '/api/')
        page = 1
        processed = set([])

        with no_ssl_verification():
            while True:
                resp = requests.get(urllib.parse.urljoin(
                    api, '/views/metadata/v1?page=' + str(page)),
                                    verify=False)
                if resp.status_code != requests.codes.ok:
                    # TODO wait? appropriate message
                    pass

                res = resp.json()
                # returns a list of datasets
                if not res:
                    break
                for datasetJSON in res:
                    if 'id' not in datasetJSON:
                        continue

                    datasetID = datasetJSON['id']
                    if datasetID not in processed:
                        dataset_ref = convert_socrata(graph, datasetJSON,
                                                      portal_api)
                        graph.add((portal_ref, DCAT.dataset, dataset_ref))
                        quality.add_quality_measures(dataset_ref, graph,
                                                     activity)
                        processed.add(datasetID)

                        if len(processed) % 1000 == 0:
                            logger.info("ProgressDSFetch: " + portal_api +
                                        ", processed= " + str(len(processed)))
                page += 1
示例#7
0
def dataset_to_ttl(datasetdata, graph, portal_uri, portal_api, portal_software,
                   activity):
    dataset_ref = dataset_converter.dict_to_dcat(datasetdata, graph,
                                                 portal_uri, portal_api,
                                                 portal_software)
    quality.add_quality_measures(dataset_ref, graph, activity)
    def fetchAndConvertToDCAT(self,
                              graph,
                              portal_ref,
                              portal_api,
                              snapshot,
                              activity,
                              timeout_attempts=5,
                              timeout=24 * 60 * 60):
        starttime = time.time()
        start = 0
        rows = 1000
        total = 0
        processed_ids = set([])
        processed_names = set([])

        try:
            with no_ssl_verification():
                session = requests.Session()
                session.verify = False
                api = ckanapi.RemoteCKAN(portal_api,
                                         get_only=True,
                                         session=session)
                response = api.action.package_search(rows=0)
                total = response["count"]
                # TODO store total

                while True:
                    response = self._get_datasets(api, timeout_attempts, rows,
                                                  start)

                    #print Portal.apiurl, start, rows, len(processed)
                    datasets = response["results"] if response else None
                    if datasets:
                        rows = len(datasets)
                        start += rows
                        for datasetJSON in datasets:
                            datasetID = datasetJSON['id']
                            try:
                                if datasetID not in processed_ids:
                                    converter = CKANConverter(
                                        graph, portal_api)
                                    dataset_ref = converter.graph_from_ckan(
                                        datasetJSON)
                                    graph.add((portal_ref, DCAT.dataset,
                                               dataset_ref))
                                    quality.add_quality_measures(
                                        dataset_ref, graph, activity)
                                    processed_ids.add(datasetID)
                                    processed_names.add(datasetJSON['name'])

                                    now = time.time()
                                    if now - starttime > timeout:
                                        raise TimeoutError(
                                            "Timeout of " + portal_api +
                                            " and " + str(timeout) +
                                            " seconds", timeout)
                            except Exception as e:
                                logger.error("CKANDSFetchDatasetBatchError: " +
                                             str(e))
                        rows = min([int(rows * 1.2), 1000])
                    else:
                        break
        except TimeoutError as e:

            raise e
        except Exception as e:
            logger.error("CKANDSFetchBatchError " + portal_api + ": " + str(e))

        if len(processed_ids) != total:
            logger.info("Not all datasets processed: fetched=" +
                        str(len(processed_ids)) + ", total=" + str(total))

            try:
                package_list, status = getPackageList(portal_api)
                tt = len(package_list)
                if total == 0:
                    # TODO store total tt
                    pass
                # TODO parameter:
                NOT_SUPPORTED_PENALITY = 100
                TIMEOUT_PENALITY = 100
                not_supported_count = 0
                timeout_counts = 0
                for entity in package_list:
                    #WAIT between two consecutive GET requests
                    if entity not in processed_ids and entity not in processed_names:
                        time.sleep(random.uniform(0.5, 1))
                        try:
                            resp, status = getPackage(apiurl=portal_api,
                                                      id=entity)
                            if resp:
                                data = resp
                                processed_names.add(entity)
                                converter = CKANConverter(graph, portal_api)
                                dataset_ref = converter.graph_from_ckan(data)
                                graph.add(
                                    (portal_ref, DCAT.dataset, dataset_ref))
                                quality.add_quality_measures(
                                    dataset_ref, graph, activity)
                                if entity not in processed_ids:
                                    processed_ids.add(entity)
                        except Exception as e:
                            logger.error('fetchDS: ' + str(e))

                            # if we get too much exceptions we assume this is not supported
                            not_supported_count += 1
                            if not_supported_count > NOT_SUPPORTED_PENALITY:
                                return

                    now = time.time()
                    if now - starttime > timeout:
                        raise TimeoutError(
                            "Timeout of " + portal_api + " and " +
                            str(timeout) + " seconds", timeout)
            except Exception as e:
                if len(processed_ids) == 0 or isinstance(e, TimeoutError):
                    raise e