Beispiel #1
0
def get_data_by_id(data_id):
    """ Return data object with the given id.

        Parameters:
            data_id:

        Returns: data object
    """
    return Data.get_by_id(data_id)
def get_pid_for_data(data_id):
    """Retrieve PID matching the document ID provided.

    Args:
        data_id:

    Returns:
    """
    # Retrieve the document passed as input and extra the PID field.
    data = Data.get_by_id(data_id)

    # Return PID value from the document and the PID_XPATH
    return get_dict_value_from_key_list(data["dict_content"],
                                        PID_XPATH.split("."))
def download_source_file(request):
    """

    Args:
        request:

    Returns:

    """
    doc_id = request.GET.get('doc_id', None)
    data = Data.get_by_id(doc_id)
    return get_file_http_response(file_content=data.xml_content,
                                  file_name=data.title,
                                  content_type="text/xml",
                                  extension="xml")
Beispiel #4
0
 def test_data_get_by_id_return_data_if_found(self):
     # Act
     result = Data.get_by_id(self.fixture.data_1.id)
     # Assert
     self.assertEqual(result, self.fixture.data_1)
Beispiel #5
0
 def test_data_get_by_id_raises_api_error_if_not_found(self):
     # Act # Assert
     with self.assertRaises(exceptions.DoesNotExist):
         Data.get_by_id(ObjectId())
def _load_data_view(node_id, nav_id, data_id, from_tree=True):
    """ Load view for a data, from a tree or a link

    Args:
        node_id:
        nav_id:
        data_id:
        from_tree:

    Returns:

    """
    if not from_tree:
        navigation_node = navigation_operations.get_navigation_node_for_document(
            node_id, data_id)
    else:
        navigation_node = navigation_api.get_by_id(node_id)

    # Initialize parameters in order to download later some information
    xml_document = Data.get_by_id(data_id)
    projection_views = json.loads(navigation_node.options["projection_view"])

    view_data = {
        "header": xml_document.title,
        "type": "leaf",
        "views": [],
        "download": []
    }
    # Initialize parameters in order to download later some information
    # dict of doc_id and queries done of cross documents : {id_doc1: [list of queries1], id_doc2: [list of queries2]}
    dict_id_and_queries_cross_docs = dict()
    # dict of doc_id and queries results for a cross document : {id_doc: [list of queries results]}
    dict_id_and_queryresults_cross_docs = dict()

    # dict of queried parameter and associated result for the queries done on the main doc : { queried parameter: value}
    dict_tags_values_main_doc = dict()

    values_of_items_from_main_doc = []
    list_values_of_items_from_main_doc = []

    # Send the annotation to the processor and collect the data
    for projection_view in projection_views:
        result_data = {"title": projection_view["title"], "data": None}

        # FIXME better handling of x-queries
        # Get info from other doc (without the main queried document)
        if "query" in projection_view.keys():
            doc_projections = []
            # Get the names of the tags tag need to be displayed
            for value in projection_view["data"]:
                doc_projections.append(value.get('path'))

            result_data["data"] = parser_processview.process_cross_query(
                nav_id, data_id, projection_view["query"],
                projection_view["data"])
            # Get all the queried documents (without the main queried document)
            queried_docs = parser_processview.ids_docs_to_querys
            for id_doc in queried_docs:
                other_doc_query = {"_id": ObjectId(id_doc)}
                # list of queries done on the current document
                query_list = list()
                # list of queries results done on the current document
                result_list = list()

                for projection in doc_projections:
                    # Get the MongoDB query path for the parameter that need to be displayed
                    # eg: query_path = dict_content.a.b.c.d.e
                    query_path = {
                        doc_projections[doc_projections.index(projection)]: 1
                    }
                    # Get the Data corresponding to the id
                    queried_data = Data.execute_query(other_doc_query).only(
                        query_path.keys()[0])
                    # Add the query to the query list for the current doc
                    query_list.append(query_path.keys()[0].replace(
                        "dict_content.", ""))
                    try:
                        # Get the result of the query
                        result_query = get_projection(queried_data[0])
                        # Add the result of the query to the result list for the current doc
                        result_list.append(str(result_query))
                    except:
                        pass
                dict_id_and_queries_cross_docs[id_doc] = query_list
                dict_id_and_queryresults_cross_docs[id_doc] = result_list
        # Get info from main doc
        else:
            # list of queries done on the current document (Main doc)
            query_list = []
            doc_projections = [
                value.get('path') for value in projection_view["data"]
            ]
            query_list = [
                doc_projections[doc_projections.index(projection)]
                for projection in doc_projections
            ]
            # Get all results of the queries. type(result_data["data"]) = dict or instance of dict
            result_data["data"] = parser_processview.processview(
                nav_id, data_id, projection_view["data"])

            for query_path, dict_result in zip(query_list,
                                               result_data["data"]):
                # eg: query_path = a.b.c.d
                # We have only one value as result for the query
                dict_result_value = dict_result.get("value", None)
                if dict_result_value is not None:
                    tag = query_path.split(".")[
                        -1]  # Get only d (the needed tag)
                    if tag in dict_tags_values_main_doc:
                        v = dict_tags_values_main_doc[tag]
                        if isinstance(v, list):
                            dict_tags_values_main_doc[tag].append(
                                dict_result_value)
                        else:
                            dict_tags_values_main_doc[tag] = [
                                dict_tags_values_main_doc[tag],
                                dict_result_value
                            ]
                    else:
                        dict_tags_values_main_doc[tag] = dict_result_value

                # We have multiple values for this result: all the chemical components
                # (dict_result[key] is an inclusion of dicts)
                dict_result_item, dict_result_items = [
                    dict_result.get(_, None) for _ in ["item", "items"]
                ]

                if dict_result_item or dict_result_items:
                    dict_result_item_v = dict_result_item if dict_result_item is not None else dict_result_items
                    #dict_result_item_v = [dict_result_item, dict_result_items][dict_result_item not None]
                    # From the inclusion of dict, process the dict into a list and get all the needed values
                    # values_of_items_from_main_doc = list[list[value1 for dict i,value2 for dict 2, ..]]
                    # eg: values_of_items_from_main_doc= [l1,l2]
                    # l1 = [["location", "NIST"], ["Build location X", "59"], "EWI_Build1"]]
                    # l2 = [["location", "NIST"], ["Build location X", "47"], "EWI_Build2"]]
                    get_values_items(dict_result_item_v,
                                     values_of_items_from_main_doc)
                    for list_of_values in values_of_items_from_main_doc:
                        for value in list_of_values:
                            # We have a list :value= [displayed parameter, value]
                            # eg : ["Build location X", "59"]
                            if len(value) == 2:
                                # list_tag_of_items_from_main_doc.append(value[0])
                                list_values_of_items_from_main_doc.append(
                                    value[1])  # Get the value. eg: 59
                            # We have only one value (last value in the list. eg: EWI_Build1 in l1)
                            else:
                                list_values_of_items_from_main_doc.append(
                                    value)
        view_data["views"].append(result_data)

    # Get the displayed data as an XML format in order to download it later #

    # STEP 1: Build the XML based on initial tags for the crossed documents:
    # Go into the dict of doc_id and queries of cross documents and build the xml for each document
    #  dict_id_and_queries_cross_docs = {id_doc1: [list of queries1], id_doc2: [list of queries2]}
    xml_cross_queries_string = ""
    for key in dict_id_and_queries_cross_docs:  # key = doc_id
        # Get all queries for the current doc_id.
        # eg: query_list = ["a.b.c.d","a.b.c.e","a.b.f.g"]
        query_list = dict_id_and_queries_cross_docs[key]
        # For the doc_id get all the results of the queries done
        # results = ["D","E","G"]
        results = dict_id_and_queryresults_cross_docs[key]
        # Build a xml string for the doc associated to doc_id thanks to the list of queries and the result list
        xml_string = queryNode.tree_to_xml_string(
            queryNode.aggregate_query(query_list, results))
        xml_object = XSDTree.fromstring(xml_string + "</data>")
        # Add the XML part to create an XML resulting of tag and values of crossed documents
        xml_cross_queries_string += XSDTree.tostring(xml_object, pretty=True)

    # STEP 2: Build the XML for the main document with only the needed tags:
    # Get the Data associated to the main document
    data = Data.get_by_id(data_id)
    # Get the XML content
    file_content = data.xml_content
    xml_main_doc = XSDTree.fromstring(file_content)
    # Transform all the result value into a string to help while testing equality of values with the original XML
    for key, value in dict_tags_values_main_doc.items():
        if isinstance(value, list):
            dict_tags_values_main_doc[key] = map(
                lambda x: x if isinstance(x, unicode) else str(x), value)
        else:
            try:
                dict_tags_values_main_doc[key] = str(value)
            except:
                dict_tags_values_main_doc[key] = u''.join(value).encode(
                    'utf-8')
        v = dict_tags_values_main_doc[key]

    # Process the XML structure that represents the main document to keep only the needed tags and information
    for child in xml_main_doc.iter():
        # Transform all values into a string
        try:
            text = str(child.text)
        except:
            text = u''.join(child.text).encode('utf-8')
        # If the xml tag is in our dict of tags and values from the main document
        # and its value = dict_tags_values_main_doc[child.tag] we keep the text in the XML structure
        # else we remove the text
        if child.tag in dict_tags_values_main_doc.keys():
            # Fixme  # if text != str(dict_tags_values_main_doc[child.tag]) or dict_tags_values_main_doc[child.tag] not in text: (caution: does not keep all needed values if we replace by this line)
            if isinstance(dict_tags_values_main_doc[child.tag], list):
                display_value = False
                for value in dict_tags_values_main_doc[child.tag]:
                    if value == text or value in text:
                        display_value = True
                        break
                if not display_value:
                    child.text = ""
            else:
                if text == str(
                        dict_tags_values_main_doc[child.tag]
                ) or dict_tags_values_main_doc[child.tag] in text:
                    pass
                else:
                    child.text = ""
        else:
            # If text is in our list of items of the main doc we keep the value and remove it from our list of items
            if text in list_values_of_items_from_main_doc:
                list_values_of_items_from_main_doc.remove(text)
            else:
                display_text = False
                # v = processed name of the tag as appears in the rendered data after clicking a doc of the tree
                # If this name is in our list of items from the main doc we keep the value (text) in the XML tree
                # else we remove this value
                for v in list_values_of_items_from_main_doc:
                    if v in text:
                        display_text = True
                        break
                if not display_text:
                    child.text = ""

    xml_f_main_doc = xml_main_doc
    # Remove empty leafs of the tree (all child where child.text="")
    while check_empty_nodes(xml_main_doc):
        remove_empty_nodes(xml_f_main_doc)

    # Build the final XML string result of the main doc and the crossed docs
    xml_main_doc = XSDTree.tostring(xml_f_main_doc, pretty="TRUE")
    xml = xml_main_doc + xml_cross_queries_string
    xml_final = "<xml>\n" + xml + "</xml>"
    xml_final = u''.join(xml_final).encode('utf-8')
    xml_final = str(xml_final)
    view_data["download"] = xml_final
    return view_data
Beispiel #7
0
def processview(navigation_root_id, document_id, json_content):
    """
    # Function Name: processview(document_id,json_content)
    # Inputs:        document_id - Document ID
    #                json_content - name/path combinations to query
    # Outputs:       Query results page
    # Exceptions:    None
    # Description:   Processes view name/path combination and outputs name/value stored in MongoDB based on document_id

    Parameters:
        navigation_root_id:
        document_id:
        json_content:

    Returns:

    """
    object1 = Data.get_by_id(document_id).dict_content
    tempobject1 = object1

    pviewoutlist = []
    if object1 is None:
        # FIXME error case not working if data is arrayType
        for item in json_content:
            pviewdict = {"name": item["name"], "error": item["path"]}

            pviewoutlist.append(pviewdict)
    else:
        for item in json_content:
            if "type" in item and item["type"] == "array":
                pviewoutlist.append(
                    {"items": process_array_type(object1, item)})
            else:
                pathflag = True
                pviewname = item['name']
                path = item['path']
                pviewdict = {}
                pathstring = path.split('.')

                if 'dict_content' in pathstring:
                    pathstring.remove("dict_content")

                for pstringitem in pathstring:
                    tempobject1 = tempobject1.get(pstringitem)

                    if tempobject1 is None:
                        pathflag = False
                        pviewdict['name'] = pviewname
                        pviewdict['error'] = path

                        my_dict[pviewname] = path
                        my_list.append(my_dict)

                        pviewoutlist.append(pviewdict)
                        break

                if pathflag:
                    pviewdict['name'] = pviewname
                    pviewdict['value'] = tempobject1

                    my_dict[pviewname] = tempobject1

                    if "link" in item:
                        linked_node = get_navigation_node_by_name(
                            navigation_root_id, item["link"])

                        if linked_node is None:
                            pviewdict["link"] = "error"
                            my_dict["link"] = "error"
                        else:
                            pviewdict["link"] = "%s %s" % (str(
                                linked_node.pk), document_id)
                            my_dict["link"] = pviewdict["link"]
                    pviewoutlist.append(pviewdict)
                    global i
                    if i == 0:
                        my_list.append(my_dict)
                        i += 1
                    else:
                        for l in my_list:
                            if l != my_dict:
                                my_list.append(my_dict)

                tempobject1 = object1

    return pviewoutlist
def get_navigation_node_for_document(node_id, document_id):
    """ Get the navigation node from a given data id

    Args:
        node_id:
        document_id:

    Returns:

    """
    navigation_node = navigation_api.get_by_id(node_id)
    original_node = navigation_api.get_by_id(node_id)

    if "projection" in navigation_node.options and navigation_node.options[
            'projection'] is not None:
        # **************
        # FIXME duplicate of parser code
        # Get projection

        filters = []
        while True:
            # add filter to the list of filters
            if 'filter' in navigation_node.options and navigation_node.options[
                    'filter'] is not None:
                filters.append(navigation_node.options['filter'])

            # if no parent, stops the filter lookup
            if navigation_node.parent is None:
                break
            else:  # if parent, continue the filter lookup at the parent level
                navigation_node = navigation_api.get_by_id(
                    navigation_node.parent)

        # get the documents matching the query
        doc = Data.get_by_id(document_id)
        template_id = doc.template.id

        documents_id = [
            str(get_projection(document)) for document in query.execute_query(
                template_id, filters, '{"id": 1}')
        ]
        #  End fixme
        # **************

        # If the document is one of the document we return the navigation node, else return None
        if document_id in documents_id:
            return original_node
        else:
            return None
    else:
        navigation_children = navigation_node.children

        if len(navigation_children) == 0:
            return None
        else:
            navigation_nodes = [
                get_navigation_node_for_document(child_id, document_id)
                for child_id in navigation_children
            ]
            navigation_nodes = [
                nav_node for nav_node in navigation_nodes
                if nav_node is not None
            ]

            if len(navigation_nodes) != 1:
                return None
            else:
                return navigation_nodes[0]
def load_view(request):
    """Load view entry point

    Args:
        request:

    Returns:

    """
    # nav_id parameter is mandatory, if it doesn't exist we delete other parameters to raise and error
    if "nav_id" not in request.POST:
        error = {"message": "Request is malformed. nav_id is mandatory"}
        return HttpResponse(json.dumps(error), status=HTTP_400_BAD_REQUEST)
    # retrieve all the parameters from the request
    nav_id = request.POST.get("nav_id", None)
    node_id = request.POST.get("node_id", None)
    doc_id = request.POST.get("doc_id", None)
    ref_node_id = request.POST.get("ref_node_id", None)
    load_doc = None
    # Get all DataCached objects:
    alldatacached = DataCached.get_all()
    # create dict of keys values for the cache
    listof = {}
    for datacached in alldatacached:
        dict_keys_docids = datacached.cached_documents_dict
        for dict_key_docid in dict_keys_docids:
            listof.update(dict_key_docid)
    # click on a document in the tree
    if node_id is not None and doc_id is not None:
        node_name = get_node_name(node_id)
        id_docleaf_cached = str(node_name) + "_" + str(doc_id)
        # Get the document from the cache if this one had ever been accessed
        for key, value in list(listof.items()):
            if id_docleaf_cached == key:
                # from leaf cache
                load_doc = leaf_cache.get(key)
                if load_doc:
                    break
                else:
                    if id_docleaf_cached == value:
                        # from link cache
                        load_doc = link_cache.get(key)
                        leaf_cache.set(id_docleaf_cached, load_doc)
                        break
            elif id_docleaf_cached == value:
                # from link cache
                load_doc = link_cache.get(key)
                leaf_cache.set(id_docleaf_cached, load_doc)
                break
        if not load_doc:
            # Else :Query the database, process the documents
            load_doc = _load_data_view(node_id, nav_id, doc_id)
            leaf_cache.set(id_docleaf_cached, load_doc)
            doc = Data.get_by_id(doc_id)
            # create or update the Datacached model representing a Node and its cached files
            dict_key_docid = {}
            dict_key_docid[id_docleaf_cached] = id_docleaf_cached
            data_cached_api.upsert_data_cache_object(node_name, doc,
                                                     dict_key_docid)

    # click on a node in the tree
    elif node_id is not None:
        if node_id in branch_cache:
            load_doc = branch_cache.get(node_id)
        else:
            load_doc = _load_branch_view(request)
            branch_cache.set(node_id, load_doc)

    # click on a link in a document (build id)
    elif ref_node_id is not None and doc_id is not None:
        navigation_node = None
        id_doclink_cached = ref_node_id + "_" + doc_id
        # The file was cached by clicking a link from a document
        if id_doclink_cached in list(
                listof.keys()) or id_doclink_cached in link_cache:
            load_doc = link_cache.get(id_doclink_cached)
        else:
            navigation_node = navigation_operations.get_navigation_node_for_document(
                ref_node_id, doc_id)
            nodename = navigation_node.name
            nodename_index = nodename.find("#")
            node_name = nodename[nodename_index + 1:]
            id_doc_cached = node_name + "_" + str(doc_id)
            # The file was cached by using the cache manager tool
            if id_doc_cached in list(listof.keys()):
                load_doc = leaf_cache.get(id_doc_cached)
                link_cache.set(id_doclink_cached, load_doc)
        if not load_doc:
            # The file was never cached
            if not navigation_node:
                navigation_node = (
                    navigation_operations.get_navigation_node_for_document(
                        ref_node_id, doc_id))
            load_doc = _load_data_view(navigation_node.id, nav_id, doc_id)
            link_cache.set(id_doclink_cached, load_doc)

            doc = Data.get_by_id(doc_id)
            dict_key_docid = {}
            dict_key_docid[id_doclink_cached] = id_doc_cached
            # create or update the DataCached model representing a Node and its cached files
            data_cached_api.upsert_data_cache_object(node_name, doc,
                                                     dict_key_docid)

    else:
        error = {"message": "Request is malformed."}
        return HttpResponse(json.dumps(error), status=HTTP_400_BAD_REQUEST)

    if load_doc is not None:
        return render(request,
                      "core_explore_tree_app/user/explore_tree/view.html",
                      load_doc)