Пример #1
0
    def test_datafile_init_invalid(self):
        """Test Datafile.init() with invalid data."""
        pdv = Datafile()

        # invalid data
        for data in ["invalid_set_types"]:
            with pytest.raises(AssertionError):
                pdv.set(data)
Пример #2
0
    def test_datafile_json_format_wrong_valid(self, import_datafile_min_dict):
        """Test Datafile.json() with non-valid format and valid data.

        Parameters
        ----------
        import_datafile_min_dict : dict
            Fixture, which returns a flat dataset dict().

        """
        data = import_datafile_min_dict
        df = Datafile()
        df.set(data)
        data = df.json('wrong')

        assert not data
Пример #3
0
    def test_datafile_dict_format_wrong(self, import_datafile_full_dict):
        """Test Datafile.dict() with non-valid format.

        Parameters
        ----------
        import_datafile_min_dict : dict
            Fixture, which returns a flat dataset dict().

        """
        data = import_datafile_full_dict
        df = Datafile()
        df.set(data)
        data = df.dict('wrong')

        assert not data
Пример #4
0
    def test_datafile_init(self):
        """Test Datafile.__init__()."""
        df = Datafile()

        assert not df.pid
        assert not df.filename
        assert not df.description
        assert not df.restrict

        df = Datafile('tests/data/datafile.txt', 'doi:10.11587/EVMUHP')

        assert df.pid == 'doi:10.11587/EVMUHP'
        assert df.filename == 'tests/data/datafile.txt'
        assert not df.description
        assert not df.restrict
Пример #5
0
    def test_datafile_init_valid(self):
        """Test Datafile.__init__() with valid data."""
        # specific
        data = [
            (Datafile(), {}),
            (Datafile(dict_flat_set_min()), object_data_min()),
            (Datafile(dict_flat_set_full()), object_data_full()),
            (Datafile({}), {}),
        ]

        for pdv, data_eval in data:
            for key, val in data_eval.items():
                assert getattr(pdv, key) == data_eval[key]
            assert len(pdv.__dict__) - len(
                object_data_init()) == len(data_eval)
Пример #6
0
    def test_datafile_json_all_valid(self, import_datafile_full_dict):
        """Test Datafile.json() with format=`all` and valid data.

        Parameters
        ----------
        import_datafile_full_dict : dict
            Fixture, which returns a flat dataset dict().

        """
        data = import_datafile_full_dict
        df = Datafile()
        df.set(data)
        data = df.json('all')

        assert data
        assert isinstance(data, str)
Пример #7
0
    def test_datafile_set_dv_up(self, import_datafile_full_dict):
        """Test Datafile.set() with format=`dv_up`.

        Parameters
        ----------
        import_datafile_full_dict : dict
            Fixture, which returns a flat datafile dict().

        """
        data = import_datafile_full_dict
        df = Datafile()
        df.set(data)

        assert df.pid == 'doi:10.11587/EVMUHP'
        assert df.filename == 'tests/data/datafile.txt'
        assert df.description == 'Test datafile'
        assert not df.restrict
Пример #8
0
    def test_datafile_is_valid_valid(self, import_datafile_full_dict):
        """Test Datafile.is_valid() with valid data.

        Parameters
        ----------
        import_datafile_min_dict : dict
            Fixture, which returns a flat datafile dict().

        """
        data = import_datafile_full_dict
        df = Datafile()
        df.set(data)

        assert df.pid == 'doi:10.11587/EVMUHP'
        assert df.filename == 'tests/data/datafile.txt'
        assert df.description == 'Test datafile'
        assert not df.restrict
        assert df.is_valid()
Пример #9
0
def data_object():
    """Get Datafile object.

    Returns
    -------
    pydataverse.models.Datafile
        :class:`Datafile` object.
    """
    return Datafile()
Пример #10
0
    def test_datafile_dict_all_valid(self, import_datafile_full_dict):
        """Test Datafile.dict() with format=`all` and valid data.

        Parameters
        ----------
        import_datafile_full_dict : dict
            Fixture, which returns a flat dataset dict().

        """
        data = import_datafile_full_dict
        df = Datafile()
        df.set(data)
        data = df.dict('all')

        assert data
        assert isinstance(data, dict)
        assert data['pid'] == 'doi:10.11587/EVMUHP'
        assert data['filename'] == 'tests/data/datafile.txt'
        assert data['description'] == 'Test datafile'
        assert not data['restrict']
Пример #11
0
    def test_datafile_json_dv_up_valid_not(self, import_datafile_min_dict):
        """Test Datafile.json() with format=`dv_up` and non-valid data.

        Parameters
        ----------
        import_datafile_min_dict : dict
            Fixture, which returns a flat dataset dict().

        """
        data = import_datafile_min_dict
        df = Datafile()
        df.set(data)
        df.filename = None

        assert not df.is_valid()
        print(df.json('dv_up'))
        assert not df.json('dv_up')
Пример #12
0
    def test_datafile_dict_dv_up_valid_not(self, import_datafile_min_dict):
        """Test Datafile.dict() with format=`dv_up` and non-valid data.

        Parameters
        ----------
        import_datafile_min_dict : dict
            Fixture, which returns a flat dataset dict().

        """
        data = import_datafile_min_dict
        df = Datafile()
        df.set(data)
        df.pid = None

        assert not df.is_valid()
        assert df.filename == 'tests/data/datafile.txt'
Пример #13
0
def create_testdata(config_file: str, force: bool) -> None:
    """Create testdata defined in a config file.

    Creates a pre-defined set of testdata on your
    instance. By default, the function uses the
    AUSSDA test data repository, which is so far not
    publicly available. If `PRODUCTION` is `true`,
    this function will not execute, as long as you
    not add `--force` to the function call. This is
    to protect from unwanted changes on a production
    instance.

    """
    # Init
    if config.PRODUCTION and not force:
        print(
            "Create testdata on a PRODUCTION instance not allowed. Use --force to force it."
        )
        sys.exit()
    pid_idx = []
    users = read_json(config.USER_FILENAME)
    workflow = read_json(os.path.join(ROOT_DIR, config_file))

    # Dataverses
    for dv_conf in workflow["dataverses"]:
        dv_alias = None
        if "create" in dv_conf:
            api = NativeApi(
                config.BASE_URL,
                users[dv_conf["create"]["user-handle"]]["api-token"])
            dv = Dataverse()
            dv_filename = os.path.join(ROOT_DIR,
                                       dv_conf["create"]["metadata-filename"])
            dv.from_json(read_file(dv_filename))
            if "update" in dv_conf["create"]:
                for key, val in dv_conf["create"]["update"].items():
                    kwargs = {key: val}
                    dv.set(kwargs)
            dv_alias = dv.get()["alias"]
            resp = api.create_dataverse(dv_conf["create"]["parent"], dv.json())

        if "publish" in dv_conf:
            api = NativeApi(
                config.BASE_URL,
                users[dv_conf["publish"]["user-handle"]]["api-token"])
            if not dv_alias and "alias" in dv_conf["publish"]:
                dv_alias = dv_conf["publish"]["alias"]
            resp = api.publish_dataverse(dv_alias)

    # Datasets
    for ds_conf in workflow["datasets"]:
        pid = None
        if "create" in ds_conf:
            api = NativeApi(
                config.BASE_URL,
                users[ds_conf["create"]["user-handle"]]["api-token"])
            ds = Dataset()
            ds_filename = os.path.join(ROOT_DIR,
                                       ds_conf["create"]["metadata-filename"])
            ds.from_json(read_file(ds_filename))
            if "update" in ds_conf["create"]:
                for key, val in ds_conf["create"]["update"].items():
                    kwargs = {key: val}
                    ds.set(kwargs)
            resp = api.create_dataset(dv_alias, ds.json())
            pid = resp.json()["data"]["persistentId"]
            pid_idx.append(pid)

        if "publish" in ds_conf:
            if not pid:
                print("ERROR: PID missing!")
                sys.exit()
            api = NativeApi(
                config.BASE_URL,
                users[ds_conf["publish"]["user-handle"]]["api-token"])
            resp = api.publish_dataset(pid, release_type="major")

    # Datafiles
    for dataset_id, ds_datafiles in workflow["datafiles"].items():
        if int(dataset_id) == workflow["datasets"][int(dataset_id)]["id"]:
            pid = pid_idx[int(dataset_id)]
        else:
            print("ERROR: Dataset ID not matching.")
            sys.exit()
        for df_conf in ds_datafiles:
            if "upload" in df_conf:
                api = NativeApi(
                    config.BASE_URL,
                    users[df_conf["upload"]["user-handle"]]["api-token"],
                )
                metadata = read_json(df_conf["upload"]["metadata-filename"])
                df = Datafile()
                df.set(metadata)
                if "update" in df_conf["upload"]:
                    for key, val in df_conf["upload"]["update"].items():
                        kwargs = {key: val}
                        df.set(kwargs)
                df.set({"pid": pid})
                filename = df_conf["upload"]["filename"]
                resp = api.upload_datafile(pid, filename, df.json())
                if filename[-4:] == ".sav" or filename[-4:] == ".dta":
                    sleep(30)
                else:
                    sleep(3)
        if "publish-dataset" in df_conf:
            api = NativeApi(
                config.BASE_URL,
                users[df_conf["publish-dataset"]["user-handle"]]["api-token"],
            )
            if df_conf["publish-dataset"]:
                resp = api.publish_dataset(pid, release_type="major")