def test_chunksizes(tmpnetcdf_filename):
    nco = create_netcdf(tmpnetcdf_filename)

    x = numpy.arange(3, dtype='float32')
    y = numpy.arange(5, dtype='float32')

    coord1 = create_coordinate(nco, 'x', x, 'm')
    coord2 = create_coordinate(nco, 'y', y, 'm')

    assert coord1 is not None and coord2 is not None

    no_chunks = create_variable(nco, 'no_chunks',
                                Variable(numpy.dtype('int16'), None, ('x', 'y'), None))

    min_max_chunks = create_variable(nco, 'min_max_chunks',
                                     Variable(numpy.dtype('int16'), None, ('x', 'y'), None),
                                     chunksizes=(2, 50))

    assert no_chunks is not None
    assert min_max_chunks is not None

    strings = numpy.array(["AAa", 'bbb', 'CcC'], dtype='S')
    strings = xr.DataArray(strings, dims=['x'], coords={'x': x})
    create_variable(nco, 'strings_unchunked', strings)
    create_variable(nco, 'strings_chunked', strings, chunksizes=(1,))

    nco.close()

    with netCDF4.Dataset(tmpnetcdf_filename) as nco:
        assert nco['no_chunks'].chunking() == 'contiguous'
        assert nco['min_max_chunks'].chunking() == [2, 5]
        assert nco['strings_unchunked'].chunking() == 'contiguous'
        assert nco['strings_chunked'].chunking() == [1, 3]
Exemple #2
0
def make_fake_netcdf_dataset(nc_name, doc_text):
    t = np.asarray([parse_time('2001-01-29 07:06:05.432')], dtype=np.datetime64)
    npdata = np.asarray([doc_text], dtype='S')

    with create_netcdf(nc_name) as nco:
        create_coordinate(nco, 'time', t, 'seconds since 1970-01-01 00:00:00')

        nc_dataset = create_variable(nco, 'dataset',
                                     Variable(npdata.dtype, None, ('time',), None))
        nc_dataset[:] = netcdfy_data(npdata)
        assert 'dataset_nchar' in nco.dimensions
def test_create_string_variable(tmpnetcdf_filename, s1, s2, s3):
    str_var = 'str_var'
    nco = create_netcdf(tmpnetcdf_filename)
    coord = create_coordinate(nco, 'greg', numpy.array([1.0, 3.0, 9.0]), 'cubic gregs')
    assert coord is not None

    dtype = numpy.dtype('S100')
    data = numpy.array([s1, s2, s3], dtype=dtype)

    var = create_variable(nco, str_var, Variable(dtype, None, ('greg',), None))
    var[:] = netcdfy_data(data)
    nco.close()

    with netCDF4.Dataset(tmpnetcdf_filename) as nco:
        assert str_var in nco.variables

    for returned, expected in zip(read_strings_from_netcdf(tmpnetcdf_filename, variable=str_var), (s1, s2, s3)):
        assert returned == expected
Exemple #4
0
def make_fake_netcdf_dataset(nc_name, doc_text):
    from datacube.drivers.netcdf.writer import (
        Variable,
        create_variable,
        create_coordinate,
        netcdfy_data,
        create_netcdf
    )
    from datacube.utils.dates import parse_time
    import numpy as np

    t = np.asarray([parse_time('2001-01-29 07:06:05.432')], dtype=np.datetime64)
    npdata = np.asarray([doc_text], dtype='S')

    with create_netcdf(nc_name) as nco:
        create_coordinate(nco, 'time', t, 'seconds since 1970-01-01 00:00:00')

        nc_dataset = create_variable(nco, 'dataset',
                                     Variable(npdata.dtype, None, ('time',), None))
        nc_dataset[:] = netcdfy_data(npdata)
        assert 'dataset_nchar' in nco.dimensions