def test_should_copy_all_object_from_latest_to_tag(self): numeric_version = 0.8 with self.app.app_context(): bit_store = BitStore('test_pub', 'test_package') s3 = boto3.client('s3') bucket_name = self.app.config['S3_BUCKET_NAME'] s3.create_bucket(Bucket=bucket_name) read_me_key = bit_store.build_s3_key('test.md') data_key = bit_store.build_s3_key('data.csv') metadata_key = bit_store.build_s3_key('datapackage.json') s3.put_object(Bucket=bucket_name, Key=read_me_key, Body='readme') s3.put_object(Bucket=bucket_name, Key=data_key, Body='data') s3.put_object(Bucket=bucket_name, Key=metadata_key, Body='metedata') bit_store.copy_to_new_version(numeric_version) bit_store_numeric = BitStore('test_pub', 'test_package', numeric_version) objects_nu = s3.list_objects( Bucket=bucket_name, Prefix=bit_store_numeric.build_s3_versioned_prefix()) objects_old = s3.list_objects( Bucket=bucket_name, Prefix=bit_store.build_s3_versioned_prefix()) self.assertEqual(len(objects_nu['Contents']), len(objects_old['Contents']))
def test_change_acl(self): with self.app.app_context(): bit_store = BitStore('test_pub', 'test_package') s3 = boto3.client('s3') bucket_name = self.app.config['S3_BUCKET_NAME'] s3.create_bucket(Bucket=bucket_name) read_me_key = bit_store.build_s3_key('test.md') data_key = bit_store.build_s3_key('data.csv') metadata_key = bit_store.build_s3_key('datapackage.json') s3.put_object(Bucket=bucket_name, Key=read_me_key, Body='') s3.put_object(Bucket=bucket_name, Key=data_key, Body='') s3.put_object(Bucket=bucket_name, Key=metadata_key, Body='') bit_store.change_acl("private") res = s3.get_object_acl(Bucket=bucket_name, Key=read_me_key) grant = res['Grants'][0]['Permission'] self.assertEqual(grant, 'FULL_CONTROL')
def test_return_none_if_no_readme_found(self): with self.app.app_context(): bit_store = BitStore('test_pub', 'test_package') s3 = boto3.client('s3') bucket_name = self.app.config['S3_BUCKET_NAME'] s3.create_bucket(Bucket=bucket_name) read_me_key = bit_store.build_s3_key('test.md') s3.put_object(Bucket=bucket_name, Key=read_me_key, Body='') self.assertEqual(bit_store.get_readme_object_key(), None)
def test_get_metadata_body(self): with self.app.app_context(): s3 = boto3.client('s3') bucket_name = self.app.config['S3_BUCKET_NAME'] s3.create_bucket(Bucket=bucket_name) metadata = BitStore(publisher="pub_test", package="test_package", body='hi') s3.put_object(Bucket=bucket_name, Key=metadata.build_s3_key('datapackage.json'), Body=metadata.body) self.assertEqual(metadata.body, metadata.get_metadata_body())
def test_throw_403_if_not_owner_or_member_of_publisher(self): s3 = boto3.client('s3') s3.create_bucket(Bucket=self.bucket_name) bit_store = BitStore('test_pub', 'test_package') read_me_key = bit_store.build_s3_key('test.md') data_key = bit_store.build_s3_key('data.csv') metadata_key = bit_store.build_s3_key('datapackage.json') s3.put_object(Bucket=self.bucket_name, Key=read_me_key, Body='readme') s3.put_object(Bucket=self.bucket_name, Key=data_key, Body='data') s3.put_object(Bucket=self.bucket_name, Key=metadata_key, Body='metedata') response = self.client.post(self.jwt_url, data=json.dumps({ 'username': self.user_not_allowed_name, 'secret': 'super_secret' }), content_type='application/json') data = json.loads(response.data) jwt_not_allowed = data['token'] auth_not_allowed = "bearer %s" % jwt_not_allowed response = self.client.post( self.url, data=json.dumps({'version': 'tag_one'}), content_type='application/json', headers=dict(Authorization=auth_not_allowed)) self.assertEqual(response.status_code, 403) with self.app.app_context(): data_latest = MetaDataDB.query.join(Publisher). \ filter(Publisher.name == self.publisher_name, MetaDataDB.name == self.package).all() self.assertEqual(1, len(data_latest)) bit_store_tagged = BitStore('test_pub', 'test_package', 'tag_one') objects_nu = s3.list_objects( Bucket=self.bucket_name, Prefix=bit_store_tagged.build_s3_versioned_prefix()) self.assertTrue('Contents' not in objects_nu)
def test_delete_data_package(self): with self.app.app_context(): bit_store = BitStore('test_pub', 'test_package') s3 = boto3.client('s3') bucket_name = self.app.config['S3_BUCKET_NAME'] s3.create_bucket(Bucket=bucket_name) read_me_key = bit_store.build_s3_key('test.md') data_key = bit_store.build_s3_key('data.csv') metadata_key = bit_store.build_s3_key('datapackage.json') s3.put_object(Bucket=bucket_name, Key=read_me_key, Body='readme') s3.put_object(Bucket=bucket_name, Key=data_key, Body='data') s3.put_object(Bucket=bucket_name, Key=metadata_key, Body='metedata') status = bit_store.delete_data_package() read_me_res = s3.list_objects(Bucket=bucket_name, Prefix=read_me_key) self.assertTrue('Contents' not in read_me_res) data_res = s3.list_objects(Bucket=bucket_name, Prefix=data_key) self.assertTrue('Contents' not in data_res) self.assertTrue(status)
def test_save(self): with self.app.app_context(): s3 = boto3.client('s3') bucket_name = self.app.config['S3_BUCKET_NAME'] s3.create_bucket(Bucket=bucket_name) metadata = BitStore(publisher="pub_test", package="test_package", body='hi') key = metadata.build_s3_key('datapackage.json') metadata.save() obs_list = list(s3.list_objects(Bucket=bucket_name, Prefix=key).\ get('Contents')) self.assertEqual(1, len(obs_list)) self.assertEqual(key, obs_list[0]['Key'])
def populate_data(publisher_name): data = json.loads(open('fixtures/datapackage.json').read()) data_csv = open('fixtures//data/demo-resource.csv').read() readme = open('fixtures/README.md').read() publisher = Publisher.query.filter_by(name=publisher_name).one() metadata = MetaDataDB(name="demo-package") metadata.descriptor, metadata.status, metadata.private, metadata.readme \ = json.dumps(data), 'active', False, readme publisher.packages.append(metadata) db.session.add(publisher) db.session.commit() bitstore = BitStore(publisher_name, package='demo-package') key = bitstore.build_s3_key('demo-resource.csv') bucket_name = app.config['S3_BUCKET_NAME'] s3_client = app.config['S3'] s3_client.put_object(Bucket=bucket_name, Key=key, Body=data_csv)
def get_resource(publisher, package, resource): """ DPR resource get operation. This API is responsible for getting resource from S3. --- tags: - package parameters: - in: path name: publisher type: string required: true description: publisher name - in: path name: package type: string required: true description: package name - to retrieve the data package metadata - in: path name: resource type: string required: true description: resource index or name responses: 200: description: Get Data package for one key schema: id: get_data_package properties: data: type: string description: The resource 500: description: Internal Server Error """ try: path = request.path metadata = BitStore(publisher, package) if path.endswith('csv'): resource_key = metadata.build_s3_key(resource + '.csv') data = metadata.get_s3_object(resource_key) def generate(): for row in data.splitlines(): yield row + '\n' return Response(generate()), 200 else: resource_key = metadata.build_s3_key(resource + '.csv') data = metadata.get_s3_object(resource_key) data = csv.DictReader(data.splitlines()) # taking first and adding at the end to avoid last comma first_row = next(data) def generate(): yield '[' for row in data: yield json.dumps(row) + ',' yield json.dumps(first_row) + ']' return Response(generate(), content_type='application/json'), 200 except Exception as e: return handle_error('GENERIC_ERROR', e.message, 500)
def test_metadata_s3_key(self): metadata = BitStore(publisher="pub_test", package="test_package") expected = "{t}/pub_test/test_package/_v/latest/datapackage.json".\ format(t=metadata.prefix) self.assertEqual(expected, metadata.build_s3_key('datapackage.json'))