def main(): service = auth.build_bq_client() # Load configuration with the destination specified. load_config = { 'destinationTable': { 'projectId': auth.PROJECT_ID, 'datasetId': 'ch06', # You can update this for each example. 'tableId': 'example_bad_data' } } # Setup the job here. # load[property] = value load_config['schema'] = { 'fields': [ {'name':'string_f', 'type':'STRING'}, {'name':'boolean_f', 'type':'BOOLEAN'}, {'name':'integer_f', 'type':'INTEGER', 'mode':'REQUIRED'}, {'name':'float_f', 'type':'FLOAT'}, {'name':'timestamp_f', 'type':'TIMESTAMP'} ] } load_config['sourceUris'] = [ 'gs://bigquery-e2e/chapters/06/sample_bad.csv', ] # End of job configuration. run_load.start_and_wait(service.jobs(), auth.PROJECT_ID, load_config)
def main(): service = auth.build_bq_client() # Load configuration with the destination specified. load_config = { 'destinationTable': { 'projectId': auth.PROJECT_ID, 'datasetId': 'ch06', # You can update this for each example. 'tableId': 'example_resumable' } } # Setup the job here. # load[property] = value load_config['schema'] = { 'fields': [ {'name':'string_f', 'type':'STRING'}, {'name':'boolean_f', 'type':'BOOLEAN'}, {'name':'integer_f', 'type':'INTEGER'}, {'name':'float_f', 'type':'FLOAT'}, {'name':'timestamp_f', 'type':'TIMESTAMP'} ] } upload = MediaFileUpload('sample.csv', mimetype='application/octet-stream', # This enables resumable uploads. resumable=True) # End of job configuration. run_load.start_and_wait(service.jobs(), auth.PROJECT_ID, load_config, media_body=upload)
def main(): service = auth.build_bq_client() # Load configuration with the destination specified. load_config = { 'destinationTable': { 'projectId': auth.PROJECT_ID, 'datasetId': 'ch06', # You can update this for each example. 'tableId': 'example_resumable' } } # Setup the job here. # load[property] = value load_config['schema'] = { 'fields': [{ 'name': 'string_f', 'type': 'STRING' }, { 'name': 'boolean_f', 'type': 'BOOLEAN' }, { 'name': 'integer_f', 'type': 'INTEGER' }, { 'name': 'float_f', 'type': 'FLOAT' }, { 'name': 'timestamp_f', 'type': 'TIMESTAMP' }] } upload = MediaFileUpload( 'sample.csv', mimetype='application/octet-stream', # This enables resumable uploads. resumable=True) # End of job configuration. run_load.start_and_wait(service.jobs(), auth.PROJECT_ID, load_config, media_body=upload)
def main(): service = auth.build_bq_client() # Load configuration with the destination specified. load_config = { 'destinationTable': { 'projectId': 'publicdata', 'datasetId': 'samples', 'tableId': 'mypersonaltable' } } # Setup the job here. # load[property] = value load_config['sourceUris'] = [ 'gs://bigquery-e2e/chapters/06/sample.csv', ] # End of job configuration. run_load.start_and_wait(service.jobs(), auth.PROJECT_ID, load_config)