Imported Upstream version 1.27.0
[platform/upstream/grpc.git] / tools / gcp / utils / big_query_utils.py
index 168b48d..6b61756 100755 (executable)
@@ -33,8 +33,10 @@ def create_big_query():
     """Authenticates with cloud platform and gets a BiqQuery service object
   """
     creds = GoogleCredentials.get_application_default()
-    return discovery.build(
-        'bigquery', 'v2', credentials=creds, cache_discovery=False)
+    return discovery.build('bigquery',
+                           'v2',
+                           credentials=creds,
+                           cache_discovery=False)
 
 
 def create_dataset(biq_query, project_id, dataset_id):
@@ -47,16 +49,16 @@ def create_dataset(biq_query, project_id, dataset_id):
     }
 
     try:
-        dataset_req = biq_query.datasets().insert(
-            projectId=project_id, body=body)
+        dataset_req = biq_query.datasets().insert(projectId=project_id,
+                                                  body=body)
         dataset_req.execute(num_retries=NUM_RETRIES)
     except HttpError as http_error:
         if http_error.resp.status == 409:
             print('Warning: The dataset %s already exists' % dataset_id)
         else:
             # Note: For more debugging info, print "http_error.content"
-            print('Error in creating dataset: %s. Err: %s' % (dataset_id,
-                                                              http_error))
+            print('Error in creating dataset: %s. Err: %s' %
+                  (dataset_id, http_error))
             is_success = False
     return is_success
 
@@ -121,16 +123,17 @@ def create_table2(big_query,
         }
 
     try:
-        table_req = big_query.tables().insert(
-            projectId=project_id, datasetId=dataset_id, body=body)
+        table_req = big_query.tables().insert(projectId=project_id,
+                                              datasetId=dataset_id,
+                                              body=body)
         res = table_req.execute(num_retries=NUM_RETRIES)
         print('Successfully created %s "%s"' % (res['kind'], res['id']))
     except HttpError as http_error:
         if http_error.resp.status == 409:
             print('Warning: Table %s already exists' % table_id)
         else:
-            print('Error in creating table: %s. Err: %s' % (table_id,
-                                                            http_error))
+            print('Error in creating table: %s. Err: %s' %
+                  (table_id, http_error))
             is_success = False
     return is_success
 
@@ -150,11 +153,10 @@ def patch_table(big_query, project_id, dataset_id, table_id, fields_schema):
     }
 
     try:
-        table_req = big_query.tables().patch(
-            projectId=project_id,
-            datasetId=dataset_id,
-            tableId=table_id,
-            body=body)
+        table_req = big_query.tables().patch(projectId=project_id,
+                                             datasetId=dataset_id,
+                                             tableId=table_id,
+                                             body=body)
         res = table_req.execute(num_retries=NUM_RETRIES)
         print('Successfully patched %s "%s"' % (res['kind'], res['id']))
     except HttpError as http_error:
@@ -167,11 +169,10 @@ def insert_rows(big_query, project_id, dataset_id, table_id, rows_list):
     is_success = True
     body = {'rows': rows_list}
     try:
-        insert_req = big_query.tabledata().insertAll(
-            projectId=project_id,
-            datasetId=dataset_id,
-            tableId=table_id,
-            body=body)
+        insert_req = big_query.tabledata().insertAll(projectId=project_id,
+                                                     datasetId=dataset_id,
+                                                     tableId=table_id,
+                                                     body=body)
         res = insert_req.execute(num_retries=NUM_RETRIES)
         if res.get('insertErrors', None):
             print('Error inserting rows! Response: %s' % res)