#!/usr/bin/env python """CLI for bigquery, version v2.""" # NOTE: This file is autogenerated and should not be edited by hand. import code import os import platform import sys from apitools.base.protorpclite import message_types from apitools.base.protorpclite import messages from google.apputils import appcommands import gflags as flags import apitools.base.py as apitools_base from apitools.base.py import cli as apitools_base_cli import bigquery_v2_client as client_lib import bigquery_v2_messages as messages def _DeclareBigqueryFlags(): """Declare global flags in an idempotent way.""" if 'api_endpoint' in flags.FLAGS: return flags.DEFINE_string( 'api_endpoint', u'https://www.googleapis.com/bigquery/v2/', 'URL of the API endpoint to use.', short_name='bigquery_url') flags.DEFINE_string( 'history_file', u'~/.bigquery.v2.history', 'File with interactive shell history.') flags.DEFINE_multistring( 'add_header', [], 'Additional http headers (as key=value strings). ' 'Can be specified multiple times.') flags.DEFINE_string( 'service_account_json_keyfile', '', 'Filename for a JSON service account key downloaded' ' from the Developer Console.') flags.DEFINE_enum( 'alt', u'json', [u'json'], u'Data format for the response.') flags.DEFINE_string( 'fields', None, u'Selector specifying which fields to include in a partial response.') flags.DEFINE_string( 'key', None, u'API key. Your API key identifies your project and provides you with ' u'API access, quota, and reports. Required unless you provide an OAuth ' u'2.0 token.') flags.DEFINE_string( 'oauth_token', None, u'OAuth 2.0 token for the current user.') flags.DEFINE_boolean( 'prettyPrint', 'True', u'Returns response with indentations and line breaks.') flags.DEFINE_string( 'quotaUser', None, u'Available to use for quota purposes for server-side applications. Can' u' be any arbitrary string assigned to a user, but should not exceed 40' u' characters. Overrides userIp if both are provided.') flags.DEFINE_string( 'trace', None, 'A tracing token of the form "token:" to include in api ' 'requests.') flags.DEFINE_string( 'userIp', None, u'IP address of the site where the request originates. Use this if you ' u'want to enforce per-user limits.') FLAGS = flags.FLAGS apitools_base_cli.DeclareBaseFlags() _DeclareBigqueryFlags() def GetGlobalParamsFromFlags(): """Return a StandardQueryParameters based on flags.""" result = messages.StandardQueryParameters() if FLAGS['alt'].present: result.alt = messages.StandardQueryParameters.AltValueValuesEnum(FLAGS.alt) if FLAGS['fields'].present: result.fields = FLAGS.fields.decode('utf8') if FLAGS['key'].present: result.key = FLAGS.key.decode('utf8') if FLAGS['oauth_token'].present: result.oauth_token = FLAGS.oauth_token.decode('utf8') if FLAGS['prettyPrint'].present: result.prettyPrint = FLAGS.prettyPrint if FLAGS['quotaUser'].present: result.quotaUser = FLAGS.quotaUser.decode('utf8') if FLAGS['trace'].present: result.trace = FLAGS.trace.decode('utf8') if FLAGS['userIp'].present: result.userIp = FLAGS.userIp.decode('utf8') return result def GetClientFromFlags(): """Return a client object, configured from flags.""" log_request = FLAGS.log_request or FLAGS.log_request_response log_response = FLAGS.log_response or FLAGS.log_request_response api_endpoint = apitools_base.NormalizeApiEndpoint(FLAGS.api_endpoint) additional_http_headers = dict(x.split('=', 1) for x in FLAGS.add_header) credentials_args = { 'service_account_json_keyfile': os.path.expanduser(FLAGS.service_account_json_keyfile) } try: client = client_lib.BigqueryV2( api_endpoint, log_request=log_request, log_response=log_response, credentials_args=credentials_args, additional_http_headers=additional_http_headers) except apitools_base.CredentialsError as e: print 'Error creating credentials: %s' % e sys.exit(1) return client class PyShell(appcommands.Cmd): def Run(self, _): """Run an interactive python shell with the client.""" client = GetClientFromFlags() params = GetGlobalParamsFromFlags() for field in params.all_fields(): value = params.get_assigned_value(field.name) if value != field.default: client.AddGlobalParam(field.name, value) banner = """ == bigquery interactive console == client: a bigquery client apitools_base: base apitools module messages: the generated messages module """ local_vars = { 'apitools_base': apitools_base, 'client': client, 'client_lib': client_lib, 'messages': messages, } if platform.system() == 'Linux': console = apitools_base_cli.ConsoleWithReadline( local_vars, histfile=FLAGS.history_file) else: console = code.InteractiveConsole(local_vars) try: console.interact(banner) except SystemExit as e: return e.code class DatasetsDelete(apitools_base_cli.NewCmd): """Command wrapping datasets.Delete.""" usage = """datasets_delete """ def __init__(self, name, fv): super(DatasetsDelete, self).__init__(name, fv) flags.DEFINE_boolean( 'deleteContents', None, u'If True, delete all the tables in the dataset. If False and the ' u'dataset contains tables, the request will fail. Default is False', flag_values=fv) def RunWithArgs(self, projectId, datasetId): """Deletes the dataset specified by the datasetId value. Before you can delete a dataset, you must delete all its tables, either manually or by specifying deleteContents. Immediately after deletion, you can create another dataset with the same name. Args: projectId: Project ID of the dataset being deleted datasetId: Dataset ID of dataset being deleted Flags: deleteContents: If True, delete all the tables in the dataset. If False and the dataset contains tables, the request will fail. Default is False """ client = GetClientFromFlags() global_params = GetGlobalParamsFromFlags() request = messages.BigqueryDatasetsDeleteRequest( projectId=projectId.decode('utf8'), datasetId=datasetId.decode('utf8'), ) if FLAGS['deleteContents'].present: request.deleteContents = FLAGS.deleteContents result = client.datasets.Delete( request, global_params=global_params) print apitools_base_cli.FormatOutput(result) class DatasetsGet(apitools_base_cli.NewCmd): """Command wrapping datasets.Get.""" usage = """datasets_get """ def __init__(self, name, fv): super(DatasetsGet, self).__init__(name, fv) def RunWithArgs(self, projectId, datasetId): """Returns the dataset specified by datasetID. Args: projectId: Project ID of the requested dataset datasetId: Dataset ID of the requested dataset """ client = GetClientFromFlags() global_params = GetGlobalParamsFromFlags() request = messages.BigqueryDatasetsGetRequest( projectId=projectId.decode('utf8'), datasetId=datasetId.decode('utf8'), ) result = client.datasets.Get( request, global_params=global_params) print apitools_base_cli.FormatOutput(result) class DatasetsInsert(apitools_base_cli.NewCmd): """Command wrapping datasets.Insert.""" usage = """datasets_insert """ def __init__(self, name, fv): super(DatasetsInsert, self).__init__(name, fv) flags.DEFINE_string( 'dataset', None, u'A Dataset resource to be passed as the request body.', flag_values=fv) def RunWithArgs(self, projectId): """Creates a new empty dataset. Args: projectId: Project ID of the new dataset Flags: dataset: A Dataset resource to be passed as the request body. """ client = GetClientFromFlags() global_params = GetGlobalParamsFromFlags() request = messages.BigqueryDatasetsInsertRequest( projectId=projectId.decode('utf8'), ) if FLAGS['dataset'].present: request.dataset = apitools_base.JsonToMessage(messages.Dataset, FLAGS.dataset) result = client.datasets.Insert( request, global_params=global_params) print apitools_base_cli.FormatOutput(result) class DatasetsList(apitools_base_cli.NewCmd): """Command wrapping datasets.List.""" usage = """datasets_list """ def __init__(self, name, fv): super(DatasetsList, self).__init__(name, fv) flags.DEFINE_boolean( 'all', None, u'Whether to list all datasets, including hidden ones', flag_values=fv) flags.DEFINE_string( 'filter', None, u'An expression for filtering the results of the request by label. ' u'The syntax is "labels.[:]". Multiple filters can be ANDed together ' u'by connecting with a space. Example: "labels.department:receiving ' u'labels.active". See https://cloud.google.com/bigquery/docs' u'/labeling-datasets#filtering_datasets_using_labels for details.', flag_values=fv) flags.DEFINE_integer( 'maxResults', None, u'The maximum number of results to return', flag_values=fv) flags.DEFINE_string( 'pageToken', None, u'Page token, returned by a previous call, to request the next page ' u'of results', flag_values=fv) def RunWithArgs(self, projectId): """Lists all datasets in the specified project to which you have been granted the READER dataset role. Args: projectId: Project ID of the datasets to be listed Flags: all: Whether to list all datasets, including hidden ones filter: An expression for filtering the results of the request by label. The syntax is "labels.[:]". Multiple filters can be ANDed together by connecting with a space. Example: "labels.department:receiving labels.active". See https://cloud.google.com/bigquery/docs/labeling- datasets#filtering_datasets_using_labels for details. maxResults: The maximum number of results to return pageToken: Page token, returned by a previous call, to request the next page of results """ client = GetClientFromFlags() global_params = GetGlobalParamsFromFlags() request = messages.BigqueryDatasetsListRequest( projectId=projectId.decode('utf8'), ) if FLAGS['all'].present: request.all = FLAGS.all if FLAGS['filter'].present: request.filter = FLAGS.filter.decode('utf8') if FLAGS['maxResults'].present: request.maxResults = FLAGS.maxResults if FLAGS['pageToken'].present: request.pageToken = FLAGS.pageToken.decode('utf8') result = client.datasets.List( request, global_params=global_params) print apitools_base_cli.FormatOutput(result) class DatasetsPatch(apitools_base_cli.NewCmd): """Command wrapping datasets.Patch.""" usage = """datasets_patch """ def __init__(self, name, fv): super(DatasetsPatch, self).__init__(name, fv) flags.DEFINE_string( 'dataset', None, u'A Dataset resource to be passed as the request body.', flag_values=fv) def RunWithArgs(self, projectId, datasetId): """Updates information in an existing dataset. The update method replaces the entire dataset resource, whereas the patch method only replaces fields that are provided in the submitted dataset resource. This method supports patch semantics. Args: projectId: Project ID of the dataset being updated datasetId: Dataset ID of the dataset being updated Flags: dataset: A Dataset resource to be passed as the request body. """ client = GetClientFromFlags() global_params = GetGlobalParamsFromFlags() request = messages.BigqueryDatasetsPatchRequest( projectId=projectId.decode('utf8'), datasetId=datasetId.decode('utf8'), ) if FLAGS['dataset'].present: request.dataset = apitools_base.JsonToMessage(messages.Dataset, FLAGS.dataset) result = client.datasets.Patch( request, global_params=global_params) print apitools_base_cli.FormatOutput(result) class DatasetsUpdate(apitools_base_cli.NewCmd): """Command wrapping datasets.Update.""" usage = """datasets_update """ def __init__(self, name, fv): super(DatasetsUpdate, self).__init__(name, fv) flags.DEFINE_string( 'dataset', None, u'A Dataset resource to be passed as the request body.', flag_values=fv) def RunWithArgs(self, projectId, datasetId): """Updates information in an existing dataset. The update method replaces the entire dataset resource, whereas the patch method only replaces fields that are provided in the submitted dataset resource. Args: projectId: Project ID of the dataset being updated datasetId: Dataset ID of the dataset being updated Flags: dataset: A Dataset resource to be passed as the request body. """ client = GetClientFromFlags() global_params = GetGlobalParamsFromFlags() request = messages.BigqueryDatasetsUpdateRequest( projectId=projectId.decode('utf8'), datasetId=datasetId.decode('utf8'), ) if FLAGS['dataset'].present: request.dataset = apitools_base.JsonToMessage(messages.Dataset, FLAGS.dataset) result = client.datasets.Update( request, global_params=global_params) print apitools_base_cli.FormatOutput(result) class JobsCancel(apitools_base_cli.NewCmd): """Command wrapping jobs.Cancel.""" usage = """jobs_cancel """ def __init__(self, name, fv): super(JobsCancel, self).__init__(name, fv) def RunWithArgs(self, projectId, jobId): """Requests that a job be cancelled. This call will return immediately, and the client will need to poll for the job status to see if the cancel completed successfully. Cancelled jobs may still incur costs. Args: projectId: [Required] Project ID of the job to cancel jobId: [Required] Job ID of the job to cancel """ client = GetClientFromFlags() global_params = GetGlobalParamsFromFlags() request = messages.BigqueryJobsCancelRequest( projectId=projectId.decode('utf8'), jobId=jobId.decode('utf8'), ) result = client.jobs.Cancel( request, global_params=global_params) print apitools_base_cli.FormatOutput(result) class JobsGet(apitools_base_cli.NewCmd): """Command wrapping jobs.Get.""" usage = """jobs_get """ def __init__(self, name, fv): super(JobsGet, self).__init__(name, fv) def RunWithArgs(self, projectId, jobId): """Returns information about a specific job. Job information is available for a six month period after creation. Requires that you're the person who ran the job, or have the Is Owner project role. Args: projectId: [Required] Project ID of the requested job jobId: [Required] Job ID of the requested job """ client = GetClientFromFlags() global_params = GetGlobalParamsFromFlags() request = messages.BigqueryJobsGetRequest( projectId=projectId.decode('utf8'), jobId=jobId.decode('utf8'), ) result = client.jobs.Get( request, global_params=global_params) print apitools_base_cli.FormatOutput(result) class JobsGetQueryResults(apitools_base_cli.NewCmd): """Command wrapping jobs.GetQueryResults.""" usage = """jobs_getQueryResults """ def __init__(self, name, fv): super(JobsGetQueryResults, self).__init__(name, fv) flags.DEFINE_integer( 'maxResults', None, u'Maximum number of results to read', flag_values=fv) flags.DEFINE_string( 'pageToken', None, u'Page token, returned by a previous call, to request the next page ' u'of results', flag_values=fv) flags.DEFINE_string( 'startIndex', None, u'Zero-based index of the starting row', flag_values=fv) flags.DEFINE_integer( 'timeoutMs', None, u'How long to wait for the query to complete, in milliseconds, before' u' returning. Default is 10 seconds. If the timeout passes before the' u" job completes, the 'jobComplete' field in the response will be " u'false', flag_values=fv) def RunWithArgs(self, projectId, jobId): """Retrieves the results of a query job. Args: projectId: [Required] Project ID of the query job jobId: [Required] Job ID of the query job Flags: maxResults: Maximum number of results to read pageToken: Page token, returned by a previous call, to request the next page of results startIndex: Zero-based index of the starting row timeoutMs: How long to wait for the query to complete, in milliseconds, before returning. Default is 10 seconds. If the timeout passes before the job completes, the 'jobComplete' field in the response will be false """ client = GetClientFromFlags() global_params = GetGlobalParamsFromFlags() request = messages.BigqueryJobsGetQueryResultsRequest( projectId=projectId.decode('utf8'), jobId=jobId.decode('utf8'), ) if FLAGS['maxResults'].present: request.maxResults = FLAGS.maxResults if FLAGS['pageToken'].present: request.pageToken = FLAGS.pageToken.decode('utf8') if FLAGS['startIndex'].present: request.startIndex = int(FLAGS.startIndex) if FLAGS['timeoutMs'].present: request.timeoutMs = FLAGS.timeoutMs result = client.jobs.GetQueryResults( request, global_params=global_params) print apitools_base_cli.FormatOutput(result) class JobsInsert(apitools_base_cli.NewCmd): """Command wrapping jobs.Insert.""" usage = """jobs_insert """ def __init__(self, name, fv): super(JobsInsert, self).__init__(name, fv) flags.DEFINE_string( 'job', None, u'A Job resource to be passed as the request body.', flag_values=fv) flags.DEFINE_string( 'upload_filename', '', 'Filename to use for upload.', flag_values=fv) flags.DEFINE_string( 'upload_mime_type', '', 'MIME type to use for the upload. Only needed if the extension on ' '--upload_filename does not determine the correct (or any) MIME ' 'type.', flag_values=fv) def RunWithArgs(self, projectId): """Starts a new asynchronous job. Requires the Can View project role. Args: projectId: Project ID of the project that will be billed for the job Flags: job: A Job resource to be passed as the request body. upload_filename: Filename to use for upload. upload_mime_type: MIME type to use for the upload. Only needed if the extension on --upload_filename does not determine the correct (or any) MIME type. """ client = GetClientFromFlags() global_params = GetGlobalParamsFromFlags() request = messages.BigqueryJobsInsertRequest( projectId=projectId.decode('utf8'), ) if FLAGS['job'].present: request.job = apitools_base.JsonToMessage(messages.Job, FLAGS.job) upload = None if FLAGS.upload_filename: upload = apitools_base.Upload.FromFile( FLAGS.upload_filename, FLAGS.upload_mime_type, progress_callback=apitools_base.UploadProgressPrinter, finish_callback=apitools_base.UploadCompletePrinter) result = client.jobs.Insert( request, global_params=global_params, upload=upload) print apitools_base_cli.FormatOutput(result) class JobsList(apitools_base_cli.NewCmd): """Command wrapping jobs.List.""" usage = """jobs_list """ def __init__(self, name, fv): super(JobsList, self).__init__(name, fv) flags.DEFINE_boolean( 'allUsers', None, u'Whether to display jobs owned by all users in the project. Default ' u'false', flag_values=fv) flags.DEFINE_integer( 'maxResults', None, u'Maximum number of results to return', flag_values=fv) flags.DEFINE_string( 'pageToken', None, u'Page token, returned by a previous call, to request the next page ' u'of results', flag_values=fv) flags.DEFINE_enum( 'projection', u'full', [u'full', u'minimal'], u'Restrict information returned to a set of selected fields', flag_values=fv) flags.DEFINE_enum( 'stateFilter', u'done', [u'done', u'pending', u'running'], u'Filter for job state', flag_values=fv) def RunWithArgs(self, projectId): """Lists all jobs that you started in the specified project. Job information is available for a six month period after creation. The job list is sorted in reverse chronological order, by job creation time. Requires the Can View project role, or the Is Owner project role if you set the allUsers property. Args: projectId: Project ID of the jobs to list Flags: allUsers: Whether to display jobs owned by all users in the project. Default false maxResults: Maximum number of results to return pageToken: Page token, returned by a previous call, to request the next page of results projection: Restrict information returned to a set of selected fields stateFilter: Filter for job state """ client = GetClientFromFlags() global_params = GetGlobalParamsFromFlags() request = messages.BigqueryJobsListRequest( projectId=projectId.decode('utf8'), ) if FLAGS['allUsers'].present: request.allUsers = FLAGS.allUsers if FLAGS['maxResults'].present: request.maxResults = FLAGS.maxResults if FLAGS['pageToken'].present: request.pageToken = FLAGS.pageToken.decode('utf8') if FLAGS['projection'].present: request.projection = messages.BigqueryJobsListRequest.ProjectionValueValuesEnum(FLAGS.projection) if FLAGS['stateFilter'].present: request.stateFilter = [messages.BigqueryJobsListRequest.StateFilterValueValuesEnum(x) for x in FLAGS.stateFilter] result = client.jobs.List( request, global_params=global_params) print apitools_base_cli.FormatOutput(result) class JobsQuery(apitools_base_cli.NewCmd): """Command wrapping jobs.Query.""" usage = """jobs_query """ def __init__(self, name, fv): super(JobsQuery, self).__init__(name, fv) flags.DEFINE_string( 'queryRequest', None, u'A QueryRequest resource to be passed as the request body.', flag_values=fv) def RunWithArgs(self, projectId): """Runs a BigQuery SQL query synchronously and returns query results if the query completes within a specified timeout. Args: projectId: Project ID of the project billed for the query Flags: queryRequest: A QueryRequest resource to be passed as the request body. """ client = GetClientFromFlags() global_params = GetGlobalParamsFromFlags() request = messages.BigqueryJobsQueryRequest( projectId=projectId.decode('utf8'), ) if FLAGS['queryRequest'].present: request.queryRequest = apitools_base.JsonToMessage(messages.QueryRequest, FLAGS.queryRequest) result = client.jobs.Query( request, global_params=global_params) print apitools_base_cli.FormatOutput(result) class ProjectsList(apitools_base_cli.NewCmd): """Command wrapping projects.List.""" usage = """projects_list""" def __init__(self, name, fv): super(ProjectsList, self).__init__(name, fv) flags.DEFINE_integer( 'maxResults', None, u'Maximum number of results to return', flag_values=fv) flags.DEFINE_string( 'pageToken', None, u'Page token, returned by a previous call, to request the next page ' u'of results', flag_values=fv) def RunWithArgs(self): """Lists all projects to which you have been granted any project role. Flags: maxResults: Maximum number of results to return pageToken: Page token, returned by a previous call, to request the next page of results """ client = GetClientFromFlags() global_params = GetGlobalParamsFromFlags() request = messages.BigqueryProjectsListRequest( ) if FLAGS['maxResults'].present: request.maxResults = FLAGS.maxResults if FLAGS['pageToken'].present: request.pageToken = FLAGS.pageToken.decode('utf8') result = client.projects.List( request, global_params=global_params) print apitools_base_cli.FormatOutput(result) class TabledataInsertAll(apitools_base_cli.NewCmd): """Command wrapping tabledata.InsertAll.""" usage = """tabledata_insertAll """ def __init__(self, name, fv): super(TabledataInsertAll, self).__init__(name, fv) flags.DEFINE_string( 'tableDataInsertAllRequest', None, u'A TableDataInsertAllRequest resource to be passed as the request ' u'body.', flag_values=fv) def RunWithArgs(self, projectId, datasetId, tableId): """Streams data into BigQuery one record at a time without needing to run a load job. Requires the WRITER dataset role. Args: projectId: Project ID of the destination table. datasetId: Dataset ID of the destination table. tableId: Table ID of the destination table. Flags: tableDataInsertAllRequest: A TableDataInsertAllRequest resource to be passed as the request body. """ client = GetClientFromFlags() global_params = GetGlobalParamsFromFlags() request = messages.BigqueryTabledataInsertAllRequest( projectId=projectId.decode('utf8'), datasetId=datasetId.decode('utf8'), tableId=tableId.decode('utf8'), ) if FLAGS['tableDataInsertAllRequest'].present: request.tableDataInsertAllRequest = apitools_base.JsonToMessage(messages.TableDataInsertAllRequest, FLAGS.tableDataInsertAllRequest) result = client.tabledata.InsertAll( request, global_params=global_params) print apitools_base_cli.FormatOutput(result) class TabledataList(apitools_base_cli.NewCmd): """Command wrapping tabledata.List.""" usage = """tabledata_list """ def __init__(self, name, fv): super(TabledataList, self).__init__(name, fv) flags.DEFINE_integer( 'maxResults', None, u'Maximum number of results to return', flag_values=fv) flags.DEFINE_string( 'pageToken', None, u'Page token, returned by a previous call, identifying the result set', flag_values=fv) flags.DEFINE_string( 'startIndex', None, u'Zero-based index of the starting row to read', flag_values=fv) def RunWithArgs(self, projectId, datasetId, tableId): """Retrieves table data from a specified set of rows. Requires the READER dataset role. Args: projectId: Project ID of the table to read datasetId: Dataset ID of the table to read tableId: Table ID of the table to read Flags: maxResults: Maximum number of results to return pageToken: Page token, returned by a previous call, identifying the result set startIndex: Zero-based index of the starting row to read """ client = GetClientFromFlags() global_params = GetGlobalParamsFromFlags() request = messages.BigqueryTabledataListRequest( projectId=projectId.decode('utf8'), datasetId=datasetId.decode('utf8'), tableId=tableId.decode('utf8'), ) if FLAGS['maxResults'].present: request.maxResults = FLAGS.maxResults if FLAGS['pageToken'].present: request.pageToken = FLAGS.pageToken.decode('utf8') if FLAGS['startIndex'].present: request.startIndex = int(FLAGS.startIndex) result = client.tabledata.List( request, global_params=global_params) print apitools_base_cli.FormatOutput(result) class TablesDelete(apitools_base_cli.NewCmd): """Command wrapping tables.Delete.""" usage = """tables_delete """ def __init__(self, name, fv): super(TablesDelete, self).__init__(name, fv) def RunWithArgs(self, projectId, datasetId, tableId): """Deletes the table specified by tableId from the dataset. If the table contains data, all the data will be deleted. Args: projectId: Project ID of the table to delete datasetId: Dataset ID of the table to delete tableId: Table ID of the table to delete """ client = GetClientFromFlags() global_params = GetGlobalParamsFromFlags() request = messages.BigqueryTablesDeleteRequest( projectId=projectId.decode('utf8'), datasetId=datasetId.decode('utf8'), tableId=tableId.decode('utf8'), ) result = client.tables.Delete( request, global_params=global_params) print apitools_base_cli.FormatOutput(result) class TablesGet(apitools_base_cli.NewCmd): """Command wrapping tables.Get.""" usage = """tables_get """ def __init__(self, name, fv): super(TablesGet, self).__init__(name, fv) def RunWithArgs(self, projectId, datasetId, tableId): """Gets the specified table resource by table ID. This method does not return the data in the table, it only returns the table resource, which describes the structure of this table. Args: projectId: Project ID of the requested table datasetId: Dataset ID of the requested table tableId: Table ID of the requested table """ client = GetClientFromFlags() global_params = GetGlobalParamsFromFlags() request = messages.BigqueryTablesGetRequest( projectId=projectId.decode('utf8'), datasetId=datasetId.decode('utf8'), tableId=tableId.decode('utf8'), ) result = client.tables.Get( request, global_params=global_params) print apitools_base_cli.FormatOutput(result) class TablesInsert(apitools_base_cli.NewCmd): """Command wrapping tables.Insert.""" usage = """tables_insert """ def __init__(self, name, fv): super(TablesInsert, self).__init__(name, fv) flags.DEFINE_string( 'table', None, u'A Table resource to be passed as the request body.', flag_values=fv) def RunWithArgs(self, projectId, datasetId): """Creates a new, empty table in the dataset. Args: projectId: Project ID of the new table datasetId: Dataset ID of the new table Flags: table: A Table resource to be passed as the request body. """ client = GetClientFromFlags() global_params = GetGlobalParamsFromFlags() request = messages.BigqueryTablesInsertRequest( projectId=projectId.decode('utf8'), datasetId=datasetId.decode('utf8'), ) if FLAGS['table'].present: request.table = apitools_base.JsonToMessage(messages.Table, FLAGS.table) result = client.tables.Insert( request, global_params=global_params) print apitools_base_cli.FormatOutput(result) class TablesList(apitools_base_cli.NewCmd): """Command wrapping tables.List.""" usage = """tables_list """ def __init__(self, name, fv): super(TablesList, self).__init__(name, fv) flags.DEFINE_integer( 'maxResults', None, u'Maximum number of results to return', flag_values=fv) flags.DEFINE_string( 'pageToken', None, u'Page token, returned by a previous call, to request the next page ' u'of results', flag_values=fv) def RunWithArgs(self, projectId, datasetId): """Lists all tables in the specified dataset. Requires the READER dataset role. Args: projectId: Project ID of the tables to list datasetId: Dataset ID of the tables to list Flags: maxResults: Maximum number of results to return pageToken: Page token, returned by a previous call, to request the next page of results """ client = GetClientFromFlags() global_params = GetGlobalParamsFromFlags() request = messages.BigqueryTablesListRequest( projectId=projectId.decode('utf8'), datasetId=datasetId.decode('utf8'), ) if FLAGS['maxResults'].present: request.maxResults = FLAGS.maxResults if FLAGS['pageToken'].present: request.pageToken = FLAGS.pageToken.decode('utf8') result = client.tables.List( request, global_params=global_params) print apitools_base_cli.FormatOutput(result) class TablesPatch(apitools_base_cli.NewCmd): """Command wrapping tables.Patch.""" usage = """tables_patch """ def __init__(self, name, fv): super(TablesPatch, self).__init__(name, fv) flags.DEFINE_string( 'table', None, u'A Table resource to be passed as the request body.', flag_values=fv) def RunWithArgs(self, projectId, datasetId, tableId): """Updates information in an existing table. The update method replaces the entire table resource, whereas the patch method only replaces fields that are provided in the submitted table resource. This method supports patch semantics. Args: projectId: Project ID of the table to update datasetId: Dataset ID of the table to update tableId: Table ID of the table to update Flags: table: A Table resource to be passed as the request body. """ client = GetClientFromFlags() global_params = GetGlobalParamsFromFlags() request = messages.BigqueryTablesPatchRequest( projectId=projectId.decode('utf8'), datasetId=datasetId.decode('utf8'), tableId=tableId.decode('utf8'), ) if FLAGS['table'].present: request.table = apitools_base.JsonToMessage(messages.Table, FLAGS.table) result = client.tables.Patch( request, global_params=global_params) print apitools_base_cli.FormatOutput(result) class TablesUpdate(apitools_base_cli.NewCmd): """Command wrapping tables.Update.""" usage = """tables_update """ def __init__(self, name, fv): super(TablesUpdate, self).__init__(name, fv) flags.DEFINE_string( 'table', None, u'A Table resource to be passed as the request body.', flag_values=fv) def RunWithArgs(self, projectId, datasetId, tableId): """Updates information in an existing table. The update method replaces the entire table resource, whereas the patch method only replaces fields that are provided in the submitted table resource. Args: projectId: Project ID of the table to update datasetId: Dataset ID of the table to update tableId: Table ID of the table to update Flags: table: A Table resource to be passed as the request body. """ client = GetClientFromFlags() global_params = GetGlobalParamsFromFlags() request = messages.BigqueryTablesUpdateRequest( projectId=projectId.decode('utf8'), datasetId=datasetId.decode('utf8'), tableId=tableId.decode('utf8'), ) if FLAGS['table'].present: request.table = apitools_base.JsonToMessage(messages.Table, FLAGS.table) result = client.tables.Update( request, global_params=global_params) print apitools_base_cli.FormatOutput(result) def main(_): appcommands.AddCmd('pyshell', PyShell) appcommands.AddCmd('datasets_delete', DatasetsDelete) appcommands.AddCmd('datasets_get', DatasetsGet) appcommands.AddCmd('datasets_insert', DatasetsInsert) appcommands.AddCmd('datasets_list', DatasetsList) appcommands.AddCmd('datasets_patch', DatasetsPatch) appcommands.AddCmd('datasets_update', DatasetsUpdate) appcommands.AddCmd('jobs_cancel', JobsCancel) appcommands.AddCmd('jobs_get', JobsGet) appcommands.AddCmd('jobs_getQueryResults', JobsGetQueryResults) appcommands.AddCmd('jobs_insert', JobsInsert) appcommands.AddCmd('jobs_list', JobsList) appcommands.AddCmd('jobs_query', JobsQuery) appcommands.AddCmd('projects_list', ProjectsList) appcommands.AddCmd('tabledata_insertAll', TabledataInsertAll) appcommands.AddCmd('tabledata_list', TabledataList) appcommands.AddCmd('tables_delete', TablesDelete) appcommands.AddCmd('tables_get', TablesGet) appcommands.AddCmd('tables_insert', TablesInsert) appcommands.AddCmd('tables_list', TablesList) appcommands.AddCmd('tables_patch', TablesPatch) appcommands.AddCmd('tables_update', TablesUpdate) apitools_base_cli.SetupLogger() if hasattr(appcommands, 'SetDefaultCommand'): appcommands.SetDefaultCommand('pyshell') run_main = apitools_base_cli.run_main if __name__ == '__main__': appcommands.Run()