Cloud Machine Learning Engine . projects . jobs

Instance Methods

cancel(name, body=None, x__xgafv=None)

Cancels a running job.

create(parent, body, x__xgafv=None)

Creates a training or a batch prediction job.

get(name, x__xgafv=None)

Describes a job.

getIamPolicy(resource, x__xgafv=None)

Gets the access control policy for a resource.

list(parent, pageToken=None, x__xgafv=None, pageSize=None, filter=None)

Lists the jobs in the project.

list_next(previous_request, previous_response)

Retrieves the next page of results.

patch(name, body, updateMask=None, x__xgafv=None)

Updates a specific job resource.

setIamPolicy(resource, body, x__xgafv=None)

Sets the access control policy on the specified resource. Replaces any

testIamPermissions(resource, body, x__xgafv=None)

Returns permissions that a caller has on the specified resource.

Method Details

cancel(name, body=None, x__xgafv=None)
Cancels a running job.

Args:
  name: string, Required. The name of the job to cancel. (required)
  body: object, The request body.
    The object takes the form of:

{ # Request message for the CancelJob method.
  }

  x__xgafv: string, V1 error format.
    Allowed values
      1 - v1 error format
      2 - v2 error format

Returns:
  An object of the form:

    { # A generic empty message that you can re-use to avoid defining duplicated
      # empty messages in your APIs. A typical example is to use it as the request
      # or the response type of an API method. For instance:
      #
      #     service Foo {
      #       rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty);
      #     }
      #
      # The JSON representation for `Empty` is empty JSON object `{}`.
  }
create(parent, body, x__xgafv=None)
Creates a training or a batch prediction job.

Args:
  parent: string, Required. The project name. (required)
  body: object, The request body. (required)
    The object takes the form of:

{ # Represents a training or prediction job.
  "errorMessage": "A String", # Output only. The details of a failure or a cancellation.
  "trainingOutput": { # Represents results of a training job. Output only. # The current training job result.
    "completedTrialCount": "A String", # The number of hyperparameter tuning trials that completed successfully.
        # Only set for hyperparameter tuning jobs.
    "trials": [ # Results for individual Hyperparameter trials.
        # Only set for hyperparameter tuning jobs.
      { # Represents the result of a single hyperparameter tuning trial from a
          # training job. The TrainingOutput object that is returned on successful
          # completion of a training job with hyperparameter tuning includes a list
          # of HyperparameterOutput objects, one for each successful trial.
        "hyperparameters": { # The hyperparameters given to this trial.
          "a_key": "A String",
        },
        "finalMetric": { # An observed value of a metric. # The final objective metric seen for this trial.
          "trainingStep": "A String", # The global training step for this metric.
          "objectiveValue": 3.14, # The objective value at this training step.
        },
        "allMetrics": [ # All recorded object metrics for this trial. This field is not currently
            # populated.
          { # An observed value of a metric.
            "trainingStep": "A String", # The global training step for this metric.
            "objectiveValue": 3.14, # The objective value at this training step.
          },
        ],
        "isTrialStoppedEarly": True or False, # True if the trial is stopped early.
        "trialId": "A String", # The trial id for these results.
        "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
            # Only set for trials of built-in algorithms jobs that have succeeded.
          "framework": "A String", # Framework on which the built-in algorithm was trained.
          "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
              # saves the trained model. Only set for successful jobs that don't use
              # hyperparameter tuning.
          "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
              # trained.
          "pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
        },
      },
    ],
    "isHyperparameterTuningJob": True or False, # Whether this job is a hyperparameter tuning job.
    "isBuiltInAlgorithmJob": True or False, # Whether this job is a built-in Algorithm job.
    "consumedMLUnits": 3.14, # The amount of ML units consumed by the job.
    "hyperparameterMetricTag": "A String", # The TensorFlow summary tag name used for optimizing hyperparameter tuning
        # trials. See
        # [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag)
        # for more information. Only set for hyperparameter tuning jobs.
    "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
        # Only set for built-in algorithms jobs.
      "framework": "A String", # Framework on which the built-in algorithm was trained.
      "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
          # saves the trained model. Only set for successful jobs that don't use
          # hyperparameter tuning.
      "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
          # trained.
      "pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
    },
  },
  "predictionInput": { # Represents input parameters for a prediction job. # Input parameters to create a prediction job.
    "modelName": "A String", # Use this field if you want to use the default version for the specified
        # model. The string must use the following format:
        #
        # `"projects/YOUR_PROJECT/models/YOUR_MODEL"`
    "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for this batch
        # prediction. If not set, AI Platform will pick the runtime version used
        # during the CreateVersion request for this model version, or choose the
        # latest stable version when model version information is not available
        # such as when the model is specified by uri.
    "signatureName": "A String", # Optional. The name of the signature defined in the SavedModel to use for
        # this job. Please refer to
        # [SavedModel](https://tensorflow.github.io/serving/serving_basic.html)
        # for information about how to use signatures.
        #
        # Defaults to
        # [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants)
        # , which is "serving_default".
    "batchSize": "A String", # Optional. Number of records per batch, defaults to 64.
        # The service will buffer batch_size number of records in memory before
        # invoking one Tensorflow prediction call internally. So take the record
        # size and memory available into consideration when setting this parameter.
    "inputPaths": [ # Required. The Cloud Storage location of the input data files. May contain
        # wildcards.
      "A String",
    ],
    "maxWorkerCount": "A String", # Optional. The maximum number of workers to be used for parallel processing.
        # Defaults to 10 if not specified.
    "uri": "A String", # Use this field if you want to specify a Google Cloud Storage path for
        # the model to use.
    "outputPath": "A String", # Required. The output Google Cloud Storage location.
    "dataFormat": "A String", # Required. The format of the input data files.
    "versionName": "A String", # Use this field if you want to specify a version of the model to use. The
        # string is formatted the same way as `model_version`, with the addition
        # of the version information:
        #
        # `"projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION"`
    "region": "A String", # Required. The Google Compute Engine region to run the prediction job in.
        # See the available regions
        # for AI Platform services.
    "outputDataFormat": "A String", # Optional. Format of the output data files, defaults to JSON.
  },
  "trainingInput": { # Represents input parameters for a training job. When using the # Input parameters to create a training job.
      # gcloud command to submit your training job, you can specify
      # the input parameters as command-line arguments and/or in a YAML configuration
      # file referenced from the --config command-line argument. For
      # details, see the guide to
      # submitting a training
      # job.
    "workerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
        # job's worker nodes.
        #
        # The supported values are the same as those described in the entry for
        # `masterType`.
        #
        # This value must be consistent with the category of machine type that
        # `masterType` uses. In other words, both must be AI Platform machine
        # types or both must be Compute Engine machine types.
        #
        # If you use `cloud_tpu` for this value, see special instructions for
        # [configuring a custom TPU
        # machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
        #
        # This value must be present when `scaleTier` is set to `CUSTOM` and
        # `workerCount` is greater than zero.
    "parameterServerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers.
        #
        # You should only set `parameterServerConfig.acceleratorConfig` if
        # `parameterServerConfigType` is set to a Compute Engine machine type. [Learn
        # about restrictions on accelerator configurations for
        # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
        #
        # Set `parameterServerConfig.imageUri` only if you build a custom image for
        # your parameter server. If `parameterServerConfig.imageUri` has not been
        # set, AI Platform uses the value of `masterConfig.imageUri`.
        # Learn more about [configuring custom
        # containers](/ml-engine/docs/distributed-training-containers).
      "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
          # [Learn about restrictions on accelerator configurations for
          # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
        "count": "A String", # The number of accelerators to attach to each machine running the job.
        "type": "A String", # The type of accelerator to use.
      },
      "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
          # Registry. Learn more about [configuring custom
          # containers](/ml-engine/docs/distributed-training-containers).
    },
    "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for training. If not
        # set, AI Platform uses the default stable version, 1.0. For more
        # information, see the
        # runtime version list
        # and
        # how to manage runtime versions.
    "scaleTier": "A String", # Required. Specifies the machine types, the number of replicas for workers
        # and parameter servers.
    "masterType": "A String", # Optional. Specifies the type of virtual machine to use for your training
        # job's master worker.
        #
        # The following types are supported:
        #
        # 
#
standard
#
# A basic machine configuration suitable for training simple models with # small to moderate datasets. #
#
large_model
#
# A machine with a lot of memory, specially suited for parameter servers # when your model is large (having many hidden layers or layers with very # large numbers of nodes). #
#
complex_model_s
#
# A machine suitable for the master and workers of the cluster when your # model requires more computation than the standard machine can handle # satisfactorily. #
#
complex_model_m
#
# A machine with roughly twice the number of cores and roughly double the # memory of complex_model_s. #
#
complex_model_l
#
# A machine with roughly twice the number of cores and roughly double the # memory of complex_model_m. #
#
standard_gpu
#
# A machine equivalent to standard that # also includes a single NVIDIA Tesla K80 GPU. See more about # using GPUs to # train your model. #
#
complex_model_m_gpu
#
# A machine equivalent to complex_model_m that also includes # four NVIDIA Tesla K80 GPUs. #
#
complex_model_l_gpu
#
# A machine equivalent to complex_model_l that also includes # eight NVIDIA Tesla K80 GPUs. #
#
standard_p100
#
# A machine equivalent to standard that # also includes a single NVIDIA Tesla P100 GPU. #
#
complex_model_m_p100
#
# A machine equivalent to complex_model_m that also includes # four NVIDIA Tesla P100 GPUs. #
#
standard_v100
#
# A machine equivalent to standard that # also includes a single NVIDIA Tesla V100 GPU. #
#
large_model_v100
#
# A machine equivalent to large_model that # also includes a single NVIDIA Tesla V100 GPU. #
#
complex_model_m_v100
#
# A machine equivalent to complex_model_m that # also includes four NVIDIA Tesla V100 GPUs. #
#
complex_model_l_v100
#
# A machine equivalent to complex_model_l that # also includes eight NVIDIA Tesla V100 GPUs. #
#
cloud_tpu
#
# A TPU VM including one Cloud TPU. See more about # using TPUs to train # your model. #
#
# # You may also use certain Compute Engine machine types directly in this # field. The following types are supported: # # - `n1-standard-4` # - `n1-standard-8` # - `n1-standard-16` # - `n1-standard-32` # - `n1-standard-64` # - `n1-standard-96` # - `n1-highmem-2` # - `n1-highmem-4` # - `n1-highmem-8` # - `n1-highmem-16` # - `n1-highmem-32` # - `n1-highmem-64` # - `n1-highmem-96` # - `n1-highcpu-16` # - `n1-highcpu-32` # - `n1-highcpu-64` # - `n1-highcpu-96` # # See more about [using Compute Engine machine # types](/ml-engine/docs/tensorflow/machine-types#compute-engine-machine-types). # # You must set this value when `scaleTier` is set to `CUSTOM`. "hyperparameters": { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune. "maxTrials": 42, # Optional. How many training trials should be attempted to optimize # the specified hyperparameters. # # Defaults to one. "goal": "A String", # Required. The type of goal to use for tuning. Available types are # `MAXIMIZE` and `MINIMIZE`. # # Defaults to `MAXIMIZE`. "algorithm": "A String", # Optional. The search algorithm specified for the hyperparameter # tuning job. # Uses the default AI Platform hyperparameter tuning # algorithm if unspecified. "maxFailedTrials": 42, # Optional. The number of failed trials that need to be seen before failing # the hyperparameter tuning job. You can specify this field to override the # default failing criteria for AI Platform hyperparameter tuning jobs. # # Defaults to zero, which means the service decides when a hyperparameter # job should fail. "enableTrialEarlyStopping": True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial # early stopping. "resumePreviousJobId": "A String", # Optional. The prior hyperparameter tuning job id that users hope to # continue with. The job id will be used to find the corresponding vizier # study guid and resume the study. "params": [ # Required. The set of parameters to tune. { # Represents a single hyperparameter to optimize. "maxValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field # should be unset if type is `CATEGORICAL`. This value should be integers if # type is `INTEGER`. "categoricalValues": [ # Required if type is `CATEGORICAL`. The list of possible categories. "A String", ], "discreteValues": [ # Required if type is `DISCRETE`. # A list of feasible points. # The list should be in strictly increasing order. For instance, this # parameter might have possible settings of 1.5, 2.5, and 4.0. This list # should not contain more than 1,000 values. 3.14, ], "parameterName": "A String", # Required. The parameter name must be unique amongst all ParameterConfigs in # a HyperparameterSpec message. E.g., "learning_rate". "minValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field # should be unset if type is `CATEGORICAL`. This value should be integers if # type is INTEGER. "type": "A String", # Required. The type of the parameter. "scaleType": "A String", # Optional. How the parameter should be scaled to the hypercube. # Leave unset for categorical parameters. # Some kind of scaling is strongly recommended for real or integral # parameters (e.g., `UNIT_LINEAR_SCALE`). }, ], "hyperparameterMetricTag": "A String", # Optional. The TensorFlow summary tag name to use for optimizing trials. For # current versions of TensorFlow, this tag name should exactly match what is # shown in TensorBoard, including all scopes. For versions of TensorFlow # prior to 0.12, this should be only the tag passed to tf.Summary. # By default, "training/hptuning/metric" will be used. "maxParallelTrials": 42, # Optional. The number of training trials to run concurrently. # You can reduce the time it takes to perform hyperparameter tuning by adding # trials in parallel. However, each trail only benefits from the information # gained in completed trials. That means that a trial does not get access to # the results of trials running at the same time, which could reduce the # quality of the overall optimization. # # Each trial will use the same scale tier and machine types. # # Defaults to one. }, "region": "A String", # Required. The Google Compute Engine region to run the training job in. # See the available regions # for AI Platform services. "args": [ # Optional. Command line arguments to pass to the program. "A String", ], "pythonModule": "A String", # Required. The Python module name to run after installing the packages. "pythonVersion": "A String", # Optional. The version of Python used in training. If not set, the default # version is '2.7'. Python '3.5' is available when `runtime_version` is set # to '1.4' and above. Python '2.7' works with all supported # runtime versions. "jobDir": "A String", # Optional. A Google Cloud Storage path in which to store training outputs # and other data needed for training. This path is passed to your TensorFlow # program as the '--job-dir' command-line argument. The benefit of specifying # this field is that Cloud ML validates the path for use in training. "packageUris": [ # Required. The Google Cloud Storage location of the packages with # the training program and any additional dependencies. # The maximum number of package URIs is 100. "A String", ], "workerCount": "A String", # Optional. The number of worker replicas to use for the training job. Each # replica in the cluster will be of the type specified in `worker_type`. # # This value can only be used when `scale_tier` is set to `CUSTOM`. If you # set this value, you must also set `worker_type`. # # The default value is zero. "parameterServerType": "A String", # Optional. Specifies the type of virtual machine to use for your training # job's parameter server. # # The supported values are the same as those described in the entry for # `master_type`. # # This value must be consistent with the category of machine type that # `masterType` uses. In other words, both must be AI Platform machine # types or both must be Compute Engine machine types. # # This value must be present when `scaleTier` is set to `CUSTOM` and # `parameter_server_count` is greater than zero. "workerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers. # # You should only set `workerConfig.acceleratorConfig` if `workerType` is set # to a Compute Engine machine type. [Learn about restrictions on accelerator # configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) # # Set `workerConfig.imageUri` only if you build a custom image for your # worker. If `workerConfig.imageUri` has not been set, AI Platform uses # the value of `masterConfig.imageUri`. Learn more about # [configuring custom # containers](/ml-engine/docs/distributed-training-containers). "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica. # [Learn about restrictions on accelerator configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) "count": "A String", # The number of accelerators to attach to each machine running the job. "type": "A String", # The type of accelerator to use. }, "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container # Registry. Learn more about [configuring custom # containers](/ml-engine/docs/distributed-training-containers). }, "maxRunningTime": "A String", # Optional. The maximum job running time. The default is 7 days. "masterConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker. # # You should only set `masterConfig.acceleratorConfig` if `masterType` is set # to a Compute Engine machine type. Learn about [restrictions on accelerator # configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) # # Set `masterConfig.imageUri` only if you build a custom image. Only one of # `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more about # [configuring custom # containers](/ml-engine/docs/distributed-training-containers). "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica. # [Learn about restrictions on accelerator configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) "count": "A String", # The number of accelerators to attach to each machine running the job. "type": "A String", # The type of accelerator to use. }, "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container # Registry. Learn more about [configuring custom # containers](/ml-engine/docs/distributed-training-containers). }, "parameterServerCount": "A String", # Optional. The number of parameter server replicas to use for the training # job. Each replica in the cluster will be of the type specified in # `parameter_server_type`. # # This value can only be used when `scale_tier` is set to `CUSTOM`.If you # set this value, you must also set `parameter_server_type`. # # The default value is zero. }, "jobId": "A String", # Required. The user-specified id of the job. "labels": { # Optional. One or more labels that you can add, to organize your jobs. # Each label is a key-value pair, where both the key and the value are # arbitrary strings that you supply. # For more information, see the documentation on # using labels. "a_key": "A String", }, "state": "A String", # Output only. The detailed state of a job. "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help # prevent simultaneous updates of a job from overwriting each other. # It is strongly suggested that systems make use of the `etag` in the # read-modify-write cycle to perform job updates in order to avoid race # conditions: An `etag` is returned in the response to `GetJob`, and # systems are expected to put that etag in the request to `UpdateJob` to # ensure that their change will be applied to the same version of the job. "startTime": "A String", # Output only. When the job processing was started. "endTime": "A String", # Output only. When the job processing was completed. "predictionOutput": { # Represents results of a prediction job. # The current prediction job result. "outputPath": "A String", # The output Google Cloud Storage location provided at the job creation time. "nodeHours": 3.14, # Node hours used by the batch prediction job. "predictionCount": "A String", # The number of generated predictions. "errorCount": "A String", # The number of data instances which resulted in errors. }, "createTime": "A String", # Output only. When the job was created. } x__xgafv: string, V1 error format. Allowed values 1 - v1 error format 2 - v2 error format Returns: An object of the form: { # Represents a training or prediction job. "errorMessage": "A String", # Output only. The details of a failure or a cancellation. "trainingOutput": { # Represents results of a training job. Output only. # The current training job result. "completedTrialCount": "A String", # The number of hyperparameter tuning trials that completed successfully. # Only set for hyperparameter tuning jobs. "trials": [ # Results for individual Hyperparameter trials. # Only set for hyperparameter tuning jobs. { # Represents the result of a single hyperparameter tuning trial from a # training job. The TrainingOutput object that is returned on successful # completion of a training job with hyperparameter tuning includes a list # of HyperparameterOutput objects, one for each successful trial. "hyperparameters": { # The hyperparameters given to this trial. "a_key": "A String", }, "finalMetric": { # An observed value of a metric. # The final objective metric seen for this trial. "trainingStep": "A String", # The global training step for this metric. "objectiveValue": 3.14, # The objective value at this training step. }, "allMetrics": [ # All recorded object metrics for this trial. This field is not currently # populated. { # An observed value of a metric. "trainingStep": "A String", # The global training step for this metric. "objectiveValue": 3.14, # The objective value at this training step. }, ], "isTrialStoppedEarly": True or False, # True if the trial is stopped early. "trialId": "A String", # The trial id for these results. "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs. # Only set for trials of built-in algorithms jobs that have succeeded. "framework": "A String", # Framework on which the built-in algorithm was trained. "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job # saves the trained model. Only set for successful jobs that don't use # hyperparameter tuning. "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was # trained. "pythonVersion": "A String", # Python version on which the built-in algorithm was trained. }, }, ], "isHyperparameterTuningJob": True or False, # Whether this job is a hyperparameter tuning job. "isBuiltInAlgorithmJob": True or False, # Whether this job is a built-in Algorithm job. "consumedMLUnits": 3.14, # The amount of ML units consumed by the job. "hyperparameterMetricTag": "A String", # The TensorFlow summary tag name used for optimizing hyperparameter tuning # trials. See # [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag) # for more information. Only set for hyperparameter tuning jobs. "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs. # Only set for built-in algorithms jobs. "framework": "A String", # Framework on which the built-in algorithm was trained. "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job # saves the trained model. Only set for successful jobs that don't use # hyperparameter tuning. "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was # trained. "pythonVersion": "A String", # Python version on which the built-in algorithm was trained. }, }, "predictionInput": { # Represents input parameters for a prediction job. # Input parameters to create a prediction job. "modelName": "A String", # Use this field if you want to use the default version for the specified # model. The string must use the following format: # # `"projects/YOUR_PROJECT/models/YOUR_MODEL"` "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for this batch # prediction. If not set, AI Platform will pick the runtime version used # during the CreateVersion request for this model version, or choose the # latest stable version when model version information is not available # such as when the model is specified by uri. "signatureName": "A String", # Optional. The name of the signature defined in the SavedModel to use for # this job. Please refer to # [SavedModel](https://tensorflow.github.io/serving/serving_basic.html) # for information about how to use signatures. # # Defaults to # [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants) # , which is "serving_default". "batchSize": "A String", # Optional. Number of records per batch, defaults to 64. # The service will buffer batch_size number of records in memory before # invoking one Tensorflow prediction call internally. So take the record # size and memory available into consideration when setting this parameter. "inputPaths": [ # Required. The Cloud Storage location of the input data files. May contain # wildcards. "A String", ], "maxWorkerCount": "A String", # Optional. The maximum number of workers to be used for parallel processing. # Defaults to 10 if not specified. "uri": "A String", # Use this field if you want to specify a Google Cloud Storage path for # the model to use. "outputPath": "A String", # Required. The output Google Cloud Storage location. "dataFormat": "A String", # Required. The format of the input data files. "versionName": "A String", # Use this field if you want to specify a version of the model to use. The # string is formatted the same way as `model_version`, with the addition # of the version information: # # `"projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION"` "region": "A String", # Required. The Google Compute Engine region to run the prediction job in. # See the available regions # for AI Platform services. "outputDataFormat": "A String", # Optional. Format of the output data files, defaults to JSON. }, "trainingInput": { # Represents input parameters for a training job. When using the # Input parameters to create a training job. # gcloud command to submit your training job, you can specify # the input parameters as command-line arguments and/or in a YAML configuration # file referenced from the --config command-line argument. For # details, see the guide to # submitting a training # job. "workerType": "A String", # Optional. Specifies the type of virtual machine to use for your training # job's worker nodes. # # The supported values are the same as those described in the entry for # `masterType`. # # This value must be consistent with the category of machine type that # `masterType` uses. In other words, both must be AI Platform machine # types or both must be Compute Engine machine types. # # If you use `cloud_tpu` for this value, see special instructions for # [configuring a custom TPU # machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine). # # This value must be present when `scaleTier` is set to `CUSTOM` and # `workerCount` is greater than zero. "parameterServerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers. # # You should only set `parameterServerConfig.acceleratorConfig` if # `parameterServerConfigType` is set to a Compute Engine machine type. [Learn # about restrictions on accelerator configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) # # Set `parameterServerConfig.imageUri` only if you build a custom image for # your parameter server. If `parameterServerConfig.imageUri` has not been # set, AI Platform uses the value of `masterConfig.imageUri`. # Learn more about [configuring custom # containers](/ml-engine/docs/distributed-training-containers). "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica. # [Learn about restrictions on accelerator configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) "count": "A String", # The number of accelerators to attach to each machine running the job. "type": "A String", # The type of accelerator to use. }, "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container # Registry. Learn more about [configuring custom # containers](/ml-engine/docs/distributed-training-containers). }, "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for training. If not # set, AI Platform uses the default stable version, 1.0. For more # information, see the # runtime version list # and # how to manage runtime versions. "scaleTier": "A String", # Required. Specifies the machine types, the number of replicas for workers # and parameter servers. "masterType": "A String", # Optional. Specifies the type of virtual machine to use for your training # job's master worker. # # The following types are supported: # #
#
standard
#
# A basic machine configuration suitable for training simple models with # small to moderate datasets. #
#
large_model
#
# A machine with a lot of memory, specially suited for parameter servers # when your model is large (having many hidden layers or layers with very # large numbers of nodes). #
#
complex_model_s
#
# A machine suitable for the master and workers of the cluster when your # model requires more computation than the standard machine can handle # satisfactorily. #
#
complex_model_m
#
# A machine with roughly twice the number of cores and roughly double the # memory of complex_model_s. #
#
complex_model_l
#
# A machine with roughly twice the number of cores and roughly double the # memory of complex_model_m. #
#
standard_gpu
#
# A machine equivalent to standard that # also includes a single NVIDIA Tesla K80 GPU. See more about # using GPUs to # train your model. #
#
complex_model_m_gpu
#
# A machine equivalent to complex_model_m that also includes # four NVIDIA Tesla K80 GPUs. #
#
complex_model_l_gpu
#
# A machine equivalent to complex_model_l that also includes # eight NVIDIA Tesla K80 GPUs. #
#
standard_p100
#
# A machine equivalent to standard that # also includes a single NVIDIA Tesla P100 GPU. #
#
complex_model_m_p100
#
# A machine equivalent to complex_model_m that also includes # four NVIDIA Tesla P100 GPUs. #
#
standard_v100
#
# A machine equivalent to standard that # also includes a single NVIDIA Tesla V100 GPU. #
#
large_model_v100
#
# A machine equivalent to large_model that # also includes a single NVIDIA Tesla V100 GPU. #
#
complex_model_m_v100
#
# A machine equivalent to complex_model_m that # also includes four NVIDIA Tesla V100 GPUs. #
#
complex_model_l_v100
#
# A machine equivalent to complex_model_l that # also includes eight NVIDIA Tesla V100 GPUs. #
#
cloud_tpu
#
# A TPU VM including one Cloud TPU. See more about # using TPUs to train # your model. #
#
# # You may also use certain Compute Engine machine types directly in this # field. The following types are supported: # # - `n1-standard-4` # - `n1-standard-8` # - `n1-standard-16` # - `n1-standard-32` # - `n1-standard-64` # - `n1-standard-96` # - `n1-highmem-2` # - `n1-highmem-4` # - `n1-highmem-8` # - `n1-highmem-16` # - `n1-highmem-32` # - `n1-highmem-64` # - `n1-highmem-96` # - `n1-highcpu-16` # - `n1-highcpu-32` # - `n1-highcpu-64` # - `n1-highcpu-96` # # See more about [using Compute Engine machine # types](/ml-engine/docs/tensorflow/machine-types#compute-engine-machine-types). # # You must set this value when `scaleTier` is set to `CUSTOM`. "hyperparameters": { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune. "maxTrials": 42, # Optional. How many training trials should be attempted to optimize # the specified hyperparameters. # # Defaults to one. "goal": "A String", # Required. The type of goal to use for tuning. Available types are # `MAXIMIZE` and `MINIMIZE`. # # Defaults to `MAXIMIZE`. "algorithm": "A String", # Optional. The search algorithm specified for the hyperparameter # tuning job. # Uses the default AI Platform hyperparameter tuning # algorithm if unspecified. "maxFailedTrials": 42, # Optional. The number of failed trials that need to be seen before failing # the hyperparameter tuning job. You can specify this field to override the # default failing criteria for AI Platform hyperparameter tuning jobs. # # Defaults to zero, which means the service decides when a hyperparameter # job should fail. "enableTrialEarlyStopping": True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial # early stopping. "resumePreviousJobId": "A String", # Optional. The prior hyperparameter tuning job id that users hope to # continue with. The job id will be used to find the corresponding vizier # study guid and resume the study. "params": [ # Required. The set of parameters to tune. { # Represents a single hyperparameter to optimize. "maxValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field # should be unset if type is `CATEGORICAL`. This value should be integers if # type is `INTEGER`. "categoricalValues": [ # Required if type is `CATEGORICAL`. The list of possible categories. "A String", ], "discreteValues": [ # Required if type is `DISCRETE`. # A list of feasible points. # The list should be in strictly increasing order. For instance, this # parameter might have possible settings of 1.5, 2.5, and 4.0. This list # should not contain more than 1,000 values. 3.14, ], "parameterName": "A String", # Required. The parameter name must be unique amongst all ParameterConfigs in # a HyperparameterSpec message. E.g., "learning_rate". "minValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field # should be unset if type is `CATEGORICAL`. This value should be integers if # type is INTEGER. "type": "A String", # Required. The type of the parameter. "scaleType": "A String", # Optional. How the parameter should be scaled to the hypercube. # Leave unset for categorical parameters. # Some kind of scaling is strongly recommended for real or integral # parameters (e.g., `UNIT_LINEAR_SCALE`). }, ], "hyperparameterMetricTag": "A String", # Optional. The TensorFlow summary tag name to use for optimizing trials. For # current versions of TensorFlow, this tag name should exactly match what is # shown in TensorBoard, including all scopes. For versions of TensorFlow # prior to 0.12, this should be only the tag passed to tf.Summary. # By default, "training/hptuning/metric" will be used. "maxParallelTrials": 42, # Optional. The number of training trials to run concurrently. # You can reduce the time it takes to perform hyperparameter tuning by adding # trials in parallel. However, each trail only benefits from the information # gained in completed trials. That means that a trial does not get access to # the results of trials running at the same time, which could reduce the # quality of the overall optimization. # # Each trial will use the same scale tier and machine types. # # Defaults to one. }, "region": "A String", # Required. The Google Compute Engine region to run the training job in. # See the available regions # for AI Platform services. "args": [ # Optional. Command line arguments to pass to the program. "A String", ], "pythonModule": "A String", # Required. The Python module name to run after installing the packages. "pythonVersion": "A String", # Optional. The version of Python used in training. If not set, the default # version is '2.7'. Python '3.5' is available when `runtime_version` is set # to '1.4' and above. Python '2.7' works with all supported # runtime versions. "jobDir": "A String", # Optional. A Google Cloud Storage path in which to store training outputs # and other data needed for training. This path is passed to your TensorFlow # program as the '--job-dir' command-line argument. The benefit of specifying # this field is that Cloud ML validates the path for use in training. "packageUris": [ # Required. The Google Cloud Storage location of the packages with # the training program and any additional dependencies. # The maximum number of package URIs is 100. "A String", ], "workerCount": "A String", # Optional. The number of worker replicas to use for the training job. Each # replica in the cluster will be of the type specified in `worker_type`. # # This value can only be used when `scale_tier` is set to `CUSTOM`. If you # set this value, you must also set `worker_type`. # # The default value is zero. "parameterServerType": "A String", # Optional. Specifies the type of virtual machine to use for your training # job's parameter server. # # The supported values are the same as those described in the entry for # `master_type`. # # This value must be consistent with the category of machine type that # `masterType` uses. In other words, both must be AI Platform machine # types or both must be Compute Engine machine types. # # This value must be present when `scaleTier` is set to `CUSTOM` and # `parameter_server_count` is greater than zero. "workerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers. # # You should only set `workerConfig.acceleratorConfig` if `workerType` is set # to a Compute Engine machine type. [Learn about restrictions on accelerator # configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) # # Set `workerConfig.imageUri` only if you build a custom image for your # worker. If `workerConfig.imageUri` has not been set, AI Platform uses # the value of `masterConfig.imageUri`. Learn more about # [configuring custom # containers](/ml-engine/docs/distributed-training-containers). "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica. # [Learn about restrictions on accelerator configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) "count": "A String", # The number of accelerators to attach to each machine running the job. "type": "A String", # The type of accelerator to use. }, "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container # Registry. Learn more about [configuring custom # containers](/ml-engine/docs/distributed-training-containers). }, "maxRunningTime": "A String", # Optional. The maximum job running time. The default is 7 days. "masterConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker. # # You should only set `masterConfig.acceleratorConfig` if `masterType` is set # to a Compute Engine machine type. Learn about [restrictions on accelerator # configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) # # Set `masterConfig.imageUri` only if you build a custom image. Only one of # `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more about # [configuring custom # containers](/ml-engine/docs/distributed-training-containers). "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica. # [Learn about restrictions on accelerator configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) "count": "A String", # The number of accelerators to attach to each machine running the job. "type": "A String", # The type of accelerator to use. }, "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container # Registry. Learn more about [configuring custom # containers](/ml-engine/docs/distributed-training-containers). }, "parameterServerCount": "A String", # Optional. The number of parameter server replicas to use for the training # job. Each replica in the cluster will be of the type specified in # `parameter_server_type`. # # This value can only be used when `scale_tier` is set to `CUSTOM`.If you # set this value, you must also set `parameter_server_type`. # # The default value is zero. }, "jobId": "A String", # Required. The user-specified id of the job. "labels": { # Optional. One or more labels that you can add, to organize your jobs. # Each label is a key-value pair, where both the key and the value are # arbitrary strings that you supply. # For more information, see the documentation on # using labels. "a_key": "A String", }, "state": "A String", # Output only. The detailed state of a job. "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help # prevent simultaneous updates of a job from overwriting each other. # It is strongly suggested that systems make use of the `etag` in the # read-modify-write cycle to perform job updates in order to avoid race # conditions: An `etag` is returned in the response to `GetJob`, and # systems are expected to put that etag in the request to `UpdateJob` to # ensure that their change will be applied to the same version of the job. "startTime": "A String", # Output only. When the job processing was started. "endTime": "A String", # Output only. When the job processing was completed. "predictionOutput": { # Represents results of a prediction job. # The current prediction job result. "outputPath": "A String", # The output Google Cloud Storage location provided at the job creation time. "nodeHours": 3.14, # Node hours used by the batch prediction job. "predictionCount": "A String", # The number of generated predictions. "errorCount": "A String", # The number of data instances which resulted in errors. }, "createTime": "A String", # Output only. When the job was created. }
get(name, x__xgafv=None)
Describes a job.

Args:
  name: string, Required. The name of the job to get the description of. (required)
  x__xgafv: string, V1 error format.
    Allowed values
      1 - v1 error format
      2 - v2 error format

Returns:
  An object of the form:

    { # Represents a training or prediction job.
    "errorMessage": "A String", # Output only. The details of a failure or a cancellation.
    "trainingOutput": { # Represents results of a training job. Output only. # The current training job result.
      "completedTrialCount": "A String", # The number of hyperparameter tuning trials that completed successfully.
          # Only set for hyperparameter tuning jobs.
      "trials": [ # Results for individual Hyperparameter trials.
          # Only set for hyperparameter tuning jobs.
        { # Represents the result of a single hyperparameter tuning trial from a
            # training job. The TrainingOutput object that is returned on successful
            # completion of a training job with hyperparameter tuning includes a list
            # of HyperparameterOutput objects, one for each successful trial.
          "hyperparameters": { # The hyperparameters given to this trial.
            "a_key": "A String",
          },
          "finalMetric": { # An observed value of a metric. # The final objective metric seen for this trial.
            "trainingStep": "A String", # The global training step for this metric.
            "objectiveValue": 3.14, # The objective value at this training step.
          },
          "allMetrics": [ # All recorded object metrics for this trial. This field is not currently
              # populated.
            { # An observed value of a metric.
              "trainingStep": "A String", # The global training step for this metric.
              "objectiveValue": 3.14, # The objective value at this training step.
            },
          ],
          "isTrialStoppedEarly": True or False, # True if the trial is stopped early.
          "trialId": "A String", # The trial id for these results.
          "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
              # Only set for trials of built-in algorithms jobs that have succeeded.
            "framework": "A String", # Framework on which the built-in algorithm was trained.
            "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
                # saves the trained model. Only set for successful jobs that don't use
                # hyperparameter tuning.
            "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
                # trained.
            "pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
          },
        },
      ],
      "isHyperparameterTuningJob": True or False, # Whether this job is a hyperparameter tuning job.
      "isBuiltInAlgorithmJob": True or False, # Whether this job is a built-in Algorithm job.
      "consumedMLUnits": 3.14, # The amount of ML units consumed by the job.
      "hyperparameterMetricTag": "A String", # The TensorFlow summary tag name used for optimizing hyperparameter tuning
          # trials. See
          # [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag)
          # for more information. Only set for hyperparameter tuning jobs.
      "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
          # Only set for built-in algorithms jobs.
        "framework": "A String", # Framework on which the built-in algorithm was trained.
        "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
            # saves the trained model. Only set for successful jobs that don't use
            # hyperparameter tuning.
        "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
            # trained.
        "pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
      },
    },
    "predictionInput": { # Represents input parameters for a prediction job. # Input parameters to create a prediction job.
      "modelName": "A String", # Use this field if you want to use the default version for the specified
          # model. The string must use the following format:
          #
          # `"projects/YOUR_PROJECT/models/YOUR_MODEL"`
      "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for this batch
          # prediction. If not set, AI Platform will pick the runtime version used
          # during the CreateVersion request for this model version, or choose the
          # latest stable version when model version information is not available
          # such as when the model is specified by uri.
      "signatureName": "A String", # Optional. The name of the signature defined in the SavedModel to use for
          # this job. Please refer to
          # [SavedModel](https://tensorflow.github.io/serving/serving_basic.html)
          # for information about how to use signatures.
          #
          # Defaults to
          # [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants)
          # , which is "serving_default".
      "batchSize": "A String", # Optional. Number of records per batch, defaults to 64.
          # The service will buffer batch_size number of records in memory before
          # invoking one Tensorflow prediction call internally. So take the record
          # size and memory available into consideration when setting this parameter.
      "inputPaths": [ # Required. The Cloud Storage location of the input data files. May contain
          # wildcards.
        "A String",
      ],
      "maxWorkerCount": "A String", # Optional. The maximum number of workers to be used for parallel processing.
          # Defaults to 10 if not specified.
      "uri": "A String", # Use this field if you want to specify a Google Cloud Storage path for
          # the model to use.
      "outputPath": "A String", # Required. The output Google Cloud Storage location.
      "dataFormat": "A String", # Required. The format of the input data files.
      "versionName": "A String", # Use this field if you want to specify a version of the model to use. The
          # string is formatted the same way as `model_version`, with the addition
          # of the version information:
          #
          # `"projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION"`
      "region": "A String", # Required. The Google Compute Engine region to run the prediction job in.
          # See the available regions
          # for AI Platform services.
      "outputDataFormat": "A String", # Optional. Format of the output data files, defaults to JSON.
    },
    "trainingInput": { # Represents input parameters for a training job. When using the # Input parameters to create a training job.
        # gcloud command to submit your training job, you can specify
        # the input parameters as command-line arguments and/or in a YAML configuration
        # file referenced from the --config command-line argument. For
        # details, see the guide to
        # submitting a training
        # job.
      "workerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
          # job's worker nodes.
          #
          # The supported values are the same as those described in the entry for
          # `masterType`.
          #
          # This value must be consistent with the category of machine type that
          # `masterType` uses. In other words, both must be AI Platform machine
          # types or both must be Compute Engine machine types.
          #
          # If you use `cloud_tpu` for this value, see special instructions for
          # [configuring a custom TPU
          # machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
          #
          # This value must be present when `scaleTier` is set to `CUSTOM` and
          # `workerCount` is greater than zero.
      "parameterServerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers.
          #
          # You should only set `parameterServerConfig.acceleratorConfig` if
          # `parameterServerConfigType` is set to a Compute Engine machine type. [Learn
          # about restrictions on accelerator configurations for
          # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
          #
          # Set `parameterServerConfig.imageUri` only if you build a custom image for
          # your parameter server. If `parameterServerConfig.imageUri` has not been
          # set, AI Platform uses the value of `masterConfig.imageUri`.
          # Learn more about [configuring custom
          # containers](/ml-engine/docs/distributed-training-containers).
        "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
            # [Learn about restrictions on accelerator configurations for
            # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
          "count": "A String", # The number of accelerators to attach to each machine running the job.
          "type": "A String", # The type of accelerator to use.
        },
        "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
            # Registry. Learn more about [configuring custom
            # containers](/ml-engine/docs/distributed-training-containers).
      },
      "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for training. If not
          # set, AI Platform uses the default stable version, 1.0. For more
          # information, see the
          # runtime version list
          # and
          # how to manage runtime versions.
      "scaleTier": "A String", # Required. Specifies the machine types, the number of replicas for workers
          # and parameter servers.
      "masterType": "A String", # Optional. Specifies the type of virtual machine to use for your training
          # job's master worker.
          #
          # The following types are supported:
          #
          # 
#
standard
#
# A basic machine configuration suitable for training simple models with # small to moderate datasets. #
#
large_model
#
# A machine with a lot of memory, specially suited for parameter servers # when your model is large (having many hidden layers or layers with very # large numbers of nodes). #
#
complex_model_s
#
# A machine suitable for the master and workers of the cluster when your # model requires more computation than the standard machine can handle # satisfactorily. #
#
complex_model_m
#
# A machine with roughly twice the number of cores and roughly double the # memory of complex_model_s. #
#
complex_model_l
#
# A machine with roughly twice the number of cores and roughly double the # memory of complex_model_m. #
#
standard_gpu
#
# A machine equivalent to standard that # also includes a single NVIDIA Tesla K80 GPU. See more about # using GPUs to # train your model. #
#
complex_model_m_gpu
#
# A machine equivalent to complex_model_m that also includes # four NVIDIA Tesla K80 GPUs. #
#
complex_model_l_gpu
#
# A machine equivalent to complex_model_l that also includes # eight NVIDIA Tesla K80 GPUs. #
#
standard_p100
#
# A machine equivalent to standard that # also includes a single NVIDIA Tesla P100 GPU. #
#
complex_model_m_p100
#
# A machine equivalent to complex_model_m that also includes # four NVIDIA Tesla P100 GPUs. #
#
standard_v100
#
# A machine equivalent to standard that # also includes a single NVIDIA Tesla V100 GPU. #
#
large_model_v100
#
# A machine equivalent to large_model that # also includes a single NVIDIA Tesla V100 GPU. #
#
complex_model_m_v100
#
# A machine equivalent to complex_model_m that # also includes four NVIDIA Tesla V100 GPUs. #
#
complex_model_l_v100
#
# A machine equivalent to complex_model_l that # also includes eight NVIDIA Tesla V100 GPUs. #
#
cloud_tpu
#
# A TPU VM including one Cloud TPU. See more about # using TPUs to train # your model. #
#
# # You may also use certain Compute Engine machine types directly in this # field. The following types are supported: # # - `n1-standard-4` # - `n1-standard-8` # - `n1-standard-16` # - `n1-standard-32` # - `n1-standard-64` # - `n1-standard-96` # - `n1-highmem-2` # - `n1-highmem-4` # - `n1-highmem-8` # - `n1-highmem-16` # - `n1-highmem-32` # - `n1-highmem-64` # - `n1-highmem-96` # - `n1-highcpu-16` # - `n1-highcpu-32` # - `n1-highcpu-64` # - `n1-highcpu-96` # # See more about [using Compute Engine machine # types](/ml-engine/docs/tensorflow/machine-types#compute-engine-machine-types). # # You must set this value when `scaleTier` is set to `CUSTOM`. "hyperparameters": { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune. "maxTrials": 42, # Optional. How many training trials should be attempted to optimize # the specified hyperparameters. # # Defaults to one. "goal": "A String", # Required. The type of goal to use for tuning. Available types are # `MAXIMIZE` and `MINIMIZE`. # # Defaults to `MAXIMIZE`. "algorithm": "A String", # Optional. The search algorithm specified for the hyperparameter # tuning job. # Uses the default AI Platform hyperparameter tuning # algorithm if unspecified. "maxFailedTrials": 42, # Optional. The number of failed trials that need to be seen before failing # the hyperparameter tuning job. You can specify this field to override the # default failing criteria for AI Platform hyperparameter tuning jobs. # # Defaults to zero, which means the service decides when a hyperparameter # job should fail. "enableTrialEarlyStopping": True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial # early stopping. "resumePreviousJobId": "A String", # Optional. The prior hyperparameter tuning job id that users hope to # continue with. The job id will be used to find the corresponding vizier # study guid and resume the study. "params": [ # Required. The set of parameters to tune. { # Represents a single hyperparameter to optimize. "maxValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field # should be unset if type is `CATEGORICAL`. This value should be integers if # type is `INTEGER`. "categoricalValues": [ # Required if type is `CATEGORICAL`. The list of possible categories. "A String", ], "discreteValues": [ # Required if type is `DISCRETE`. # A list of feasible points. # The list should be in strictly increasing order. For instance, this # parameter might have possible settings of 1.5, 2.5, and 4.0. This list # should not contain more than 1,000 values. 3.14, ], "parameterName": "A String", # Required. The parameter name must be unique amongst all ParameterConfigs in # a HyperparameterSpec message. E.g., "learning_rate". "minValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field # should be unset if type is `CATEGORICAL`. This value should be integers if # type is INTEGER. "type": "A String", # Required. The type of the parameter. "scaleType": "A String", # Optional. How the parameter should be scaled to the hypercube. # Leave unset for categorical parameters. # Some kind of scaling is strongly recommended for real or integral # parameters (e.g., `UNIT_LINEAR_SCALE`). }, ], "hyperparameterMetricTag": "A String", # Optional. The TensorFlow summary tag name to use for optimizing trials. For # current versions of TensorFlow, this tag name should exactly match what is # shown in TensorBoard, including all scopes. For versions of TensorFlow # prior to 0.12, this should be only the tag passed to tf.Summary. # By default, "training/hptuning/metric" will be used. "maxParallelTrials": 42, # Optional. The number of training trials to run concurrently. # You can reduce the time it takes to perform hyperparameter tuning by adding # trials in parallel. However, each trail only benefits from the information # gained in completed trials. That means that a trial does not get access to # the results of trials running at the same time, which could reduce the # quality of the overall optimization. # # Each trial will use the same scale tier and machine types. # # Defaults to one. }, "region": "A String", # Required. The Google Compute Engine region to run the training job in. # See the available regions # for AI Platform services. "args": [ # Optional. Command line arguments to pass to the program. "A String", ], "pythonModule": "A String", # Required. The Python module name to run after installing the packages. "pythonVersion": "A String", # Optional. The version of Python used in training. If not set, the default # version is '2.7'. Python '3.5' is available when `runtime_version` is set # to '1.4' and above. Python '2.7' works with all supported # runtime versions. "jobDir": "A String", # Optional. A Google Cloud Storage path in which to store training outputs # and other data needed for training. This path is passed to your TensorFlow # program as the '--job-dir' command-line argument. The benefit of specifying # this field is that Cloud ML validates the path for use in training. "packageUris": [ # Required. The Google Cloud Storage location of the packages with # the training program and any additional dependencies. # The maximum number of package URIs is 100. "A String", ], "workerCount": "A String", # Optional. The number of worker replicas to use for the training job. Each # replica in the cluster will be of the type specified in `worker_type`. # # This value can only be used when `scale_tier` is set to `CUSTOM`. If you # set this value, you must also set `worker_type`. # # The default value is zero. "parameterServerType": "A String", # Optional. Specifies the type of virtual machine to use for your training # job's parameter server. # # The supported values are the same as those described in the entry for # `master_type`. # # This value must be consistent with the category of machine type that # `masterType` uses. In other words, both must be AI Platform machine # types or both must be Compute Engine machine types. # # This value must be present when `scaleTier` is set to `CUSTOM` and # `parameter_server_count` is greater than zero. "workerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers. # # You should only set `workerConfig.acceleratorConfig` if `workerType` is set # to a Compute Engine machine type. [Learn about restrictions on accelerator # configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) # # Set `workerConfig.imageUri` only if you build a custom image for your # worker. If `workerConfig.imageUri` has not been set, AI Platform uses # the value of `masterConfig.imageUri`. Learn more about # [configuring custom # containers](/ml-engine/docs/distributed-training-containers). "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica. # [Learn about restrictions on accelerator configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) "count": "A String", # The number of accelerators to attach to each machine running the job. "type": "A String", # The type of accelerator to use. }, "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container # Registry. Learn more about [configuring custom # containers](/ml-engine/docs/distributed-training-containers). }, "maxRunningTime": "A String", # Optional. The maximum job running time. The default is 7 days. "masterConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker. # # You should only set `masterConfig.acceleratorConfig` if `masterType` is set # to a Compute Engine machine type. Learn about [restrictions on accelerator # configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) # # Set `masterConfig.imageUri` only if you build a custom image. Only one of # `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more about # [configuring custom # containers](/ml-engine/docs/distributed-training-containers). "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica. # [Learn about restrictions on accelerator configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) "count": "A String", # The number of accelerators to attach to each machine running the job. "type": "A String", # The type of accelerator to use. }, "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container # Registry. Learn more about [configuring custom # containers](/ml-engine/docs/distributed-training-containers). }, "parameterServerCount": "A String", # Optional. The number of parameter server replicas to use for the training # job. Each replica in the cluster will be of the type specified in # `parameter_server_type`. # # This value can only be used when `scale_tier` is set to `CUSTOM`.If you # set this value, you must also set `parameter_server_type`. # # The default value is zero. }, "jobId": "A String", # Required. The user-specified id of the job. "labels": { # Optional. One or more labels that you can add, to organize your jobs. # Each label is a key-value pair, where both the key and the value are # arbitrary strings that you supply. # For more information, see the documentation on # using labels. "a_key": "A String", }, "state": "A String", # Output only. The detailed state of a job. "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help # prevent simultaneous updates of a job from overwriting each other. # It is strongly suggested that systems make use of the `etag` in the # read-modify-write cycle to perform job updates in order to avoid race # conditions: An `etag` is returned in the response to `GetJob`, and # systems are expected to put that etag in the request to `UpdateJob` to # ensure that their change will be applied to the same version of the job. "startTime": "A String", # Output only. When the job processing was started. "endTime": "A String", # Output only. When the job processing was completed. "predictionOutput": { # Represents results of a prediction job. # The current prediction job result. "outputPath": "A String", # The output Google Cloud Storage location provided at the job creation time. "nodeHours": 3.14, # Node hours used by the batch prediction job. "predictionCount": "A String", # The number of generated predictions. "errorCount": "A String", # The number of data instances which resulted in errors. }, "createTime": "A String", # Output only. When the job was created. }
getIamPolicy(resource, x__xgafv=None)
Gets the access control policy for a resource.
Returns an empty policy if the resource exists and does not have a policy
set.

Args:
  resource: string, REQUIRED: The resource for which the policy is being requested.
See the operation documentation for the appropriate value for this field. (required)
  x__xgafv: string, V1 error format.
    Allowed values
      1 - v1 error format
      2 - v2 error format

Returns:
  An object of the form:

    { # Defines an Identity and Access Management (IAM) policy. It is used to
      # specify access control policies for Cloud Platform resources.
      #
      #
      # A `Policy` consists of a list of `bindings`. A `binding` binds a list of
      # `members` to a `role`, where the members can be user accounts, Google groups,
      # Google domains, and service accounts. A `role` is a named list of permissions
      # defined by IAM.
      #
      # **JSON Example**
      #
      #     {
      #       "bindings": [
      #         {
      #           "role": "roles/owner",
      #           "members": [
      #             "user:mike@example.com",
      #             "group:admins@example.com",
      #             "domain:google.com",
      #             "serviceAccount:my-other-app@appspot.gserviceaccount.com"
      #           ]
      #         },
      #         {
      #           "role": "roles/viewer",
      #           "members": ["user:sean@example.com"]
      #         }
      #       ]
      #     }
      #
      # **YAML Example**
      #
      #     bindings:
      #     - members:
      #       - user:mike@example.com
      #       - group:admins@example.com
      #       - domain:google.com
      #       - serviceAccount:my-other-app@appspot.gserviceaccount.com
      #       role: roles/owner
      #     - members:
      #       - user:sean@example.com
      #       role: roles/viewer
      #
      #
      # For a description of IAM and its features, see the
      # [IAM developer's guide](https://cloud.google.com/iam/docs).
    "bindings": [ # Associates a list of `members` to a `role`.
        # `bindings` with no members will result in an error.
      { # Associates `members` with a `role`.
        "role": "A String", # Role that is assigned to `members`.
            # For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
        "members": [ # Specifies the identities requesting access for a Cloud Platform resource.
            # `members` can have the following values:
            #
            # * `allUsers`: A special identifier that represents anyone who is
            #    on the internet; with or without a Google account.
            #
            # * `allAuthenticatedUsers`: A special identifier that represents anyone
            #    who is authenticated with a Google account or a service account.
            #
            # * `user:{emailid}`: An email address that represents a specific Google
            #    account. For example, `alice@gmail.com` .
            #
            #
            # * `serviceAccount:{emailid}`: An email address that represents a service
            #    account. For example, `my-other-app@appspot.gserviceaccount.com`.
            #
            # * `group:{emailid}`: An email address that represents a Google group.
            #    For example, `admins@example.com`.
            #
            #
            # * `domain:{domain}`: The G Suite domain (primary) that represents all the
            #    users of that domain. For example, `google.com` or `example.com`.
            #
          "A String",
        ],
        "condition": { # Represents an expression text. Example: # The condition that is associated with this binding.
            # NOTE: An unsatisfied condition will not allow user access via current
            # binding. Different bindings, including their conditions, are examined
            # independently.
            #
            #     title: "User account presence"
            #     description: "Determines whether the request has a user account"
            #     expression: "size(request.user) > 0"
          "description": "A String", # An optional description of the expression. This is a longer text which
              # describes the expression, e.g. when hovered over it in a UI.
          "expression": "A String", # Textual representation of an expression in
              # Common Expression Language syntax.
              #
              # The application context of the containing message determines which
              # well-known feature set of CEL is supported.
          "location": "A String", # An optional string indicating the location of the expression for error
              # reporting, e.g. a file name and a position in the file.
          "title": "A String", # An optional title for the expression, i.e. a short string describing
              # its purpose. This can be used e.g. in UIs which allow to enter the
              # expression.
        },
      },
    ],
    "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help
        # prevent simultaneous updates of a policy from overwriting each other.
        # It is strongly suggested that systems make use of the `etag` in the
        # read-modify-write cycle to perform policy updates in order to avoid race
        # conditions: An `etag` is returned in the response to `getIamPolicy`, and
        # systems are expected to put that etag in the request to `setIamPolicy` to
        # ensure that their change will be applied to the same version of the policy.
        #
        # If no `etag` is provided in the call to `setIamPolicy`, then the existing
        # policy is overwritten blindly.
    "version": 42, # Deprecated.
    "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
      { # Specifies the audit configuration for a service.
          # The configuration determines which permission types are logged, and what
          # identities, if any, are exempted from logging.
          # An AuditConfig must have one or more AuditLogConfigs.
          #
          # If there are AuditConfigs for both `allServices` and a specific service,
          # the union of the two AuditConfigs is used for that service: the log_types
          # specified in each AuditConfig are enabled, and the exempted_members in each
          # AuditLogConfig are exempted.
          #
          # Example Policy with multiple AuditConfigs:
          #
          #     {
          #       "audit_configs": [
          #         {
          #           "service": "allServices"
          #           "audit_log_configs": [
          #             {
          #               "log_type": "DATA_READ",
          #               "exempted_members": [
          #                 "user:foo@gmail.com"
          #               ]
          #             },
          #             {
          #               "log_type": "DATA_WRITE",
          #             },
          #             {
          #               "log_type": "ADMIN_READ",
          #             }
          #           ]
          #         },
          #         {
          #           "service": "fooservice.googleapis.com"
          #           "audit_log_configs": [
          #             {
          #               "log_type": "DATA_READ",
          #             },
          #             {
          #               "log_type": "DATA_WRITE",
          #               "exempted_members": [
          #                 "user:bar@gmail.com"
          #               ]
          #             }
          #           ]
          #         }
          #       ]
          #     }
          #
          # For fooservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ
          # logging. It also exempts foo@gmail.com from DATA_READ logging, and
          # bar@gmail.com from DATA_WRITE logging.
        "auditLogConfigs": [ # The configuration for logging of each type of permission.
          { # Provides the configuration for logging a type of permissions.
              # Example:
              #
              #     {
              #       "audit_log_configs": [
              #         {
              #           "log_type": "DATA_READ",
              #           "exempted_members": [
              #             "user:foo@gmail.com"
              #           ]
              #         },
              #         {
              #           "log_type": "DATA_WRITE",
              #         }
              #       ]
              #     }
              #
              # This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting
              # foo@gmail.com from DATA_READ logging.
            "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of
                # permission.
                # Follows the same format of Binding.members.
              "A String",
            ],
            "logType": "A String", # The log type that this config enables.
          },
        ],
        "service": "A String", # Specifies a service that will be enabled for audit logging.
            # For example, `storage.googleapis.com`, `cloudsql.googleapis.com`.
            # `allServices` is a special value that covers all services.
      },
    ],
  }
list(parent, pageToken=None, x__xgafv=None, pageSize=None, filter=None)
Lists the jobs in the project.

If there are no jobs that match the request parameters, the list
request returns an empty response body: {}.

Args:
  parent: string, Required. The name of the project for which to list jobs. (required)
  pageToken: string, Optional. A page token to request the next page of results.

You get the token from the `next_page_token` field of the response from
the previous call.
  x__xgafv: string, V1 error format.
    Allowed values
      1 - v1 error format
      2 - v2 error format
  pageSize: integer, Optional. The number of jobs to retrieve per "page" of results. If there
are more remaining results than this number, the response message will
contain a valid value in the `next_page_token` field.

The default value is 20, and the maximum page size is 100.
  filter: string, Optional. Specifies the subset of jobs to retrieve.
You can filter on the value of one or more attributes of the job object.
For example, retrieve jobs with a job identifier that starts with 'census':

gcloud ai-platform jobs list --filter='jobId:census*'

List all failed jobs with names that start with 'rnn':

gcloud ai-platform jobs list --filter='jobId:rnn* AND state:FAILED'

For more examples, see the guide to monitoring jobs. Returns: An object of the form: { # Response message for the ListJobs method. "nextPageToken": "A String", # Optional. Pass this token as the `page_token` field of the request for a # subsequent call. "jobs": [ # The list of jobs. { # Represents a training or prediction job. "errorMessage": "A String", # Output only. The details of a failure or a cancellation. "trainingOutput": { # Represents results of a training job. Output only. # The current training job result. "completedTrialCount": "A String", # The number of hyperparameter tuning trials that completed successfully. # Only set for hyperparameter tuning jobs. "trials": [ # Results for individual Hyperparameter trials. # Only set for hyperparameter tuning jobs. { # Represents the result of a single hyperparameter tuning trial from a # training job. The TrainingOutput object that is returned on successful # completion of a training job with hyperparameter tuning includes a list # of HyperparameterOutput objects, one for each successful trial. "hyperparameters": { # The hyperparameters given to this trial. "a_key": "A String", }, "finalMetric": { # An observed value of a metric. # The final objective metric seen for this trial. "trainingStep": "A String", # The global training step for this metric. "objectiveValue": 3.14, # The objective value at this training step. }, "allMetrics": [ # All recorded object metrics for this trial. This field is not currently # populated. { # An observed value of a metric. "trainingStep": "A String", # The global training step for this metric. "objectiveValue": 3.14, # The objective value at this training step. }, ], "isTrialStoppedEarly": True or False, # True if the trial is stopped early. "trialId": "A String", # The trial id for these results. "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs. # Only set for trials of built-in algorithms jobs that have succeeded. "framework": "A String", # Framework on which the built-in algorithm was trained. "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job # saves the trained model. Only set for successful jobs that don't use # hyperparameter tuning. "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was # trained. "pythonVersion": "A String", # Python version on which the built-in algorithm was trained. }, }, ], "isHyperparameterTuningJob": True or False, # Whether this job is a hyperparameter tuning job. "isBuiltInAlgorithmJob": True or False, # Whether this job is a built-in Algorithm job. "consumedMLUnits": 3.14, # The amount of ML units consumed by the job. "hyperparameterMetricTag": "A String", # The TensorFlow summary tag name used for optimizing hyperparameter tuning # trials. See # [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag) # for more information. Only set for hyperparameter tuning jobs. "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs. # Only set for built-in algorithms jobs. "framework": "A String", # Framework on which the built-in algorithm was trained. "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job # saves the trained model. Only set for successful jobs that don't use # hyperparameter tuning. "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was # trained. "pythonVersion": "A String", # Python version on which the built-in algorithm was trained. }, }, "predictionInput": { # Represents input parameters for a prediction job. # Input parameters to create a prediction job. "modelName": "A String", # Use this field if you want to use the default version for the specified # model. The string must use the following format: # # `"projects/YOUR_PROJECT/models/YOUR_MODEL"` "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for this batch # prediction. If not set, AI Platform will pick the runtime version used # during the CreateVersion request for this model version, or choose the # latest stable version when model version information is not available # such as when the model is specified by uri. "signatureName": "A String", # Optional. The name of the signature defined in the SavedModel to use for # this job. Please refer to # [SavedModel](https://tensorflow.github.io/serving/serving_basic.html) # for information about how to use signatures. # # Defaults to # [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants) # , which is "serving_default". "batchSize": "A String", # Optional. Number of records per batch, defaults to 64. # The service will buffer batch_size number of records in memory before # invoking one Tensorflow prediction call internally. So take the record # size and memory available into consideration when setting this parameter. "inputPaths": [ # Required. The Cloud Storage location of the input data files. May contain # wildcards. "A String", ], "maxWorkerCount": "A String", # Optional. The maximum number of workers to be used for parallel processing. # Defaults to 10 if not specified. "uri": "A String", # Use this field if you want to specify a Google Cloud Storage path for # the model to use. "outputPath": "A String", # Required. The output Google Cloud Storage location. "dataFormat": "A String", # Required. The format of the input data files. "versionName": "A String", # Use this field if you want to specify a version of the model to use. The # string is formatted the same way as `model_version`, with the addition # of the version information: # # `"projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION"` "region": "A String", # Required. The Google Compute Engine region to run the prediction job in. # See the available regions # for AI Platform services. "outputDataFormat": "A String", # Optional. Format of the output data files, defaults to JSON. }, "trainingInput": { # Represents input parameters for a training job. When using the # Input parameters to create a training job. # gcloud command to submit your training job, you can specify # the input parameters as command-line arguments and/or in a YAML configuration # file referenced from the --config command-line argument. For # details, see the guide to # submitting a training # job. "workerType": "A String", # Optional. Specifies the type of virtual machine to use for your training # job's worker nodes. # # The supported values are the same as those described in the entry for # `masterType`. # # This value must be consistent with the category of machine type that # `masterType` uses. In other words, both must be AI Platform machine # types or both must be Compute Engine machine types. # # If you use `cloud_tpu` for this value, see special instructions for # [configuring a custom TPU # machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine). # # This value must be present when `scaleTier` is set to `CUSTOM` and # `workerCount` is greater than zero. "parameterServerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers. # # You should only set `parameterServerConfig.acceleratorConfig` if # `parameterServerConfigType` is set to a Compute Engine machine type. [Learn # about restrictions on accelerator configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) # # Set `parameterServerConfig.imageUri` only if you build a custom image for # your parameter server. If `parameterServerConfig.imageUri` has not been # set, AI Platform uses the value of `masterConfig.imageUri`. # Learn more about [configuring custom # containers](/ml-engine/docs/distributed-training-containers). "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica. # [Learn about restrictions on accelerator configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) "count": "A String", # The number of accelerators to attach to each machine running the job. "type": "A String", # The type of accelerator to use. }, "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container # Registry. Learn more about [configuring custom # containers](/ml-engine/docs/distributed-training-containers). }, "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for training. If not # set, AI Platform uses the default stable version, 1.0. For more # information, see the # runtime version list # and # how to manage runtime versions. "scaleTier": "A String", # Required. Specifies the machine types, the number of replicas for workers # and parameter servers. "masterType": "A String", # Optional. Specifies the type of virtual machine to use for your training # job's master worker. # # The following types are supported: # #

#
standard
#
# A basic machine configuration suitable for training simple models with # small to moderate datasets. #
#
large_model
#
# A machine with a lot of memory, specially suited for parameter servers # when your model is large (having many hidden layers or layers with very # large numbers of nodes). #
#
complex_model_s
#
# A machine suitable for the master and workers of the cluster when your # model requires more computation than the standard machine can handle # satisfactorily. #
#
complex_model_m
#
# A machine with roughly twice the number of cores and roughly double the # memory of complex_model_s. #
#
complex_model_l
#
# A machine with roughly twice the number of cores and roughly double the # memory of complex_model_m. #
#
standard_gpu
#
# A machine equivalent to standard that # also includes a single NVIDIA Tesla K80 GPU. See more about # using GPUs to # train your model. #
#
complex_model_m_gpu
#
# A machine equivalent to complex_model_m that also includes # four NVIDIA Tesla K80 GPUs. #
#
complex_model_l_gpu
#
# A machine equivalent to complex_model_l that also includes # eight NVIDIA Tesla K80 GPUs. #
#
standard_p100
#
# A machine equivalent to standard that # also includes a single NVIDIA Tesla P100 GPU. #
#
complex_model_m_p100
#
# A machine equivalent to complex_model_m that also includes # four NVIDIA Tesla P100 GPUs. #
#
standard_v100
#
# A machine equivalent to standard that # also includes a single NVIDIA Tesla V100 GPU. #
#
large_model_v100
#
# A machine equivalent to large_model that # also includes a single NVIDIA Tesla V100 GPU. #
#
complex_model_m_v100
#
# A machine equivalent to complex_model_m that # also includes four NVIDIA Tesla V100 GPUs. #
#
complex_model_l_v100
#
# A machine equivalent to complex_model_l that # also includes eight NVIDIA Tesla V100 GPUs. #
#
cloud_tpu
#
# A TPU VM including one Cloud TPU. See more about # using TPUs to train # your model. #
#
# # You may also use certain Compute Engine machine types directly in this # field. The following types are supported: # # - `n1-standard-4` # - `n1-standard-8` # - `n1-standard-16` # - `n1-standard-32` # - `n1-standard-64` # - `n1-standard-96` # - `n1-highmem-2` # - `n1-highmem-4` # - `n1-highmem-8` # - `n1-highmem-16` # - `n1-highmem-32` # - `n1-highmem-64` # - `n1-highmem-96` # - `n1-highcpu-16` # - `n1-highcpu-32` # - `n1-highcpu-64` # - `n1-highcpu-96` # # See more about [using Compute Engine machine # types](/ml-engine/docs/tensorflow/machine-types#compute-engine-machine-types). # # You must set this value when `scaleTier` is set to `CUSTOM`. "hyperparameters": { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune. "maxTrials": 42, # Optional. How many training trials should be attempted to optimize # the specified hyperparameters. # # Defaults to one. "goal": "A String", # Required. The type of goal to use for tuning. Available types are # `MAXIMIZE` and `MINIMIZE`. # # Defaults to `MAXIMIZE`. "algorithm": "A String", # Optional. The search algorithm specified for the hyperparameter # tuning job. # Uses the default AI Platform hyperparameter tuning # algorithm if unspecified. "maxFailedTrials": 42, # Optional. The number of failed trials that need to be seen before failing # the hyperparameter tuning job. You can specify this field to override the # default failing criteria for AI Platform hyperparameter tuning jobs. # # Defaults to zero, which means the service decides when a hyperparameter # job should fail. "enableTrialEarlyStopping": True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial # early stopping. "resumePreviousJobId": "A String", # Optional. The prior hyperparameter tuning job id that users hope to # continue with. The job id will be used to find the corresponding vizier # study guid and resume the study. "params": [ # Required. The set of parameters to tune. { # Represents a single hyperparameter to optimize. "maxValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field # should be unset if type is `CATEGORICAL`. This value should be integers if # type is `INTEGER`. "categoricalValues": [ # Required if type is `CATEGORICAL`. The list of possible categories. "A String", ], "discreteValues": [ # Required if type is `DISCRETE`. # A list of feasible points. # The list should be in strictly increasing order. For instance, this # parameter might have possible settings of 1.5, 2.5, and 4.0. This list # should not contain more than 1,000 values. 3.14, ], "parameterName": "A String", # Required. The parameter name must be unique amongst all ParameterConfigs in # a HyperparameterSpec message. E.g., "learning_rate". "minValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field # should be unset if type is `CATEGORICAL`. This value should be integers if # type is INTEGER. "type": "A String", # Required. The type of the parameter. "scaleType": "A String", # Optional. How the parameter should be scaled to the hypercube. # Leave unset for categorical parameters. # Some kind of scaling is strongly recommended for real or integral # parameters (e.g., `UNIT_LINEAR_SCALE`). }, ], "hyperparameterMetricTag": "A String", # Optional. The TensorFlow summary tag name to use for optimizing trials. For # current versions of TensorFlow, this tag name should exactly match what is # shown in TensorBoard, including all scopes. For versions of TensorFlow # prior to 0.12, this should be only the tag passed to tf.Summary. # By default, "training/hptuning/metric" will be used. "maxParallelTrials": 42, # Optional. The number of training trials to run concurrently. # You can reduce the time it takes to perform hyperparameter tuning by adding # trials in parallel. However, each trail only benefits from the information # gained in completed trials. That means that a trial does not get access to # the results of trials running at the same time, which could reduce the # quality of the overall optimization. # # Each trial will use the same scale tier and machine types. # # Defaults to one. }, "region": "A String", # Required. The Google Compute Engine region to run the training job in. # See the available regions # for AI Platform services. "args": [ # Optional. Command line arguments to pass to the program. "A String", ], "pythonModule": "A String", # Required. The Python module name to run after installing the packages. "pythonVersion": "A String", # Optional. The version of Python used in training. If not set, the default # version is '2.7'. Python '3.5' is available when `runtime_version` is set # to '1.4' and above. Python '2.7' works with all supported # runtime versions. "jobDir": "A String", # Optional. A Google Cloud Storage path in which to store training outputs # and other data needed for training. This path is passed to your TensorFlow # program as the '--job-dir' command-line argument. The benefit of specifying # this field is that Cloud ML validates the path for use in training. "packageUris": [ # Required. The Google Cloud Storage location of the packages with # the training program and any additional dependencies. # The maximum number of package URIs is 100. "A String", ], "workerCount": "A String", # Optional. The number of worker replicas to use for the training job. Each # replica in the cluster will be of the type specified in `worker_type`. # # This value can only be used when `scale_tier` is set to `CUSTOM`. If you # set this value, you must also set `worker_type`. # # The default value is zero. "parameterServerType": "A String", # Optional. Specifies the type of virtual machine to use for your training # job's parameter server. # # The supported values are the same as those described in the entry for # `master_type`. # # This value must be consistent with the category of machine type that # `masterType` uses. In other words, both must be AI Platform machine # types or both must be Compute Engine machine types. # # This value must be present when `scaleTier` is set to `CUSTOM` and # `parameter_server_count` is greater than zero. "workerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers. # # You should only set `workerConfig.acceleratorConfig` if `workerType` is set # to a Compute Engine machine type. [Learn about restrictions on accelerator # configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) # # Set `workerConfig.imageUri` only if you build a custom image for your # worker. If `workerConfig.imageUri` has not been set, AI Platform uses # the value of `masterConfig.imageUri`. Learn more about # [configuring custom # containers](/ml-engine/docs/distributed-training-containers). "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica. # [Learn about restrictions on accelerator configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) "count": "A String", # The number of accelerators to attach to each machine running the job. "type": "A String", # The type of accelerator to use. }, "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container # Registry. Learn more about [configuring custom # containers](/ml-engine/docs/distributed-training-containers). }, "maxRunningTime": "A String", # Optional. The maximum job running time. The default is 7 days. "masterConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker. # # You should only set `masterConfig.acceleratorConfig` if `masterType` is set # to a Compute Engine machine type. Learn about [restrictions on accelerator # configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) # # Set `masterConfig.imageUri` only if you build a custom image. Only one of # `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more about # [configuring custom # containers](/ml-engine/docs/distributed-training-containers). "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica. # [Learn about restrictions on accelerator configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) "count": "A String", # The number of accelerators to attach to each machine running the job. "type": "A String", # The type of accelerator to use. }, "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container # Registry. Learn more about [configuring custom # containers](/ml-engine/docs/distributed-training-containers). }, "parameterServerCount": "A String", # Optional. The number of parameter server replicas to use for the training # job. Each replica in the cluster will be of the type specified in # `parameter_server_type`. # # This value can only be used when `scale_tier` is set to `CUSTOM`.If you # set this value, you must also set `parameter_server_type`. # # The default value is zero. }, "jobId": "A String", # Required. The user-specified id of the job. "labels": { # Optional. One or more labels that you can add, to organize your jobs. # Each label is a key-value pair, where both the key and the value are # arbitrary strings that you supply. # For more information, see the documentation on # using labels. "a_key": "A String", }, "state": "A String", # Output only. The detailed state of a job. "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help # prevent simultaneous updates of a job from overwriting each other. # It is strongly suggested that systems make use of the `etag` in the # read-modify-write cycle to perform job updates in order to avoid race # conditions: An `etag` is returned in the response to `GetJob`, and # systems are expected to put that etag in the request to `UpdateJob` to # ensure that their change will be applied to the same version of the job. "startTime": "A String", # Output only. When the job processing was started. "endTime": "A String", # Output only. When the job processing was completed. "predictionOutput": { # Represents results of a prediction job. # The current prediction job result. "outputPath": "A String", # The output Google Cloud Storage location provided at the job creation time. "nodeHours": 3.14, # Node hours used by the batch prediction job. "predictionCount": "A String", # The number of generated predictions. "errorCount": "A String", # The number of data instances which resulted in errors. }, "createTime": "A String", # Output only. When the job was created. }, ], }
list_next(previous_request, previous_response)
Retrieves the next page of results.

Args:
  previous_request: The request for the previous page. (required)
  previous_response: The response from the request for the previous page. (required)

Returns:
  A request object that you can call 'execute()' on to request the next
  page. Returns None if there are no more items in the collection.
    
patch(name, body, updateMask=None, x__xgafv=None)
Updates a specific job resource.

Currently the only supported fields to update are `labels`.

Args:
  name: string, Required. The job name. (required)
  body: object, The request body. (required)
    The object takes the form of:

{ # Represents a training or prediction job.
  "errorMessage": "A String", # Output only. The details of a failure or a cancellation.
  "trainingOutput": { # Represents results of a training job. Output only. # The current training job result.
    "completedTrialCount": "A String", # The number of hyperparameter tuning trials that completed successfully.
        # Only set for hyperparameter tuning jobs.
    "trials": [ # Results for individual Hyperparameter trials.
        # Only set for hyperparameter tuning jobs.
      { # Represents the result of a single hyperparameter tuning trial from a
          # training job. The TrainingOutput object that is returned on successful
          # completion of a training job with hyperparameter tuning includes a list
          # of HyperparameterOutput objects, one for each successful trial.
        "hyperparameters": { # The hyperparameters given to this trial.
          "a_key": "A String",
        },
        "finalMetric": { # An observed value of a metric. # The final objective metric seen for this trial.
          "trainingStep": "A String", # The global training step for this metric.
          "objectiveValue": 3.14, # The objective value at this training step.
        },
        "allMetrics": [ # All recorded object metrics for this trial. This field is not currently
            # populated.
          { # An observed value of a metric.
            "trainingStep": "A String", # The global training step for this metric.
            "objectiveValue": 3.14, # The objective value at this training step.
          },
        ],
        "isTrialStoppedEarly": True or False, # True if the trial is stopped early.
        "trialId": "A String", # The trial id for these results.
        "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
            # Only set for trials of built-in algorithms jobs that have succeeded.
          "framework": "A String", # Framework on which the built-in algorithm was trained.
          "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
              # saves the trained model. Only set for successful jobs that don't use
              # hyperparameter tuning.
          "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
              # trained.
          "pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
        },
      },
    ],
    "isHyperparameterTuningJob": True or False, # Whether this job is a hyperparameter tuning job.
    "isBuiltInAlgorithmJob": True or False, # Whether this job is a built-in Algorithm job.
    "consumedMLUnits": 3.14, # The amount of ML units consumed by the job.
    "hyperparameterMetricTag": "A String", # The TensorFlow summary tag name used for optimizing hyperparameter tuning
        # trials. See
        # [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag)
        # for more information. Only set for hyperparameter tuning jobs.
    "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
        # Only set for built-in algorithms jobs.
      "framework": "A String", # Framework on which the built-in algorithm was trained.
      "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
          # saves the trained model. Only set for successful jobs that don't use
          # hyperparameter tuning.
      "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
          # trained.
      "pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
    },
  },
  "predictionInput": { # Represents input parameters for a prediction job. # Input parameters to create a prediction job.
    "modelName": "A String", # Use this field if you want to use the default version for the specified
        # model. The string must use the following format:
        #
        # `"projects/YOUR_PROJECT/models/YOUR_MODEL"`
    "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for this batch
        # prediction. If not set, AI Platform will pick the runtime version used
        # during the CreateVersion request for this model version, or choose the
        # latest stable version when model version information is not available
        # such as when the model is specified by uri.
    "signatureName": "A String", # Optional. The name of the signature defined in the SavedModel to use for
        # this job. Please refer to
        # [SavedModel](https://tensorflow.github.io/serving/serving_basic.html)
        # for information about how to use signatures.
        #
        # Defaults to
        # [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants)
        # , which is "serving_default".
    "batchSize": "A String", # Optional. Number of records per batch, defaults to 64.
        # The service will buffer batch_size number of records in memory before
        # invoking one Tensorflow prediction call internally. So take the record
        # size and memory available into consideration when setting this parameter.
    "inputPaths": [ # Required. The Cloud Storage location of the input data files. May contain
        # wildcards.
      "A String",
    ],
    "maxWorkerCount": "A String", # Optional. The maximum number of workers to be used for parallel processing.
        # Defaults to 10 if not specified.
    "uri": "A String", # Use this field if you want to specify a Google Cloud Storage path for
        # the model to use.
    "outputPath": "A String", # Required. The output Google Cloud Storage location.
    "dataFormat": "A String", # Required. The format of the input data files.
    "versionName": "A String", # Use this field if you want to specify a version of the model to use. The
        # string is formatted the same way as `model_version`, with the addition
        # of the version information:
        #
        # `"projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION"`
    "region": "A String", # Required. The Google Compute Engine region to run the prediction job in.
        # See the available regions
        # for AI Platform services.
    "outputDataFormat": "A String", # Optional. Format of the output data files, defaults to JSON.
  },
  "trainingInput": { # Represents input parameters for a training job. When using the # Input parameters to create a training job.
      # gcloud command to submit your training job, you can specify
      # the input parameters as command-line arguments and/or in a YAML configuration
      # file referenced from the --config command-line argument. For
      # details, see the guide to
      # submitting a training
      # job.
    "workerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
        # job's worker nodes.
        #
        # The supported values are the same as those described in the entry for
        # `masterType`.
        #
        # This value must be consistent with the category of machine type that
        # `masterType` uses. In other words, both must be AI Platform machine
        # types or both must be Compute Engine machine types.
        #
        # If you use `cloud_tpu` for this value, see special instructions for
        # [configuring a custom TPU
        # machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
        #
        # This value must be present when `scaleTier` is set to `CUSTOM` and
        # `workerCount` is greater than zero.
    "parameterServerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers.
        #
        # You should only set `parameterServerConfig.acceleratorConfig` if
        # `parameterServerConfigType` is set to a Compute Engine machine type. [Learn
        # about restrictions on accelerator configurations for
        # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
        #
        # Set `parameterServerConfig.imageUri` only if you build a custom image for
        # your parameter server. If `parameterServerConfig.imageUri` has not been
        # set, AI Platform uses the value of `masterConfig.imageUri`.
        # Learn more about [configuring custom
        # containers](/ml-engine/docs/distributed-training-containers).
      "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
          # [Learn about restrictions on accelerator configurations for
          # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
        "count": "A String", # The number of accelerators to attach to each machine running the job.
        "type": "A String", # The type of accelerator to use.
      },
      "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
          # Registry. Learn more about [configuring custom
          # containers](/ml-engine/docs/distributed-training-containers).
    },
    "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for training. If not
        # set, AI Platform uses the default stable version, 1.0. For more
        # information, see the
        # runtime version list
        # and
        # how to manage runtime versions.
    "scaleTier": "A String", # Required. Specifies the machine types, the number of replicas for workers
        # and parameter servers.
    "masterType": "A String", # Optional. Specifies the type of virtual machine to use for your training
        # job's master worker.
        #
        # The following types are supported:
        #
        # 
#
standard
#
# A basic machine configuration suitable for training simple models with # small to moderate datasets. #
#
large_model
#
# A machine with a lot of memory, specially suited for parameter servers # when your model is large (having many hidden layers or layers with very # large numbers of nodes). #
#
complex_model_s
#
# A machine suitable for the master and workers of the cluster when your # model requires more computation than the standard machine can handle # satisfactorily. #
#
complex_model_m
#
# A machine with roughly twice the number of cores and roughly double the # memory of complex_model_s. #
#
complex_model_l
#
# A machine with roughly twice the number of cores and roughly double the # memory of complex_model_m. #
#
standard_gpu
#
# A machine equivalent to standard that # also includes a single NVIDIA Tesla K80 GPU. See more about # using GPUs to # train your model. #
#
complex_model_m_gpu
#
# A machine equivalent to complex_model_m that also includes # four NVIDIA Tesla K80 GPUs. #
#
complex_model_l_gpu
#
# A machine equivalent to complex_model_l that also includes # eight NVIDIA Tesla K80 GPUs. #
#
standard_p100
#
# A machine equivalent to standard that # also includes a single NVIDIA Tesla P100 GPU. #
#
complex_model_m_p100
#
# A machine equivalent to complex_model_m that also includes # four NVIDIA Tesla P100 GPUs. #
#
standard_v100
#
# A machine equivalent to standard that # also includes a single NVIDIA Tesla V100 GPU. #
#
large_model_v100
#
# A machine equivalent to large_model that # also includes a single NVIDIA Tesla V100 GPU. #
#
complex_model_m_v100
#
# A machine equivalent to complex_model_m that # also includes four NVIDIA Tesla V100 GPUs. #
#
complex_model_l_v100
#
# A machine equivalent to complex_model_l that # also includes eight NVIDIA Tesla V100 GPUs. #
#
cloud_tpu
#
# A TPU VM including one Cloud TPU. See more about # using TPUs to train # your model. #
#
# # You may also use certain Compute Engine machine types directly in this # field. The following types are supported: # # - `n1-standard-4` # - `n1-standard-8` # - `n1-standard-16` # - `n1-standard-32` # - `n1-standard-64` # - `n1-standard-96` # - `n1-highmem-2` # - `n1-highmem-4` # - `n1-highmem-8` # - `n1-highmem-16` # - `n1-highmem-32` # - `n1-highmem-64` # - `n1-highmem-96` # - `n1-highcpu-16` # - `n1-highcpu-32` # - `n1-highcpu-64` # - `n1-highcpu-96` # # See more about [using Compute Engine machine # types](/ml-engine/docs/tensorflow/machine-types#compute-engine-machine-types). # # You must set this value when `scaleTier` is set to `CUSTOM`. "hyperparameters": { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune. "maxTrials": 42, # Optional. How many training trials should be attempted to optimize # the specified hyperparameters. # # Defaults to one. "goal": "A String", # Required. The type of goal to use for tuning. Available types are # `MAXIMIZE` and `MINIMIZE`. # # Defaults to `MAXIMIZE`. "algorithm": "A String", # Optional. The search algorithm specified for the hyperparameter # tuning job. # Uses the default AI Platform hyperparameter tuning # algorithm if unspecified. "maxFailedTrials": 42, # Optional. The number of failed trials that need to be seen before failing # the hyperparameter tuning job. You can specify this field to override the # default failing criteria for AI Platform hyperparameter tuning jobs. # # Defaults to zero, which means the service decides when a hyperparameter # job should fail. "enableTrialEarlyStopping": True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial # early stopping. "resumePreviousJobId": "A String", # Optional. The prior hyperparameter tuning job id that users hope to # continue with. The job id will be used to find the corresponding vizier # study guid and resume the study. "params": [ # Required. The set of parameters to tune. { # Represents a single hyperparameter to optimize. "maxValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field # should be unset if type is `CATEGORICAL`. This value should be integers if # type is `INTEGER`. "categoricalValues": [ # Required if type is `CATEGORICAL`. The list of possible categories. "A String", ], "discreteValues": [ # Required if type is `DISCRETE`. # A list of feasible points. # The list should be in strictly increasing order. For instance, this # parameter might have possible settings of 1.5, 2.5, and 4.0. This list # should not contain more than 1,000 values. 3.14, ], "parameterName": "A String", # Required. The parameter name must be unique amongst all ParameterConfigs in # a HyperparameterSpec message. E.g., "learning_rate". "minValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field # should be unset if type is `CATEGORICAL`. This value should be integers if # type is INTEGER. "type": "A String", # Required. The type of the parameter. "scaleType": "A String", # Optional. How the parameter should be scaled to the hypercube. # Leave unset for categorical parameters. # Some kind of scaling is strongly recommended for real or integral # parameters (e.g., `UNIT_LINEAR_SCALE`). }, ], "hyperparameterMetricTag": "A String", # Optional. The TensorFlow summary tag name to use for optimizing trials. For # current versions of TensorFlow, this tag name should exactly match what is # shown in TensorBoard, including all scopes. For versions of TensorFlow # prior to 0.12, this should be only the tag passed to tf.Summary. # By default, "training/hptuning/metric" will be used. "maxParallelTrials": 42, # Optional. The number of training trials to run concurrently. # You can reduce the time it takes to perform hyperparameter tuning by adding # trials in parallel. However, each trail only benefits from the information # gained in completed trials. That means that a trial does not get access to # the results of trials running at the same time, which could reduce the # quality of the overall optimization. # # Each trial will use the same scale tier and machine types. # # Defaults to one. }, "region": "A String", # Required. The Google Compute Engine region to run the training job in. # See the available regions # for AI Platform services. "args": [ # Optional. Command line arguments to pass to the program. "A String", ], "pythonModule": "A String", # Required. The Python module name to run after installing the packages. "pythonVersion": "A String", # Optional. The version of Python used in training. If not set, the default # version is '2.7'. Python '3.5' is available when `runtime_version` is set # to '1.4' and above. Python '2.7' works with all supported # runtime versions. "jobDir": "A String", # Optional. A Google Cloud Storage path in which to store training outputs # and other data needed for training. This path is passed to your TensorFlow # program as the '--job-dir' command-line argument. The benefit of specifying # this field is that Cloud ML validates the path for use in training. "packageUris": [ # Required. The Google Cloud Storage location of the packages with # the training program and any additional dependencies. # The maximum number of package URIs is 100. "A String", ], "workerCount": "A String", # Optional. The number of worker replicas to use for the training job. Each # replica in the cluster will be of the type specified in `worker_type`. # # This value can only be used when `scale_tier` is set to `CUSTOM`. If you # set this value, you must also set `worker_type`. # # The default value is zero. "parameterServerType": "A String", # Optional. Specifies the type of virtual machine to use for your training # job's parameter server. # # The supported values are the same as those described in the entry for # `master_type`. # # This value must be consistent with the category of machine type that # `masterType` uses. In other words, both must be AI Platform machine # types or both must be Compute Engine machine types. # # This value must be present when `scaleTier` is set to `CUSTOM` and # `parameter_server_count` is greater than zero. "workerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers. # # You should only set `workerConfig.acceleratorConfig` if `workerType` is set # to a Compute Engine machine type. [Learn about restrictions on accelerator # configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) # # Set `workerConfig.imageUri` only if you build a custom image for your # worker. If `workerConfig.imageUri` has not been set, AI Platform uses # the value of `masterConfig.imageUri`. Learn more about # [configuring custom # containers](/ml-engine/docs/distributed-training-containers). "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica. # [Learn about restrictions on accelerator configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) "count": "A String", # The number of accelerators to attach to each machine running the job. "type": "A String", # The type of accelerator to use. }, "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container # Registry. Learn more about [configuring custom # containers](/ml-engine/docs/distributed-training-containers). }, "maxRunningTime": "A String", # Optional. The maximum job running time. The default is 7 days. "masterConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker. # # You should only set `masterConfig.acceleratorConfig` if `masterType` is set # to a Compute Engine machine type. Learn about [restrictions on accelerator # configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) # # Set `masterConfig.imageUri` only if you build a custom image. Only one of # `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more about # [configuring custom # containers](/ml-engine/docs/distributed-training-containers). "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica. # [Learn about restrictions on accelerator configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) "count": "A String", # The number of accelerators to attach to each machine running the job. "type": "A String", # The type of accelerator to use. }, "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container # Registry. Learn more about [configuring custom # containers](/ml-engine/docs/distributed-training-containers). }, "parameterServerCount": "A String", # Optional. The number of parameter server replicas to use for the training # job. Each replica in the cluster will be of the type specified in # `parameter_server_type`. # # This value can only be used when `scale_tier` is set to `CUSTOM`.If you # set this value, you must also set `parameter_server_type`. # # The default value is zero. }, "jobId": "A String", # Required. The user-specified id of the job. "labels": { # Optional. One or more labels that you can add, to organize your jobs. # Each label is a key-value pair, where both the key and the value are # arbitrary strings that you supply. # For more information, see the documentation on # using labels. "a_key": "A String", }, "state": "A String", # Output only. The detailed state of a job. "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help # prevent simultaneous updates of a job from overwriting each other. # It is strongly suggested that systems make use of the `etag` in the # read-modify-write cycle to perform job updates in order to avoid race # conditions: An `etag` is returned in the response to `GetJob`, and # systems are expected to put that etag in the request to `UpdateJob` to # ensure that their change will be applied to the same version of the job. "startTime": "A String", # Output only. When the job processing was started. "endTime": "A String", # Output only. When the job processing was completed. "predictionOutput": { # Represents results of a prediction job. # The current prediction job result. "outputPath": "A String", # The output Google Cloud Storage location provided at the job creation time. "nodeHours": 3.14, # Node hours used by the batch prediction job. "predictionCount": "A String", # The number of generated predictions. "errorCount": "A String", # The number of data instances which resulted in errors. }, "createTime": "A String", # Output only. When the job was created. } updateMask: string, Required. Specifies the path, relative to `Job`, of the field to update. To adopt etag mechanism, include `etag` field in the mask, and include the `etag` value in your job resource. For example, to change the labels of a job, the `update_mask` parameter would be specified as `labels`, `etag`, and the `PATCH` request body would specify the new value, as follows: { "labels": { "owner": "Google", "color": "Blue" } "etag": "33a64df551425fcc55e4d42a148795d9f25f89d4" } If `etag` matches the one on the server, the labels of the job will be replaced with the given ones, and the server end `etag` will be recalculated. Currently the only supported update masks are `labels` and `etag`. x__xgafv: string, V1 error format. Allowed values 1 - v1 error format 2 - v2 error format Returns: An object of the form: { # Represents a training or prediction job. "errorMessage": "A String", # Output only. The details of a failure or a cancellation. "trainingOutput": { # Represents results of a training job. Output only. # The current training job result. "completedTrialCount": "A String", # The number of hyperparameter tuning trials that completed successfully. # Only set for hyperparameter tuning jobs. "trials": [ # Results for individual Hyperparameter trials. # Only set for hyperparameter tuning jobs. { # Represents the result of a single hyperparameter tuning trial from a # training job. The TrainingOutput object that is returned on successful # completion of a training job with hyperparameter tuning includes a list # of HyperparameterOutput objects, one for each successful trial. "hyperparameters": { # The hyperparameters given to this trial. "a_key": "A String", }, "finalMetric": { # An observed value of a metric. # The final objective metric seen for this trial. "trainingStep": "A String", # The global training step for this metric. "objectiveValue": 3.14, # The objective value at this training step. }, "allMetrics": [ # All recorded object metrics for this trial. This field is not currently # populated. { # An observed value of a metric. "trainingStep": "A String", # The global training step for this metric. "objectiveValue": 3.14, # The objective value at this training step. }, ], "isTrialStoppedEarly": True or False, # True if the trial is stopped early. "trialId": "A String", # The trial id for these results. "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs. # Only set for trials of built-in algorithms jobs that have succeeded. "framework": "A String", # Framework on which the built-in algorithm was trained. "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job # saves the trained model. Only set for successful jobs that don't use # hyperparameter tuning. "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was # trained. "pythonVersion": "A String", # Python version on which the built-in algorithm was trained. }, }, ], "isHyperparameterTuningJob": True or False, # Whether this job is a hyperparameter tuning job. "isBuiltInAlgorithmJob": True or False, # Whether this job is a built-in Algorithm job. "consumedMLUnits": 3.14, # The amount of ML units consumed by the job. "hyperparameterMetricTag": "A String", # The TensorFlow summary tag name used for optimizing hyperparameter tuning # trials. See # [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag) # for more information. Only set for hyperparameter tuning jobs. "builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs. # Only set for built-in algorithms jobs. "framework": "A String", # Framework on which the built-in algorithm was trained. "modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job # saves the trained model. Only set for successful jobs that don't use # hyperparameter tuning. "runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was # trained. "pythonVersion": "A String", # Python version on which the built-in algorithm was trained. }, }, "predictionInput": { # Represents input parameters for a prediction job. # Input parameters to create a prediction job. "modelName": "A String", # Use this field if you want to use the default version for the specified # model. The string must use the following format: # # `"projects/YOUR_PROJECT/models/YOUR_MODEL"` "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for this batch # prediction. If not set, AI Platform will pick the runtime version used # during the CreateVersion request for this model version, or choose the # latest stable version when model version information is not available # such as when the model is specified by uri. "signatureName": "A String", # Optional. The name of the signature defined in the SavedModel to use for # this job. Please refer to # [SavedModel](https://tensorflow.github.io/serving/serving_basic.html) # for information about how to use signatures. # # Defaults to # [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants) # , which is "serving_default". "batchSize": "A String", # Optional. Number of records per batch, defaults to 64. # The service will buffer batch_size number of records in memory before # invoking one Tensorflow prediction call internally. So take the record # size and memory available into consideration when setting this parameter. "inputPaths": [ # Required. The Cloud Storage location of the input data files. May contain # wildcards. "A String", ], "maxWorkerCount": "A String", # Optional. The maximum number of workers to be used for parallel processing. # Defaults to 10 if not specified. "uri": "A String", # Use this field if you want to specify a Google Cloud Storage path for # the model to use. "outputPath": "A String", # Required. The output Google Cloud Storage location. "dataFormat": "A String", # Required. The format of the input data files. "versionName": "A String", # Use this field if you want to specify a version of the model to use. The # string is formatted the same way as `model_version`, with the addition # of the version information: # # `"projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION"` "region": "A String", # Required. The Google Compute Engine region to run the prediction job in. # See the available regions # for AI Platform services. "outputDataFormat": "A String", # Optional. Format of the output data files, defaults to JSON. }, "trainingInput": { # Represents input parameters for a training job. When using the # Input parameters to create a training job. # gcloud command to submit your training job, you can specify # the input parameters as command-line arguments and/or in a YAML configuration # file referenced from the --config command-line argument. For # details, see the guide to # submitting a training # job. "workerType": "A String", # Optional. Specifies the type of virtual machine to use for your training # job's worker nodes. # # The supported values are the same as those described in the entry for # `masterType`. # # This value must be consistent with the category of machine type that # `masterType` uses. In other words, both must be AI Platform machine # types or both must be Compute Engine machine types. # # If you use `cloud_tpu` for this value, see special instructions for # [configuring a custom TPU # machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine). # # This value must be present when `scaleTier` is set to `CUSTOM` and # `workerCount` is greater than zero. "parameterServerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers. # # You should only set `parameterServerConfig.acceleratorConfig` if # `parameterServerConfigType` is set to a Compute Engine machine type. [Learn # about restrictions on accelerator configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) # # Set `parameterServerConfig.imageUri` only if you build a custom image for # your parameter server. If `parameterServerConfig.imageUri` has not been # set, AI Platform uses the value of `masterConfig.imageUri`. # Learn more about [configuring custom # containers](/ml-engine/docs/distributed-training-containers). "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica. # [Learn about restrictions on accelerator configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) "count": "A String", # The number of accelerators to attach to each machine running the job. "type": "A String", # The type of accelerator to use. }, "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container # Registry. Learn more about [configuring custom # containers](/ml-engine/docs/distributed-training-containers). }, "runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for training. If not # set, AI Platform uses the default stable version, 1.0. For more # information, see the # runtime version list # and # how to manage runtime versions. "scaleTier": "A String", # Required. Specifies the machine types, the number of replicas for workers # and parameter servers. "masterType": "A String", # Optional. Specifies the type of virtual machine to use for your training # job's master worker. # # The following types are supported: # #
#
standard
#
# A basic machine configuration suitable for training simple models with # small to moderate datasets. #
#
large_model
#
# A machine with a lot of memory, specially suited for parameter servers # when your model is large (having many hidden layers or layers with very # large numbers of nodes). #
#
complex_model_s
#
# A machine suitable for the master and workers of the cluster when your # model requires more computation than the standard machine can handle # satisfactorily. #
#
complex_model_m
#
# A machine with roughly twice the number of cores and roughly double the # memory of complex_model_s. #
#
complex_model_l
#
# A machine with roughly twice the number of cores and roughly double the # memory of complex_model_m. #
#
standard_gpu
#
# A machine equivalent to standard that # also includes a single NVIDIA Tesla K80 GPU. See more about # using GPUs to # train your model. #
#
complex_model_m_gpu
#
# A machine equivalent to complex_model_m that also includes # four NVIDIA Tesla K80 GPUs. #
#
complex_model_l_gpu
#
# A machine equivalent to complex_model_l that also includes # eight NVIDIA Tesla K80 GPUs. #
#
standard_p100
#
# A machine equivalent to standard that # also includes a single NVIDIA Tesla P100 GPU. #
#
complex_model_m_p100
#
# A machine equivalent to complex_model_m that also includes # four NVIDIA Tesla P100 GPUs. #
#
standard_v100
#
# A machine equivalent to standard that # also includes a single NVIDIA Tesla V100 GPU. #
#
large_model_v100
#
# A machine equivalent to large_model that # also includes a single NVIDIA Tesla V100 GPU. #
#
complex_model_m_v100
#
# A machine equivalent to complex_model_m that # also includes four NVIDIA Tesla V100 GPUs. #
#
complex_model_l_v100
#
# A machine equivalent to complex_model_l that # also includes eight NVIDIA Tesla V100 GPUs. #
#
cloud_tpu
#
# A TPU VM including one Cloud TPU. See more about # using TPUs to train # your model. #
#
# # You may also use certain Compute Engine machine types directly in this # field. The following types are supported: # # - `n1-standard-4` # - `n1-standard-8` # - `n1-standard-16` # - `n1-standard-32` # - `n1-standard-64` # - `n1-standard-96` # - `n1-highmem-2` # - `n1-highmem-4` # - `n1-highmem-8` # - `n1-highmem-16` # - `n1-highmem-32` # - `n1-highmem-64` # - `n1-highmem-96` # - `n1-highcpu-16` # - `n1-highcpu-32` # - `n1-highcpu-64` # - `n1-highcpu-96` # # See more about [using Compute Engine machine # types](/ml-engine/docs/tensorflow/machine-types#compute-engine-machine-types). # # You must set this value when `scaleTier` is set to `CUSTOM`. "hyperparameters": { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune. "maxTrials": 42, # Optional. How many training trials should be attempted to optimize # the specified hyperparameters. # # Defaults to one. "goal": "A String", # Required. The type of goal to use for tuning. Available types are # `MAXIMIZE` and `MINIMIZE`. # # Defaults to `MAXIMIZE`. "algorithm": "A String", # Optional. The search algorithm specified for the hyperparameter # tuning job. # Uses the default AI Platform hyperparameter tuning # algorithm if unspecified. "maxFailedTrials": 42, # Optional. The number of failed trials that need to be seen before failing # the hyperparameter tuning job. You can specify this field to override the # default failing criteria for AI Platform hyperparameter tuning jobs. # # Defaults to zero, which means the service decides when a hyperparameter # job should fail. "enableTrialEarlyStopping": True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial # early stopping. "resumePreviousJobId": "A String", # Optional. The prior hyperparameter tuning job id that users hope to # continue with. The job id will be used to find the corresponding vizier # study guid and resume the study. "params": [ # Required. The set of parameters to tune. { # Represents a single hyperparameter to optimize. "maxValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field # should be unset if type is `CATEGORICAL`. This value should be integers if # type is `INTEGER`. "categoricalValues": [ # Required if type is `CATEGORICAL`. The list of possible categories. "A String", ], "discreteValues": [ # Required if type is `DISCRETE`. # A list of feasible points. # The list should be in strictly increasing order. For instance, this # parameter might have possible settings of 1.5, 2.5, and 4.0. This list # should not contain more than 1,000 values. 3.14, ], "parameterName": "A String", # Required. The parameter name must be unique amongst all ParameterConfigs in # a HyperparameterSpec message. E.g., "learning_rate". "minValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field # should be unset if type is `CATEGORICAL`. This value should be integers if # type is INTEGER. "type": "A String", # Required. The type of the parameter. "scaleType": "A String", # Optional. How the parameter should be scaled to the hypercube. # Leave unset for categorical parameters. # Some kind of scaling is strongly recommended for real or integral # parameters (e.g., `UNIT_LINEAR_SCALE`). }, ], "hyperparameterMetricTag": "A String", # Optional. The TensorFlow summary tag name to use for optimizing trials. For # current versions of TensorFlow, this tag name should exactly match what is # shown in TensorBoard, including all scopes. For versions of TensorFlow # prior to 0.12, this should be only the tag passed to tf.Summary. # By default, "training/hptuning/metric" will be used. "maxParallelTrials": 42, # Optional. The number of training trials to run concurrently. # You can reduce the time it takes to perform hyperparameter tuning by adding # trials in parallel. However, each trail only benefits from the information # gained in completed trials. That means that a trial does not get access to # the results of trials running at the same time, which could reduce the # quality of the overall optimization. # # Each trial will use the same scale tier and machine types. # # Defaults to one. }, "region": "A String", # Required. The Google Compute Engine region to run the training job in. # See the available regions # for AI Platform services. "args": [ # Optional. Command line arguments to pass to the program. "A String", ], "pythonModule": "A String", # Required. The Python module name to run after installing the packages. "pythonVersion": "A String", # Optional. The version of Python used in training. If not set, the default # version is '2.7'. Python '3.5' is available when `runtime_version` is set # to '1.4' and above. Python '2.7' works with all supported # runtime versions. "jobDir": "A String", # Optional. A Google Cloud Storage path in which to store training outputs # and other data needed for training. This path is passed to your TensorFlow # program as the '--job-dir' command-line argument. The benefit of specifying # this field is that Cloud ML validates the path for use in training. "packageUris": [ # Required. The Google Cloud Storage location of the packages with # the training program and any additional dependencies. # The maximum number of package URIs is 100. "A String", ], "workerCount": "A String", # Optional. The number of worker replicas to use for the training job. Each # replica in the cluster will be of the type specified in `worker_type`. # # This value can only be used when `scale_tier` is set to `CUSTOM`. If you # set this value, you must also set `worker_type`. # # The default value is zero. "parameterServerType": "A String", # Optional. Specifies the type of virtual machine to use for your training # job's parameter server. # # The supported values are the same as those described in the entry for # `master_type`. # # This value must be consistent with the category of machine type that # `masterType` uses. In other words, both must be AI Platform machine # types or both must be Compute Engine machine types. # # This value must be present when `scaleTier` is set to `CUSTOM` and # `parameter_server_count` is greater than zero. "workerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers. # # You should only set `workerConfig.acceleratorConfig` if `workerType` is set # to a Compute Engine machine type. [Learn about restrictions on accelerator # configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) # # Set `workerConfig.imageUri` only if you build a custom image for your # worker. If `workerConfig.imageUri` has not been set, AI Platform uses # the value of `masterConfig.imageUri`. Learn more about # [configuring custom # containers](/ml-engine/docs/distributed-training-containers). "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica. # [Learn about restrictions on accelerator configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) "count": "A String", # The number of accelerators to attach to each machine running the job. "type": "A String", # The type of accelerator to use. }, "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container # Registry. Learn more about [configuring custom # containers](/ml-engine/docs/distributed-training-containers). }, "maxRunningTime": "A String", # Optional. The maximum job running time. The default is 7 days. "masterConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker. # # You should only set `masterConfig.acceleratorConfig` if `masterType` is set # to a Compute Engine machine type. Learn about [restrictions on accelerator # configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) # # Set `masterConfig.imageUri` only if you build a custom image. Only one of # `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more about # [configuring custom # containers](/ml-engine/docs/distributed-training-containers). "acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica. # [Learn about restrictions on accelerator configurations for # training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu) "count": "A String", # The number of accelerators to attach to each machine running the job. "type": "A String", # The type of accelerator to use. }, "imageUri": "A String", # The Docker image to run on the replica. This image must be in Container # Registry. Learn more about [configuring custom # containers](/ml-engine/docs/distributed-training-containers). }, "parameterServerCount": "A String", # Optional. The number of parameter server replicas to use for the training # job. Each replica in the cluster will be of the type specified in # `parameter_server_type`. # # This value can only be used when `scale_tier` is set to `CUSTOM`.If you # set this value, you must also set `parameter_server_type`. # # The default value is zero. }, "jobId": "A String", # Required. The user-specified id of the job. "labels": { # Optional. One or more labels that you can add, to organize your jobs. # Each label is a key-value pair, where both the key and the value are # arbitrary strings that you supply. # For more information, see the documentation on # using labels. "a_key": "A String", }, "state": "A String", # Output only. The detailed state of a job. "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help # prevent simultaneous updates of a job from overwriting each other. # It is strongly suggested that systems make use of the `etag` in the # read-modify-write cycle to perform job updates in order to avoid race # conditions: An `etag` is returned in the response to `GetJob`, and # systems are expected to put that etag in the request to `UpdateJob` to # ensure that their change will be applied to the same version of the job. "startTime": "A String", # Output only. When the job processing was started. "endTime": "A String", # Output only. When the job processing was completed. "predictionOutput": { # Represents results of a prediction job. # The current prediction job result. "outputPath": "A String", # The output Google Cloud Storage location provided at the job creation time. "nodeHours": 3.14, # Node hours used by the batch prediction job. "predictionCount": "A String", # The number of generated predictions. "errorCount": "A String", # The number of data instances which resulted in errors. }, "createTime": "A String", # Output only. When the job was created. }
setIamPolicy(resource, body, x__xgafv=None)
Sets the access control policy on the specified resource. Replaces any
existing policy.

Args:
  resource: string, REQUIRED: The resource for which the policy is being specified.
See the operation documentation for the appropriate value for this field. (required)
  body: object, The request body. (required)
    The object takes the form of:

{ # Request message for `SetIamPolicy` method.
    "policy": { # Defines an Identity and Access Management (IAM) policy. It is used to # REQUIRED: The complete policy to be applied to the `resource`. The size of
        # the policy is limited to a few 10s of KB. An empty policy is a
        # valid policy but certain Cloud Platform services (such as Projects)
        # might reject them.
        # specify access control policies for Cloud Platform resources.
        #
        #
        # A `Policy` consists of a list of `bindings`. A `binding` binds a list of
        # `members` to a `role`, where the members can be user accounts, Google groups,
        # Google domains, and service accounts. A `role` is a named list of permissions
        # defined by IAM.
        #
        # **JSON Example**
        #
        #     {
        #       "bindings": [
        #         {
        #           "role": "roles/owner",
        #           "members": [
        #             "user:mike@example.com",
        #             "group:admins@example.com",
        #             "domain:google.com",
        #             "serviceAccount:my-other-app@appspot.gserviceaccount.com"
        #           ]
        #         },
        #         {
        #           "role": "roles/viewer",
        #           "members": ["user:sean@example.com"]
        #         }
        #       ]
        #     }
        #
        # **YAML Example**
        #
        #     bindings:
        #     - members:
        #       - user:mike@example.com
        #       - group:admins@example.com
        #       - domain:google.com
        #       - serviceAccount:my-other-app@appspot.gserviceaccount.com
        #       role: roles/owner
        #     - members:
        #       - user:sean@example.com
        #       role: roles/viewer
        #
        #
        # For a description of IAM and its features, see the
        # [IAM developer's guide](https://cloud.google.com/iam/docs).
      "bindings": [ # Associates a list of `members` to a `role`.
          # `bindings` with no members will result in an error.
        { # Associates `members` with a `role`.
          "role": "A String", # Role that is assigned to `members`.
              # For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
          "members": [ # Specifies the identities requesting access for a Cloud Platform resource.
              # `members` can have the following values:
              #
              # * `allUsers`: A special identifier that represents anyone who is
              #    on the internet; with or without a Google account.
              #
              # * `allAuthenticatedUsers`: A special identifier that represents anyone
              #    who is authenticated with a Google account or a service account.
              #
              # * `user:{emailid}`: An email address that represents a specific Google
              #    account. For example, `alice@gmail.com` .
              #
              #
              # * `serviceAccount:{emailid}`: An email address that represents a service
              #    account. For example, `my-other-app@appspot.gserviceaccount.com`.
              #
              # * `group:{emailid}`: An email address that represents a Google group.
              #    For example, `admins@example.com`.
              #
              #
              # * `domain:{domain}`: The G Suite domain (primary) that represents all the
              #    users of that domain. For example, `google.com` or `example.com`.
              #
            "A String",
          ],
          "condition": { # Represents an expression text. Example: # The condition that is associated with this binding.
              # NOTE: An unsatisfied condition will not allow user access via current
              # binding. Different bindings, including their conditions, are examined
              # independently.
              #
              #     title: "User account presence"
              #     description: "Determines whether the request has a user account"
              #     expression: "size(request.user) > 0"
            "description": "A String", # An optional description of the expression. This is a longer text which
                # describes the expression, e.g. when hovered over it in a UI.
            "expression": "A String", # Textual representation of an expression in
                # Common Expression Language syntax.
                #
                # The application context of the containing message determines which
                # well-known feature set of CEL is supported.
            "location": "A String", # An optional string indicating the location of the expression for error
                # reporting, e.g. a file name and a position in the file.
            "title": "A String", # An optional title for the expression, i.e. a short string describing
                # its purpose. This can be used e.g. in UIs which allow to enter the
                # expression.
          },
        },
      ],
      "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help
          # prevent simultaneous updates of a policy from overwriting each other.
          # It is strongly suggested that systems make use of the `etag` in the
          # read-modify-write cycle to perform policy updates in order to avoid race
          # conditions: An `etag` is returned in the response to `getIamPolicy`, and
          # systems are expected to put that etag in the request to `setIamPolicy` to
          # ensure that their change will be applied to the same version of the policy.
          #
          # If no `etag` is provided in the call to `setIamPolicy`, then the existing
          # policy is overwritten blindly.
      "version": 42, # Deprecated.
      "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
        { # Specifies the audit configuration for a service.
            # The configuration determines which permission types are logged, and what
            # identities, if any, are exempted from logging.
            # An AuditConfig must have one or more AuditLogConfigs.
            #
            # If there are AuditConfigs for both `allServices` and a specific service,
            # the union of the two AuditConfigs is used for that service: the log_types
            # specified in each AuditConfig are enabled, and the exempted_members in each
            # AuditLogConfig are exempted.
            #
            # Example Policy with multiple AuditConfigs:
            #
            #     {
            #       "audit_configs": [
            #         {
            #           "service": "allServices"
            #           "audit_log_configs": [
            #             {
            #               "log_type": "DATA_READ",
            #               "exempted_members": [
            #                 "user:foo@gmail.com"
            #               ]
            #             },
            #             {
            #               "log_type": "DATA_WRITE",
            #             },
            #             {
            #               "log_type": "ADMIN_READ",
            #             }
            #           ]
            #         },
            #         {
            #           "service": "fooservice.googleapis.com"
            #           "audit_log_configs": [
            #             {
            #               "log_type": "DATA_READ",
            #             },
            #             {
            #               "log_type": "DATA_WRITE",
            #               "exempted_members": [
            #                 "user:bar@gmail.com"
            #               ]
            #             }
            #           ]
            #         }
            #       ]
            #     }
            #
            # For fooservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ
            # logging. It also exempts foo@gmail.com from DATA_READ logging, and
            # bar@gmail.com from DATA_WRITE logging.
          "auditLogConfigs": [ # The configuration for logging of each type of permission.
            { # Provides the configuration for logging a type of permissions.
                # Example:
                #
                #     {
                #       "audit_log_configs": [
                #         {
                #           "log_type": "DATA_READ",
                #           "exempted_members": [
                #             "user:foo@gmail.com"
                #           ]
                #         },
                #         {
                #           "log_type": "DATA_WRITE",
                #         }
                #       ]
                #     }
                #
                # This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting
                # foo@gmail.com from DATA_READ logging.
              "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of
                  # permission.
                  # Follows the same format of Binding.members.
                "A String",
              ],
              "logType": "A String", # The log type that this config enables.
            },
          ],
          "service": "A String", # Specifies a service that will be enabled for audit logging.
              # For example, `storage.googleapis.com`, `cloudsql.googleapis.com`.
              # `allServices` is a special value that covers all services.
        },
      ],
    },
    "updateMask": "A String", # OPTIONAL: A FieldMask specifying which fields of the policy to modify. Only
        # the fields in the mask will be modified. If no mask is provided, the
        # following default mask is used:
        # paths: "bindings, etag"
        # This field is only used by Cloud IAM.
  }

  x__xgafv: string, V1 error format.
    Allowed values
      1 - v1 error format
      2 - v2 error format

Returns:
  An object of the form:

    { # Defines an Identity and Access Management (IAM) policy. It is used to
      # specify access control policies for Cloud Platform resources.
      #
      #
      # A `Policy` consists of a list of `bindings`. A `binding` binds a list of
      # `members` to a `role`, where the members can be user accounts, Google groups,
      # Google domains, and service accounts. A `role` is a named list of permissions
      # defined by IAM.
      #
      # **JSON Example**
      #
      #     {
      #       "bindings": [
      #         {
      #           "role": "roles/owner",
      #           "members": [
      #             "user:mike@example.com",
      #             "group:admins@example.com",
      #             "domain:google.com",
      #             "serviceAccount:my-other-app@appspot.gserviceaccount.com"
      #           ]
      #         },
      #         {
      #           "role": "roles/viewer",
      #           "members": ["user:sean@example.com"]
      #         }
      #       ]
      #     }
      #
      # **YAML Example**
      #
      #     bindings:
      #     - members:
      #       - user:mike@example.com
      #       - group:admins@example.com
      #       - domain:google.com
      #       - serviceAccount:my-other-app@appspot.gserviceaccount.com
      #       role: roles/owner
      #     - members:
      #       - user:sean@example.com
      #       role: roles/viewer
      #
      #
      # For a description of IAM and its features, see the
      # [IAM developer's guide](https://cloud.google.com/iam/docs).
    "bindings": [ # Associates a list of `members` to a `role`.
        # `bindings` with no members will result in an error.
      { # Associates `members` with a `role`.
        "role": "A String", # Role that is assigned to `members`.
            # For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
        "members": [ # Specifies the identities requesting access for a Cloud Platform resource.
            # `members` can have the following values:
            #
            # * `allUsers`: A special identifier that represents anyone who is
            #    on the internet; with or without a Google account.
            #
            # * `allAuthenticatedUsers`: A special identifier that represents anyone
            #    who is authenticated with a Google account or a service account.
            #
            # * `user:{emailid}`: An email address that represents a specific Google
            #    account. For example, `alice@gmail.com` .
            #
            #
            # * `serviceAccount:{emailid}`: An email address that represents a service
            #    account. For example, `my-other-app@appspot.gserviceaccount.com`.
            #
            # * `group:{emailid}`: An email address that represents a Google group.
            #    For example, `admins@example.com`.
            #
            #
            # * `domain:{domain}`: The G Suite domain (primary) that represents all the
            #    users of that domain. For example, `google.com` or `example.com`.
            #
          "A String",
        ],
        "condition": { # Represents an expression text. Example: # The condition that is associated with this binding.
            # NOTE: An unsatisfied condition will not allow user access via current
            # binding. Different bindings, including their conditions, are examined
            # independently.
            #
            #     title: "User account presence"
            #     description: "Determines whether the request has a user account"
            #     expression: "size(request.user) > 0"
          "description": "A String", # An optional description of the expression. This is a longer text which
              # describes the expression, e.g. when hovered over it in a UI.
          "expression": "A String", # Textual representation of an expression in
              # Common Expression Language syntax.
              #
              # The application context of the containing message determines which
              # well-known feature set of CEL is supported.
          "location": "A String", # An optional string indicating the location of the expression for error
              # reporting, e.g. a file name and a position in the file.
          "title": "A String", # An optional title for the expression, i.e. a short string describing
              # its purpose. This can be used e.g. in UIs which allow to enter the
              # expression.
        },
      },
    ],
    "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help
        # prevent simultaneous updates of a policy from overwriting each other.
        # It is strongly suggested that systems make use of the `etag` in the
        # read-modify-write cycle to perform policy updates in order to avoid race
        # conditions: An `etag` is returned in the response to `getIamPolicy`, and
        # systems are expected to put that etag in the request to `setIamPolicy` to
        # ensure that their change will be applied to the same version of the policy.
        #
        # If no `etag` is provided in the call to `setIamPolicy`, then the existing
        # policy is overwritten blindly.
    "version": 42, # Deprecated.
    "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
      { # Specifies the audit configuration for a service.
          # The configuration determines which permission types are logged, and what
          # identities, if any, are exempted from logging.
          # An AuditConfig must have one or more AuditLogConfigs.
          #
          # If there are AuditConfigs for both `allServices` and a specific service,
          # the union of the two AuditConfigs is used for that service: the log_types
          # specified in each AuditConfig are enabled, and the exempted_members in each
          # AuditLogConfig are exempted.
          #
          # Example Policy with multiple AuditConfigs:
          #
          #     {
          #       "audit_configs": [
          #         {
          #           "service": "allServices"
          #           "audit_log_configs": [
          #             {
          #               "log_type": "DATA_READ",
          #               "exempted_members": [
          #                 "user:foo@gmail.com"
          #               ]
          #             },
          #             {
          #               "log_type": "DATA_WRITE",
          #             },
          #             {
          #               "log_type": "ADMIN_READ",
          #             }
          #           ]
          #         },
          #         {
          #           "service": "fooservice.googleapis.com"
          #           "audit_log_configs": [
          #             {
          #               "log_type": "DATA_READ",
          #             },
          #             {
          #               "log_type": "DATA_WRITE",
          #               "exempted_members": [
          #                 "user:bar@gmail.com"
          #               ]
          #             }
          #           ]
          #         }
          #       ]
          #     }
          #
          # For fooservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ
          # logging. It also exempts foo@gmail.com from DATA_READ logging, and
          # bar@gmail.com from DATA_WRITE logging.
        "auditLogConfigs": [ # The configuration for logging of each type of permission.
          { # Provides the configuration for logging a type of permissions.
              # Example:
              #
              #     {
              #       "audit_log_configs": [
              #         {
              #           "log_type": "DATA_READ",
              #           "exempted_members": [
              #             "user:foo@gmail.com"
              #           ]
              #         },
              #         {
              #           "log_type": "DATA_WRITE",
              #         }
              #       ]
              #     }
              #
              # This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting
              # foo@gmail.com from DATA_READ logging.
            "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of
                # permission.
                # Follows the same format of Binding.members.
              "A String",
            ],
            "logType": "A String", # The log type that this config enables.
          },
        ],
        "service": "A String", # Specifies a service that will be enabled for audit logging.
            # For example, `storage.googleapis.com`, `cloudsql.googleapis.com`.
            # `allServices` is a special value that covers all services.
      },
    ],
  }
testIamPermissions(resource, body, x__xgafv=None)
Returns permissions that a caller has on the specified resource.
If the resource does not exist, this will return an empty set of
permissions, not a NOT_FOUND error.

Note: This operation is designed to be used for building permission-aware
UIs and command-line tools, not for authorization checking. This operation
may "fail open" without warning.

Args:
  resource: string, REQUIRED: The resource for which the policy detail is being requested.
See the operation documentation for the appropriate value for this field. (required)
  body: object, The request body. (required)
    The object takes the form of:

{ # Request message for `TestIamPermissions` method.
    "permissions": [ # The set of permissions to check for the `resource`. Permissions with
        # wildcards (such as '*' or 'storage.*') are not allowed. For more
        # information see
        # [IAM Overview](https://cloud.google.com/iam/docs/overview#permissions).
      "A String",
    ],
  }

  x__xgafv: string, V1 error format.
    Allowed values
      1 - v1 error format
      2 - v2 error format

Returns:
  An object of the form:

    { # Response message for `TestIamPermissions` method.
    "permissions": [ # A subset of `TestPermissionsRequest.permissions` that the caller is
        # allowed.
      "A String",
    ],
  }