You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
3648 lines
196 KiB
3648 lines
196 KiB
<html><body>
|
|
<style>
|
|
|
|
body, h1, h2, h3, div, span, p, pre, a {
|
|
margin: 0;
|
|
padding: 0;
|
|
border: 0;
|
|
font-weight: inherit;
|
|
font-style: inherit;
|
|
font-size: 100%;
|
|
font-family: inherit;
|
|
vertical-align: baseline;
|
|
}
|
|
|
|
body {
|
|
font-size: 13px;
|
|
padding: 1em;
|
|
}
|
|
|
|
h1 {
|
|
font-size: 26px;
|
|
margin-bottom: 1em;
|
|
}
|
|
|
|
h2 {
|
|
font-size: 24px;
|
|
margin-bottom: 1em;
|
|
}
|
|
|
|
h3 {
|
|
font-size: 20px;
|
|
margin-bottom: 1em;
|
|
margin-top: 1em;
|
|
}
|
|
|
|
pre, code {
|
|
line-height: 1.5;
|
|
font-family: Monaco, 'DejaVu Sans Mono', 'Bitstream Vera Sans Mono', 'Lucida Console', monospace;
|
|
}
|
|
|
|
pre {
|
|
margin-top: 0.5em;
|
|
}
|
|
|
|
h1, h2, h3, p {
|
|
font-family: Arial, sans serif;
|
|
}
|
|
|
|
h1, h2, h3 {
|
|
border-bottom: solid #CCC 1px;
|
|
}
|
|
|
|
.toc_element {
|
|
margin-top: 0.5em;
|
|
}
|
|
|
|
.firstline {
|
|
margin-left: 2 em;
|
|
}
|
|
|
|
.method {
|
|
margin-top: 1em;
|
|
border: solid 1px #CCC;
|
|
padding: 1em;
|
|
background: #EEE;
|
|
}
|
|
|
|
.details {
|
|
font-weight: bold;
|
|
font-size: 14px;
|
|
}
|
|
|
|
</style>
|
|
|
|
<h1><a href="ml_v1.html">Cloud Machine Learning Engine</a> . <a href="ml_v1.projects.html">projects</a> . <a href="ml_v1.projects.jobs.html">jobs</a></h1>
|
|
<h2>Instance Methods</h2>
|
|
<p class="toc_element">
|
|
<code><a href="#cancel">cancel(name, body=None, x__xgafv=None)</a></code></p>
|
|
<p class="firstline">Cancels a running job.</p>
|
|
<p class="toc_element">
|
|
<code><a href="#create">create(parent, body, x__xgafv=None)</a></code></p>
|
|
<p class="firstline">Creates a training or a batch prediction job.</p>
|
|
<p class="toc_element">
|
|
<code><a href="#get">get(name, x__xgafv=None)</a></code></p>
|
|
<p class="firstline">Describes a job.</p>
|
|
<p class="toc_element">
|
|
<code><a href="#getIamPolicy">getIamPolicy(resource, x__xgafv=None)</a></code></p>
|
|
<p class="firstline">Gets the access control policy for a resource.</p>
|
|
<p class="toc_element">
|
|
<code><a href="#list">list(parent, pageToken=None, x__xgafv=None, pageSize=None, filter=None)</a></code></p>
|
|
<p class="firstline">Lists the jobs in the project.</p>
|
|
<p class="toc_element">
|
|
<code><a href="#list_next">list_next(previous_request, previous_response)</a></code></p>
|
|
<p class="firstline">Retrieves the next page of results.</p>
|
|
<p class="toc_element">
|
|
<code><a href="#patch">patch(name, body, updateMask=None, x__xgafv=None)</a></code></p>
|
|
<p class="firstline">Updates a specific job resource.</p>
|
|
<p class="toc_element">
|
|
<code><a href="#setIamPolicy">setIamPolicy(resource, body, x__xgafv=None)</a></code></p>
|
|
<p class="firstline">Sets the access control policy on the specified resource. Replaces any</p>
|
|
<p class="toc_element">
|
|
<code><a href="#testIamPermissions">testIamPermissions(resource, body, x__xgafv=None)</a></code></p>
|
|
<p class="firstline">Returns permissions that a caller has on the specified resource.</p>
|
|
<h3>Method Details</h3>
|
|
<div class="method">
|
|
<code class="details" id="cancel">cancel(name, body=None, x__xgafv=None)</code>
|
|
<pre>Cancels a running job.
|
|
|
|
Args:
|
|
name: string, Required. The name of the job to cancel. (required)
|
|
body: object, The request body.
|
|
The object takes the form of:
|
|
|
|
{ # Request message for the CancelJob method.
|
|
}
|
|
|
|
x__xgafv: string, V1 error format.
|
|
Allowed values
|
|
1 - v1 error format
|
|
2 - v2 error format
|
|
|
|
Returns:
|
|
An object of the form:
|
|
|
|
{ # A generic empty message that you can re-use to avoid defining duplicated
|
|
# empty messages in your APIs. A typical example is to use it as the request
|
|
# or the response type of an API method. For instance:
|
|
#
|
|
# service Foo {
|
|
# rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty);
|
|
# }
|
|
#
|
|
# The JSON representation for `Empty` is empty JSON object `{}`.
|
|
}</pre>
|
|
</div>
|
|
|
|
<div class="method">
|
|
<code class="details" id="create">create(parent, body, x__xgafv=None)</code>
|
|
<pre>Creates a training or a batch prediction job.
|
|
|
|
Args:
|
|
parent: string, Required. The project name. (required)
|
|
body: object, The request body. (required)
|
|
The object takes the form of:
|
|
|
|
{ # Represents a training or prediction job.
|
|
"errorMessage": "A String", # Output only. The details of a failure or a cancellation.
|
|
"trainingOutput": { # Represents results of a training job. Output only. # The current training job result.
|
|
"completedTrialCount": "A String", # The number of hyperparameter tuning trials that completed successfully.
|
|
# Only set for hyperparameter tuning jobs.
|
|
"trials": [ # Results for individual Hyperparameter trials.
|
|
# Only set for hyperparameter tuning jobs.
|
|
{ # Represents the result of a single hyperparameter tuning trial from a
|
|
# training job. The TrainingOutput object that is returned on successful
|
|
# completion of a training job with hyperparameter tuning includes a list
|
|
# of HyperparameterOutput objects, one for each successful trial.
|
|
"hyperparameters": { # The hyperparameters given to this trial.
|
|
"a_key": "A String",
|
|
},
|
|
"finalMetric": { # An observed value of a metric. # The final objective metric seen for this trial.
|
|
"trainingStep": "A String", # The global training step for this metric.
|
|
"objectiveValue": 3.14, # The objective value at this training step.
|
|
},
|
|
"allMetrics": [ # All recorded object metrics for this trial. This field is not currently
|
|
# populated.
|
|
{ # An observed value of a metric.
|
|
"trainingStep": "A String", # The global training step for this metric.
|
|
"objectiveValue": 3.14, # The objective value at this training step.
|
|
},
|
|
],
|
|
"isTrialStoppedEarly": True or False, # True if the trial is stopped early.
|
|
"trialId": "A String", # The trial id for these results.
|
|
"builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
|
|
# Only set for trials of built-in algorithms jobs that have succeeded.
|
|
"framework": "A String", # Framework on which the built-in algorithm was trained.
|
|
"modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
|
|
# saves the trained model. Only set for successful jobs that don't use
|
|
# hyperparameter tuning.
|
|
"runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
|
|
# trained.
|
|
"pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
|
|
},
|
|
},
|
|
],
|
|
"isHyperparameterTuningJob": True or False, # Whether this job is a hyperparameter tuning job.
|
|
"isBuiltInAlgorithmJob": True or False, # Whether this job is a built-in Algorithm job.
|
|
"consumedMLUnits": 3.14, # The amount of ML units consumed by the job.
|
|
"hyperparameterMetricTag": "A String", # The TensorFlow summary tag name used for optimizing hyperparameter tuning
|
|
# trials. See
|
|
# [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag)
|
|
# for more information. Only set for hyperparameter tuning jobs.
|
|
"builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
|
|
# Only set for built-in algorithms jobs.
|
|
"framework": "A String", # Framework on which the built-in algorithm was trained.
|
|
"modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
|
|
# saves the trained model. Only set for successful jobs that don't use
|
|
# hyperparameter tuning.
|
|
"runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
|
|
# trained.
|
|
"pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
|
|
},
|
|
},
|
|
"predictionInput": { # Represents input parameters for a prediction job. # Input parameters to create a prediction job.
|
|
"modelName": "A String", # Use this field if you want to use the default version for the specified
|
|
# model. The string must use the following format:
|
|
#
|
|
# `"projects/YOUR_PROJECT/models/YOUR_MODEL"`
|
|
"runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for this batch
|
|
# prediction. If not set, AI Platform will pick the runtime version used
|
|
# during the CreateVersion request for this model version, or choose the
|
|
# latest stable version when model version information is not available
|
|
# such as when the model is specified by uri.
|
|
"signatureName": "A String", # Optional. The name of the signature defined in the SavedModel to use for
|
|
# this job. Please refer to
|
|
# [SavedModel](https://tensorflow.github.io/serving/serving_basic.html)
|
|
# for information about how to use signatures.
|
|
#
|
|
# Defaults to
|
|
# [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants)
|
|
# , which is "serving_default".
|
|
"batchSize": "A String", # Optional. Number of records per batch, defaults to 64.
|
|
# The service will buffer batch_size number of records in memory before
|
|
# invoking one Tensorflow prediction call internally. So take the record
|
|
# size and memory available into consideration when setting this parameter.
|
|
"inputPaths": [ # Required. The Cloud Storage location of the input data files. May contain
|
|
# <a href="/storage/docs/gsutil/addlhelp/WildcardNames">wildcards</a>.
|
|
"A String",
|
|
],
|
|
"maxWorkerCount": "A String", # Optional. The maximum number of workers to be used for parallel processing.
|
|
# Defaults to 10 if not specified.
|
|
"uri": "A String", # Use this field if you want to specify a Google Cloud Storage path for
|
|
# the model to use.
|
|
"outputPath": "A String", # Required. The output Google Cloud Storage location.
|
|
"dataFormat": "A String", # Required. The format of the input data files.
|
|
"versionName": "A String", # Use this field if you want to specify a version of the model to use. The
|
|
# string is formatted the same way as `model_version`, with the addition
|
|
# of the version information:
|
|
#
|
|
# `"projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION"`
|
|
"region": "A String", # Required. The Google Compute Engine region to run the prediction job in.
|
|
# See the <a href="/ml-engine/docs/tensorflow/regions">available regions</a>
|
|
# for AI Platform services.
|
|
"outputDataFormat": "A String", # Optional. Format of the output data files, defaults to JSON.
|
|
},
|
|
"trainingInput": { # Represents input parameters for a training job. When using the # Input parameters to create a training job.
|
|
# gcloud command to submit your training job, you can specify
|
|
# the input parameters as command-line arguments and/or in a YAML configuration
|
|
# file referenced from the --config command-line argument. For
|
|
# details, see the guide to
|
|
# <a href="/ml-engine/docs/tensorflow/training-jobs">submitting a training
|
|
# job</a>.
|
|
"workerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
|
|
# job's worker nodes.
|
|
#
|
|
# The supported values are the same as those described in the entry for
|
|
# `masterType`.
|
|
#
|
|
# This value must be consistent with the category of machine type that
|
|
# `masterType` uses. In other words, both must be AI Platform machine
|
|
# types or both must be Compute Engine machine types.
|
|
#
|
|
# If you use `cloud_tpu` for this value, see special instructions for
|
|
# [configuring a custom TPU
|
|
# machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
|
|
#
|
|
# This value must be present when `scaleTier` is set to `CUSTOM` and
|
|
# `workerCount` is greater than zero.
|
|
"parameterServerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers.
|
|
#
|
|
# You should only set `parameterServerConfig.acceleratorConfig` if
|
|
# `parameterServerConfigType` is set to a Compute Engine machine type. [Learn
|
|
# about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
#
|
|
# Set `parameterServerConfig.imageUri` only if you build a custom image for
|
|
# your parameter server. If `parameterServerConfig.imageUri` has not been
|
|
# set, AI Platform uses the value of `masterConfig.imageUri`.
|
|
# Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
"acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
|
|
# [Learn about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
"count": "A String", # The number of accelerators to attach to each machine running the job.
|
|
"type": "A String", # The type of accelerator to use.
|
|
},
|
|
"imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
|
|
# Registry. Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
},
|
|
"runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for training. If not
|
|
# set, AI Platform uses the default stable version, 1.0. For more
|
|
# information, see the
|
|
# <a href="/ml-engine/docs/runtime-version-list">runtime version list</a>
|
|
# and
|
|
# <a href="/ml-engine/docs/versioning">how to manage runtime versions</a>.
|
|
"scaleTier": "A String", # Required. Specifies the machine types, the number of replicas for workers
|
|
# and parameter servers.
|
|
"masterType": "A String", # Optional. Specifies the type of virtual machine to use for your training
|
|
# job's master worker.
|
|
#
|
|
# The following types are supported:
|
|
#
|
|
# <dl>
|
|
# <dt>standard</dt>
|
|
# <dd>
|
|
# A basic machine configuration suitable for training simple models with
|
|
# small to moderate datasets.
|
|
# </dd>
|
|
# <dt>large_model</dt>
|
|
# <dd>
|
|
# A machine with a lot of memory, specially suited for parameter servers
|
|
# when your model is large (having many hidden layers or layers with very
|
|
# large numbers of nodes).
|
|
# </dd>
|
|
# <dt>complex_model_s</dt>
|
|
# <dd>
|
|
# A machine suitable for the master and workers of the cluster when your
|
|
# model requires more computation than the standard machine can handle
|
|
# satisfactorily.
|
|
# </dd>
|
|
# <dt>complex_model_m</dt>
|
|
# <dd>
|
|
# A machine with roughly twice the number of cores and roughly double the
|
|
# memory of <i>complex_model_s</i>.
|
|
# </dd>
|
|
# <dt>complex_model_l</dt>
|
|
# <dd>
|
|
# A machine with roughly twice the number of cores and roughly double the
|
|
# memory of <i>complex_model_m</i>.
|
|
# </dd>
|
|
# <dt>standard_gpu</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>standard</i> that
|
|
# also includes a single NVIDIA Tesla K80 GPU. See more about
|
|
# <a href="/ml-engine/docs/tensorflow/using-gpus">using GPUs to
|
|
# train your model</a>.
|
|
# </dd>
|
|
# <dt>complex_model_m_gpu</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_m</i> that also includes
|
|
# four NVIDIA Tesla K80 GPUs.
|
|
# </dd>
|
|
# <dt>complex_model_l_gpu</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_l</i> that also includes
|
|
# eight NVIDIA Tesla K80 GPUs.
|
|
# </dd>
|
|
# <dt>standard_p100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>standard</i> that
|
|
# also includes a single NVIDIA Tesla P100 GPU.
|
|
# </dd>
|
|
# <dt>complex_model_m_p100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_m</i> that also includes
|
|
# four NVIDIA Tesla P100 GPUs.
|
|
# </dd>
|
|
# <dt>standard_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>standard</i> that
|
|
# also includes a single NVIDIA Tesla V100 GPU.
|
|
# </dd>
|
|
# <dt>large_model_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>large_model</i> that
|
|
# also includes a single NVIDIA Tesla V100 GPU.
|
|
# </dd>
|
|
# <dt>complex_model_m_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_m</i> that
|
|
# also includes four NVIDIA Tesla V100 GPUs.
|
|
# </dd>
|
|
# <dt>complex_model_l_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_l</i> that
|
|
# also includes eight NVIDIA Tesla V100 GPUs.
|
|
# </dd>
|
|
# <dt>cloud_tpu</dt>
|
|
# <dd>
|
|
# A TPU VM including one Cloud TPU. See more about
|
|
# <a href="/ml-engine/docs/tensorflow/using-tpus">using TPUs to train
|
|
# your model</a>.
|
|
# </dd>
|
|
# </dl>
|
|
#
|
|
# You may also use certain Compute Engine machine types directly in this
|
|
# field. The following types are supported:
|
|
#
|
|
# - `n1-standard-4`
|
|
# - `n1-standard-8`
|
|
# - `n1-standard-16`
|
|
# - `n1-standard-32`
|
|
# - `n1-standard-64`
|
|
# - `n1-standard-96`
|
|
# - `n1-highmem-2`
|
|
# - `n1-highmem-4`
|
|
# - `n1-highmem-8`
|
|
# - `n1-highmem-16`
|
|
# - `n1-highmem-32`
|
|
# - `n1-highmem-64`
|
|
# - `n1-highmem-96`
|
|
# - `n1-highcpu-16`
|
|
# - `n1-highcpu-32`
|
|
# - `n1-highcpu-64`
|
|
# - `n1-highcpu-96`
|
|
#
|
|
# See more about [using Compute Engine machine
|
|
# types](/ml-engine/docs/tensorflow/machine-types#compute-engine-machine-types).
|
|
#
|
|
# You must set this value when `scaleTier` is set to `CUSTOM`.
|
|
"hyperparameters": { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune.
|
|
"maxTrials": 42, # Optional. How many training trials should be attempted to optimize
|
|
# the specified hyperparameters.
|
|
#
|
|
# Defaults to one.
|
|
"goal": "A String", # Required. The type of goal to use for tuning. Available types are
|
|
# `MAXIMIZE` and `MINIMIZE`.
|
|
#
|
|
# Defaults to `MAXIMIZE`.
|
|
"algorithm": "A String", # Optional. The search algorithm specified for the hyperparameter
|
|
# tuning job.
|
|
# Uses the default AI Platform hyperparameter tuning
|
|
# algorithm if unspecified.
|
|
"maxFailedTrials": 42, # Optional. The number of failed trials that need to be seen before failing
|
|
# the hyperparameter tuning job. You can specify this field to override the
|
|
# default failing criteria for AI Platform hyperparameter tuning jobs.
|
|
#
|
|
# Defaults to zero, which means the service decides when a hyperparameter
|
|
# job should fail.
|
|
"enableTrialEarlyStopping": True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial
|
|
# early stopping.
|
|
"resumePreviousJobId": "A String", # Optional. The prior hyperparameter tuning job id that users hope to
|
|
# continue with. The job id will be used to find the corresponding vizier
|
|
# study guid and resume the study.
|
|
"params": [ # Required. The set of parameters to tune.
|
|
{ # Represents a single hyperparameter to optimize.
|
|
"maxValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
|
|
# should be unset if type is `CATEGORICAL`. This value should be integers if
|
|
# type is `INTEGER`.
|
|
"categoricalValues": [ # Required if type is `CATEGORICAL`. The list of possible categories.
|
|
"A String",
|
|
],
|
|
"discreteValues": [ # Required if type is `DISCRETE`.
|
|
# A list of feasible points.
|
|
# The list should be in strictly increasing order. For instance, this
|
|
# parameter might have possible settings of 1.5, 2.5, and 4.0. This list
|
|
# should not contain more than 1,000 values.
|
|
3.14,
|
|
],
|
|
"parameterName": "A String", # Required. The parameter name must be unique amongst all ParameterConfigs in
|
|
# a HyperparameterSpec message. E.g., "learning_rate".
|
|
"minValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
|
|
# should be unset if type is `CATEGORICAL`. This value should be integers if
|
|
# type is INTEGER.
|
|
"type": "A String", # Required. The type of the parameter.
|
|
"scaleType": "A String", # Optional. How the parameter should be scaled to the hypercube.
|
|
# Leave unset for categorical parameters.
|
|
# Some kind of scaling is strongly recommended for real or integral
|
|
# parameters (e.g., `UNIT_LINEAR_SCALE`).
|
|
},
|
|
],
|
|
"hyperparameterMetricTag": "A String", # Optional. The TensorFlow summary tag name to use for optimizing trials. For
|
|
# current versions of TensorFlow, this tag name should exactly match what is
|
|
# shown in TensorBoard, including all scopes. For versions of TensorFlow
|
|
# prior to 0.12, this should be only the tag passed to tf.Summary.
|
|
# By default, "training/hptuning/metric" will be used.
|
|
"maxParallelTrials": 42, # Optional. The number of training trials to run concurrently.
|
|
# You can reduce the time it takes to perform hyperparameter tuning by adding
|
|
# trials in parallel. However, each trail only benefits from the information
|
|
# gained in completed trials. That means that a trial does not get access to
|
|
# the results of trials running at the same time, which could reduce the
|
|
# quality of the overall optimization.
|
|
#
|
|
# Each trial will use the same scale tier and machine types.
|
|
#
|
|
# Defaults to one.
|
|
},
|
|
"region": "A String", # Required. The Google Compute Engine region to run the training job in.
|
|
# See the <a href="/ml-engine/docs/tensorflow/regions">available regions</a>
|
|
# for AI Platform services.
|
|
"args": [ # Optional. Command line arguments to pass to the program.
|
|
"A String",
|
|
],
|
|
"pythonModule": "A String", # Required. The Python module name to run after installing the packages.
|
|
"pythonVersion": "A String", # Optional. The version of Python used in training. If not set, the default
|
|
# version is '2.7'. Python '3.5' is available when `runtime_version` is set
|
|
# to '1.4' and above. Python '2.7' works with all supported
|
|
# <a href="/ml-engine/docs/runtime-version-list">runtime versions</a>.
|
|
"jobDir": "A String", # Optional. A Google Cloud Storage path in which to store training outputs
|
|
# and other data needed for training. This path is passed to your TensorFlow
|
|
# program as the '--job-dir' command-line argument. The benefit of specifying
|
|
# this field is that Cloud ML validates the path for use in training.
|
|
"packageUris": [ # Required. The Google Cloud Storage location of the packages with
|
|
# the training program and any additional dependencies.
|
|
# The maximum number of package URIs is 100.
|
|
"A String",
|
|
],
|
|
"workerCount": "A String", # Optional. The number of worker replicas to use for the training job. Each
|
|
# replica in the cluster will be of the type specified in `worker_type`.
|
|
#
|
|
# This value can only be used when `scale_tier` is set to `CUSTOM`. If you
|
|
# set this value, you must also set `worker_type`.
|
|
#
|
|
# The default value is zero.
|
|
"parameterServerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
|
|
# job's parameter server.
|
|
#
|
|
# The supported values are the same as those described in the entry for
|
|
# `master_type`.
|
|
#
|
|
# This value must be consistent with the category of machine type that
|
|
# `masterType` uses. In other words, both must be AI Platform machine
|
|
# types or both must be Compute Engine machine types.
|
|
#
|
|
# This value must be present when `scaleTier` is set to `CUSTOM` and
|
|
# `parameter_server_count` is greater than zero.
|
|
"workerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers.
|
|
#
|
|
# You should only set `workerConfig.acceleratorConfig` if `workerType` is set
|
|
# to a Compute Engine machine type. [Learn about restrictions on accelerator
|
|
# configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
#
|
|
# Set `workerConfig.imageUri` only if you build a custom image for your
|
|
# worker. If `workerConfig.imageUri` has not been set, AI Platform uses
|
|
# the value of `masterConfig.imageUri`. Learn more about
|
|
# [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
"acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
|
|
# [Learn about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
"count": "A String", # The number of accelerators to attach to each machine running the job.
|
|
"type": "A String", # The type of accelerator to use.
|
|
},
|
|
"imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
|
|
# Registry. Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
},
|
|
"maxRunningTime": "A String", # Optional. The maximum job running time. The default is 7 days.
|
|
"masterConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker.
|
|
#
|
|
# You should only set `masterConfig.acceleratorConfig` if `masterType` is set
|
|
# to a Compute Engine machine type. Learn about [restrictions on accelerator
|
|
# configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
#
|
|
# Set `masterConfig.imageUri` only if you build a custom image. Only one of
|
|
# `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more about
|
|
# [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
"acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
|
|
# [Learn about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
"count": "A String", # The number of accelerators to attach to each machine running the job.
|
|
"type": "A String", # The type of accelerator to use.
|
|
},
|
|
"imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
|
|
# Registry. Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
},
|
|
"parameterServerCount": "A String", # Optional. The number of parameter server replicas to use for the training
|
|
# job. Each replica in the cluster will be of the type specified in
|
|
# `parameter_server_type`.
|
|
#
|
|
# This value can only be used when `scale_tier` is set to `CUSTOM`.If you
|
|
# set this value, you must also set `parameter_server_type`.
|
|
#
|
|
# The default value is zero.
|
|
},
|
|
"jobId": "A String", # Required. The user-specified id of the job.
|
|
"labels": { # Optional. One or more labels that you can add, to organize your jobs.
|
|
# Each label is a key-value pair, where both the key and the value are
|
|
# arbitrary strings that you supply.
|
|
# For more information, see the documentation on
|
|
# <a href="/ml-engine/docs/tensorflow/resource-labels">using labels</a>.
|
|
"a_key": "A String",
|
|
},
|
|
"state": "A String", # Output only. The detailed state of a job.
|
|
"etag": "A String", # `etag` is used for optimistic concurrency control as a way to help
|
|
# prevent simultaneous updates of a job from overwriting each other.
|
|
# It is strongly suggested that systems make use of the `etag` in the
|
|
# read-modify-write cycle to perform job updates in order to avoid race
|
|
# conditions: An `etag` is returned in the response to `GetJob`, and
|
|
# systems are expected to put that etag in the request to `UpdateJob` to
|
|
# ensure that their change will be applied to the same version of the job.
|
|
"startTime": "A String", # Output only. When the job processing was started.
|
|
"endTime": "A String", # Output only. When the job processing was completed.
|
|
"predictionOutput": { # Represents results of a prediction job. # The current prediction job result.
|
|
"outputPath": "A String", # The output Google Cloud Storage location provided at the job creation time.
|
|
"nodeHours": 3.14, # Node hours used by the batch prediction job.
|
|
"predictionCount": "A String", # The number of generated predictions.
|
|
"errorCount": "A String", # The number of data instances which resulted in errors.
|
|
},
|
|
"createTime": "A String", # Output only. When the job was created.
|
|
}
|
|
|
|
x__xgafv: string, V1 error format.
|
|
Allowed values
|
|
1 - v1 error format
|
|
2 - v2 error format
|
|
|
|
Returns:
|
|
An object of the form:
|
|
|
|
{ # Represents a training or prediction job.
|
|
"errorMessage": "A String", # Output only. The details of a failure or a cancellation.
|
|
"trainingOutput": { # Represents results of a training job. Output only. # The current training job result.
|
|
"completedTrialCount": "A String", # The number of hyperparameter tuning trials that completed successfully.
|
|
# Only set for hyperparameter tuning jobs.
|
|
"trials": [ # Results for individual Hyperparameter trials.
|
|
# Only set for hyperparameter tuning jobs.
|
|
{ # Represents the result of a single hyperparameter tuning trial from a
|
|
# training job. The TrainingOutput object that is returned on successful
|
|
# completion of a training job with hyperparameter tuning includes a list
|
|
# of HyperparameterOutput objects, one for each successful trial.
|
|
"hyperparameters": { # The hyperparameters given to this trial.
|
|
"a_key": "A String",
|
|
},
|
|
"finalMetric": { # An observed value of a metric. # The final objective metric seen for this trial.
|
|
"trainingStep": "A String", # The global training step for this metric.
|
|
"objectiveValue": 3.14, # The objective value at this training step.
|
|
},
|
|
"allMetrics": [ # All recorded object metrics for this trial. This field is not currently
|
|
# populated.
|
|
{ # An observed value of a metric.
|
|
"trainingStep": "A String", # The global training step for this metric.
|
|
"objectiveValue": 3.14, # The objective value at this training step.
|
|
},
|
|
],
|
|
"isTrialStoppedEarly": True or False, # True if the trial is stopped early.
|
|
"trialId": "A String", # The trial id for these results.
|
|
"builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
|
|
# Only set for trials of built-in algorithms jobs that have succeeded.
|
|
"framework": "A String", # Framework on which the built-in algorithm was trained.
|
|
"modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
|
|
# saves the trained model. Only set for successful jobs that don't use
|
|
# hyperparameter tuning.
|
|
"runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
|
|
# trained.
|
|
"pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
|
|
},
|
|
},
|
|
],
|
|
"isHyperparameterTuningJob": True or False, # Whether this job is a hyperparameter tuning job.
|
|
"isBuiltInAlgorithmJob": True or False, # Whether this job is a built-in Algorithm job.
|
|
"consumedMLUnits": 3.14, # The amount of ML units consumed by the job.
|
|
"hyperparameterMetricTag": "A String", # The TensorFlow summary tag name used for optimizing hyperparameter tuning
|
|
# trials. See
|
|
# [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag)
|
|
# for more information. Only set for hyperparameter tuning jobs.
|
|
"builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
|
|
# Only set for built-in algorithms jobs.
|
|
"framework": "A String", # Framework on which the built-in algorithm was trained.
|
|
"modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
|
|
# saves the trained model. Only set for successful jobs that don't use
|
|
# hyperparameter tuning.
|
|
"runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
|
|
# trained.
|
|
"pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
|
|
},
|
|
},
|
|
"predictionInput": { # Represents input parameters for a prediction job. # Input parameters to create a prediction job.
|
|
"modelName": "A String", # Use this field if you want to use the default version for the specified
|
|
# model. The string must use the following format:
|
|
#
|
|
# `"projects/YOUR_PROJECT/models/YOUR_MODEL"`
|
|
"runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for this batch
|
|
# prediction. If not set, AI Platform will pick the runtime version used
|
|
# during the CreateVersion request for this model version, or choose the
|
|
# latest stable version when model version information is not available
|
|
# such as when the model is specified by uri.
|
|
"signatureName": "A String", # Optional. The name of the signature defined in the SavedModel to use for
|
|
# this job. Please refer to
|
|
# [SavedModel](https://tensorflow.github.io/serving/serving_basic.html)
|
|
# for information about how to use signatures.
|
|
#
|
|
# Defaults to
|
|
# [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants)
|
|
# , which is "serving_default".
|
|
"batchSize": "A String", # Optional. Number of records per batch, defaults to 64.
|
|
# The service will buffer batch_size number of records in memory before
|
|
# invoking one Tensorflow prediction call internally. So take the record
|
|
# size and memory available into consideration when setting this parameter.
|
|
"inputPaths": [ # Required. The Cloud Storage location of the input data files. May contain
|
|
# <a href="/storage/docs/gsutil/addlhelp/WildcardNames">wildcards</a>.
|
|
"A String",
|
|
],
|
|
"maxWorkerCount": "A String", # Optional. The maximum number of workers to be used for parallel processing.
|
|
# Defaults to 10 if not specified.
|
|
"uri": "A String", # Use this field if you want to specify a Google Cloud Storage path for
|
|
# the model to use.
|
|
"outputPath": "A String", # Required. The output Google Cloud Storage location.
|
|
"dataFormat": "A String", # Required. The format of the input data files.
|
|
"versionName": "A String", # Use this field if you want to specify a version of the model to use. The
|
|
# string is formatted the same way as `model_version`, with the addition
|
|
# of the version information:
|
|
#
|
|
# `"projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION"`
|
|
"region": "A String", # Required. The Google Compute Engine region to run the prediction job in.
|
|
# See the <a href="/ml-engine/docs/tensorflow/regions">available regions</a>
|
|
# for AI Platform services.
|
|
"outputDataFormat": "A String", # Optional. Format of the output data files, defaults to JSON.
|
|
},
|
|
"trainingInput": { # Represents input parameters for a training job. When using the # Input parameters to create a training job.
|
|
# gcloud command to submit your training job, you can specify
|
|
# the input parameters as command-line arguments and/or in a YAML configuration
|
|
# file referenced from the --config command-line argument. For
|
|
# details, see the guide to
|
|
# <a href="/ml-engine/docs/tensorflow/training-jobs">submitting a training
|
|
# job</a>.
|
|
"workerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
|
|
# job's worker nodes.
|
|
#
|
|
# The supported values are the same as those described in the entry for
|
|
# `masterType`.
|
|
#
|
|
# This value must be consistent with the category of machine type that
|
|
# `masterType` uses. In other words, both must be AI Platform machine
|
|
# types or both must be Compute Engine machine types.
|
|
#
|
|
# If you use `cloud_tpu` for this value, see special instructions for
|
|
# [configuring a custom TPU
|
|
# machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
|
|
#
|
|
# This value must be present when `scaleTier` is set to `CUSTOM` and
|
|
# `workerCount` is greater than zero.
|
|
"parameterServerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers.
|
|
#
|
|
# You should only set `parameterServerConfig.acceleratorConfig` if
|
|
# `parameterServerConfigType` is set to a Compute Engine machine type. [Learn
|
|
# about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
#
|
|
# Set `parameterServerConfig.imageUri` only if you build a custom image for
|
|
# your parameter server. If `parameterServerConfig.imageUri` has not been
|
|
# set, AI Platform uses the value of `masterConfig.imageUri`.
|
|
# Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
"acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
|
|
# [Learn about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
"count": "A String", # The number of accelerators to attach to each machine running the job.
|
|
"type": "A String", # The type of accelerator to use.
|
|
},
|
|
"imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
|
|
# Registry. Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
},
|
|
"runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for training. If not
|
|
# set, AI Platform uses the default stable version, 1.0. For more
|
|
# information, see the
|
|
# <a href="/ml-engine/docs/runtime-version-list">runtime version list</a>
|
|
# and
|
|
# <a href="/ml-engine/docs/versioning">how to manage runtime versions</a>.
|
|
"scaleTier": "A String", # Required. Specifies the machine types, the number of replicas for workers
|
|
# and parameter servers.
|
|
"masterType": "A String", # Optional. Specifies the type of virtual machine to use for your training
|
|
# job's master worker.
|
|
#
|
|
# The following types are supported:
|
|
#
|
|
# <dl>
|
|
# <dt>standard</dt>
|
|
# <dd>
|
|
# A basic machine configuration suitable for training simple models with
|
|
# small to moderate datasets.
|
|
# </dd>
|
|
# <dt>large_model</dt>
|
|
# <dd>
|
|
# A machine with a lot of memory, specially suited for parameter servers
|
|
# when your model is large (having many hidden layers or layers with very
|
|
# large numbers of nodes).
|
|
# </dd>
|
|
# <dt>complex_model_s</dt>
|
|
# <dd>
|
|
# A machine suitable for the master and workers of the cluster when your
|
|
# model requires more computation than the standard machine can handle
|
|
# satisfactorily.
|
|
# </dd>
|
|
# <dt>complex_model_m</dt>
|
|
# <dd>
|
|
# A machine with roughly twice the number of cores and roughly double the
|
|
# memory of <i>complex_model_s</i>.
|
|
# </dd>
|
|
# <dt>complex_model_l</dt>
|
|
# <dd>
|
|
# A machine with roughly twice the number of cores and roughly double the
|
|
# memory of <i>complex_model_m</i>.
|
|
# </dd>
|
|
# <dt>standard_gpu</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>standard</i> that
|
|
# also includes a single NVIDIA Tesla K80 GPU. See more about
|
|
# <a href="/ml-engine/docs/tensorflow/using-gpus">using GPUs to
|
|
# train your model</a>.
|
|
# </dd>
|
|
# <dt>complex_model_m_gpu</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_m</i> that also includes
|
|
# four NVIDIA Tesla K80 GPUs.
|
|
# </dd>
|
|
# <dt>complex_model_l_gpu</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_l</i> that also includes
|
|
# eight NVIDIA Tesla K80 GPUs.
|
|
# </dd>
|
|
# <dt>standard_p100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>standard</i> that
|
|
# also includes a single NVIDIA Tesla P100 GPU.
|
|
# </dd>
|
|
# <dt>complex_model_m_p100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_m</i> that also includes
|
|
# four NVIDIA Tesla P100 GPUs.
|
|
# </dd>
|
|
# <dt>standard_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>standard</i> that
|
|
# also includes a single NVIDIA Tesla V100 GPU.
|
|
# </dd>
|
|
# <dt>large_model_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>large_model</i> that
|
|
# also includes a single NVIDIA Tesla V100 GPU.
|
|
# </dd>
|
|
# <dt>complex_model_m_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_m</i> that
|
|
# also includes four NVIDIA Tesla V100 GPUs.
|
|
# </dd>
|
|
# <dt>complex_model_l_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_l</i> that
|
|
# also includes eight NVIDIA Tesla V100 GPUs.
|
|
# </dd>
|
|
# <dt>cloud_tpu</dt>
|
|
# <dd>
|
|
# A TPU VM including one Cloud TPU. See more about
|
|
# <a href="/ml-engine/docs/tensorflow/using-tpus">using TPUs to train
|
|
# your model</a>.
|
|
# </dd>
|
|
# </dl>
|
|
#
|
|
# You may also use certain Compute Engine machine types directly in this
|
|
# field. The following types are supported:
|
|
#
|
|
# - `n1-standard-4`
|
|
# - `n1-standard-8`
|
|
# - `n1-standard-16`
|
|
# - `n1-standard-32`
|
|
# - `n1-standard-64`
|
|
# - `n1-standard-96`
|
|
# - `n1-highmem-2`
|
|
# - `n1-highmem-4`
|
|
# - `n1-highmem-8`
|
|
# - `n1-highmem-16`
|
|
# - `n1-highmem-32`
|
|
# - `n1-highmem-64`
|
|
# - `n1-highmem-96`
|
|
# - `n1-highcpu-16`
|
|
# - `n1-highcpu-32`
|
|
# - `n1-highcpu-64`
|
|
# - `n1-highcpu-96`
|
|
#
|
|
# See more about [using Compute Engine machine
|
|
# types](/ml-engine/docs/tensorflow/machine-types#compute-engine-machine-types).
|
|
#
|
|
# You must set this value when `scaleTier` is set to `CUSTOM`.
|
|
"hyperparameters": { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune.
|
|
"maxTrials": 42, # Optional. How many training trials should be attempted to optimize
|
|
# the specified hyperparameters.
|
|
#
|
|
# Defaults to one.
|
|
"goal": "A String", # Required. The type of goal to use for tuning. Available types are
|
|
# `MAXIMIZE` and `MINIMIZE`.
|
|
#
|
|
# Defaults to `MAXIMIZE`.
|
|
"algorithm": "A String", # Optional. The search algorithm specified for the hyperparameter
|
|
# tuning job.
|
|
# Uses the default AI Platform hyperparameter tuning
|
|
# algorithm if unspecified.
|
|
"maxFailedTrials": 42, # Optional. The number of failed trials that need to be seen before failing
|
|
# the hyperparameter tuning job. You can specify this field to override the
|
|
# default failing criteria for AI Platform hyperparameter tuning jobs.
|
|
#
|
|
# Defaults to zero, which means the service decides when a hyperparameter
|
|
# job should fail.
|
|
"enableTrialEarlyStopping": True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial
|
|
# early stopping.
|
|
"resumePreviousJobId": "A String", # Optional. The prior hyperparameter tuning job id that users hope to
|
|
# continue with. The job id will be used to find the corresponding vizier
|
|
# study guid and resume the study.
|
|
"params": [ # Required. The set of parameters to tune.
|
|
{ # Represents a single hyperparameter to optimize.
|
|
"maxValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
|
|
# should be unset if type is `CATEGORICAL`. This value should be integers if
|
|
# type is `INTEGER`.
|
|
"categoricalValues": [ # Required if type is `CATEGORICAL`. The list of possible categories.
|
|
"A String",
|
|
],
|
|
"discreteValues": [ # Required if type is `DISCRETE`.
|
|
# A list of feasible points.
|
|
# The list should be in strictly increasing order. For instance, this
|
|
# parameter might have possible settings of 1.5, 2.5, and 4.0. This list
|
|
# should not contain more than 1,000 values.
|
|
3.14,
|
|
],
|
|
"parameterName": "A String", # Required. The parameter name must be unique amongst all ParameterConfigs in
|
|
# a HyperparameterSpec message. E.g., "learning_rate".
|
|
"minValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
|
|
# should be unset if type is `CATEGORICAL`. This value should be integers if
|
|
# type is INTEGER.
|
|
"type": "A String", # Required. The type of the parameter.
|
|
"scaleType": "A String", # Optional. How the parameter should be scaled to the hypercube.
|
|
# Leave unset for categorical parameters.
|
|
# Some kind of scaling is strongly recommended for real or integral
|
|
# parameters (e.g., `UNIT_LINEAR_SCALE`).
|
|
},
|
|
],
|
|
"hyperparameterMetricTag": "A String", # Optional. The TensorFlow summary tag name to use for optimizing trials. For
|
|
# current versions of TensorFlow, this tag name should exactly match what is
|
|
# shown in TensorBoard, including all scopes. For versions of TensorFlow
|
|
# prior to 0.12, this should be only the tag passed to tf.Summary.
|
|
# By default, "training/hptuning/metric" will be used.
|
|
"maxParallelTrials": 42, # Optional. The number of training trials to run concurrently.
|
|
# You can reduce the time it takes to perform hyperparameter tuning by adding
|
|
# trials in parallel. However, each trail only benefits from the information
|
|
# gained in completed trials. That means that a trial does not get access to
|
|
# the results of trials running at the same time, which could reduce the
|
|
# quality of the overall optimization.
|
|
#
|
|
# Each trial will use the same scale tier and machine types.
|
|
#
|
|
# Defaults to one.
|
|
},
|
|
"region": "A String", # Required. The Google Compute Engine region to run the training job in.
|
|
# See the <a href="/ml-engine/docs/tensorflow/regions">available regions</a>
|
|
# for AI Platform services.
|
|
"args": [ # Optional. Command line arguments to pass to the program.
|
|
"A String",
|
|
],
|
|
"pythonModule": "A String", # Required. The Python module name to run after installing the packages.
|
|
"pythonVersion": "A String", # Optional. The version of Python used in training. If not set, the default
|
|
# version is '2.7'. Python '3.5' is available when `runtime_version` is set
|
|
# to '1.4' and above. Python '2.7' works with all supported
|
|
# <a href="/ml-engine/docs/runtime-version-list">runtime versions</a>.
|
|
"jobDir": "A String", # Optional. A Google Cloud Storage path in which to store training outputs
|
|
# and other data needed for training. This path is passed to your TensorFlow
|
|
# program as the '--job-dir' command-line argument. The benefit of specifying
|
|
# this field is that Cloud ML validates the path for use in training.
|
|
"packageUris": [ # Required. The Google Cloud Storage location of the packages with
|
|
# the training program and any additional dependencies.
|
|
# The maximum number of package URIs is 100.
|
|
"A String",
|
|
],
|
|
"workerCount": "A String", # Optional. The number of worker replicas to use for the training job. Each
|
|
# replica in the cluster will be of the type specified in `worker_type`.
|
|
#
|
|
# This value can only be used when `scale_tier` is set to `CUSTOM`. If you
|
|
# set this value, you must also set `worker_type`.
|
|
#
|
|
# The default value is zero.
|
|
"parameterServerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
|
|
# job's parameter server.
|
|
#
|
|
# The supported values are the same as those described in the entry for
|
|
# `master_type`.
|
|
#
|
|
# This value must be consistent with the category of machine type that
|
|
# `masterType` uses. In other words, both must be AI Platform machine
|
|
# types or both must be Compute Engine machine types.
|
|
#
|
|
# This value must be present when `scaleTier` is set to `CUSTOM` and
|
|
# `parameter_server_count` is greater than zero.
|
|
"workerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers.
|
|
#
|
|
# You should only set `workerConfig.acceleratorConfig` if `workerType` is set
|
|
# to a Compute Engine machine type. [Learn about restrictions on accelerator
|
|
# configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
#
|
|
# Set `workerConfig.imageUri` only if you build a custom image for your
|
|
# worker. If `workerConfig.imageUri` has not been set, AI Platform uses
|
|
# the value of `masterConfig.imageUri`. Learn more about
|
|
# [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
"acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
|
|
# [Learn about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
"count": "A String", # The number of accelerators to attach to each machine running the job.
|
|
"type": "A String", # The type of accelerator to use.
|
|
},
|
|
"imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
|
|
# Registry. Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
},
|
|
"maxRunningTime": "A String", # Optional. The maximum job running time. The default is 7 days.
|
|
"masterConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker.
|
|
#
|
|
# You should only set `masterConfig.acceleratorConfig` if `masterType` is set
|
|
# to a Compute Engine machine type. Learn about [restrictions on accelerator
|
|
# configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
#
|
|
# Set `masterConfig.imageUri` only if you build a custom image. Only one of
|
|
# `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more about
|
|
# [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
"acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
|
|
# [Learn about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
"count": "A String", # The number of accelerators to attach to each machine running the job.
|
|
"type": "A String", # The type of accelerator to use.
|
|
},
|
|
"imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
|
|
# Registry. Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
},
|
|
"parameterServerCount": "A String", # Optional. The number of parameter server replicas to use for the training
|
|
# job. Each replica in the cluster will be of the type specified in
|
|
# `parameter_server_type`.
|
|
#
|
|
# This value can only be used when `scale_tier` is set to `CUSTOM`.If you
|
|
# set this value, you must also set `parameter_server_type`.
|
|
#
|
|
# The default value is zero.
|
|
},
|
|
"jobId": "A String", # Required. The user-specified id of the job.
|
|
"labels": { # Optional. One or more labels that you can add, to organize your jobs.
|
|
# Each label is a key-value pair, where both the key and the value are
|
|
# arbitrary strings that you supply.
|
|
# For more information, see the documentation on
|
|
# <a href="/ml-engine/docs/tensorflow/resource-labels">using labels</a>.
|
|
"a_key": "A String",
|
|
},
|
|
"state": "A String", # Output only. The detailed state of a job.
|
|
"etag": "A String", # `etag` is used for optimistic concurrency control as a way to help
|
|
# prevent simultaneous updates of a job from overwriting each other.
|
|
# It is strongly suggested that systems make use of the `etag` in the
|
|
# read-modify-write cycle to perform job updates in order to avoid race
|
|
# conditions: An `etag` is returned in the response to `GetJob`, and
|
|
# systems are expected to put that etag in the request to `UpdateJob` to
|
|
# ensure that their change will be applied to the same version of the job.
|
|
"startTime": "A String", # Output only. When the job processing was started.
|
|
"endTime": "A String", # Output only. When the job processing was completed.
|
|
"predictionOutput": { # Represents results of a prediction job. # The current prediction job result.
|
|
"outputPath": "A String", # The output Google Cloud Storage location provided at the job creation time.
|
|
"nodeHours": 3.14, # Node hours used by the batch prediction job.
|
|
"predictionCount": "A String", # The number of generated predictions.
|
|
"errorCount": "A String", # The number of data instances which resulted in errors.
|
|
},
|
|
"createTime": "A String", # Output only. When the job was created.
|
|
}</pre>
|
|
</div>
|
|
|
|
<div class="method">
|
|
<code class="details" id="get">get(name, x__xgafv=None)</code>
|
|
<pre>Describes a job.
|
|
|
|
Args:
|
|
name: string, Required. The name of the job to get the description of. (required)
|
|
x__xgafv: string, V1 error format.
|
|
Allowed values
|
|
1 - v1 error format
|
|
2 - v2 error format
|
|
|
|
Returns:
|
|
An object of the form:
|
|
|
|
{ # Represents a training or prediction job.
|
|
"errorMessage": "A String", # Output only. The details of a failure or a cancellation.
|
|
"trainingOutput": { # Represents results of a training job. Output only. # The current training job result.
|
|
"completedTrialCount": "A String", # The number of hyperparameter tuning trials that completed successfully.
|
|
# Only set for hyperparameter tuning jobs.
|
|
"trials": [ # Results for individual Hyperparameter trials.
|
|
# Only set for hyperparameter tuning jobs.
|
|
{ # Represents the result of a single hyperparameter tuning trial from a
|
|
# training job. The TrainingOutput object that is returned on successful
|
|
# completion of a training job with hyperparameter tuning includes a list
|
|
# of HyperparameterOutput objects, one for each successful trial.
|
|
"hyperparameters": { # The hyperparameters given to this trial.
|
|
"a_key": "A String",
|
|
},
|
|
"finalMetric": { # An observed value of a metric. # The final objective metric seen for this trial.
|
|
"trainingStep": "A String", # The global training step for this metric.
|
|
"objectiveValue": 3.14, # The objective value at this training step.
|
|
},
|
|
"allMetrics": [ # All recorded object metrics for this trial. This field is not currently
|
|
# populated.
|
|
{ # An observed value of a metric.
|
|
"trainingStep": "A String", # The global training step for this metric.
|
|
"objectiveValue": 3.14, # The objective value at this training step.
|
|
},
|
|
],
|
|
"isTrialStoppedEarly": True or False, # True if the trial is stopped early.
|
|
"trialId": "A String", # The trial id for these results.
|
|
"builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
|
|
# Only set for trials of built-in algorithms jobs that have succeeded.
|
|
"framework": "A String", # Framework on which the built-in algorithm was trained.
|
|
"modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
|
|
# saves the trained model. Only set for successful jobs that don't use
|
|
# hyperparameter tuning.
|
|
"runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
|
|
# trained.
|
|
"pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
|
|
},
|
|
},
|
|
],
|
|
"isHyperparameterTuningJob": True or False, # Whether this job is a hyperparameter tuning job.
|
|
"isBuiltInAlgorithmJob": True or False, # Whether this job is a built-in Algorithm job.
|
|
"consumedMLUnits": 3.14, # The amount of ML units consumed by the job.
|
|
"hyperparameterMetricTag": "A String", # The TensorFlow summary tag name used for optimizing hyperparameter tuning
|
|
# trials. See
|
|
# [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag)
|
|
# for more information. Only set for hyperparameter tuning jobs.
|
|
"builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
|
|
# Only set for built-in algorithms jobs.
|
|
"framework": "A String", # Framework on which the built-in algorithm was trained.
|
|
"modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
|
|
# saves the trained model. Only set for successful jobs that don't use
|
|
# hyperparameter tuning.
|
|
"runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
|
|
# trained.
|
|
"pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
|
|
},
|
|
},
|
|
"predictionInput": { # Represents input parameters for a prediction job. # Input parameters to create a prediction job.
|
|
"modelName": "A String", # Use this field if you want to use the default version for the specified
|
|
# model. The string must use the following format:
|
|
#
|
|
# `"projects/YOUR_PROJECT/models/YOUR_MODEL"`
|
|
"runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for this batch
|
|
# prediction. If not set, AI Platform will pick the runtime version used
|
|
# during the CreateVersion request for this model version, or choose the
|
|
# latest stable version when model version information is not available
|
|
# such as when the model is specified by uri.
|
|
"signatureName": "A String", # Optional. The name of the signature defined in the SavedModel to use for
|
|
# this job. Please refer to
|
|
# [SavedModel](https://tensorflow.github.io/serving/serving_basic.html)
|
|
# for information about how to use signatures.
|
|
#
|
|
# Defaults to
|
|
# [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants)
|
|
# , which is "serving_default".
|
|
"batchSize": "A String", # Optional. Number of records per batch, defaults to 64.
|
|
# The service will buffer batch_size number of records in memory before
|
|
# invoking one Tensorflow prediction call internally. So take the record
|
|
# size and memory available into consideration when setting this parameter.
|
|
"inputPaths": [ # Required. The Cloud Storage location of the input data files. May contain
|
|
# <a href="/storage/docs/gsutil/addlhelp/WildcardNames">wildcards</a>.
|
|
"A String",
|
|
],
|
|
"maxWorkerCount": "A String", # Optional. The maximum number of workers to be used for parallel processing.
|
|
# Defaults to 10 if not specified.
|
|
"uri": "A String", # Use this field if you want to specify a Google Cloud Storage path for
|
|
# the model to use.
|
|
"outputPath": "A String", # Required. The output Google Cloud Storage location.
|
|
"dataFormat": "A String", # Required. The format of the input data files.
|
|
"versionName": "A String", # Use this field if you want to specify a version of the model to use. The
|
|
# string is formatted the same way as `model_version`, with the addition
|
|
# of the version information:
|
|
#
|
|
# `"projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION"`
|
|
"region": "A String", # Required. The Google Compute Engine region to run the prediction job in.
|
|
# See the <a href="/ml-engine/docs/tensorflow/regions">available regions</a>
|
|
# for AI Platform services.
|
|
"outputDataFormat": "A String", # Optional. Format of the output data files, defaults to JSON.
|
|
},
|
|
"trainingInput": { # Represents input parameters for a training job. When using the # Input parameters to create a training job.
|
|
# gcloud command to submit your training job, you can specify
|
|
# the input parameters as command-line arguments and/or in a YAML configuration
|
|
# file referenced from the --config command-line argument. For
|
|
# details, see the guide to
|
|
# <a href="/ml-engine/docs/tensorflow/training-jobs">submitting a training
|
|
# job</a>.
|
|
"workerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
|
|
# job's worker nodes.
|
|
#
|
|
# The supported values are the same as those described in the entry for
|
|
# `masterType`.
|
|
#
|
|
# This value must be consistent with the category of machine type that
|
|
# `masterType` uses. In other words, both must be AI Platform machine
|
|
# types or both must be Compute Engine machine types.
|
|
#
|
|
# If you use `cloud_tpu` for this value, see special instructions for
|
|
# [configuring a custom TPU
|
|
# machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
|
|
#
|
|
# This value must be present when `scaleTier` is set to `CUSTOM` and
|
|
# `workerCount` is greater than zero.
|
|
"parameterServerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers.
|
|
#
|
|
# You should only set `parameterServerConfig.acceleratorConfig` if
|
|
# `parameterServerConfigType` is set to a Compute Engine machine type. [Learn
|
|
# about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
#
|
|
# Set `parameterServerConfig.imageUri` only if you build a custom image for
|
|
# your parameter server. If `parameterServerConfig.imageUri` has not been
|
|
# set, AI Platform uses the value of `masterConfig.imageUri`.
|
|
# Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
"acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
|
|
# [Learn about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
"count": "A String", # The number of accelerators to attach to each machine running the job.
|
|
"type": "A String", # The type of accelerator to use.
|
|
},
|
|
"imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
|
|
# Registry. Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
},
|
|
"runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for training. If not
|
|
# set, AI Platform uses the default stable version, 1.0. For more
|
|
# information, see the
|
|
# <a href="/ml-engine/docs/runtime-version-list">runtime version list</a>
|
|
# and
|
|
# <a href="/ml-engine/docs/versioning">how to manage runtime versions</a>.
|
|
"scaleTier": "A String", # Required. Specifies the machine types, the number of replicas for workers
|
|
# and parameter servers.
|
|
"masterType": "A String", # Optional. Specifies the type of virtual machine to use for your training
|
|
# job's master worker.
|
|
#
|
|
# The following types are supported:
|
|
#
|
|
# <dl>
|
|
# <dt>standard</dt>
|
|
# <dd>
|
|
# A basic machine configuration suitable for training simple models with
|
|
# small to moderate datasets.
|
|
# </dd>
|
|
# <dt>large_model</dt>
|
|
# <dd>
|
|
# A machine with a lot of memory, specially suited for parameter servers
|
|
# when your model is large (having many hidden layers or layers with very
|
|
# large numbers of nodes).
|
|
# </dd>
|
|
# <dt>complex_model_s</dt>
|
|
# <dd>
|
|
# A machine suitable for the master and workers of the cluster when your
|
|
# model requires more computation than the standard machine can handle
|
|
# satisfactorily.
|
|
# </dd>
|
|
# <dt>complex_model_m</dt>
|
|
# <dd>
|
|
# A machine with roughly twice the number of cores and roughly double the
|
|
# memory of <i>complex_model_s</i>.
|
|
# </dd>
|
|
# <dt>complex_model_l</dt>
|
|
# <dd>
|
|
# A machine with roughly twice the number of cores and roughly double the
|
|
# memory of <i>complex_model_m</i>.
|
|
# </dd>
|
|
# <dt>standard_gpu</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>standard</i> that
|
|
# also includes a single NVIDIA Tesla K80 GPU. See more about
|
|
# <a href="/ml-engine/docs/tensorflow/using-gpus">using GPUs to
|
|
# train your model</a>.
|
|
# </dd>
|
|
# <dt>complex_model_m_gpu</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_m</i> that also includes
|
|
# four NVIDIA Tesla K80 GPUs.
|
|
# </dd>
|
|
# <dt>complex_model_l_gpu</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_l</i> that also includes
|
|
# eight NVIDIA Tesla K80 GPUs.
|
|
# </dd>
|
|
# <dt>standard_p100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>standard</i> that
|
|
# also includes a single NVIDIA Tesla P100 GPU.
|
|
# </dd>
|
|
# <dt>complex_model_m_p100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_m</i> that also includes
|
|
# four NVIDIA Tesla P100 GPUs.
|
|
# </dd>
|
|
# <dt>standard_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>standard</i> that
|
|
# also includes a single NVIDIA Tesla V100 GPU.
|
|
# </dd>
|
|
# <dt>large_model_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>large_model</i> that
|
|
# also includes a single NVIDIA Tesla V100 GPU.
|
|
# </dd>
|
|
# <dt>complex_model_m_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_m</i> that
|
|
# also includes four NVIDIA Tesla V100 GPUs.
|
|
# </dd>
|
|
# <dt>complex_model_l_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_l</i> that
|
|
# also includes eight NVIDIA Tesla V100 GPUs.
|
|
# </dd>
|
|
# <dt>cloud_tpu</dt>
|
|
# <dd>
|
|
# A TPU VM including one Cloud TPU. See more about
|
|
# <a href="/ml-engine/docs/tensorflow/using-tpus">using TPUs to train
|
|
# your model</a>.
|
|
# </dd>
|
|
# </dl>
|
|
#
|
|
# You may also use certain Compute Engine machine types directly in this
|
|
# field. The following types are supported:
|
|
#
|
|
# - `n1-standard-4`
|
|
# - `n1-standard-8`
|
|
# - `n1-standard-16`
|
|
# - `n1-standard-32`
|
|
# - `n1-standard-64`
|
|
# - `n1-standard-96`
|
|
# - `n1-highmem-2`
|
|
# - `n1-highmem-4`
|
|
# - `n1-highmem-8`
|
|
# - `n1-highmem-16`
|
|
# - `n1-highmem-32`
|
|
# - `n1-highmem-64`
|
|
# - `n1-highmem-96`
|
|
# - `n1-highcpu-16`
|
|
# - `n1-highcpu-32`
|
|
# - `n1-highcpu-64`
|
|
# - `n1-highcpu-96`
|
|
#
|
|
# See more about [using Compute Engine machine
|
|
# types](/ml-engine/docs/tensorflow/machine-types#compute-engine-machine-types).
|
|
#
|
|
# You must set this value when `scaleTier` is set to `CUSTOM`.
|
|
"hyperparameters": { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune.
|
|
"maxTrials": 42, # Optional. How many training trials should be attempted to optimize
|
|
# the specified hyperparameters.
|
|
#
|
|
# Defaults to one.
|
|
"goal": "A String", # Required. The type of goal to use for tuning. Available types are
|
|
# `MAXIMIZE` and `MINIMIZE`.
|
|
#
|
|
# Defaults to `MAXIMIZE`.
|
|
"algorithm": "A String", # Optional. The search algorithm specified for the hyperparameter
|
|
# tuning job.
|
|
# Uses the default AI Platform hyperparameter tuning
|
|
# algorithm if unspecified.
|
|
"maxFailedTrials": 42, # Optional. The number of failed trials that need to be seen before failing
|
|
# the hyperparameter tuning job. You can specify this field to override the
|
|
# default failing criteria for AI Platform hyperparameter tuning jobs.
|
|
#
|
|
# Defaults to zero, which means the service decides when a hyperparameter
|
|
# job should fail.
|
|
"enableTrialEarlyStopping": True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial
|
|
# early stopping.
|
|
"resumePreviousJobId": "A String", # Optional. The prior hyperparameter tuning job id that users hope to
|
|
# continue with. The job id will be used to find the corresponding vizier
|
|
# study guid and resume the study.
|
|
"params": [ # Required. The set of parameters to tune.
|
|
{ # Represents a single hyperparameter to optimize.
|
|
"maxValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
|
|
# should be unset if type is `CATEGORICAL`. This value should be integers if
|
|
# type is `INTEGER`.
|
|
"categoricalValues": [ # Required if type is `CATEGORICAL`. The list of possible categories.
|
|
"A String",
|
|
],
|
|
"discreteValues": [ # Required if type is `DISCRETE`.
|
|
# A list of feasible points.
|
|
# The list should be in strictly increasing order. For instance, this
|
|
# parameter might have possible settings of 1.5, 2.5, and 4.0. This list
|
|
# should not contain more than 1,000 values.
|
|
3.14,
|
|
],
|
|
"parameterName": "A String", # Required. The parameter name must be unique amongst all ParameterConfigs in
|
|
# a HyperparameterSpec message. E.g., "learning_rate".
|
|
"minValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
|
|
# should be unset if type is `CATEGORICAL`. This value should be integers if
|
|
# type is INTEGER.
|
|
"type": "A String", # Required. The type of the parameter.
|
|
"scaleType": "A String", # Optional. How the parameter should be scaled to the hypercube.
|
|
# Leave unset for categorical parameters.
|
|
# Some kind of scaling is strongly recommended for real or integral
|
|
# parameters (e.g., `UNIT_LINEAR_SCALE`).
|
|
},
|
|
],
|
|
"hyperparameterMetricTag": "A String", # Optional. The TensorFlow summary tag name to use for optimizing trials. For
|
|
# current versions of TensorFlow, this tag name should exactly match what is
|
|
# shown in TensorBoard, including all scopes. For versions of TensorFlow
|
|
# prior to 0.12, this should be only the tag passed to tf.Summary.
|
|
# By default, "training/hptuning/metric" will be used.
|
|
"maxParallelTrials": 42, # Optional. The number of training trials to run concurrently.
|
|
# You can reduce the time it takes to perform hyperparameter tuning by adding
|
|
# trials in parallel. However, each trail only benefits from the information
|
|
# gained in completed trials. That means that a trial does not get access to
|
|
# the results of trials running at the same time, which could reduce the
|
|
# quality of the overall optimization.
|
|
#
|
|
# Each trial will use the same scale tier and machine types.
|
|
#
|
|
# Defaults to one.
|
|
},
|
|
"region": "A String", # Required. The Google Compute Engine region to run the training job in.
|
|
# See the <a href="/ml-engine/docs/tensorflow/regions">available regions</a>
|
|
# for AI Platform services.
|
|
"args": [ # Optional. Command line arguments to pass to the program.
|
|
"A String",
|
|
],
|
|
"pythonModule": "A String", # Required. The Python module name to run after installing the packages.
|
|
"pythonVersion": "A String", # Optional. The version of Python used in training. If not set, the default
|
|
# version is '2.7'. Python '3.5' is available when `runtime_version` is set
|
|
# to '1.4' and above. Python '2.7' works with all supported
|
|
# <a href="/ml-engine/docs/runtime-version-list">runtime versions</a>.
|
|
"jobDir": "A String", # Optional. A Google Cloud Storage path in which to store training outputs
|
|
# and other data needed for training. This path is passed to your TensorFlow
|
|
# program as the '--job-dir' command-line argument. The benefit of specifying
|
|
# this field is that Cloud ML validates the path for use in training.
|
|
"packageUris": [ # Required. The Google Cloud Storage location of the packages with
|
|
# the training program and any additional dependencies.
|
|
# The maximum number of package URIs is 100.
|
|
"A String",
|
|
],
|
|
"workerCount": "A String", # Optional. The number of worker replicas to use for the training job. Each
|
|
# replica in the cluster will be of the type specified in `worker_type`.
|
|
#
|
|
# This value can only be used when `scale_tier` is set to `CUSTOM`. If you
|
|
# set this value, you must also set `worker_type`.
|
|
#
|
|
# The default value is zero.
|
|
"parameterServerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
|
|
# job's parameter server.
|
|
#
|
|
# The supported values are the same as those described in the entry for
|
|
# `master_type`.
|
|
#
|
|
# This value must be consistent with the category of machine type that
|
|
# `masterType` uses. In other words, both must be AI Platform machine
|
|
# types or both must be Compute Engine machine types.
|
|
#
|
|
# This value must be present when `scaleTier` is set to `CUSTOM` and
|
|
# `parameter_server_count` is greater than zero.
|
|
"workerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers.
|
|
#
|
|
# You should only set `workerConfig.acceleratorConfig` if `workerType` is set
|
|
# to a Compute Engine machine type. [Learn about restrictions on accelerator
|
|
# configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
#
|
|
# Set `workerConfig.imageUri` only if you build a custom image for your
|
|
# worker. If `workerConfig.imageUri` has not been set, AI Platform uses
|
|
# the value of `masterConfig.imageUri`. Learn more about
|
|
# [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
"acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
|
|
# [Learn about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
"count": "A String", # The number of accelerators to attach to each machine running the job.
|
|
"type": "A String", # The type of accelerator to use.
|
|
},
|
|
"imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
|
|
# Registry. Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
},
|
|
"maxRunningTime": "A String", # Optional. The maximum job running time. The default is 7 days.
|
|
"masterConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker.
|
|
#
|
|
# You should only set `masterConfig.acceleratorConfig` if `masterType` is set
|
|
# to a Compute Engine machine type. Learn about [restrictions on accelerator
|
|
# configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
#
|
|
# Set `masterConfig.imageUri` only if you build a custom image. Only one of
|
|
# `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more about
|
|
# [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
"acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
|
|
# [Learn about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
"count": "A String", # The number of accelerators to attach to each machine running the job.
|
|
"type": "A String", # The type of accelerator to use.
|
|
},
|
|
"imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
|
|
# Registry. Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
},
|
|
"parameterServerCount": "A String", # Optional. The number of parameter server replicas to use for the training
|
|
# job. Each replica in the cluster will be of the type specified in
|
|
# `parameter_server_type`.
|
|
#
|
|
# This value can only be used when `scale_tier` is set to `CUSTOM`.If you
|
|
# set this value, you must also set `parameter_server_type`.
|
|
#
|
|
# The default value is zero.
|
|
},
|
|
"jobId": "A String", # Required. The user-specified id of the job.
|
|
"labels": { # Optional. One or more labels that you can add, to organize your jobs.
|
|
# Each label is a key-value pair, where both the key and the value are
|
|
# arbitrary strings that you supply.
|
|
# For more information, see the documentation on
|
|
# <a href="/ml-engine/docs/tensorflow/resource-labels">using labels</a>.
|
|
"a_key": "A String",
|
|
},
|
|
"state": "A String", # Output only. The detailed state of a job.
|
|
"etag": "A String", # `etag` is used for optimistic concurrency control as a way to help
|
|
# prevent simultaneous updates of a job from overwriting each other.
|
|
# It is strongly suggested that systems make use of the `etag` in the
|
|
# read-modify-write cycle to perform job updates in order to avoid race
|
|
# conditions: An `etag` is returned in the response to `GetJob`, and
|
|
# systems are expected to put that etag in the request to `UpdateJob` to
|
|
# ensure that their change will be applied to the same version of the job.
|
|
"startTime": "A String", # Output only. When the job processing was started.
|
|
"endTime": "A String", # Output only. When the job processing was completed.
|
|
"predictionOutput": { # Represents results of a prediction job. # The current prediction job result.
|
|
"outputPath": "A String", # The output Google Cloud Storage location provided at the job creation time.
|
|
"nodeHours": 3.14, # Node hours used by the batch prediction job.
|
|
"predictionCount": "A String", # The number of generated predictions.
|
|
"errorCount": "A String", # The number of data instances which resulted in errors.
|
|
},
|
|
"createTime": "A String", # Output only. When the job was created.
|
|
}</pre>
|
|
</div>
|
|
|
|
<div class="method">
|
|
<code class="details" id="getIamPolicy">getIamPolicy(resource, x__xgafv=None)</code>
|
|
<pre>Gets the access control policy for a resource.
|
|
Returns an empty policy if the resource exists and does not have a policy
|
|
set.
|
|
|
|
Args:
|
|
resource: string, REQUIRED: The resource for which the policy is being requested.
|
|
See the operation documentation for the appropriate value for this field. (required)
|
|
x__xgafv: string, V1 error format.
|
|
Allowed values
|
|
1 - v1 error format
|
|
2 - v2 error format
|
|
|
|
Returns:
|
|
An object of the form:
|
|
|
|
{ # Defines an Identity and Access Management (IAM) policy. It is used to
|
|
# specify access control policies for Cloud Platform resources.
|
|
#
|
|
#
|
|
# A `Policy` consists of a list of `bindings`. A `binding` binds a list of
|
|
# `members` to a `role`, where the members can be user accounts, Google groups,
|
|
# Google domains, and service accounts. A `role` is a named list of permissions
|
|
# defined by IAM.
|
|
#
|
|
# **JSON Example**
|
|
#
|
|
# {
|
|
# "bindings": [
|
|
# {
|
|
# "role": "roles/owner",
|
|
# "members": [
|
|
# "user:mike@example.com",
|
|
# "group:admins@example.com",
|
|
# "domain:google.com",
|
|
# "serviceAccount:my-other-app@appspot.gserviceaccount.com"
|
|
# ]
|
|
# },
|
|
# {
|
|
# "role": "roles/viewer",
|
|
# "members": ["user:sean@example.com"]
|
|
# }
|
|
# ]
|
|
# }
|
|
#
|
|
# **YAML Example**
|
|
#
|
|
# bindings:
|
|
# - members:
|
|
# - user:mike@example.com
|
|
# - group:admins@example.com
|
|
# - domain:google.com
|
|
# - serviceAccount:my-other-app@appspot.gserviceaccount.com
|
|
# role: roles/owner
|
|
# - members:
|
|
# - user:sean@example.com
|
|
# role: roles/viewer
|
|
#
|
|
#
|
|
# For a description of IAM and its features, see the
|
|
# [IAM developer's guide](https://cloud.google.com/iam/docs).
|
|
"bindings": [ # Associates a list of `members` to a `role`.
|
|
# `bindings` with no members will result in an error.
|
|
{ # Associates `members` with a `role`.
|
|
"role": "A String", # Role that is assigned to `members`.
|
|
# For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
|
|
"members": [ # Specifies the identities requesting access for a Cloud Platform resource.
|
|
# `members` can have the following values:
|
|
#
|
|
# * `allUsers`: A special identifier that represents anyone who is
|
|
# on the internet; with or without a Google account.
|
|
#
|
|
# * `allAuthenticatedUsers`: A special identifier that represents anyone
|
|
# who is authenticated with a Google account or a service account.
|
|
#
|
|
# * `user:{emailid}`: An email address that represents a specific Google
|
|
# account. For example, `alice@gmail.com` .
|
|
#
|
|
#
|
|
# * `serviceAccount:{emailid}`: An email address that represents a service
|
|
# account. For example, `my-other-app@appspot.gserviceaccount.com`.
|
|
#
|
|
# * `group:{emailid}`: An email address that represents a Google group.
|
|
# For example, `admins@example.com`.
|
|
#
|
|
#
|
|
# * `domain:{domain}`: The G Suite domain (primary) that represents all the
|
|
# users of that domain. For example, `google.com` or `example.com`.
|
|
#
|
|
"A String",
|
|
],
|
|
"condition": { # Represents an expression text. Example: # The condition that is associated with this binding.
|
|
# NOTE: An unsatisfied condition will not allow user access via current
|
|
# binding. Different bindings, including their conditions, are examined
|
|
# independently.
|
|
#
|
|
# title: "User account presence"
|
|
# description: "Determines whether the request has a user account"
|
|
# expression: "size(request.user) > 0"
|
|
"description": "A String", # An optional description of the expression. This is a longer text which
|
|
# describes the expression, e.g. when hovered over it in a UI.
|
|
"expression": "A String", # Textual representation of an expression in
|
|
# Common Expression Language syntax.
|
|
#
|
|
# The application context of the containing message determines which
|
|
# well-known feature set of CEL is supported.
|
|
"location": "A String", # An optional string indicating the location of the expression for error
|
|
# reporting, e.g. a file name and a position in the file.
|
|
"title": "A String", # An optional title for the expression, i.e. a short string describing
|
|
# its purpose. This can be used e.g. in UIs which allow to enter the
|
|
# expression.
|
|
},
|
|
},
|
|
],
|
|
"etag": "A String", # `etag` is used for optimistic concurrency control as a way to help
|
|
# prevent simultaneous updates of a policy from overwriting each other.
|
|
# It is strongly suggested that systems make use of the `etag` in the
|
|
# read-modify-write cycle to perform policy updates in order to avoid race
|
|
# conditions: An `etag` is returned in the response to `getIamPolicy`, and
|
|
# systems are expected to put that etag in the request to `setIamPolicy` to
|
|
# ensure that their change will be applied to the same version of the policy.
|
|
#
|
|
# If no `etag` is provided in the call to `setIamPolicy`, then the existing
|
|
# policy is overwritten blindly.
|
|
"version": 42, # Deprecated.
|
|
"auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
|
|
{ # Specifies the audit configuration for a service.
|
|
# The configuration determines which permission types are logged, and what
|
|
# identities, if any, are exempted from logging.
|
|
# An AuditConfig must have one or more AuditLogConfigs.
|
|
#
|
|
# If there are AuditConfigs for both `allServices` and a specific service,
|
|
# the union of the two AuditConfigs is used for that service: the log_types
|
|
# specified in each AuditConfig are enabled, and the exempted_members in each
|
|
# AuditLogConfig are exempted.
|
|
#
|
|
# Example Policy with multiple AuditConfigs:
|
|
#
|
|
# {
|
|
# "audit_configs": [
|
|
# {
|
|
# "service": "allServices"
|
|
# "audit_log_configs": [
|
|
# {
|
|
# "log_type": "DATA_READ",
|
|
# "exempted_members": [
|
|
# "user:foo@gmail.com"
|
|
# ]
|
|
# },
|
|
# {
|
|
# "log_type": "DATA_WRITE",
|
|
# },
|
|
# {
|
|
# "log_type": "ADMIN_READ",
|
|
# }
|
|
# ]
|
|
# },
|
|
# {
|
|
# "service": "fooservice.googleapis.com"
|
|
# "audit_log_configs": [
|
|
# {
|
|
# "log_type": "DATA_READ",
|
|
# },
|
|
# {
|
|
# "log_type": "DATA_WRITE",
|
|
# "exempted_members": [
|
|
# "user:bar@gmail.com"
|
|
# ]
|
|
# }
|
|
# ]
|
|
# }
|
|
# ]
|
|
# }
|
|
#
|
|
# For fooservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ
|
|
# logging. It also exempts foo@gmail.com from DATA_READ logging, and
|
|
# bar@gmail.com from DATA_WRITE logging.
|
|
"auditLogConfigs": [ # The configuration for logging of each type of permission.
|
|
{ # Provides the configuration for logging a type of permissions.
|
|
# Example:
|
|
#
|
|
# {
|
|
# "audit_log_configs": [
|
|
# {
|
|
# "log_type": "DATA_READ",
|
|
# "exempted_members": [
|
|
# "user:foo@gmail.com"
|
|
# ]
|
|
# },
|
|
# {
|
|
# "log_type": "DATA_WRITE",
|
|
# }
|
|
# ]
|
|
# }
|
|
#
|
|
# This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting
|
|
# foo@gmail.com from DATA_READ logging.
|
|
"exemptedMembers": [ # Specifies the identities that do not cause logging for this type of
|
|
# permission.
|
|
# Follows the same format of Binding.members.
|
|
"A String",
|
|
],
|
|
"logType": "A String", # The log type that this config enables.
|
|
},
|
|
],
|
|
"service": "A String", # Specifies a service that will be enabled for audit logging.
|
|
# For example, `storage.googleapis.com`, `cloudsql.googleapis.com`.
|
|
# `allServices` is a special value that covers all services.
|
|
},
|
|
],
|
|
}</pre>
|
|
</div>
|
|
|
|
<div class="method">
|
|
<code class="details" id="list">list(parent, pageToken=None, x__xgafv=None, pageSize=None, filter=None)</code>
|
|
<pre>Lists the jobs in the project.
|
|
|
|
If there are no jobs that match the request parameters, the list
|
|
request returns an empty response body: {}.
|
|
|
|
Args:
|
|
parent: string, Required. The name of the project for which to list jobs. (required)
|
|
pageToken: string, Optional. A page token to request the next page of results.
|
|
|
|
You get the token from the `next_page_token` field of the response from
|
|
the previous call.
|
|
x__xgafv: string, V1 error format.
|
|
Allowed values
|
|
1 - v1 error format
|
|
2 - v2 error format
|
|
pageSize: integer, Optional. The number of jobs to retrieve per "page" of results. If there
|
|
are more remaining results than this number, the response message will
|
|
contain a valid value in the `next_page_token` field.
|
|
|
|
The default value is 20, and the maximum page size is 100.
|
|
filter: string, Optional. Specifies the subset of jobs to retrieve.
|
|
You can filter on the value of one or more attributes of the job object.
|
|
For example, retrieve jobs with a job identifier that starts with 'census':
|
|
<p><code>gcloud ai-platform jobs list --filter='jobId:census*'</code>
|
|
<p>List all failed jobs with names that start with 'rnn':
|
|
<p><code>gcloud ai-platform jobs list --filter='jobId:rnn*
|
|
AND state:FAILED'</code>
|
|
<p>For more examples, see the guide to
|
|
<a href="/ml-engine/docs/tensorflow/monitor-training">monitoring jobs</a>.
|
|
|
|
Returns:
|
|
An object of the form:
|
|
|
|
{ # Response message for the ListJobs method.
|
|
"nextPageToken": "A String", # Optional. Pass this token as the `page_token` field of the request for a
|
|
# subsequent call.
|
|
"jobs": [ # The list of jobs.
|
|
{ # Represents a training or prediction job.
|
|
"errorMessage": "A String", # Output only. The details of a failure or a cancellation.
|
|
"trainingOutput": { # Represents results of a training job. Output only. # The current training job result.
|
|
"completedTrialCount": "A String", # The number of hyperparameter tuning trials that completed successfully.
|
|
# Only set for hyperparameter tuning jobs.
|
|
"trials": [ # Results for individual Hyperparameter trials.
|
|
# Only set for hyperparameter tuning jobs.
|
|
{ # Represents the result of a single hyperparameter tuning trial from a
|
|
# training job. The TrainingOutput object that is returned on successful
|
|
# completion of a training job with hyperparameter tuning includes a list
|
|
# of HyperparameterOutput objects, one for each successful trial.
|
|
"hyperparameters": { # The hyperparameters given to this trial.
|
|
"a_key": "A String",
|
|
},
|
|
"finalMetric": { # An observed value of a metric. # The final objective metric seen for this trial.
|
|
"trainingStep": "A String", # The global training step for this metric.
|
|
"objectiveValue": 3.14, # The objective value at this training step.
|
|
},
|
|
"allMetrics": [ # All recorded object metrics for this trial. This field is not currently
|
|
# populated.
|
|
{ # An observed value of a metric.
|
|
"trainingStep": "A String", # The global training step for this metric.
|
|
"objectiveValue": 3.14, # The objective value at this training step.
|
|
},
|
|
],
|
|
"isTrialStoppedEarly": True or False, # True if the trial is stopped early.
|
|
"trialId": "A String", # The trial id for these results.
|
|
"builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
|
|
# Only set for trials of built-in algorithms jobs that have succeeded.
|
|
"framework": "A String", # Framework on which the built-in algorithm was trained.
|
|
"modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
|
|
# saves the trained model. Only set for successful jobs that don't use
|
|
# hyperparameter tuning.
|
|
"runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
|
|
# trained.
|
|
"pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
|
|
},
|
|
},
|
|
],
|
|
"isHyperparameterTuningJob": True or False, # Whether this job is a hyperparameter tuning job.
|
|
"isBuiltInAlgorithmJob": True or False, # Whether this job is a built-in Algorithm job.
|
|
"consumedMLUnits": 3.14, # The amount of ML units consumed by the job.
|
|
"hyperparameterMetricTag": "A String", # The TensorFlow summary tag name used for optimizing hyperparameter tuning
|
|
# trials. See
|
|
# [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag)
|
|
# for more information. Only set for hyperparameter tuning jobs.
|
|
"builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
|
|
# Only set for built-in algorithms jobs.
|
|
"framework": "A String", # Framework on which the built-in algorithm was trained.
|
|
"modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
|
|
# saves the trained model. Only set for successful jobs that don't use
|
|
# hyperparameter tuning.
|
|
"runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
|
|
# trained.
|
|
"pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
|
|
},
|
|
},
|
|
"predictionInput": { # Represents input parameters for a prediction job. # Input parameters to create a prediction job.
|
|
"modelName": "A String", # Use this field if you want to use the default version for the specified
|
|
# model. The string must use the following format:
|
|
#
|
|
# `"projects/YOUR_PROJECT/models/YOUR_MODEL"`
|
|
"runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for this batch
|
|
# prediction. If not set, AI Platform will pick the runtime version used
|
|
# during the CreateVersion request for this model version, or choose the
|
|
# latest stable version when model version information is not available
|
|
# such as when the model is specified by uri.
|
|
"signatureName": "A String", # Optional. The name of the signature defined in the SavedModel to use for
|
|
# this job. Please refer to
|
|
# [SavedModel](https://tensorflow.github.io/serving/serving_basic.html)
|
|
# for information about how to use signatures.
|
|
#
|
|
# Defaults to
|
|
# [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants)
|
|
# , which is "serving_default".
|
|
"batchSize": "A String", # Optional. Number of records per batch, defaults to 64.
|
|
# The service will buffer batch_size number of records in memory before
|
|
# invoking one Tensorflow prediction call internally. So take the record
|
|
# size and memory available into consideration when setting this parameter.
|
|
"inputPaths": [ # Required. The Cloud Storage location of the input data files. May contain
|
|
# <a href="/storage/docs/gsutil/addlhelp/WildcardNames">wildcards</a>.
|
|
"A String",
|
|
],
|
|
"maxWorkerCount": "A String", # Optional. The maximum number of workers to be used for parallel processing.
|
|
# Defaults to 10 if not specified.
|
|
"uri": "A String", # Use this field if you want to specify a Google Cloud Storage path for
|
|
# the model to use.
|
|
"outputPath": "A String", # Required. The output Google Cloud Storage location.
|
|
"dataFormat": "A String", # Required. The format of the input data files.
|
|
"versionName": "A String", # Use this field if you want to specify a version of the model to use. The
|
|
# string is formatted the same way as `model_version`, with the addition
|
|
# of the version information:
|
|
#
|
|
# `"projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION"`
|
|
"region": "A String", # Required. The Google Compute Engine region to run the prediction job in.
|
|
# See the <a href="/ml-engine/docs/tensorflow/regions">available regions</a>
|
|
# for AI Platform services.
|
|
"outputDataFormat": "A String", # Optional. Format of the output data files, defaults to JSON.
|
|
},
|
|
"trainingInput": { # Represents input parameters for a training job. When using the # Input parameters to create a training job.
|
|
# gcloud command to submit your training job, you can specify
|
|
# the input parameters as command-line arguments and/or in a YAML configuration
|
|
# file referenced from the --config command-line argument. For
|
|
# details, see the guide to
|
|
# <a href="/ml-engine/docs/tensorflow/training-jobs">submitting a training
|
|
# job</a>.
|
|
"workerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
|
|
# job's worker nodes.
|
|
#
|
|
# The supported values are the same as those described in the entry for
|
|
# `masterType`.
|
|
#
|
|
# This value must be consistent with the category of machine type that
|
|
# `masterType` uses. In other words, both must be AI Platform machine
|
|
# types or both must be Compute Engine machine types.
|
|
#
|
|
# If you use `cloud_tpu` for this value, see special instructions for
|
|
# [configuring a custom TPU
|
|
# machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
|
|
#
|
|
# This value must be present when `scaleTier` is set to `CUSTOM` and
|
|
# `workerCount` is greater than zero.
|
|
"parameterServerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers.
|
|
#
|
|
# You should only set `parameterServerConfig.acceleratorConfig` if
|
|
# `parameterServerConfigType` is set to a Compute Engine machine type. [Learn
|
|
# about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
#
|
|
# Set `parameterServerConfig.imageUri` only if you build a custom image for
|
|
# your parameter server. If `parameterServerConfig.imageUri` has not been
|
|
# set, AI Platform uses the value of `masterConfig.imageUri`.
|
|
# Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
"acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
|
|
# [Learn about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
"count": "A String", # The number of accelerators to attach to each machine running the job.
|
|
"type": "A String", # The type of accelerator to use.
|
|
},
|
|
"imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
|
|
# Registry. Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
},
|
|
"runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for training. If not
|
|
# set, AI Platform uses the default stable version, 1.0. For more
|
|
# information, see the
|
|
# <a href="/ml-engine/docs/runtime-version-list">runtime version list</a>
|
|
# and
|
|
# <a href="/ml-engine/docs/versioning">how to manage runtime versions</a>.
|
|
"scaleTier": "A String", # Required. Specifies the machine types, the number of replicas for workers
|
|
# and parameter servers.
|
|
"masterType": "A String", # Optional. Specifies the type of virtual machine to use for your training
|
|
# job's master worker.
|
|
#
|
|
# The following types are supported:
|
|
#
|
|
# <dl>
|
|
# <dt>standard</dt>
|
|
# <dd>
|
|
# A basic machine configuration suitable for training simple models with
|
|
# small to moderate datasets.
|
|
# </dd>
|
|
# <dt>large_model</dt>
|
|
# <dd>
|
|
# A machine with a lot of memory, specially suited for parameter servers
|
|
# when your model is large (having many hidden layers or layers with very
|
|
# large numbers of nodes).
|
|
# </dd>
|
|
# <dt>complex_model_s</dt>
|
|
# <dd>
|
|
# A machine suitable for the master and workers of the cluster when your
|
|
# model requires more computation than the standard machine can handle
|
|
# satisfactorily.
|
|
# </dd>
|
|
# <dt>complex_model_m</dt>
|
|
# <dd>
|
|
# A machine with roughly twice the number of cores and roughly double the
|
|
# memory of <i>complex_model_s</i>.
|
|
# </dd>
|
|
# <dt>complex_model_l</dt>
|
|
# <dd>
|
|
# A machine with roughly twice the number of cores and roughly double the
|
|
# memory of <i>complex_model_m</i>.
|
|
# </dd>
|
|
# <dt>standard_gpu</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>standard</i> that
|
|
# also includes a single NVIDIA Tesla K80 GPU. See more about
|
|
# <a href="/ml-engine/docs/tensorflow/using-gpus">using GPUs to
|
|
# train your model</a>.
|
|
# </dd>
|
|
# <dt>complex_model_m_gpu</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_m</i> that also includes
|
|
# four NVIDIA Tesla K80 GPUs.
|
|
# </dd>
|
|
# <dt>complex_model_l_gpu</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_l</i> that also includes
|
|
# eight NVIDIA Tesla K80 GPUs.
|
|
# </dd>
|
|
# <dt>standard_p100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>standard</i> that
|
|
# also includes a single NVIDIA Tesla P100 GPU.
|
|
# </dd>
|
|
# <dt>complex_model_m_p100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_m</i> that also includes
|
|
# four NVIDIA Tesla P100 GPUs.
|
|
# </dd>
|
|
# <dt>standard_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>standard</i> that
|
|
# also includes a single NVIDIA Tesla V100 GPU.
|
|
# </dd>
|
|
# <dt>large_model_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>large_model</i> that
|
|
# also includes a single NVIDIA Tesla V100 GPU.
|
|
# </dd>
|
|
# <dt>complex_model_m_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_m</i> that
|
|
# also includes four NVIDIA Tesla V100 GPUs.
|
|
# </dd>
|
|
# <dt>complex_model_l_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_l</i> that
|
|
# also includes eight NVIDIA Tesla V100 GPUs.
|
|
# </dd>
|
|
# <dt>cloud_tpu</dt>
|
|
# <dd>
|
|
# A TPU VM including one Cloud TPU. See more about
|
|
# <a href="/ml-engine/docs/tensorflow/using-tpus">using TPUs to train
|
|
# your model</a>.
|
|
# </dd>
|
|
# </dl>
|
|
#
|
|
# You may also use certain Compute Engine machine types directly in this
|
|
# field. The following types are supported:
|
|
#
|
|
# - `n1-standard-4`
|
|
# - `n1-standard-8`
|
|
# - `n1-standard-16`
|
|
# - `n1-standard-32`
|
|
# - `n1-standard-64`
|
|
# - `n1-standard-96`
|
|
# - `n1-highmem-2`
|
|
# - `n1-highmem-4`
|
|
# - `n1-highmem-8`
|
|
# - `n1-highmem-16`
|
|
# - `n1-highmem-32`
|
|
# - `n1-highmem-64`
|
|
# - `n1-highmem-96`
|
|
# - `n1-highcpu-16`
|
|
# - `n1-highcpu-32`
|
|
# - `n1-highcpu-64`
|
|
# - `n1-highcpu-96`
|
|
#
|
|
# See more about [using Compute Engine machine
|
|
# types](/ml-engine/docs/tensorflow/machine-types#compute-engine-machine-types).
|
|
#
|
|
# You must set this value when `scaleTier` is set to `CUSTOM`.
|
|
"hyperparameters": { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune.
|
|
"maxTrials": 42, # Optional. How many training trials should be attempted to optimize
|
|
# the specified hyperparameters.
|
|
#
|
|
# Defaults to one.
|
|
"goal": "A String", # Required. The type of goal to use for tuning. Available types are
|
|
# `MAXIMIZE` and `MINIMIZE`.
|
|
#
|
|
# Defaults to `MAXIMIZE`.
|
|
"algorithm": "A String", # Optional. The search algorithm specified for the hyperparameter
|
|
# tuning job.
|
|
# Uses the default AI Platform hyperparameter tuning
|
|
# algorithm if unspecified.
|
|
"maxFailedTrials": 42, # Optional. The number of failed trials that need to be seen before failing
|
|
# the hyperparameter tuning job. You can specify this field to override the
|
|
# default failing criteria for AI Platform hyperparameter tuning jobs.
|
|
#
|
|
# Defaults to zero, which means the service decides when a hyperparameter
|
|
# job should fail.
|
|
"enableTrialEarlyStopping": True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial
|
|
# early stopping.
|
|
"resumePreviousJobId": "A String", # Optional. The prior hyperparameter tuning job id that users hope to
|
|
# continue with. The job id will be used to find the corresponding vizier
|
|
# study guid and resume the study.
|
|
"params": [ # Required. The set of parameters to tune.
|
|
{ # Represents a single hyperparameter to optimize.
|
|
"maxValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
|
|
# should be unset if type is `CATEGORICAL`. This value should be integers if
|
|
# type is `INTEGER`.
|
|
"categoricalValues": [ # Required if type is `CATEGORICAL`. The list of possible categories.
|
|
"A String",
|
|
],
|
|
"discreteValues": [ # Required if type is `DISCRETE`.
|
|
# A list of feasible points.
|
|
# The list should be in strictly increasing order. For instance, this
|
|
# parameter might have possible settings of 1.5, 2.5, and 4.0. This list
|
|
# should not contain more than 1,000 values.
|
|
3.14,
|
|
],
|
|
"parameterName": "A String", # Required. The parameter name must be unique amongst all ParameterConfigs in
|
|
# a HyperparameterSpec message. E.g., "learning_rate".
|
|
"minValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
|
|
# should be unset if type is `CATEGORICAL`. This value should be integers if
|
|
# type is INTEGER.
|
|
"type": "A String", # Required. The type of the parameter.
|
|
"scaleType": "A String", # Optional. How the parameter should be scaled to the hypercube.
|
|
# Leave unset for categorical parameters.
|
|
# Some kind of scaling is strongly recommended for real or integral
|
|
# parameters (e.g., `UNIT_LINEAR_SCALE`).
|
|
},
|
|
],
|
|
"hyperparameterMetricTag": "A String", # Optional. The TensorFlow summary tag name to use for optimizing trials. For
|
|
# current versions of TensorFlow, this tag name should exactly match what is
|
|
# shown in TensorBoard, including all scopes. For versions of TensorFlow
|
|
# prior to 0.12, this should be only the tag passed to tf.Summary.
|
|
# By default, "training/hptuning/metric" will be used.
|
|
"maxParallelTrials": 42, # Optional. The number of training trials to run concurrently.
|
|
# You can reduce the time it takes to perform hyperparameter tuning by adding
|
|
# trials in parallel. However, each trail only benefits from the information
|
|
# gained in completed trials. That means that a trial does not get access to
|
|
# the results of trials running at the same time, which could reduce the
|
|
# quality of the overall optimization.
|
|
#
|
|
# Each trial will use the same scale tier and machine types.
|
|
#
|
|
# Defaults to one.
|
|
},
|
|
"region": "A String", # Required. The Google Compute Engine region to run the training job in.
|
|
# See the <a href="/ml-engine/docs/tensorflow/regions">available regions</a>
|
|
# for AI Platform services.
|
|
"args": [ # Optional. Command line arguments to pass to the program.
|
|
"A String",
|
|
],
|
|
"pythonModule": "A String", # Required. The Python module name to run after installing the packages.
|
|
"pythonVersion": "A String", # Optional. The version of Python used in training. If not set, the default
|
|
# version is '2.7'. Python '3.5' is available when `runtime_version` is set
|
|
# to '1.4' and above. Python '2.7' works with all supported
|
|
# <a href="/ml-engine/docs/runtime-version-list">runtime versions</a>.
|
|
"jobDir": "A String", # Optional. A Google Cloud Storage path in which to store training outputs
|
|
# and other data needed for training. This path is passed to your TensorFlow
|
|
# program as the '--job-dir' command-line argument. The benefit of specifying
|
|
# this field is that Cloud ML validates the path for use in training.
|
|
"packageUris": [ # Required. The Google Cloud Storage location of the packages with
|
|
# the training program and any additional dependencies.
|
|
# The maximum number of package URIs is 100.
|
|
"A String",
|
|
],
|
|
"workerCount": "A String", # Optional. The number of worker replicas to use for the training job. Each
|
|
# replica in the cluster will be of the type specified in `worker_type`.
|
|
#
|
|
# This value can only be used when `scale_tier` is set to `CUSTOM`. If you
|
|
# set this value, you must also set `worker_type`.
|
|
#
|
|
# The default value is zero.
|
|
"parameterServerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
|
|
# job's parameter server.
|
|
#
|
|
# The supported values are the same as those described in the entry for
|
|
# `master_type`.
|
|
#
|
|
# This value must be consistent with the category of machine type that
|
|
# `masterType` uses. In other words, both must be AI Platform machine
|
|
# types or both must be Compute Engine machine types.
|
|
#
|
|
# This value must be present when `scaleTier` is set to `CUSTOM` and
|
|
# `parameter_server_count` is greater than zero.
|
|
"workerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers.
|
|
#
|
|
# You should only set `workerConfig.acceleratorConfig` if `workerType` is set
|
|
# to a Compute Engine machine type. [Learn about restrictions on accelerator
|
|
# configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
#
|
|
# Set `workerConfig.imageUri` only if you build a custom image for your
|
|
# worker. If `workerConfig.imageUri` has not been set, AI Platform uses
|
|
# the value of `masterConfig.imageUri`. Learn more about
|
|
# [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
"acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
|
|
# [Learn about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
"count": "A String", # The number of accelerators to attach to each machine running the job.
|
|
"type": "A String", # The type of accelerator to use.
|
|
},
|
|
"imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
|
|
# Registry. Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
},
|
|
"maxRunningTime": "A String", # Optional. The maximum job running time. The default is 7 days.
|
|
"masterConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker.
|
|
#
|
|
# You should only set `masterConfig.acceleratorConfig` if `masterType` is set
|
|
# to a Compute Engine machine type. Learn about [restrictions on accelerator
|
|
# configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
#
|
|
# Set `masterConfig.imageUri` only if you build a custom image. Only one of
|
|
# `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more about
|
|
# [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
"acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
|
|
# [Learn about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
"count": "A String", # The number of accelerators to attach to each machine running the job.
|
|
"type": "A String", # The type of accelerator to use.
|
|
},
|
|
"imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
|
|
# Registry. Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
},
|
|
"parameterServerCount": "A String", # Optional. The number of parameter server replicas to use for the training
|
|
# job. Each replica in the cluster will be of the type specified in
|
|
# `parameter_server_type`.
|
|
#
|
|
# This value can only be used when `scale_tier` is set to `CUSTOM`.If you
|
|
# set this value, you must also set `parameter_server_type`.
|
|
#
|
|
# The default value is zero.
|
|
},
|
|
"jobId": "A String", # Required. The user-specified id of the job.
|
|
"labels": { # Optional. One or more labels that you can add, to organize your jobs.
|
|
# Each label is a key-value pair, where both the key and the value are
|
|
# arbitrary strings that you supply.
|
|
# For more information, see the documentation on
|
|
# <a href="/ml-engine/docs/tensorflow/resource-labels">using labels</a>.
|
|
"a_key": "A String",
|
|
},
|
|
"state": "A String", # Output only. The detailed state of a job.
|
|
"etag": "A String", # `etag` is used for optimistic concurrency control as a way to help
|
|
# prevent simultaneous updates of a job from overwriting each other.
|
|
# It is strongly suggested that systems make use of the `etag` in the
|
|
# read-modify-write cycle to perform job updates in order to avoid race
|
|
# conditions: An `etag` is returned in the response to `GetJob`, and
|
|
# systems are expected to put that etag in the request to `UpdateJob` to
|
|
# ensure that their change will be applied to the same version of the job.
|
|
"startTime": "A String", # Output only. When the job processing was started.
|
|
"endTime": "A String", # Output only. When the job processing was completed.
|
|
"predictionOutput": { # Represents results of a prediction job. # The current prediction job result.
|
|
"outputPath": "A String", # The output Google Cloud Storage location provided at the job creation time.
|
|
"nodeHours": 3.14, # Node hours used by the batch prediction job.
|
|
"predictionCount": "A String", # The number of generated predictions.
|
|
"errorCount": "A String", # The number of data instances which resulted in errors.
|
|
},
|
|
"createTime": "A String", # Output only. When the job was created.
|
|
},
|
|
],
|
|
}</pre>
|
|
</div>
|
|
|
|
<div class="method">
|
|
<code class="details" id="list_next">list_next(previous_request, previous_response)</code>
|
|
<pre>Retrieves the next page of results.
|
|
|
|
Args:
|
|
previous_request: The request for the previous page. (required)
|
|
previous_response: The response from the request for the previous page. (required)
|
|
|
|
Returns:
|
|
A request object that you can call 'execute()' on to request the next
|
|
page. Returns None if there are no more items in the collection.
|
|
</pre>
|
|
</div>
|
|
|
|
<div class="method">
|
|
<code class="details" id="patch">patch(name, body, updateMask=None, x__xgafv=None)</code>
|
|
<pre>Updates a specific job resource.
|
|
|
|
Currently the only supported fields to update are `labels`.
|
|
|
|
Args:
|
|
name: string, Required. The job name. (required)
|
|
body: object, The request body. (required)
|
|
The object takes the form of:
|
|
|
|
{ # Represents a training or prediction job.
|
|
"errorMessage": "A String", # Output only. The details of a failure or a cancellation.
|
|
"trainingOutput": { # Represents results of a training job. Output only. # The current training job result.
|
|
"completedTrialCount": "A String", # The number of hyperparameter tuning trials that completed successfully.
|
|
# Only set for hyperparameter tuning jobs.
|
|
"trials": [ # Results for individual Hyperparameter trials.
|
|
# Only set for hyperparameter tuning jobs.
|
|
{ # Represents the result of a single hyperparameter tuning trial from a
|
|
# training job. The TrainingOutput object that is returned on successful
|
|
# completion of a training job with hyperparameter tuning includes a list
|
|
# of HyperparameterOutput objects, one for each successful trial.
|
|
"hyperparameters": { # The hyperparameters given to this trial.
|
|
"a_key": "A String",
|
|
},
|
|
"finalMetric": { # An observed value of a metric. # The final objective metric seen for this trial.
|
|
"trainingStep": "A String", # The global training step for this metric.
|
|
"objectiveValue": 3.14, # The objective value at this training step.
|
|
},
|
|
"allMetrics": [ # All recorded object metrics for this trial. This field is not currently
|
|
# populated.
|
|
{ # An observed value of a metric.
|
|
"trainingStep": "A String", # The global training step for this metric.
|
|
"objectiveValue": 3.14, # The objective value at this training step.
|
|
},
|
|
],
|
|
"isTrialStoppedEarly": True or False, # True if the trial is stopped early.
|
|
"trialId": "A String", # The trial id for these results.
|
|
"builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
|
|
# Only set for trials of built-in algorithms jobs that have succeeded.
|
|
"framework": "A String", # Framework on which the built-in algorithm was trained.
|
|
"modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
|
|
# saves the trained model. Only set for successful jobs that don't use
|
|
# hyperparameter tuning.
|
|
"runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
|
|
# trained.
|
|
"pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
|
|
},
|
|
},
|
|
],
|
|
"isHyperparameterTuningJob": True or False, # Whether this job is a hyperparameter tuning job.
|
|
"isBuiltInAlgorithmJob": True or False, # Whether this job is a built-in Algorithm job.
|
|
"consumedMLUnits": 3.14, # The amount of ML units consumed by the job.
|
|
"hyperparameterMetricTag": "A String", # The TensorFlow summary tag name used for optimizing hyperparameter tuning
|
|
# trials. See
|
|
# [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag)
|
|
# for more information. Only set for hyperparameter tuning jobs.
|
|
"builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
|
|
# Only set for built-in algorithms jobs.
|
|
"framework": "A String", # Framework on which the built-in algorithm was trained.
|
|
"modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
|
|
# saves the trained model. Only set for successful jobs that don't use
|
|
# hyperparameter tuning.
|
|
"runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
|
|
# trained.
|
|
"pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
|
|
},
|
|
},
|
|
"predictionInput": { # Represents input parameters for a prediction job. # Input parameters to create a prediction job.
|
|
"modelName": "A String", # Use this field if you want to use the default version for the specified
|
|
# model. The string must use the following format:
|
|
#
|
|
# `"projects/YOUR_PROJECT/models/YOUR_MODEL"`
|
|
"runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for this batch
|
|
# prediction. If not set, AI Platform will pick the runtime version used
|
|
# during the CreateVersion request for this model version, or choose the
|
|
# latest stable version when model version information is not available
|
|
# such as when the model is specified by uri.
|
|
"signatureName": "A String", # Optional. The name of the signature defined in the SavedModel to use for
|
|
# this job. Please refer to
|
|
# [SavedModel](https://tensorflow.github.io/serving/serving_basic.html)
|
|
# for information about how to use signatures.
|
|
#
|
|
# Defaults to
|
|
# [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants)
|
|
# , which is "serving_default".
|
|
"batchSize": "A String", # Optional. Number of records per batch, defaults to 64.
|
|
# The service will buffer batch_size number of records in memory before
|
|
# invoking one Tensorflow prediction call internally. So take the record
|
|
# size and memory available into consideration when setting this parameter.
|
|
"inputPaths": [ # Required. The Cloud Storage location of the input data files. May contain
|
|
# <a href="/storage/docs/gsutil/addlhelp/WildcardNames">wildcards</a>.
|
|
"A String",
|
|
],
|
|
"maxWorkerCount": "A String", # Optional. The maximum number of workers to be used for parallel processing.
|
|
# Defaults to 10 if not specified.
|
|
"uri": "A String", # Use this field if you want to specify a Google Cloud Storage path for
|
|
# the model to use.
|
|
"outputPath": "A String", # Required. The output Google Cloud Storage location.
|
|
"dataFormat": "A String", # Required. The format of the input data files.
|
|
"versionName": "A String", # Use this field if you want to specify a version of the model to use. The
|
|
# string is formatted the same way as `model_version`, with the addition
|
|
# of the version information:
|
|
#
|
|
# `"projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION"`
|
|
"region": "A String", # Required. The Google Compute Engine region to run the prediction job in.
|
|
# See the <a href="/ml-engine/docs/tensorflow/regions">available regions</a>
|
|
# for AI Platform services.
|
|
"outputDataFormat": "A String", # Optional. Format of the output data files, defaults to JSON.
|
|
},
|
|
"trainingInput": { # Represents input parameters for a training job. When using the # Input parameters to create a training job.
|
|
# gcloud command to submit your training job, you can specify
|
|
# the input parameters as command-line arguments and/or in a YAML configuration
|
|
# file referenced from the --config command-line argument. For
|
|
# details, see the guide to
|
|
# <a href="/ml-engine/docs/tensorflow/training-jobs">submitting a training
|
|
# job</a>.
|
|
"workerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
|
|
# job's worker nodes.
|
|
#
|
|
# The supported values are the same as those described in the entry for
|
|
# `masterType`.
|
|
#
|
|
# This value must be consistent with the category of machine type that
|
|
# `masterType` uses. In other words, both must be AI Platform machine
|
|
# types or both must be Compute Engine machine types.
|
|
#
|
|
# If you use `cloud_tpu` for this value, see special instructions for
|
|
# [configuring a custom TPU
|
|
# machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
|
|
#
|
|
# This value must be present when `scaleTier` is set to `CUSTOM` and
|
|
# `workerCount` is greater than zero.
|
|
"parameterServerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers.
|
|
#
|
|
# You should only set `parameterServerConfig.acceleratorConfig` if
|
|
# `parameterServerConfigType` is set to a Compute Engine machine type. [Learn
|
|
# about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
#
|
|
# Set `parameterServerConfig.imageUri` only if you build a custom image for
|
|
# your parameter server. If `parameterServerConfig.imageUri` has not been
|
|
# set, AI Platform uses the value of `masterConfig.imageUri`.
|
|
# Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
"acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
|
|
# [Learn about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
"count": "A String", # The number of accelerators to attach to each machine running the job.
|
|
"type": "A String", # The type of accelerator to use.
|
|
},
|
|
"imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
|
|
# Registry. Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
},
|
|
"runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for training. If not
|
|
# set, AI Platform uses the default stable version, 1.0. For more
|
|
# information, see the
|
|
# <a href="/ml-engine/docs/runtime-version-list">runtime version list</a>
|
|
# and
|
|
# <a href="/ml-engine/docs/versioning">how to manage runtime versions</a>.
|
|
"scaleTier": "A String", # Required. Specifies the machine types, the number of replicas for workers
|
|
# and parameter servers.
|
|
"masterType": "A String", # Optional. Specifies the type of virtual machine to use for your training
|
|
# job's master worker.
|
|
#
|
|
# The following types are supported:
|
|
#
|
|
# <dl>
|
|
# <dt>standard</dt>
|
|
# <dd>
|
|
# A basic machine configuration suitable for training simple models with
|
|
# small to moderate datasets.
|
|
# </dd>
|
|
# <dt>large_model</dt>
|
|
# <dd>
|
|
# A machine with a lot of memory, specially suited for parameter servers
|
|
# when your model is large (having many hidden layers or layers with very
|
|
# large numbers of nodes).
|
|
# </dd>
|
|
# <dt>complex_model_s</dt>
|
|
# <dd>
|
|
# A machine suitable for the master and workers of the cluster when your
|
|
# model requires more computation than the standard machine can handle
|
|
# satisfactorily.
|
|
# </dd>
|
|
# <dt>complex_model_m</dt>
|
|
# <dd>
|
|
# A machine with roughly twice the number of cores and roughly double the
|
|
# memory of <i>complex_model_s</i>.
|
|
# </dd>
|
|
# <dt>complex_model_l</dt>
|
|
# <dd>
|
|
# A machine with roughly twice the number of cores and roughly double the
|
|
# memory of <i>complex_model_m</i>.
|
|
# </dd>
|
|
# <dt>standard_gpu</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>standard</i> that
|
|
# also includes a single NVIDIA Tesla K80 GPU. See more about
|
|
# <a href="/ml-engine/docs/tensorflow/using-gpus">using GPUs to
|
|
# train your model</a>.
|
|
# </dd>
|
|
# <dt>complex_model_m_gpu</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_m</i> that also includes
|
|
# four NVIDIA Tesla K80 GPUs.
|
|
# </dd>
|
|
# <dt>complex_model_l_gpu</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_l</i> that also includes
|
|
# eight NVIDIA Tesla K80 GPUs.
|
|
# </dd>
|
|
# <dt>standard_p100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>standard</i> that
|
|
# also includes a single NVIDIA Tesla P100 GPU.
|
|
# </dd>
|
|
# <dt>complex_model_m_p100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_m</i> that also includes
|
|
# four NVIDIA Tesla P100 GPUs.
|
|
# </dd>
|
|
# <dt>standard_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>standard</i> that
|
|
# also includes a single NVIDIA Tesla V100 GPU.
|
|
# </dd>
|
|
# <dt>large_model_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>large_model</i> that
|
|
# also includes a single NVIDIA Tesla V100 GPU.
|
|
# </dd>
|
|
# <dt>complex_model_m_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_m</i> that
|
|
# also includes four NVIDIA Tesla V100 GPUs.
|
|
# </dd>
|
|
# <dt>complex_model_l_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_l</i> that
|
|
# also includes eight NVIDIA Tesla V100 GPUs.
|
|
# </dd>
|
|
# <dt>cloud_tpu</dt>
|
|
# <dd>
|
|
# A TPU VM including one Cloud TPU. See more about
|
|
# <a href="/ml-engine/docs/tensorflow/using-tpus">using TPUs to train
|
|
# your model</a>.
|
|
# </dd>
|
|
# </dl>
|
|
#
|
|
# You may also use certain Compute Engine machine types directly in this
|
|
# field. The following types are supported:
|
|
#
|
|
# - `n1-standard-4`
|
|
# - `n1-standard-8`
|
|
# - `n1-standard-16`
|
|
# - `n1-standard-32`
|
|
# - `n1-standard-64`
|
|
# - `n1-standard-96`
|
|
# - `n1-highmem-2`
|
|
# - `n1-highmem-4`
|
|
# - `n1-highmem-8`
|
|
# - `n1-highmem-16`
|
|
# - `n1-highmem-32`
|
|
# - `n1-highmem-64`
|
|
# - `n1-highmem-96`
|
|
# - `n1-highcpu-16`
|
|
# - `n1-highcpu-32`
|
|
# - `n1-highcpu-64`
|
|
# - `n1-highcpu-96`
|
|
#
|
|
# See more about [using Compute Engine machine
|
|
# types](/ml-engine/docs/tensorflow/machine-types#compute-engine-machine-types).
|
|
#
|
|
# You must set this value when `scaleTier` is set to `CUSTOM`.
|
|
"hyperparameters": { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune.
|
|
"maxTrials": 42, # Optional. How many training trials should be attempted to optimize
|
|
# the specified hyperparameters.
|
|
#
|
|
# Defaults to one.
|
|
"goal": "A String", # Required. The type of goal to use for tuning. Available types are
|
|
# `MAXIMIZE` and `MINIMIZE`.
|
|
#
|
|
# Defaults to `MAXIMIZE`.
|
|
"algorithm": "A String", # Optional. The search algorithm specified for the hyperparameter
|
|
# tuning job.
|
|
# Uses the default AI Platform hyperparameter tuning
|
|
# algorithm if unspecified.
|
|
"maxFailedTrials": 42, # Optional. The number of failed trials that need to be seen before failing
|
|
# the hyperparameter tuning job. You can specify this field to override the
|
|
# default failing criteria for AI Platform hyperparameter tuning jobs.
|
|
#
|
|
# Defaults to zero, which means the service decides when a hyperparameter
|
|
# job should fail.
|
|
"enableTrialEarlyStopping": True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial
|
|
# early stopping.
|
|
"resumePreviousJobId": "A String", # Optional. The prior hyperparameter tuning job id that users hope to
|
|
# continue with. The job id will be used to find the corresponding vizier
|
|
# study guid and resume the study.
|
|
"params": [ # Required. The set of parameters to tune.
|
|
{ # Represents a single hyperparameter to optimize.
|
|
"maxValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
|
|
# should be unset if type is `CATEGORICAL`. This value should be integers if
|
|
# type is `INTEGER`.
|
|
"categoricalValues": [ # Required if type is `CATEGORICAL`. The list of possible categories.
|
|
"A String",
|
|
],
|
|
"discreteValues": [ # Required if type is `DISCRETE`.
|
|
# A list of feasible points.
|
|
# The list should be in strictly increasing order. For instance, this
|
|
# parameter might have possible settings of 1.5, 2.5, and 4.0. This list
|
|
# should not contain more than 1,000 values.
|
|
3.14,
|
|
],
|
|
"parameterName": "A String", # Required. The parameter name must be unique amongst all ParameterConfigs in
|
|
# a HyperparameterSpec message. E.g., "learning_rate".
|
|
"minValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
|
|
# should be unset if type is `CATEGORICAL`. This value should be integers if
|
|
# type is INTEGER.
|
|
"type": "A String", # Required. The type of the parameter.
|
|
"scaleType": "A String", # Optional. How the parameter should be scaled to the hypercube.
|
|
# Leave unset for categorical parameters.
|
|
# Some kind of scaling is strongly recommended for real or integral
|
|
# parameters (e.g., `UNIT_LINEAR_SCALE`).
|
|
},
|
|
],
|
|
"hyperparameterMetricTag": "A String", # Optional. The TensorFlow summary tag name to use for optimizing trials. For
|
|
# current versions of TensorFlow, this tag name should exactly match what is
|
|
# shown in TensorBoard, including all scopes. For versions of TensorFlow
|
|
# prior to 0.12, this should be only the tag passed to tf.Summary.
|
|
# By default, "training/hptuning/metric" will be used.
|
|
"maxParallelTrials": 42, # Optional. The number of training trials to run concurrently.
|
|
# You can reduce the time it takes to perform hyperparameter tuning by adding
|
|
# trials in parallel. However, each trail only benefits from the information
|
|
# gained in completed trials. That means that a trial does not get access to
|
|
# the results of trials running at the same time, which could reduce the
|
|
# quality of the overall optimization.
|
|
#
|
|
# Each trial will use the same scale tier and machine types.
|
|
#
|
|
# Defaults to one.
|
|
},
|
|
"region": "A String", # Required. The Google Compute Engine region to run the training job in.
|
|
# See the <a href="/ml-engine/docs/tensorflow/regions">available regions</a>
|
|
# for AI Platform services.
|
|
"args": [ # Optional. Command line arguments to pass to the program.
|
|
"A String",
|
|
],
|
|
"pythonModule": "A String", # Required. The Python module name to run after installing the packages.
|
|
"pythonVersion": "A String", # Optional. The version of Python used in training. If not set, the default
|
|
# version is '2.7'. Python '3.5' is available when `runtime_version` is set
|
|
# to '1.4' and above. Python '2.7' works with all supported
|
|
# <a href="/ml-engine/docs/runtime-version-list">runtime versions</a>.
|
|
"jobDir": "A String", # Optional. A Google Cloud Storage path in which to store training outputs
|
|
# and other data needed for training. This path is passed to your TensorFlow
|
|
# program as the '--job-dir' command-line argument. The benefit of specifying
|
|
# this field is that Cloud ML validates the path for use in training.
|
|
"packageUris": [ # Required. The Google Cloud Storage location of the packages with
|
|
# the training program and any additional dependencies.
|
|
# The maximum number of package URIs is 100.
|
|
"A String",
|
|
],
|
|
"workerCount": "A String", # Optional. The number of worker replicas to use for the training job. Each
|
|
# replica in the cluster will be of the type specified in `worker_type`.
|
|
#
|
|
# This value can only be used when `scale_tier` is set to `CUSTOM`. If you
|
|
# set this value, you must also set `worker_type`.
|
|
#
|
|
# The default value is zero.
|
|
"parameterServerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
|
|
# job's parameter server.
|
|
#
|
|
# The supported values are the same as those described in the entry for
|
|
# `master_type`.
|
|
#
|
|
# This value must be consistent with the category of machine type that
|
|
# `masterType` uses. In other words, both must be AI Platform machine
|
|
# types or both must be Compute Engine machine types.
|
|
#
|
|
# This value must be present when `scaleTier` is set to `CUSTOM` and
|
|
# `parameter_server_count` is greater than zero.
|
|
"workerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers.
|
|
#
|
|
# You should only set `workerConfig.acceleratorConfig` if `workerType` is set
|
|
# to a Compute Engine machine type. [Learn about restrictions on accelerator
|
|
# configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
#
|
|
# Set `workerConfig.imageUri` only if you build a custom image for your
|
|
# worker. If `workerConfig.imageUri` has not been set, AI Platform uses
|
|
# the value of `masterConfig.imageUri`. Learn more about
|
|
# [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
"acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
|
|
# [Learn about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
"count": "A String", # The number of accelerators to attach to each machine running the job.
|
|
"type": "A String", # The type of accelerator to use.
|
|
},
|
|
"imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
|
|
# Registry. Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
},
|
|
"maxRunningTime": "A String", # Optional. The maximum job running time. The default is 7 days.
|
|
"masterConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker.
|
|
#
|
|
# You should only set `masterConfig.acceleratorConfig` if `masterType` is set
|
|
# to a Compute Engine machine type. Learn about [restrictions on accelerator
|
|
# configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
#
|
|
# Set `masterConfig.imageUri` only if you build a custom image. Only one of
|
|
# `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more about
|
|
# [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
"acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
|
|
# [Learn about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
"count": "A String", # The number of accelerators to attach to each machine running the job.
|
|
"type": "A String", # The type of accelerator to use.
|
|
},
|
|
"imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
|
|
# Registry. Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
},
|
|
"parameterServerCount": "A String", # Optional. The number of parameter server replicas to use for the training
|
|
# job. Each replica in the cluster will be of the type specified in
|
|
# `parameter_server_type`.
|
|
#
|
|
# This value can only be used when `scale_tier` is set to `CUSTOM`.If you
|
|
# set this value, you must also set `parameter_server_type`.
|
|
#
|
|
# The default value is zero.
|
|
},
|
|
"jobId": "A String", # Required. The user-specified id of the job.
|
|
"labels": { # Optional. One or more labels that you can add, to organize your jobs.
|
|
# Each label is a key-value pair, where both the key and the value are
|
|
# arbitrary strings that you supply.
|
|
# For more information, see the documentation on
|
|
# <a href="/ml-engine/docs/tensorflow/resource-labels">using labels</a>.
|
|
"a_key": "A String",
|
|
},
|
|
"state": "A String", # Output only. The detailed state of a job.
|
|
"etag": "A String", # `etag` is used for optimistic concurrency control as a way to help
|
|
# prevent simultaneous updates of a job from overwriting each other.
|
|
# It is strongly suggested that systems make use of the `etag` in the
|
|
# read-modify-write cycle to perform job updates in order to avoid race
|
|
# conditions: An `etag` is returned in the response to `GetJob`, and
|
|
# systems are expected to put that etag in the request to `UpdateJob` to
|
|
# ensure that their change will be applied to the same version of the job.
|
|
"startTime": "A String", # Output only. When the job processing was started.
|
|
"endTime": "A String", # Output only. When the job processing was completed.
|
|
"predictionOutput": { # Represents results of a prediction job. # The current prediction job result.
|
|
"outputPath": "A String", # The output Google Cloud Storage location provided at the job creation time.
|
|
"nodeHours": 3.14, # Node hours used by the batch prediction job.
|
|
"predictionCount": "A String", # The number of generated predictions.
|
|
"errorCount": "A String", # The number of data instances which resulted in errors.
|
|
},
|
|
"createTime": "A String", # Output only. When the job was created.
|
|
}
|
|
|
|
updateMask: string, Required. Specifies the path, relative to `Job`, of the field to update.
|
|
To adopt etag mechanism, include `etag` field in the mask, and include the
|
|
`etag` value in your job resource.
|
|
|
|
For example, to change the labels of a job, the `update_mask` parameter
|
|
would be specified as `labels`, `etag`, and the
|
|
`PATCH` request body would specify the new value, as follows:
|
|
{
|
|
"labels": {
|
|
"owner": "Google",
|
|
"color": "Blue"
|
|
}
|
|
"etag": "33a64df551425fcc55e4d42a148795d9f25f89d4"
|
|
}
|
|
If `etag` matches the one on the server, the labels of the job will be
|
|
replaced with the given ones, and the server end `etag` will be
|
|
recalculated.
|
|
|
|
Currently the only supported update masks are `labels` and `etag`.
|
|
x__xgafv: string, V1 error format.
|
|
Allowed values
|
|
1 - v1 error format
|
|
2 - v2 error format
|
|
|
|
Returns:
|
|
An object of the form:
|
|
|
|
{ # Represents a training or prediction job.
|
|
"errorMessage": "A String", # Output only. The details of a failure or a cancellation.
|
|
"trainingOutput": { # Represents results of a training job. Output only. # The current training job result.
|
|
"completedTrialCount": "A String", # The number of hyperparameter tuning trials that completed successfully.
|
|
# Only set for hyperparameter tuning jobs.
|
|
"trials": [ # Results for individual Hyperparameter trials.
|
|
# Only set for hyperparameter tuning jobs.
|
|
{ # Represents the result of a single hyperparameter tuning trial from a
|
|
# training job. The TrainingOutput object that is returned on successful
|
|
# completion of a training job with hyperparameter tuning includes a list
|
|
# of HyperparameterOutput objects, one for each successful trial.
|
|
"hyperparameters": { # The hyperparameters given to this trial.
|
|
"a_key": "A String",
|
|
},
|
|
"finalMetric": { # An observed value of a metric. # The final objective metric seen for this trial.
|
|
"trainingStep": "A String", # The global training step for this metric.
|
|
"objectiveValue": 3.14, # The objective value at this training step.
|
|
},
|
|
"allMetrics": [ # All recorded object metrics for this trial. This field is not currently
|
|
# populated.
|
|
{ # An observed value of a metric.
|
|
"trainingStep": "A String", # The global training step for this metric.
|
|
"objectiveValue": 3.14, # The objective value at this training step.
|
|
},
|
|
],
|
|
"isTrialStoppedEarly": True or False, # True if the trial is stopped early.
|
|
"trialId": "A String", # The trial id for these results.
|
|
"builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
|
|
# Only set for trials of built-in algorithms jobs that have succeeded.
|
|
"framework": "A String", # Framework on which the built-in algorithm was trained.
|
|
"modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
|
|
# saves the trained model. Only set for successful jobs that don't use
|
|
# hyperparameter tuning.
|
|
"runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
|
|
# trained.
|
|
"pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
|
|
},
|
|
},
|
|
],
|
|
"isHyperparameterTuningJob": True or False, # Whether this job is a hyperparameter tuning job.
|
|
"isBuiltInAlgorithmJob": True or False, # Whether this job is a built-in Algorithm job.
|
|
"consumedMLUnits": 3.14, # The amount of ML units consumed by the job.
|
|
"hyperparameterMetricTag": "A String", # The TensorFlow summary tag name used for optimizing hyperparameter tuning
|
|
# trials. See
|
|
# [`HyperparameterSpec.hyperparameterMetricTag`](#HyperparameterSpec.FIELDS.hyperparameter_metric_tag)
|
|
# for more information. Only set for hyperparameter tuning jobs.
|
|
"builtInAlgorithmOutput": { # Represents output related to a built-in algorithm Job. # Details related to built-in algorithms jobs.
|
|
# Only set for built-in algorithms jobs.
|
|
"framework": "A String", # Framework on which the built-in algorithm was trained.
|
|
"modelPath": "A String", # The Cloud Storage path to the `model/` directory where the training job
|
|
# saves the trained model. Only set for successful jobs that don't use
|
|
# hyperparameter tuning.
|
|
"runtimeVersion": "A String", # AI Platform runtime version on which the built-in algorithm was
|
|
# trained.
|
|
"pythonVersion": "A String", # Python version on which the built-in algorithm was trained.
|
|
},
|
|
},
|
|
"predictionInput": { # Represents input parameters for a prediction job. # Input parameters to create a prediction job.
|
|
"modelName": "A String", # Use this field if you want to use the default version for the specified
|
|
# model. The string must use the following format:
|
|
#
|
|
# `"projects/YOUR_PROJECT/models/YOUR_MODEL"`
|
|
"runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for this batch
|
|
# prediction. If not set, AI Platform will pick the runtime version used
|
|
# during the CreateVersion request for this model version, or choose the
|
|
# latest stable version when model version information is not available
|
|
# such as when the model is specified by uri.
|
|
"signatureName": "A String", # Optional. The name of the signature defined in the SavedModel to use for
|
|
# this job. Please refer to
|
|
# [SavedModel](https://tensorflow.github.io/serving/serving_basic.html)
|
|
# for information about how to use signatures.
|
|
#
|
|
# Defaults to
|
|
# [DEFAULT_SERVING_SIGNATURE_DEF_KEY](https://www.tensorflow.org/api_docs/python/tf/saved_model/signature_constants)
|
|
# , which is "serving_default".
|
|
"batchSize": "A String", # Optional. Number of records per batch, defaults to 64.
|
|
# The service will buffer batch_size number of records in memory before
|
|
# invoking one Tensorflow prediction call internally. So take the record
|
|
# size and memory available into consideration when setting this parameter.
|
|
"inputPaths": [ # Required. The Cloud Storage location of the input data files. May contain
|
|
# <a href="/storage/docs/gsutil/addlhelp/WildcardNames">wildcards</a>.
|
|
"A String",
|
|
],
|
|
"maxWorkerCount": "A String", # Optional. The maximum number of workers to be used for parallel processing.
|
|
# Defaults to 10 if not specified.
|
|
"uri": "A String", # Use this field if you want to specify a Google Cloud Storage path for
|
|
# the model to use.
|
|
"outputPath": "A String", # Required. The output Google Cloud Storage location.
|
|
"dataFormat": "A String", # Required. The format of the input data files.
|
|
"versionName": "A String", # Use this field if you want to specify a version of the model to use. The
|
|
# string is formatted the same way as `model_version`, with the addition
|
|
# of the version information:
|
|
#
|
|
# `"projects/YOUR_PROJECT/models/YOUR_MODEL/versions/YOUR_VERSION"`
|
|
"region": "A String", # Required. The Google Compute Engine region to run the prediction job in.
|
|
# See the <a href="/ml-engine/docs/tensorflow/regions">available regions</a>
|
|
# for AI Platform services.
|
|
"outputDataFormat": "A String", # Optional. Format of the output data files, defaults to JSON.
|
|
},
|
|
"trainingInput": { # Represents input parameters for a training job. When using the # Input parameters to create a training job.
|
|
# gcloud command to submit your training job, you can specify
|
|
# the input parameters as command-line arguments and/or in a YAML configuration
|
|
# file referenced from the --config command-line argument. For
|
|
# details, see the guide to
|
|
# <a href="/ml-engine/docs/tensorflow/training-jobs">submitting a training
|
|
# job</a>.
|
|
"workerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
|
|
# job's worker nodes.
|
|
#
|
|
# The supported values are the same as those described in the entry for
|
|
# `masterType`.
|
|
#
|
|
# This value must be consistent with the category of machine type that
|
|
# `masterType` uses. In other words, both must be AI Platform machine
|
|
# types or both must be Compute Engine machine types.
|
|
#
|
|
# If you use `cloud_tpu` for this value, see special instructions for
|
|
# [configuring a custom TPU
|
|
# machine](/ml-engine/docs/tensorflow/using-tpus#configuring_a_custom_tpu_machine).
|
|
#
|
|
# This value must be present when `scaleTier` is set to `CUSTOM` and
|
|
# `workerCount` is greater than zero.
|
|
"parameterServerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for parameter servers.
|
|
#
|
|
# You should only set `parameterServerConfig.acceleratorConfig` if
|
|
# `parameterServerConfigType` is set to a Compute Engine machine type. [Learn
|
|
# about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
#
|
|
# Set `parameterServerConfig.imageUri` only if you build a custom image for
|
|
# your parameter server. If `parameterServerConfig.imageUri` has not been
|
|
# set, AI Platform uses the value of `masterConfig.imageUri`.
|
|
# Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
"acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
|
|
# [Learn about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
"count": "A String", # The number of accelerators to attach to each machine running the job.
|
|
"type": "A String", # The type of accelerator to use.
|
|
},
|
|
"imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
|
|
# Registry. Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
},
|
|
"runtimeVersion": "A String", # Optional. The AI Platform runtime version to use for training. If not
|
|
# set, AI Platform uses the default stable version, 1.0. For more
|
|
# information, see the
|
|
# <a href="/ml-engine/docs/runtime-version-list">runtime version list</a>
|
|
# and
|
|
# <a href="/ml-engine/docs/versioning">how to manage runtime versions</a>.
|
|
"scaleTier": "A String", # Required. Specifies the machine types, the number of replicas for workers
|
|
# and parameter servers.
|
|
"masterType": "A String", # Optional. Specifies the type of virtual machine to use for your training
|
|
# job's master worker.
|
|
#
|
|
# The following types are supported:
|
|
#
|
|
# <dl>
|
|
# <dt>standard</dt>
|
|
# <dd>
|
|
# A basic machine configuration suitable for training simple models with
|
|
# small to moderate datasets.
|
|
# </dd>
|
|
# <dt>large_model</dt>
|
|
# <dd>
|
|
# A machine with a lot of memory, specially suited for parameter servers
|
|
# when your model is large (having many hidden layers or layers with very
|
|
# large numbers of nodes).
|
|
# </dd>
|
|
# <dt>complex_model_s</dt>
|
|
# <dd>
|
|
# A machine suitable for the master and workers of the cluster when your
|
|
# model requires more computation than the standard machine can handle
|
|
# satisfactorily.
|
|
# </dd>
|
|
# <dt>complex_model_m</dt>
|
|
# <dd>
|
|
# A machine with roughly twice the number of cores and roughly double the
|
|
# memory of <i>complex_model_s</i>.
|
|
# </dd>
|
|
# <dt>complex_model_l</dt>
|
|
# <dd>
|
|
# A machine with roughly twice the number of cores and roughly double the
|
|
# memory of <i>complex_model_m</i>.
|
|
# </dd>
|
|
# <dt>standard_gpu</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>standard</i> that
|
|
# also includes a single NVIDIA Tesla K80 GPU. See more about
|
|
# <a href="/ml-engine/docs/tensorflow/using-gpus">using GPUs to
|
|
# train your model</a>.
|
|
# </dd>
|
|
# <dt>complex_model_m_gpu</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_m</i> that also includes
|
|
# four NVIDIA Tesla K80 GPUs.
|
|
# </dd>
|
|
# <dt>complex_model_l_gpu</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_l</i> that also includes
|
|
# eight NVIDIA Tesla K80 GPUs.
|
|
# </dd>
|
|
# <dt>standard_p100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>standard</i> that
|
|
# also includes a single NVIDIA Tesla P100 GPU.
|
|
# </dd>
|
|
# <dt>complex_model_m_p100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_m</i> that also includes
|
|
# four NVIDIA Tesla P100 GPUs.
|
|
# </dd>
|
|
# <dt>standard_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>standard</i> that
|
|
# also includes a single NVIDIA Tesla V100 GPU.
|
|
# </dd>
|
|
# <dt>large_model_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>large_model</i> that
|
|
# also includes a single NVIDIA Tesla V100 GPU.
|
|
# </dd>
|
|
# <dt>complex_model_m_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_m</i> that
|
|
# also includes four NVIDIA Tesla V100 GPUs.
|
|
# </dd>
|
|
# <dt>complex_model_l_v100</dt>
|
|
# <dd>
|
|
# A machine equivalent to <i>complex_model_l</i> that
|
|
# also includes eight NVIDIA Tesla V100 GPUs.
|
|
# </dd>
|
|
# <dt>cloud_tpu</dt>
|
|
# <dd>
|
|
# A TPU VM including one Cloud TPU. See more about
|
|
# <a href="/ml-engine/docs/tensorflow/using-tpus">using TPUs to train
|
|
# your model</a>.
|
|
# </dd>
|
|
# </dl>
|
|
#
|
|
# You may also use certain Compute Engine machine types directly in this
|
|
# field. The following types are supported:
|
|
#
|
|
# - `n1-standard-4`
|
|
# - `n1-standard-8`
|
|
# - `n1-standard-16`
|
|
# - `n1-standard-32`
|
|
# - `n1-standard-64`
|
|
# - `n1-standard-96`
|
|
# - `n1-highmem-2`
|
|
# - `n1-highmem-4`
|
|
# - `n1-highmem-8`
|
|
# - `n1-highmem-16`
|
|
# - `n1-highmem-32`
|
|
# - `n1-highmem-64`
|
|
# - `n1-highmem-96`
|
|
# - `n1-highcpu-16`
|
|
# - `n1-highcpu-32`
|
|
# - `n1-highcpu-64`
|
|
# - `n1-highcpu-96`
|
|
#
|
|
# See more about [using Compute Engine machine
|
|
# types](/ml-engine/docs/tensorflow/machine-types#compute-engine-machine-types).
|
|
#
|
|
# You must set this value when `scaleTier` is set to `CUSTOM`.
|
|
"hyperparameters": { # Represents a set of hyperparameters to optimize. # Optional. The set of Hyperparameters to tune.
|
|
"maxTrials": 42, # Optional. How many training trials should be attempted to optimize
|
|
# the specified hyperparameters.
|
|
#
|
|
# Defaults to one.
|
|
"goal": "A String", # Required. The type of goal to use for tuning. Available types are
|
|
# `MAXIMIZE` and `MINIMIZE`.
|
|
#
|
|
# Defaults to `MAXIMIZE`.
|
|
"algorithm": "A String", # Optional. The search algorithm specified for the hyperparameter
|
|
# tuning job.
|
|
# Uses the default AI Platform hyperparameter tuning
|
|
# algorithm if unspecified.
|
|
"maxFailedTrials": 42, # Optional. The number of failed trials that need to be seen before failing
|
|
# the hyperparameter tuning job. You can specify this field to override the
|
|
# default failing criteria for AI Platform hyperparameter tuning jobs.
|
|
#
|
|
# Defaults to zero, which means the service decides when a hyperparameter
|
|
# job should fail.
|
|
"enableTrialEarlyStopping": True or False, # Optional. Indicates if the hyperparameter tuning job enables auto trial
|
|
# early stopping.
|
|
"resumePreviousJobId": "A String", # Optional. The prior hyperparameter tuning job id that users hope to
|
|
# continue with. The job id will be used to find the corresponding vizier
|
|
# study guid and resume the study.
|
|
"params": [ # Required. The set of parameters to tune.
|
|
{ # Represents a single hyperparameter to optimize.
|
|
"maxValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
|
|
# should be unset if type is `CATEGORICAL`. This value should be integers if
|
|
# type is `INTEGER`.
|
|
"categoricalValues": [ # Required if type is `CATEGORICAL`. The list of possible categories.
|
|
"A String",
|
|
],
|
|
"discreteValues": [ # Required if type is `DISCRETE`.
|
|
# A list of feasible points.
|
|
# The list should be in strictly increasing order. For instance, this
|
|
# parameter might have possible settings of 1.5, 2.5, and 4.0. This list
|
|
# should not contain more than 1,000 values.
|
|
3.14,
|
|
],
|
|
"parameterName": "A String", # Required. The parameter name must be unique amongst all ParameterConfigs in
|
|
# a HyperparameterSpec message. E.g., "learning_rate".
|
|
"minValue": 3.14, # Required if type is `DOUBLE` or `INTEGER`. This field
|
|
# should be unset if type is `CATEGORICAL`. This value should be integers if
|
|
# type is INTEGER.
|
|
"type": "A String", # Required. The type of the parameter.
|
|
"scaleType": "A String", # Optional. How the parameter should be scaled to the hypercube.
|
|
# Leave unset for categorical parameters.
|
|
# Some kind of scaling is strongly recommended for real or integral
|
|
# parameters (e.g., `UNIT_LINEAR_SCALE`).
|
|
},
|
|
],
|
|
"hyperparameterMetricTag": "A String", # Optional. The TensorFlow summary tag name to use for optimizing trials. For
|
|
# current versions of TensorFlow, this tag name should exactly match what is
|
|
# shown in TensorBoard, including all scopes. For versions of TensorFlow
|
|
# prior to 0.12, this should be only the tag passed to tf.Summary.
|
|
# By default, "training/hptuning/metric" will be used.
|
|
"maxParallelTrials": 42, # Optional. The number of training trials to run concurrently.
|
|
# You can reduce the time it takes to perform hyperparameter tuning by adding
|
|
# trials in parallel. However, each trail only benefits from the information
|
|
# gained in completed trials. That means that a trial does not get access to
|
|
# the results of trials running at the same time, which could reduce the
|
|
# quality of the overall optimization.
|
|
#
|
|
# Each trial will use the same scale tier and machine types.
|
|
#
|
|
# Defaults to one.
|
|
},
|
|
"region": "A String", # Required. The Google Compute Engine region to run the training job in.
|
|
# See the <a href="/ml-engine/docs/tensorflow/regions">available regions</a>
|
|
# for AI Platform services.
|
|
"args": [ # Optional. Command line arguments to pass to the program.
|
|
"A String",
|
|
],
|
|
"pythonModule": "A String", # Required. The Python module name to run after installing the packages.
|
|
"pythonVersion": "A String", # Optional. The version of Python used in training. If not set, the default
|
|
# version is '2.7'. Python '3.5' is available when `runtime_version` is set
|
|
# to '1.4' and above. Python '2.7' works with all supported
|
|
# <a href="/ml-engine/docs/runtime-version-list">runtime versions</a>.
|
|
"jobDir": "A String", # Optional. A Google Cloud Storage path in which to store training outputs
|
|
# and other data needed for training. This path is passed to your TensorFlow
|
|
# program as the '--job-dir' command-line argument. The benefit of specifying
|
|
# this field is that Cloud ML validates the path for use in training.
|
|
"packageUris": [ # Required. The Google Cloud Storage location of the packages with
|
|
# the training program and any additional dependencies.
|
|
# The maximum number of package URIs is 100.
|
|
"A String",
|
|
],
|
|
"workerCount": "A String", # Optional. The number of worker replicas to use for the training job. Each
|
|
# replica in the cluster will be of the type specified in `worker_type`.
|
|
#
|
|
# This value can only be used when `scale_tier` is set to `CUSTOM`. If you
|
|
# set this value, you must also set `worker_type`.
|
|
#
|
|
# The default value is zero.
|
|
"parameterServerType": "A String", # Optional. Specifies the type of virtual machine to use for your training
|
|
# job's parameter server.
|
|
#
|
|
# The supported values are the same as those described in the entry for
|
|
# `master_type`.
|
|
#
|
|
# This value must be consistent with the category of machine type that
|
|
# `masterType` uses. In other words, both must be AI Platform machine
|
|
# types or both must be Compute Engine machine types.
|
|
#
|
|
# This value must be present when `scaleTier` is set to `CUSTOM` and
|
|
# `parameter_server_count` is greater than zero.
|
|
"workerConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for workers.
|
|
#
|
|
# You should only set `workerConfig.acceleratorConfig` if `workerType` is set
|
|
# to a Compute Engine machine type. [Learn about restrictions on accelerator
|
|
# configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
#
|
|
# Set `workerConfig.imageUri` only if you build a custom image for your
|
|
# worker. If `workerConfig.imageUri` has not been set, AI Platform uses
|
|
# the value of `masterConfig.imageUri`. Learn more about
|
|
# [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
"acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
|
|
# [Learn about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
"count": "A String", # The number of accelerators to attach to each machine running the job.
|
|
"type": "A String", # The type of accelerator to use.
|
|
},
|
|
"imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
|
|
# Registry. Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
},
|
|
"maxRunningTime": "A String", # Optional. The maximum job running time. The default is 7 days.
|
|
"masterConfig": { # Represents the configuration for a replica in a cluster. # Optional. The configuration for your master worker.
|
|
#
|
|
# You should only set `masterConfig.acceleratorConfig` if `masterType` is set
|
|
# to a Compute Engine machine type. Learn about [restrictions on accelerator
|
|
# configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
#
|
|
# Set `masterConfig.imageUri` only if you build a custom image. Only one of
|
|
# `masterConfig.imageUri` and `runtimeVersion` should be set. Learn more about
|
|
# [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
"acceleratorConfig": { # Represents a hardware accelerator request config. # Represents the type and number of accelerators used by the replica.
|
|
# [Learn about restrictions on accelerator configurations for
|
|
# training.](/ml-engine/docs/tensorflow/using-gpus#compute-engine-machine-types-with-gpu)
|
|
"count": "A String", # The number of accelerators to attach to each machine running the job.
|
|
"type": "A String", # The type of accelerator to use.
|
|
},
|
|
"imageUri": "A String", # The Docker image to run on the replica. This image must be in Container
|
|
# Registry. Learn more about [configuring custom
|
|
# containers](/ml-engine/docs/distributed-training-containers).
|
|
},
|
|
"parameterServerCount": "A String", # Optional. The number of parameter server replicas to use for the training
|
|
# job. Each replica in the cluster will be of the type specified in
|
|
# `parameter_server_type`.
|
|
#
|
|
# This value can only be used when `scale_tier` is set to `CUSTOM`.If you
|
|
# set this value, you must also set `parameter_server_type`.
|
|
#
|
|
# The default value is zero.
|
|
},
|
|
"jobId": "A String", # Required. The user-specified id of the job.
|
|
"labels": { # Optional. One or more labels that you can add, to organize your jobs.
|
|
# Each label is a key-value pair, where both the key and the value are
|
|
# arbitrary strings that you supply.
|
|
# For more information, see the documentation on
|
|
# <a href="/ml-engine/docs/tensorflow/resource-labels">using labels</a>.
|
|
"a_key": "A String",
|
|
},
|
|
"state": "A String", # Output only. The detailed state of a job.
|
|
"etag": "A String", # `etag` is used for optimistic concurrency control as a way to help
|
|
# prevent simultaneous updates of a job from overwriting each other.
|
|
# It is strongly suggested that systems make use of the `etag` in the
|
|
# read-modify-write cycle to perform job updates in order to avoid race
|
|
# conditions: An `etag` is returned in the response to `GetJob`, and
|
|
# systems are expected to put that etag in the request to `UpdateJob` to
|
|
# ensure that their change will be applied to the same version of the job.
|
|
"startTime": "A String", # Output only. When the job processing was started.
|
|
"endTime": "A String", # Output only. When the job processing was completed.
|
|
"predictionOutput": { # Represents results of a prediction job. # The current prediction job result.
|
|
"outputPath": "A String", # The output Google Cloud Storage location provided at the job creation time.
|
|
"nodeHours": 3.14, # Node hours used by the batch prediction job.
|
|
"predictionCount": "A String", # The number of generated predictions.
|
|
"errorCount": "A String", # The number of data instances which resulted in errors.
|
|
},
|
|
"createTime": "A String", # Output only. When the job was created.
|
|
}</pre>
|
|
</div>
|
|
|
|
<div class="method">
|
|
<code class="details" id="setIamPolicy">setIamPolicy(resource, body, x__xgafv=None)</code>
|
|
<pre>Sets the access control policy on the specified resource. Replaces any
|
|
existing policy.
|
|
|
|
Args:
|
|
resource: string, REQUIRED: The resource for which the policy is being specified.
|
|
See the operation documentation for the appropriate value for this field. (required)
|
|
body: object, The request body. (required)
|
|
The object takes the form of:
|
|
|
|
{ # Request message for `SetIamPolicy` method.
|
|
"policy": { # Defines an Identity and Access Management (IAM) policy. It is used to # REQUIRED: The complete policy to be applied to the `resource`. The size of
|
|
# the policy is limited to a few 10s of KB. An empty policy is a
|
|
# valid policy but certain Cloud Platform services (such as Projects)
|
|
# might reject them.
|
|
# specify access control policies for Cloud Platform resources.
|
|
#
|
|
#
|
|
# A `Policy` consists of a list of `bindings`. A `binding` binds a list of
|
|
# `members` to a `role`, where the members can be user accounts, Google groups,
|
|
# Google domains, and service accounts. A `role` is a named list of permissions
|
|
# defined by IAM.
|
|
#
|
|
# **JSON Example**
|
|
#
|
|
# {
|
|
# "bindings": [
|
|
# {
|
|
# "role": "roles/owner",
|
|
# "members": [
|
|
# "user:mike@example.com",
|
|
# "group:admins@example.com",
|
|
# "domain:google.com",
|
|
# "serviceAccount:my-other-app@appspot.gserviceaccount.com"
|
|
# ]
|
|
# },
|
|
# {
|
|
# "role": "roles/viewer",
|
|
# "members": ["user:sean@example.com"]
|
|
# }
|
|
# ]
|
|
# }
|
|
#
|
|
# **YAML Example**
|
|
#
|
|
# bindings:
|
|
# - members:
|
|
# - user:mike@example.com
|
|
# - group:admins@example.com
|
|
# - domain:google.com
|
|
# - serviceAccount:my-other-app@appspot.gserviceaccount.com
|
|
# role: roles/owner
|
|
# - members:
|
|
# - user:sean@example.com
|
|
# role: roles/viewer
|
|
#
|
|
#
|
|
# For a description of IAM and its features, see the
|
|
# [IAM developer's guide](https://cloud.google.com/iam/docs).
|
|
"bindings": [ # Associates a list of `members` to a `role`.
|
|
# `bindings` with no members will result in an error.
|
|
{ # Associates `members` with a `role`.
|
|
"role": "A String", # Role that is assigned to `members`.
|
|
# For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
|
|
"members": [ # Specifies the identities requesting access for a Cloud Platform resource.
|
|
# `members` can have the following values:
|
|
#
|
|
# * `allUsers`: A special identifier that represents anyone who is
|
|
# on the internet; with or without a Google account.
|
|
#
|
|
# * `allAuthenticatedUsers`: A special identifier that represents anyone
|
|
# who is authenticated with a Google account or a service account.
|
|
#
|
|
# * `user:{emailid}`: An email address that represents a specific Google
|
|
# account. For example, `alice@gmail.com` .
|
|
#
|
|
#
|
|
# * `serviceAccount:{emailid}`: An email address that represents a service
|
|
# account. For example, `my-other-app@appspot.gserviceaccount.com`.
|
|
#
|
|
# * `group:{emailid}`: An email address that represents a Google group.
|
|
# For example, `admins@example.com`.
|
|
#
|
|
#
|
|
# * `domain:{domain}`: The G Suite domain (primary) that represents all the
|
|
# users of that domain. For example, `google.com` or `example.com`.
|
|
#
|
|
"A String",
|
|
],
|
|
"condition": { # Represents an expression text. Example: # The condition that is associated with this binding.
|
|
# NOTE: An unsatisfied condition will not allow user access via current
|
|
# binding. Different bindings, including their conditions, are examined
|
|
# independently.
|
|
#
|
|
# title: "User account presence"
|
|
# description: "Determines whether the request has a user account"
|
|
# expression: "size(request.user) > 0"
|
|
"description": "A String", # An optional description of the expression. This is a longer text which
|
|
# describes the expression, e.g. when hovered over it in a UI.
|
|
"expression": "A String", # Textual representation of an expression in
|
|
# Common Expression Language syntax.
|
|
#
|
|
# The application context of the containing message determines which
|
|
# well-known feature set of CEL is supported.
|
|
"location": "A String", # An optional string indicating the location of the expression for error
|
|
# reporting, e.g. a file name and a position in the file.
|
|
"title": "A String", # An optional title for the expression, i.e. a short string describing
|
|
# its purpose. This can be used e.g. in UIs which allow to enter the
|
|
# expression.
|
|
},
|
|
},
|
|
],
|
|
"etag": "A String", # `etag` is used for optimistic concurrency control as a way to help
|
|
# prevent simultaneous updates of a policy from overwriting each other.
|
|
# It is strongly suggested that systems make use of the `etag` in the
|
|
# read-modify-write cycle to perform policy updates in order to avoid race
|
|
# conditions: An `etag` is returned in the response to `getIamPolicy`, and
|
|
# systems are expected to put that etag in the request to `setIamPolicy` to
|
|
# ensure that their change will be applied to the same version of the policy.
|
|
#
|
|
# If no `etag` is provided in the call to `setIamPolicy`, then the existing
|
|
# policy is overwritten blindly.
|
|
"version": 42, # Deprecated.
|
|
"auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
|
|
{ # Specifies the audit configuration for a service.
|
|
# The configuration determines which permission types are logged, and what
|
|
# identities, if any, are exempted from logging.
|
|
# An AuditConfig must have one or more AuditLogConfigs.
|
|
#
|
|
# If there are AuditConfigs for both `allServices` and a specific service,
|
|
# the union of the two AuditConfigs is used for that service: the log_types
|
|
# specified in each AuditConfig are enabled, and the exempted_members in each
|
|
# AuditLogConfig are exempted.
|
|
#
|
|
# Example Policy with multiple AuditConfigs:
|
|
#
|
|
# {
|
|
# "audit_configs": [
|
|
# {
|
|
# "service": "allServices"
|
|
# "audit_log_configs": [
|
|
# {
|
|
# "log_type": "DATA_READ",
|
|
# "exempted_members": [
|
|
# "user:foo@gmail.com"
|
|
# ]
|
|
# },
|
|
# {
|
|
# "log_type": "DATA_WRITE",
|
|
# },
|
|
# {
|
|
# "log_type": "ADMIN_READ",
|
|
# }
|
|
# ]
|
|
# },
|
|
# {
|
|
# "service": "fooservice.googleapis.com"
|
|
# "audit_log_configs": [
|
|
# {
|
|
# "log_type": "DATA_READ",
|
|
# },
|
|
# {
|
|
# "log_type": "DATA_WRITE",
|
|
# "exempted_members": [
|
|
# "user:bar@gmail.com"
|
|
# ]
|
|
# }
|
|
# ]
|
|
# }
|
|
# ]
|
|
# }
|
|
#
|
|
# For fooservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ
|
|
# logging. It also exempts foo@gmail.com from DATA_READ logging, and
|
|
# bar@gmail.com from DATA_WRITE logging.
|
|
"auditLogConfigs": [ # The configuration for logging of each type of permission.
|
|
{ # Provides the configuration for logging a type of permissions.
|
|
# Example:
|
|
#
|
|
# {
|
|
# "audit_log_configs": [
|
|
# {
|
|
# "log_type": "DATA_READ",
|
|
# "exempted_members": [
|
|
# "user:foo@gmail.com"
|
|
# ]
|
|
# },
|
|
# {
|
|
# "log_type": "DATA_WRITE",
|
|
# }
|
|
# ]
|
|
# }
|
|
#
|
|
# This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting
|
|
# foo@gmail.com from DATA_READ logging.
|
|
"exemptedMembers": [ # Specifies the identities that do not cause logging for this type of
|
|
# permission.
|
|
# Follows the same format of Binding.members.
|
|
"A String",
|
|
],
|
|
"logType": "A String", # The log type that this config enables.
|
|
},
|
|
],
|
|
"service": "A String", # Specifies a service that will be enabled for audit logging.
|
|
# For example, `storage.googleapis.com`, `cloudsql.googleapis.com`.
|
|
# `allServices` is a special value that covers all services.
|
|
},
|
|
],
|
|
},
|
|
"updateMask": "A String", # OPTIONAL: A FieldMask specifying which fields of the policy to modify. Only
|
|
# the fields in the mask will be modified. If no mask is provided, the
|
|
# following default mask is used:
|
|
# paths: "bindings, etag"
|
|
# This field is only used by Cloud IAM.
|
|
}
|
|
|
|
x__xgafv: string, V1 error format.
|
|
Allowed values
|
|
1 - v1 error format
|
|
2 - v2 error format
|
|
|
|
Returns:
|
|
An object of the form:
|
|
|
|
{ # Defines an Identity and Access Management (IAM) policy. It is used to
|
|
# specify access control policies for Cloud Platform resources.
|
|
#
|
|
#
|
|
# A `Policy` consists of a list of `bindings`. A `binding` binds a list of
|
|
# `members` to a `role`, where the members can be user accounts, Google groups,
|
|
# Google domains, and service accounts. A `role` is a named list of permissions
|
|
# defined by IAM.
|
|
#
|
|
# **JSON Example**
|
|
#
|
|
# {
|
|
# "bindings": [
|
|
# {
|
|
# "role": "roles/owner",
|
|
# "members": [
|
|
# "user:mike@example.com",
|
|
# "group:admins@example.com",
|
|
# "domain:google.com",
|
|
# "serviceAccount:my-other-app@appspot.gserviceaccount.com"
|
|
# ]
|
|
# },
|
|
# {
|
|
# "role": "roles/viewer",
|
|
# "members": ["user:sean@example.com"]
|
|
# }
|
|
# ]
|
|
# }
|
|
#
|
|
# **YAML Example**
|
|
#
|
|
# bindings:
|
|
# - members:
|
|
# - user:mike@example.com
|
|
# - group:admins@example.com
|
|
# - domain:google.com
|
|
# - serviceAccount:my-other-app@appspot.gserviceaccount.com
|
|
# role: roles/owner
|
|
# - members:
|
|
# - user:sean@example.com
|
|
# role: roles/viewer
|
|
#
|
|
#
|
|
# For a description of IAM and its features, see the
|
|
# [IAM developer's guide](https://cloud.google.com/iam/docs).
|
|
"bindings": [ # Associates a list of `members` to a `role`.
|
|
# `bindings` with no members will result in an error.
|
|
{ # Associates `members` with a `role`.
|
|
"role": "A String", # Role that is assigned to `members`.
|
|
# For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
|
|
"members": [ # Specifies the identities requesting access for a Cloud Platform resource.
|
|
# `members` can have the following values:
|
|
#
|
|
# * `allUsers`: A special identifier that represents anyone who is
|
|
# on the internet; with or without a Google account.
|
|
#
|
|
# * `allAuthenticatedUsers`: A special identifier that represents anyone
|
|
# who is authenticated with a Google account or a service account.
|
|
#
|
|
# * `user:{emailid}`: An email address that represents a specific Google
|
|
# account. For example, `alice@gmail.com` .
|
|
#
|
|
#
|
|
# * `serviceAccount:{emailid}`: An email address that represents a service
|
|
# account. For example, `my-other-app@appspot.gserviceaccount.com`.
|
|
#
|
|
# * `group:{emailid}`: An email address that represents a Google group.
|
|
# For example, `admins@example.com`.
|
|
#
|
|
#
|
|
# * `domain:{domain}`: The G Suite domain (primary) that represents all the
|
|
# users of that domain. For example, `google.com` or `example.com`.
|
|
#
|
|
"A String",
|
|
],
|
|
"condition": { # Represents an expression text. Example: # The condition that is associated with this binding.
|
|
# NOTE: An unsatisfied condition will not allow user access via current
|
|
# binding. Different bindings, including their conditions, are examined
|
|
# independently.
|
|
#
|
|
# title: "User account presence"
|
|
# description: "Determines whether the request has a user account"
|
|
# expression: "size(request.user) > 0"
|
|
"description": "A String", # An optional description of the expression. This is a longer text which
|
|
# describes the expression, e.g. when hovered over it in a UI.
|
|
"expression": "A String", # Textual representation of an expression in
|
|
# Common Expression Language syntax.
|
|
#
|
|
# The application context of the containing message determines which
|
|
# well-known feature set of CEL is supported.
|
|
"location": "A String", # An optional string indicating the location of the expression for error
|
|
# reporting, e.g. a file name and a position in the file.
|
|
"title": "A String", # An optional title for the expression, i.e. a short string describing
|
|
# its purpose. This can be used e.g. in UIs which allow to enter the
|
|
# expression.
|
|
},
|
|
},
|
|
],
|
|
"etag": "A String", # `etag` is used for optimistic concurrency control as a way to help
|
|
# prevent simultaneous updates of a policy from overwriting each other.
|
|
# It is strongly suggested that systems make use of the `etag` in the
|
|
# read-modify-write cycle to perform policy updates in order to avoid race
|
|
# conditions: An `etag` is returned in the response to `getIamPolicy`, and
|
|
# systems are expected to put that etag in the request to `setIamPolicy` to
|
|
# ensure that their change will be applied to the same version of the policy.
|
|
#
|
|
# If no `etag` is provided in the call to `setIamPolicy`, then the existing
|
|
# policy is overwritten blindly.
|
|
"version": 42, # Deprecated.
|
|
"auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
|
|
{ # Specifies the audit configuration for a service.
|
|
# The configuration determines which permission types are logged, and what
|
|
# identities, if any, are exempted from logging.
|
|
# An AuditConfig must have one or more AuditLogConfigs.
|
|
#
|
|
# If there are AuditConfigs for both `allServices` and a specific service,
|
|
# the union of the two AuditConfigs is used for that service: the log_types
|
|
# specified in each AuditConfig are enabled, and the exempted_members in each
|
|
# AuditLogConfig are exempted.
|
|
#
|
|
# Example Policy with multiple AuditConfigs:
|
|
#
|
|
# {
|
|
# "audit_configs": [
|
|
# {
|
|
# "service": "allServices"
|
|
# "audit_log_configs": [
|
|
# {
|
|
# "log_type": "DATA_READ",
|
|
# "exempted_members": [
|
|
# "user:foo@gmail.com"
|
|
# ]
|
|
# },
|
|
# {
|
|
# "log_type": "DATA_WRITE",
|
|
# },
|
|
# {
|
|
# "log_type": "ADMIN_READ",
|
|
# }
|
|
# ]
|
|
# },
|
|
# {
|
|
# "service": "fooservice.googleapis.com"
|
|
# "audit_log_configs": [
|
|
# {
|
|
# "log_type": "DATA_READ",
|
|
# },
|
|
# {
|
|
# "log_type": "DATA_WRITE",
|
|
# "exempted_members": [
|
|
# "user:bar@gmail.com"
|
|
# ]
|
|
# }
|
|
# ]
|
|
# }
|
|
# ]
|
|
# }
|
|
#
|
|
# For fooservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ
|
|
# logging. It also exempts foo@gmail.com from DATA_READ logging, and
|
|
# bar@gmail.com from DATA_WRITE logging.
|
|
"auditLogConfigs": [ # The configuration for logging of each type of permission.
|
|
{ # Provides the configuration for logging a type of permissions.
|
|
# Example:
|
|
#
|
|
# {
|
|
# "audit_log_configs": [
|
|
# {
|
|
# "log_type": "DATA_READ",
|
|
# "exempted_members": [
|
|
# "user:foo@gmail.com"
|
|
# ]
|
|
# },
|
|
# {
|
|
# "log_type": "DATA_WRITE",
|
|
# }
|
|
# ]
|
|
# }
|
|
#
|
|
# This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting
|
|
# foo@gmail.com from DATA_READ logging.
|
|
"exemptedMembers": [ # Specifies the identities that do not cause logging for this type of
|
|
# permission.
|
|
# Follows the same format of Binding.members.
|
|
"A String",
|
|
],
|
|
"logType": "A String", # The log type that this config enables.
|
|
},
|
|
],
|
|
"service": "A String", # Specifies a service that will be enabled for audit logging.
|
|
# For example, `storage.googleapis.com`, `cloudsql.googleapis.com`.
|
|
# `allServices` is a special value that covers all services.
|
|
},
|
|
],
|
|
}</pre>
|
|
</div>
|
|
|
|
<div class="method">
|
|
<code class="details" id="testIamPermissions">testIamPermissions(resource, body, x__xgafv=None)</code>
|
|
<pre>Returns permissions that a caller has on the specified resource.
|
|
If the resource does not exist, this will return an empty set of
|
|
permissions, not a NOT_FOUND error.
|
|
|
|
Note: This operation is designed to be used for building permission-aware
|
|
UIs and command-line tools, not for authorization checking. This operation
|
|
may "fail open" without warning.
|
|
|
|
Args:
|
|
resource: string, REQUIRED: The resource for which the policy detail is being requested.
|
|
See the operation documentation for the appropriate value for this field. (required)
|
|
body: object, The request body. (required)
|
|
The object takes the form of:
|
|
|
|
{ # Request message for `TestIamPermissions` method.
|
|
"permissions": [ # The set of permissions to check for the `resource`. Permissions with
|
|
# wildcards (such as '*' or 'storage.*') are not allowed. For more
|
|
# information see
|
|
# [IAM Overview](https://cloud.google.com/iam/docs/overview#permissions).
|
|
"A String",
|
|
],
|
|
}
|
|
|
|
x__xgafv: string, V1 error format.
|
|
Allowed values
|
|
1 - v1 error format
|
|
2 - v2 error format
|
|
|
|
Returns:
|
|
An object of the form:
|
|
|
|
{ # Response message for `TestIamPermissions` method.
|
|
"permissions": [ # A subset of `TestPermissionsRequest.permissions` that the caller is
|
|
# allowed.
|
|
"A String",
|
|
],
|
|
}</pre>
|
|
</div>
|
|
|
|
</body></html> |