batch_prediction_job.proto 16 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301
  1. // Copyright 2022 Google LLC
  2. //
  3. // Licensed under the Apache License, Version 2.0 (the "License");
  4. // you may not use this file except in compliance with the License.
  5. // You may obtain a copy of the License at
  6. //
  7. // http://www.apache.org/licenses/LICENSE-2.0
  8. //
  9. // Unless required by applicable law or agreed to in writing, software
  10. // distributed under the License is distributed on an "AS IS" BASIS,
  11. // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  12. // See the License for the specific language governing permissions and
  13. // limitations under the License.
  14. syntax = "proto3";
  15. package google.cloud.aiplatform.v1;
  16. import "google/api/field_behavior.proto";
  17. import "google/api/resource.proto";
  18. import "google/cloud/aiplatform/v1/completion_stats.proto";
  19. import "google/cloud/aiplatform/v1/encryption_spec.proto";
  20. import "google/cloud/aiplatform/v1/explanation.proto";
  21. import "google/cloud/aiplatform/v1/io.proto";
  22. import "google/cloud/aiplatform/v1/job_state.proto";
  23. import "google/cloud/aiplatform/v1/machine_resources.proto";
  24. import "google/cloud/aiplatform/v1/manual_batch_tuning_parameters.proto";
  25. import "google/cloud/aiplatform/v1/unmanaged_container_model.proto";
  26. import "google/protobuf/struct.proto";
  27. import "google/protobuf/timestamp.proto";
  28. import "google/rpc/status.proto";
  29. option csharp_namespace = "Google.Cloud.AIPlatform.V1";
  30. option go_package = "google.golang.org/genproto/googleapis/cloud/aiplatform/v1;aiplatform";
  31. option java_multiple_files = true;
  32. option java_outer_classname = "BatchPredictionJobProto";
  33. option java_package = "com.google.cloud.aiplatform.v1";
  34. option php_namespace = "Google\\Cloud\\AIPlatform\\V1";
  35. option ruby_package = "Google::Cloud::AIPlatform::V1";
  36. // A job that uses a [Model][google.cloud.aiplatform.v1.BatchPredictionJob.model] to produce predictions
  37. // on multiple [input instances][google.cloud.aiplatform.v1.BatchPredictionJob.input_config]. If
  38. // predictions for significant portion of the instances fail, the job may finish
  39. // without attempting predictions for all remaining instances.
  40. message BatchPredictionJob {
  41. option (google.api.resource) = {
  42. type: "aiplatform.googleapis.com/BatchPredictionJob"
  43. pattern: "projects/{project}/locations/{location}/batchPredictionJobs/{batch_prediction_job}"
  44. };
  45. // Configures the input to [BatchPredictionJob][google.cloud.aiplatform.v1.BatchPredictionJob].
  46. // See [Model.supported_input_storage_formats][google.cloud.aiplatform.v1.Model.supported_input_storage_formats] for Model's supported input
  47. // formats, and how instances should be expressed via any of them.
  48. message InputConfig {
  49. // Required. The source of the input.
  50. oneof source {
  51. // The Cloud Storage location for the input instances.
  52. GcsSource gcs_source = 2;
  53. // The BigQuery location of the input table.
  54. // The schema of the table should be in the format described by the given
  55. // context OpenAPI Schema, if one is provided. The table may contain
  56. // additional columns that are not described by the schema, and they will
  57. // be ignored.
  58. BigQuerySource bigquery_source = 3;
  59. }
  60. // Required. The format in which instances are given, must be one of the
  61. // [Model's][google.cloud.aiplatform.v1.BatchPredictionJob.model]
  62. // [supported_input_storage_formats][google.cloud.aiplatform.v1.Model.supported_input_storage_formats].
  63. string instances_format = 1 [(google.api.field_behavior) = REQUIRED];
  64. }
  65. // Configures the output of [BatchPredictionJob][google.cloud.aiplatform.v1.BatchPredictionJob].
  66. // See [Model.supported_output_storage_formats][google.cloud.aiplatform.v1.Model.supported_output_storage_formats] for supported output
  67. // formats, and how predictions are expressed via any of them.
  68. message OutputConfig {
  69. // Required. The destination of the output.
  70. oneof destination {
  71. // The Cloud Storage location of the directory where the output is
  72. // to be written to. In the given directory a new directory is created.
  73. // Its name is `prediction-<model-display-name>-<job-create-time>`,
  74. // where timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format.
  75. // Inside of it files `predictions_0001.<extension>`,
  76. // `predictions_0002.<extension>`, ..., `predictions_N.<extension>`
  77. // are created where `<extension>` depends on chosen
  78. // [predictions_format][google.cloud.aiplatform.v1.BatchPredictionJob.OutputConfig.predictions_format], and N may equal 0001 and depends on the total
  79. // number of successfully predicted instances.
  80. // If the Model has both [instance][google.cloud.aiplatform.v1.PredictSchemata.instance_schema_uri]
  81. // and [prediction][google.cloud.aiplatform.v1.PredictSchemata.parameters_schema_uri] schemata
  82. // defined then each such file contains predictions as per the
  83. // [predictions_format][google.cloud.aiplatform.v1.BatchPredictionJob.OutputConfig.predictions_format].
  84. // If prediction for any instance failed (partially or completely), then
  85. // an additional `errors_0001.<extension>`, `errors_0002.<extension>`,...,
  86. // `errors_N.<extension>` files are created (N depends on total number
  87. // of failed predictions). These files contain the failed instances,
  88. // as per their schema, followed by an additional `error` field which as
  89. // value has [google.rpc.Status][google.rpc.Status]
  90. // containing only `code` and `message` fields.
  91. GcsDestination gcs_destination = 2;
  92. // The BigQuery project or dataset location where the output is to be
  93. // written to. If project is provided, a new dataset is created with name
  94. // `prediction_<model-display-name>_<job-create-time>`
  95. // where <model-display-name> is made
  96. // BigQuery-dataset-name compatible (for example, most special characters
  97. // become underscores), and timestamp is in
  98. // YYYY_MM_DDThh_mm_ss_sssZ "based on ISO-8601" format. In the dataset
  99. // two tables will be created, `predictions`, and `errors`.
  100. // If the Model has both [instance][google.cloud.aiplatform.v1.PredictSchemata.instance_schema_uri]
  101. // and [prediction][google.cloud.aiplatform.v1.PredictSchemata.parameters_schema_uri] schemata
  102. // defined then the tables have columns as follows: The `predictions`
  103. // table contains instances for which the prediction succeeded, it
  104. // has columns as per a concatenation of the Model's instance and
  105. // prediction schemata. The `errors` table contains rows for which the
  106. // prediction has failed, it has instance columns, as per the
  107. // instance schema, followed by a single "errors" column, which as values
  108. // has [google.rpc.Status][google.rpc.Status]
  109. // represented as a STRUCT, and containing only `code` and `message`.
  110. BigQueryDestination bigquery_destination = 3;
  111. }
  112. // Required. The format in which Vertex AI gives the predictions, must be one of the
  113. // [Model's][google.cloud.aiplatform.v1.BatchPredictionJob.model]
  114. // [supported_output_storage_formats][google.cloud.aiplatform.v1.Model.supported_output_storage_formats].
  115. string predictions_format = 1 [(google.api.field_behavior) = REQUIRED];
  116. }
  117. // Further describes this job's output.
  118. // Supplements [output_config][google.cloud.aiplatform.v1.BatchPredictionJob.output_config].
  119. message OutputInfo {
  120. // The output location into which prediction output is written.
  121. oneof output_location {
  122. // Output only. The full path of the Cloud Storage directory created, into which
  123. // the prediction output is written.
  124. string gcs_output_directory = 1 [(google.api.field_behavior) = OUTPUT_ONLY];
  125. // Output only. The path of the BigQuery dataset created, in
  126. // `bq://projectId.bqDatasetId`
  127. // format, into which the prediction output is written.
  128. string bigquery_output_dataset = 2 [(google.api.field_behavior) = OUTPUT_ONLY];
  129. }
  130. // Output only. The name of the BigQuery table created, in
  131. // `predictions_<timestamp>`
  132. // format, into which the prediction output is written.
  133. // Can be used by UI to generate the BigQuery output path, for example.
  134. string bigquery_output_table = 4 [(google.api.field_behavior) = OUTPUT_ONLY];
  135. }
  136. // Output only. Resource name of the BatchPredictionJob.
  137. string name = 1 [(google.api.field_behavior) = OUTPUT_ONLY];
  138. // Required. The user-defined name of this BatchPredictionJob.
  139. string display_name = 2 [(google.api.field_behavior) = REQUIRED];
  140. // The name of the Model resource that produces the predictions via this job,
  141. // must share the same ancestor Location.
  142. // Starting this job has no impact on any existing deployments of the Model
  143. // and their resources.
  144. // Exactly one of model and unmanaged_container_model must be set.
  145. //
  146. // The model resource name may contain version id or version alias to specify
  147. // the version, if no version is specified, the default version will be used.
  148. string model = 3 [(google.api.resource_reference) = {
  149. type: "aiplatform.googleapis.com/Model"
  150. }];
  151. // Output only. The version ID of the Model that produces the predictions via this job.
  152. string model_version_id = 30 [(google.api.field_behavior) = OUTPUT_ONLY];
  153. // Contains model information necessary to perform batch prediction without
  154. // requiring uploading to model registry.
  155. // Exactly one of model and unmanaged_container_model must be set.
  156. UnmanagedContainerModel unmanaged_container_model = 28;
  157. // Required. Input configuration of the instances on which predictions are performed.
  158. // The schema of any single instance may be specified via
  159. // the [Model's][google.cloud.aiplatform.v1.BatchPredictionJob.model]
  160. // [PredictSchemata's][google.cloud.aiplatform.v1.Model.predict_schemata]
  161. // [instance_schema_uri][google.cloud.aiplatform.v1.PredictSchemata.instance_schema_uri].
  162. InputConfig input_config = 4 [(google.api.field_behavior) = REQUIRED];
  163. // The parameters that govern the predictions. The schema of the parameters
  164. // may be specified via the [Model's][google.cloud.aiplatform.v1.BatchPredictionJob.model]
  165. // [PredictSchemata's][google.cloud.aiplatform.v1.Model.predict_schemata]
  166. // [parameters_schema_uri][google.cloud.aiplatform.v1.PredictSchemata.parameters_schema_uri].
  167. google.protobuf.Value model_parameters = 5;
  168. // Required. The Configuration specifying where output predictions should
  169. // be written.
  170. // The schema of any single prediction may be specified as a concatenation
  171. // of [Model's][google.cloud.aiplatform.v1.BatchPredictionJob.model]
  172. // [PredictSchemata's][google.cloud.aiplatform.v1.Model.predict_schemata]
  173. // [instance_schema_uri][google.cloud.aiplatform.v1.PredictSchemata.instance_schema_uri]
  174. // and
  175. // [prediction_schema_uri][google.cloud.aiplatform.v1.PredictSchemata.prediction_schema_uri].
  176. OutputConfig output_config = 6 [(google.api.field_behavior) = REQUIRED];
  177. // The config of resources used by the Model during the batch prediction. If
  178. // the Model [supports][google.cloud.aiplatform.v1.Model.supported_deployment_resources_types]
  179. // DEDICATED_RESOURCES this config may be provided (and the job will use these
  180. // resources), if the Model doesn't support AUTOMATIC_RESOURCES, this config
  181. // must be provided.
  182. BatchDedicatedResources dedicated_resources = 7;
  183. // The service account that the DeployedModel's container runs as. If not
  184. // specified, a system generated one will be used, which
  185. // has minimal permissions and the custom container, if used, may not have
  186. // enough permission to access other GCP resources.
  187. //
  188. // Users deploying the Model must have the `iam.serviceAccounts.actAs`
  189. // permission on this service account.
  190. string service_account = 29;
  191. // Immutable. Parameters configuring the batch behavior. Currently only applicable when
  192. // [dedicated_resources][google.cloud.aiplatform.v1.BatchPredictionJob.dedicated_resources] are used (in other cases Vertex AI does
  193. // the tuning itself).
  194. ManualBatchTuningParameters manual_batch_tuning_parameters = 8 [(google.api.field_behavior) = IMMUTABLE];
  195. // Generate explanation with the batch prediction results.
  196. //
  197. // When set to `true`, the batch prediction output changes based on the
  198. // `predictions_format` field of the
  199. // [BatchPredictionJob.output_config][google.cloud.aiplatform.v1.BatchPredictionJob.output_config] object:
  200. //
  201. // * `bigquery`: output includes a column named `explanation`. The value
  202. // is a struct that conforms to the [Explanation][google.cloud.aiplatform.v1.Explanation] object.
  203. // * `jsonl`: The JSON objects on each line include an additional entry
  204. // keyed `explanation`. The value of the entry is a JSON object that
  205. // conforms to the [Explanation][google.cloud.aiplatform.v1.Explanation] object.
  206. // * `csv`: Generating explanations for CSV format is not supported.
  207. //
  208. // If this field is set to true, either the [Model.explanation_spec][google.cloud.aiplatform.v1.Model.explanation_spec] or
  209. // [explanation_spec][google.cloud.aiplatform.v1.BatchPredictionJob.explanation_spec] must be populated.
  210. bool generate_explanation = 23;
  211. // Explanation configuration for this BatchPredictionJob. Can be
  212. // specified only if [generate_explanation][google.cloud.aiplatform.v1.BatchPredictionJob.generate_explanation] is set to `true`.
  213. //
  214. // This value overrides the value of [Model.explanation_spec][google.cloud.aiplatform.v1.Model.explanation_spec]. All fields of
  215. // [explanation_spec][google.cloud.aiplatform.v1.BatchPredictionJob.explanation_spec] are optional in the request. If a field of the
  216. // [explanation_spec][google.cloud.aiplatform.v1.BatchPredictionJob.explanation_spec] object is not populated, the corresponding field of
  217. // the [Model.explanation_spec][google.cloud.aiplatform.v1.Model.explanation_spec] object is inherited.
  218. ExplanationSpec explanation_spec = 25;
  219. // Output only. Information further describing the output of this job.
  220. OutputInfo output_info = 9 [(google.api.field_behavior) = OUTPUT_ONLY];
  221. // Output only. The detailed state of the job.
  222. JobState state = 10 [(google.api.field_behavior) = OUTPUT_ONLY];
  223. // Output only. Only populated when the job's state is JOB_STATE_FAILED or
  224. // JOB_STATE_CANCELLED.
  225. google.rpc.Status error = 11 [(google.api.field_behavior) = OUTPUT_ONLY];
  226. // Output only. Partial failures encountered.
  227. // For example, single files that can't be read.
  228. // This field never exceeds 20 entries.
  229. // Status details fields contain standard GCP error details.
  230. repeated google.rpc.Status partial_failures = 12 [(google.api.field_behavior) = OUTPUT_ONLY];
  231. // Output only. Information about resources that had been consumed by this job.
  232. // Provided in real time at best effort basis, as well as a final value
  233. // once the job completes.
  234. //
  235. // Note: This field currently may be not populated for batch predictions that
  236. // use AutoML Models.
  237. ResourcesConsumed resources_consumed = 13 [(google.api.field_behavior) = OUTPUT_ONLY];
  238. // Output only. Statistics on completed and failed prediction instances.
  239. CompletionStats completion_stats = 14 [(google.api.field_behavior) = OUTPUT_ONLY];
  240. // Output only. Time when the BatchPredictionJob was created.
  241. google.protobuf.Timestamp create_time = 15 [(google.api.field_behavior) = OUTPUT_ONLY];
  242. // Output only. Time when the BatchPredictionJob for the first time entered the
  243. // `JOB_STATE_RUNNING` state.
  244. google.protobuf.Timestamp start_time = 16 [(google.api.field_behavior) = OUTPUT_ONLY];
  245. // Output only. Time when the BatchPredictionJob entered any of the following states:
  246. // `JOB_STATE_SUCCEEDED`, `JOB_STATE_FAILED`, `JOB_STATE_CANCELLED`.
  247. google.protobuf.Timestamp end_time = 17 [(google.api.field_behavior) = OUTPUT_ONLY];
  248. // Output only. Time when the BatchPredictionJob was most recently updated.
  249. google.protobuf.Timestamp update_time = 18 [(google.api.field_behavior) = OUTPUT_ONLY];
  250. // The labels with user-defined metadata to organize BatchPredictionJobs.
  251. //
  252. // Label keys and values can be no longer than 64 characters
  253. // (Unicode codepoints), can only contain lowercase letters, numeric
  254. // characters, underscores and dashes. International characters are allowed.
  255. //
  256. // See https://goo.gl/xmQnxf for more information and examples of labels.
  257. map<string, string> labels = 19;
  258. // Customer-managed encryption key options for a BatchPredictionJob. If this
  259. // is set, then all resources created by the BatchPredictionJob will be
  260. // encrypted with the provided encryption key.
  261. EncryptionSpec encryption_spec = 24;
  262. }