1 /* 2 * Copyright 2020 Google LLC 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * https://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 // Generated by the protocol buffer compiler. DO NOT EDIT! 17 // source: google/cloud/retail/v2/import_config.proto 18 19 package com.google.cloud.retail.v2; 20 21 public interface BigQuerySourceOrBuilder 22 extends 23 // @@protoc_insertion_point(interface_extends:google.cloud.retail.v2.BigQuerySource) 24 com.google.protobuf.MessageOrBuilder { 25 26 /** 27 * 28 * 29 * <pre> 30 * BigQuery time partitioned table's _PARTITIONDATE in YYYY-MM-DD format. 31 * Only supported in 32 * [ImportProductsRequest][google.cloud.retail.v2.ImportProductsRequest]. 33 * </pre> 34 * 35 * <code>.google.type.Date partition_date = 6;</code> 36 * 37 * @return Whether the partitionDate field is set. 38 */ hasPartitionDate()39 boolean hasPartitionDate(); 40 /** 41 * 42 * 43 * <pre> 44 * BigQuery time partitioned table's _PARTITIONDATE in YYYY-MM-DD format. 45 * Only supported in 46 * [ImportProductsRequest][google.cloud.retail.v2.ImportProductsRequest]. 47 * </pre> 48 * 49 * <code>.google.type.Date partition_date = 6;</code> 50 * 51 * @return The partitionDate. 52 */ getPartitionDate()53 com.google.type.Date getPartitionDate(); 54 /** 55 * 56 * 57 * <pre> 58 * BigQuery time partitioned table's _PARTITIONDATE in YYYY-MM-DD format. 59 * Only supported in 60 * [ImportProductsRequest][google.cloud.retail.v2.ImportProductsRequest]. 61 * </pre> 62 * 63 * <code>.google.type.Date partition_date = 6;</code> 64 */ getPartitionDateOrBuilder()65 com.google.type.DateOrBuilder getPartitionDateOrBuilder(); 66 67 /** 68 * 69 * 70 * <pre> 71 * The project ID (can be project # or ID) that the BigQuery source is in with 72 * a length limit of 128 characters. If not specified, inherits the project 73 * ID from the parent request. 74 * </pre> 75 * 76 * <code>string project_id = 5;</code> 77 * 78 * @return The projectId. 79 */ getProjectId()80 java.lang.String getProjectId(); 81 /** 82 * 83 * 84 * <pre> 85 * The project ID (can be project # or ID) that the BigQuery source is in with 86 * a length limit of 128 characters. If not specified, inherits the project 87 * ID from the parent request. 88 * </pre> 89 * 90 * <code>string project_id = 5;</code> 91 * 92 * @return The bytes for projectId. 93 */ getProjectIdBytes()94 com.google.protobuf.ByteString getProjectIdBytes(); 95 96 /** 97 * 98 * 99 * <pre> 100 * Required. The BigQuery data set to copy the data from with a length limit 101 * of 1,024 characters. 102 * </pre> 103 * 104 * <code>string dataset_id = 1 [(.google.api.field_behavior) = REQUIRED];</code> 105 * 106 * @return The datasetId. 107 */ getDatasetId()108 java.lang.String getDatasetId(); 109 /** 110 * 111 * 112 * <pre> 113 * Required. The BigQuery data set to copy the data from with a length limit 114 * of 1,024 characters. 115 * </pre> 116 * 117 * <code>string dataset_id = 1 [(.google.api.field_behavior) = REQUIRED];</code> 118 * 119 * @return The bytes for datasetId. 120 */ getDatasetIdBytes()121 com.google.protobuf.ByteString getDatasetIdBytes(); 122 123 /** 124 * 125 * 126 * <pre> 127 * Required. The BigQuery table to copy the data from with a length limit of 128 * 1,024 characters. 129 * </pre> 130 * 131 * <code>string table_id = 2 [(.google.api.field_behavior) = REQUIRED];</code> 132 * 133 * @return The tableId. 134 */ getTableId()135 java.lang.String getTableId(); 136 /** 137 * 138 * 139 * <pre> 140 * Required. The BigQuery table to copy the data from with a length limit of 141 * 1,024 characters. 142 * </pre> 143 * 144 * <code>string table_id = 2 [(.google.api.field_behavior) = REQUIRED];</code> 145 * 146 * @return The bytes for tableId. 147 */ getTableIdBytes()148 com.google.protobuf.ByteString getTableIdBytes(); 149 150 /** 151 * 152 * 153 * <pre> 154 * Intermediate Cloud Storage directory used for the import with a length 155 * limit of 2,000 characters. Can be specified if one wants to have the 156 * BigQuery export to a specific Cloud Storage directory. 157 * </pre> 158 * 159 * <code>string gcs_staging_dir = 3;</code> 160 * 161 * @return The gcsStagingDir. 162 */ getGcsStagingDir()163 java.lang.String getGcsStagingDir(); 164 /** 165 * 166 * 167 * <pre> 168 * Intermediate Cloud Storage directory used for the import with a length 169 * limit of 2,000 characters. Can be specified if one wants to have the 170 * BigQuery export to a specific Cloud Storage directory. 171 * </pre> 172 * 173 * <code>string gcs_staging_dir = 3;</code> 174 * 175 * @return The bytes for gcsStagingDir. 176 */ getGcsStagingDirBytes()177 com.google.protobuf.ByteString getGcsStagingDirBytes(); 178 179 /** 180 * 181 * 182 * <pre> 183 * The schema to use when parsing the data from the source. 184 * Supported values for product imports: 185 * * `product` (default): One JSON [Product][google.cloud.retail.v2.Product] 186 * per line. Each product must 187 * have a valid [Product.id][google.cloud.retail.v2.Product.id]. 188 * * `product_merchant_center`: See [Importing catalog data from Merchant 189 * Center](https://cloud.google.com/retail/recommendations-ai/docs/upload-catalog#mc). 190 * Supported values for user events imports: 191 * * `user_event` (default): One JSON 192 * [UserEvent][google.cloud.retail.v2.UserEvent] per line. 193 * * `user_event_ga360`: 194 * The schema is available here: 195 * https://support.google.com/analytics/answer/3437719. 196 * * `user_event_ga4`: 197 * The schema is available here: 198 * https://support.google.com/analytics/answer/7029846. 199 * Supported values for autocomplete imports: 200 * * `suggestions` (default): One JSON completion suggestion per line. 201 * * `denylist`: One JSON deny suggestion per line. 202 * * `allowlist`: One JSON allow suggestion per line. 203 * </pre> 204 * 205 * <code>string data_schema = 4;</code> 206 * 207 * @return The dataSchema. 208 */ getDataSchema()209 java.lang.String getDataSchema(); 210 /** 211 * 212 * 213 * <pre> 214 * The schema to use when parsing the data from the source. 215 * Supported values for product imports: 216 * * `product` (default): One JSON [Product][google.cloud.retail.v2.Product] 217 * per line. Each product must 218 * have a valid [Product.id][google.cloud.retail.v2.Product.id]. 219 * * `product_merchant_center`: See [Importing catalog data from Merchant 220 * Center](https://cloud.google.com/retail/recommendations-ai/docs/upload-catalog#mc). 221 * Supported values for user events imports: 222 * * `user_event` (default): One JSON 223 * [UserEvent][google.cloud.retail.v2.UserEvent] per line. 224 * * `user_event_ga360`: 225 * The schema is available here: 226 * https://support.google.com/analytics/answer/3437719. 227 * * `user_event_ga4`: 228 * The schema is available here: 229 * https://support.google.com/analytics/answer/7029846. 230 * Supported values for autocomplete imports: 231 * * `suggestions` (default): One JSON completion suggestion per line. 232 * * `denylist`: One JSON deny suggestion per line. 233 * * `allowlist`: One JSON allow suggestion per line. 234 * </pre> 235 * 236 * <code>string data_schema = 4;</code> 237 * 238 * @return The bytes for dataSchema. 239 */ getDataSchemaBytes()240 com.google.protobuf.ByteString getDataSchemaBytes(); 241 getPartitionCase()242 public com.google.cloud.retail.v2.BigQuerySource.PartitionCase getPartitionCase(); 243 } 244