Google Cloud Native is in preview. Google Cloud Classic is fully supported.
google-native.aiplatform/v1beta1.Endpoint
Explore with Pulumi AI
Google Cloud Native is in preview. Google Cloud Classic is fully supported.
Creates an Endpoint. Auto-naming is currently not supported for this resource.
Create Endpoint Resource
Resources are created with functions called constructors. To learn more about declaring and configuring resources, see Resources.
Constructor syntax
new Endpoint(name: string, args: EndpointArgs, opts?: CustomResourceOptions);
@overload
def Endpoint(resource_name: str,
args: EndpointArgs,
opts: Optional[ResourceOptions] = None)
@overload
def Endpoint(resource_name: str,
opts: Optional[ResourceOptions] = None,
display_name: Optional[str] = None,
description: Optional[str] = None,
enable_private_service_connect: Optional[bool] = None,
encryption_spec: Optional[GoogleCloudAiplatformV1beta1EncryptionSpecArgs] = None,
endpoint_id: Optional[str] = None,
etag: Optional[str] = None,
labels: Optional[Mapping[str, str]] = None,
location: Optional[str] = None,
network: Optional[str] = None,
predict_request_response_logging_config: Optional[GoogleCloudAiplatformV1beta1PredictRequestResponseLoggingConfigArgs] = None,
project: Optional[str] = None,
traffic_split: Optional[Mapping[str, str]] = None)
func NewEndpoint(ctx *Context, name string, args EndpointArgs, opts ...ResourceOption) (*Endpoint, error)
public Endpoint(string name, EndpointArgs args, CustomResourceOptions? opts = null)
public Endpoint(String name, EndpointArgs args)
public Endpoint(String name, EndpointArgs args, CustomResourceOptions options)
type: google-native:aiplatform/v1beta1:Endpoint
properties: # The arguments to resource properties.
options: # Bag of options to control resource's behavior.
Parameters
- name string
- The unique name of the resource.
- args EndpointArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- resource_name str
- The unique name of the resource.
- args EndpointArgs
- The arguments to resource properties.
- opts ResourceOptions
- Bag of options to control resource's behavior.
- ctx Context
- Context object for the current deployment.
- name string
- The unique name of the resource.
- args EndpointArgs
- The arguments to resource properties.
- opts ResourceOption
- Bag of options to control resource's behavior.
- name string
- The unique name of the resource.
- args EndpointArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- name String
- The unique name of the resource.
- args EndpointArgs
- The arguments to resource properties.
- options CustomResourceOptions
- Bag of options to control resource's behavior.
Constructor example
The following reference example uses placeholder values for all input properties.
var google_nativeEndpointResource = new GoogleNative.Aiplatform.V1Beta1.Endpoint("google-nativeEndpointResource", new()
{
DisplayName = "string",
Description = "string",
EncryptionSpec = new GoogleNative.Aiplatform.V1Beta1.Inputs.GoogleCloudAiplatformV1beta1EncryptionSpecArgs
{
KmsKeyName = "string",
},
EndpointId = "string",
Etag = "string",
Labels =
{
{ "string", "string" },
},
Location = "string",
Network = "string",
PredictRequestResponseLoggingConfig = new GoogleNative.Aiplatform.V1Beta1.Inputs.GoogleCloudAiplatformV1beta1PredictRequestResponseLoggingConfigArgs
{
BigqueryDestination = new GoogleNative.Aiplatform.V1Beta1.Inputs.GoogleCloudAiplatformV1beta1BigQueryDestinationArgs
{
OutputUri = "string",
},
Enabled = false,
SamplingRate = 0,
},
Project = "string",
TrafficSplit =
{
{ "string", "string" },
},
});
example, err := aiplatformv1beta1.NewEndpoint(ctx, "google-nativeEndpointResource", &aiplatformv1beta1.EndpointArgs{
DisplayName: pulumi.String("string"),
Description: pulumi.String("string"),
EncryptionSpec: &aiplatform.GoogleCloudAiplatformV1beta1EncryptionSpecArgs{
KmsKeyName: pulumi.String("string"),
},
EndpointId: pulumi.String("string"),
Etag: pulumi.String("string"),
Labels: pulumi.StringMap{
"string": pulumi.String("string"),
},
Location: pulumi.String("string"),
Network: pulumi.String("string"),
PredictRequestResponseLoggingConfig: &aiplatform.GoogleCloudAiplatformV1beta1PredictRequestResponseLoggingConfigArgs{
BigqueryDestination: &aiplatform.GoogleCloudAiplatformV1beta1BigQueryDestinationArgs{
OutputUri: pulumi.String("string"),
},
Enabled: pulumi.Bool(false),
SamplingRate: pulumi.Float64(0),
},
Project: pulumi.String("string"),
TrafficSplit: pulumi.StringMap{
"string": pulumi.String("string"),
},
})
var google_nativeEndpointResource = new Endpoint("google-nativeEndpointResource", EndpointArgs.builder()
.displayName("string")
.description("string")
.encryptionSpec(GoogleCloudAiplatformV1beta1EncryptionSpecArgs.builder()
.kmsKeyName("string")
.build())
.endpointId("string")
.etag("string")
.labels(Map.of("string", "string"))
.location("string")
.network("string")
.predictRequestResponseLoggingConfig(GoogleCloudAiplatformV1beta1PredictRequestResponseLoggingConfigArgs.builder()
.bigqueryDestination(GoogleCloudAiplatformV1beta1BigQueryDestinationArgs.builder()
.outputUri("string")
.build())
.enabled(false)
.samplingRate(0)
.build())
.project("string")
.trafficSplit(Map.of("string", "string"))
.build());
google_native_endpoint_resource = google_native.aiplatform.v1beta1.Endpoint("google-nativeEndpointResource",
display_name="string",
description="string",
encryption_spec={
"kms_key_name": "string",
},
endpoint_id="string",
etag="string",
labels={
"string": "string",
},
location="string",
network="string",
predict_request_response_logging_config={
"bigquery_destination": {
"output_uri": "string",
},
"enabled": False,
"sampling_rate": 0,
},
project="string",
traffic_split={
"string": "string",
})
const google_nativeEndpointResource = new google_native.aiplatform.v1beta1.Endpoint("google-nativeEndpointResource", {
displayName: "string",
description: "string",
encryptionSpec: {
kmsKeyName: "string",
},
endpointId: "string",
etag: "string",
labels: {
string: "string",
},
location: "string",
network: "string",
predictRequestResponseLoggingConfig: {
bigqueryDestination: {
outputUri: "string",
},
enabled: false,
samplingRate: 0,
},
project: "string",
trafficSplit: {
string: "string",
},
});
type: google-native:aiplatform/v1beta1:Endpoint
properties:
description: string
displayName: string
encryptionSpec:
kmsKeyName: string
endpointId: string
etag: string
labels:
string: string
location: string
network: string
predictRequestResponseLoggingConfig:
bigqueryDestination:
outputUri: string
enabled: false
samplingRate: 0
project: string
trafficSplit:
string: string
Endpoint Resource Properties
To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.
Inputs
In Python, inputs that are objects can be passed either as argument classes or as dictionary literals.
The Endpoint resource accepts the following input properties:
- Display
Name string - The display name of the Endpoint. The name can be up to 128 characters long and can consist of any UTF-8 characters.
- Description string
- The description of the Endpoint.
- Enable
Private boolService Connect - Deprecated: If true, expose the Endpoint via private service connect. Only one of the fields, network or enable_private_service_connect, can be set.
- Encryption
Spec Pulumi.Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Encryption Spec - Customer-managed encryption key spec for an Endpoint. If set, this Endpoint and all sub-resources of this Endpoint will be secured by this key.
- Endpoint
Id string - Immutable. The ID to use for endpoint, which will become the final component of the endpoint resource name. If not provided, Vertex AI will generate a value for this ID. If the first character is a letter, this value may be up to 63 characters, and valid characters are
[a-z0-9-]
. The last character must be a letter or number. If the first character is a number, this value may be up to 9 characters, and valid characters are[0-9]
with no leading zeros. When using HTTP/JSON, this field is populated based on a query string argument, such as?endpoint_id=12345
. This is the fallback for fields that are not included in either the URI or the body. - Etag string
- Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
- Labels Dictionary<string, string>
- The labels with user-defined metadata to organize your Endpoints. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels.
- Location string
- Network string
- Optional. The full name of the Google Compute Engine network to which the Endpoint should be peered. Private services access must already be configured for the network. If left unspecified, the Endpoint is not peered with any network. Only one of the fields, network or enable_private_service_connect, can be set. Format:
projects/{project}/global/networks/{network}
. Where{project}
is a project number, as in12345
, and{network}
is network name. - Predict
Request Pulumi.Response Logging Config Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Predict Request Response Logging Config - Configures the request-response logging for online prediction.
- Project string
- Traffic
Split Dictionary<string, string> - A map from a DeployedModel's ID to the percentage of this Endpoint's traffic that should be forwarded to that DeployedModel. If a DeployedModel's ID is not listed in this map, then it receives no traffic. The traffic percentage values must add up to 100, or map must be empty if the Endpoint is to not accept any traffic at a moment.
- Display
Name string - The display name of the Endpoint. The name can be up to 128 characters long and can consist of any UTF-8 characters.
- Description string
- The description of the Endpoint.
- Enable
Private boolService Connect - Deprecated: If true, expose the Endpoint via private service connect. Only one of the fields, network or enable_private_service_connect, can be set.
- Encryption
Spec GoogleCloud Aiplatform V1beta1Encryption Spec Args - Customer-managed encryption key spec for an Endpoint. If set, this Endpoint and all sub-resources of this Endpoint will be secured by this key.
- Endpoint
Id string - Immutable. The ID to use for endpoint, which will become the final component of the endpoint resource name. If not provided, Vertex AI will generate a value for this ID. If the first character is a letter, this value may be up to 63 characters, and valid characters are
[a-z0-9-]
. The last character must be a letter or number. If the first character is a number, this value may be up to 9 characters, and valid characters are[0-9]
with no leading zeros. When using HTTP/JSON, this field is populated based on a query string argument, such as?endpoint_id=12345
. This is the fallback for fields that are not included in either the URI or the body. - Etag string
- Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
- Labels map[string]string
- The labels with user-defined metadata to organize your Endpoints. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels.
- Location string
- Network string
- Optional. The full name of the Google Compute Engine network to which the Endpoint should be peered. Private services access must already be configured for the network. If left unspecified, the Endpoint is not peered with any network. Only one of the fields, network or enable_private_service_connect, can be set. Format:
projects/{project}/global/networks/{network}
. Where{project}
is a project number, as in12345
, and{network}
is network name. - Predict
Request GoogleResponse Logging Config Cloud Aiplatform V1beta1Predict Request Response Logging Config Args - Configures the request-response logging for online prediction.
- Project string
- Traffic
Split map[string]string - A map from a DeployedModel's ID to the percentage of this Endpoint's traffic that should be forwarded to that DeployedModel. If a DeployedModel's ID is not listed in this map, then it receives no traffic. The traffic percentage values must add up to 100, or map must be empty if the Endpoint is to not accept any traffic at a moment.
- display
Name String - The display name of the Endpoint. The name can be up to 128 characters long and can consist of any UTF-8 characters.
- description String
- The description of the Endpoint.
- enable
Private BooleanService Connect - Deprecated: If true, expose the Endpoint via private service connect. Only one of the fields, network or enable_private_service_connect, can be set.
- encryption
Spec GoogleCloud Aiplatform V1beta1Encryption Spec - Customer-managed encryption key spec for an Endpoint. If set, this Endpoint and all sub-resources of this Endpoint will be secured by this key.
- endpoint
Id String - Immutable. The ID to use for endpoint, which will become the final component of the endpoint resource name. If not provided, Vertex AI will generate a value for this ID. If the first character is a letter, this value may be up to 63 characters, and valid characters are
[a-z0-9-]
. The last character must be a letter or number. If the first character is a number, this value may be up to 9 characters, and valid characters are[0-9]
with no leading zeros. When using HTTP/JSON, this field is populated based on a query string argument, such as?endpoint_id=12345
. This is the fallback for fields that are not included in either the URI or the body. - etag String
- Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
- labels Map<String,String>
- The labels with user-defined metadata to organize your Endpoints. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels.
- location String
- network String
- Optional. The full name of the Google Compute Engine network to which the Endpoint should be peered. Private services access must already be configured for the network. If left unspecified, the Endpoint is not peered with any network. Only one of the fields, network or enable_private_service_connect, can be set. Format:
projects/{project}/global/networks/{network}
. Where{project}
is a project number, as in12345
, and{network}
is network name. - predict
Request GoogleResponse Logging Config Cloud Aiplatform V1beta1Predict Request Response Logging Config - Configures the request-response logging for online prediction.
- project String
- traffic
Split Map<String,String> - A map from a DeployedModel's ID to the percentage of this Endpoint's traffic that should be forwarded to that DeployedModel. If a DeployedModel's ID is not listed in this map, then it receives no traffic. The traffic percentage values must add up to 100, or map must be empty if the Endpoint is to not accept any traffic at a moment.
- display
Name string - The display name of the Endpoint. The name can be up to 128 characters long and can consist of any UTF-8 characters.
- description string
- The description of the Endpoint.
- enable
Private booleanService Connect - Deprecated: If true, expose the Endpoint via private service connect. Only one of the fields, network or enable_private_service_connect, can be set.
- encryption
Spec GoogleCloud Aiplatform V1beta1Encryption Spec - Customer-managed encryption key spec for an Endpoint. If set, this Endpoint and all sub-resources of this Endpoint will be secured by this key.
- endpoint
Id string - Immutable. The ID to use for endpoint, which will become the final component of the endpoint resource name. If not provided, Vertex AI will generate a value for this ID. If the first character is a letter, this value may be up to 63 characters, and valid characters are
[a-z0-9-]
. The last character must be a letter or number. If the first character is a number, this value may be up to 9 characters, and valid characters are[0-9]
with no leading zeros. When using HTTP/JSON, this field is populated based on a query string argument, such as?endpoint_id=12345
. This is the fallback for fields that are not included in either the URI or the body. - etag string
- Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
- labels {[key: string]: string}
- The labels with user-defined metadata to organize your Endpoints. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels.
- location string
- network string
- Optional. The full name of the Google Compute Engine network to which the Endpoint should be peered. Private services access must already be configured for the network. If left unspecified, the Endpoint is not peered with any network. Only one of the fields, network or enable_private_service_connect, can be set. Format:
projects/{project}/global/networks/{network}
. Where{project}
is a project number, as in12345
, and{network}
is network name. - predict
Request GoogleResponse Logging Config Cloud Aiplatform V1beta1Predict Request Response Logging Config - Configures the request-response logging for online prediction.
- project string
- traffic
Split {[key: string]: string} - A map from a DeployedModel's ID to the percentage of this Endpoint's traffic that should be forwarded to that DeployedModel. If a DeployedModel's ID is not listed in this map, then it receives no traffic. The traffic percentage values must add up to 100, or map must be empty if the Endpoint is to not accept any traffic at a moment.
- display_
name str - The display name of the Endpoint. The name can be up to 128 characters long and can consist of any UTF-8 characters.
- description str
- The description of the Endpoint.
- enable_
private_ boolservice_ connect - Deprecated: If true, expose the Endpoint via private service connect. Only one of the fields, network or enable_private_service_connect, can be set.
- encryption_
spec GoogleCloud Aiplatform V1beta1Encryption Spec Args - Customer-managed encryption key spec for an Endpoint. If set, this Endpoint and all sub-resources of this Endpoint will be secured by this key.
- endpoint_
id str - Immutable. The ID to use for endpoint, which will become the final component of the endpoint resource name. If not provided, Vertex AI will generate a value for this ID. If the first character is a letter, this value may be up to 63 characters, and valid characters are
[a-z0-9-]
. The last character must be a letter or number. If the first character is a number, this value may be up to 9 characters, and valid characters are[0-9]
with no leading zeros. When using HTTP/JSON, this field is populated based on a query string argument, such as?endpoint_id=12345
. This is the fallback for fields that are not included in either the URI or the body. - etag str
- Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
- labels Mapping[str, str]
- The labels with user-defined metadata to organize your Endpoints. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels.
- location str
- network str
- Optional. The full name of the Google Compute Engine network to which the Endpoint should be peered. Private services access must already be configured for the network. If left unspecified, the Endpoint is not peered with any network. Only one of the fields, network or enable_private_service_connect, can be set. Format:
projects/{project}/global/networks/{network}
. Where{project}
is a project number, as in12345
, and{network}
is network name. - predict_
request_ Googleresponse_ logging_ config Cloud Aiplatform V1beta1Predict Request Response Logging Config Args - Configures the request-response logging for online prediction.
- project str
- traffic_
split Mapping[str, str] - A map from a DeployedModel's ID to the percentage of this Endpoint's traffic that should be forwarded to that DeployedModel. If a DeployedModel's ID is not listed in this map, then it receives no traffic. The traffic percentage values must add up to 100, or map must be empty if the Endpoint is to not accept any traffic at a moment.
- display
Name String - The display name of the Endpoint. The name can be up to 128 characters long and can consist of any UTF-8 characters.
- description String
- The description of the Endpoint.
- enable
Private BooleanService Connect - Deprecated: If true, expose the Endpoint via private service connect. Only one of the fields, network or enable_private_service_connect, can be set.
- encryption
Spec Property Map - Customer-managed encryption key spec for an Endpoint. If set, this Endpoint and all sub-resources of this Endpoint will be secured by this key.
- endpoint
Id String - Immutable. The ID to use for endpoint, which will become the final component of the endpoint resource name. If not provided, Vertex AI will generate a value for this ID. If the first character is a letter, this value may be up to 63 characters, and valid characters are
[a-z0-9-]
. The last character must be a letter or number. If the first character is a number, this value may be up to 9 characters, and valid characters are[0-9]
with no leading zeros. When using HTTP/JSON, this field is populated based on a query string argument, such as?endpoint_id=12345
. This is the fallback for fields that are not included in either the URI or the body. - etag String
- Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
- labels Map<String>
- The labels with user-defined metadata to organize your Endpoints. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels.
- location String
- network String
- Optional. The full name of the Google Compute Engine network to which the Endpoint should be peered. Private services access must already be configured for the network. If left unspecified, the Endpoint is not peered with any network. Only one of the fields, network or enable_private_service_connect, can be set. Format:
projects/{project}/global/networks/{network}
. Where{project}
is a project number, as in12345
, and{network}
is network name. - predict
Request Property MapResponse Logging Config - Configures the request-response logging for online prediction.
- project String
- traffic
Split Map<String> - A map from a DeployedModel's ID to the percentage of this Endpoint's traffic that should be forwarded to that DeployedModel. If a DeployedModel's ID is not listed in this map, then it receives no traffic. The traffic percentage values must add up to 100, or map must be empty if the Endpoint is to not accept any traffic at a moment.
Outputs
All input properties are implicitly available as output properties. Additionally, the Endpoint resource produces the following output properties:
- Create
Time string - Timestamp when this Endpoint was created.
- Deployed
Models List<Pulumi.Google Native. Aiplatform. V1Beta1. Outputs. Google Cloud Aiplatform V1beta1Deployed Model Response> - The models deployed in this Endpoint. To add or remove DeployedModels use EndpointService.DeployModel and EndpointService.UndeployModel respectively.
- Id string
- The provider-assigned unique ID for this managed resource.
- Model
Deployment stringMonitoring Job - Resource name of the Model Monitoring job associated with this Endpoint if monitoring is enabled by JobService.CreateModelDeploymentMonitoringJob. Format:
projects/{project}/locations/{location}/modelDeploymentMonitoringJobs/{model_deployment_monitoring_job}
- Name string
- The resource name of the Endpoint.
- Update
Time string - Timestamp when this Endpoint was last updated.
- Create
Time string - Timestamp when this Endpoint was created.
- Deployed
Models []GoogleCloud Aiplatform V1beta1Deployed Model Response - The models deployed in this Endpoint. To add or remove DeployedModels use EndpointService.DeployModel and EndpointService.UndeployModel respectively.
- Id string
- The provider-assigned unique ID for this managed resource.
- Model
Deployment stringMonitoring Job - Resource name of the Model Monitoring job associated with this Endpoint if monitoring is enabled by JobService.CreateModelDeploymentMonitoringJob. Format:
projects/{project}/locations/{location}/modelDeploymentMonitoringJobs/{model_deployment_monitoring_job}
- Name string
- The resource name of the Endpoint.
- Update
Time string - Timestamp when this Endpoint was last updated.
- create
Time String - Timestamp when this Endpoint was created.
- deployed
Models List<GoogleCloud Aiplatform V1beta1Deployed Model Response> - The models deployed in this Endpoint. To add or remove DeployedModels use EndpointService.DeployModel and EndpointService.UndeployModel respectively.
- id String
- The provider-assigned unique ID for this managed resource.
- model
Deployment StringMonitoring Job - Resource name of the Model Monitoring job associated with this Endpoint if monitoring is enabled by JobService.CreateModelDeploymentMonitoringJob. Format:
projects/{project}/locations/{location}/modelDeploymentMonitoringJobs/{model_deployment_monitoring_job}
- name String
- The resource name of the Endpoint.
- update
Time String - Timestamp when this Endpoint was last updated.
- create
Time string - Timestamp when this Endpoint was created.
- deployed
Models GoogleCloud Aiplatform V1beta1Deployed Model Response[] - The models deployed in this Endpoint. To add or remove DeployedModels use EndpointService.DeployModel and EndpointService.UndeployModel respectively.
- id string
- The provider-assigned unique ID for this managed resource.
- model
Deployment stringMonitoring Job - Resource name of the Model Monitoring job associated with this Endpoint if monitoring is enabled by JobService.CreateModelDeploymentMonitoringJob. Format:
projects/{project}/locations/{location}/modelDeploymentMonitoringJobs/{model_deployment_monitoring_job}
- name string
- The resource name of the Endpoint.
- update
Time string - Timestamp when this Endpoint was last updated.
- create_
time str - Timestamp when this Endpoint was created.
- deployed_
models Sequence[GoogleCloud Aiplatform V1beta1Deployed Model Response] - The models deployed in this Endpoint. To add or remove DeployedModels use EndpointService.DeployModel and EndpointService.UndeployModel respectively.
- id str
- The provider-assigned unique ID for this managed resource.
- model_
deployment_ strmonitoring_ job - Resource name of the Model Monitoring job associated with this Endpoint if monitoring is enabled by JobService.CreateModelDeploymentMonitoringJob. Format:
projects/{project}/locations/{location}/modelDeploymentMonitoringJobs/{model_deployment_monitoring_job}
- name str
- The resource name of the Endpoint.
- update_
time str - Timestamp when this Endpoint was last updated.
- create
Time String - Timestamp when this Endpoint was created.
- deployed
Models List<Property Map> - The models deployed in this Endpoint. To add or remove DeployedModels use EndpointService.DeployModel and EndpointService.UndeployModel respectively.
- id String
- The provider-assigned unique ID for this managed resource.
- model
Deployment StringMonitoring Job - Resource name of the Model Monitoring job associated with this Endpoint if monitoring is enabled by JobService.CreateModelDeploymentMonitoringJob. Format:
projects/{project}/locations/{location}/modelDeploymentMonitoringJobs/{model_deployment_monitoring_job}
- name String
- The resource name of the Endpoint.
- update
Time String - Timestamp when this Endpoint was last updated.
Supporting Types
GoogleCloudAiplatformV1beta1AutomaticResourcesResponse, GoogleCloudAiplatformV1beta1AutomaticResourcesResponseArgs
- Max
Replica intCount - Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, a no upper bound for scaling under heavy traffic will be assume, though Vertex AI may be unable to scale beyond certain replica number.
- Min
Replica intCount - Immutable. The minimum number of replicas this DeployedModel will be always deployed on. If traffic against it increases, it may dynamically be deployed onto more replicas up to max_replica_count, and as traffic decreases, some of these extra replicas may be freed. If the requested value is too large, the deployment will error.
- Max
Replica intCount - Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, a no upper bound for scaling under heavy traffic will be assume, though Vertex AI may be unable to scale beyond certain replica number.
- Min
Replica intCount - Immutable. The minimum number of replicas this DeployedModel will be always deployed on. If traffic against it increases, it may dynamically be deployed onto more replicas up to max_replica_count, and as traffic decreases, some of these extra replicas may be freed. If the requested value is too large, the deployment will error.
- max
Replica IntegerCount - Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, a no upper bound for scaling under heavy traffic will be assume, though Vertex AI may be unable to scale beyond certain replica number.
- min
Replica IntegerCount - Immutable. The minimum number of replicas this DeployedModel will be always deployed on. If traffic against it increases, it may dynamically be deployed onto more replicas up to max_replica_count, and as traffic decreases, some of these extra replicas may be freed. If the requested value is too large, the deployment will error.
- max
Replica numberCount - Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, a no upper bound for scaling under heavy traffic will be assume, though Vertex AI may be unable to scale beyond certain replica number.
- min
Replica numberCount - Immutable. The minimum number of replicas this DeployedModel will be always deployed on. If traffic against it increases, it may dynamically be deployed onto more replicas up to max_replica_count, and as traffic decreases, some of these extra replicas may be freed. If the requested value is too large, the deployment will error.
- max_
replica_ intcount - Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, a no upper bound for scaling under heavy traffic will be assume, though Vertex AI may be unable to scale beyond certain replica number.
- min_
replica_ intcount - Immutable. The minimum number of replicas this DeployedModel will be always deployed on. If traffic against it increases, it may dynamically be deployed onto more replicas up to max_replica_count, and as traffic decreases, some of these extra replicas may be freed. If the requested value is too large, the deployment will error.
- max
Replica NumberCount - Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, a no upper bound for scaling under heavy traffic will be assume, though Vertex AI may be unable to scale beyond certain replica number.
- min
Replica NumberCount - Immutable. The minimum number of replicas this DeployedModel will be always deployed on. If traffic against it increases, it may dynamically be deployed onto more replicas up to max_replica_count, and as traffic decreases, some of these extra replicas may be freed. If the requested value is too large, the deployment will error.
GoogleCloudAiplatformV1beta1AutoscalingMetricSpecResponse, GoogleCloudAiplatformV1beta1AutoscalingMetricSpecResponseArgs
- Metric
Name string - The resource metric name. Supported metrics: * For Online Prediction: *
aiplatform.googleapis.com/prediction/online/accelerator/duty_cycle
*aiplatform.googleapis.com/prediction/online/cpu/utilization
- Target int
- The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided.
- Metric
Name string - The resource metric name. Supported metrics: * For Online Prediction: *
aiplatform.googleapis.com/prediction/online/accelerator/duty_cycle
*aiplatform.googleapis.com/prediction/online/cpu/utilization
- Target int
- The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided.
- metric
Name String - The resource metric name. Supported metrics: * For Online Prediction: *
aiplatform.googleapis.com/prediction/online/accelerator/duty_cycle
*aiplatform.googleapis.com/prediction/online/cpu/utilization
- target Integer
- The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided.
- metric
Name string - The resource metric name. Supported metrics: * For Online Prediction: *
aiplatform.googleapis.com/prediction/online/accelerator/duty_cycle
*aiplatform.googleapis.com/prediction/online/cpu/utilization
- target number
- The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided.
- metric_
name str - The resource metric name. Supported metrics: * For Online Prediction: *
aiplatform.googleapis.com/prediction/online/accelerator/duty_cycle
*aiplatform.googleapis.com/prediction/online/cpu/utilization
- target int
- The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided.
- metric
Name String - The resource metric name. Supported metrics: * For Online Prediction: *
aiplatform.googleapis.com/prediction/online/accelerator/duty_cycle
*aiplatform.googleapis.com/prediction/online/cpu/utilization
- target Number
- The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided.
GoogleCloudAiplatformV1beta1BigQueryDestination, GoogleCloudAiplatformV1beta1BigQueryDestinationArgs
- Output
Uri string - BigQuery URI to a project or table, up to 2000 characters long. When only the project is specified, the Dataset and Table is created. When the full table reference is specified, the Dataset must exist and table must not exist. Accepted forms: * BigQuery path. For example:
bq://projectId
orbq://projectId.bqDatasetId
orbq://projectId.bqDatasetId.bqTableId
.
- Output
Uri string - BigQuery URI to a project or table, up to 2000 characters long. When only the project is specified, the Dataset and Table is created. When the full table reference is specified, the Dataset must exist and table must not exist. Accepted forms: * BigQuery path. For example:
bq://projectId
orbq://projectId.bqDatasetId
orbq://projectId.bqDatasetId.bqTableId
.
- output
Uri String - BigQuery URI to a project or table, up to 2000 characters long. When only the project is specified, the Dataset and Table is created. When the full table reference is specified, the Dataset must exist and table must not exist. Accepted forms: * BigQuery path. For example:
bq://projectId
orbq://projectId.bqDatasetId
orbq://projectId.bqDatasetId.bqTableId
.
- output
Uri string - BigQuery URI to a project or table, up to 2000 characters long. When only the project is specified, the Dataset and Table is created. When the full table reference is specified, the Dataset must exist and table must not exist. Accepted forms: * BigQuery path. For example:
bq://projectId
orbq://projectId.bqDatasetId
orbq://projectId.bqDatasetId.bqTableId
.
- output_
uri str - BigQuery URI to a project or table, up to 2000 characters long. When only the project is specified, the Dataset and Table is created. When the full table reference is specified, the Dataset must exist and table must not exist. Accepted forms: * BigQuery path. For example:
bq://projectId
orbq://projectId.bqDatasetId
orbq://projectId.bqDatasetId.bqTableId
.
- output
Uri String - BigQuery URI to a project or table, up to 2000 characters long. When only the project is specified, the Dataset and Table is created. When the full table reference is specified, the Dataset must exist and table must not exist. Accepted forms: * BigQuery path. For example:
bq://projectId
orbq://projectId.bqDatasetId
orbq://projectId.bqDatasetId.bqTableId
.
GoogleCloudAiplatformV1beta1BigQueryDestinationResponse, GoogleCloudAiplatformV1beta1BigQueryDestinationResponseArgs
- Output
Uri string - BigQuery URI to a project or table, up to 2000 characters long. When only the project is specified, the Dataset and Table is created. When the full table reference is specified, the Dataset must exist and table must not exist. Accepted forms: * BigQuery path. For example:
bq://projectId
orbq://projectId.bqDatasetId
orbq://projectId.bqDatasetId.bqTableId
.
- Output
Uri string - BigQuery URI to a project or table, up to 2000 characters long. When only the project is specified, the Dataset and Table is created. When the full table reference is specified, the Dataset must exist and table must not exist. Accepted forms: * BigQuery path. For example:
bq://projectId
orbq://projectId.bqDatasetId
orbq://projectId.bqDatasetId.bqTableId
.
- output
Uri String - BigQuery URI to a project or table, up to 2000 characters long. When only the project is specified, the Dataset and Table is created. When the full table reference is specified, the Dataset must exist and table must not exist. Accepted forms: * BigQuery path. For example:
bq://projectId
orbq://projectId.bqDatasetId
orbq://projectId.bqDatasetId.bqTableId
.
- output
Uri string - BigQuery URI to a project or table, up to 2000 characters long. When only the project is specified, the Dataset and Table is created. When the full table reference is specified, the Dataset must exist and table must not exist. Accepted forms: * BigQuery path. For example:
bq://projectId
orbq://projectId.bqDatasetId
orbq://projectId.bqDatasetId.bqTableId
.
- output_
uri str - BigQuery URI to a project or table, up to 2000 characters long. When only the project is specified, the Dataset and Table is created. When the full table reference is specified, the Dataset must exist and table must not exist. Accepted forms: * BigQuery path. For example:
bq://projectId
orbq://projectId.bqDatasetId
orbq://projectId.bqDatasetId.bqTableId
.
- output
Uri String - BigQuery URI to a project or table, up to 2000 characters long. When only the project is specified, the Dataset and Table is created. When the full table reference is specified, the Dataset must exist and table must not exist. Accepted forms: * BigQuery path. For example:
bq://projectId
orbq://projectId.bqDatasetId
orbq://projectId.bqDatasetId.bqTableId
.
GoogleCloudAiplatformV1beta1BlurBaselineConfigResponse, GoogleCloudAiplatformV1beta1BlurBaselineConfigResponseArgs
- Max
Blur doubleSigma - The standard deviation of the blur kernel for the blurred baseline. The same blurring parameter is used for both the height and the width dimension. If not set, the method defaults to the zero (i.e. black for images) baseline.
- Max
Blur float64Sigma - The standard deviation of the blur kernel for the blurred baseline. The same blurring parameter is used for both the height and the width dimension. If not set, the method defaults to the zero (i.e. black for images) baseline.
- max
Blur DoubleSigma - The standard deviation of the blur kernel for the blurred baseline. The same blurring parameter is used for both the height and the width dimension. If not set, the method defaults to the zero (i.e. black for images) baseline.
- max
Blur numberSigma - The standard deviation of the blur kernel for the blurred baseline. The same blurring parameter is used for both the height and the width dimension. If not set, the method defaults to the zero (i.e. black for images) baseline.
- max_
blur_ floatsigma - The standard deviation of the blur kernel for the blurred baseline. The same blurring parameter is used for both the height and the width dimension. If not set, the method defaults to the zero (i.e. black for images) baseline.
- max
Blur NumberSigma - The standard deviation of the blur kernel for the blurred baseline. The same blurring parameter is used for both the height and the width dimension. If not set, the method defaults to the zero (i.e. black for images) baseline.
GoogleCloudAiplatformV1beta1DedicatedResourcesResponse, GoogleCloudAiplatformV1beta1DedicatedResourcesResponseArgs
- Autoscaling
Metric List<Pulumi.Specs Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Autoscaling Metric Spec Response> - Immutable. The metric specifications that overrides a resource utilization metric (CPU utilization, accelerator's duty cycle, and so on) target value (default to 60 if not set). At most one entry is allowed per metric. If machine_spec.accelerator_count is above 0, the autoscaling will be based on both CPU utilization and accelerator's duty cycle metrics and scale up when either metrics exceeds its target value while scale down if both metrics are under their target value. The default target value is 60 for both metrics. If machine_spec.accelerator_count is 0, the autoscaling will be based on CPU utilization metric only with default target value 60 if not explicitly set. For example, in the case of Online Prediction, if you want to override target CPU utilization to 80, you should set autoscaling_metric_specs.metric_name to
aiplatform.googleapis.com/prediction/online/cpu/utilization
and autoscaling_metric_specs.target to80
. - Machine
Spec Pulumi.Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Machine Spec Response - Immutable. The specification of a single machine used by the prediction.
- Max
Replica intCount - Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, will use min_replica_count as the default value. The value of this field impacts the charge against Vertex CPU and GPU quotas. Specifically, you will be charged for (max_replica_count * number of cores in the selected machine type) and (max_replica_count * number of GPUs per replica in the selected machine type).
- Min
Replica intCount - Immutable. The minimum number of machine replicas this DeployedModel will be always deployed on. This value must be greater than or equal to 1. If traffic against the DeployedModel increases, it may dynamically be deployed onto more replicas, and as traffic decreases, some of these extra replicas may be freed.
- Autoscaling
Metric []GoogleSpecs Cloud Aiplatform V1beta1Autoscaling Metric Spec Response - Immutable. The metric specifications that overrides a resource utilization metric (CPU utilization, accelerator's duty cycle, and so on) target value (default to 60 if not set). At most one entry is allowed per metric. If machine_spec.accelerator_count is above 0, the autoscaling will be based on both CPU utilization and accelerator's duty cycle metrics and scale up when either metrics exceeds its target value while scale down if both metrics are under their target value. The default target value is 60 for both metrics. If machine_spec.accelerator_count is 0, the autoscaling will be based on CPU utilization metric only with default target value 60 if not explicitly set. For example, in the case of Online Prediction, if you want to override target CPU utilization to 80, you should set autoscaling_metric_specs.metric_name to
aiplatform.googleapis.com/prediction/online/cpu/utilization
and autoscaling_metric_specs.target to80
. - Machine
Spec GoogleCloud Aiplatform V1beta1Machine Spec Response - Immutable. The specification of a single machine used by the prediction.
- Max
Replica intCount - Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, will use min_replica_count as the default value. The value of this field impacts the charge against Vertex CPU and GPU quotas. Specifically, you will be charged for (max_replica_count * number of cores in the selected machine type) and (max_replica_count * number of GPUs per replica in the selected machine type).
- Min
Replica intCount - Immutable. The minimum number of machine replicas this DeployedModel will be always deployed on. This value must be greater than or equal to 1. If traffic against the DeployedModel increases, it may dynamically be deployed onto more replicas, and as traffic decreases, some of these extra replicas may be freed.
- autoscaling
Metric List<GoogleSpecs Cloud Aiplatform V1beta1Autoscaling Metric Spec Response> - Immutable. The metric specifications that overrides a resource utilization metric (CPU utilization, accelerator's duty cycle, and so on) target value (default to 60 if not set). At most one entry is allowed per metric. If machine_spec.accelerator_count is above 0, the autoscaling will be based on both CPU utilization and accelerator's duty cycle metrics and scale up when either metrics exceeds its target value while scale down if both metrics are under their target value. The default target value is 60 for both metrics. If machine_spec.accelerator_count is 0, the autoscaling will be based on CPU utilization metric only with default target value 60 if not explicitly set. For example, in the case of Online Prediction, if you want to override target CPU utilization to 80, you should set autoscaling_metric_specs.metric_name to
aiplatform.googleapis.com/prediction/online/cpu/utilization
and autoscaling_metric_specs.target to80
. - machine
Spec GoogleCloud Aiplatform V1beta1Machine Spec Response - Immutable. The specification of a single machine used by the prediction.
- max
Replica IntegerCount - Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, will use min_replica_count as the default value. The value of this field impacts the charge against Vertex CPU and GPU quotas. Specifically, you will be charged for (max_replica_count * number of cores in the selected machine type) and (max_replica_count * number of GPUs per replica in the selected machine type).
- min
Replica IntegerCount - Immutable. The minimum number of machine replicas this DeployedModel will be always deployed on. This value must be greater than or equal to 1. If traffic against the DeployedModel increases, it may dynamically be deployed onto more replicas, and as traffic decreases, some of these extra replicas may be freed.
- autoscaling
Metric GoogleSpecs Cloud Aiplatform V1beta1Autoscaling Metric Spec Response[] - Immutable. The metric specifications that overrides a resource utilization metric (CPU utilization, accelerator's duty cycle, and so on) target value (default to 60 if not set). At most one entry is allowed per metric. If machine_spec.accelerator_count is above 0, the autoscaling will be based on both CPU utilization and accelerator's duty cycle metrics and scale up when either metrics exceeds its target value while scale down if both metrics are under their target value. The default target value is 60 for both metrics. If machine_spec.accelerator_count is 0, the autoscaling will be based on CPU utilization metric only with default target value 60 if not explicitly set. For example, in the case of Online Prediction, if you want to override target CPU utilization to 80, you should set autoscaling_metric_specs.metric_name to
aiplatform.googleapis.com/prediction/online/cpu/utilization
and autoscaling_metric_specs.target to80
. - machine
Spec GoogleCloud Aiplatform V1beta1Machine Spec Response - Immutable. The specification of a single machine used by the prediction.
- max
Replica numberCount - Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, will use min_replica_count as the default value. The value of this field impacts the charge against Vertex CPU and GPU quotas. Specifically, you will be charged for (max_replica_count * number of cores in the selected machine type) and (max_replica_count * number of GPUs per replica in the selected machine type).
- min
Replica numberCount - Immutable. The minimum number of machine replicas this DeployedModel will be always deployed on. This value must be greater than or equal to 1. If traffic against the DeployedModel increases, it may dynamically be deployed onto more replicas, and as traffic decreases, some of these extra replicas may be freed.
- autoscaling_
metric_ Sequence[Googlespecs Cloud Aiplatform V1beta1Autoscaling Metric Spec Response] - Immutable. The metric specifications that overrides a resource utilization metric (CPU utilization, accelerator's duty cycle, and so on) target value (default to 60 if not set). At most one entry is allowed per metric. If machine_spec.accelerator_count is above 0, the autoscaling will be based on both CPU utilization and accelerator's duty cycle metrics and scale up when either metrics exceeds its target value while scale down if both metrics are under their target value. The default target value is 60 for both metrics. If machine_spec.accelerator_count is 0, the autoscaling will be based on CPU utilization metric only with default target value 60 if not explicitly set. For example, in the case of Online Prediction, if you want to override target CPU utilization to 80, you should set autoscaling_metric_specs.metric_name to
aiplatform.googleapis.com/prediction/online/cpu/utilization
and autoscaling_metric_specs.target to80
. - machine_
spec GoogleCloud Aiplatform V1beta1Machine Spec Response - Immutable. The specification of a single machine used by the prediction.
- max_
replica_ intcount - Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, will use min_replica_count as the default value. The value of this field impacts the charge against Vertex CPU and GPU quotas. Specifically, you will be charged for (max_replica_count * number of cores in the selected machine type) and (max_replica_count * number of GPUs per replica in the selected machine type).
- min_
replica_ intcount - Immutable. The minimum number of machine replicas this DeployedModel will be always deployed on. This value must be greater than or equal to 1. If traffic against the DeployedModel increases, it may dynamically be deployed onto more replicas, and as traffic decreases, some of these extra replicas may be freed.
- autoscaling
Metric List<Property Map>Specs - Immutable. The metric specifications that overrides a resource utilization metric (CPU utilization, accelerator's duty cycle, and so on) target value (default to 60 if not set). At most one entry is allowed per metric. If machine_spec.accelerator_count is above 0, the autoscaling will be based on both CPU utilization and accelerator's duty cycle metrics and scale up when either metrics exceeds its target value while scale down if both metrics are under their target value. The default target value is 60 for both metrics. If machine_spec.accelerator_count is 0, the autoscaling will be based on CPU utilization metric only with default target value 60 if not explicitly set. For example, in the case of Online Prediction, if you want to override target CPU utilization to 80, you should set autoscaling_metric_specs.metric_name to
aiplatform.googleapis.com/prediction/online/cpu/utilization
and autoscaling_metric_specs.target to80
. - machine
Spec Property Map - Immutable. The specification of a single machine used by the prediction.
- max
Replica NumberCount - Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, will use min_replica_count as the default value. The value of this field impacts the charge against Vertex CPU and GPU quotas. Specifically, you will be charged for (max_replica_count * number of cores in the selected machine type) and (max_replica_count * number of GPUs per replica in the selected machine type).
- min
Replica NumberCount - Immutable. The minimum number of machine replicas this DeployedModel will be always deployed on. This value must be greater than or equal to 1. If traffic against the DeployedModel increases, it may dynamically be deployed onto more replicas, and as traffic decreases, some of these extra replicas may be freed.
GoogleCloudAiplatformV1beta1DeployedModelResponse, GoogleCloudAiplatformV1beta1DeployedModelResponseArgs
- Automatic
Resources Pulumi.Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Automatic Resources Response - A description of resources that to large degree are decided by Vertex AI, and require only a modest additional configuration.
- Create
Time string - Timestamp when the DeployedModel was created.
- Dedicated
Resources Pulumi.Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Dedicated Resources Response - A description of resources that are dedicated to the DeployedModel, and that need a higher degree of manual configuration.
- Disable
Explanations bool - If true, deploy the model without explainable feature, regardless the existence of Model.explanation_spec or explanation_spec.
- Display
Name string - The display name of the DeployedModel. If not provided upon creation, the Model's display_name is used.
- Enable
Access boolLogging - If true, online prediction access logs are sent to Cloud Logging. These logs are like standard server access logs, containing information like timestamp and latency for each prediction request. Note that logs may incur a cost, especially if your project receives prediction requests at a high queries per second rate (QPS). Estimate your costs before enabling this option.
- Enable
Container boolLogging - If true, the container of the DeployedModel instances will send
stderr
andstdout
streams to Cloud Logging. Only supported for custom-trained Models and AutoML Tabular Models. - Explanation
Spec Pulumi.Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Explanation Spec Response - Explanation configuration for this DeployedModel. When deploying a Model using EndpointService.DeployModel, this value overrides the value of Model.explanation_spec. All fields of explanation_spec are optional in the request. If a field of explanation_spec is not populated, the value of the same field of Model.explanation_spec is inherited. If the corresponding Model.explanation_spec is not populated, all fields of the explanation_spec will be used for the explanation configuration.
- Model string
- The resource name of the Model that this is the deployment of. Note that the Model may be in a different location than the DeployedModel's Endpoint. The resource name may contain version id or version alias to specify the version. Example:
projects/{project}/locations/{location}/models/{model}@2
orprojects/{project}/locations/{location}/models/{model}@golden
if no version is specified, the default version will be deployed. - Model
Version stringId - The version ID of the model that is deployed.
- Private
Endpoints Pulumi.Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Private Endpoints Response - Provide paths for users to send predict/explain/health requests directly to the deployed model services running on Cloud via private services access. This field is populated if network is configured.
- Service
Account string - The service account that the DeployedModel's container runs as. Specify the email address of the service account. If this service account is not specified, the container runs as a service account that doesn't have access to the resource project. Users deploying the Model must have the
iam.serviceAccounts.actAs
permission on this service account. - string
- The resource name of the shared DeploymentResourcePool to deploy on. Format:
projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}
- Automatic
Resources GoogleCloud Aiplatform V1beta1Automatic Resources Response - A description of resources that to large degree are decided by Vertex AI, and require only a modest additional configuration.
- Create
Time string - Timestamp when the DeployedModel was created.
- Dedicated
Resources GoogleCloud Aiplatform V1beta1Dedicated Resources Response - A description of resources that are dedicated to the DeployedModel, and that need a higher degree of manual configuration.
- Disable
Explanations bool - If true, deploy the model without explainable feature, regardless the existence of Model.explanation_spec or explanation_spec.
- Display
Name string - The display name of the DeployedModel. If not provided upon creation, the Model's display_name is used.
- Enable
Access boolLogging - If true, online prediction access logs are sent to Cloud Logging. These logs are like standard server access logs, containing information like timestamp and latency for each prediction request. Note that logs may incur a cost, especially if your project receives prediction requests at a high queries per second rate (QPS). Estimate your costs before enabling this option.
- Enable
Container boolLogging - If true, the container of the DeployedModel instances will send
stderr
andstdout
streams to Cloud Logging. Only supported for custom-trained Models and AutoML Tabular Models. - Explanation
Spec GoogleCloud Aiplatform V1beta1Explanation Spec Response - Explanation configuration for this DeployedModel. When deploying a Model using EndpointService.DeployModel, this value overrides the value of Model.explanation_spec. All fields of explanation_spec are optional in the request. If a field of explanation_spec is not populated, the value of the same field of Model.explanation_spec is inherited. If the corresponding Model.explanation_spec is not populated, all fields of the explanation_spec will be used for the explanation configuration.
- Model string
- The resource name of the Model that this is the deployment of. Note that the Model may be in a different location than the DeployedModel's Endpoint. The resource name may contain version id or version alias to specify the version. Example:
projects/{project}/locations/{location}/models/{model}@2
orprojects/{project}/locations/{location}/models/{model}@golden
if no version is specified, the default version will be deployed. - Model
Version stringId - The version ID of the model that is deployed.
- Private
Endpoints GoogleCloud Aiplatform V1beta1Private Endpoints Response - Provide paths for users to send predict/explain/health requests directly to the deployed model services running on Cloud via private services access. This field is populated if network is configured.
- Service
Account string - The service account that the DeployedModel's container runs as. Specify the email address of the service account. If this service account is not specified, the container runs as a service account that doesn't have access to the resource project. Users deploying the Model must have the
iam.serviceAccounts.actAs
permission on this service account. - string
- The resource name of the shared DeploymentResourcePool to deploy on. Format:
projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}
- automatic
Resources GoogleCloud Aiplatform V1beta1Automatic Resources Response - A description of resources that to large degree are decided by Vertex AI, and require only a modest additional configuration.
- create
Time String - Timestamp when the DeployedModel was created.
- dedicated
Resources GoogleCloud Aiplatform V1beta1Dedicated Resources Response - A description of resources that are dedicated to the DeployedModel, and that need a higher degree of manual configuration.
- disable
Explanations Boolean - If true, deploy the model without explainable feature, regardless the existence of Model.explanation_spec or explanation_spec.
- display
Name String - The display name of the DeployedModel. If not provided upon creation, the Model's display_name is used.
- enable
Access BooleanLogging - If true, online prediction access logs are sent to Cloud Logging. These logs are like standard server access logs, containing information like timestamp and latency for each prediction request. Note that logs may incur a cost, especially if your project receives prediction requests at a high queries per second rate (QPS). Estimate your costs before enabling this option.
- enable
Container BooleanLogging - If true, the container of the DeployedModel instances will send
stderr
andstdout
streams to Cloud Logging. Only supported for custom-trained Models and AutoML Tabular Models. - explanation
Spec GoogleCloud Aiplatform V1beta1Explanation Spec Response - Explanation configuration for this DeployedModel. When deploying a Model using EndpointService.DeployModel, this value overrides the value of Model.explanation_spec. All fields of explanation_spec are optional in the request. If a field of explanation_spec is not populated, the value of the same field of Model.explanation_spec is inherited. If the corresponding Model.explanation_spec is not populated, all fields of the explanation_spec will be used for the explanation configuration.
- model String
- The resource name of the Model that this is the deployment of. Note that the Model may be in a different location than the DeployedModel's Endpoint. The resource name may contain version id or version alias to specify the version. Example:
projects/{project}/locations/{location}/models/{model}@2
orprojects/{project}/locations/{location}/models/{model}@golden
if no version is specified, the default version will be deployed. - model
Version StringId - The version ID of the model that is deployed.
- private
Endpoints GoogleCloud Aiplatform V1beta1Private Endpoints Response - Provide paths for users to send predict/explain/health requests directly to the deployed model services running on Cloud via private services access. This field is populated if network is configured.
- service
Account String - The service account that the DeployedModel's container runs as. Specify the email address of the service account. If this service account is not specified, the container runs as a service account that doesn't have access to the resource project. Users deploying the Model must have the
iam.serviceAccounts.actAs
permission on this service account. - String
- The resource name of the shared DeploymentResourcePool to deploy on. Format:
projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}
- automatic
Resources GoogleCloud Aiplatform V1beta1Automatic Resources Response - A description of resources that to large degree are decided by Vertex AI, and require only a modest additional configuration.
- create
Time string - Timestamp when the DeployedModel was created.
- dedicated
Resources GoogleCloud Aiplatform V1beta1Dedicated Resources Response - A description of resources that are dedicated to the DeployedModel, and that need a higher degree of manual configuration.
- disable
Explanations boolean - If true, deploy the model without explainable feature, regardless the existence of Model.explanation_spec or explanation_spec.
- display
Name string - The display name of the DeployedModel. If not provided upon creation, the Model's display_name is used.
- enable
Access booleanLogging - If true, online prediction access logs are sent to Cloud Logging. These logs are like standard server access logs, containing information like timestamp and latency for each prediction request. Note that logs may incur a cost, especially if your project receives prediction requests at a high queries per second rate (QPS). Estimate your costs before enabling this option.
- enable
Container booleanLogging - If true, the container of the DeployedModel instances will send
stderr
andstdout
streams to Cloud Logging. Only supported for custom-trained Models and AutoML Tabular Models. - explanation
Spec GoogleCloud Aiplatform V1beta1Explanation Spec Response - Explanation configuration for this DeployedModel. When deploying a Model using EndpointService.DeployModel, this value overrides the value of Model.explanation_spec. All fields of explanation_spec are optional in the request. If a field of explanation_spec is not populated, the value of the same field of Model.explanation_spec is inherited. If the corresponding Model.explanation_spec is not populated, all fields of the explanation_spec will be used for the explanation configuration.
- model string
- The resource name of the Model that this is the deployment of. Note that the Model may be in a different location than the DeployedModel's Endpoint. The resource name may contain version id or version alias to specify the version. Example:
projects/{project}/locations/{location}/models/{model}@2
orprojects/{project}/locations/{location}/models/{model}@golden
if no version is specified, the default version will be deployed. - model
Version stringId - The version ID of the model that is deployed.
- private
Endpoints GoogleCloud Aiplatform V1beta1Private Endpoints Response - Provide paths for users to send predict/explain/health requests directly to the deployed model services running on Cloud via private services access. This field is populated if network is configured.
- service
Account string - The service account that the DeployedModel's container runs as. Specify the email address of the service account. If this service account is not specified, the container runs as a service account that doesn't have access to the resource project. Users deploying the Model must have the
iam.serviceAccounts.actAs
permission on this service account. - string
- The resource name of the shared DeploymentResourcePool to deploy on. Format:
projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}
- automatic_
resources GoogleCloud Aiplatform V1beta1Automatic Resources Response - A description of resources that to large degree are decided by Vertex AI, and require only a modest additional configuration.
- create_
time str - Timestamp when the DeployedModel was created.
- dedicated_
resources GoogleCloud Aiplatform V1beta1Dedicated Resources Response - A description of resources that are dedicated to the DeployedModel, and that need a higher degree of manual configuration.
- disable_
explanations bool - If true, deploy the model without explainable feature, regardless the existence of Model.explanation_spec or explanation_spec.
- display_
name str - The display name of the DeployedModel. If not provided upon creation, the Model's display_name is used.
- enable_
access_ boollogging - If true, online prediction access logs are sent to Cloud Logging. These logs are like standard server access logs, containing information like timestamp and latency for each prediction request. Note that logs may incur a cost, especially if your project receives prediction requests at a high queries per second rate (QPS). Estimate your costs before enabling this option.
- enable_
container_ boollogging - If true, the container of the DeployedModel instances will send
stderr
andstdout
streams to Cloud Logging. Only supported for custom-trained Models and AutoML Tabular Models. - explanation_
spec GoogleCloud Aiplatform V1beta1Explanation Spec Response - Explanation configuration for this DeployedModel. When deploying a Model using EndpointService.DeployModel, this value overrides the value of Model.explanation_spec. All fields of explanation_spec are optional in the request. If a field of explanation_spec is not populated, the value of the same field of Model.explanation_spec is inherited. If the corresponding Model.explanation_spec is not populated, all fields of the explanation_spec will be used for the explanation configuration.
- model str
- The resource name of the Model that this is the deployment of. Note that the Model may be in a different location than the DeployedModel's Endpoint. The resource name may contain version id or version alias to specify the version. Example:
projects/{project}/locations/{location}/models/{model}@2
orprojects/{project}/locations/{location}/models/{model}@golden
if no version is specified, the default version will be deployed. - model_
version_ strid - The version ID of the model that is deployed.
- private_
endpoints GoogleCloud Aiplatform V1beta1Private Endpoints Response - Provide paths for users to send predict/explain/health requests directly to the deployed model services running on Cloud via private services access. This field is populated if network is configured.
- service_
account str - The service account that the DeployedModel's container runs as. Specify the email address of the service account. If this service account is not specified, the container runs as a service account that doesn't have access to the resource project. Users deploying the Model must have the
iam.serviceAccounts.actAs
permission on this service account. - str
- The resource name of the shared DeploymentResourcePool to deploy on. Format:
projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}
- automatic
Resources Property Map - A description of resources that to large degree are decided by Vertex AI, and require only a modest additional configuration.
- create
Time String - Timestamp when the DeployedModel was created.
- dedicated
Resources Property Map - A description of resources that are dedicated to the DeployedModel, and that need a higher degree of manual configuration.
- disable
Explanations Boolean - If true, deploy the model without explainable feature, regardless the existence of Model.explanation_spec or explanation_spec.
- display
Name String - The display name of the DeployedModel. If not provided upon creation, the Model's display_name is used.
- enable
Access BooleanLogging - If true, online prediction access logs are sent to Cloud Logging. These logs are like standard server access logs, containing information like timestamp and latency for each prediction request. Note that logs may incur a cost, especially if your project receives prediction requests at a high queries per second rate (QPS). Estimate your costs before enabling this option.
- enable
Container BooleanLogging - If true, the container of the DeployedModel instances will send
stderr
andstdout
streams to Cloud Logging. Only supported for custom-trained Models and AutoML Tabular Models. - explanation
Spec Property Map - Explanation configuration for this DeployedModel. When deploying a Model using EndpointService.DeployModel, this value overrides the value of Model.explanation_spec. All fields of explanation_spec are optional in the request. If a field of explanation_spec is not populated, the value of the same field of Model.explanation_spec is inherited. If the corresponding Model.explanation_spec is not populated, all fields of the explanation_spec will be used for the explanation configuration.
- model String
- The resource name of the Model that this is the deployment of. Note that the Model may be in a different location than the DeployedModel's Endpoint. The resource name may contain version id or version alias to specify the version. Example:
projects/{project}/locations/{location}/models/{model}@2
orprojects/{project}/locations/{location}/models/{model}@golden
if no version is specified, the default version will be deployed. - model
Version StringId - The version ID of the model that is deployed.
- private
Endpoints Property Map - Provide paths for users to send predict/explain/health requests directly to the deployed model services running on Cloud via private services access. This field is populated if network is configured.
- service
Account String - The service account that the DeployedModel's container runs as. Specify the email address of the service account. If this service account is not specified, the container runs as a service account that doesn't have access to the resource project. Users deploying the Model must have the
iam.serviceAccounts.actAs
permission on this service account. - String
- The resource name of the shared DeploymentResourcePool to deploy on. Format:
projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}
GoogleCloudAiplatformV1beta1EncryptionSpec, GoogleCloudAiplatformV1beta1EncryptionSpecArgs
- Kms
Key stringName - The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form:
projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key
. The key needs to be in the same region as where the compute resource is created.
- Kms
Key stringName - The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form:
projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key
. The key needs to be in the same region as where the compute resource is created.
- kms
Key StringName - The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form:
projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key
. The key needs to be in the same region as where the compute resource is created.
- kms
Key stringName - The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form:
projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key
. The key needs to be in the same region as where the compute resource is created.
- kms_
key_ strname - The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form:
projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key
. The key needs to be in the same region as where the compute resource is created.
- kms
Key StringName - The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form:
projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key
. The key needs to be in the same region as where the compute resource is created.
GoogleCloudAiplatformV1beta1EncryptionSpecResponse, GoogleCloudAiplatformV1beta1EncryptionSpecResponseArgs
- Kms
Key stringName - The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form:
projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key
. The key needs to be in the same region as where the compute resource is created.
- Kms
Key stringName - The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form:
projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key
. The key needs to be in the same region as where the compute resource is created.
- kms
Key StringName - The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form:
projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key
. The key needs to be in the same region as where the compute resource is created.
- kms
Key stringName - The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form:
projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key
. The key needs to be in the same region as where the compute resource is created.
- kms_
key_ strname - The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form:
projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key
. The key needs to be in the same region as where the compute resource is created.
- kms
Key StringName - The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form:
projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key
. The key needs to be in the same region as where the compute resource is created.
GoogleCloudAiplatformV1beta1ExamplesExampleGcsSourceResponse, GoogleCloudAiplatformV1beta1ExamplesExampleGcsSourceResponseArgs
- Data
Format string - The format in which instances are given, if not specified, assume it's JSONL format. Currently only JSONL format is supported.
- Gcs
Source Pulumi.Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Gcs Source Response - The Cloud Storage location for the input instances.
- Data
Format string - The format in which instances are given, if not specified, assume it's JSONL format. Currently only JSONL format is supported.
- Gcs
Source GoogleCloud Aiplatform V1beta1Gcs Source Response - The Cloud Storage location for the input instances.
- data
Format String - The format in which instances are given, if not specified, assume it's JSONL format. Currently only JSONL format is supported.
- gcs
Source GoogleCloud Aiplatform V1beta1Gcs Source Response - The Cloud Storage location for the input instances.
- data
Format string - The format in which instances are given, if not specified, assume it's JSONL format. Currently only JSONL format is supported.
- gcs
Source GoogleCloud Aiplatform V1beta1Gcs Source Response - The Cloud Storage location for the input instances.
- data_
format str - The format in which instances are given, if not specified, assume it's JSONL format. Currently only JSONL format is supported.
- gcs_
source GoogleCloud Aiplatform V1beta1Gcs Source Response - The Cloud Storage location for the input instances.
- data
Format String - The format in which instances are given, if not specified, assume it's JSONL format. Currently only JSONL format is supported.
- gcs
Source Property Map - The Cloud Storage location for the input instances.
GoogleCloudAiplatformV1beta1ExamplesResponse, GoogleCloudAiplatformV1beta1ExamplesResponseArgs
- Example
Gcs Pulumi.Source Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Examples Example Gcs Source Response - The Cloud Storage input instances.
- Gcs
Source Pulumi.Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Gcs Source Response - The Cloud Storage locations that contain the instances to be indexed for approximate nearest neighbor search.
- Nearest
Neighbor objectSearch Config - The full configuration for the generated index, the semantics are the same as metadata and should match NearestNeighborSearchConfig.
- Neighbor
Count int - The number of neighbors to return when querying for examples.
- Presets
Pulumi.
Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Presets Response - Simplified preset configuration, which automatically sets configuration values based on the desired query speed-precision trade-off and modality.
- Example
Gcs GoogleSource Cloud Aiplatform V1beta1Examples Example Gcs Source Response - The Cloud Storage input instances.
- Gcs
Source GoogleCloud Aiplatform V1beta1Gcs Source Response - The Cloud Storage locations that contain the instances to be indexed for approximate nearest neighbor search.
- Nearest
Neighbor interface{}Search Config - The full configuration for the generated index, the semantics are the same as metadata and should match NearestNeighborSearchConfig.
- Neighbor
Count int - The number of neighbors to return when querying for examples.
- Presets
Google
Cloud Aiplatform V1beta1Presets Response - Simplified preset configuration, which automatically sets configuration values based on the desired query speed-precision trade-off and modality.
- example
Gcs GoogleSource Cloud Aiplatform V1beta1Examples Example Gcs Source Response - The Cloud Storage input instances.
- gcs
Source GoogleCloud Aiplatform V1beta1Gcs Source Response - The Cloud Storage locations that contain the instances to be indexed for approximate nearest neighbor search.
- nearest
Neighbor ObjectSearch Config - The full configuration for the generated index, the semantics are the same as metadata and should match NearestNeighborSearchConfig.
- neighbor
Count Integer - The number of neighbors to return when querying for examples.
- presets
Google
Cloud Aiplatform V1beta1Presets Response - Simplified preset configuration, which automatically sets configuration values based on the desired query speed-precision trade-off and modality.
- example
Gcs GoogleSource Cloud Aiplatform V1beta1Examples Example Gcs Source Response - The Cloud Storage input instances.
- gcs
Source GoogleCloud Aiplatform V1beta1Gcs Source Response - The Cloud Storage locations that contain the instances to be indexed for approximate nearest neighbor search.
- nearest
Neighbor anySearch Config - The full configuration for the generated index, the semantics are the same as metadata and should match NearestNeighborSearchConfig.
- neighbor
Count number - The number of neighbors to return when querying for examples.
- presets
Google
Cloud Aiplatform V1beta1Presets Response - Simplified preset configuration, which automatically sets configuration values based on the desired query speed-precision trade-off and modality.
- example_
gcs_ Googlesource Cloud Aiplatform V1beta1Examples Example Gcs Source Response - The Cloud Storage input instances.
- gcs_
source GoogleCloud Aiplatform V1beta1Gcs Source Response - The Cloud Storage locations that contain the instances to be indexed for approximate nearest neighbor search.
- nearest_
neighbor_ Anysearch_ config - The full configuration for the generated index, the semantics are the same as metadata and should match NearestNeighborSearchConfig.
- neighbor_
count int - The number of neighbors to return when querying for examples.
- presets
Google
Cloud Aiplatform V1beta1Presets Response - Simplified preset configuration, which automatically sets configuration values based on the desired query speed-precision trade-off and modality.
- example
Gcs Property MapSource - The Cloud Storage input instances.
- gcs
Source Property Map - The Cloud Storage locations that contain the instances to be indexed for approximate nearest neighbor search.
- nearest
Neighbor AnySearch Config - The full configuration for the generated index, the semantics are the same as metadata and should match NearestNeighborSearchConfig.
- neighbor
Count Number - The number of neighbors to return when querying for examples.
- presets Property Map
- Simplified preset configuration, which automatically sets configuration values based on the desired query speed-precision trade-off and modality.
GoogleCloudAiplatformV1beta1ExplanationMetadataResponse, GoogleCloudAiplatformV1beta1ExplanationMetadataResponseArgs
- Feature
Attributions stringSchema Uri - Points to a YAML file stored on Google Cloud Storage describing the format of the feature attributions. The schema is defined as an OpenAPI 3.0.2 Schema Object. AutoML tabular Models always have this field populated by Vertex AI. Note: The URI given on output may be different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.
- Inputs Dictionary<string, string>
- Map from feature names to feature input metadata. Keys are the name of the features. Values are the specification of the feature. An empty InputMetadata is valid. It describes a text feature which has the name specified as the key in ExplanationMetadata.inputs. The baseline of the empty feature is chosen by Vertex AI. For Vertex AI-provided Tensorflow images, the key can be any friendly name of the feature. Once specified, featureAttributions are keyed by this key (if not grouped with another feature). For custom images, the key must match with the key in instance.
- Latent
Space stringSource - Name of the source to generate embeddings for example based explanations.
- Outputs Dictionary<string, string>
- Map from output names to output metadata. For Vertex AI-provided Tensorflow images, keys can be any user defined string that consists of any UTF-8 characters. For custom images, keys are the name of the output field in the prediction to be explained. Currently only one key is allowed.
- Feature
Attributions stringSchema Uri - Points to a YAML file stored on Google Cloud Storage describing the format of the feature attributions. The schema is defined as an OpenAPI 3.0.2 Schema Object. AutoML tabular Models always have this field populated by Vertex AI. Note: The URI given on output may be different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.
- Inputs map[string]string
- Map from feature names to feature input metadata. Keys are the name of the features. Values are the specification of the feature. An empty InputMetadata is valid. It describes a text feature which has the name specified as the key in ExplanationMetadata.inputs. The baseline of the empty feature is chosen by Vertex AI. For Vertex AI-provided Tensorflow images, the key can be any friendly name of the feature. Once specified, featureAttributions are keyed by this key (if not grouped with another feature). For custom images, the key must match with the key in instance.
- Latent
Space stringSource - Name of the source to generate embeddings for example based explanations.
- Outputs map[string]string
- Map from output names to output metadata. For Vertex AI-provided Tensorflow images, keys can be any user defined string that consists of any UTF-8 characters. For custom images, keys are the name of the output field in the prediction to be explained. Currently only one key is allowed.
- feature
Attributions StringSchema Uri - Points to a YAML file stored on Google Cloud Storage describing the format of the feature attributions. The schema is defined as an OpenAPI 3.0.2 Schema Object. AutoML tabular Models always have this field populated by Vertex AI. Note: The URI given on output may be different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.
- inputs Map<String,String>
- Map from feature names to feature input metadata. Keys are the name of the features. Values are the specification of the feature. An empty InputMetadata is valid. It describes a text feature which has the name specified as the key in ExplanationMetadata.inputs. The baseline of the empty feature is chosen by Vertex AI. For Vertex AI-provided Tensorflow images, the key can be any friendly name of the feature. Once specified, featureAttributions are keyed by this key (if not grouped with another feature). For custom images, the key must match with the key in instance.
- latent
Space StringSource - Name of the source to generate embeddings for example based explanations.
- outputs Map<String,String>
- Map from output names to output metadata. For Vertex AI-provided Tensorflow images, keys can be any user defined string that consists of any UTF-8 characters. For custom images, keys are the name of the output field in the prediction to be explained. Currently only one key is allowed.
- feature
Attributions stringSchema Uri - Points to a YAML file stored on Google Cloud Storage describing the format of the feature attributions. The schema is defined as an OpenAPI 3.0.2 Schema Object. AutoML tabular Models always have this field populated by Vertex AI. Note: The URI given on output may be different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.
- inputs {[key: string]: string}
- Map from feature names to feature input metadata. Keys are the name of the features. Values are the specification of the feature. An empty InputMetadata is valid. It describes a text feature which has the name specified as the key in ExplanationMetadata.inputs. The baseline of the empty feature is chosen by Vertex AI. For Vertex AI-provided Tensorflow images, the key can be any friendly name of the feature. Once specified, featureAttributions are keyed by this key (if not grouped with another feature). For custom images, the key must match with the key in instance.
- latent
Space stringSource - Name of the source to generate embeddings for example based explanations.
- outputs {[key: string]: string}
- Map from output names to output metadata. For Vertex AI-provided Tensorflow images, keys can be any user defined string that consists of any UTF-8 characters. For custom images, keys are the name of the output field in the prediction to be explained. Currently only one key is allowed.
- feature_
attributions_ strschema_ uri - Points to a YAML file stored on Google Cloud Storage describing the format of the feature attributions. The schema is defined as an OpenAPI 3.0.2 Schema Object. AutoML tabular Models always have this field populated by Vertex AI. Note: The URI given on output may be different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.
- inputs Mapping[str, str]
- Map from feature names to feature input metadata. Keys are the name of the features. Values are the specification of the feature. An empty InputMetadata is valid. It describes a text feature which has the name specified as the key in ExplanationMetadata.inputs. The baseline of the empty feature is chosen by Vertex AI. For Vertex AI-provided Tensorflow images, the key can be any friendly name of the feature. Once specified, featureAttributions are keyed by this key (if not grouped with another feature). For custom images, the key must match with the key in instance.
- latent_
space_ strsource - Name of the source to generate embeddings for example based explanations.
- outputs Mapping[str, str]
- Map from output names to output metadata. For Vertex AI-provided Tensorflow images, keys can be any user defined string that consists of any UTF-8 characters. For custom images, keys are the name of the output field in the prediction to be explained. Currently only one key is allowed.
- feature
Attributions StringSchema Uri - Points to a YAML file stored on Google Cloud Storage describing the format of the feature attributions. The schema is defined as an OpenAPI 3.0.2 Schema Object. AutoML tabular Models always have this field populated by Vertex AI. Note: The URI given on output may be different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.
- inputs Map<String>
- Map from feature names to feature input metadata. Keys are the name of the features. Values are the specification of the feature. An empty InputMetadata is valid. It describes a text feature which has the name specified as the key in ExplanationMetadata.inputs. The baseline of the empty feature is chosen by Vertex AI. For Vertex AI-provided Tensorflow images, the key can be any friendly name of the feature. Once specified, featureAttributions are keyed by this key (if not grouped with another feature). For custom images, the key must match with the key in instance.
- latent
Space StringSource - Name of the source to generate embeddings for example based explanations.
- outputs Map<String>
- Map from output names to output metadata. For Vertex AI-provided Tensorflow images, keys can be any user defined string that consists of any UTF-8 characters. For custom images, keys are the name of the output field in the prediction to be explained. Currently only one key is allowed.
GoogleCloudAiplatformV1beta1ExplanationParametersResponse, GoogleCloudAiplatformV1beta1ExplanationParametersResponseArgs
- Examples
Pulumi.
Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Examples Response - Example-based explanations that returns the nearest neighbors from the provided dataset.
- Integrated
Gradients Pulumi.Attribution Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Integrated Gradients Attribution Response - An attribution method that computes Aumann-Shapley values taking advantage of the model's fully differentiable structure. Refer to this paper for more details: https://arxiv.org/abs/1703.01365
- Output
Indices List<object> - If populated, only returns attributions that have output_index contained in output_indices. It must be an ndarray of integers, with the same shape of the output it's explaining. If not populated, returns attributions for top_k indices of outputs. If neither top_k nor output_indices is populated, returns the argmax index of the outputs. Only applicable to Models that predict multiple outputs (e,g, multi-class Models that predict multiple classes).
- Sampled
Shapley Pulumi.Attribution Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Sampled Shapley Attribution Response - An attribution method that approximates Shapley values for features that contribute to the label being predicted. A sampling strategy is used to approximate the value rather than considering all subsets of features. Refer to this paper for model details: https://arxiv.org/abs/1306.4265.
- Top
K int - If populated, returns attributions for top K indices of outputs (defaults to 1). Only applies to Models that predicts more than one outputs (e,g, multi-class Models). When set to -1, returns explanations for all outputs.
- Xrai
Attribution Pulumi.Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Xrai Attribution Response - An attribution method that redistributes Integrated Gradients attribution to segmented regions, taking advantage of the model's fully differentiable structure. Refer to this paper for more details: https://arxiv.org/abs/1906.02825 XRAI currently performs better on natural images, like a picture of a house or an animal. If the images are taken in artificial environments, like a lab or manufacturing line, or from diagnostic equipment, like x-rays or quality-control cameras, use Integrated Gradients instead.
- Examples
Google
Cloud Aiplatform V1beta1Examples Response - Example-based explanations that returns the nearest neighbors from the provided dataset.
- Integrated
Gradients GoogleAttribution Cloud Aiplatform V1beta1Integrated Gradients Attribution Response - An attribution method that computes Aumann-Shapley values taking advantage of the model's fully differentiable structure. Refer to this paper for more details: https://arxiv.org/abs/1703.01365
- Output
Indices []interface{} - If populated, only returns attributions that have output_index contained in output_indices. It must be an ndarray of integers, with the same shape of the output it's explaining. If not populated, returns attributions for top_k indices of outputs. If neither top_k nor output_indices is populated, returns the argmax index of the outputs. Only applicable to Models that predict multiple outputs (e,g, multi-class Models that predict multiple classes).
- Sampled
Shapley GoogleAttribution Cloud Aiplatform V1beta1Sampled Shapley Attribution Response - An attribution method that approximates Shapley values for features that contribute to the label being predicted. A sampling strategy is used to approximate the value rather than considering all subsets of features. Refer to this paper for model details: https://arxiv.org/abs/1306.4265.
- Top
K int - If populated, returns attributions for top K indices of outputs (defaults to 1). Only applies to Models that predicts more than one outputs (e,g, multi-class Models). When set to -1, returns explanations for all outputs.
- Xrai
Attribution GoogleCloud Aiplatform V1beta1Xrai Attribution Response - An attribution method that redistributes Integrated Gradients attribution to segmented regions, taking advantage of the model's fully differentiable structure. Refer to this paper for more details: https://arxiv.org/abs/1906.02825 XRAI currently performs better on natural images, like a picture of a house or an animal. If the images are taken in artificial environments, like a lab or manufacturing line, or from diagnostic equipment, like x-rays or quality-control cameras, use Integrated Gradients instead.
- examples
Google
Cloud Aiplatform V1beta1Examples Response - Example-based explanations that returns the nearest neighbors from the provided dataset.
- integrated
Gradients GoogleAttribution Cloud Aiplatform V1beta1Integrated Gradients Attribution Response - An attribution method that computes Aumann-Shapley values taking advantage of the model's fully differentiable structure. Refer to this paper for more details: https://arxiv.org/abs/1703.01365
- output
Indices List<Object> - If populated, only returns attributions that have output_index contained in output_indices. It must be an ndarray of integers, with the same shape of the output it's explaining. If not populated, returns attributions for top_k indices of outputs. If neither top_k nor output_indices is populated, returns the argmax index of the outputs. Only applicable to Models that predict multiple outputs (e,g, multi-class Models that predict multiple classes).
- sampled
Shapley GoogleAttribution Cloud Aiplatform V1beta1Sampled Shapley Attribution Response - An attribution method that approximates Shapley values for features that contribute to the label being predicted. A sampling strategy is used to approximate the value rather than considering all subsets of features. Refer to this paper for model details: https://arxiv.org/abs/1306.4265.
- top
K Integer - If populated, returns attributions for top K indices of outputs (defaults to 1). Only applies to Models that predicts more than one outputs (e,g, multi-class Models). When set to -1, returns explanations for all outputs.
- xrai
Attribution GoogleCloud Aiplatform V1beta1Xrai Attribution Response - An attribution method that redistributes Integrated Gradients attribution to segmented regions, taking advantage of the model's fully differentiable structure. Refer to this paper for more details: https://arxiv.org/abs/1906.02825 XRAI currently performs better on natural images, like a picture of a house or an animal. If the images are taken in artificial environments, like a lab or manufacturing line, or from diagnostic equipment, like x-rays or quality-control cameras, use Integrated Gradients instead.
- examples
Google
Cloud Aiplatform V1beta1Examples Response - Example-based explanations that returns the nearest neighbors from the provided dataset.
- integrated
Gradients GoogleAttribution Cloud Aiplatform V1beta1Integrated Gradients Attribution Response - An attribution method that computes Aumann-Shapley values taking advantage of the model's fully differentiable structure. Refer to this paper for more details: https://arxiv.org/abs/1703.01365
- output
Indices any[] - If populated, only returns attributions that have output_index contained in output_indices. It must be an ndarray of integers, with the same shape of the output it's explaining. If not populated, returns attributions for top_k indices of outputs. If neither top_k nor output_indices is populated, returns the argmax index of the outputs. Only applicable to Models that predict multiple outputs (e,g, multi-class Models that predict multiple classes).
- sampled
Shapley GoogleAttribution Cloud Aiplatform V1beta1Sampled Shapley Attribution Response - An attribution method that approximates Shapley values for features that contribute to the label being predicted. A sampling strategy is used to approximate the value rather than considering all subsets of features. Refer to this paper for model details: https://arxiv.org/abs/1306.4265.
- top
K number - If populated, returns attributions for top K indices of outputs (defaults to 1). Only applies to Models that predicts more than one outputs (e,g, multi-class Models). When set to -1, returns explanations for all outputs.
- xrai
Attribution GoogleCloud Aiplatform V1beta1Xrai Attribution Response - An attribution method that redistributes Integrated Gradients attribution to segmented regions, taking advantage of the model's fully differentiable structure. Refer to this paper for more details: https://arxiv.org/abs/1906.02825 XRAI currently performs better on natural images, like a picture of a house or an animal. If the images are taken in artificial environments, like a lab or manufacturing line, or from diagnostic equipment, like x-rays or quality-control cameras, use Integrated Gradients instead.
- examples
Google
Cloud Aiplatform V1beta1Examples Response - Example-based explanations that returns the nearest neighbors from the provided dataset.
- integrated_
gradients_ Googleattribution Cloud Aiplatform V1beta1Integrated Gradients Attribution Response - An attribution method that computes Aumann-Shapley values taking advantage of the model's fully differentiable structure. Refer to this paper for more details: https://arxiv.org/abs/1703.01365
- output_
indices Sequence[Any] - If populated, only returns attributions that have output_index contained in output_indices. It must be an ndarray of integers, with the same shape of the output it's explaining. If not populated, returns attributions for top_k indices of outputs. If neither top_k nor output_indices is populated, returns the argmax index of the outputs. Only applicable to Models that predict multiple outputs (e,g, multi-class Models that predict multiple classes).
- sampled_
shapley_ Googleattribution Cloud Aiplatform V1beta1Sampled Shapley Attribution Response - An attribution method that approximates Shapley values for features that contribute to the label being predicted. A sampling strategy is used to approximate the value rather than considering all subsets of features. Refer to this paper for model details: https://arxiv.org/abs/1306.4265.
- top_
k int - If populated, returns attributions for top K indices of outputs (defaults to 1). Only applies to Models that predicts more than one outputs (e,g, multi-class Models). When set to -1, returns explanations for all outputs.
- xrai_
attribution GoogleCloud Aiplatform V1beta1Xrai Attribution Response - An attribution method that redistributes Integrated Gradients attribution to segmented regions, taking advantage of the model's fully differentiable structure. Refer to this paper for more details: https://arxiv.org/abs/1906.02825 XRAI currently performs better on natural images, like a picture of a house or an animal. If the images are taken in artificial environments, like a lab or manufacturing line, or from diagnostic equipment, like x-rays or quality-control cameras, use Integrated Gradients instead.
- examples Property Map
- Example-based explanations that returns the nearest neighbors from the provided dataset.
- integrated
Gradients Property MapAttribution - An attribution method that computes Aumann-Shapley values taking advantage of the model's fully differentiable structure. Refer to this paper for more details: https://arxiv.org/abs/1703.01365
- output
Indices List<Any> - If populated, only returns attributions that have output_index contained in output_indices. It must be an ndarray of integers, with the same shape of the output it's explaining. If not populated, returns attributions for top_k indices of outputs. If neither top_k nor output_indices is populated, returns the argmax index of the outputs. Only applicable to Models that predict multiple outputs (e,g, multi-class Models that predict multiple classes).
- sampled
Shapley Property MapAttribution - An attribution method that approximates Shapley values for features that contribute to the label being predicted. A sampling strategy is used to approximate the value rather than considering all subsets of features. Refer to this paper for model details: https://arxiv.org/abs/1306.4265.
- top
K Number - If populated, returns attributions for top K indices of outputs (defaults to 1). Only applies to Models that predicts more than one outputs (e,g, multi-class Models). When set to -1, returns explanations for all outputs.
- xrai
Attribution Property Map - An attribution method that redistributes Integrated Gradients attribution to segmented regions, taking advantage of the model's fully differentiable structure. Refer to this paper for more details: https://arxiv.org/abs/1906.02825 XRAI currently performs better on natural images, like a picture of a house or an animal. If the images are taken in artificial environments, like a lab or manufacturing line, or from diagnostic equipment, like x-rays or quality-control cameras, use Integrated Gradients instead.
GoogleCloudAiplatformV1beta1ExplanationSpecResponse, GoogleCloudAiplatformV1beta1ExplanationSpecResponseArgs
- Metadata
Pulumi.
Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Explanation Metadata Response - Optional. Metadata describing the Model's input and output for explanation.
- Parameters
Pulumi.
Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Explanation Parameters Response - Parameters that configure explaining of the Model's predictions.
- Metadata
Google
Cloud Aiplatform V1beta1Explanation Metadata Response - Optional. Metadata describing the Model's input and output for explanation.
- Parameters
Google
Cloud Aiplatform V1beta1Explanation Parameters Response - Parameters that configure explaining of the Model's predictions.
- metadata
Google
Cloud Aiplatform V1beta1Explanation Metadata Response - Optional. Metadata describing the Model's input and output for explanation.
- parameters
Google
Cloud Aiplatform V1beta1Explanation Parameters Response - Parameters that configure explaining of the Model's predictions.
- metadata
Google
Cloud Aiplatform V1beta1Explanation Metadata Response - Optional. Metadata describing the Model's input and output for explanation.
- parameters
Google
Cloud Aiplatform V1beta1Explanation Parameters Response - Parameters that configure explaining of the Model's predictions.
- metadata
Google
Cloud Aiplatform V1beta1Explanation Metadata Response - Optional. Metadata describing the Model's input and output for explanation.
- parameters
Google
Cloud Aiplatform V1beta1Explanation Parameters Response - Parameters that configure explaining of the Model's predictions.
- metadata Property Map
- Optional. Metadata describing the Model's input and output for explanation.
- parameters Property Map
- Parameters that configure explaining of the Model's predictions.
GoogleCloudAiplatformV1beta1FeatureNoiseSigmaNoiseSigmaForFeatureResponse, GoogleCloudAiplatformV1beta1FeatureNoiseSigmaNoiseSigmaForFeatureResponseArgs
- Name string
- The name of the input feature for which noise sigma is provided. The features are defined in explanation metadata inputs.
- Sigma double
- This represents the standard deviation of the Gaussian kernel that will be used to add noise to the feature prior to computing gradients. Similar to noise_sigma but represents the noise added to the current feature. Defaults to 0.1.
- Name string
- The name of the input feature for which noise sigma is provided. The features are defined in explanation metadata inputs.
- Sigma float64
- This represents the standard deviation of the Gaussian kernel that will be used to add noise to the feature prior to computing gradients. Similar to noise_sigma but represents the noise added to the current feature. Defaults to 0.1.
- name String
- The name of the input feature for which noise sigma is provided. The features are defined in explanation metadata inputs.
- sigma Double
- This represents the standard deviation of the Gaussian kernel that will be used to add noise to the feature prior to computing gradients. Similar to noise_sigma but represents the noise added to the current feature. Defaults to 0.1.
- name string
- The name of the input feature for which noise sigma is provided. The features are defined in explanation metadata inputs.
- sigma number
- This represents the standard deviation of the Gaussian kernel that will be used to add noise to the feature prior to computing gradients. Similar to noise_sigma but represents the noise added to the current feature. Defaults to 0.1.
- name str
- The name of the input feature for which noise sigma is provided. The features are defined in explanation metadata inputs.
- sigma float
- This represents the standard deviation of the Gaussian kernel that will be used to add noise to the feature prior to computing gradients. Similar to noise_sigma but represents the noise added to the current feature. Defaults to 0.1.
- name String
- The name of the input feature for which noise sigma is provided. The features are defined in explanation metadata inputs.
- sigma Number
- This represents the standard deviation of the Gaussian kernel that will be used to add noise to the feature prior to computing gradients. Similar to noise_sigma but represents the noise added to the current feature. Defaults to 0.1.
GoogleCloudAiplatformV1beta1FeatureNoiseSigmaResponse, GoogleCloudAiplatformV1beta1FeatureNoiseSigmaResponseArgs
- Noise
Sigma List<Pulumi.Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Feature Noise Sigma Noise Sigma For Feature Response> - Noise sigma per feature. No noise is added to features that are not set.
- Noise
Sigma []GoogleCloud Aiplatform V1beta1Feature Noise Sigma Noise Sigma For Feature Response - Noise sigma per feature. No noise is added to features that are not set.
- noise
Sigma List<GoogleCloud Aiplatform V1beta1Feature Noise Sigma Noise Sigma For Feature Response> - Noise sigma per feature. No noise is added to features that are not set.
- noise
Sigma GoogleCloud Aiplatform V1beta1Feature Noise Sigma Noise Sigma For Feature Response[] - Noise sigma per feature. No noise is added to features that are not set.
- noise_
sigma Sequence[GoogleCloud Aiplatform V1beta1Feature Noise Sigma Noise Sigma For Feature Response] - Noise sigma per feature. No noise is added to features that are not set.
- noise
Sigma List<Property Map> - Noise sigma per feature. No noise is added to features that are not set.
GoogleCloudAiplatformV1beta1GcsSourceResponse, GoogleCloudAiplatformV1beta1GcsSourceResponseArgs
- Uris List<string>
- Google Cloud Storage URI(-s) to the input file(s). May contain wildcards. For more information on wildcards, see https://cloud.google.com/storage/docs/gsutil/addlhelp/WildcardNames.
- Uris []string
- Google Cloud Storage URI(-s) to the input file(s). May contain wildcards. For more information on wildcards, see https://cloud.google.com/storage/docs/gsutil/addlhelp/WildcardNames.
- uris List<String>
- Google Cloud Storage URI(-s) to the input file(s). May contain wildcards. For more information on wildcards, see https://cloud.google.com/storage/docs/gsutil/addlhelp/WildcardNames.
- uris string[]
- Google Cloud Storage URI(-s) to the input file(s). May contain wildcards. For more information on wildcards, see https://cloud.google.com/storage/docs/gsutil/addlhelp/WildcardNames.
- uris Sequence[str]
- Google Cloud Storage URI(-s) to the input file(s). May contain wildcards. For more information on wildcards, see https://cloud.google.com/storage/docs/gsutil/addlhelp/WildcardNames.
- uris List<String>
- Google Cloud Storage URI(-s) to the input file(s). May contain wildcards. For more information on wildcards, see https://cloud.google.com/storage/docs/gsutil/addlhelp/WildcardNames.
GoogleCloudAiplatformV1beta1IntegratedGradientsAttributionResponse, GoogleCloudAiplatformV1beta1IntegratedGradientsAttributionResponseArgs
- Blur
Baseline Pulumi.Config Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Blur Baseline Config Response - Config for IG with blur baseline. When enabled, a linear path from the maximally blurred image to the input image is created. Using a blurred baseline instead of zero (black image) is motivated by the BlurIG approach explained here: https://arxiv.org/abs/2004.03383
- Smooth
Grad Pulumi.Config Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Smooth Grad Config Response - Config for SmoothGrad approximation of gradients. When enabled, the gradients are approximated by averaging the gradients from noisy samples in the vicinity of the inputs. Adding noise can help improve the computed gradients. Refer to this paper for more details: https://arxiv.org/pdf/1706.03825.pdf
- Step
Count int - The number of steps for approximating the path integral. A good value to start is 50 and gradually increase until the sum to diff property is within the desired error range. Valid range of its value is [1, 100], inclusively.
- Blur
Baseline GoogleConfig Cloud Aiplatform V1beta1Blur Baseline Config Response - Config for IG with blur baseline. When enabled, a linear path from the maximally blurred image to the input image is created. Using a blurred baseline instead of zero (black image) is motivated by the BlurIG approach explained here: https://arxiv.org/abs/2004.03383
- Smooth
Grad GoogleConfig Cloud Aiplatform V1beta1Smooth Grad Config Response - Config for SmoothGrad approximation of gradients. When enabled, the gradients are approximated by averaging the gradients from noisy samples in the vicinity of the inputs. Adding noise can help improve the computed gradients. Refer to this paper for more details: https://arxiv.org/pdf/1706.03825.pdf
- Step
Count int - The number of steps for approximating the path integral. A good value to start is 50 and gradually increase until the sum to diff property is within the desired error range. Valid range of its value is [1, 100], inclusively.
- blur
Baseline GoogleConfig Cloud Aiplatform V1beta1Blur Baseline Config Response - Config for IG with blur baseline. When enabled, a linear path from the maximally blurred image to the input image is created. Using a blurred baseline instead of zero (black image) is motivated by the BlurIG approach explained here: https://arxiv.org/abs/2004.03383
- smooth
Grad GoogleConfig Cloud Aiplatform V1beta1Smooth Grad Config Response - Config for SmoothGrad approximation of gradients. When enabled, the gradients are approximated by averaging the gradients from noisy samples in the vicinity of the inputs. Adding noise can help improve the computed gradients. Refer to this paper for more details: https://arxiv.org/pdf/1706.03825.pdf
- step
Count Integer - The number of steps for approximating the path integral. A good value to start is 50 and gradually increase until the sum to diff property is within the desired error range. Valid range of its value is [1, 100], inclusively.
- blur
Baseline GoogleConfig Cloud Aiplatform V1beta1Blur Baseline Config Response - Config for IG with blur baseline. When enabled, a linear path from the maximally blurred image to the input image is created. Using a blurred baseline instead of zero (black image) is motivated by the BlurIG approach explained here: https://arxiv.org/abs/2004.03383
- smooth
Grad GoogleConfig Cloud Aiplatform V1beta1Smooth Grad Config Response - Config for SmoothGrad approximation of gradients. When enabled, the gradients are approximated by averaging the gradients from noisy samples in the vicinity of the inputs. Adding noise can help improve the computed gradients. Refer to this paper for more details: https://arxiv.org/pdf/1706.03825.pdf
- step
Count number - The number of steps for approximating the path integral. A good value to start is 50 and gradually increase until the sum to diff property is within the desired error range. Valid range of its value is [1, 100], inclusively.
- blur_
baseline_ Googleconfig Cloud Aiplatform V1beta1Blur Baseline Config Response - Config for IG with blur baseline. When enabled, a linear path from the maximally blurred image to the input image is created. Using a blurred baseline instead of zero (black image) is motivated by the BlurIG approach explained here: https://arxiv.org/abs/2004.03383
- smooth_
grad_ Googleconfig Cloud Aiplatform V1beta1Smooth Grad Config Response - Config for SmoothGrad approximation of gradients. When enabled, the gradients are approximated by averaging the gradients from noisy samples in the vicinity of the inputs. Adding noise can help improve the computed gradients. Refer to this paper for more details: https://arxiv.org/pdf/1706.03825.pdf
- step_
count int - The number of steps for approximating the path integral. A good value to start is 50 and gradually increase until the sum to diff property is within the desired error range. Valid range of its value is [1, 100], inclusively.
- blur
Baseline Property MapConfig - Config for IG with blur baseline. When enabled, a linear path from the maximally blurred image to the input image is created. Using a blurred baseline instead of zero (black image) is motivated by the BlurIG approach explained here: https://arxiv.org/abs/2004.03383
- smooth
Grad Property MapConfig - Config for SmoothGrad approximation of gradients. When enabled, the gradients are approximated by averaging the gradients from noisy samples in the vicinity of the inputs. Adding noise can help improve the computed gradients. Refer to this paper for more details: https://arxiv.org/pdf/1706.03825.pdf
- step
Count Number - The number of steps for approximating the path integral. A good value to start is 50 and gradually increase until the sum to diff property is within the desired error range. Valid range of its value is [1, 100], inclusively.
GoogleCloudAiplatformV1beta1MachineSpecResponse, GoogleCloudAiplatformV1beta1MachineSpecResponseArgs
- Accelerator
Count int - The number of accelerators to attach to the machine.
- Accelerator
Type string - Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count.
- Machine
Type string - Immutable. The type of the machine. See the list of machine types supported for prediction See the list of machine types supported for custom training. For DeployedModel this field is optional, and the default value is
n1-standard-2
. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. - Tpu
Topology string - Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1").
- Accelerator
Count int - The number of accelerators to attach to the machine.
- Accelerator
Type string - Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count.
- Machine
Type string - Immutable. The type of the machine. See the list of machine types supported for prediction See the list of machine types supported for custom training. For DeployedModel this field is optional, and the default value is
n1-standard-2
. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. - Tpu
Topology string - Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1").
- accelerator
Count Integer - The number of accelerators to attach to the machine.
- accelerator
Type String - Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count.
- machine
Type String - Immutable. The type of the machine. See the list of machine types supported for prediction See the list of machine types supported for custom training. For DeployedModel this field is optional, and the default value is
n1-standard-2
. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. - tpu
Topology String - Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1").
- accelerator
Count number - The number of accelerators to attach to the machine.
- accelerator
Type string - Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count.
- machine
Type string - Immutable. The type of the machine. See the list of machine types supported for prediction See the list of machine types supported for custom training. For DeployedModel this field is optional, and the default value is
n1-standard-2
. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. - tpu
Topology string - Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1").
- accelerator_
count int - The number of accelerators to attach to the machine.
- accelerator_
type str - Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count.
- machine_
type str - Immutable. The type of the machine. See the list of machine types supported for prediction See the list of machine types supported for custom training. For DeployedModel this field is optional, and the default value is
n1-standard-2
. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. - tpu_
topology str - Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1").
- accelerator
Count Number - The number of accelerators to attach to the machine.
- accelerator
Type String - Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count.
- machine
Type String - Immutable. The type of the machine. See the list of machine types supported for prediction See the list of machine types supported for custom training. For DeployedModel this field is optional, and the default value is
n1-standard-2
. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. - tpu
Topology String - Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1").
GoogleCloudAiplatformV1beta1PredictRequestResponseLoggingConfig, GoogleCloudAiplatformV1beta1PredictRequestResponseLoggingConfigArgs
- Bigquery
Destination Pulumi.Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Big Query Destination - BigQuery table for logging. If only given a project, a new dataset will be created with name
logging__
where will be made BigQuery-dataset-name compatible (e.g. most special characters will become underscores). If no table name is given, a new table will be created with namerequest_response_logging
- Enabled bool
- If logging is enabled or not.
- Sampling
Rate double - Percentage of requests to be logged, expressed as a fraction in range(0,1].
- Bigquery
Destination GoogleCloud Aiplatform V1beta1Big Query Destination - BigQuery table for logging. If only given a project, a new dataset will be created with name
logging__
where will be made BigQuery-dataset-name compatible (e.g. most special characters will become underscores). If no table name is given, a new table will be created with namerequest_response_logging
- Enabled bool
- If logging is enabled or not.
- Sampling
Rate float64 - Percentage of requests to be logged, expressed as a fraction in range(0,1].
- bigquery
Destination GoogleCloud Aiplatform V1beta1Big Query Destination - BigQuery table for logging. If only given a project, a new dataset will be created with name
logging__
where will be made BigQuery-dataset-name compatible (e.g. most special characters will become underscores). If no table name is given, a new table will be created with namerequest_response_logging
- enabled Boolean
- If logging is enabled or not.
- sampling
Rate Double - Percentage of requests to be logged, expressed as a fraction in range(0,1].
- bigquery
Destination GoogleCloud Aiplatform V1beta1Big Query Destination - BigQuery table for logging. If only given a project, a new dataset will be created with name
logging__
where will be made BigQuery-dataset-name compatible (e.g. most special characters will become underscores). If no table name is given, a new table will be created with namerequest_response_logging
- enabled boolean
- If logging is enabled or not.
- sampling
Rate number - Percentage of requests to be logged, expressed as a fraction in range(0,1].
- bigquery_
destination GoogleCloud Aiplatform V1beta1Big Query Destination - BigQuery table for logging. If only given a project, a new dataset will be created with name
logging__
where will be made BigQuery-dataset-name compatible (e.g. most special characters will become underscores). If no table name is given, a new table will be created with namerequest_response_logging
- enabled bool
- If logging is enabled or not.
- sampling_
rate float - Percentage of requests to be logged, expressed as a fraction in range(0,1].
- bigquery
Destination Property Map - BigQuery table for logging. If only given a project, a new dataset will be created with name
logging__
where will be made BigQuery-dataset-name compatible (e.g. most special characters will become underscores). If no table name is given, a new table will be created with namerequest_response_logging
- enabled Boolean
- If logging is enabled or not.
- sampling
Rate Number - Percentage of requests to be logged, expressed as a fraction in range(0,1].
GoogleCloudAiplatformV1beta1PredictRequestResponseLoggingConfigResponse, GoogleCloudAiplatformV1beta1PredictRequestResponseLoggingConfigResponseArgs
- Bigquery
Destination Pulumi.Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Big Query Destination Response - BigQuery table for logging. If only given a project, a new dataset will be created with name
logging__
where will be made BigQuery-dataset-name compatible (e.g. most special characters will become underscores). If no table name is given, a new table will be created with namerequest_response_logging
- Enabled bool
- If logging is enabled or not.
- Sampling
Rate double - Percentage of requests to be logged, expressed as a fraction in range(0,1].
- Bigquery
Destination GoogleCloud Aiplatform V1beta1Big Query Destination Response - BigQuery table for logging. If only given a project, a new dataset will be created with name
logging__
where will be made BigQuery-dataset-name compatible (e.g. most special characters will become underscores). If no table name is given, a new table will be created with namerequest_response_logging
- Enabled bool
- If logging is enabled or not.
- Sampling
Rate float64 - Percentage of requests to be logged, expressed as a fraction in range(0,1].
- bigquery
Destination GoogleCloud Aiplatform V1beta1Big Query Destination Response - BigQuery table for logging. If only given a project, a new dataset will be created with name
logging__
where will be made BigQuery-dataset-name compatible (e.g. most special characters will become underscores). If no table name is given, a new table will be created with namerequest_response_logging
- enabled Boolean
- If logging is enabled or not.
- sampling
Rate Double - Percentage of requests to be logged, expressed as a fraction in range(0,1].
- bigquery
Destination GoogleCloud Aiplatform V1beta1Big Query Destination Response - BigQuery table for logging. If only given a project, a new dataset will be created with name
logging__
where will be made BigQuery-dataset-name compatible (e.g. most special characters will become underscores). If no table name is given, a new table will be created with namerequest_response_logging
- enabled boolean
- If logging is enabled or not.
- sampling
Rate number - Percentage of requests to be logged, expressed as a fraction in range(0,1].
- bigquery_
destination GoogleCloud Aiplatform V1beta1Big Query Destination Response - BigQuery table for logging. If only given a project, a new dataset will be created with name
logging__
where will be made BigQuery-dataset-name compatible (e.g. most special characters will become underscores). If no table name is given, a new table will be created with namerequest_response_logging
- enabled bool
- If logging is enabled or not.
- sampling_
rate float - Percentage of requests to be logged, expressed as a fraction in range(0,1].
- bigquery
Destination Property Map - BigQuery table for logging. If only given a project, a new dataset will be created with name
logging__
where will be made BigQuery-dataset-name compatible (e.g. most special characters will become underscores). If no table name is given, a new table will be created with namerequest_response_logging
- enabled Boolean
- If logging is enabled or not.
- sampling
Rate Number - Percentage of requests to be logged, expressed as a fraction in range(0,1].
GoogleCloudAiplatformV1beta1PresetsResponse, GoogleCloudAiplatformV1beta1PresetsResponseArgs
- Modality string
- The modality of the uploaded model, which automatically configures the distance measurement and feature normalization for the underlying example index and queries. If your model does not precisely fit one of these types, it is okay to choose the closest type.
- Query string
- Preset option controlling parameters for speed-precision trade-off when querying for examples. If omitted, defaults to
PRECISE
.
- Modality string
- The modality of the uploaded model, which automatically configures the distance measurement and feature normalization for the underlying example index and queries. If your model does not precisely fit one of these types, it is okay to choose the closest type.
- Query string
- Preset option controlling parameters for speed-precision trade-off when querying for examples. If omitted, defaults to
PRECISE
.
- modality String
- The modality of the uploaded model, which automatically configures the distance measurement and feature normalization for the underlying example index and queries. If your model does not precisely fit one of these types, it is okay to choose the closest type.
- query String
- Preset option controlling parameters for speed-precision trade-off when querying for examples. If omitted, defaults to
PRECISE
.
- modality string
- The modality of the uploaded model, which automatically configures the distance measurement and feature normalization for the underlying example index and queries. If your model does not precisely fit one of these types, it is okay to choose the closest type.
- query string
- Preset option controlling parameters for speed-precision trade-off when querying for examples. If omitted, defaults to
PRECISE
.
- modality str
- The modality of the uploaded model, which automatically configures the distance measurement and feature normalization for the underlying example index and queries. If your model does not precisely fit one of these types, it is okay to choose the closest type.
- query str
- Preset option controlling parameters for speed-precision trade-off when querying for examples. If omitted, defaults to
PRECISE
.
- modality String
- The modality of the uploaded model, which automatically configures the distance measurement and feature normalization for the underlying example index and queries. If your model does not precisely fit one of these types, it is okay to choose the closest type.
- query String
- Preset option controlling parameters for speed-precision trade-off when querying for examples. If omitted, defaults to
PRECISE
.
GoogleCloudAiplatformV1beta1PrivateEndpointsResponse, GoogleCloudAiplatformV1beta1PrivateEndpointsResponseArgs
- Explain
Http stringUri - Http(s) path to send explain requests.
- Health
Http stringUri - Http(s) path to send health check requests.
- Predict
Http stringUri - Http(s) path to send prediction requests.
- Service
Attachment string - The name of the service attachment resource. Populated if private service connect is enabled.
- Explain
Http stringUri - Http(s) path to send explain requests.
- Health
Http stringUri - Http(s) path to send health check requests.
- Predict
Http stringUri - Http(s) path to send prediction requests.
- Service
Attachment string - The name of the service attachment resource. Populated if private service connect is enabled.
- explain
Http StringUri - Http(s) path to send explain requests.
- health
Http StringUri - Http(s) path to send health check requests.
- predict
Http StringUri - Http(s) path to send prediction requests.
- service
Attachment String - The name of the service attachment resource. Populated if private service connect is enabled.
- explain
Http stringUri - Http(s) path to send explain requests.
- health
Http stringUri - Http(s) path to send health check requests.
- predict
Http stringUri - Http(s) path to send prediction requests.
- service
Attachment string - The name of the service attachment resource. Populated if private service connect is enabled.
- explain_
http_ struri - Http(s) path to send explain requests.
- health_
http_ struri - Http(s) path to send health check requests.
- predict_
http_ struri - Http(s) path to send prediction requests.
- service_
attachment str - The name of the service attachment resource. Populated if private service connect is enabled.
- explain
Http StringUri - Http(s) path to send explain requests.
- health
Http StringUri - Http(s) path to send health check requests.
- predict
Http StringUri - Http(s) path to send prediction requests.
- service
Attachment String - The name of the service attachment resource. Populated if private service connect is enabled.
GoogleCloudAiplatformV1beta1SampledShapleyAttributionResponse, GoogleCloudAiplatformV1beta1SampledShapleyAttributionResponseArgs
- Path
Count int - The number of feature permutations to consider when approximating the Shapley values. Valid range of its value is [1, 50], inclusively.
- Path
Count int - The number of feature permutations to consider when approximating the Shapley values. Valid range of its value is [1, 50], inclusively.
- path
Count Integer - The number of feature permutations to consider when approximating the Shapley values. Valid range of its value is [1, 50], inclusively.
- path
Count number - The number of feature permutations to consider when approximating the Shapley values. Valid range of its value is [1, 50], inclusively.
- path_
count int - The number of feature permutations to consider when approximating the Shapley values. Valid range of its value is [1, 50], inclusively.
- path
Count Number - The number of feature permutations to consider when approximating the Shapley values. Valid range of its value is [1, 50], inclusively.
GoogleCloudAiplatformV1beta1SmoothGradConfigResponse, GoogleCloudAiplatformV1beta1SmoothGradConfigResponseArgs
- Feature
Noise Pulumi.Sigma Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Feature Noise Sigma Response - This is similar to noise_sigma, but provides additional flexibility. A separate noise sigma can be provided for each feature, which is useful if their distributions are different. No noise is added to features that are not set. If this field is unset, noise_sigma will be used for all features.
- Noise
Sigma double - This is a single float value and will be used to add noise to all the features. Use this field when all features are normalized to have the same distribution: scale to range [0, 1], [-1, 1] or z-scoring, where features are normalized to have 0-mean and 1-variance. Learn more about normalization. For best results the recommended value is about 10% - 20% of the standard deviation of the input feature. Refer to section 3.2 of the SmoothGrad paper: https://arxiv.org/pdf/1706.03825.pdf. Defaults to 0.1. If the distribution is different per feature, set feature_noise_sigma instead for each feature.
- Noisy
Sample intCount - The number of gradient samples to use for approximation. The higher this number, the more accurate the gradient is, but the runtime complexity increases by this factor as well. Valid range of its value is [1, 50]. Defaults to 3.
- Feature
Noise GoogleSigma Cloud Aiplatform V1beta1Feature Noise Sigma Response - This is similar to noise_sigma, but provides additional flexibility. A separate noise sigma can be provided for each feature, which is useful if their distributions are different. No noise is added to features that are not set. If this field is unset, noise_sigma will be used for all features.
- Noise
Sigma float64 - This is a single float value and will be used to add noise to all the features. Use this field when all features are normalized to have the same distribution: scale to range [0, 1], [-1, 1] or z-scoring, where features are normalized to have 0-mean and 1-variance. Learn more about normalization. For best results the recommended value is about 10% - 20% of the standard deviation of the input feature. Refer to section 3.2 of the SmoothGrad paper: https://arxiv.org/pdf/1706.03825.pdf. Defaults to 0.1. If the distribution is different per feature, set feature_noise_sigma instead for each feature.
- Noisy
Sample intCount - The number of gradient samples to use for approximation. The higher this number, the more accurate the gradient is, but the runtime complexity increases by this factor as well. Valid range of its value is [1, 50]. Defaults to 3.
- feature
Noise GoogleSigma Cloud Aiplatform V1beta1Feature Noise Sigma Response - This is similar to noise_sigma, but provides additional flexibility. A separate noise sigma can be provided for each feature, which is useful if their distributions are different. No noise is added to features that are not set. If this field is unset, noise_sigma will be used for all features.
- noise
Sigma Double - This is a single float value and will be used to add noise to all the features. Use this field when all features are normalized to have the same distribution: scale to range [0, 1], [-1, 1] or z-scoring, where features are normalized to have 0-mean and 1-variance. Learn more about normalization. For best results the recommended value is about 10% - 20% of the standard deviation of the input feature. Refer to section 3.2 of the SmoothGrad paper: https://arxiv.org/pdf/1706.03825.pdf. Defaults to 0.1. If the distribution is different per feature, set feature_noise_sigma instead for each feature.
- noisy
Sample IntegerCount - The number of gradient samples to use for approximation. The higher this number, the more accurate the gradient is, but the runtime complexity increases by this factor as well. Valid range of its value is [1, 50]. Defaults to 3.
- feature
Noise GoogleSigma Cloud Aiplatform V1beta1Feature Noise Sigma Response - This is similar to noise_sigma, but provides additional flexibility. A separate noise sigma can be provided for each feature, which is useful if their distributions are different. No noise is added to features that are not set. If this field is unset, noise_sigma will be used for all features.
- noise
Sigma number - This is a single float value and will be used to add noise to all the features. Use this field when all features are normalized to have the same distribution: scale to range [0, 1], [-1, 1] or z-scoring, where features are normalized to have 0-mean and 1-variance. Learn more about normalization. For best results the recommended value is about 10% - 20% of the standard deviation of the input feature. Refer to section 3.2 of the SmoothGrad paper: https://arxiv.org/pdf/1706.03825.pdf. Defaults to 0.1. If the distribution is different per feature, set feature_noise_sigma instead for each feature.
- noisy
Sample numberCount - The number of gradient samples to use for approximation. The higher this number, the more accurate the gradient is, but the runtime complexity increases by this factor as well. Valid range of its value is [1, 50]. Defaults to 3.
- feature_
noise_ Googlesigma Cloud Aiplatform V1beta1Feature Noise Sigma Response - This is similar to noise_sigma, but provides additional flexibility. A separate noise sigma can be provided for each feature, which is useful if their distributions are different. No noise is added to features that are not set. If this field is unset, noise_sigma will be used for all features.
- noise_
sigma float - This is a single float value and will be used to add noise to all the features. Use this field when all features are normalized to have the same distribution: scale to range [0, 1], [-1, 1] or z-scoring, where features are normalized to have 0-mean and 1-variance. Learn more about normalization. For best results the recommended value is about 10% - 20% of the standard deviation of the input feature. Refer to section 3.2 of the SmoothGrad paper: https://arxiv.org/pdf/1706.03825.pdf. Defaults to 0.1. If the distribution is different per feature, set feature_noise_sigma instead for each feature.
- noisy_
sample_ intcount - The number of gradient samples to use for approximation. The higher this number, the more accurate the gradient is, but the runtime complexity increases by this factor as well. Valid range of its value is [1, 50]. Defaults to 3.
- feature
Noise Property MapSigma - This is similar to noise_sigma, but provides additional flexibility. A separate noise sigma can be provided for each feature, which is useful if their distributions are different. No noise is added to features that are not set. If this field is unset, noise_sigma will be used for all features.
- noise
Sigma Number - This is a single float value and will be used to add noise to all the features. Use this field when all features are normalized to have the same distribution: scale to range [0, 1], [-1, 1] or z-scoring, where features are normalized to have 0-mean and 1-variance. Learn more about normalization. For best results the recommended value is about 10% - 20% of the standard deviation of the input feature. Refer to section 3.2 of the SmoothGrad paper: https://arxiv.org/pdf/1706.03825.pdf. Defaults to 0.1. If the distribution is different per feature, set feature_noise_sigma instead for each feature.
- noisy
Sample NumberCount - The number of gradient samples to use for approximation. The higher this number, the more accurate the gradient is, but the runtime complexity increases by this factor as well. Valid range of its value is [1, 50]. Defaults to 3.
GoogleCloudAiplatformV1beta1XraiAttributionResponse, GoogleCloudAiplatformV1beta1XraiAttributionResponseArgs
- Blur
Baseline Pulumi.Config Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Blur Baseline Config Response - Config for XRAI with blur baseline. When enabled, a linear path from the maximally blurred image to the input image is created. Using a blurred baseline instead of zero (black image) is motivated by the BlurIG approach explained here: https://arxiv.org/abs/2004.03383
- Smooth
Grad Pulumi.Config Google Native. Aiplatform. V1Beta1. Inputs. Google Cloud Aiplatform V1beta1Smooth Grad Config Response - Config for SmoothGrad approximation of gradients. When enabled, the gradients are approximated by averaging the gradients from noisy samples in the vicinity of the inputs. Adding noise can help improve the computed gradients. Refer to this paper for more details: https://arxiv.org/pdf/1706.03825.pdf
- Step
Count int - The number of steps for approximating the path integral. A good value to start is 50 and gradually increase until the sum to diff property is met within the desired error range. Valid range of its value is [1, 100], inclusively.
- Blur
Baseline GoogleConfig Cloud Aiplatform V1beta1Blur Baseline Config Response - Config for XRAI with blur baseline. When enabled, a linear path from the maximally blurred image to the input image is created. Using a blurred baseline instead of zero (black image) is motivated by the BlurIG approach explained here: https://arxiv.org/abs/2004.03383
- Smooth
Grad GoogleConfig Cloud Aiplatform V1beta1Smooth Grad Config Response - Config for SmoothGrad approximation of gradients. When enabled, the gradients are approximated by averaging the gradients from noisy samples in the vicinity of the inputs. Adding noise can help improve the computed gradients. Refer to this paper for more details: https://arxiv.org/pdf/1706.03825.pdf
- Step
Count int - The number of steps for approximating the path integral. A good value to start is 50 and gradually increase until the sum to diff property is met within the desired error range. Valid range of its value is [1, 100], inclusively.
- blur
Baseline GoogleConfig Cloud Aiplatform V1beta1Blur Baseline Config Response - Config for XRAI with blur baseline. When enabled, a linear path from the maximally blurred image to the input image is created. Using a blurred baseline instead of zero (black image) is motivated by the BlurIG approach explained here: https://arxiv.org/abs/2004.03383
- smooth
Grad GoogleConfig Cloud Aiplatform V1beta1Smooth Grad Config Response - Config for SmoothGrad approximation of gradients. When enabled, the gradients are approximated by averaging the gradients from noisy samples in the vicinity of the inputs. Adding noise can help improve the computed gradients. Refer to this paper for more details: https://arxiv.org/pdf/1706.03825.pdf
- step
Count Integer - The number of steps for approximating the path integral. A good value to start is 50 and gradually increase until the sum to diff property is met within the desired error range. Valid range of its value is [1, 100], inclusively.
- blur
Baseline GoogleConfig Cloud Aiplatform V1beta1Blur Baseline Config Response - Config for XRAI with blur baseline. When enabled, a linear path from the maximally blurred image to the input image is created. Using a blurred baseline instead of zero (black image) is motivated by the BlurIG approach explained here: https://arxiv.org/abs/2004.03383
- smooth
Grad GoogleConfig Cloud Aiplatform V1beta1Smooth Grad Config Response - Config for SmoothGrad approximation of gradients. When enabled, the gradients are approximated by averaging the gradients from noisy samples in the vicinity of the inputs. Adding noise can help improve the computed gradients. Refer to this paper for more details: https://arxiv.org/pdf/1706.03825.pdf
- step
Count number - The number of steps for approximating the path integral. A good value to start is 50 and gradually increase until the sum to diff property is met within the desired error range. Valid range of its value is [1, 100], inclusively.
- blur_
baseline_ Googleconfig Cloud Aiplatform V1beta1Blur Baseline Config Response - Config for XRAI with blur baseline. When enabled, a linear path from the maximally blurred image to the input image is created. Using a blurred baseline instead of zero (black image) is motivated by the BlurIG approach explained here: https://arxiv.org/abs/2004.03383
- smooth_
grad_ Googleconfig Cloud Aiplatform V1beta1Smooth Grad Config Response - Config for SmoothGrad approximation of gradients. When enabled, the gradients are approximated by averaging the gradients from noisy samples in the vicinity of the inputs. Adding noise can help improve the computed gradients. Refer to this paper for more details: https://arxiv.org/pdf/1706.03825.pdf
- step_
count int - The number of steps for approximating the path integral. A good value to start is 50 and gradually increase until the sum to diff property is met within the desired error range. Valid range of its value is [1, 100], inclusively.
- blur
Baseline Property MapConfig - Config for XRAI with blur baseline. When enabled, a linear path from the maximally blurred image to the input image is created. Using a blurred baseline instead of zero (black image) is motivated by the BlurIG approach explained here: https://arxiv.org/abs/2004.03383
- smooth
Grad Property MapConfig - Config for SmoothGrad approximation of gradients. When enabled, the gradients are approximated by averaging the gradients from noisy samples in the vicinity of the inputs. Adding noise can help improve the computed gradients. Refer to this paper for more details: https://arxiv.org/pdf/1706.03825.pdf
- step
Count Number - The number of steps for approximating the path integral. A good value to start is 50 and gradually increase until the sum to diff property is met within the desired error range. Valid range of its value is [1, 100], inclusively.
Package Details
- Repository
- Google Cloud Native pulumi/pulumi-google-native
- License
- Apache-2.0
Google Cloud Native is in preview. Google Cloud Classic is fully supported.