confluentcloud.KafkaClientQuota
Explore with Pulumi AI
Example Usage
import * as pulumi from "@pulumi/pulumi";
import * as confluentcloud from "@pulumi/confluentcloud";
const example = new confluentcloud.KafkaClientQuota("example", {
displayName: "test-quota",
description: "Test Quota",
throughput: {
ingressByteRate: "100",
egressByteRate: "200",
},
principals: [
appManager.id,
appManager2.id,
],
kafkaCluster: {
id: dedicated.id,
},
environment: {
id: development.id,
},
});
import pulumi
import pulumi_confluentcloud as confluentcloud
example = confluentcloud.KafkaClientQuota("example",
display_name="test-quota",
description="Test Quota",
throughput={
"ingress_byte_rate": "100",
"egress_byte_rate": "200",
},
principals=[
app_manager["id"],
app_manager2["id"],
],
kafka_cluster={
"id": dedicated["id"],
},
environment={
"id": development["id"],
})
package main
import (
"github.com/pulumi/pulumi-confluentcloud/sdk/v2/go/confluentcloud"
"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
)
func main() {
pulumi.Run(func(ctx *pulumi.Context) error {
_, err := confluentcloud.NewKafkaClientQuota(ctx, "example", &confluentcloud.KafkaClientQuotaArgs{
DisplayName: pulumi.String("test-quota"),
Description: pulumi.String("Test Quota"),
Throughput: &confluentcloud.KafkaClientQuotaThroughputArgs{
IngressByteRate: pulumi.String("100"),
EgressByteRate: pulumi.String("200"),
},
Principals: pulumi.StringArray{
appManager.Id,
appManager2.Id,
},
KafkaCluster: &confluentcloud.KafkaClientQuotaKafkaClusterArgs{
Id: pulumi.Any(dedicated.Id),
},
Environment: &confluentcloud.KafkaClientQuotaEnvironmentArgs{
Id: pulumi.Any(development.Id),
},
})
if err != nil {
return err
}
return nil
})
}
using System.Collections.Generic;
using System.Linq;
using Pulumi;
using ConfluentCloud = Pulumi.ConfluentCloud;
return await Deployment.RunAsync(() =>
{
var example = new ConfluentCloud.KafkaClientQuota("example", new()
{
DisplayName = "test-quota",
Description = "Test Quota",
Throughput = new ConfluentCloud.Inputs.KafkaClientQuotaThroughputArgs
{
IngressByteRate = "100",
EgressByteRate = "200",
},
Principals = new[]
{
appManager.Id,
appManager2.Id,
},
KafkaCluster = new ConfluentCloud.Inputs.KafkaClientQuotaKafkaClusterArgs
{
Id = dedicated.Id,
},
Environment = new ConfluentCloud.Inputs.KafkaClientQuotaEnvironmentArgs
{
Id = development.Id,
},
});
});
package generated_program;
import com.pulumi.Context;
import com.pulumi.Pulumi;
import com.pulumi.core.Output;
import com.pulumi.confluentcloud.KafkaClientQuota;
import com.pulumi.confluentcloud.KafkaClientQuotaArgs;
import com.pulumi.confluentcloud.inputs.KafkaClientQuotaThroughputArgs;
import com.pulumi.confluentcloud.inputs.KafkaClientQuotaKafkaClusterArgs;
import com.pulumi.confluentcloud.inputs.KafkaClientQuotaEnvironmentArgs;
import java.util.List;
import java.util.ArrayList;
import java.util.Map;
import java.io.File;
import java.nio.file.Files;
import java.nio.file.Paths;
public class App {
public static void main(String[] args) {
Pulumi.run(App::stack);
}
public static void stack(Context ctx) {
var example = new KafkaClientQuota("example", KafkaClientQuotaArgs.builder()
.displayName("test-quota")
.description("Test Quota")
.throughput(KafkaClientQuotaThroughputArgs.builder()
.ingressByteRate("100")
.egressByteRate("200")
.build())
.principals(
appManager.id(),
appManager2.id())
.kafkaCluster(KafkaClientQuotaKafkaClusterArgs.builder()
.id(dedicated.id())
.build())
.environment(KafkaClientQuotaEnvironmentArgs.builder()
.id(development.id())
.build())
.build());
}
}
resources:
example:
type: confluentcloud:KafkaClientQuota
properties:
displayName: test-quota
description: Test Quota
throughput:
ingressByteRate: '100'
egressByteRate: '200'
principals:
- ${appManager.id}
- ${appManager2.id}
kafkaCluster:
id: ${dedicated.id}
environment:
id: ${development.id}
Create KafkaClientQuota Resource
Resources are created with functions called constructors. To learn more about declaring and configuring resources, see Resources.
Constructor syntax
new KafkaClientQuota(name: string, args: KafkaClientQuotaArgs, opts?: CustomResourceOptions);
@overload
def KafkaClientQuota(resource_name: str,
args: KafkaClientQuotaArgs,
opts: Optional[ResourceOptions] = None)
@overload
def KafkaClientQuota(resource_name: str,
opts: Optional[ResourceOptions] = None,
display_name: Optional[str] = None,
environment: Optional[KafkaClientQuotaEnvironmentArgs] = None,
kafka_cluster: Optional[KafkaClientQuotaKafkaClusterArgs] = None,
principals: Optional[Sequence[str]] = None,
throughput: Optional[KafkaClientQuotaThroughputArgs] = None,
description: Optional[str] = None)
func NewKafkaClientQuota(ctx *Context, name string, args KafkaClientQuotaArgs, opts ...ResourceOption) (*KafkaClientQuota, error)
public KafkaClientQuota(string name, KafkaClientQuotaArgs args, CustomResourceOptions? opts = null)
public KafkaClientQuota(String name, KafkaClientQuotaArgs args)
public KafkaClientQuota(String name, KafkaClientQuotaArgs args, CustomResourceOptions options)
type: confluentcloud:KafkaClientQuota
properties: # The arguments to resource properties.
options: # Bag of options to control resource's behavior.
Parameters
- name string
- The unique name of the resource.
- args KafkaClientQuotaArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- resource_name str
- The unique name of the resource.
- args KafkaClientQuotaArgs
- The arguments to resource properties.
- opts ResourceOptions
- Bag of options to control resource's behavior.
- ctx Context
- Context object for the current deployment.
- name string
- The unique name of the resource.
- args KafkaClientQuotaArgs
- The arguments to resource properties.
- opts ResourceOption
- Bag of options to control resource's behavior.
- name string
- The unique name of the resource.
- args KafkaClientQuotaArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- name String
- The unique name of the resource.
- args KafkaClientQuotaArgs
- The arguments to resource properties.
- options CustomResourceOptions
- Bag of options to control resource's behavior.
Constructor example
The following reference example uses placeholder values for all input properties.
var kafkaClientQuotaResource = new ConfluentCloud.KafkaClientQuota("kafkaClientQuotaResource", new()
{
DisplayName = "string",
Environment = new ConfluentCloud.Inputs.KafkaClientQuotaEnvironmentArgs
{
Id = "string",
},
KafkaCluster = new ConfluentCloud.Inputs.KafkaClientQuotaKafkaClusterArgs
{
Id = "string",
},
Principals = new[]
{
"string",
},
Throughput = new ConfluentCloud.Inputs.KafkaClientQuotaThroughputArgs
{
EgressByteRate = "string",
IngressByteRate = "string",
},
Description = "string",
});
example, err := confluentcloud.NewKafkaClientQuota(ctx, "kafkaClientQuotaResource", &confluentcloud.KafkaClientQuotaArgs{
DisplayName: pulumi.String("string"),
Environment: &confluentcloud.KafkaClientQuotaEnvironmentArgs{
Id: pulumi.String("string"),
},
KafkaCluster: &confluentcloud.KafkaClientQuotaKafkaClusterArgs{
Id: pulumi.String("string"),
},
Principals: pulumi.StringArray{
pulumi.String("string"),
},
Throughput: &confluentcloud.KafkaClientQuotaThroughputArgs{
EgressByteRate: pulumi.String("string"),
IngressByteRate: pulumi.String("string"),
},
Description: pulumi.String("string"),
})
var kafkaClientQuotaResource = new KafkaClientQuota("kafkaClientQuotaResource", KafkaClientQuotaArgs.builder()
.displayName("string")
.environment(KafkaClientQuotaEnvironmentArgs.builder()
.id("string")
.build())
.kafkaCluster(KafkaClientQuotaKafkaClusterArgs.builder()
.id("string")
.build())
.principals("string")
.throughput(KafkaClientQuotaThroughputArgs.builder()
.egressByteRate("string")
.ingressByteRate("string")
.build())
.description("string")
.build());
kafka_client_quota_resource = confluentcloud.KafkaClientQuota("kafkaClientQuotaResource",
display_name="string",
environment={
"id": "string",
},
kafka_cluster={
"id": "string",
},
principals=["string"],
throughput={
"egress_byte_rate": "string",
"ingress_byte_rate": "string",
},
description="string")
const kafkaClientQuotaResource = new confluentcloud.KafkaClientQuota("kafkaClientQuotaResource", {
displayName: "string",
environment: {
id: "string",
},
kafkaCluster: {
id: "string",
},
principals: ["string"],
throughput: {
egressByteRate: "string",
ingressByteRate: "string",
},
description: "string",
});
type: confluentcloud:KafkaClientQuota
properties:
description: string
displayName: string
environment:
id: string
kafkaCluster:
id: string
principals:
- string
throughput:
egressByteRate: string
ingressByteRate: string
KafkaClientQuota Resource Properties
To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.
Inputs
In Python, inputs that are objects can be passed either as argument classes or as dictionary literals.
The KafkaClientQuota resource accepts the following input properties:
- Display
Name string - The name of the Kafka Client Quota.
- Environment
Pulumi.
Confluent Cloud. Inputs. Kafka Client Quota Environment - Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- Kafka
Cluster Pulumi.Confluent Cloud. Inputs. Kafka Client Quota Kafka Cluster - Principals List<string>
- The list of principals (i.e., service accounts or identity pools) to apply the Kafka Client Quota to. Use the special name,
"<default>"
, to represent the default quota for all users and service accounts. - Throughput
Pulumi.
Confluent Cloud. Inputs. Kafka Client Quota Throughput - Block for representing a Kafka Quota.
- Description string
- The description of the Kafka Client Quota.
- Display
Name string - The name of the Kafka Client Quota.
- Environment
Kafka
Client Quota Environment Args - Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- Kafka
Cluster KafkaClient Quota Kafka Cluster Args - Principals []string
- The list of principals (i.e., service accounts or identity pools) to apply the Kafka Client Quota to. Use the special name,
"<default>"
, to represent the default quota for all users and service accounts. - Throughput
Kafka
Client Quota Throughput Args - Block for representing a Kafka Quota.
- Description string
- The description of the Kafka Client Quota.
- display
Name String - The name of the Kafka Client Quota.
- environment
Kafka
Client Quota Environment - Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- kafka
Cluster KafkaClient Quota Kafka Cluster - principals List<String>
- The list of principals (i.e., service accounts or identity pools) to apply the Kafka Client Quota to. Use the special name,
"<default>"
, to represent the default quota for all users and service accounts. - throughput
Kafka
Client Quota Throughput - Block for representing a Kafka Quota.
- description String
- The description of the Kafka Client Quota.
- display
Name string - The name of the Kafka Client Quota.
- environment
Kafka
Client Quota Environment - Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- kafka
Cluster KafkaClient Quota Kafka Cluster - principals string[]
- The list of principals (i.e., service accounts or identity pools) to apply the Kafka Client Quota to. Use the special name,
"<default>"
, to represent the default quota for all users and service accounts. - throughput
Kafka
Client Quota Throughput - Block for representing a Kafka Quota.
- description string
- The description of the Kafka Client Quota.
- display_
name str - The name of the Kafka Client Quota.
- environment
Kafka
Client Quota Environment Args - Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- kafka_
cluster KafkaClient Quota Kafka Cluster Args - principals Sequence[str]
- The list of principals (i.e., service accounts or identity pools) to apply the Kafka Client Quota to. Use the special name,
"<default>"
, to represent the default quota for all users and service accounts. - throughput
Kafka
Client Quota Throughput Args - Block for representing a Kafka Quota.
- description str
- The description of the Kafka Client Quota.
- display
Name String - The name of the Kafka Client Quota.
- environment Property Map
- Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- kafka
Cluster Property Map - principals List<String>
- The list of principals (i.e., service accounts or identity pools) to apply the Kafka Client Quota to. Use the special name,
"<default>"
, to represent the default quota for all users and service accounts. - throughput Property Map
- Block for representing a Kafka Quota.
- description String
- The description of the Kafka Client Quota.
Outputs
All input properties are implicitly available as output properties. Additionally, the KafkaClientQuota resource produces the following output properties:
- Id string
- The provider-assigned unique ID for this managed resource.
- Id string
- The provider-assigned unique ID for this managed resource.
- id String
- The provider-assigned unique ID for this managed resource.
- id string
- The provider-assigned unique ID for this managed resource.
- id str
- The provider-assigned unique ID for this managed resource.
- id String
- The provider-assigned unique ID for this managed resource.
Look up Existing KafkaClientQuota Resource
Get an existing KafkaClientQuota resource’s state with the given name, ID, and optional extra properties used to qualify the lookup.
public static get(name: string, id: Input<ID>, state?: KafkaClientQuotaState, opts?: CustomResourceOptions): KafkaClientQuota
@staticmethod
def get(resource_name: str,
id: str,
opts: Optional[ResourceOptions] = None,
description: Optional[str] = None,
display_name: Optional[str] = None,
environment: Optional[KafkaClientQuotaEnvironmentArgs] = None,
kafka_cluster: Optional[KafkaClientQuotaKafkaClusterArgs] = None,
principals: Optional[Sequence[str]] = None,
throughput: Optional[KafkaClientQuotaThroughputArgs] = None) -> KafkaClientQuota
func GetKafkaClientQuota(ctx *Context, name string, id IDInput, state *KafkaClientQuotaState, opts ...ResourceOption) (*KafkaClientQuota, error)
public static KafkaClientQuota Get(string name, Input<string> id, KafkaClientQuotaState? state, CustomResourceOptions? opts = null)
public static KafkaClientQuota get(String name, Output<String> id, KafkaClientQuotaState state, CustomResourceOptions options)
Resource lookup is not supported in YAML
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- resource_name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- Description string
- The description of the Kafka Client Quota.
- Display
Name string - The name of the Kafka Client Quota.
- Environment
Pulumi.
Confluent Cloud. Inputs. Kafka Client Quota Environment - Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- Kafka
Cluster Pulumi.Confluent Cloud. Inputs. Kafka Client Quota Kafka Cluster - Principals List<string>
- The list of principals (i.e., service accounts or identity pools) to apply the Kafka Client Quota to. Use the special name,
"<default>"
, to represent the default quota for all users and service accounts. - Throughput
Pulumi.
Confluent Cloud. Inputs. Kafka Client Quota Throughput - Block for representing a Kafka Quota.
- Description string
- The description of the Kafka Client Quota.
- Display
Name string - The name of the Kafka Client Quota.
- Environment
Kafka
Client Quota Environment Args - Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- Kafka
Cluster KafkaClient Quota Kafka Cluster Args - Principals []string
- The list of principals (i.e., service accounts or identity pools) to apply the Kafka Client Quota to. Use the special name,
"<default>"
, to represent the default quota for all users and service accounts. - Throughput
Kafka
Client Quota Throughput Args - Block for representing a Kafka Quota.
- description String
- The description of the Kafka Client Quota.
- display
Name String - The name of the Kafka Client Quota.
- environment
Kafka
Client Quota Environment - Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- kafka
Cluster KafkaClient Quota Kafka Cluster - principals List<String>
- The list of principals (i.e., service accounts or identity pools) to apply the Kafka Client Quota to. Use the special name,
"<default>"
, to represent the default quota for all users and service accounts. - throughput
Kafka
Client Quota Throughput - Block for representing a Kafka Quota.
- description string
- The description of the Kafka Client Quota.
- display
Name string - The name of the Kafka Client Quota.
- environment
Kafka
Client Quota Environment - Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- kafka
Cluster KafkaClient Quota Kafka Cluster - principals string[]
- The list of principals (i.e., service accounts or identity pools) to apply the Kafka Client Quota to. Use the special name,
"<default>"
, to represent the default quota for all users and service accounts. - throughput
Kafka
Client Quota Throughput - Block for representing a Kafka Quota.
- description str
- The description of the Kafka Client Quota.
- display_
name str - The name of the Kafka Client Quota.
- environment
Kafka
Client Quota Environment Args - Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- kafka_
cluster KafkaClient Quota Kafka Cluster Args - principals Sequence[str]
- The list of principals (i.e., service accounts or identity pools) to apply the Kafka Client Quota to. Use the special name,
"<default>"
, to represent the default quota for all users and service accounts. - throughput
Kafka
Client Quota Throughput Args - Block for representing a Kafka Quota.
- description String
- The description of the Kafka Client Quota.
- display
Name String - The name of the Kafka Client Quota.
- environment Property Map
- Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- kafka
Cluster Property Map - principals List<String>
- The list of principals (i.e., service accounts or identity pools) to apply the Kafka Client Quota to. Use the special name,
"<default>"
, to represent the default quota for all users and service accounts. - throughput Property Map
- Block for representing a Kafka Quota.
Supporting Types
KafkaClientQuotaEnvironment, KafkaClientQuotaEnvironmentArgs
- Id string
The ID of the Environment that the corresponding Kafka Cluster belongs to, for example,
env-abc123
.Note: Each principal assigned to a quota receives the full amount of the quota, meaning the quota is not shared by the principals it is assigned. For example, if a 10 MBps ingress quota is applied to Principals 1 and 2, Principal 1 can produce at most 10 MBps, independently of Principal 2.
Note: Define a throughput maximum, but do not guarantee a throughput floor. Applications are rate-limited through the use of the Kafka throttling mechanism. Kafka asks the client to wait before sending more data and mutes the channel, which appears as latency to the client application.
- Id string
The ID of the Environment that the corresponding Kafka Cluster belongs to, for example,
env-abc123
.Note: Each principal assigned to a quota receives the full amount of the quota, meaning the quota is not shared by the principals it is assigned. For example, if a 10 MBps ingress quota is applied to Principals 1 and 2, Principal 1 can produce at most 10 MBps, independently of Principal 2.
Note: Define a throughput maximum, but do not guarantee a throughput floor. Applications are rate-limited through the use of the Kafka throttling mechanism. Kafka asks the client to wait before sending more data and mutes the channel, which appears as latency to the client application.
- id String
The ID of the Environment that the corresponding Kafka Cluster belongs to, for example,
env-abc123
.Note: Each principal assigned to a quota receives the full amount of the quota, meaning the quota is not shared by the principals it is assigned. For example, if a 10 MBps ingress quota is applied to Principals 1 and 2, Principal 1 can produce at most 10 MBps, independently of Principal 2.
Note: Define a throughput maximum, but do not guarantee a throughput floor. Applications are rate-limited through the use of the Kafka throttling mechanism. Kafka asks the client to wait before sending more data and mutes the channel, which appears as latency to the client application.
- id string
The ID of the Environment that the corresponding Kafka Cluster belongs to, for example,
env-abc123
.Note: Each principal assigned to a quota receives the full amount of the quota, meaning the quota is not shared by the principals it is assigned. For example, if a 10 MBps ingress quota is applied to Principals 1 and 2, Principal 1 can produce at most 10 MBps, independently of Principal 2.
Note: Define a throughput maximum, but do not guarantee a throughput floor. Applications are rate-limited through the use of the Kafka throttling mechanism. Kafka asks the client to wait before sending more data and mutes the channel, which appears as latency to the client application.
- id str
The ID of the Environment that the corresponding Kafka Cluster belongs to, for example,
env-abc123
.Note: Each principal assigned to a quota receives the full amount of the quota, meaning the quota is not shared by the principals it is assigned. For example, if a 10 MBps ingress quota is applied to Principals 1 and 2, Principal 1 can produce at most 10 MBps, independently of Principal 2.
Note: Define a throughput maximum, but do not guarantee a throughput floor. Applications are rate-limited through the use of the Kafka throttling mechanism. Kafka asks the client to wait before sending more data and mutes the channel, which appears as latency to the client application.
- id String
The ID of the Environment that the corresponding Kafka Cluster belongs to, for example,
env-abc123
.Note: Each principal assigned to a quota receives the full amount of the quota, meaning the quota is not shared by the principals it is assigned. For example, if a 10 MBps ingress quota is applied to Principals 1 and 2, Principal 1 can produce at most 10 MBps, independently of Principal 2.
Note: Define a throughput maximum, but do not guarantee a throughput floor. Applications are rate-limited through the use of the Kafka throttling mechanism. Kafka asks the client to wait before sending more data and mutes the channel, which appears as latency to the client application.
KafkaClientQuotaKafkaCluster, KafkaClientQuotaKafkaClusterArgs
- Id string
- The ID of the Kafka Cluster where the Kafka Client Quota is applied, for example,
lkc-abc123
.
- Id string
- The ID of the Kafka Cluster where the Kafka Client Quota is applied, for example,
lkc-abc123
.
- id String
- The ID of the Kafka Cluster where the Kafka Client Quota is applied, for example,
lkc-abc123
.
- id string
- The ID of the Kafka Cluster where the Kafka Client Quota is applied, for example,
lkc-abc123
.
- id str
- The ID of the Kafka Cluster where the Kafka Client Quota is applied, for example,
lkc-abc123
.
- id String
- The ID of the Kafka Cluster where the Kafka Client Quota is applied, for example,
lkc-abc123
.
KafkaClientQuotaThroughput, KafkaClientQuotaThroughputArgs
- Egress
Byte stringRate - The egress throughput limit in bytes per second.
- Ingress
Byte stringRate - The ingress throughput limit in bytes per second.
- Egress
Byte stringRate - The egress throughput limit in bytes per second.
- Ingress
Byte stringRate - The ingress throughput limit in bytes per second.
- egress
Byte StringRate - The egress throughput limit in bytes per second.
- ingress
Byte StringRate - The ingress throughput limit in bytes per second.
- egress
Byte stringRate - The egress throughput limit in bytes per second.
- ingress
Byte stringRate - The ingress throughput limit in bytes per second.
- egress_
byte_ strrate - The egress throughput limit in bytes per second.
- ingress_
byte_ strrate - The ingress throughput limit in bytes per second.
- egress
Byte StringRate - The egress throughput limit in bytes per second.
- ingress
Byte StringRate - The ingress throughput limit in bytes per second.
Import
You can import a Kafka Client Quota by using Kafka Client Quota ID. The following example shows how to import a Kafka Client Quota ID:
$ export CONFLUENT_CLOUD_API_KEY="<cloud_api_key>"
$ export CONFLUENT_CLOUD_API_SECRET="<cloud_api_secret>"
$ pulumi import confluentcloud:index/kafkaClientQuota:KafkaClientQuota example cq-abc123
!> Warning: Do not forget to delete terminal command history afterwards for security purposes.
To learn more about importing existing cloud resources, see Importing resources.
Package Details
- Repository
- Confluent Cloud pulumi/pulumi-confluentcloud
- License
- Apache-2.0
- Notes
- This Pulumi package is based on the
confluent
Terraform Provider.