confluentcloud.KafkaTopic
Explore with Pulumi AI
Import
You can import a Kafka topic by using the Kafka cluster ID and Kafka topic name in the format <Kafka cluster ID>/<Kafka topic name>
, for example:
Option #1: Manage multiple Kafka clusters in the same Terraform workspace
$ export IMPORT_KAFKA_API_KEY="<kafka_api_key>"
$ export IMPORT_KAFKA_API_SECRET="<kafka_api_secret>"
$ export IMPORT_KAFKA_REST_ENDPOINT="<kafka_rest_endpoint>"
$ pulumi import confluentcloud:index/kafkaTopic:KafkaTopic my_topic lkc-abc123/orders-123
Option #2: Manage a single Kafka cluster in the same Terraform workspace
$ pulumi import confluentcloud:index/kafkaTopic:KafkaTopic my_topic lkc-abc123/orders-123
resource “confluent_kafka_topic” “orders” {
kafka_cluster {
id = confluent_kafka_cluster.basic-cluster.id
}
topic_name = “orders”
partitions_count = 4
rest_endpoint = confluent_kafka_cluster.basic-cluster.rest_endpoint
https://docs.confluent.io/cloud/current/client-apps/topics/manage.html#ak-topic-configurations-for-all-ccloud-cluster-types
config = {
"cleanup.policy" = "delete"
"delete.retention.ms" = "86400000"
"max.compaction.lag.ms" = "9223372036854775807"
"max.message.bytes" = "2097164"
"message.timestamp.after.max.ms" = "9223372036854775807"
"message.timestamp.before.max.ms" = "9223372036854775807"
"message.timestamp.difference.max.ms" = "9223372036854775807"
"message.timestamp.type" = "CreateTime"
"min.compaction.lag.ms" = "0"
"min.insync.replicas" = "2"
"retention.bytes" = "-1"
"retention.ms" = "604800000"
"segment.bytes" = "104857600"
"segment.ms" = "604800000"
}
credentials {
key = confluent_api_key.app-manager-kafka-api-key.id
secret = confluent_api_key.app-manager-kafka-api-key.secret
}
}
!> Warning: Do not forget to delete terminal command history afterwards for security purposes.
Create KafkaTopic Resource
Resources are created with functions called constructors. To learn more about declaring and configuring resources, see Resources.
Constructor syntax
new KafkaTopic(name: string, args: KafkaTopicArgs, opts?: CustomResourceOptions);
@overload
def KafkaTopic(resource_name: str,
args: KafkaTopicArgs,
opts: Optional[ResourceOptions] = None)
@overload
def KafkaTopic(resource_name: str,
opts: Optional[ResourceOptions] = None,
topic_name: Optional[str] = None,
config: Optional[Mapping[str, str]] = None,
credentials: Optional[KafkaTopicCredentialsArgs] = None,
http_endpoint: Optional[str] = None,
kafka_cluster: Optional[KafkaTopicKafkaClusterArgs] = None,
partitions_count: Optional[int] = None,
rest_endpoint: Optional[str] = None)
func NewKafkaTopic(ctx *Context, name string, args KafkaTopicArgs, opts ...ResourceOption) (*KafkaTopic, error)
public KafkaTopic(string name, KafkaTopicArgs args, CustomResourceOptions? opts = null)
public KafkaTopic(String name, KafkaTopicArgs args)
public KafkaTopic(String name, KafkaTopicArgs args, CustomResourceOptions options)
type: confluentcloud:KafkaTopic
properties: # The arguments to resource properties.
options: # Bag of options to control resource's behavior.
Parameters
- name string
- The unique name of the resource.
- args KafkaTopicArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- resource_name str
- The unique name of the resource.
- args KafkaTopicArgs
- The arguments to resource properties.
- opts ResourceOptions
- Bag of options to control resource's behavior.
- ctx Context
- Context object for the current deployment.
- name string
- The unique name of the resource.
- args KafkaTopicArgs
- The arguments to resource properties.
- opts ResourceOption
- Bag of options to control resource's behavior.
- name string
- The unique name of the resource.
- args KafkaTopicArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- name String
- The unique name of the resource.
- args KafkaTopicArgs
- The arguments to resource properties.
- options CustomResourceOptions
- Bag of options to control resource's behavior.
Constructor example
The following reference example uses placeholder values for all input properties.
var kafkaTopicResource = new ConfluentCloud.KafkaTopic("kafkaTopicResource", new()
{
TopicName = "string",
Config =
{
{ "string", "string" },
},
Credentials = new ConfluentCloud.Inputs.KafkaTopicCredentialsArgs
{
Key = "string",
Secret = "string",
},
KafkaCluster = new ConfluentCloud.Inputs.KafkaTopicKafkaClusterArgs
{
Id = "string",
},
PartitionsCount = 0,
RestEndpoint = "string",
});
example, err := confluentcloud.NewKafkaTopic(ctx, "kafkaTopicResource", &confluentcloud.KafkaTopicArgs{
TopicName: pulumi.String("string"),
Config: pulumi.StringMap{
"string": pulumi.String("string"),
},
Credentials: &confluentcloud.KafkaTopicCredentialsArgs{
Key: pulumi.String("string"),
Secret: pulumi.String("string"),
},
KafkaCluster: &confluentcloud.KafkaTopicKafkaClusterArgs{
Id: pulumi.String("string"),
},
PartitionsCount: pulumi.Int(0),
RestEndpoint: pulumi.String("string"),
})
var kafkaTopicResource = new KafkaTopic("kafkaTopicResource", KafkaTopicArgs.builder()
.topicName("string")
.config(Map.of("string", "string"))
.credentials(KafkaTopicCredentialsArgs.builder()
.key("string")
.secret("string")
.build())
.kafkaCluster(KafkaTopicKafkaClusterArgs.builder()
.id("string")
.build())
.partitionsCount(0)
.restEndpoint("string")
.build());
kafka_topic_resource = confluentcloud.KafkaTopic("kafkaTopicResource",
topic_name="string",
config={
"string": "string",
},
credentials=confluentcloud.KafkaTopicCredentialsArgs(
key="string",
secret="string",
),
kafka_cluster=confluentcloud.KafkaTopicKafkaClusterArgs(
id="string",
),
partitions_count=0,
rest_endpoint="string")
const kafkaTopicResource = new confluentcloud.KafkaTopic("kafkaTopicResource", {
topicName: "string",
config: {
string: "string",
},
credentials: {
key: "string",
secret: "string",
},
kafkaCluster: {
id: "string",
},
partitionsCount: 0,
restEndpoint: "string",
});
type: confluentcloud:KafkaTopic
properties:
config:
string: string
credentials:
key: string
secret: string
kafkaCluster:
id: string
partitionsCount: 0
restEndpoint: string
topicName: string
KafkaTopic Resource Properties
To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.
Inputs
The KafkaTopic resource accepts the following input properties:
- Topic
Name string - The name of the topic, for example,
orders-1
. The topic name can be up to 249 characters in length, and can include the following characters: a-z, A-Z, 0-9, . (dot), _ (underscore), and - (dash). As a best practice, we recommend against using any personally identifiable information (PII) when naming your topic. - Config Dictionary<string, string>
- The custom topic settings to set:
- Credentials
Pulumi.
Confluent Cloud. Inputs. Kafka Topic Credentials - The Cluster API Credentials.
- Http
Endpoint string - Kafka
Cluster Pulumi.Confluent Cloud. Inputs. Kafka Topic Kafka Cluster - Partitions
Count int - The number of partitions to create in the topic. Defaults to
6
. - Rest
Endpoint string - The REST endpoint of the Kafka cluster, for example,
https://pkc-00000.us-central1.gcp.confluent.cloud:443
).
- Topic
Name string - The name of the topic, for example,
orders-1
. The topic name can be up to 249 characters in length, and can include the following characters: a-z, A-Z, 0-9, . (dot), _ (underscore), and - (dash). As a best practice, we recommend against using any personally identifiable information (PII) when naming your topic. - Config map[string]string
- The custom topic settings to set:
- Credentials
Kafka
Topic Credentials Args - The Cluster API Credentials.
- Http
Endpoint string - Kafka
Cluster KafkaTopic Kafka Cluster Args - Partitions
Count int - The number of partitions to create in the topic. Defaults to
6
. - Rest
Endpoint string - The REST endpoint of the Kafka cluster, for example,
https://pkc-00000.us-central1.gcp.confluent.cloud:443
).
- topic
Name String - The name of the topic, for example,
orders-1
. The topic name can be up to 249 characters in length, and can include the following characters: a-z, A-Z, 0-9, . (dot), _ (underscore), and - (dash). As a best practice, we recommend against using any personally identifiable information (PII) when naming your topic. - config Map<String,String>
- The custom topic settings to set:
- credentials
Kafka
Topic Credentials - The Cluster API Credentials.
- http
Endpoint String - kafka
Cluster KafkaTopic Kafka Cluster - partitions
Count Integer - The number of partitions to create in the topic. Defaults to
6
. - rest
Endpoint String - The REST endpoint of the Kafka cluster, for example,
https://pkc-00000.us-central1.gcp.confluent.cloud:443
).
- topic
Name string - The name of the topic, for example,
orders-1
. The topic name can be up to 249 characters in length, and can include the following characters: a-z, A-Z, 0-9, . (dot), _ (underscore), and - (dash). As a best practice, we recommend against using any personally identifiable information (PII) when naming your topic. - config {[key: string]: string}
- The custom topic settings to set:
- credentials
Kafka
Topic Credentials - The Cluster API Credentials.
- http
Endpoint string - kafka
Cluster KafkaTopic Kafka Cluster - partitions
Count number - The number of partitions to create in the topic. Defaults to
6
. - rest
Endpoint string - The REST endpoint of the Kafka cluster, for example,
https://pkc-00000.us-central1.gcp.confluent.cloud:443
).
- topic_
name str - The name of the topic, for example,
orders-1
. The topic name can be up to 249 characters in length, and can include the following characters: a-z, A-Z, 0-9, . (dot), _ (underscore), and - (dash). As a best practice, we recommend against using any personally identifiable information (PII) when naming your topic. - config Mapping[str, str]
- The custom topic settings to set:
- credentials
Kafka
Topic Credentials Args - The Cluster API Credentials.
- http_
endpoint str - kafka_
cluster KafkaTopic Kafka Cluster Args - partitions_
count int - The number of partitions to create in the topic. Defaults to
6
. - rest_
endpoint str - The REST endpoint of the Kafka cluster, for example,
https://pkc-00000.us-central1.gcp.confluent.cloud:443
).
- topic
Name String - The name of the topic, for example,
orders-1
. The topic name can be up to 249 characters in length, and can include the following characters: a-z, A-Z, 0-9, . (dot), _ (underscore), and - (dash). As a best practice, we recommend against using any personally identifiable information (PII) when naming your topic. - config Map<String>
- The custom topic settings to set:
- credentials Property Map
- The Cluster API Credentials.
- http
Endpoint String - kafka
Cluster Property Map - partitions
Count Number - The number of partitions to create in the topic. Defaults to
6
. - rest
Endpoint String - The REST endpoint of the Kafka cluster, for example,
https://pkc-00000.us-central1.gcp.confluent.cloud:443
).
Outputs
All input properties are implicitly available as output properties. Additionally, the KafkaTopic resource produces the following output properties:
- Id string
- The provider-assigned unique ID for this managed resource.
- Id string
- The provider-assigned unique ID for this managed resource.
- id String
- The provider-assigned unique ID for this managed resource.
- id string
- The provider-assigned unique ID for this managed resource.
- id str
- The provider-assigned unique ID for this managed resource.
- id String
- The provider-assigned unique ID for this managed resource.
Look up Existing KafkaTopic Resource
Get an existing KafkaTopic resource’s state with the given name, ID, and optional extra properties used to qualify the lookup.
public static get(name: string, id: Input<ID>, state?: KafkaTopicState, opts?: CustomResourceOptions): KafkaTopic
@staticmethod
def get(resource_name: str,
id: str,
opts: Optional[ResourceOptions] = None,
config: Optional[Mapping[str, str]] = None,
credentials: Optional[KafkaTopicCredentialsArgs] = None,
http_endpoint: Optional[str] = None,
kafka_cluster: Optional[KafkaTopicKafkaClusterArgs] = None,
partitions_count: Optional[int] = None,
rest_endpoint: Optional[str] = None,
topic_name: Optional[str] = None) -> KafkaTopic
func GetKafkaTopic(ctx *Context, name string, id IDInput, state *KafkaTopicState, opts ...ResourceOption) (*KafkaTopic, error)
public static KafkaTopic Get(string name, Input<string> id, KafkaTopicState? state, CustomResourceOptions? opts = null)
public static KafkaTopic get(String name, Output<String> id, KafkaTopicState state, CustomResourceOptions options)
Resource lookup is not supported in YAML
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- resource_name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- Config Dictionary<string, string>
- The custom topic settings to set:
- Credentials
Pulumi.
Confluent Cloud. Inputs. Kafka Topic Credentials - The Cluster API Credentials.
- Http
Endpoint string - Kafka
Cluster Pulumi.Confluent Cloud. Inputs. Kafka Topic Kafka Cluster - Partitions
Count int - The number of partitions to create in the topic. Defaults to
6
. - Rest
Endpoint string - The REST endpoint of the Kafka cluster, for example,
https://pkc-00000.us-central1.gcp.confluent.cloud:443
). - Topic
Name string - The name of the topic, for example,
orders-1
. The topic name can be up to 249 characters in length, and can include the following characters: a-z, A-Z, 0-9, . (dot), _ (underscore), and - (dash). As a best practice, we recommend against using any personally identifiable information (PII) when naming your topic.
- Config map[string]string
- The custom topic settings to set:
- Credentials
Kafka
Topic Credentials Args - The Cluster API Credentials.
- Http
Endpoint string - Kafka
Cluster KafkaTopic Kafka Cluster Args - Partitions
Count int - The number of partitions to create in the topic. Defaults to
6
. - Rest
Endpoint string - The REST endpoint of the Kafka cluster, for example,
https://pkc-00000.us-central1.gcp.confluent.cloud:443
). - Topic
Name string - The name of the topic, for example,
orders-1
. The topic name can be up to 249 characters in length, and can include the following characters: a-z, A-Z, 0-9, . (dot), _ (underscore), and - (dash). As a best practice, we recommend against using any personally identifiable information (PII) when naming your topic.
- config Map<String,String>
- The custom topic settings to set:
- credentials
Kafka
Topic Credentials - The Cluster API Credentials.
- http
Endpoint String - kafka
Cluster KafkaTopic Kafka Cluster - partitions
Count Integer - The number of partitions to create in the topic. Defaults to
6
. - rest
Endpoint String - The REST endpoint of the Kafka cluster, for example,
https://pkc-00000.us-central1.gcp.confluent.cloud:443
). - topic
Name String - The name of the topic, for example,
orders-1
. The topic name can be up to 249 characters in length, and can include the following characters: a-z, A-Z, 0-9, . (dot), _ (underscore), and - (dash). As a best practice, we recommend against using any personally identifiable information (PII) when naming your topic.
- config {[key: string]: string}
- The custom topic settings to set:
- credentials
Kafka
Topic Credentials - The Cluster API Credentials.
- http
Endpoint string - kafka
Cluster KafkaTopic Kafka Cluster - partitions
Count number - The number of partitions to create in the topic. Defaults to
6
. - rest
Endpoint string - The REST endpoint of the Kafka cluster, for example,
https://pkc-00000.us-central1.gcp.confluent.cloud:443
). - topic
Name string - The name of the topic, for example,
orders-1
. The topic name can be up to 249 characters in length, and can include the following characters: a-z, A-Z, 0-9, . (dot), _ (underscore), and - (dash). As a best practice, we recommend against using any personally identifiable information (PII) when naming your topic.
- config Mapping[str, str]
- The custom topic settings to set:
- credentials
Kafka
Topic Credentials Args - The Cluster API Credentials.
- http_
endpoint str - kafka_
cluster KafkaTopic Kafka Cluster Args - partitions_
count int - The number of partitions to create in the topic. Defaults to
6
. - rest_
endpoint str - The REST endpoint of the Kafka cluster, for example,
https://pkc-00000.us-central1.gcp.confluent.cloud:443
). - topic_
name str - The name of the topic, for example,
orders-1
. The topic name can be up to 249 characters in length, and can include the following characters: a-z, A-Z, 0-9, . (dot), _ (underscore), and - (dash). As a best practice, we recommend against using any personally identifiable information (PII) when naming your topic.
- config Map<String>
- The custom topic settings to set:
- credentials Property Map
- The Cluster API Credentials.
- http
Endpoint String - kafka
Cluster Property Map - partitions
Count Number - The number of partitions to create in the topic. Defaults to
6
. - rest
Endpoint String - The REST endpoint of the Kafka cluster, for example,
https://pkc-00000.us-central1.gcp.confluent.cloud:443
). - topic
Name String - The name of the topic, for example,
orders-1
. The topic name can be up to 249 characters in length, and can include the following characters: a-z, A-Z, 0-9, . (dot), _ (underscore), and - (dash). As a best practice, we recommend against using any personally identifiable information (PII) when naming your topic.
Supporting Types
KafkaTopicCredentials, KafkaTopicCredentialsArgs
- Key string
- The Kafka API Key.
- Secret string
The Kafka API Secret.
Note: A Kafka API key consists of a key and a secret. Kafka API keys are required to interact with Kafka clusters in Confluent Cloud. Each Kafka API key is valid for one specific Kafka cluster.
Note: Use Option #2 to simplify the key rotation process. When using Option #1, to rotate a Kafka API key, create a new Kafka API key, update the
credentials
block in all configuration files to use the new Kafka API key, runpulumi up -target="confluent_kafka_topic.orders"
, and remove the old Kafka API key. Alternatively, in case the old Kafka API Key was deleted already, you might need to runpulumi preview -refresh=false -target="confluent_kafka_topic.orders" -out=rotate-kafka-api-key
andpulumi up rotate-kafka-api-key
instead.
- Key string
- The Kafka API Key.
- Secret string
The Kafka API Secret.
Note: A Kafka API key consists of a key and a secret. Kafka API keys are required to interact with Kafka clusters in Confluent Cloud. Each Kafka API key is valid for one specific Kafka cluster.
Note: Use Option #2 to simplify the key rotation process. When using Option #1, to rotate a Kafka API key, create a new Kafka API key, update the
credentials
block in all configuration files to use the new Kafka API key, runpulumi up -target="confluent_kafka_topic.orders"
, and remove the old Kafka API key. Alternatively, in case the old Kafka API Key was deleted already, you might need to runpulumi preview -refresh=false -target="confluent_kafka_topic.orders" -out=rotate-kafka-api-key
andpulumi up rotate-kafka-api-key
instead.
- key String
- The Kafka API Key.
- secret String
The Kafka API Secret.
Note: A Kafka API key consists of a key and a secret. Kafka API keys are required to interact with Kafka clusters in Confluent Cloud. Each Kafka API key is valid for one specific Kafka cluster.
Note: Use Option #2 to simplify the key rotation process. When using Option #1, to rotate a Kafka API key, create a new Kafka API key, update the
credentials
block in all configuration files to use the new Kafka API key, runpulumi up -target="confluent_kafka_topic.orders"
, and remove the old Kafka API key. Alternatively, in case the old Kafka API Key was deleted already, you might need to runpulumi preview -refresh=false -target="confluent_kafka_topic.orders" -out=rotate-kafka-api-key
andpulumi up rotate-kafka-api-key
instead.
- key string
- The Kafka API Key.
- secret string
The Kafka API Secret.
Note: A Kafka API key consists of a key and a secret. Kafka API keys are required to interact with Kafka clusters in Confluent Cloud. Each Kafka API key is valid for one specific Kafka cluster.
Note: Use Option #2 to simplify the key rotation process. When using Option #1, to rotate a Kafka API key, create a new Kafka API key, update the
credentials
block in all configuration files to use the new Kafka API key, runpulumi up -target="confluent_kafka_topic.orders"
, and remove the old Kafka API key. Alternatively, in case the old Kafka API Key was deleted already, you might need to runpulumi preview -refresh=false -target="confluent_kafka_topic.orders" -out=rotate-kafka-api-key
andpulumi up rotate-kafka-api-key
instead.
- key str
- The Kafka API Key.
- secret str
The Kafka API Secret.
Note: A Kafka API key consists of a key and a secret. Kafka API keys are required to interact with Kafka clusters in Confluent Cloud. Each Kafka API key is valid for one specific Kafka cluster.
Note: Use Option #2 to simplify the key rotation process. When using Option #1, to rotate a Kafka API key, create a new Kafka API key, update the
credentials
block in all configuration files to use the new Kafka API key, runpulumi up -target="confluent_kafka_topic.orders"
, and remove the old Kafka API key. Alternatively, in case the old Kafka API Key was deleted already, you might need to runpulumi preview -refresh=false -target="confluent_kafka_topic.orders" -out=rotate-kafka-api-key
andpulumi up rotate-kafka-api-key
instead.
- key String
- The Kafka API Key.
- secret String
The Kafka API Secret.
Note: A Kafka API key consists of a key and a secret. Kafka API keys are required to interact with Kafka clusters in Confluent Cloud. Each Kafka API key is valid for one specific Kafka cluster.
Note: Use Option #2 to simplify the key rotation process. When using Option #1, to rotate a Kafka API key, create a new Kafka API key, update the
credentials
block in all configuration files to use the new Kafka API key, runpulumi up -target="confluent_kafka_topic.orders"
, and remove the old Kafka API key. Alternatively, in case the old Kafka API Key was deleted already, you might need to runpulumi preview -refresh=false -target="confluent_kafka_topic.orders" -out=rotate-kafka-api-key
andpulumi up rotate-kafka-api-key
instead.
KafkaTopicKafkaCluster, KafkaTopicKafkaClusterArgs
- Id string
- The ID of the Kafka cluster, for example,
lkc-abc123
.
- Id string
- The ID of the Kafka cluster, for example,
lkc-abc123
.
- id String
- The ID of the Kafka cluster, for example,
lkc-abc123
.
- id string
- The ID of the Kafka cluster, for example,
lkc-abc123
.
- id str
- The ID of the Kafka cluster, for example,
lkc-abc123
.
- id String
- The ID of the Kafka cluster, for example,
lkc-abc123
.
Package Details
- Repository
- Confluent Cloud pulumi/pulumi-confluentcloud
- License
- Apache-2.0
- Notes
- This Pulumi package is based on the
confluent
Terraform Provider.