confluentcloud.TableflowTopic
Explore with Pulumi AI
- Example Usage
- Option #1: Manage multiple Tableflow Topics in the same Pulumi Stack
- Option #2: Manage a single Tableflow Topic in the same Pulumi Stack
- Getting Started
- Create TableflowTopic Resource
- Constructor syntax
- Constructor example
- TableflowTopic Resource Properties
- Inputs
- Outputs
- Look up Existing TableflowTopic Resource
- Supporting Types
- Import
- Package Details
Example Usage
Option #1: Manage multiple Tableflow Topics in the same Pulumi Stack
import * as pulumi from "@pulumi/pulumi";
import * as confluentcloud from "@pulumi/confluentcloud";
const example = new confluentcloud.TableflowTopic("example", {
managedStorages: [{}],
environment: {
id: staging.id,
},
kafkaCluster: {
id: stagingConfluentKafkaCluster.id,
},
displayName: orders.topicName,
tableFormats: [
"ICEBERG",
"DELTA",
],
credentials: {
key: env_admin_tableflow_api_key.id,
secret: env_admin_tableflow_api_key.secret,
},
});
import pulumi
import pulumi_confluentcloud as confluentcloud
example = confluentcloud.TableflowTopic("example",
managed_storages=[{}],
environment={
"id": staging["id"],
},
kafka_cluster={
"id": staging_confluent_kafka_cluster["id"],
},
display_name=orders["topicName"],
table_formats=[
"ICEBERG",
"DELTA",
],
credentials={
"key": env_admin_tableflow_api_key["id"],
"secret": env_admin_tableflow_api_key["secret"],
})
package main
import (
"github.com/pulumi/pulumi-confluentcloud/sdk/v2/go/confluentcloud"
"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
)
func main() {
pulumi.Run(func(ctx *pulumi.Context) error {
_, err := confluentcloud.NewTableflowTopic(ctx, "example", &confluentcloud.TableflowTopicArgs{
ManagedStorages: confluentcloud.TableflowTopicManagedStorageArray{
&confluentcloud.TableflowTopicManagedStorageArgs{},
},
Environment: &confluentcloud.TableflowTopicEnvironmentArgs{
Id: pulumi.Any(staging.Id),
},
KafkaCluster: &confluentcloud.TableflowTopicKafkaClusterArgs{
Id: pulumi.Any(stagingConfluentKafkaCluster.Id),
},
DisplayName: pulumi.Any(orders.TopicName),
TableFormats: pulumi.StringArray{
pulumi.String("ICEBERG"),
pulumi.String("DELTA"),
},
Credentials: &confluentcloud.TableflowTopicCredentialsArgs{
Key: pulumi.Any(env_admin_tableflow_api_key.Id),
Secret: pulumi.Any(env_admin_tableflow_api_key.Secret),
},
})
if err != nil {
return err
}
return nil
})
}
using System.Collections.Generic;
using System.Linq;
using Pulumi;
using ConfluentCloud = Pulumi.ConfluentCloud;
return await Deployment.RunAsync(() =>
{
var example = new ConfluentCloud.TableflowTopic("example", new()
{
ManagedStorages = new[]
{
null,
},
Environment = new ConfluentCloud.Inputs.TableflowTopicEnvironmentArgs
{
Id = staging.Id,
},
KafkaCluster = new ConfluentCloud.Inputs.TableflowTopicKafkaClusterArgs
{
Id = stagingConfluentKafkaCluster.Id,
},
DisplayName = orders.TopicName,
TableFormats = new[]
{
"ICEBERG",
"DELTA",
},
Credentials = new ConfluentCloud.Inputs.TableflowTopicCredentialsArgs
{
Key = env_admin_tableflow_api_key.Id,
Secret = env_admin_tableflow_api_key.Secret,
},
});
});
package generated_program;
import com.pulumi.Context;
import com.pulumi.Pulumi;
import com.pulumi.core.Output;
import com.pulumi.confluentcloud.TableflowTopic;
import com.pulumi.confluentcloud.TableflowTopicArgs;
import com.pulumi.confluentcloud.inputs.TableflowTopicManagedStorageArgs;
import com.pulumi.confluentcloud.inputs.TableflowTopicEnvironmentArgs;
import com.pulumi.confluentcloud.inputs.TableflowTopicKafkaClusterArgs;
import com.pulumi.confluentcloud.inputs.TableflowTopicCredentialsArgs;
import java.util.List;
import java.util.ArrayList;
import java.util.Map;
import java.io.File;
import java.nio.file.Files;
import java.nio.file.Paths;
public class App {
public static void main(String[] args) {
Pulumi.run(App::stack);
}
public static void stack(Context ctx) {
var example = new TableflowTopic("example", TableflowTopicArgs.builder()
.managedStorages()
.environment(TableflowTopicEnvironmentArgs.builder()
.id(staging.id())
.build())
.kafkaCluster(TableflowTopicKafkaClusterArgs.builder()
.id(stagingConfluentKafkaCluster.id())
.build())
.displayName(orders.topicName())
.tableFormats(
"ICEBERG",
"DELTA")
.credentials(TableflowTopicCredentialsArgs.builder()
.key(env_admin_tableflow_api_key.id())
.secret(env_admin_tableflow_api_key.secret())
.build())
.build());
}
}
resources:
example:
type: confluentcloud:TableflowTopic
properties:
managedStorages:
- {}
environment:
id: ${staging.id}
kafkaCluster:
id: ${stagingConfluentKafkaCluster.id}
displayName: ${orders.topicName}
tableFormats:
- ICEBERG
- DELTA
credentials:
key: ${["env-admin-tableflow-api-key"].id}
secret: ${["env-admin-tableflow-api-key"].secret}
Option #2: Manage a single Tableflow Topic in the same Pulumi Stack
import * as pulumi from "@pulumi/pulumi";
import * as confluentcloud from "@pulumi/confluentcloud";
const example = new confluentcloud.TableflowTopic("example", {
environment: {
id: staging.id,
},
kafkaCluster: {
id: stagingConfluentKafkaCluster.id,
},
displayName: orders.topicName,
byobAws: {
bucketName: "bucket_1",
providerIntegrationId: main.id,
},
});
import pulumi
import pulumi_confluentcloud as confluentcloud
example = confluentcloud.TableflowTopic("example",
environment={
"id": staging["id"],
},
kafka_cluster={
"id": staging_confluent_kafka_cluster["id"],
},
display_name=orders["topicName"],
byob_aws={
"bucket_name": "bucket_1",
"provider_integration_id": main["id"],
})
package main
import (
"github.com/pulumi/pulumi-confluentcloud/sdk/v2/go/confluentcloud"
"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
)
func main() {
pulumi.Run(func(ctx *pulumi.Context) error {
_, err := confluentcloud.NewTableflowTopic(ctx, "example", &confluentcloud.TableflowTopicArgs{
Environment: &confluentcloud.TableflowTopicEnvironmentArgs{
Id: pulumi.Any(staging.Id),
},
KafkaCluster: &confluentcloud.TableflowTopicKafkaClusterArgs{
Id: pulumi.Any(stagingConfluentKafkaCluster.Id),
},
DisplayName: pulumi.Any(orders.TopicName),
ByobAws: &confluentcloud.TableflowTopicByobAwsArgs{
BucketName: pulumi.String("bucket_1"),
ProviderIntegrationId: pulumi.Any(main.Id),
},
})
if err != nil {
return err
}
return nil
})
}
using System.Collections.Generic;
using System.Linq;
using Pulumi;
using ConfluentCloud = Pulumi.ConfluentCloud;
return await Deployment.RunAsync(() =>
{
var example = new ConfluentCloud.TableflowTopic("example", new()
{
Environment = new ConfluentCloud.Inputs.TableflowTopicEnvironmentArgs
{
Id = staging.Id,
},
KafkaCluster = new ConfluentCloud.Inputs.TableflowTopicKafkaClusterArgs
{
Id = stagingConfluentKafkaCluster.Id,
},
DisplayName = orders.TopicName,
ByobAws = new ConfluentCloud.Inputs.TableflowTopicByobAwsArgs
{
BucketName = "bucket_1",
ProviderIntegrationId = main.Id,
},
});
});
package generated_program;
import com.pulumi.Context;
import com.pulumi.Pulumi;
import com.pulumi.core.Output;
import com.pulumi.confluentcloud.TableflowTopic;
import com.pulumi.confluentcloud.TableflowTopicArgs;
import com.pulumi.confluentcloud.inputs.TableflowTopicEnvironmentArgs;
import com.pulumi.confluentcloud.inputs.TableflowTopicKafkaClusterArgs;
import com.pulumi.confluentcloud.inputs.TableflowTopicByobAwsArgs;
import java.util.List;
import java.util.ArrayList;
import java.util.Map;
import java.io.File;
import java.nio.file.Files;
import java.nio.file.Paths;
public class App {
public static void main(String[] args) {
Pulumi.run(App::stack);
}
public static void stack(Context ctx) {
var example = new TableflowTopic("example", TableflowTopicArgs.builder()
.environment(TableflowTopicEnvironmentArgs.builder()
.id(staging.id())
.build())
.kafkaCluster(TableflowTopicKafkaClusterArgs.builder()
.id(stagingConfluentKafkaCluster.id())
.build())
.displayName(orders.topicName())
.byobAws(TableflowTopicByobAwsArgs.builder()
.bucketName("bucket_1")
.providerIntegrationId(main.id())
.build())
.build());
}
}
resources:
example:
type: confluentcloud:TableflowTopic
properties:
environment:
id: ${staging.id}
kafkaCluster:
id: ${stagingConfluentKafkaCluster.id}
displayName: ${orders.topicName}
byobAws:
bucketName: bucket_1
providerIntegrationId: ${main.id}
Getting Started
The following end-to-end examples might help to get started with confluentcloud.TableflowTopic
resource:
confluent-managed-storage
: Tableflow topic with Confluent-managed storage.
Create TableflowTopic Resource
Resources are created with functions called constructors. To learn more about declaring and configuring resources, see Resources.
Constructor syntax
new TableflowTopic(name: string, args: TableflowTopicArgs, opts?: CustomResourceOptions);
@overload
def TableflowTopic(resource_name: str,
args: TableflowTopicArgs,
opts: Optional[ResourceOptions] = None)
@overload
def TableflowTopic(resource_name: str,
opts: Optional[ResourceOptions] = None,
display_name: Optional[str] = None,
environment: Optional[TableflowTopicEnvironmentArgs] = None,
kafka_cluster: Optional[TableflowTopicKafkaClusterArgs] = None,
byob_aws: Optional[TableflowTopicByobAwsArgs] = None,
credentials: Optional[TableflowTopicCredentialsArgs] = None,
managed_storages: Optional[Sequence[TableflowTopicManagedStorageArgs]] = None,
record_failure_strategy: Optional[str] = None,
retention_ms: Optional[str] = None,
table_formats: Optional[Sequence[str]] = None)
func NewTableflowTopic(ctx *Context, name string, args TableflowTopicArgs, opts ...ResourceOption) (*TableflowTopic, error)
public TableflowTopic(string name, TableflowTopicArgs args, CustomResourceOptions? opts = null)
public TableflowTopic(String name, TableflowTopicArgs args)
public TableflowTopic(String name, TableflowTopicArgs args, CustomResourceOptions options)
type: confluentcloud:TableflowTopic
properties: # The arguments to resource properties.
options: # Bag of options to control resource's behavior.
Parameters
- name
This property is required. string - The unique name of the resource.
- args
This property is required. TableflowTopicArgs - The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- resource_name
This property is required. str - The unique name of the resource.
- args
This property is required. TableflowTopicArgs - The arguments to resource properties.
- opts ResourceOptions
- Bag of options to control resource's behavior.
- ctx Context
- Context object for the current deployment.
- name
This property is required. string - The unique name of the resource.
- args
This property is required. TableflowTopicArgs - The arguments to resource properties.
- opts ResourceOption
- Bag of options to control resource's behavior.
- name
This property is required. string - The unique name of the resource.
- args
This property is required. TableflowTopicArgs - The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- name
This property is required. String - The unique name of the resource.
- args
This property is required. TableflowTopicArgs - The arguments to resource properties.
- options CustomResourceOptions
- Bag of options to control resource's behavior.
Constructor example
The following reference example uses placeholder values for all input properties.
var tableflowTopicResource = new ConfluentCloud.TableflowTopic("tableflowTopicResource", new()
{
DisplayName = "string",
Environment = new ConfluentCloud.Inputs.TableflowTopicEnvironmentArgs
{
Id = "string",
},
KafkaCluster = new ConfluentCloud.Inputs.TableflowTopicKafkaClusterArgs
{
Id = "string",
},
ByobAws = new ConfluentCloud.Inputs.TableflowTopicByobAwsArgs
{
BucketName = "string",
ProviderIntegrationId = "string",
BucketRegion = "string",
},
Credentials = new ConfluentCloud.Inputs.TableflowTopicCredentialsArgs
{
Key = "string",
Secret = "string",
},
ManagedStorages = new[]
{
null,
},
RecordFailureStrategy = "string",
RetentionMs = "string",
TableFormats = new[]
{
"string",
},
});
example, err := confluentcloud.NewTableflowTopic(ctx, "tableflowTopicResource", &confluentcloud.TableflowTopicArgs{
DisplayName: pulumi.String("string"),
Environment: &confluentcloud.TableflowTopicEnvironmentArgs{
Id: pulumi.String("string"),
},
KafkaCluster: &confluentcloud.TableflowTopicKafkaClusterArgs{
Id: pulumi.String("string"),
},
ByobAws: &confluentcloud.TableflowTopicByobAwsArgs{
BucketName: pulumi.String("string"),
ProviderIntegrationId: pulumi.String("string"),
BucketRegion: pulumi.String("string"),
},
Credentials: &confluentcloud.TableflowTopicCredentialsArgs{
Key: pulumi.String("string"),
Secret: pulumi.String("string"),
},
ManagedStorages: confluentcloud.TableflowTopicManagedStorageArray{
&confluentcloud.TableflowTopicManagedStorageArgs{},
},
RecordFailureStrategy: pulumi.String("string"),
RetentionMs: pulumi.String("string"),
TableFormats: pulumi.StringArray{
pulumi.String("string"),
},
})
var tableflowTopicResource = new TableflowTopic("tableflowTopicResource", TableflowTopicArgs.builder()
.displayName("string")
.environment(TableflowTopicEnvironmentArgs.builder()
.id("string")
.build())
.kafkaCluster(TableflowTopicKafkaClusterArgs.builder()
.id("string")
.build())
.byobAws(TableflowTopicByobAwsArgs.builder()
.bucketName("string")
.providerIntegrationId("string")
.bucketRegion("string")
.build())
.credentials(TableflowTopicCredentialsArgs.builder()
.key("string")
.secret("string")
.build())
.managedStorages()
.recordFailureStrategy("string")
.retentionMs("string")
.tableFormats("string")
.build());
tableflow_topic_resource = confluentcloud.TableflowTopic("tableflowTopicResource",
display_name="string",
environment={
"id": "string",
},
kafka_cluster={
"id": "string",
},
byob_aws={
"bucket_name": "string",
"provider_integration_id": "string",
"bucket_region": "string",
},
credentials={
"key": "string",
"secret": "string",
},
managed_storages=[{}],
record_failure_strategy="string",
retention_ms="string",
table_formats=["string"])
const tableflowTopicResource = new confluentcloud.TableflowTopic("tableflowTopicResource", {
displayName: "string",
environment: {
id: "string",
},
kafkaCluster: {
id: "string",
},
byobAws: {
bucketName: "string",
providerIntegrationId: "string",
bucketRegion: "string",
},
credentials: {
key: "string",
secret: "string",
},
managedStorages: [{}],
recordFailureStrategy: "string",
retentionMs: "string",
tableFormats: ["string"],
});
type: confluentcloud:TableflowTopic
properties:
byobAws:
bucketName: string
bucketRegion: string
providerIntegrationId: string
credentials:
key: string
secret: string
displayName: string
environment:
id: string
kafkaCluster:
id: string
managedStorages:
- {}
recordFailureStrategy: string
retentionMs: string
tableFormats:
- string
TableflowTopic Resource Properties
To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.
Inputs
In Python, inputs that are objects can be passed either as argument classes or as dictionary literals.
The TableflowTopic resource accepts the following input properties:
- Display
Name This property is required. Changes to this property will trigger replacement.
- The name of the Kafka topic for which Tableflow is enabled.
- Environment
This property is required. Changes to this property will trigger replacement.
Confluent Cloud. Inputs. Tableflow Topic Environment - Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- Kafka
Cluster This property is required. Changes to this property will trigger replacement.
Confluent Cloud. Inputs. Tableflow Topic Kafka Cluster - Byob
Aws Changes to this property will trigger replacement.
Confluent Cloud. Inputs. Tableflow Topic Byob Aws - (Optional Configuration Block) supports the following:
- Credentials
Pulumi.
Confluent Cloud. Inputs. Tableflow Topic Credentials - The Cluster API Credentials.
- Managed
Storages List<Pulumi.Confluent Cloud. Inputs. Tableflow Topic Managed Storage> - The configuration of the Confluent managed bucket.
- Record
Failure stringStrategy - The strategy to handle record failures in the Tableflow enabled topic during materialization. Accepted values are
SKIP
,SUSPEND
. ForSKIP
, we skip the bad records and move to the next record. ForSUSPEND
, we suspend the materialization of the topic. - Retention
Ms string - The max age of snapshots (Iceberg) or versions (Delta) (snapshot/version expiration) to keep on the table in milliseconds for the Tableflow enabled topic.
- Table
Formats List<string> - The supported table formats for the Tableflow-enabled topic. Accepted values are
DELTA
,ICEBERG
.
- Display
Name This property is required. Changes to this property will trigger replacement.
- The name of the Kafka topic for which Tableflow is enabled.
- Environment
This property is required. Changes to this property will trigger replacement.
Topic Environment Args - Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- Kafka
Cluster This property is required. Changes to this property will trigger replacement.
Topic Kafka Cluster Args - Byob
Aws Changes to this property will trigger replacement.
Topic Byob Aws Args - (Optional Configuration Block) supports the following:
- Credentials
Tableflow
Topic Credentials Args - The Cluster API Credentials.
- Managed
Storages []TableflowTopic Managed Storage Args - The configuration of the Confluent managed bucket.
- Record
Failure stringStrategy - The strategy to handle record failures in the Tableflow enabled topic during materialization. Accepted values are
SKIP
,SUSPEND
. ForSKIP
, we skip the bad records and move to the next record. ForSUSPEND
, we suspend the materialization of the topic. - Retention
Ms string - The max age of snapshots (Iceberg) or versions (Delta) (snapshot/version expiration) to keep on the table in milliseconds for the Tableflow enabled topic.
- Table
Formats []string - The supported table formats for the Tableflow-enabled topic. Accepted values are
DELTA
,ICEBERG
.
- display
Name This property is required. Changes to this property will trigger replacement.
- The name of the Kafka topic for which Tableflow is enabled.
- environment
This property is required. Changes to this property will trigger replacement.
Topic Environment - Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- kafka
Cluster This property is required. Changes to this property will trigger replacement.
Topic Kafka Cluster - byob
Aws Changes to this property will trigger replacement.
Topic Byob Aws - (Optional Configuration Block) supports the following:
- credentials
Tableflow
Topic Credentials - The Cluster API Credentials.
- managed
Storages List<TableflowTopic Managed Storage> - The configuration of the Confluent managed bucket.
- record
Failure StringStrategy - The strategy to handle record failures in the Tableflow enabled topic during materialization. Accepted values are
SKIP
,SUSPEND
. ForSKIP
, we skip the bad records and move to the next record. ForSUSPEND
, we suspend the materialization of the topic. - retention
Ms String - The max age of snapshots (Iceberg) or versions (Delta) (snapshot/version expiration) to keep on the table in milliseconds for the Tableflow enabled topic.
- table
Formats List<String> - The supported table formats for the Tableflow-enabled topic. Accepted values are
DELTA
,ICEBERG
.
- display
Name This property is required. Changes to this property will trigger replacement.
- The name of the Kafka topic for which Tableflow is enabled.
- environment
This property is required. Changes to this property will trigger replacement.
Topic Environment - Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- kafka
Cluster This property is required. Changes to this property will trigger replacement.
Topic Kafka Cluster - byob
Aws Changes to this property will trigger replacement.
Topic Byob Aws - (Optional Configuration Block) supports the following:
- credentials
Tableflow
Topic Credentials - The Cluster API Credentials.
- managed
Storages TableflowTopic Managed Storage[] - The configuration of the Confluent managed bucket.
- record
Failure stringStrategy - The strategy to handle record failures in the Tableflow enabled topic during materialization. Accepted values are
SKIP
,SUSPEND
. ForSKIP
, we skip the bad records and move to the next record. ForSUSPEND
, we suspend the materialization of the topic. - retention
Ms string - The max age of snapshots (Iceberg) or versions (Delta) (snapshot/version expiration) to keep on the table in milliseconds for the Tableflow enabled topic.
- table
Formats string[] - The supported table formats for the Tableflow-enabled topic. Accepted values are
DELTA
,ICEBERG
.
- display_
name This property is required. Changes to this property will trigger replacement.
- The name of the Kafka topic for which Tableflow is enabled.
- environment
This property is required. Changes to this property will trigger replacement.
Topic Environment Args - Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- kafka_
cluster This property is required. Changes to this property will trigger replacement.
Topic Kafka Cluster Args - byob_
aws Changes to this property will trigger replacement.
Topic Byob Aws Args - (Optional Configuration Block) supports the following:
- credentials
Tableflow
Topic Credentials Args - The Cluster API Credentials.
- managed_
storages Sequence[TableflowTopic Managed Storage Args] - The configuration of the Confluent managed bucket.
- record_
failure_ strstrategy - The strategy to handle record failures in the Tableflow enabled topic during materialization. Accepted values are
SKIP
,SUSPEND
. ForSKIP
, we skip the bad records and move to the next record. ForSUSPEND
, we suspend the materialization of the topic. - retention_
ms str - The max age of snapshots (Iceberg) or versions (Delta) (snapshot/version expiration) to keep on the table in milliseconds for the Tableflow enabled topic.
- table_
formats Sequence[str] - The supported table formats for the Tableflow-enabled topic. Accepted values are
DELTA
,ICEBERG
.
- display
Name This property is required. Changes to this property will trigger replacement.
- The name of the Kafka topic for which Tableflow is enabled.
- environment
This property is required. Changes to this property will trigger replacement.
- Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- kafka
Cluster This property is required. Changes to this property will trigger replacement.
- byob
Aws Changes to this property will trigger replacement.
- (Optional Configuration Block) supports the following:
- credentials Property Map
- The Cluster API Credentials.
- managed
Storages List<Property Map> - The configuration of the Confluent managed bucket.
- record
Failure StringStrategy - The strategy to handle record failures in the Tableflow enabled topic during materialization. Accepted values are
SKIP
,SUSPEND
. ForSKIP
, we skip the bad records and move to the next record. ForSUSPEND
, we suspend the materialization of the topic. - retention
Ms String - The max age of snapshots (Iceberg) or versions (Delta) (snapshot/version expiration) to keep on the table in milliseconds for the Tableflow enabled topic.
- table
Formats List<String> - The supported table formats for the Tableflow-enabled topic. Accepted values are
DELTA
,ICEBERG
.
Outputs
All input properties are implicitly available as output properties. Additionally, the TableflowTopic resource produces the following output properties:
- Enable
Compaction bool - (Optional Boolean) This flag determines whether to enable compaction for the Tableflow enabled topic.
- Enable
Partitioning bool - (Optional Boolean) This flag determines whether to enable partitioning for the Tableflow enabled topic.
- Id string
- The provider-assigned unique ID for this managed resource.
- Suspended bool
- (Optional Boolean) Indicates whether the Tableflow should be suspended.
- Enable
Compaction bool - (Optional Boolean) This flag determines whether to enable compaction for the Tableflow enabled topic.
- Enable
Partitioning bool - (Optional Boolean) This flag determines whether to enable partitioning for the Tableflow enabled topic.
- Id string
- The provider-assigned unique ID for this managed resource.
- Suspended bool
- (Optional Boolean) Indicates whether the Tableflow should be suspended.
- enable
Compaction Boolean - (Optional Boolean) This flag determines whether to enable compaction for the Tableflow enabled topic.
- enable
Partitioning Boolean - (Optional Boolean) This flag determines whether to enable partitioning for the Tableflow enabled topic.
- id String
- The provider-assigned unique ID for this managed resource.
- suspended Boolean
- (Optional Boolean) Indicates whether the Tableflow should be suspended.
- enable
Compaction boolean - (Optional Boolean) This flag determines whether to enable compaction for the Tableflow enabled topic.
- enable
Partitioning boolean - (Optional Boolean) This flag determines whether to enable partitioning for the Tableflow enabled topic.
- id string
- The provider-assigned unique ID for this managed resource.
- suspended boolean
- (Optional Boolean) Indicates whether the Tableflow should be suspended.
- enable_
compaction bool - (Optional Boolean) This flag determines whether to enable compaction for the Tableflow enabled topic.
- enable_
partitioning bool - (Optional Boolean) This flag determines whether to enable partitioning for the Tableflow enabled topic.
- id str
- The provider-assigned unique ID for this managed resource.
- suspended bool
- (Optional Boolean) Indicates whether the Tableflow should be suspended.
- enable
Compaction Boolean - (Optional Boolean) This flag determines whether to enable compaction for the Tableflow enabled topic.
- enable
Partitioning Boolean - (Optional Boolean) This flag determines whether to enable partitioning for the Tableflow enabled topic.
- id String
- The provider-assigned unique ID for this managed resource.
- suspended Boolean
- (Optional Boolean) Indicates whether the Tableflow should be suspended.
Look up Existing TableflowTopic Resource
Get an existing TableflowTopic resource’s state with the given name, ID, and optional extra properties used to qualify the lookup.
public static get(name: string, id: Input<ID>, state?: TableflowTopicState, opts?: CustomResourceOptions): TableflowTopic
@staticmethod
def get(resource_name: str,
id: str,
opts: Optional[ResourceOptions] = None,
byob_aws: Optional[TableflowTopicByobAwsArgs] = None,
credentials: Optional[TableflowTopicCredentialsArgs] = None,
display_name: Optional[str] = None,
enable_compaction: Optional[bool] = None,
enable_partitioning: Optional[bool] = None,
environment: Optional[TableflowTopicEnvironmentArgs] = None,
kafka_cluster: Optional[TableflowTopicKafkaClusterArgs] = None,
managed_storages: Optional[Sequence[TableflowTopicManagedStorageArgs]] = None,
record_failure_strategy: Optional[str] = None,
retention_ms: Optional[str] = None,
suspended: Optional[bool] = None,
table_formats: Optional[Sequence[str]] = None) -> TableflowTopic
func GetTableflowTopic(ctx *Context, name string, id IDInput, state *TableflowTopicState, opts ...ResourceOption) (*TableflowTopic, error)
public static TableflowTopic Get(string name, Input<string> id, TableflowTopicState? state, CustomResourceOptions? opts = null)
public static TableflowTopic get(String name, Output<String> id, TableflowTopicState state, CustomResourceOptions options)
resources: _: type: confluentcloud:TableflowTopic get: id: ${id}
- name
This property is required. - The unique name of the resulting resource.
- id
This property is required. - The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- resource_name
This property is required. - The unique name of the resulting resource.
- id
This property is required. - The unique provider ID of the resource to lookup.
- name
This property is required. - The unique name of the resulting resource.
- id
This property is required. - The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
This property is required. - The unique name of the resulting resource.
- id
This property is required. - The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
This property is required. - The unique name of the resulting resource.
- id
This property is required. - The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- Byob
Aws Changes to this property will trigger replacement.
Confluent Cloud. Inputs. Tableflow Topic Byob Aws - (Optional Configuration Block) supports the following:
- Credentials
Pulumi.
Confluent Cloud. Inputs. Tableflow Topic Credentials - The Cluster API Credentials.
- Display
Name Changes to this property will trigger replacement.
- The name of the Kafka topic for which Tableflow is enabled.
- Enable
Compaction bool - (Optional Boolean) This flag determines whether to enable compaction for the Tableflow enabled topic.
- Enable
Partitioning bool - (Optional Boolean) This flag determines whether to enable partitioning for the Tableflow enabled topic.
- Environment
Changes to this property will trigger replacement.
Confluent Cloud. Inputs. Tableflow Topic Environment - Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- Kafka
Cluster Changes to this property will trigger replacement.
Confluent Cloud. Inputs. Tableflow Topic Kafka Cluster - Managed
Storages List<Pulumi.Confluent Cloud. Inputs. Tableflow Topic Managed Storage> - The configuration of the Confluent managed bucket.
- Record
Failure stringStrategy - The strategy to handle record failures in the Tableflow enabled topic during materialization. Accepted values are
SKIP
,SUSPEND
. ForSKIP
, we skip the bad records and move to the next record. ForSUSPEND
, we suspend the materialization of the topic. - Retention
Ms string - The max age of snapshots (Iceberg) or versions (Delta) (snapshot/version expiration) to keep on the table in milliseconds for the Tableflow enabled topic.
- Suspended bool
- (Optional Boolean) Indicates whether the Tableflow should be suspended.
- Table
Formats List<string> - The supported table formats for the Tableflow-enabled topic. Accepted values are
DELTA
,ICEBERG
.
- Byob
Aws Changes to this property will trigger replacement.
Topic Byob Aws Args - (Optional Configuration Block) supports the following:
- Credentials
Tableflow
Topic Credentials Args - The Cluster API Credentials.
- Display
Name Changes to this property will trigger replacement.
- The name of the Kafka topic for which Tableflow is enabled.
- Enable
Compaction bool - (Optional Boolean) This flag determines whether to enable compaction for the Tableflow enabled topic.
- Enable
Partitioning bool - (Optional Boolean) This flag determines whether to enable partitioning for the Tableflow enabled topic.
- Environment
Changes to this property will trigger replacement.
Topic Environment Args - Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- Kafka
Cluster Changes to this property will trigger replacement.
Topic Kafka Cluster Args - Managed
Storages []TableflowTopic Managed Storage Args - The configuration of the Confluent managed bucket.
- Record
Failure stringStrategy - The strategy to handle record failures in the Tableflow enabled topic during materialization. Accepted values are
SKIP
,SUSPEND
. ForSKIP
, we skip the bad records and move to the next record. ForSUSPEND
, we suspend the materialization of the topic. - Retention
Ms string - The max age of snapshots (Iceberg) or versions (Delta) (snapshot/version expiration) to keep on the table in milliseconds for the Tableflow enabled topic.
- Suspended bool
- (Optional Boolean) Indicates whether the Tableflow should be suspended.
- Table
Formats []string - The supported table formats for the Tableflow-enabled topic. Accepted values are
DELTA
,ICEBERG
.
- byob
Aws Changes to this property will trigger replacement.
Topic Byob Aws - (Optional Configuration Block) supports the following:
- credentials
Tableflow
Topic Credentials - The Cluster API Credentials.
- display
Name Changes to this property will trigger replacement.
- The name of the Kafka topic for which Tableflow is enabled.
- enable
Compaction Boolean - (Optional Boolean) This flag determines whether to enable compaction for the Tableflow enabled topic.
- enable
Partitioning Boolean - (Optional Boolean) This flag determines whether to enable partitioning for the Tableflow enabled topic.
- environment
Changes to this property will trigger replacement.
Topic Environment - Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- kafka
Cluster Changes to this property will trigger replacement.
Topic Kafka Cluster - managed
Storages List<TableflowTopic Managed Storage> - The configuration of the Confluent managed bucket.
- record
Failure StringStrategy - The strategy to handle record failures in the Tableflow enabled topic during materialization. Accepted values are
SKIP
,SUSPEND
. ForSKIP
, we skip the bad records and move to the next record. ForSUSPEND
, we suspend the materialization of the topic. - retention
Ms String - The max age of snapshots (Iceberg) or versions (Delta) (snapshot/version expiration) to keep on the table in milliseconds for the Tableflow enabled topic.
- suspended Boolean
- (Optional Boolean) Indicates whether the Tableflow should be suspended.
- table
Formats List<String> - The supported table formats for the Tableflow-enabled topic. Accepted values are
DELTA
,ICEBERG
.
- byob
Aws Changes to this property will trigger replacement.
Topic Byob Aws - (Optional Configuration Block) supports the following:
- credentials
Tableflow
Topic Credentials - The Cluster API Credentials.
- display
Name Changes to this property will trigger replacement.
- The name of the Kafka topic for which Tableflow is enabled.
- enable
Compaction boolean - (Optional Boolean) This flag determines whether to enable compaction for the Tableflow enabled topic.
- enable
Partitioning boolean - (Optional Boolean) This flag determines whether to enable partitioning for the Tableflow enabled topic.
- environment
Changes to this property will trigger replacement.
Topic Environment - Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- kafka
Cluster Changes to this property will trigger replacement.
Topic Kafka Cluster - managed
Storages TableflowTopic Managed Storage[] - The configuration of the Confluent managed bucket.
- record
Failure stringStrategy - The strategy to handle record failures in the Tableflow enabled topic during materialization. Accepted values are
SKIP
,SUSPEND
. ForSKIP
, we skip the bad records and move to the next record. ForSUSPEND
, we suspend the materialization of the topic. - retention
Ms string - The max age of snapshots (Iceberg) or versions (Delta) (snapshot/version expiration) to keep on the table in milliseconds for the Tableflow enabled topic.
- suspended boolean
- (Optional Boolean) Indicates whether the Tableflow should be suspended.
- table
Formats string[] - The supported table formats for the Tableflow-enabled topic. Accepted values are
DELTA
,ICEBERG
.
- byob_
aws Changes to this property will trigger replacement.
Topic Byob Aws Args - (Optional Configuration Block) supports the following:
- credentials
Tableflow
Topic Credentials Args - The Cluster API Credentials.
- display_
name Changes to this property will trigger replacement.
- The name of the Kafka topic for which Tableflow is enabled.
- enable_
compaction bool - (Optional Boolean) This flag determines whether to enable compaction for the Tableflow enabled topic.
- enable_
partitioning bool - (Optional Boolean) This flag determines whether to enable partitioning for the Tableflow enabled topic.
- environment
Changes to this property will trigger replacement.
Topic Environment Args - Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- kafka_
cluster Changes to this property will trigger replacement.
Topic Kafka Cluster Args - managed_
storages Sequence[TableflowTopic Managed Storage Args] - The configuration of the Confluent managed bucket.
- record_
failure_ strstrategy - The strategy to handle record failures in the Tableflow enabled topic during materialization. Accepted values are
SKIP
,SUSPEND
. ForSKIP
, we skip the bad records and move to the next record. ForSUSPEND
, we suspend the materialization of the topic. - retention_
ms str - The max age of snapshots (Iceberg) or versions (Delta) (snapshot/version expiration) to keep on the table in milliseconds for the Tableflow enabled topic.
- suspended bool
- (Optional Boolean) Indicates whether the Tableflow should be suspended.
- table_
formats Sequence[str] - The supported table formats for the Tableflow-enabled topic. Accepted values are
DELTA
,ICEBERG
.
- byob
Aws Changes to this property will trigger replacement.
- (Optional Configuration Block) supports the following:
- credentials Property Map
- The Cluster API Credentials.
- display
Name Changes to this property will trigger replacement.
- The name of the Kafka topic for which Tableflow is enabled.
- enable
Compaction Boolean - (Optional Boolean) This flag determines whether to enable compaction for the Tableflow enabled topic.
- enable
Partitioning Boolean - (Optional Boolean) This flag determines whether to enable partitioning for the Tableflow enabled topic.
- environment
Changes to this property will trigger replacement.
- Environment objects represent an isolated namespace for your Confluent resources for organizational purposes.
- kafka
Cluster Changes to this property will trigger replacement.
- managed
Storages List<Property Map> - The configuration of the Confluent managed bucket.
- record
Failure StringStrategy - The strategy to handle record failures in the Tableflow enabled topic during materialization. Accepted values are
SKIP
,SUSPEND
. ForSKIP
, we skip the bad records and move to the next record. ForSUSPEND
, we suspend the materialization of the topic. - retention
Ms String - The max age of snapshots (Iceberg) or versions (Delta) (snapshot/version expiration) to keep on the table in milliseconds for the Tableflow enabled topic.
- suspended Boolean
- (Optional Boolean) Indicates whether the Tableflow should be suspended.
- table
Formats List<String> - The supported table formats for the Tableflow-enabled topic. Accepted values are
DELTA
,ICEBERG
.
Supporting Types
TableflowTopicByobAws, TableflowTopicByobAwsArgs
- Bucket
Name This property is required. Changes to this property will trigger replacement.
- The bucket name.
- Provider
Integration Id This property is required. Changes to this property will trigger replacement.
- The provider integration id.
- Bucket
Region string - (Required String) The bucket region.
- Bucket
Name This property is required. Changes to this property will trigger replacement.
- The bucket name.
- Provider
Integration Id This property is required. Changes to this property will trigger replacement.
- The provider integration id.
- Bucket
Region string - (Required String) The bucket region.
- bucket
Name This property is required. Changes to this property will trigger replacement.
- The bucket name.
- provider
Integration Id This property is required. Changes to this property will trigger replacement.
- The provider integration id.
- bucket
Region String - (Required String) The bucket region.
- bucket
Name This property is required. Changes to this property will trigger replacement.
- The bucket name.
- provider
Integration Id This property is required. Changes to this property will trigger replacement.
- The provider integration id.
- bucket
Region string - (Required String) The bucket region.
- bucket_
name This property is required. Changes to this property will trigger replacement.
- The bucket name.
- provider_
integration_ id This property is required. Changes to this property will trigger replacement.
- The provider integration id.
- bucket_
region str - (Required String) The bucket region.
- bucket
Name This property is required. Changes to this property will trigger replacement.
- The bucket name.
- provider
Integration Id This property is required. Changes to this property will trigger replacement.
- The provider integration id.
- bucket
Region String - (Required String) The bucket region.
TableflowTopicCredentials, TableflowTopicCredentialsArgs
TableflowTopicEnvironment, TableflowTopicEnvironmentArgs
- Id
This property is required. Changes to this property will trigger replacement.
- The ID of the Environment, for example,
env-abc123
.
- Id
This property is required. Changes to this property will trigger replacement.
- The ID of the Environment, for example,
env-abc123
.
- id
This property is required. Changes to this property will trigger replacement.
- The ID of the Environment, for example,
env-abc123
.
- id
This property is required. Changes to this property will trigger replacement.
- The ID of the Environment, for example,
env-abc123
.
- id
This property is required. Changes to this property will trigger replacement.
- The ID of the Environment, for example,
env-abc123
.
- id
This property is required. Changes to this property will trigger replacement.
- The ID of the Environment, for example,
env-abc123
.
TableflowTopicKafkaCluster, TableflowTopicKafkaClusterArgs
- Id
This property is required. Changes to this property will trigger replacement.
- The ID of the Kafka cluster, for example,
lkc-abc123
.
- Id
This property is required. Changes to this property will trigger replacement.
- The ID of the Kafka cluster, for example,
lkc-abc123
.
- id
This property is required. Changes to this property will trigger replacement.
- The ID of the Kafka cluster, for example,
lkc-abc123
.
- id
This property is required. Changes to this property will trigger replacement.
- The ID of the Kafka cluster, for example,
lkc-abc123
.
- id
This property is required. Changes to this property will trigger replacement.
- The ID of the Kafka cluster, for example,
lkc-abc123
.
- id
This property is required. Changes to this property will trigger replacement.
- The ID of the Kafka cluster, for example,
lkc-abc123
.
Import
You can import a Tableflow Topic by using the Tableflow Topic name, Environment ID, and Kafka Cluster ID, in the format <Environment ID>/<Kafka Cluster ID>/<Tableflow Topic name>
, for example:
Option #1: Manage multiple Tableflow Topics in the same Pulumi Stack
$ export IMPORT_TABLEFLOW_API_KEY="<tableflow_api_key>"
$ export IMPORT_TABLEFLOW_API_SECRET="<tableflow_api_secret>"
$ pulumi import confluentcloud:index/tableflowTopic:TableflowTopic example env-abc123/lkc-abc123/orders
Option #2: Manage a single Tableflow Topic in the same Pulumi Stack
$ pulumi import confluentcloud:index/tableflowTopic:TableflowTopic example env-abc123/lkc-abc123/orders
!> Warning: Do not forget to delete terminal command history afterwards for security purposes.
To learn more about importing existing cloud resources, see Importing resources.
Package Details
- Repository
- Confluent Cloud pulumi/pulumi-confluentcloud
- License
- Apache-2.0
- Notes
- This Pulumi package is based on the
confluent
Terraform Provider.
- Example Usage
- Option #1: Manage multiple Tableflow Topics in the same Pulumi Stack
- Option #2: Manage a single Tableflow Topic in the same Pulumi Stack
- Getting Started
- Create TableflowTopic Resource
- Constructor syntax
- Constructor example
- TableflowTopic Resource Properties
- Inputs
- Outputs
- Look up Existing TableflowTopic Resource
- Supporting Types
- Import
- Package Details