1. Packages
  2. Dbtcloud Provider
  3. API Docs
  4. GlobalConnection
dbt Cloud v0.1.20 published on Friday, Sep 27, 2024 by Pulumi

dbtcloud.GlobalConnection

Explore with Pulumi AI

dbtcloud logo
dbt Cloud v0.1.20 published on Friday, Sep 27, 2024 by Pulumi

    This resource can be used to create global connections as introduced in dbt Cloud in August 2024.

    Those connections are not linked to a specific project and can be linked to environments from different projects by using the connection_id field in the dbtcloud.Environment resource.

    All connections types are supported, and the old resources dbtcloud.Connection, dbtcloud.BigQueryConnection and dbtcloud.FabricConnection are now flagged as deprecated and will be removed from the next major version of the provider.

    Example Usage

    Coming soon!
    
    Coming soon!
    
    Coming soon!
    
    Coming soon!
    
    package generated_program;
    
    import com.pulumi.Context;
    import com.pulumi.Pulumi;
    import com.pulumi.core.Output;
    import com.pulumi.dbtcloud.GlobalConnection;
    import com.pulumi.dbtcloud.GlobalConnectionArgs;
    import com.pulumi.dbtcloud.inputs.GlobalConnectionApacheSparkArgs;
    import com.pulumi.dbtcloud.inputs.GlobalConnectionAthenaArgs;
    import com.pulumi.dbtcloud.inputs.GlobalConnectionBigqueryArgs;
    import com.pulumi.dbtcloud.inputs.GlobalConnectionDatabricksArgs;
    import com.pulumi.dbtcloud.inputs.GlobalConnectionFabricArgs;
    import com.pulumi.dbtcloud.inputs.GlobalConnectionPostgresArgs;
    import com.pulumi.dbtcloud.inputs.GlobalConnectionRedshiftArgs;
    import com.pulumi.dbtcloud.inputs.GlobalConnectionSnowflakeArgs;
    import com.pulumi.dbtcloud.inputs.GlobalConnectionStarburstArgs;
    import com.pulumi.dbtcloud.inputs.GlobalConnectionSynapseArgs;
    import java.util.List;
    import java.util.ArrayList;
    import java.util.Map;
    import java.io.File;
    import java.nio.file.Files;
    import java.nio.file.Paths;
    
    public class App {
        public static void main(String[] args) {
            Pulumi.run(App::stack);
        }
    
        public static void stack(Context ctx) {
            var apacheSpark = new GlobalConnection("apacheSpark", GlobalConnectionArgs.builder()
                .name("My Apache Spark connection")
                .apacheSpark(GlobalConnectionApacheSparkArgs.builder()
                    .method("http")
                    .host("my-spark-host.com")
                    .cluster("my-cluster")
                    .connect_timeout(100)
                    .build())
                .build());
    
            var athena = new GlobalConnection("athena", GlobalConnectionArgs.builder()
                .name("My Athena connection")
                .athena(GlobalConnectionAthenaArgs.builder()
                    .region_name("us-east-1")
                    .database("mydatabase")
                    .s3_staging_dir("my_dir")
                    .work_group("my_work_group")
                    .build())
                .build());
    
            var bigquery = new GlobalConnection("bigquery", GlobalConnectionArgs.builder()
                .name("My BigQuery connection")
                .bigquery(GlobalConnectionBigqueryArgs.builder()
                    .gcp_project_id("my-gcp-project-id")
                    .timeout_seconds(1000)
                    .private_key_id("my-private-key-id")
                    .private_key("ABCDEFGHIJKL")
                    .client_email("my_client_email")
                    .client_id("my_client_id")
                    .auth_uri("my_auth_uri")
                    .token_uri("my_token_uri")
                    .auth_provider_x509_cert_url("my_auth_provider_x509_cert_url")
                    .client_x509_cert_url("my_client_x509_cert_url")
                    .application_id("oauth_application_id")
                    .application_secret("oauth_secret_id")
                    .build())
                .build());
    
            var databricks = new GlobalConnection("databricks", GlobalConnectionArgs.builder()
                .name("My Databricks connection")
                .databricks(GlobalConnectionDatabricksArgs.builder()
                    .host("my-databricks-host.cloud.databricks.com")
                    .http_path("/sql/my/http/path")
                    .catalog("dbt_catalog")
                    .client_id("yourclientid")
                    .client_secret("yourclientsecret")
                    .build())
                .build());
    
            var fabric = new GlobalConnection("fabric", GlobalConnectionArgs.builder()
                .name("My Fabric connection")
                .fabric(GlobalConnectionFabricArgs.builder()
                    .server("my-fabric-server.com")
                    .database("mydb")
                    .port(1234)
                    .retries(3)
                    .login_timeout(60)
                    .query_timeout(3600)
                    .build())
                .build());
    
            var postgres = new GlobalConnection("postgres", GlobalConnectionArgs.builder()
                .name("My PostgreSQL connection")
                .postgres(GlobalConnectionPostgresArgs.builder()
                    .hostname("my-postgresql-server.com")
                    .port(5432)
                    .dbname("my_database")
                    .build())
                .build());
    
            var redshift = new GlobalConnection("redshift", GlobalConnectionArgs.builder()
                .name("My Redshift connection")
                .redshift(GlobalConnectionRedshiftArgs.builder()
                    .hostname("my-redshift-connection.com")
                    .port(5432)
                    .dbname("my_database")
                    .build())
                .build());
    
            var snowflake = new GlobalConnection("snowflake", GlobalConnectionArgs.builder()
                .name("My Snowflake connection")
                .privateLinkEndpointId(myPrivateLink.id())
                .snowflake(GlobalConnectionSnowflakeArgs.builder()
                    .account("my-snowflake-account")
                    .database("MY_DATABASE")
                    .warehouse("MY_WAREHOUSE")
                    .client_session_keep_alive(false)
                    .allow_sso(true)
                    .oauth_client_id("yourclientid")
                    .oauth_client_secret("yourclientsecret")
                    .build())
                .build());
    
            var starburst = new GlobalConnection("starburst", GlobalConnectionArgs.builder()
                .name("My Starburst connection")
                .starburst(GlobalConnectionStarburstArgs.builder()
                    .host("my-starburst-host.com")
                    .database("mydb")
                    .build())
                .build());
    
            var synapse = new GlobalConnection("synapse", GlobalConnectionArgs.builder()
                .name("My Synapse connection")
                .synapse(GlobalConnectionSynapseArgs.builder()
                    .host("my-synapse-server.com")
                    .database("mydb")
                    .port(1234)
                    .retries(3)
                    .login_timeout(60)
                    .query_timeout(3600)
                    .build())
                .build());
    
        }
    }
    
    resources:
      apacheSpark:
        type: dbtcloud:GlobalConnection
        name: apache_spark
        properties:
          name: My Apache Spark connection
          apacheSpark:
            method: http
            host: my-spark-host.com
            cluster: my-cluster
            connect_timeout: 100
      athena:
        type: dbtcloud:GlobalConnection
        properties:
          name: My Athena connection
          athena:
            region_name: us-east-1
            database: mydatabase
            s3_staging_dir: my_dir
            work_group: my_work_group
      bigquery:
        type: dbtcloud:GlobalConnection
        properties:
          name: My BigQuery connection
          bigquery:
            gcp_project_id: my-gcp-project-id
            timeout_seconds: 1000
            private_key_id: my-private-key-id
            private_key: ABCDEFGHIJKL
            client_email: my_client_email
            client_id: my_client_id
            auth_uri: my_auth_uri
            token_uri: my_token_uri
            auth_provider_x509_cert_url: my_auth_provider_x509_cert_url
            client_x509_cert_url: my_client_x509_cert_url
            application_id: oauth_application_id
            application_secret: oauth_secret_id
      databricks:
        type: dbtcloud:GlobalConnection
        properties:
          name: My Databricks connection
          databricks:
            host: my-databricks-host.cloud.databricks.com
            http_path: /sql/my/http/path
            catalog: dbt_catalog
            client_id: yourclientid
            client_secret: yourclientsecret
      fabric:
        type: dbtcloud:GlobalConnection
        properties:
          name: My Fabric connection
          fabric:
            server: my-fabric-server.com
            database: mydb
            port: 1234
            retries: 3
            login_timeout: 60
            query_timeout: 3600
      postgres:
        type: dbtcloud:GlobalConnection
        properties:
          name: My PostgreSQL connection
          postgres:
            hostname: my-postgresql-server.com
            port: 5432
            dbname: my_database
      redshift:
        type: dbtcloud:GlobalConnection
        properties:
          name: My Redshift connection
          redshift:
            hostname: my-redshift-connection.com
            port: 5432
            dbname: my_database
      snowflake:
        type: dbtcloud:GlobalConnection
        properties:
          name: My Snowflake connection
          privateLinkEndpointId: ${myPrivateLink.id}
          snowflake:
            account: my-snowflake-account
            database: MY_DATABASE
            warehouse: MY_WAREHOUSE
            client_session_keep_alive: false
            allow_sso: true
            oauth_client_id: yourclientid
            oauth_client_secret: yourclientsecret
      starburst:
        type: dbtcloud:GlobalConnection
        properties:
          name: My Starburst connection
          starburst:
            host: my-starburst-host.com
            database: mydb
      synapse:
        type: dbtcloud:GlobalConnection
        properties:
          name: My Synapse connection
          synapse:
            host: my-synapse-server.com
            database: mydb
            port: 1234
            retries: 3
            login_timeout: 60
            query_timeout: 3600
    

    Create GlobalConnection Resource

    Resources are created with functions called constructors. To learn more about declaring and configuring resources, see Resources.

    Constructor syntax

    new GlobalConnection(name: string, args?: GlobalConnectionArgs, opts?: CustomResourceOptions);
    @overload
    def GlobalConnection(resource_name: str,
                         args: Optional[GlobalConnectionArgs] = None,
                         opts: Optional[ResourceOptions] = None)
    
    @overload
    def GlobalConnection(resource_name: str,
                         opts: Optional[ResourceOptions] = None,
                         apache_spark: Optional[GlobalConnectionApacheSparkArgs] = None,
                         athena: Optional[GlobalConnectionAthenaArgs] = None,
                         bigquery: Optional[GlobalConnectionBigqueryArgs] = None,
                         databricks: Optional[GlobalConnectionDatabricksArgs] = None,
                         fabric: Optional[GlobalConnectionFabricArgs] = None,
                         name: Optional[str] = None,
                         postgres: Optional[GlobalConnectionPostgresArgs] = None,
                         private_link_endpoint_id: Optional[str] = None,
                         redshift: Optional[GlobalConnectionRedshiftArgs] = None,
                         snowflake: Optional[GlobalConnectionSnowflakeArgs] = None,
                         starburst: Optional[GlobalConnectionStarburstArgs] = None,
                         synapse: Optional[GlobalConnectionSynapseArgs] = None)
    func NewGlobalConnection(ctx *Context, name string, args *GlobalConnectionArgs, opts ...ResourceOption) (*GlobalConnection, error)
    public GlobalConnection(string name, GlobalConnectionArgs? args = null, CustomResourceOptions? opts = null)
    public GlobalConnection(String name, GlobalConnectionArgs args)
    public GlobalConnection(String name, GlobalConnectionArgs args, CustomResourceOptions options)
    
    type: dbtcloud:GlobalConnection
    properties: # The arguments to resource properties.
    options: # Bag of options to control resource's behavior.
    
    

    Parameters

    name string
    The unique name of the resource.
    args GlobalConnectionArgs
    The arguments to resource properties.
    opts CustomResourceOptions
    Bag of options to control resource's behavior.
    resource_name str
    The unique name of the resource.
    args GlobalConnectionArgs
    The arguments to resource properties.
    opts ResourceOptions
    Bag of options to control resource's behavior.
    ctx Context
    Context object for the current deployment.
    name string
    The unique name of the resource.
    args GlobalConnectionArgs
    The arguments to resource properties.
    opts ResourceOption
    Bag of options to control resource's behavior.
    name string
    The unique name of the resource.
    args GlobalConnectionArgs
    The arguments to resource properties.
    opts CustomResourceOptions
    Bag of options to control resource's behavior.
    name String
    The unique name of the resource.
    args GlobalConnectionArgs
    The arguments to resource properties.
    options CustomResourceOptions
    Bag of options to control resource's behavior.

    Constructor example

    The following reference example uses placeholder values for all input properties.

    var globalConnectionResource = new DbtCloud.GlobalConnection("globalConnectionResource", new()
    {
        ApacheSpark = new DbtCloud.Inputs.GlobalConnectionApacheSparkArgs
        {
            Cluster = "string",
            Host = "string",
            Method = "string",
            Auth = "string",
            ConnectRetries = 0,
            ConnectTimeout = 0,
            Organization = "string",
            Port = 0,
            User = "string",
        },
        Athena = new DbtCloud.Inputs.GlobalConnectionAthenaArgs
        {
            Database = "string",
            RegionName = "string",
            S3StagingDir = "string",
            NumBoto3Retries = 0,
            NumIcebergRetries = 0,
            NumRetries = 0,
            PollInterval = 0,
            S3DataDir = "string",
            S3DataNaming = "string",
            S3TmpTableDir = "string",
            SparkWorkGroup = "string",
            WorkGroup = "string",
        },
        Bigquery = new DbtCloud.Inputs.GlobalConnectionBigqueryArgs
        {
            PrivateKey = "string",
            TokenUri = "string",
            AuthProviderX509CertUrl = "string",
            AuthUri = "string",
            ClientEmail = "string",
            ClientId = "string",
            ClientX509CertUrl = "string",
            GcpProjectId = "string",
            PrivateKeyId = "string",
            Scopes = new[]
            {
                "string",
            },
            MaximumBytesBilled = 0,
            ApplicationSecret = "string",
            JobRetryDeadlineSeconds = 0,
            JobCreationTimeoutSeconds = 0,
            ImpersonateServiceAccount = "string",
            Location = "string",
            ExecutionProject = "string",
            Priority = "string",
            ApplicationId = "string",
            DataprocRegion = "string",
            Retries = 0,
            GcsBucket = "string",
            TimeoutSeconds = 0,
            DataprocClusterName = "string",
        },
        Databricks = new DbtCloud.Inputs.GlobalConnectionDatabricksArgs
        {
            Host = "string",
            HttpPath = "string",
            Catalog = "string",
            ClientId = "string",
            ClientSecret = "string",
        },
        Fabric = new DbtCloud.Inputs.GlobalConnectionFabricArgs
        {
            Database = "string",
            Server = "string",
            LoginTimeout = 0,
            Port = 0,
            QueryTimeout = 0,
            Retries = 0,
        },
        Name = "string",
        Postgres = new DbtCloud.Inputs.GlobalConnectionPostgresArgs
        {
            Dbname = "string",
            Hostname = "string",
            Port = 0,
            SshTunnel = new DbtCloud.Inputs.GlobalConnectionPostgresSshTunnelArgs
            {
                Hostname = "string",
                Port = 0,
                Username = "string",
                Id = 0,
                PublicKey = "string",
            },
        },
        PrivateLinkEndpointId = "string",
        Redshift = new DbtCloud.Inputs.GlobalConnectionRedshiftArgs
        {
            Dbname = "string",
            Hostname = "string",
            Port = 0,
            SshTunnel = new DbtCloud.Inputs.GlobalConnectionRedshiftSshTunnelArgs
            {
                Hostname = "string",
                Port = 0,
                Username = "string",
                Id = 0,
                PublicKey = "string",
            },
        },
        Snowflake = new DbtCloud.Inputs.GlobalConnectionSnowflakeArgs
        {
            Account = "string",
            Database = "string",
            Warehouse = "string",
            AllowSso = false,
            ClientSessionKeepAlive = false,
            OauthClientId = "string",
            OauthClientSecret = "string",
            Role = "string",
        },
        Starburst = new DbtCloud.Inputs.GlobalConnectionStarburstArgs
        {
            Host = "string",
            Method = "string",
            Port = 0,
        },
        Synapse = new DbtCloud.Inputs.GlobalConnectionSynapseArgs
        {
            Database = "string",
            Host = "string",
            LoginTimeout = 0,
            Port = 0,
            QueryTimeout = 0,
            Retries = 0,
        },
    });
    
    example, err := dbtcloud.NewGlobalConnection(ctx, "globalConnectionResource", &dbtcloud.GlobalConnectionArgs{
    	ApacheSpark: &dbtcloud.GlobalConnectionApacheSparkArgs{
    		Cluster:        pulumi.String("string"),
    		Host:           pulumi.String("string"),
    		Method:         pulumi.String("string"),
    		Auth:           pulumi.String("string"),
    		ConnectRetries: pulumi.Int(0),
    		ConnectTimeout: pulumi.Int(0),
    		Organization:   pulumi.String("string"),
    		Port:           pulumi.Int(0),
    		User:           pulumi.String("string"),
    	},
    	Athena: &dbtcloud.GlobalConnectionAthenaArgs{
    		Database:          pulumi.String("string"),
    		RegionName:        pulumi.String("string"),
    		S3StagingDir:      pulumi.String("string"),
    		NumBoto3Retries:   pulumi.Int(0),
    		NumIcebergRetries: pulumi.Int(0),
    		NumRetries:        pulumi.Int(0),
    		PollInterval:      pulumi.Int(0),
    		S3DataDir:         pulumi.String("string"),
    		S3DataNaming:      pulumi.String("string"),
    		S3TmpTableDir:     pulumi.String("string"),
    		SparkWorkGroup:    pulumi.String("string"),
    		WorkGroup:         pulumi.String("string"),
    	},
    	Bigquery: &dbtcloud.GlobalConnectionBigqueryArgs{
    		PrivateKey:              pulumi.String("string"),
    		TokenUri:                pulumi.String("string"),
    		AuthProviderX509CertUrl: pulumi.String("string"),
    		AuthUri:                 pulumi.String("string"),
    		ClientEmail:             pulumi.String("string"),
    		ClientId:                pulumi.String("string"),
    		ClientX509CertUrl:       pulumi.String("string"),
    		GcpProjectId:            pulumi.String("string"),
    		PrivateKeyId:            pulumi.String("string"),
    		Scopes: pulumi.StringArray{
    			pulumi.String("string"),
    		},
    		MaximumBytesBilled:        pulumi.Int(0),
    		ApplicationSecret:         pulumi.String("string"),
    		JobRetryDeadlineSeconds:   pulumi.Int(0),
    		JobCreationTimeoutSeconds: pulumi.Int(0),
    		ImpersonateServiceAccount: pulumi.String("string"),
    		Location:                  pulumi.String("string"),
    		ExecutionProject:          pulumi.String("string"),
    		Priority:                  pulumi.String("string"),
    		ApplicationId:             pulumi.String("string"),
    		DataprocRegion:            pulumi.String("string"),
    		Retries:                   pulumi.Int(0),
    		GcsBucket:                 pulumi.String("string"),
    		TimeoutSeconds:            pulumi.Int(0),
    		DataprocClusterName:       pulumi.String("string"),
    	},
    	Databricks: &dbtcloud.GlobalConnectionDatabricksArgs{
    		Host:         pulumi.String("string"),
    		HttpPath:     pulumi.String("string"),
    		Catalog:      pulumi.String("string"),
    		ClientId:     pulumi.String("string"),
    		ClientSecret: pulumi.String("string"),
    	},
    	Fabric: &dbtcloud.GlobalConnectionFabricArgs{
    		Database:     pulumi.String("string"),
    		Server:       pulumi.String("string"),
    		LoginTimeout: pulumi.Int(0),
    		Port:         pulumi.Int(0),
    		QueryTimeout: pulumi.Int(0),
    		Retries:      pulumi.Int(0),
    	},
    	Name: pulumi.String("string"),
    	Postgres: &dbtcloud.GlobalConnectionPostgresArgs{
    		Dbname:   pulumi.String("string"),
    		Hostname: pulumi.String("string"),
    		Port:     pulumi.Int(0),
    		SshTunnel: &dbtcloud.GlobalConnectionPostgresSshTunnelArgs{
    			Hostname:  pulumi.String("string"),
    			Port:      pulumi.Int(0),
    			Username:  pulumi.String("string"),
    			Id:        pulumi.Int(0),
    			PublicKey: pulumi.String("string"),
    		},
    	},
    	PrivateLinkEndpointId: pulumi.String("string"),
    	Redshift: &dbtcloud.GlobalConnectionRedshiftArgs{
    		Dbname:   pulumi.String("string"),
    		Hostname: pulumi.String("string"),
    		Port:     pulumi.Int(0),
    		SshTunnel: &dbtcloud.GlobalConnectionRedshiftSshTunnelArgs{
    			Hostname:  pulumi.String("string"),
    			Port:      pulumi.Int(0),
    			Username:  pulumi.String("string"),
    			Id:        pulumi.Int(0),
    			PublicKey: pulumi.String("string"),
    		},
    	},
    	Snowflake: &dbtcloud.GlobalConnectionSnowflakeArgs{
    		Account:                pulumi.String("string"),
    		Database:               pulumi.String("string"),
    		Warehouse:              pulumi.String("string"),
    		AllowSso:               pulumi.Bool(false),
    		ClientSessionKeepAlive: pulumi.Bool(false),
    		OauthClientId:          pulumi.String("string"),
    		OauthClientSecret:      pulumi.String("string"),
    		Role:                   pulumi.String("string"),
    	},
    	Starburst: &dbtcloud.GlobalConnectionStarburstArgs{
    		Host:   pulumi.String("string"),
    		Method: pulumi.String("string"),
    		Port:   pulumi.Int(0),
    	},
    	Synapse: &dbtcloud.GlobalConnectionSynapseArgs{
    		Database:     pulumi.String("string"),
    		Host:         pulumi.String("string"),
    		LoginTimeout: pulumi.Int(0),
    		Port:         pulumi.Int(0),
    		QueryTimeout: pulumi.Int(0),
    		Retries:      pulumi.Int(0),
    	},
    })
    
    var globalConnectionResource = new GlobalConnection("globalConnectionResource", GlobalConnectionArgs.builder()
        .apacheSpark(GlobalConnectionApacheSparkArgs.builder()
            .cluster("string")
            .host("string")
            .method("string")
            .auth("string")
            .connectRetries(0)
            .connectTimeout(0)
            .organization("string")
            .port(0)
            .user("string")
            .build())
        .athena(GlobalConnectionAthenaArgs.builder()
            .database("string")
            .regionName("string")
            .s3StagingDir("string")
            .numBoto3Retries(0)
            .numIcebergRetries(0)
            .numRetries(0)
            .pollInterval(0)
            .s3DataDir("string")
            .s3DataNaming("string")
            .s3TmpTableDir("string")
            .sparkWorkGroup("string")
            .workGroup("string")
            .build())
        .bigquery(GlobalConnectionBigqueryArgs.builder()
            .privateKey("string")
            .tokenUri("string")
            .authProviderX509CertUrl("string")
            .authUri("string")
            .clientEmail("string")
            .clientId("string")
            .clientX509CertUrl("string")
            .gcpProjectId("string")
            .privateKeyId("string")
            .scopes("string")
            .maximumBytesBilled(0)
            .applicationSecret("string")
            .jobRetryDeadlineSeconds(0)
            .jobCreationTimeoutSeconds(0)
            .impersonateServiceAccount("string")
            .location("string")
            .executionProject("string")
            .priority("string")
            .applicationId("string")
            .dataprocRegion("string")
            .retries(0)
            .gcsBucket("string")
            .timeoutSeconds(0)
            .dataprocClusterName("string")
            .build())
        .databricks(GlobalConnectionDatabricksArgs.builder()
            .host("string")
            .httpPath("string")
            .catalog("string")
            .clientId("string")
            .clientSecret("string")
            .build())
        .fabric(GlobalConnectionFabricArgs.builder()
            .database("string")
            .server("string")
            .loginTimeout(0)
            .port(0)
            .queryTimeout(0)
            .retries(0)
            .build())
        .name("string")
        .postgres(GlobalConnectionPostgresArgs.builder()
            .dbname("string")
            .hostname("string")
            .port(0)
            .sshTunnel(GlobalConnectionPostgresSshTunnelArgs.builder()
                .hostname("string")
                .port(0)
                .username("string")
                .id(0)
                .publicKey("string")
                .build())
            .build())
        .privateLinkEndpointId("string")
        .redshift(GlobalConnectionRedshiftArgs.builder()
            .dbname("string")
            .hostname("string")
            .port(0)
            .sshTunnel(GlobalConnectionRedshiftSshTunnelArgs.builder()
                .hostname("string")
                .port(0)
                .username("string")
                .id(0)
                .publicKey("string")
                .build())
            .build())
        .snowflake(GlobalConnectionSnowflakeArgs.builder()
            .account("string")
            .database("string")
            .warehouse("string")
            .allowSso(false)
            .clientSessionKeepAlive(false)
            .oauthClientId("string")
            .oauthClientSecret("string")
            .role("string")
            .build())
        .starburst(GlobalConnectionStarburstArgs.builder()
            .host("string")
            .method("string")
            .port(0)
            .build())
        .synapse(GlobalConnectionSynapseArgs.builder()
            .database("string")
            .host("string")
            .loginTimeout(0)
            .port(0)
            .queryTimeout(0)
            .retries(0)
            .build())
        .build());
    
    global_connection_resource = dbtcloud.GlobalConnection("globalConnectionResource",
        apache_spark=dbtcloud.GlobalConnectionApacheSparkArgs(
            cluster="string",
            host="string",
            method="string",
            auth="string",
            connect_retries=0,
            connect_timeout=0,
            organization="string",
            port=0,
            user="string",
        ),
        athena=dbtcloud.GlobalConnectionAthenaArgs(
            database="string",
            region_name="string",
            s3_staging_dir="string",
            num_boto3_retries=0,
            num_iceberg_retries=0,
            num_retries=0,
            poll_interval=0,
            s3_data_dir="string",
            s3_data_naming="string",
            s3_tmp_table_dir="string",
            spark_work_group="string",
            work_group="string",
        ),
        bigquery=dbtcloud.GlobalConnectionBigqueryArgs(
            private_key="string",
            token_uri="string",
            auth_provider_x509_cert_url="string",
            auth_uri="string",
            client_email="string",
            client_id="string",
            client_x509_cert_url="string",
            gcp_project_id="string",
            private_key_id="string",
            scopes=["string"],
            maximum_bytes_billed=0,
            application_secret="string",
            job_retry_deadline_seconds=0,
            job_creation_timeout_seconds=0,
            impersonate_service_account="string",
            location="string",
            execution_project="string",
            priority="string",
            application_id="string",
            dataproc_region="string",
            retries=0,
            gcs_bucket="string",
            timeout_seconds=0,
            dataproc_cluster_name="string",
        ),
        databricks=dbtcloud.GlobalConnectionDatabricksArgs(
            host="string",
            http_path="string",
            catalog="string",
            client_id="string",
            client_secret="string",
        ),
        fabric=dbtcloud.GlobalConnectionFabricArgs(
            database="string",
            server="string",
            login_timeout=0,
            port=0,
            query_timeout=0,
            retries=0,
        ),
        name="string",
        postgres=dbtcloud.GlobalConnectionPostgresArgs(
            dbname="string",
            hostname="string",
            port=0,
            ssh_tunnel=dbtcloud.GlobalConnectionPostgresSshTunnelArgs(
                hostname="string",
                port=0,
                username="string",
                id=0,
                public_key="string",
            ),
        ),
        private_link_endpoint_id="string",
        redshift=dbtcloud.GlobalConnectionRedshiftArgs(
            dbname="string",
            hostname="string",
            port=0,
            ssh_tunnel=dbtcloud.GlobalConnectionRedshiftSshTunnelArgs(
                hostname="string",
                port=0,
                username="string",
                id=0,
                public_key="string",
            ),
        ),
        snowflake=dbtcloud.GlobalConnectionSnowflakeArgs(
            account="string",
            database="string",
            warehouse="string",
            allow_sso=False,
            client_session_keep_alive=False,
            oauth_client_id="string",
            oauth_client_secret="string",
            role="string",
        ),
        starburst=dbtcloud.GlobalConnectionStarburstArgs(
            host="string",
            method="string",
            port=0,
        ),
        synapse=dbtcloud.GlobalConnectionSynapseArgs(
            database="string",
            host="string",
            login_timeout=0,
            port=0,
            query_timeout=0,
            retries=0,
        ))
    
    const globalConnectionResource = new dbtcloud.GlobalConnection("globalConnectionResource", {
        apacheSpark: {
            cluster: "string",
            host: "string",
            method: "string",
            auth: "string",
            connectRetries: 0,
            connectTimeout: 0,
            organization: "string",
            port: 0,
            user: "string",
        },
        athena: {
            database: "string",
            regionName: "string",
            s3StagingDir: "string",
            numBoto3Retries: 0,
            numIcebergRetries: 0,
            numRetries: 0,
            pollInterval: 0,
            s3DataDir: "string",
            s3DataNaming: "string",
            s3TmpTableDir: "string",
            sparkWorkGroup: "string",
            workGroup: "string",
        },
        bigquery: {
            privateKey: "string",
            tokenUri: "string",
            authProviderX509CertUrl: "string",
            authUri: "string",
            clientEmail: "string",
            clientId: "string",
            clientX509CertUrl: "string",
            gcpProjectId: "string",
            privateKeyId: "string",
            scopes: ["string"],
            maximumBytesBilled: 0,
            applicationSecret: "string",
            jobRetryDeadlineSeconds: 0,
            jobCreationTimeoutSeconds: 0,
            impersonateServiceAccount: "string",
            location: "string",
            executionProject: "string",
            priority: "string",
            applicationId: "string",
            dataprocRegion: "string",
            retries: 0,
            gcsBucket: "string",
            timeoutSeconds: 0,
            dataprocClusterName: "string",
        },
        databricks: {
            host: "string",
            httpPath: "string",
            catalog: "string",
            clientId: "string",
            clientSecret: "string",
        },
        fabric: {
            database: "string",
            server: "string",
            loginTimeout: 0,
            port: 0,
            queryTimeout: 0,
            retries: 0,
        },
        name: "string",
        postgres: {
            dbname: "string",
            hostname: "string",
            port: 0,
            sshTunnel: {
                hostname: "string",
                port: 0,
                username: "string",
                id: 0,
                publicKey: "string",
            },
        },
        privateLinkEndpointId: "string",
        redshift: {
            dbname: "string",
            hostname: "string",
            port: 0,
            sshTunnel: {
                hostname: "string",
                port: 0,
                username: "string",
                id: 0,
                publicKey: "string",
            },
        },
        snowflake: {
            account: "string",
            database: "string",
            warehouse: "string",
            allowSso: false,
            clientSessionKeepAlive: false,
            oauthClientId: "string",
            oauthClientSecret: "string",
            role: "string",
        },
        starburst: {
            host: "string",
            method: "string",
            port: 0,
        },
        synapse: {
            database: "string",
            host: "string",
            loginTimeout: 0,
            port: 0,
            queryTimeout: 0,
            retries: 0,
        },
    });
    
    type: dbtcloud:GlobalConnection
    properties:
        apacheSpark:
            auth: string
            cluster: string
            connectRetries: 0
            connectTimeout: 0
            host: string
            method: string
            organization: string
            port: 0
            user: string
        athena:
            database: string
            numBoto3Retries: 0
            numIcebergRetries: 0
            numRetries: 0
            pollInterval: 0
            regionName: string
            s3DataDir: string
            s3DataNaming: string
            s3StagingDir: string
            s3TmpTableDir: string
            sparkWorkGroup: string
            workGroup: string
        bigquery:
            applicationId: string
            applicationSecret: string
            authProviderX509CertUrl: string
            authUri: string
            clientEmail: string
            clientId: string
            clientX509CertUrl: string
            dataprocClusterName: string
            dataprocRegion: string
            executionProject: string
            gcpProjectId: string
            gcsBucket: string
            impersonateServiceAccount: string
            jobCreationTimeoutSeconds: 0
            jobRetryDeadlineSeconds: 0
            location: string
            maximumBytesBilled: 0
            priority: string
            privateKey: string
            privateKeyId: string
            retries: 0
            scopes:
                - string
            timeoutSeconds: 0
            tokenUri: string
        databricks:
            catalog: string
            clientId: string
            clientSecret: string
            host: string
            httpPath: string
        fabric:
            database: string
            loginTimeout: 0
            port: 0
            queryTimeout: 0
            retries: 0
            server: string
        name: string
        postgres:
            dbname: string
            hostname: string
            port: 0
            sshTunnel:
                hostname: string
                id: 0
                port: 0
                publicKey: string
                username: string
        privateLinkEndpointId: string
        redshift:
            dbname: string
            hostname: string
            port: 0
            sshTunnel:
                hostname: string
                id: 0
                port: 0
                publicKey: string
                username: string
        snowflake:
            account: string
            allowSso: false
            clientSessionKeepAlive: false
            database: string
            oauthClientId: string
            oauthClientSecret: string
            role: string
            warehouse: string
        starburst:
            host: string
            method: string
            port: 0
        synapse:
            database: string
            host: string
            loginTimeout: 0
            port: 0
            queryTimeout: 0
            retries: 0
    

    GlobalConnection Resource Properties

    To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.

    Inputs

    The GlobalConnection resource accepts the following input properties:

    ApacheSpark Pulumi.DbtCloud.Inputs.GlobalConnectionApacheSpark
    Apache Spark connection configuration.
    Athena Pulumi.DbtCloud.Inputs.GlobalConnectionAthena
    Athena connection configuration.
    Bigquery Pulumi.DbtCloud.Inputs.GlobalConnectionBigquery
    Databricks Pulumi.DbtCloud.Inputs.GlobalConnectionDatabricks
    Databricks connection configuration
    Fabric Pulumi.DbtCloud.Inputs.GlobalConnectionFabric
    Microsoft Fabric connection configuration.
    Name string
    Connection name
    Postgres Pulumi.DbtCloud.Inputs.GlobalConnectionPostgres
    PostgreSQL connection configuration.
    PrivateLinkEndpointId string
    Private Link Endpoint ID. This ID can be found using the privatelink_endpoint data source
    Redshift Pulumi.DbtCloud.Inputs.GlobalConnectionRedshift
    Redshift connection configuration
    Snowflake Pulumi.DbtCloud.Inputs.GlobalConnectionSnowflake
    Snowflake connection configuration
    Starburst Pulumi.DbtCloud.Inputs.GlobalConnectionStarburst
    Starburst/Trino connection configuration.
    Synapse Pulumi.DbtCloud.Inputs.GlobalConnectionSynapse
    Azure Synapse Analytics connection configuration.
    ApacheSpark GlobalConnectionApacheSparkArgs
    Apache Spark connection configuration.
    Athena GlobalConnectionAthenaArgs
    Athena connection configuration.
    Bigquery GlobalConnectionBigqueryArgs
    Databricks GlobalConnectionDatabricksArgs
    Databricks connection configuration
    Fabric GlobalConnectionFabricArgs
    Microsoft Fabric connection configuration.
    Name string
    Connection name
    Postgres GlobalConnectionPostgresArgs
    PostgreSQL connection configuration.
    PrivateLinkEndpointId string
    Private Link Endpoint ID. This ID can be found using the privatelink_endpoint data source
    Redshift GlobalConnectionRedshiftArgs
    Redshift connection configuration
    Snowflake GlobalConnectionSnowflakeArgs
    Snowflake connection configuration
    Starburst GlobalConnectionStarburstArgs
    Starburst/Trino connection configuration.
    Synapse GlobalConnectionSynapseArgs
    Azure Synapse Analytics connection configuration.
    apacheSpark GlobalConnectionApacheSpark
    Apache Spark connection configuration.
    athena GlobalConnectionAthena
    Athena connection configuration.
    bigquery GlobalConnectionBigquery
    databricks GlobalConnectionDatabricks
    Databricks connection configuration
    fabric GlobalConnectionFabric
    Microsoft Fabric connection configuration.
    name String
    Connection name
    postgres GlobalConnectionPostgres
    PostgreSQL connection configuration.
    privateLinkEndpointId String
    Private Link Endpoint ID. This ID can be found using the privatelink_endpoint data source
    redshift GlobalConnectionRedshift
    Redshift connection configuration
    snowflake GlobalConnectionSnowflake
    Snowflake connection configuration
    starburst GlobalConnectionStarburst
    Starburst/Trino connection configuration.
    synapse GlobalConnectionSynapse
    Azure Synapse Analytics connection configuration.
    apacheSpark GlobalConnectionApacheSpark
    Apache Spark connection configuration.
    athena GlobalConnectionAthena
    Athena connection configuration.
    bigquery GlobalConnectionBigquery
    databricks GlobalConnectionDatabricks
    Databricks connection configuration
    fabric GlobalConnectionFabric
    Microsoft Fabric connection configuration.
    name string
    Connection name
    postgres GlobalConnectionPostgres
    PostgreSQL connection configuration.
    privateLinkEndpointId string
    Private Link Endpoint ID. This ID can be found using the privatelink_endpoint data source
    redshift GlobalConnectionRedshift
    Redshift connection configuration
    snowflake GlobalConnectionSnowflake
    Snowflake connection configuration
    starburst GlobalConnectionStarburst
    Starburst/Trino connection configuration.
    synapse GlobalConnectionSynapse
    Azure Synapse Analytics connection configuration.
    apache_spark GlobalConnectionApacheSparkArgs
    Apache Spark connection configuration.
    athena GlobalConnectionAthenaArgs
    Athena connection configuration.
    bigquery GlobalConnectionBigqueryArgs
    databricks GlobalConnectionDatabricksArgs
    Databricks connection configuration
    fabric GlobalConnectionFabricArgs
    Microsoft Fabric connection configuration.
    name str
    Connection name
    postgres GlobalConnectionPostgresArgs
    PostgreSQL connection configuration.
    private_link_endpoint_id str
    Private Link Endpoint ID. This ID can be found using the privatelink_endpoint data source
    redshift GlobalConnectionRedshiftArgs
    Redshift connection configuration
    snowflake GlobalConnectionSnowflakeArgs
    Snowflake connection configuration
    starburst GlobalConnectionStarburstArgs
    Starburst/Trino connection configuration.
    synapse GlobalConnectionSynapseArgs
    Azure Synapse Analytics connection configuration.
    apacheSpark Property Map
    Apache Spark connection configuration.
    athena Property Map
    Athena connection configuration.
    bigquery Property Map
    databricks Property Map
    Databricks connection configuration
    fabric Property Map
    Microsoft Fabric connection configuration.
    name String
    Connection name
    postgres Property Map
    PostgreSQL connection configuration.
    privateLinkEndpointId String
    Private Link Endpoint ID. This ID can be found using the privatelink_endpoint data source
    redshift Property Map
    Redshift connection configuration
    snowflake Property Map
    Snowflake connection configuration
    starburst Property Map
    Starburst/Trino connection configuration.
    synapse Property Map
    Azure Synapse Analytics connection configuration.

    Outputs

    All input properties are implicitly available as output properties. Additionally, the GlobalConnection resource produces the following output properties:

    AdapterVersion string
    Version of the adapter
    Id string
    The provider-assigned unique ID for this managed resource.
    IsSshTunnelEnabled bool
    Whether the connection can use an SSH tunnel
    OauthConfigurationId int
    AdapterVersion string
    Version of the adapter
    Id string
    The provider-assigned unique ID for this managed resource.
    IsSshTunnelEnabled bool
    Whether the connection can use an SSH tunnel
    OauthConfigurationId int
    adapterVersion String
    Version of the adapter
    id String
    The provider-assigned unique ID for this managed resource.
    isSshTunnelEnabled Boolean
    Whether the connection can use an SSH tunnel
    oauthConfigurationId Integer
    adapterVersion string
    Version of the adapter
    id string
    The provider-assigned unique ID for this managed resource.
    isSshTunnelEnabled boolean
    Whether the connection can use an SSH tunnel
    oauthConfigurationId number
    adapter_version str
    Version of the adapter
    id str
    The provider-assigned unique ID for this managed resource.
    is_ssh_tunnel_enabled bool
    Whether the connection can use an SSH tunnel
    oauth_configuration_id int
    adapterVersion String
    Version of the adapter
    id String
    The provider-assigned unique ID for this managed resource.
    isSshTunnelEnabled Boolean
    Whether the connection can use an SSH tunnel
    oauthConfigurationId Number

    Look up Existing GlobalConnection Resource

    Get an existing GlobalConnection resource’s state with the given name, ID, and optional extra properties used to qualify the lookup.

    public static get(name: string, id: Input<ID>, state?: GlobalConnectionState, opts?: CustomResourceOptions): GlobalConnection
    @staticmethod
    def get(resource_name: str,
            id: str,
            opts: Optional[ResourceOptions] = None,
            adapter_version: Optional[str] = None,
            apache_spark: Optional[GlobalConnectionApacheSparkArgs] = None,
            athena: Optional[GlobalConnectionAthenaArgs] = None,
            bigquery: Optional[GlobalConnectionBigqueryArgs] = None,
            databricks: Optional[GlobalConnectionDatabricksArgs] = None,
            fabric: Optional[GlobalConnectionFabricArgs] = None,
            is_ssh_tunnel_enabled: Optional[bool] = None,
            name: Optional[str] = None,
            oauth_configuration_id: Optional[int] = None,
            postgres: Optional[GlobalConnectionPostgresArgs] = None,
            private_link_endpoint_id: Optional[str] = None,
            redshift: Optional[GlobalConnectionRedshiftArgs] = None,
            snowflake: Optional[GlobalConnectionSnowflakeArgs] = None,
            starburst: Optional[GlobalConnectionStarburstArgs] = None,
            synapse: Optional[GlobalConnectionSynapseArgs] = None) -> GlobalConnection
    func GetGlobalConnection(ctx *Context, name string, id IDInput, state *GlobalConnectionState, opts ...ResourceOption) (*GlobalConnection, error)
    public static GlobalConnection Get(string name, Input<string> id, GlobalConnectionState? state, CustomResourceOptions? opts = null)
    public static GlobalConnection get(String name, Output<String> id, GlobalConnectionState state, CustomResourceOptions options)
    Resource lookup is not supported in YAML
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    resource_name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    The following state arguments are supported:
    AdapterVersion string
    Version of the adapter
    ApacheSpark Pulumi.DbtCloud.Inputs.GlobalConnectionApacheSpark
    Apache Spark connection configuration.
    Athena Pulumi.DbtCloud.Inputs.GlobalConnectionAthena
    Athena connection configuration.
    Bigquery Pulumi.DbtCloud.Inputs.GlobalConnectionBigquery
    Databricks Pulumi.DbtCloud.Inputs.GlobalConnectionDatabricks
    Databricks connection configuration
    Fabric Pulumi.DbtCloud.Inputs.GlobalConnectionFabric
    Microsoft Fabric connection configuration.
    IsSshTunnelEnabled bool
    Whether the connection can use an SSH tunnel
    Name string
    Connection name
    OauthConfigurationId int
    Postgres Pulumi.DbtCloud.Inputs.GlobalConnectionPostgres
    PostgreSQL connection configuration.
    PrivateLinkEndpointId string
    Private Link Endpoint ID. This ID can be found using the privatelink_endpoint data source
    Redshift Pulumi.DbtCloud.Inputs.GlobalConnectionRedshift
    Redshift connection configuration
    Snowflake Pulumi.DbtCloud.Inputs.GlobalConnectionSnowflake
    Snowflake connection configuration
    Starburst Pulumi.DbtCloud.Inputs.GlobalConnectionStarburst
    Starburst/Trino connection configuration.
    Synapse Pulumi.DbtCloud.Inputs.GlobalConnectionSynapse
    Azure Synapse Analytics connection configuration.
    AdapterVersion string
    Version of the adapter
    ApacheSpark GlobalConnectionApacheSparkArgs
    Apache Spark connection configuration.
    Athena GlobalConnectionAthenaArgs
    Athena connection configuration.
    Bigquery GlobalConnectionBigqueryArgs
    Databricks GlobalConnectionDatabricksArgs
    Databricks connection configuration
    Fabric GlobalConnectionFabricArgs
    Microsoft Fabric connection configuration.
    IsSshTunnelEnabled bool
    Whether the connection can use an SSH tunnel
    Name string
    Connection name
    OauthConfigurationId int
    Postgres GlobalConnectionPostgresArgs
    PostgreSQL connection configuration.
    PrivateLinkEndpointId string
    Private Link Endpoint ID. This ID can be found using the privatelink_endpoint data source
    Redshift GlobalConnectionRedshiftArgs
    Redshift connection configuration
    Snowflake GlobalConnectionSnowflakeArgs
    Snowflake connection configuration
    Starburst GlobalConnectionStarburstArgs
    Starburst/Trino connection configuration.
    Synapse GlobalConnectionSynapseArgs
    Azure Synapse Analytics connection configuration.
    adapterVersion String
    Version of the adapter
    apacheSpark GlobalConnectionApacheSpark
    Apache Spark connection configuration.
    athena GlobalConnectionAthena
    Athena connection configuration.
    bigquery GlobalConnectionBigquery
    databricks GlobalConnectionDatabricks
    Databricks connection configuration
    fabric GlobalConnectionFabric
    Microsoft Fabric connection configuration.
    isSshTunnelEnabled Boolean
    Whether the connection can use an SSH tunnel
    name String
    Connection name
    oauthConfigurationId Integer
    postgres GlobalConnectionPostgres
    PostgreSQL connection configuration.
    privateLinkEndpointId String
    Private Link Endpoint ID. This ID can be found using the privatelink_endpoint data source
    redshift GlobalConnectionRedshift
    Redshift connection configuration
    snowflake GlobalConnectionSnowflake
    Snowflake connection configuration
    starburst GlobalConnectionStarburst
    Starburst/Trino connection configuration.
    synapse GlobalConnectionSynapse
    Azure Synapse Analytics connection configuration.
    adapterVersion string
    Version of the adapter
    apacheSpark GlobalConnectionApacheSpark
    Apache Spark connection configuration.
    athena GlobalConnectionAthena
    Athena connection configuration.
    bigquery GlobalConnectionBigquery
    databricks GlobalConnectionDatabricks
    Databricks connection configuration
    fabric GlobalConnectionFabric
    Microsoft Fabric connection configuration.
    isSshTunnelEnabled boolean
    Whether the connection can use an SSH tunnel
    name string
    Connection name
    oauthConfigurationId number
    postgres GlobalConnectionPostgres
    PostgreSQL connection configuration.
    privateLinkEndpointId string
    Private Link Endpoint ID. This ID can be found using the privatelink_endpoint data source
    redshift GlobalConnectionRedshift
    Redshift connection configuration
    snowflake GlobalConnectionSnowflake
    Snowflake connection configuration
    starburst GlobalConnectionStarburst
    Starburst/Trino connection configuration.
    synapse GlobalConnectionSynapse
    Azure Synapse Analytics connection configuration.
    adapter_version str
    Version of the adapter
    apache_spark GlobalConnectionApacheSparkArgs
    Apache Spark connection configuration.
    athena GlobalConnectionAthenaArgs
    Athena connection configuration.
    bigquery GlobalConnectionBigqueryArgs
    databricks GlobalConnectionDatabricksArgs
    Databricks connection configuration
    fabric GlobalConnectionFabricArgs
    Microsoft Fabric connection configuration.
    is_ssh_tunnel_enabled bool
    Whether the connection can use an SSH tunnel
    name str
    Connection name
    oauth_configuration_id int
    postgres GlobalConnectionPostgresArgs
    PostgreSQL connection configuration.
    private_link_endpoint_id str
    Private Link Endpoint ID. This ID can be found using the privatelink_endpoint data source
    redshift GlobalConnectionRedshiftArgs
    Redshift connection configuration
    snowflake GlobalConnectionSnowflakeArgs
    Snowflake connection configuration
    starburst GlobalConnectionStarburstArgs
    Starburst/Trino connection configuration.
    synapse GlobalConnectionSynapseArgs
    Azure Synapse Analytics connection configuration.
    adapterVersion String
    Version of the adapter
    apacheSpark Property Map
    Apache Spark connection configuration.
    athena Property Map
    Athena connection configuration.
    bigquery Property Map
    databricks Property Map
    Databricks connection configuration
    fabric Property Map
    Microsoft Fabric connection configuration.
    isSshTunnelEnabled Boolean
    Whether the connection can use an SSH tunnel
    name String
    Connection name
    oauthConfigurationId Number
    postgres Property Map
    PostgreSQL connection configuration.
    privateLinkEndpointId String
    Private Link Endpoint ID. This ID can be found using the privatelink_endpoint data source
    redshift Property Map
    Redshift connection configuration
    snowflake Property Map
    Snowflake connection configuration
    starburst Property Map
    Starburst/Trino connection configuration.
    synapse Property Map
    Azure Synapse Analytics connection configuration.

    Supporting Types

    GlobalConnectionApacheSpark, GlobalConnectionApacheSparkArgs

    Cluster string
    Spark cluster for the connection
    Host string
    Hostname of the connection
    Method string
    Authentication method for the connection (http or thrift).
    Auth string
    Auth
    ConnectRetries int
    Connection retries. Default=0
    ConnectTimeout int
    Connection time out in seconds. Default=10
    Organization string
    Organization ID
    Port int
    Port for the connection. Default=443
    User string
    User
    Cluster string
    Spark cluster for the connection
    Host string
    Hostname of the connection
    Method string
    Authentication method for the connection (http or thrift).
    Auth string
    Auth
    ConnectRetries int
    Connection retries. Default=0
    ConnectTimeout int
    Connection time out in seconds. Default=10
    Organization string
    Organization ID
    Port int
    Port for the connection. Default=443
    User string
    User
    cluster String
    Spark cluster for the connection
    host String
    Hostname of the connection
    method String
    Authentication method for the connection (http or thrift).
    auth String
    Auth
    connectRetries Integer
    Connection retries. Default=0
    connectTimeout Integer
    Connection time out in seconds. Default=10
    organization String
    Organization ID
    port Integer
    Port for the connection. Default=443
    user String
    User
    cluster string
    Spark cluster for the connection
    host string
    Hostname of the connection
    method string
    Authentication method for the connection (http or thrift).
    auth string
    Auth
    connectRetries number
    Connection retries. Default=0
    connectTimeout number
    Connection time out in seconds. Default=10
    organization string
    Organization ID
    port number
    Port for the connection. Default=443
    user string
    User
    cluster str
    Spark cluster for the connection
    host str
    Hostname of the connection
    method str
    Authentication method for the connection (http or thrift).
    auth str
    Auth
    connect_retries int
    Connection retries. Default=0
    connect_timeout int
    Connection time out in seconds. Default=10
    organization str
    Organization ID
    port int
    Port for the connection. Default=443
    user str
    User
    cluster String
    Spark cluster for the connection
    host String
    Hostname of the connection
    method String
    Authentication method for the connection (http or thrift).
    auth String
    Auth
    connectRetries Number
    Connection retries. Default=0
    connectTimeout Number
    Connection time out in seconds. Default=10
    organization String
    Organization ID
    port Number
    Port for the connection. Default=443
    user String
    User

    GlobalConnectionAthena, GlobalConnectionAthenaArgs

    Database string
    Specify the database (data catalog) to build models into (lowercase only).
    RegionName string
    AWS region of your Athena instance.
    S3StagingDir string
    S3 location to store Athena query results and metadata.
    NumBoto3Retries int
    Number of times to retry boto3 requests (e.g. deleting S3 files for materialized tables).
    NumIcebergRetries int
    Number of times to retry iceberg commit queries to fix ICEBERGCOMMITERROR.
    NumRetries int
    Number of times to retry a failing query.
    PollInterval int
    Interval in seconds to use for polling the status of query results in Athena.
    S3DataDir string
    Prefix for storing tables, if different from the connection's S3 staging directory.
    S3DataNaming string
    How to generate table paths in the S3 data directory.
    S3TmpTableDir string
    Prefix for storing temporary tables, if different from the connection's S3 data directory.
    SparkWorkGroup string
    Identifier of Athena Spark workgroup for running Python models.
    WorkGroup string
    Identifier of Athena workgroup.
    Database string
    Specify the database (data catalog) to build models into (lowercase only).
    RegionName string
    AWS region of your Athena instance.
    S3StagingDir string
    S3 location to store Athena query results and metadata.
    NumBoto3Retries int
    Number of times to retry boto3 requests (e.g. deleting S3 files for materialized tables).
    NumIcebergRetries int
    Number of times to retry iceberg commit queries to fix ICEBERGCOMMITERROR.
    NumRetries int
    Number of times to retry a failing query.
    PollInterval int
    Interval in seconds to use for polling the status of query results in Athena.
    S3DataDir string
    Prefix for storing tables, if different from the connection's S3 staging directory.
    S3DataNaming string
    How to generate table paths in the S3 data directory.
    S3TmpTableDir string
    Prefix for storing temporary tables, if different from the connection's S3 data directory.
    SparkWorkGroup string
    Identifier of Athena Spark workgroup for running Python models.
    WorkGroup string
    Identifier of Athena workgroup.
    database String
    Specify the database (data catalog) to build models into (lowercase only).
    regionName String
    AWS region of your Athena instance.
    s3StagingDir String
    S3 location to store Athena query results and metadata.
    numBoto3Retries Integer
    Number of times to retry boto3 requests (e.g. deleting S3 files for materialized tables).
    numIcebergRetries Integer
    Number of times to retry iceberg commit queries to fix ICEBERGCOMMITERROR.
    numRetries Integer
    Number of times to retry a failing query.
    pollInterval Integer
    Interval in seconds to use for polling the status of query results in Athena.
    s3DataDir String
    Prefix for storing tables, if different from the connection's S3 staging directory.
    s3DataNaming String
    How to generate table paths in the S3 data directory.
    s3TmpTableDir String
    Prefix for storing temporary tables, if different from the connection's S3 data directory.
    sparkWorkGroup String
    Identifier of Athena Spark workgroup for running Python models.
    workGroup String
    Identifier of Athena workgroup.
    database string
    Specify the database (data catalog) to build models into (lowercase only).
    regionName string
    AWS region of your Athena instance.
    s3StagingDir string
    S3 location to store Athena query results and metadata.
    numBoto3Retries number
    Number of times to retry boto3 requests (e.g. deleting S3 files for materialized tables).
    numIcebergRetries number
    Number of times to retry iceberg commit queries to fix ICEBERGCOMMITERROR.
    numRetries number
    Number of times to retry a failing query.
    pollInterval number
    Interval in seconds to use for polling the status of query results in Athena.
    s3DataDir string
    Prefix for storing tables, if different from the connection's S3 staging directory.
    s3DataNaming string
    How to generate table paths in the S3 data directory.
    s3TmpTableDir string
    Prefix for storing temporary tables, if different from the connection's S3 data directory.
    sparkWorkGroup string
    Identifier of Athena Spark workgroup for running Python models.
    workGroup string
    Identifier of Athena workgroup.
    database str
    Specify the database (data catalog) to build models into (lowercase only).
    region_name str
    AWS region of your Athena instance.
    s3_staging_dir str
    S3 location to store Athena query results and metadata.
    num_boto3_retries int
    Number of times to retry boto3 requests (e.g. deleting S3 files for materialized tables).
    num_iceberg_retries int
    Number of times to retry iceberg commit queries to fix ICEBERGCOMMITERROR.
    num_retries int
    Number of times to retry a failing query.
    poll_interval int
    Interval in seconds to use for polling the status of query results in Athena.
    s3_data_dir str
    Prefix for storing tables, if different from the connection's S3 staging directory.
    s3_data_naming str
    How to generate table paths in the S3 data directory.
    s3_tmp_table_dir str
    Prefix for storing temporary tables, if different from the connection's S3 data directory.
    spark_work_group str
    Identifier of Athena Spark workgroup for running Python models.
    work_group str
    Identifier of Athena workgroup.
    database String
    Specify the database (data catalog) to build models into (lowercase only).
    regionName String
    AWS region of your Athena instance.
    s3StagingDir String
    S3 location to store Athena query results and metadata.
    numBoto3Retries Number
    Number of times to retry boto3 requests (e.g. deleting S3 files for materialized tables).
    numIcebergRetries Number
    Number of times to retry iceberg commit queries to fix ICEBERGCOMMITERROR.
    numRetries Number
    Number of times to retry a failing query.
    pollInterval Number
    Interval in seconds to use for polling the status of query results in Athena.
    s3DataDir String
    Prefix for storing tables, if different from the connection's S3 staging directory.
    s3DataNaming String
    How to generate table paths in the S3 data directory.
    s3TmpTableDir String
    Prefix for storing temporary tables, if different from the connection's S3 data directory.
    sparkWorkGroup String
    Identifier of Athena Spark workgroup for running Python models.
    workGroup String
    Identifier of Athena workgroup.

    GlobalConnectionBigquery, GlobalConnectionBigqueryArgs

    AuthProviderX509CertUrl string
    Auth Provider X509 Cert URL for the Service Account
    AuthUri string
    Auth URI for the Service Account
    ClientEmail string
    Service Account email
    ClientId string
    Client ID of the Service Account
    ClientX509CertUrl string
    Client X509 Cert URL for the Service Account
    GcpProjectId string
    The GCP project ID to use for the connection
    PrivateKey string
    Private Key for the Service Account
    PrivateKeyId string
    Private Key ID for the Service Account
    TokenUri string
    Token URI for the Service Account
    ApplicationId string
    OAuth Client ID
    ApplicationSecret string
    OAuth Client Secret
    DataprocClusterName string
    Dataproc cluster name for PySpark workloads
    DataprocRegion string
    Google Cloud region for PySpark workloads on Dataproc
    ExecutionProject string
    Project to bill for query execution
    GcsBucket string
    URI for a Google Cloud Storage bucket to host Python code executed via Datapro
    ImpersonateServiceAccount string
    Service Account to impersonate when running queries
    JobCreationTimeoutSeconds int
    Maximum timeout for the job creation step
    JobRetryDeadlineSeconds int
    Total number of seconds to wait while retrying the same query
    Location string
    Location to create new Datasets in
    MaximumBytesBilled int
    Max number of bytes that can be billed for a given BigQuery query
    Priority string
    The priority with which to execute BigQuery queries (batch or interactive)
    Retries int
    Number of retries for queries
    Scopes List<string>
    OAuth scopes for the BigQuery connection
    TimeoutSeconds int
    Timeout in seconds for queries
    AuthProviderX509CertUrl string
    Auth Provider X509 Cert URL for the Service Account
    AuthUri string
    Auth URI for the Service Account
    ClientEmail string
    Service Account email
    ClientId string
    Client ID of the Service Account
    ClientX509CertUrl string
    Client X509 Cert URL for the Service Account
    GcpProjectId string
    The GCP project ID to use for the connection
    PrivateKey string
    Private Key for the Service Account
    PrivateKeyId string
    Private Key ID for the Service Account
    TokenUri string
    Token URI for the Service Account
    ApplicationId string
    OAuth Client ID
    ApplicationSecret string
    OAuth Client Secret
    DataprocClusterName string
    Dataproc cluster name for PySpark workloads
    DataprocRegion string
    Google Cloud region for PySpark workloads on Dataproc
    ExecutionProject string
    Project to bill for query execution
    GcsBucket string
    URI for a Google Cloud Storage bucket to host Python code executed via Datapro
    ImpersonateServiceAccount string
    Service Account to impersonate when running queries
    JobCreationTimeoutSeconds int
    Maximum timeout for the job creation step
    JobRetryDeadlineSeconds int
    Total number of seconds to wait while retrying the same query
    Location string
    Location to create new Datasets in
    MaximumBytesBilled int
    Max number of bytes that can be billed for a given BigQuery query
    Priority string
    The priority with which to execute BigQuery queries (batch or interactive)
    Retries int
    Number of retries for queries
    Scopes []string
    OAuth scopes for the BigQuery connection
    TimeoutSeconds int
    Timeout in seconds for queries
    authProviderX509CertUrl String
    Auth Provider X509 Cert URL for the Service Account
    authUri String
    Auth URI for the Service Account
    clientEmail String
    Service Account email
    clientId String
    Client ID of the Service Account
    clientX509CertUrl String
    Client X509 Cert URL for the Service Account
    gcpProjectId String
    The GCP project ID to use for the connection
    privateKey String
    Private Key for the Service Account
    privateKeyId String
    Private Key ID for the Service Account
    tokenUri String
    Token URI for the Service Account
    applicationId String
    OAuth Client ID
    applicationSecret String
    OAuth Client Secret
    dataprocClusterName String
    Dataproc cluster name for PySpark workloads
    dataprocRegion String
    Google Cloud region for PySpark workloads on Dataproc
    executionProject String
    Project to bill for query execution
    gcsBucket String
    URI for a Google Cloud Storage bucket to host Python code executed via Datapro
    impersonateServiceAccount String
    Service Account to impersonate when running queries
    jobCreationTimeoutSeconds Integer
    Maximum timeout for the job creation step
    jobRetryDeadlineSeconds Integer
    Total number of seconds to wait while retrying the same query
    location String
    Location to create new Datasets in
    maximumBytesBilled Integer
    Max number of bytes that can be billed for a given BigQuery query
    priority String
    The priority with which to execute BigQuery queries (batch or interactive)
    retries Integer
    Number of retries for queries
    scopes List<String>
    OAuth scopes for the BigQuery connection
    timeoutSeconds Integer
    Timeout in seconds for queries
    authProviderX509CertUrl string
    Auth Provider X509 Cert URL for the Service Account
    authUri string
    Auth URI for the Service Account
    clientEmail string
    Service Account email
    clientId string
    Client ID of the Service Account
    clientX509CertUrl string
    Client X509 Cert URL for the Service Account
    gcpProjectId string
    The GCP project ID to use for the connection
    privateKey string
    Private Key for the Service Account
    privateKeyId string
    Private Key ID for the Service Account
    tokenUri string
    Token URI for the Service Account
    applicationId string
    OAuth Client ID
    applicationSecret string
    OAuth Client Secret
    dataprocClusterName string
    Dataproc cluster name for PySpark workloads
    dataprocRegion string
    Google Cloud region for PySpark workloads on Dataproc
    executionProject string
    Project to bill for query execution
    gcsBucket string
    URI for a Google Cloud Storage bucket to host Python code executed via Datapro
    impersonateServiceAccount string
    Service Account to impersonate when running queries
    jobCreationTimeoutSeconds number
    Maximum timeout for the job creation step
    jobRetryDeadlineSeconds number
    Total number of seconds to wait while retrying the same query
    location string
    Location to create new Datasets in
    maximumBytesBilled number
    Max number of bytes that can be billed for a given BigQuery query
    priority string
    The priority with which to execute BigQuery queries (batch or interactive)
    retries number
    Number of retries for queries
    scopes string[]
    OAuth scopes for the BigQuery connection
    timeoutSeconds number
    Timeout in seconds for queries
    auth_provider_x509_cert_url str
    Auth Provider X509 Cert URL for the Service Account
    auth_uri str
    Auth URI for the Service Account
    client_email str
    Service Account email
    client_id str
    Client ID of the Service Account
    client_x509_cert_url str
    Client X509 Cert URL for the Service Account
    gcp_project_id str
    The GCP project ID to use for the connection
    private_key str
    Private Key for the Service Account
    private_key_id str
    Private Key ID for the Service Account
    token_uri str
    Token URI for the Service Account
    application_id str
    OAuth Client ID
    application_secret str
    OAuth Client Secret
    dataproc_cluster_name str
    Dataproc cluster name for PySpark workloads
    dataproc_region str
    Google Cloud region for PySpark workloads on Dataproc
    execution_project str
    Project to bill for query execution
    gcs_bucket str
    URI for a Google Cloud Storage bucket to host Python code executed via Datapro
    impersonate_service_account str
    Service Account to impersonate when running queries
    job_creation_timeout_seconds int
    Maximum timeout for the job creation step
    job_retry_deadline_seconds int
    Total number of seconds to wait while retrying the same query
    location str
    Location to create new Datasets in
    maximum_bytes_billed int
    Max number of bytes that can be billed for a given BigQuery query
    priority str
    The priority with which to execute BigQuery queries (batch or interactive)
    retries int
    Number of retries for queries
    scopes Sequence[str]
    OAuth scopes for the BigQuery connection
    timeout_seconds int
    Timeout in seconds for queries
    authProviderX509CertUrl String
    Auth Provider X509 Cert URL for the Service Account
    authUri String
    Auth URI for the Service Account
    clientEmail String
    Service Account email
    clientId String
    Client ID of the Service Account
    clientX509CertUrl String
    Client X509 Cert URL for the Service Account
    gcpProjectId String
    The GCP project ID to use for the connection
    privateKey String
    Private Key for the Service Account
    privateKeyId String
    Private Key ID for the Service Account
    tokenUri String
    Token URI for the Service Account
    applicationId String
    OAuth Client ID
    applicationSecret String
    OAuth Client Secret
    dataprocClusterName String
    Dataproc cluster name for PySpark workloads
    dataprocRegion String
    Google Cloud region for PySpark workloads on Dataproc
    executionProject String
    Project to bill for query execution
    gcsBucket String
    URI for a Google Cloud Storage bucket to host Python code executed via Datapro
    impersonateServiceAccount String
    Service Account to impersonate when running queries
    jobCreationTimeoutSeconds Number
    Maximum timeout for the job creation step
    jobRetryDeadlineSeconds Number
    Total number of seconds to wait while retrying the same query
    location String
    Location to create new Datasets in
    maximumBytesBilled Number
    Max number of bytes that can be billed for a given BigQuery query
    priority String
    The priority with which to execute BigQuery queries (batch or interactive)
    retries Number
    Number of retries for queries
    scopes List<String>
    OAuth scopes for the BigQuery connection
    timeoutSeconds Number
    Timeout in seconds for queries

    GlobalConnectionDatabricks, GlobalConnectionDatabricksArgs

    Host string
    The hostname of the Databricks cluster or SQL warehouse.
    HttpPath string
    The HTTP path of the Databricks cluster or SQL warehouse.
    Catalog string
    Catalog name if Unity Catalog is enabled in your Databricks workspace.
    ClientId string
    Required to enable Databricks OAuth authentication for IDE developers.
    ClientSecret string
    Required to enable Databricks OAuth authentication for IDE developers.
    Host string
    The hostname of the Databricks cluster or SQL warehouse.
    HttpPath string
    The HTTP path of the Databricks cluster or SQL warehouse.
    Catalog string
    Catalog name if Unity Catalog is enabled in your Databricks workspace.
    ClientId string
    Required to enable Databricks OAuth authentication for IDE developers.
    ClientSecret string
    Required to enable Databricks OAuth authentication for IDE developers.
    host String
    The hostname of the Databricks cluster or SQL warehouse.
    httpPath String
    The HTTP path of the Databricks cluster or SQL warehouse.
    catalog String
    Catalog name if Unity Catalog is enabled in your Databricks workspace.
    clientId String
    Required to enable Databricks OAuth authentication for IDE developers.
    clientSecret String
    Required to enable Databricks OAuth authentication for IDE developers.
    host string
    The hostname of the Databricks cluster or SQL warehouse.
    httpPath string
    The HTTP path of the Databricks cluster or SQL warehouse.
    catalog string
    Catalog name if Unity Catalog is enabled in your Databricks workspace.
    clientId string
    Required to enable Databricks OAuth authentication for IDE developers.
    clientSecret string
    Required to enable Databricks OAuth authentication for IDE developers.
    host str
    The hostname of the Databricks cluster or SQL warehouse.
    http_path str
    The HTTP path of the Databricks cluster or SQL warehouse.
    catalog str
    Catalog name if Unity Catalog is enabled in your Databricks workspace.
    client_id str
    Required to enable Databricks OAuth authentication for IDE developers.
    client_secret str
    Required to enable Databricks OAuth authentication for IDE developers.
    host String
    The hostname of the Databricks cluster or SQL warehouse.
    httpPath String
    The HTTP path of the Databricks cluster or SQL warehouse.
    catalog String
    Catalog name if Unity Catalog is enabled in your Databricks workspace.
    clientId String
    Required to enable Databricks OAuth authentication for IDE developers.
    clientSecret String
    Required to enable Databricks OAuth authentication for IDE developers.

    GlobalConnectionFabric, GlobalConnectionFabricArgs

    Database string
    The database to connect to for this connection.
    Server string
    The server hostname.
    LoginTimeout int
    The number of seconds used to establish a connection before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    Port int
    The port to connect to for this connection. Default=1433
    QueryTimeout int
    The number of seconds used to wait for a query before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    Retries int
    The number of automatic times to retry a query before failing. Defaults to 1. Queries with syntax errors will not be retried. This setting can be used to overcome intermittent network issues.
    Database string
    The database to connect to for this connection.
    Server string
    The server hostname.
    LoginTimeout int
    The number of seconds used to establish a connection before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    Port int
    The port to connect to for this connection. Default=1433
    QueryTimeout int
    The number of seconds used to wait for a query before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    Retries int
    The number of automatic times to retry a query before failing. Defaults to 1. Queries with syntax errors will not be retried. This setting can be used to overcome intermittent network issues.
    database String
    The database to connect to for this connection.
    server String
    The server hostname.
    loginTimeout Integer
    The number of seconds used to establish a connection before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    port Integer
    The port to connect to for this connection. Default=1433
    queryTimeout Integer
    The number of seconds used to wait for a query before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    retries Integer
    The number of automatic times to retry a query before failing. Defaults to 1. Queries with syntax errors will not be retried. This setting can be used to overcome intermittent network issues.
    database string
    The database to connect to for this connection.
    server string
    The server hostname.
    loginTimeout number
    The number of seconds used to establish a connection before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    port number
    The port to connect to for this connection. Default=1433
    queryTimeout number
    The number of seconds used to wait for a query before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    retries number
    The number of automatic times to retry a query before failing. Defaults to 1. Queries with syntax errors will not be retried. This setting can be used to overcome intermittent network issues.
    database str
    The database to connect to for this connection.
    server str
    The server hostname.
    login_timeout int
    The number of seconds used to establish a connection before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    port int
    The port to connect to for this connection. Default=1433
    query_timeout int
    The number of seconds used to wait for a query before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    retries int
    The number of automatic times to retry a query before failing. Defaults to 1. Queries with syntax errors will not be retried. This setting can be used to overcome intermittent network issues.
    database String
    The database to connect to for this connection.
    server String
    The server hostname.
    loginTimeout Number
    The number of seconds used to establish a connection before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    port Number
    The port to connect to for this connection. Default=1433
    queryTimeout Number
    The number of seconds used to wait for a query before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    retries Number
    The number of automatic times to retry a query before failing. Defaults to 1. Queries with syntax errors will not be retried. This setting can be used to overcome intermittent network issues.

    GlobalConnectionPostgres, GlobalConnectionPostgresArgs

    Dbname string
    The database name for this connection.
    Hostname string
    The hostname of the database.
    Port int
    The port to connect to for this connection. Default=5432
    SshTunnel Pulumi.DbtCloud.Inputs.GlobalConnectionPostgresSshTunnel
    PostgreSQL SSH Tunnel configuration
    Dbname string
    The database name for this connection.
    Hostname string
    The hostname of the database.
    Port int
    The port to connect to for this connection. Default=5432
    SshTunnel GlobalConnectionPostgresSshTunnel
    PostgreSQL SSH Tunnel configuration
    dbname String
    The database name for this connection.
    hostname String
    The hostname of the database.
    port Integer
    The port to connect to for this connection. Default=5432
    sshTunnel GlobalConnectionPostgresSshTunnel
    PostgreSQL SSH Tunnel configuration
    dbname string
    The database name for this connection.
    hostname string
    The hostname of the database.
    port number
    The port to connect to for this connection. Default=5432
    sshTunnel GlobalConnectionPostgresSshTunnel
    PostgreSQL SSH Tunnel configuration
    dbname str
    The database name for this connection.
    hostname str
    The hostname of the database.
    port int
    The port to connect to for this connection. Default=5432
    ssh_tunnel GlobalConnectionPostgresSshTunnel
    PostgreSQL SSH Tunnel configuration
    dbname String
    The database name for this connection.
    hostname String
    The hostname of the database.
    port Number
    The port to connect to for this connection. Default=5432
    sshTunnel Property Map
    PostgreSQL SSH Tunnel configuration

    GlobalConnectionPostgresSshTunnel, GlobalConnectionPostgresSshTunnelArgs

    Hostname string
    The hostname for the SSH tunnel.
    Port int
    The HTTP port for the SSH tunnel.
    Username string
    The username to use for the SSH tunnel.
    Id int
    The ID of the SSH tunnel connection.
    PublicKey string
    The SSH public key generated to allow connecting via SSH tunnel.
    Hostname string
    The hostname for the SSH tunnel.
    Port int
    The HTTP port for the SSH tunnel.
    Username string
    The username to use for the SSH tunnel.
    Id int
    The ID of the SSH tunnel connection.
    PublicKey string
    The SSH public key generated to allow connecting via SSH tunnel.
    hostname String
    The hostname for the SSH tunnel.
    port Integer
    The HTTP port for the SSH tunnel.
    username String
    The username to use for the SSH tunnel.
    id Integer
    The ID of the SSH tunnel connection.
    publicKey String
    The SSH public key generated to allow connecting via SSH tunnel.
    hostname string
    The hostname for the SSH tunnel.
    port number
    The HTTP port for the SSH tunnel.
    username string
    The username to use for the SSH tunnel.
    id number
    The ID of the SSH tunnel connection.
    publicKey string
    The SSH public key generated to allow connecting via SSH tunnel.
    hostname str
    The hostname for the SSH tunnel.
    port int
    The HTTP port for the SSH tunnel.
    username str
    The username to use for the SSH tunnel.
    id int
    The ID of the SSH tunnel connection.
    public_key str
    The SSH public key generated to allow connecting via SSH tunnel.
    hostname String
    The hostname for the SSH tunnel.
    port Number
    The HTTP port for the SSH tunnel.
    username String
    The username to use for the SSH tunnel.
    id Number
    The ID of the SSH tunnel connection.
    publicKey String
    The SSH public key generated to allow connecting via SSH tunnel.

    GlobalConnectionRedshift, GlobalConnectionRedshiftArgs

    Dbname string
    The database name for this connection.
    Hostname string
    The hostname of the data warehouse.
    Port int
    The port to connect to for this connection. Default=5432
    SshTunnel Pulumi.DbtCloud.Inputs.GlobalConnectionRedshiftSshTunnel
    Redshift SSH Tunnel configuration
    Dbname string
    The database name for this connection.
    Hostname string
    The hostname of the data warehouse.
    Port int
    The port to connect to for this connection. Default=5432
    SshTunnel GlobalConnectionRedshiftSshTunnel
    Redshift SSH Tunnel configuration
    dbname String
    The database name for this connection.
    hostname String
    The hostname of the data warehouse.
    port Integer
    The port to connect to for this connection. Default=5432
    sshTunnel GlobalConnectionRedshiftSshTunnel
    Redshift SSH Tunnel configuration
    dbname string
    The database name for this connection.
    hostname string
    The hostname of the data warehouse.
    port number
    The port to connect to for this connection. Default=5432
    sshTunnel GlobalConnectionRedshiftSshTunnel
    Redshift SSH Tunnel configuration
    dbname str
    The database name for this connection.
    hostname str
    The hostname of the data warehouse.
    port int
    The port to connect to for this connection. Default=5432
    ssh_tunnel GlobalConnectionRedshiftSshTunnel
    Redshift SSH Tunnel configuration
    dbname String
    The database name for this connection.
    hostname String
    The hostname of the data warehouse.
    port Number
    The port to connect to for this connection. Default=5432
    sshTunnel Property Map
    Redshift SSH Tunnel configuration

    GlobalConnectionRedshiftSshTunnel, GlobalConnectionRedshiftSshTunnelArgs

    Hostname string
    The hostname for the SSH tunnel.
    Port int
    The HTTP port for the SSH tunnel.
    Username string
    The username to use for the SSH tunnel.
    Id int
    The ID of the SSH tunnel connection.
    PublicKey string
    The SSH public key generated to allow connecting via SSH tunnel.
    Hostname string
    The hostname for the SSH tunnel.
    Port int
    The HTTP port for the SSH tunnel.
    Username string
    The username to use for the SSH tunnel.
    Id int
    The ID of the SSH tunnel connection.
    PublicKey string
    The SSH public key generated to allow connecting via SSH tunnel.
    hostname String
    The hostname for the SSH tunnel.
    port Integer
    The HTTP port for the SSH tunnel.
    username String
    The username to use for the SSH tunnel.
    id Integer
    The ID of the SSH tunnel connection.
    publicKey String
    The SSH public key generated to allow connecting via SSH tunnel.
    hostname string
    The hostname for the SSH tunnel.
    port number
    The HTTP port for the SSH tunnel.
    username string
    The username to use for the SSH tunnel.
    id number
    The ID of the SSH tunnel connection.
    publicKey string
    The SSH public key generated to allow connecting via SSH tunnel.
    hostname str
    The hostname for the SSH tunnel.
    port int
    The HTTP port for the SSH tunnel.
    username str
    The username to use for the SSH tunnel.
    id int
    The ID of the SSH tunnel connection.
    public_key str
    The SSH public key generated to allow connecting via SSH tunnel.
    hostname String
    The hostname for the SSH tunnel.
    port Number
    The HTTP port for the SSH tunnel.
    username String
    The username to use for the SSH tunnel.
    id Number
    The ID of the SSH tunnel connection.
    publicKey String
    The SSH public key generated to allow connecting via SSH tunnel.

    GlobalConnectionSnowflake, GlobalConnectionSnowflakeArgs

    Account string
    The Snowflake account name
    Database string
    The default database for the connection
    Warehouse string
    The default Snowflake Warehouse to use for the connection
    AllowSso bool
    Whether to allow Snowflake OAuth for the connection. If true, the oauth_client_id and oauth_client_secret fields must be set
    ClientSessionKeepAlive bool
    If true, the snowflake client will keep connections for longer than the default 4 hours. This is helpful when particularly long-running queries are executing (> 4 hours)
    OauthClientId string
    OAuth Client ID. Required to allow OAuth between dbt Cloud and Snowflake
    OauthClientSecret string
    OAuth Client Secret. Required to allow OAuth between dbt Cloud and Snowflake
    Role string
    The Snowflake role to use when running queries on the connection
    Account string
    The Snowflake account name
    Database string
    The default database for the connection
    Warehouse string
    The default Snowflake Warehouse to use for the connection
    AllowSso bool
    Whether to allow Snowflake OAuth for the connection. If true, the oauth_client_id and oauth_client_secret fields must be set
    ClientSessionKeepAlive bool
    If true, the snowflake client will keep connections for longer than the default 4 hours. This is helpful when particularly long-running queries are executing (> 4 hours)
    OauthClientId string
    OAuth Client ID. Required to allow OAuth between dbt Cloud and Snowflake
    OauthClientSecret string
    OAuth Client Secret. Required to allow OAuth between dbt Cloud and Snowflake
    Role string
    The Snowflake role to use when running queries on the connection
    account String
    The Snowflake account name
    database String
    The default database for the connection
    warehouse String
    The default Snowflake Warehouse to use for the connection
    allowSso Boolean
    Whether to allow Snowflake OAuth for the connection. If true, the oauth_client_id and oauth_client_secret fields must be set
    clientSessionKeepAlive Boolean
    If true, the snowflake client will keep connections for longer than the default 4 hours. This is helpful when particularly long-running queries are executing (> 4 hours)
    oauthClientId String
    OAuth Client ID. Required to allow OAuth between dbt Cloud and Snowflake
    oauthClientSecret String
    OAuth Client Secret. Required to allow OAuth between dbt Cloud and Snowflake
    role String
    The Snowflake role to use when running queries on the connection
    account string
    The Snowflake account name
    database string
    The default database for the connection
    warehouse string
    The default Snowflake Warehouse to use for the connection
    allowSso boolean
    Whether to allow Snowflake OAuth for the connection. If true, the oauth_client_id and oauth_client_secret fields must be set
    clientSessionKeepAlive boolean
    If true, the snowflake client will keep connections for longer than the default 4 hours. This is helpful when particularly long-running queries are executing (> 4 hours)
    oauthClientId string
    OAuth Client ID. Required to allow OAuth between dbt Cloud and Snowflake
    oauthClientSecret string
    OAuth Client Secret. Required to allow OAuth between dbt Cloud and Snowflake
    role string
    The Snowflake role to use when running queries on the connection
    account str
    The Snowflake account name
    database str
    The default database for the connection
    warehouse str
    The default Snowflake Warehouse to use for the connection
    allow_sso bool
    Whether to allow Snowflake OAuth for the connection. If true, the oauth_client_id and oauth_client_secret fields must be set
    client_session_keep_alive bool
    If true, the snowflake client will keep connections for longer than the default 4 hours. This is helpful when particularly long-running queries are executing (> 4 hours)
    oauth_client_id str
    OAuth Client ID. Required to allow OAuth between dbt Cloud and Snowflake
    oauth_client_secret str
    OAuth Client Secret. Required to allow OAuth between dbt Cloud and Snowflake
    role str
    The Snowflake role to use when running queries on the connection
    account String
    The Snowflake account name
    database String
    The default database for the connection
    warehouse String
    The default Snowflake Warehouse to use for the connection
    allowSso Boolean
    Whether to allow Snowflake OAuth for the connection. If true, the oauth_client_id and oauth_client_secret fields must be set
    clientSessionKeepAlive Boolean
    If true, the snowflake client will keep connections for longer than the default 4 hours. This is helpful when particularly long-running queries are executing (> 4 hours)
    oauthClientId String
    OAuth Client ID. Required to allow OAuth between dbt Cloud and Snowflake
    oauthClientSecret String
    OAuth Client Secret. Required to allow OAuth between dbt Cloud and Snowflake
    role String
    The Snowflake role to use when running queries on the connection

    GlobalConnectionStarburst, GlobalConnectionStarburstArgs

    Host string
    The hostname of the account to connect to.
    Method string
    The authentication method. Only LDAP for now.
    Port int
    The port to connect to for this connection. Default=443
    Host string
    The hostname of the account to connect to.
    Method string
    The authentication method. Only LDAP for now.
    Port int
    The port to connect to for this connection. Default=443
    host String
    The hostname of the account to connect to.
    method String
    The authentication method. Only LDAP for now.
    port Integer
    The port to connect to for this connection. Default=443
    host string
    The hostname of the account to connect to.
    method string
    The authentication method. Only LDAP for now.
    port number
    The port to connect to for this connection. Default=443
    host str
    The hostname of the account to connect to.
    method str
    The authentication method. Only LDAP for now.
    port int
    The port to connect to for this connection. Default=443
    host String
    The hostname of the account to connect to.
    method String
    The authentication method. Only LDAP for now.
    port Number
    The port to connect to for this connection. Default=443

    GlobalConnectionSynapse, GlobalConnectionSynapseArgs

    Database string
    The database to connect to for this connection.
    Host string
    The server hostname.
    LoginTimeout int
    The number of seconds used to establish a connection before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    Port int
    The port to connect to for this connection. Default=1433
    QueryTimeout int
    The number of seconds used to wait for a query before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    Retries int
    The number of automatic times to retry a query before failing. Defaults to 1. Queries with syntax errors will not be retried. This setting can be used to overcome intermittent network issues.
    Database string
    The database to connect to for this connection.
    Host string
    The server hostname.
    LoginTimeout int
    The number of seconds used to establish a connection before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    Port int
    The port to connect to for this connection. Default=1433
    QueryTimeout int
    The number of seconds used to wait for a query before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    Retries int
    The number of automatic times to retry a query before failing. Defaults to 1. Queries with syntax errors will not be retried. This setting can be used to overcome intermittent network issues.
    database String
    The database to connect to for this connection.
    host String
    The server hostname.
    loginTimeout Integer
    The number of seconds used to establish a connection before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    port Integer
    The port to connect to for this connection. Default=1433
    queryTimeout Integer
    The number of seconds used to wait for a query before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    retries Integer
    The number of automatic times to retry a query before failing. Defaults to 1. Queries with syntax errors will not be retried. This setting can be used to overcome intermittent network issues.
    database string
    The database to connect to for this connection.
    host string
    The server hostname.
    loginTimeout number
    The number of seconds used to establish a connection before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    port number
    The port to connect to for this connection. Default=1433
    queryTimeout number
    The number of seconds used to wait for a query before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    retries number
    The number of automatic times to retry a query before failing. Defaults to 1. Queries with syntax errors will not be retried. This setting can be used to overcome intermittent network issues.
    database str
    The database to connect to for this connection.
    host str
    The server hostname.
    login_timeout int
    The number of seconds used to establish a connection before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    port int
    The port to connect to for this connection. Default=1433
    query_timeout int
    The number of seconds used to wait for a query before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    retries int
    The number of automatic times to retry a query before failing. Defaults to 1. Queries with syntax errors will not be retried. This setting can be used to overcome intermittent network issues.
    database String
    The database to connect to for this connection.
    host String
    The server hostname.
    loginTimeout Number
    The number of seconds used to establish a connection before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    port Number
    The port to connect to for this connection. Default=1433
    queryTimeout Number
    The number of seconds used to wait for a query before failing. Defaults to 0, which means that the timeout is disabled or uses the default system settings.
    retries Number
    The number of automatic times to retry a query before failing. Defaults to 1. Queries with syntax errors will not be retried. This setting can be used to overcome intermittent network issues.

    Import

    A project-scoped connection can be imported as a global connection by specifying the connection ID

    Migrating from project-scoped connections to global connections could be done by:

    1. Adding the config for the global connection and importing it (see below)

    2. Removing the project-scoped connection from the config AND from the state

      • CAREFUL: If the connection is removed from the config but not the state, it will be destroyed on the next apply

    using import blocks (requires Terraform >= 1.5)

    import {

    to = dbtcloud_global_connection.my_connection

    id = “connection_id”

    }

    import {

    to = dbtcloud_global_connection.my_connection

    id = “1234”

    }

    using the older import command

    $ pulumi import dbtcloud:index/globalConnection:GlobalConnection my_connection "connection_id"
    
    $ pulumi import dbtcloud:index/globalConnection:GlobalConnection my_connection 1234
    

    To learn more about importing existing cloud resources, see Importing resources.

    Package Details

    Repository
    dbtcloud pulumi/pulumi-dbtcloud
    License
    Apache-2.0
    Notes
    This Pulumi package is based on the dbtcloud Terraform Provider.
    dbtcloud logo
    dbt Cloud v0.1.20 published on Friday, Sep 27, 2024 by Pulumi