Google Cloud Native is in preview. Google Cloud Classic is fully supported.
Google Cloud Native v0.32.0 published on Wednesday, Nov 29, 2023 by Pulumi
google-native.bigquery/v2.getRoutine
Explore with Pulumi AI
Google Cloud Native is in preview. Google Cloud Classic is fully supported.
Google Cloud Native v0.32.0 published on Wednesday, Nov 29, 2023 by Pulumi
Gets the specified routine resource by routine ID.
Using getRoutine
Two invocation forms are available. The direct form accepts plain arguments and either blocks until the result value is available, or returns a Promise-wrapped result. The output form accepts Input-wrapped arguments and returns an Output-wrapped result.
function getRoutine(args: GetRoutineArgs, opts?: InvokeOptions): Promise<GetRoutineResult>
function getRoutineOutput(args: GetRoutineOutputArgs, opts?: InvokeOptions): Output<GetRoutineResult>
def get_routine(dataset_id: Optional[str] = None,
project: Optional[str] = None,
read_mask: Optional[str] = None,
routine_id: Optional[str] = None,
opts: Optional[InvokeOptions] = None) -> GetRoutineResult
def get_routine_output(dataset_id: Optional[pulumi.Input[str]] = None,
project: Optional[pulumi.Input[str]] = None,
read_mask: Optional[pulumi.Input[str]] = None,
routine_id: Optional[pulumi.Input[str]] = None,
opts: Optional[InvokeOptions] = None) -> Output[GetRoutineResult]
func LookupRoutine(ctx *Context, args *LookupRoutineArgs, opts ...InvokeOption) (*LookupRoutineResult, error)
func LookupRoutineOutput(ctx *Context, args *LookupRoutineOutputArgs, opts ...InvokeOption) LookupRoutineResultOutput
> Note: This function is named LookupRoutine
in the Go SDK.
public static class GetRoutine
{
public static Task<GetRoutineResult> InvokeAsync(GetRoutineArgs args, InvokeOptions? opts = null)
public static Output<GetRoutineResult> Invoke(GetRoutineInvokeArgs args, InvokeOptions? opts = null)
}
public static CompletableFuture<GetRoutineResult> getRoutine(GetRoutineArgs args, InvokeOptions options)
// Output-based functions aren't available in Java yet
fn::invoke:
function: google-native:bigquery/v2:getRoutine
arguments:
# arguments dictionary
The following arguments are supported:
- dataset_
id str - routine_
id str - project str
- read_
mask str
getRoutine Result
The following output properties are available:
- Arguments
List<Pulumi.
Google Native. Big Query. V2. Outputs. Argument Response> - Optional.
- Creation
Time string - The time when this routine was created, in milliseconds since the epoch.
- Data
Governance stringType - Optional. If set to
DATA_MASKING
, the function is validated and made available as a masking function. For more information, see Create custom masking routines. - Definition
Body string - The body of the routine. For functions, this is the expression in the AS clause. If language=SQL, it is the substring inside (but excluding) the parentheses. For example, for the function created with the following statement:
CREATE FUNCTION JoinLines(x string, y string) as (concat(x, "\n", y))
The definition_body isconcat(x, "\n", y)
(\n is not replaced with linebreak). If language=JAVASCRIPT, it is the evaluated string in the AS clause. For example, for the function created with the following statement:CREATE FUNCTION f() RETURNS STRING LANGUAGE js AS 'return "\n";\n'
The definition_body isreturn "\n";\n
Note that both \n are replaced with linebreaks. - Description string
- Optional. The description of the routine, if defined.
- Determinism
Level string - Optional. The determinism level of the JavaScript UDF, if defined.
- Etag string
- A hash of this resource.
- Imported
Libraries List<string> - Optional. If language = "JAVASCRIPT", this field stores the path of the imported JAVASCRIPT libraries.
- Language string
- Optional. Defaults to "SQL" if remote_function_options field is absent, not set otherwise.
- Last
Modified stringTime - The time when this routine was last modified, in milliseconds since the epoch.
- Remote
Function Pulumi.Options Google Native. Big Query. V2. Outputs. Remote Function Options Response - Optional. Remote function specific options.
- Return
Table Pulumi.Type Google Native. Big Query. V2. Outputs. Standard Sql Table Type Response - Optional. Can be set only if routine_type = "TABLE_VALUED_FUNCTION". If absent, the return table type is inferred from definition_body at query time in each query that references this routine. If present, then the columns in the evaluated table result will be cast to match the column types specified in return table type, at query time.
- Return
Type Pulumi.Google Native. Big Query. V2. Outputs. Standard Sql Data Type Response - Optional if language = "SQL"; required otherwise. Cannot be set if routine_type = "TABLE_VALUED_FUNCTION". If absent, the return type is inferred from definition_body at query time in each query that references this routine. If present, then the evaluated result will be cast to the specified returned type at query time. For example, for the functions created with the following statements: *
CREATE FUNCTION Add(x FLOAT64, y FLOAT64) RETURNS FLOAT64 AS (x + y);
*CREATE FUNCTION Increment(x FLOAT64) AS (Add(x, 1));
*CREATE FUNCTION Decrement(x FLOAT64) RETURNS FLOAT64 AS (Add(x, -1));
The return_type is{type_kind: "FLOAT64"}
forAdd
andDecrement
, and is absent forIncrement
(inferred as FLOAT64 at query time). Suppose the functionAdd
is replaced byCREATE OR REPLACE FUNCTION Add(x INT64, y INT64) AS (x + y);
Then the inferred return type ofIncrement
is automatically changed to INT64 at query time, while the return type ofDecrement
remains FLOAT64. - Routine
Reference Pulumi.Google Native. Big Query. V2. Outputs. Routine Reference Response - Reference describing the ID of this routine.
- Routine
Type string - The type of routine.
- Security
Mode string - Optional. The security mode of the routine, if defined. If not defined, the security mode is automatically determined from the routine's configuration.
- Spark
Options Pulumi.Google Native. Big Query. V2. Outputs. Spark Options Response - Optional. Spark specific options.
- Strict
Mode bool - Optional. Can be set for procedures only. If true (default), the definition body will be validated in the creation and the updates of the procedure. For procedures with an argument of ANY TYPE, the definition body validtion is not supported at creation/update time, and thus this field must be set to false explicitly.
- Arguments
[]Argument
Response - Optional.
- Creation
Time string - The time when this routine was created, in milliseconds since the epoch.
- Data
Governance stringType - Optional. If set to
DATA_MASKING
, the function is validated and made available as a masking function. For more information, see Create custom masking routines. - Definition
Body string - The body of the routine. For functions, this is the expression in the AS clause. If language=SQL, it is the substring inside (but excluding) the parentheses. For example, for the function created with the following statement:
CREATE FUNCTION JoinLines(x string, y string) as (concat(x, "\n", y))
The definition_body isconcat(x, "\n", y)
(\n is not replaced with linebreak). If language=JAVASCRIPT, it is the evaluated string in the AS clause. For example, for the function created with the following statement:CREATE FUNCTION f() RETURNS STRING LANGUAGE js AS 'return "\n";\n'
The definition_body isreturn "\n";\n
Note that both \n are replaced with linebreaks. - Description string
- Optional. The description of the routine, if defined.
- Determinism
Level string - Optional. The determinism level of the JavaScript UDF, if defined.
- Etag string
- A hash of this resource.
- Imported
Libraries []string - Optional. If language = "JAVASCRIPT", this field stores the path of the imported JAVASCRIPT libraries.
- Language string
- Optional. Defaults to "SQL" if remote_function_options field is absent, not set otherwise.
- Last
Modified stringTime - The time when this routine was last modified, in milliseconds since the epoch.
- Remote
Function RemoteOptions Function Options Response - Optional. Remote function specific options.
- Return
Table StandardType Sql Table Type Response - Optional. Can be set only if routine_type = "TABLE_VALUED_FUNCTION". If absent, the return table type is inferred from definition_body at query time in each query that references this routine. If present, then the columns in the evaluated table result will be cast to match the column types specified in return table type, at query time.
- Return
Type StandardSql Data Type Response - Optional if language = "SQL"; required otherwise. Cannot be set if routine_type = "TABLE_VALUED_FUNCTION". If absent, the return type is inferred from definition_body at query time in each query that references this routine. If present, then the evaluated result will be cast to the specified returned type at query time. For example, for the functions created with the following statements: *
CREATE FUNCTION Add(x FLOAT64, y FLOAT64) RETURNS FLOAT64 AS (x + y);
*CREATE FUNCTION Increment(x FLOAT64) AS (Add(x, 1));
*CREATE FUNCTION Decrement(x FLOAT64) RETURNS FLOAT64 AS (Add(x, -1));
The return_type is{type_kind: "FLOAT64"}
forAdd
andDecrement
, and is absent forIncrement
(inferred as FLOAT64 at query time). Suppose the functionAdd
is replaced byCREATE OR REPLACE FUNCTION Add(x INT64, y INT64) AS (x + y);
Then the inferred return type ofIncrement
is automatically changed to INT64 at query time, while the return type ofDecrement
remains FLOAT64. - Routine
Reference RoutineReference Response - Reference describing the ID of this routine.
- Routine
Type string - The type of routine.
- Security
Mode string - Optional. The security mode of the routine, if defined. If not defined, the security mode is automatically determined from the routine's configuration.
- Spark
Options SparkOptions Response - Optional. Spark specific options.
- Strict
Mode bool - Optional. Can be set for procedures only. If true (default), the definition body will be validated in the creation and the updates of the procedure. For procedures with an argument of ANY TYPE, the definition body validtion is not supported at creation/update time, and thus this field must be set to false explicitly.
- arguments
List<Argument
Response> - Optional.
- creation
Time String - The time when this routine was created, in milliseconds since the epoch.
- data
Governance StringType - Optional. If set to
DATA_MASKING
, the function is validated and made available as a masking function. For more information, see Create custom masking routines. - definition
Body String - The body of the routine. For functions, this is the expression in the AS clause. If language=SQL, it is the substring inside (but excluding) the parentheses. For example, for the function created with the following statement:
CREATE FUNCTION JoinLines(x string, y string) as (concat(x, "\n", y))
The definition_body isconcat(x, "\n", y)
(\n is not replaced with linebreak). If language=JAVASCRIPT, it is the evaluated string in the AS clause. For example, for the function created with the following statement:CREATE FUNCTION f() RETURNS STRING LANGUAGE js AS 'return "\n";\n'
The definition_body isreturn "\n";\n
Note that both \n are replaced with linebreaks. - description String
- Optional. The description of the routine, if defined.
- determinism
Level String - Optional. The determinism level of the JavaScript UDF, if defined.
- etag String
- A hash of this resource.
- imported
Libraries List<String> - Optional. If language = "JAVASCRIPT", this field stores the path of the imported JAVASCRIPT libraries.
- language String
- Optional. Defaults to "SQL" if remote_function_options field is absent, not set otherwise.
- last
Modified StringTime - The time when this routine was last modified, in milliseconds since the epoch.
- remote
Function RemoteOptions Function Options Response - Optional. Remote function specific options.
- return
Table StandardType Sql Table Type Response - Optional. Can be set only if routine_type = "TABLE_VALUED_FUNCTION". If absent, the return table type is inferred from definition_body at query time in each query that references this routine. If present, then the columns in the evaluated table result will be cast to match the column types specified in return table type, at query time.
- return
Type StandardSql Data Type Response - Optional if language = "SQL"; required otherwise. Cannot be set if routine_type = "TABLE_VALUED_FUNCTION". If absent, the return type is inferred from definition_body at query time in each query that references this routine. If present, then the evaluated result will be cast to the specified returned type at query time. For example, for the functions created with the following statements: *
CREATE FUNCTION Add(x FLOAT64, y FLOAT64) RETURNS FLOAT64 AS (x + y);
*CREATE FUNCTION Increment(x FLOAT64) AS (Add(x, 1));
*CREATE FUNCTION Decrement(x FLOAT64) RETURNS FLOAT64 AS (Add(x, -1));
The return_type is{type_kind: "FLOAT64"}
forAdd
andDecrement
, and is absent forIncrement
(inferred as FLOAT64 at query time). Suppose the functionAdd
is replaced byCREATE OR REPLACE FUNCTION Add(x INT64, y INT64) AS (x + y);
Then the inferred return type ofIncrement
is automatically changed to INT64 at query time, while the return type ofDecrement
remains FLOAT64. - routine
Reference RoutineReference Response - Reference describing the ID of this routine.
- routine
Type String - The type of routine.
- security
Mode String - Optional. The security mode of the routine, if defined. If not defined, the security mode is automatically determined from the routine's configuration.
- spark
Options SparkOptions Response - Optional. Spark specific options.
- strict
Mode Boolean - Optional. Can be set for procedures only. If true (default), the definition body will be validated in the creation and the updates of the procedure. For procedures with an argument of ANY TYPE, the definition body validtion is not supported at creation/update time, and thus this field must be set to false explicitly.
- arguments
Argument
Response[] - Optional.
- creation
Time string - The time when this routine was created, in milliseconds since the epoch.
- data
Governance stringType - Optional. If set to
DATA_MASKING
, the function is validated and made available as a masking function. For more information, see Create custom masking routines. - definition
Body string - The body of the routine. For functions, this is the expression in the AS clause. If language=SQL, it is the substring inside (but excluding) the parentheses. For example, for the function created with the following statement:
CREATE FUNCTION JoinLines(x string, y string) as (concat(x, "\n", y))
The definition_body isconcat(x, "\n", y)
(\n is not replaced with linebreak). If language=JAVASCRIPT, it is the evaluated string in the AS clause. For example, for the function created with the following statement:CREATE FUNCTION f() RETURNS STRING LANGUAGE js AS 'return "\n";\n'
The definition_body isreturn "\n";\n
Note that both \n are replaced with linebreaks. - description string
- Optional. The description of the routine, if defined.
- determinism
Level string - Optional. The determinism level of the JavaScript UDF, if defined.
- etag string
- A hash of this resource.
- imported
Libraries string[] - Optional. If language = "JAVASCRIPT", this field stores the path of the imported JAVASCRIPT libraries.
- language string
- Optional. Defaults to "SQL" if remote_function_options field is absent, not set otherwise.
- last
Modified stringTime - The time when this routine was last modified, in milliseconds since the epoch.
- remote
Function RemoteOptions Function Options Response - Optional. Remote function specific options.
- return
Table StandardType Sql Table Type Response - Optional. Can be set only if routine_type = "TABLE_VALUED_FUNCTION". If absent, the return table type is inferred from definition_body at query time in each query that references this routine. If present, then the columns in the evaluated table result will be cast to match the column types specified in return table type, at query time.
- return
Type StandardSql Data Type Response - Optional if language = "SQL"; required otherwise. Cannot be set if routine_type = "TABLE_VALUED_FUNCTION". If absent, the return type is inferred from definition_body at query time in each query that references this routine. If present, then the evaluated result will be cast to the specified returned type at query time. For example, for the functions created with the following statements: *
CREATE FUNCTION Add(x FLOAT64, y FLOAT64) RETURNS FLOAT64 AS (x + y);
*CREATE FUNCTION Increment(x FLOAT64) AS (Add(x, 1));
*CREATE FUNCTION Decrement(x FLOAT64) RETURNS FLOAT64 AS (Add(x, -1));
The return_type is{type_kind: "FLOAT64"}
forAdd
andDecrement
, and is absent forIncrement
(inferred as FLOAT64 at query time). Suppose the functionAdd
is replaced byCREATE OR REPLACE FUNCTION Add(x INT64, y INT64) AS (x + y);
Then the inferred return type ofIncrement
is automatically changed to INT64 at query time, while the return type ofDecrement
remains FLOAT64. - routine
Reference RoutineReference Response - Reference describing the ID of this routine.
- routine
Type string - The type of routine.
- security
Mode string - Optional. The security mode of the routine, if defined. If not defined, the security mode is automatically determined from the routine's configuration.
- spark
Options SparkOptions Response - Optional. Spark specific options.
- strict
Mode boolean - Optional. Can be set for procedures only. If true (default), the definition body will be validated in the creation and the updates of the procedure. For procedures with an argument of ANY TYPE, the definition body validtion is not supported at creation/update time, and thus this field must be set to false explicitly.
- arguments
Sequence[Argument
Response] - Optional.
- creation_
time str - The time when this routine was created, in milliseconds since the epoch.
- data_
governance_ strtype - Optional. If set to
DATA_MASKING
, the function is validated and made available as a masking function. For more information, see Create custom masking routines. - definition_
body str - The body of the routine. For functions, this is the expression in the AS clause. If language=SQL, it is the substring inside (but excluding) the parentheses. For example, for the function created with the following statement:
CREATE FUNCTION JoinLines(x string, y string) as (concat(x, "\n", y))
The definition_body isconcat(x, "\n", y)
(\n is not replaced with linebreak). If language=JAVASCRIPT, it is the evaluated string in the AS clause. For example, for the function created with the following statement:CREATE FUNCTION f() RETURNS STRING LANGUAGE js AS 'return "\n";\n'
The definition_body isreturn "\n";\n
Note that both \n are replaced with linebreaks. - description str
- Optional. The description of the routine, if defined.
- determinism_
level str - Optional. The determinism level of the JavaScript UDF, if defined.
- etag str
- A hash of this resource.
- imported_
libraries Sequence[str] - Optional. If language = "JAVASCRIPT", this field stores the path of the imported JAVASCRIPT libraries.
- language str
- Optional. Defaults to "SQL" if remote_function_options field is absent, not set otherwise.
- last_
modified_ strtime - The time when this routine was last modified, in milliseconds since the epoch.
- remote_
function_ Remoteoptions Function Options Response - Optional. Remote function specific options.
- return_
table_ Standardtype Sql Table Type Response - Optional. Can be set only if routine_type = "TABLE_VALUED_FUNCTION". If absent, the return table type is inferred from definition_body at query time in each query that references this routine. If present, then the columns in the evaluated table result will be cast to match the column types specified in return table type, at query time.
- return_
type StandardSql Data Type Response - Optional if language = "SQL"; required otherwise. Cannot be set if routine_type = "TABLE_VALUED_FUNCTION". If absent, the return type is inferred from definition_body at query time in each query that references this routine. If present, then the evaluated result will be cast to the specified returned type at query time. For example, for the functions created with the following statements: *
CREATE FUNCTION Add(x FLOAT64, y FLOAT64) RETURNS FLOAT64 AS (x + y);
*CREATE FUNCTION Increment(x FLOAT64) AS (Add(x, 1));
*CREATE FUNCTION Decrement(x FLOAT64) RETURNS FLOAT64 AS (Add(x, -1));
The return_type is{type_kind: "FLOAT64"}
forAdd
andDecrement
, and is absent forIncrement
(inferred as FLOAT64 at query time). Suppose the functionAdd
is replaced byCREATE OR REPLACE FUNCTION Add(x INT64, y INT64) AS (x + y);
Then the inferred return type ofIncrement
is automatically changed to INT64 at query time, while the return type ofDecrement
remains FLOAT64. - routine_
reference RoutineReference Response - Reference describing the ID of this routine.
- routine_
type str - The type of routine.
- security_
mode str - Optional. The security mode of the routine, if defined. If not defined, the security mode is automatically determined from the routine's configuration.
- spark_
options SparkOptions Response - Optional. Spark specific options.
- strict_
mode bool - Optional. Can be set for procedures only. If true (default), the definition body will be validated in the creation and the updates of the procedure. For procedures with an argument of ANY TYPE, the definition body validtion is not supported at creation/update time, and thus this field must be set to false explicitly.
- arguments List<Property Map>
- Optional.
- creation
Time String - The time when this routine was created, in milliseconds since the epoch.
- data
Governance StringType - Optional. If set to
DATA_MASKING
, the function is validated and made available as a masking function. For more information, see Create custom masking routines. - definition
Body String - The body of the routine. For functions, this is the expression in the AS clause. If language=SQL, it is the substring inside (but excluding) the parentheses. For example, for the function created with the following statement:
CREATE FUNCTION JoinLines(x string, y string) as (concat(x, "\n", y))
The definition_body isconcat(x, "\n", y)
(\n is not replaced with linebreak). If language=JAVASCRIPT, it is the evaluated string in the AS clause. For example, for the function created with the following statement:CREATE FUNCTION f() RETURNS STRING LANGUAGE js AS 'return "\n";\n'
The definition_body isreturn "\n";\n
Note that both \n are replaced with linebreaks. - description String
- Optional. The description of the routine, if defined.
- determinism
Level String - Optional. The determinism level of the JavaScript UDF, if defined.
- etag String
- A hash of this resource.
- imported
Libraries List<String> - Optional. If language = "JAVASCRIPT", this field stores the path of the imported JAVASCRIPT libraries.
- language String
- Optional. Defaults to "SQL" if remote_function_options field is absent, not set otherwise.
- last
Modified StringTime - The time when this routine was last modified, in milliseconds since the epoch.
- remote
Function Property MapOptions - Optional. Remote function specific options.
- return
Table Property MapType - Optional. Can be set only if routine_type = "TABLE_VALUED_FUNCTION". If absent, the return table type is inferred from definition_body at query time in each query that references this routine. If present, then the columns in the evaluated table result will be cast to match the column types specified in return table type, at query time.
- return
Type Property Map - Optional if language = "SQL"; required otherwise. Cannot be set if routine_type = "TABLE_VALUED_FUNCTION". If absent, the return type is inferred from definition_body at query time in each query that references this routine. If present, then the evaluated result will be cast to the specified returned type at query time. For example, for the functions created with the following statements: *
CREATE FUNCTION Add(x FLOAT64, y FLOAT64) RETURNS FLOAT64 AS (x + y);
*CREATE FUNCTION Increment(x FLOAT64) AS (Add(x, 1));
*CREATE FUNCTION Decrement(x FLOAT64) RETURNS FLOAT64 AS (Add(x, -1));
The return_type is{type_kind: "FLOAT64"}
forAdd
andDecrement
, and is absent forIncrement
(inferred as FLOAT64 at query time). Suppose the functionAdd
is replaced byCREATE OR REPLACE FUNCTION Add(x INT64, y INT64) AS (x + y);
Then the inferred return type ofIncrement
is automatically changed to INT64 at query time, while the return type ofDecrement
remains FLOAT64. - routine
Reference Property Map - Reference describing the ID of this routine.
- routine
Type String - The type of routine.
- security
Mode String - Optional. The security mode of the routine, if defined. If not defined, the security mode is automatically determined from the routine's configuration.
- spark
Options Property Map - Optional. Spark specific options.
- strict
Mode Boolean - Optional. Can be set for procedures only. If true (default), the definition body will be validated in the creation and the updates of the procedure. For procedures with an argument of ANY TYPE, the definition body validtion is not supported at creation/update time, and thus this field must be set to false explicitly.
Supporting Types
ArgumentResponse
- Argument
Kind string - Optional. Defaults to FIXED_TYPE.
- Data
Type Pulumi.Google Native. Big Query. V2. Inputs. Standard Sql Data Type Response - Required unless argument_kind = ANY_TYPE.
- Is
Aggregate bool - Optional. Whether the argument is an aggregate function parameter. Must be Unset for routine types other than AGGREGATE_FUNCTION. For AGGREGATE_FUNCTION, if set to false, it is equivalent to adding "NOT AGGREGATE" clause in DDL; Otherwise, it is equivalent to omitting "NOT AGGREGATE" clause in DDL.
- Mode string
- Optional. Specifies whether the argument is input or output. Can be set for procedures only.
- Name string
- Optional. The name of this argument. Can be absent for function return argument.
- Argument
Kind string - Optional. Defaults to FIXED_TYPE.
- Data
Type StandardSql Data Type Response - Required unless argument_kind = ANY_TYPE.
- Is
Aggregate bool - Optional. Whether the argument is an aggregate function parameter. Must be Unset for routine types other than AGGREGATE_FUNCTION. For AGGREGATE_FUNCTION, if set to false, it is equivalent to adding "NOT AGGREGATE" clause in DDL; Otherwise, it is equivalent to omitting "NOT AGGREGATE" clause in DDL.
- Mode string
- Optional. Specifies whether the argument is input or output. Can be set for procedures only.
- Name string
- Optional. The name of this argument. Can be absent for function return argument.
- argument
Kind String - Optional. Defaults to FIXED_TYPE.
- data
Type StandardSql Data Type Response - Required unless argument_kind = ANY_TYPE.
- is
Aggregate Boolean - Optional. Whether the argument is an aggregate function parameter. Must be Unset for routine types other than AGGREGATE_FUNCTION. For AGGREGATE_FUNCTION, if set to false, it is equivalent to adding "NOT AGGREGATE" clause in DDL; Otherwise, it is equivalent to omitting "NOT AGGREGATE" clause in DDL.
- mode String
- Optional. Specifies whether the argument is input or output. Can be set for procedures only.
- name String
- Optional. The name of this argument. Can be absent for function return argument.
- argument
Kind string - Optional. Defaults to FIXED_TYPE.
- data
Type StandardSql Data Type Response - Required unless argument_kind = ANY_TYPE.
- is
Aggregate boolean - Optional. Whether the argument is an aggregate function parameter. Must be Unset for routine types other than AGGREGATE_FUNCTION. For AGGREGATE_FUNCTION, if set to false, it is equivalent to adding "NOT AGGREGATE" clause in DDL; Otherwise, it is equivalent to omitting "NOT AGGREGATE" clause in DDL.
- mode string
- Optional. Specifies whether the argument is input or output. Can be set for procedures only.
- name string
- Optional. The name of this argument. Can be absent for function return argument.
- argument_
kind str - Optional. Defaults to FIXED_TYPE.
- data_
type StandardSql Data Type Response - Required unless argument_kind = ANY_TYPE.
- is_
aggregate bool - Optional. Whether the argument is an aggregate function parameter. Must be Unset for routine types other than AGGREGATE_FUNCTION. For AGGREGATE_FUNCTION, if set to false, it is equivalent to adding "NOT AGGREGATE" clause in DDL; Otherwise, it is equivalent to omitting "NOT AGGREGATE" clause in DDL.
- mode str
- Optional. Specifies whether the argument is input or output. Can be set for procedures only.
- name str
- Optional. The name of this argument. Can be absent for function return argument.
- argument
Kind String - Optional. Defaults to FIXED_TYPE.
- data
Type Property Map - Required unless argument_kind = ANY_TYPE.
- is
Aggregate Boolean - Optional. Whether the argument is an aggregate function parameter. Must be Unset for routine types other than AGGREGATE_FUNCTION. For AGGREGATE_FUNCTION, if set to false, it is equivalent to adding "NOT AGGREGATE" clause in DDL; Otherwise, it is equivalent to omitting "NOT AGGREGATE" clause in DDL.
- mode String
- Optional. Specifies whether the argument is input or output. Can be set for procedures only.
- name String
- Optional. The name of this argument. Can be absent for function return argument.
RemoteFunctionOptionsResponse
- Connection string
- Fully qualified name of the user-provided connection object which holds the authentication information to send requests to the remote service. Format:
"projects/{projectId}/locations/{locationId}/connections/{connectionId}"
- Endpoint string
- Endpoint of the user-provided remote service, e.g.
https://us-east1-my_gcf_project.cloudfunctions.net/remote_add
- Max
Batching stringRows - Max number of rows in each batch sent to the remote service. If absent or if 0, BigQuery dynamically decides the number of rows in a batch.
- User
Defined Dictionary<string, string>Context - User-defined context as a set of key/value pairs, which will be sent as function invocation context together with batched arguments in the requests to the remote service. The total number of bytes of keys and values must be less than 8KB.
- Connection string
- Fully qualified name of the user-provided connection object which holds the authentication information to send requests to the remote service. Format:
"projects/{projectId}/locations/{locationId}/connections/{connectionId}"
- Endpoint string
- Endpoint of the user-provided remote service, e.g.
https://us-east1-my_gcf_project.cloudfunctions.net/remote_add
- Max
Batching stringRows - Max number of rows in each batch sent to the remote service. If absent or if 0, BigQuery dynamically decides the number of rows in a batch.
- User
Defined map[string]stringContext - User-defined context as a set of key/value pairs, which will be sent as function invocation context together with batched arguments in the requests to the remote service. The total number of bytes of keys and values must be less than 8KB.
- connection String
- Fully qualified name of the user-provided connection object which holds the authentication information to send requests to the remote service. Format:
"projects/{projectId}/locations/{locationId}/connections/{connectionId}"
- endpoint String
- Endpoint of the user-provided remote service, e.g.
https://us-east1-my_gcf_project.cloudfunctions.net/remote_add
- max
Batching StringRows - Max number of rows in each batch sent to the remote service. If absent or if 0, BigQuery dynamically decides the number of rows in a batch.
- user
Defined Map<String,String>Context - User-defined context as a set of key/value pairs, which will be sent as function invocation context together with batched arguments in the requests to the remote service. The total number of bytes of keys and values must be less than 8KB.
- connection string
- Fully qualified name of the user-provided connection object which holds the authentication information to send requests to the remote service. Format:
"projects/{projectId}/locations/{locationId}/connections/{connectionId}"
- endpoint string
- Endpoint of the user-provided remote service, e.g.
https://us-east1-my_gcf_project.cloudfunctions.net/remote_add
- max
Batching stringRows - Max number of rows in each batch sent to the remote service. If absent or if 0, BigQuery dynamically decides the number of rows in a batch.
- user
Defined {[key: string]: string}Context - User-defined context as a set of key/value pairs, which will be sent as function invocation context together with batched arguments in the requests to the remote service. The total number of bytes of keys and values must be less than 8KB.
- connection str
- Fully qualified name of the user-provided connection object which holds the authentication information to send requests to the remote service. Format:
"projects/{projectId}/locations/{locationId}/connections/{connectionId}"
- endpoint str
- Endpoint of the user-provided remote service, e.g.
https://us-east1-my_gcf_project.cloudfunctions.net/remote_add
- max_
batching_ strrows - Max number of rows in each batch sent to the remote service. If absent or if 0, BigQuery dynamically decides the number of rows in a batch.
- user_
defined_ Mapping[str, str]context - User-defined context as a set of key/value pairs, which will be sent as function invocation context together with batched arguments in the requests to the remote service. The total number of bytes of keys and values must be less than 8KB.
- connection String
- Fully qualified name of the user-provided connection object which holds the authentication information to send requests to the remote service. Format:
"projects/{projectId}/locations/{locationId}/connections/{connectionId}"
- endpoint String
- Endpoint of the user-provided remote service, e.g.
https://us-east1-my_gcf_project.cloudfunctions.net/remote_add
- max
Batching StringRows - Max number of rows in each batch sent to the remote service. If absent or if 0, BigQuery dynamically decides the number of rows in a batch.
- user
Defined Map<String>Context - User-defined context as a set of key/value pairs, which will be sent as function invocation context together with batched arguments in the requests to the remote service. The total number of bytes of keys and values must be less than 8KB.
RoutineReferenceResponse
- dataset_
id str - The ID of the dataset containing this routine.
- project str
- The ID of the project containing this routine.
- routine_
id str - The ID of the routine. The ID must contain only letters (a-z, A-Z), numbers (0-9), or underscores (_). The maximum length is 256 characters.
SparkOptionsResponse
- Archive
Uris List<string> - Archive files to be extracted into the working directory of each executor. For more information about Apache Spark, see Apache Spark.
- Connection string
- Fully qualified name of the user-provided Spark connection object. Format:
"projects/{project_id}/locations/{location_id}/connections/{connection_id}"
- Container
Image string - Custom container image for the runtime environment.
- File
Uris List<string> - Files to be placed in the working directory of each executor. For more information about Apache Spark, see Apache Spark.
- Jar
Uris List<string> - JARs to include on the driver and executor CLASSPATH. For more information about Apache Spark, see Apache Spark.
- Main
Class string - The fully qualified name of a class in jar_uris, for example, com.example.wordcount. Exactly one of main_class and main_jar_uri field should be set for Java/Scala language type.
- Main
File stringUri - The main file/jar URI of the Spark application. Exactly one of the definition_body field and the main_file_uri field must be set for Python. Exactly one of main_class and main_file_uri field should be set for Java/Scala language type.
- Properties Dictionary<string, string>
- Configuration properties as a set of key/value pairs, which will be passed on to the Spark application. For more information, see Apache Spark and the procedure option list.
- Py
File List<string>Uris - Python files to be placed on the PYTHONPATH for PySpark application. Supported file types:
.py
,.egg
, and.zip
. For more information about Apache Spark, see Apache Spark. - Runtime
Version string - Runtime version. If not specified, the default runtime version is used.
- Archive
Uris []string - Archive files to be extracted into the working directory of each executor. For more information about Apache Spark, see Apache Spark.
- Connection string
- Fully qualified name of the user-provided Spark connection object. Format:
"projects/{project_id}/locations/{location_id}/connections/{connection_id}"
- Container
Image string - Custom container image for the runtime environment.
- File
Uris []string - Files to be placed in the working directory of each executor. For more information about Apache Spark, see Apache Spark.
- Jar
Uris []string - JARs to include on the driver and executor CLASSPATH. For more information about Apache Spark, see Apache Spark.
- Main
Class string - The fully qualified name of a class in jar_uris, for example, com.example.wordcount. Exactly one of main_class and main_jar_uri field should be set for Java/Scala language type.
- Main
File stringUri - The main file/jar URI of the Spark application. Exactly one of the definition_body field and the main_file_uri field must be set for Python. Exactly one of main_class and main_file_uri field should be set for Java/Scala language type.
- Properties map[string]string
- Configuration properties as a set of key/value pairs, which will be passed on to the Spark application. For more information, see Apache Spark and the procedure option list.
- Py
File []stringUris - Python files to be placed on the PYTHONPATH for PySpark application. Supported file types:
.py
,.egg
, and.zip
. For more information about Apache Spark, see Apache Spark. - Runtime
Version string - Runtime version. If not specified, the default runtime version is used.
- archive
Uris List<String> - Archive files to be extracted into the working directory of each executor. For more information about Apache Spark, see Apache Spark.
- connection String
- Fully qualified name of the user-provided Spark connection object. Format:
"projects/{project_id}/locations/{location_id}/connections/{connection_id}"
- container
Image String - Custom container image for the runtime environment.
- file
Uris List<String> - Files to be placed in the working directory of each executor. For more information about Apache Spark, see Apache Spark.
- jar
Uris List<String> - JARs to include on the driver and executor CLASSPATH. For more information about Apache Spark, see Apache Spark.
- main
Class String - The fully qualified name of a class in jar_uris, for example, com.example.wordcount. Exactly one of main_class and main_jar_uri field should be set for Java/Scala language type.
- main
File StringUri - The main file/jar URI of the Spark application. Exactly one of the definition_body field and the main_file_uri field must be set for Python. Exactly one of main_class and main_file_uri field should be set for Java/Scala language type.
- properties Map<String,String>
- Configuration properties as a set of key/value pairs, which will be passed on to the Spark application. For more information, see Apache Spark and the procedure option list.
- py
File List<String>Uris - Python files to be placed on the PYTHONPATH for PySpark application. Supported file types:
.py
,.egg
, and.zip
. For more information about Apache Spark, see Apache Spark. - runtime
Version String - Runtime version. If not specified, the default runtime version is used.
- archive
Uris string[] - Archive files to be extracted into the working directory of each executor. For more information about Apache Spark, see Apache Spark.
- connection string
- Fully qualified name of the user-provided Spark connection object. Format:
"projects/{project_id}/locations/{location_id}/connections/{connection_id}"
- container
Image string - Custom container image for the runtime environment.
- file
Uris string[] - Files to be placed in the working directory of each executor. For more information about Apache Spark, see Apache Spark.
- jar
Uris string[] - JARs to include on the driver and executor CLASSPATH. For more information about Apache Spark, see Apache Spark.
- main
Class string - The fully qualified name of a class in jar_uris, for example, com.example.wordcount. Exactly one of main_class and main_jar_uri field should be set for Java/Scala language type.
- main
File stringUri - The main file/jar URI of the Spark application. Exactly one of the definition_body field and the main_file_uri field must be set for Python. Exactly one of main_class and main_file_uri field should be set for Java/Scala language type.
- properties {[key: string]: string}
- Configuration properties as a set of key/value pairs, which will be passed on to the Spark application. For more information, see Apache Spark and the procedure option list.
- py
File string[]Uris - Python files to be placed on the PYTHONPATH for PySpark application. Supported file types:
.py
,.egg
, and.zip
. For more information about Apache Spark, see Apache Spark. - runtime
Version string - Runtime version. If not specified, the default runtime version is used.
- archive_
uris Sequence[str] - Archive files to be extracted into the working directory of each executor. For more information about Apache Spark, see Apache Spark.
- connection str
- Fully qualified name of the user-provided Spark connection object. Format:
"projects/{project_id}/locations/{location_id}/connections/{connection_id}"
- container_
image str - Custom container image for the runtime environment.
- file_
uris Sequence[str] - Files to be placed in the working directory of each executor. For more information about Apache Spark, see Apache Spark.
- jar_
uris Sequence[str] - JARs to include on the driver and executor CLASSPATH. For more information about Apache Spark, see Apache Spark.
- main_
class str - The fully qualified name of a class in jar_uris, for example, com.example.wordcount. Exactly one of main_class and main_jar_uri field should be set for Java/Scala language type.
- main_
file_ struri - The main file/jar URI of the Spark application. Exactly one of the definition_body field and the main_file_uri field must be set for Python. Exactly one of main_class and main_file_uri field should be set for Java/Scala language type.
- properties Mapping[str, str]
- Configuration properties as a set of key/value pairs, which will be passed on to the Spark application. For more information, see Apache Spark and the procedure option list.
- py_
file_ Sequence[str]uris - Python files to be placed on the PYTHONPATH for PySpark application. Supported file types:
.py
,.egg
, and.zip
. For more information about Apache Spark, see Apache Spark. - runtime_
version str - Runtime version. If not specified, the default runtime version is used.
- archive
Uris List<String> - Archive files to be extracted into the working directory of each executor. For more information about Apache Spark, see Apache Spark.
- connection String
- Fully qualified name of the user-provided Spark connection object. Format:
"projects/{project_id}/locations/{location_id}/connections/{connection_id}"
- container
Image String - Custom container image for the runtime environment.
- file
Uris List<String> - Files to be placed in the working directory of each executor. For more information about Apache Spark, see Apache Spark.
- jar
Uris List<String> - JARs to include on the driver and executor CLASSPATH. For more information about Apache Spark, see Apache Spark.
- main
Class String - The fully qualified name of a class in jar_uris, for example, com.example.wordcount. Exactly one of main_class and main_jar_uri field should be set for Java/Scala language type.
- main
File StringUri - The main file/jar URI of the Spark application. Exactly one of the definition_body field and the main_file_uri field must be set for Python. Exactly one of main_class and main_file_uri field should be set for Java/Scala language type.
- properties Map<String>
- Configuration properties as a set of key/value pairs, which will be passed on to the Spark application. For more information, see Apache Spark and the procedure option list.
- py
File List<String>Uris - Python files to be placed on the PYTHONPATH for PySpark application. Supported file types:
.py
,.egg
, and.zip
. For more information about Apache Spark, see Apache Spark. - runtime
Version String - Runtime version. If not specified, the default runtime version is used.
StandardSqlDataTypeResponse
- Struct
Type Pulumi.Google Native. Big Query. V2. Inputs. Standard Sql Struct Type Response - The fields of this struct, in order, if type_kind = "STRUCT".
- Type
Kind string - The top level type of this field. Can be any GoogleSQL data type (e.g., "INT64", "DATE", "ARRAY").
- Array
Element Pulumi.Type Google Native. Big Query. V2. Inputs. Standard Sql Data Type Response - The type of the array's elements, if type_kind = "ARRAY".
- Range
Element Pulumi.Type Google Native. Big Query. V2. Inputs. Standard Sql Data Type Response - The type of the range's elements, if type_kind = "RANGE".
- Struct
Type StandardSql Struct Type Response - The fields of this struct, in order, if type_kind = "STRUCT".
- Type
Kind string - The top level type of this field. Can be any GoogleSQL data type (e.g., "INT64", "DATE", "ARRAY").
- Array
Element StandardType Sql Data Type Response - The type of the array's elements, if type_kind = "ARRAY".
- Range
Element StandardType Sql Data Type Response - The type of the range's elements, if type_kind = "RANGE".
- struct
Type StandardSql Struct Type Response - The fields of this struct, in order, if type_kind = "STRUCT".
- type
Kind String - The top level type of this field. Can be any GoogleSQL data type (e.g., "INT64", "DATE", "ARRAY").
- array
Element StandardType Sql Data Type Response - The type of the array's elements, if type_kind = "ARRAY".
- range
Element StandardType Sql Data Type Response - The type of the range's elements, if type_kind = "RANGE".
- struct
Type StandardSql Struct Type Response - The fields of this struct, in order, if type_kind = "STRUCT".
- type
Kind string - The top level type of this field. Can be any GoogleSQL data type (e.g., "INT64", "DATE", "ARRAY").
- array
Element StandardType Sql Data Type Response - The type of the array's elements, if type_kind = "ARRAY".
- range
Element StandardType Sql Data Type Response - The type of the range's elements, if type_kind = "RANGE".
- struct_
type StandardSql Struct Type Response - The fields of this struct, in order, if type_kind = "STRUCT".
- type_
kind str - The top level type of this field. Can be any GoogleSQL data type (e.g., "INT64", "DATE", "ARRAY").
- array_
element_ Standardtype Sql Data Type Response - The type of the array's elements, if type_kind = "ARRAY".
- range_
element_ Standardtype Sql Data Type Response - The type of the range's elements, if type_kind = "RANGE".
- struct
Type Property Map - The fields of this struct, in order, if type_kind = "STRUCT".
- type
Kind String - The top level type of this field. Can be any GoogleSQL data type (e.g., "INT64", "DATE", "ARRAY").
- array
Element Property MapType - The type of the array's elements, if type_kind = "ARRAY".
- range
Element Property MapType - The type of the range's elements, if type_kind = "RANGE".
StandardSqlFieldResponse
- Name string
- Optional. The name of this field. Can be absent for struct fields.
- Type
Pulumi.
Google Native. Big Query. V2. Inputs. Standard Sql Data Type Response - Optional. The type of this parameter. Absent if not explicitly specified (e.g., CREATE FUNCTION statement can omit the return type; in this case the output parameter does not have this "type" field).
- Name string
- Optional. The name of this field. Can be absent for struct fields.
- Type
Standard
Sql Data Type Response - Optional. The type of this parameter. Absent if not explicitly specified (e.g., CREATE FUNCTION statement can omit the return type; in this case the output parameter does not have this "type" field).
- name String
- Optional. The name of this field. Can be absent for struct fields.
- type
Standard
Sql Data Type Response - Optional. The type of this parameter. Absent if not explicitly specified (e.g., CREATE FUNCTION statement can omit the return type; in this case the output parameter does not have this "type" field).
- name string
- Optional. The name of this field. Can be absent for struct fields.
- type
Standard
Sql Data Type Response - Optional. The type of this parameter. Absent if not explicitly specified (e.g., CREATE FUNCTION statement can omit the return type; in this case the output parameter does not have this "type" field).
- name str
- Optional. The name of this field. Can be absent for struct fields.
- type
Standard
Sql Data Type Response - Optional. The type of this parameter. Absent if not explicitly specified (e.g., CREATE FUNCTION statement can omit the return type; in this case the output parameter does not have this "type" field).
- name String
- Optional. The name of this field. Can be absent for struct fields.
- type Property Map
- Optional. The type of this parameter. Absent if not explicitly specified (e.g., CREATE FUNCTION statement can omit the return type; in this case the output parameter does not have this "type" field).
StandardSqlStructTypeResponse
- Fields
List<Pulumi.
Google Native. Big Query. V2. Inputs. Standard Sql Field Response> - Fields within the struct.
- Fields
[]Standard
Sql Field Response - Fields within the struct.
- fields
List<Standard
Sql Field Response> - Fields within the struct.
- fields
Standard
Sql Field Response[] - Fields within the struct.
- fields
Sequence[Standard
Sql Field Response] - Fields within the struct.
- fields List<Property Map>
- Fields within the struct.
StandardSqlTableTypeResponse
- Columns
List<Pulumi.
Google Native. Big Query. V2. Inputs. Standard Sql Field Response> - The columns in this table type
- Columns
[]Standard
Sql Field Response - The columns in this table type
- columns
List<Standard
Sql Field Response> - The columns in this table type
- columns
Standard
Sql Field Response[] - The columns in this table type
- columns
Sequence[Standard
Sql Field Response] - The columns in this table type
- columns List<Property Map>
- The columns in this table type
Package Details
- Repository
- Google Cloud Native pulumi/pulumi-google-native
- License
- Apache-2.0
Google Cloud Native is in preview. Google Cloud Classic is fully supported.
Google Cloud Native v0.32.0 published on Wednesday, Nov 29, 2023 by Pulumi