azure-native.media.Transform
Explore with Pulumi AI
A Transform encapsulates the rules or instructions for generating desired outputs from input media, such as by transcoding or by extracting insights. After the Transform is created, it can be applied to input media by creating Jobs. API Version: 2020-05-01.
Example Usage
Create or update a Transform
using System.Collections.Generic;
using System.Linq;
using Pulumi;
using AzureNative = Pulumi.AzureNative;
return await Deployment.RunAsync(() =>
{
var transform = new AzureNative.Media.Transform("transform", new()
{
AccountName = "contosomedia",
Description = "Example Transform to illustrate create and update.",
Outputs = new[]
{
new AzureNative.Media.Inputs.TransformOutputArgs
{
Preset = new AzureNative.Media.Inputs.BuiltInStandardEncoderPresetArgs
{
OdataType = "#Microsoft.Media.BuiltInStandardEncoderPreset",
PresetName = "AdaptiveStreaming",
},
},
},
ResourceGroupName = "contosoresources",
TransformName = "createdTransform",
});
});
package main
import (
media "github.com/pulumi/pulumi-azure-native-sdk/media"
"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
)
func main() {
pulumi.Run(func(ctx *pulumi.Context) error {
_, err := media.NewTransform(ctx, "transform", &media.TransformArgs{
AccountName: pulumi.String("contosomedia"),
Description: pulumi.String("Example Transform to illustrate create and update."),
Outputs: []media.TransformOutputTypeArgs{
{
Preset: {
OdataType: "#Microsoft.Media.BuiltInStandardEncoderPreset",
PresetName: "AdaptiveStreaming",
},
},
},
ResourceGroupName: pulumi.String("contosoresources"),
TransformName: pulumi.String("createdTransform"),
})
if err != nil {
return err
}
return nil
})
}
package generated_program;
import com.pulumi.Context;
import com.pulumi.Pulumi;
import com.pulumi.core.Output;
import com.pulumi.azurenative.media.Transform;
import com.pulumi.azurenative.media.TransformArgs;
import java.util.List;
import java.util.ArrayList;
import java.util.Map;
import java.io.File;
import java.nio.file.Files;
import java.nio.file.Paths;
public class App {
public static void main(String[] args) {
Pulumi.run(App::stack);
}
public static void stack(Context ctx) {
var transform = new Transform("transform", TransformArgs.builder()
.accountName("contosomedia")
.description("Example Transform to illustrate create and update.")
.outputs(Map.of("preset", Map.ofEntries(
Map.entry("odataType", "#Microsoft.Media.BuiltInStandardEncoderPreset"),
Map.entry("presetName", "AdaptiveStreaming")
)))
.resourceGroupName("contosoresources")
.transformName("createdTransform")
.build());
}
}
import pulumi
import pulumi_azure_native as azure_native
transform = azure_native.media.Transform("transform",
account_name="contosomedia",
description="Example Transform to illustrate create and update.",
outputs=[azure_native.media.TransformOutputArgs(
preset=azure_native.media.BuiltInStandardEncoderPresetArgs(
odata_type="#Microsoft.Media.BuiltInStandardEncoderPreset",
preset_name="AdaptiveStreaming",
),
)],
resource_group_name="contosoresources",
transform_name="createdTransform")
import * as pulumi from "@pulumi/pulumi";
import * as azure_native from "@pulumi/azure-native";
const transform = new azure_native.media.Transform("transform", {
accountName: "contosomedia",
description: "Example Transform to illustrate create and update.",
outputs: [{
preset: {
odataType: "#Microsoft.Media.BuiltInStandardEncoderPreset",
presetName: "AdaptiveStreaming",
},
}],
resourceGroupName: "contosoresources",
transformName: "createdTransform",
});
resources:
transform:
type: azure-native:media:Transform
properties:
accountName: contosomedia
description: Example Transform to illustrate create and update.
outputs:
- preset:
odataType: '#Microsoft.Media.BuiltInStandardEncoderPreset'
presetName: AdaptiveStreaming
resourceGroupName: contosoresources
transformName: createdTransform
Create Transform Resource
Resources are created with functions called constructors. To learn more about declaring and configuring resources, see Resources.
Constructor syntax
new Transform(name: string, args: TransformArgs, opts?: CustomResourceOptions);
@overload
def Transform(resource_name: str,
args: TransformArgs,
opts: Optional[ResourceOptions] = None)
@overload
def Transform(resource_name: str,
opts: Optional[ResourceOptions] = None,
account_name: Optional[str] = None,
outputs: Optional[Sequence[TransformOutputArgs]] = None,
resource_group_name: Optional[str] = None,
description: Optional[str] = None,
transform_name: Optional[str] = None)
func NewTransform(ctx *Context, name string, args TransformArgs, opts ...ResourceOption) (*Transform, error)
public Transform(string name, TransformArgs args, CustomResourceOptions? opts = null)
public Transform(String name, TransformArgs args)
public Transform(String name, TransformArgs args, CustomResourceOptions options)
type: azure-native:media:Transform
properties: # The arguments to resource properties.
options: # Bag of options to control resource's behavior.
Parameters
- name string
- The unique name of the resource.
- args TransformArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- resource_name str
- The unique name of the resource.
- args TransformArgs
- The arguments to resource properties.
- opts ResourceOptions
- Bag of options to control resource's behavior.
- ctx Context
- Context object for the current deployment.
- name string
- The unique name of the resource.
- args TransformArgs
- The arguments to resource properties.
- opts ResourceOption
- Bag of options to control resource's behavior.
- name string
- The unique name of the resource.
- args TransformArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- name String
- The unique name of the resource.
- args TransformArgs
- The arguments to resource properties.
- options CustomResourceOptions
- Bag of options to control resource's behavior.
Constructor example
The following reference example uses placeholder values for all input properties.
var transformResource = new AzureNative.Media.Transform("transformResource", new()
{
AccountName = "string",
Outputs = new[]
{
{
{ "preset",
{
{ "odataType", "#Microsoft.Media.AudioAnalyzerPreset" },
{ "audioLanguage", "string" },
{ "experimentalOptions",
{
{ "string", "string" },
} },
{ "mode", "string" },
} },
{ "onError", "string" },
{ "relativePriority", "string" },
},
},
ResourceGroupName = "string",
Description = "string",
TransformName = "string",
});
example, err := media.NewTransform(ctx, "transformResource", &media.TransformArgs{
AccountName: "string",
Outputs: []map[string]interface{}{
map[string]interface{}{
"preset": map[string]interface{}{
"odataType": "#Microsoft.Media.AudioAnalyzerPreset",
"audioLanguage": "string",
"experimentalOptions": map[string]interface{}{
"string": "string",
},
"mode": "string",
},
"onError": "string",
"relativePriority": "string",
},
},
ResourceGroupName: "string",
Description: "string",
TransformName: "string",
})
var transformResource = new Transform("transformResource", TransformArgs.builder()
.accountName("string")
.outputs(%!v(PANIC=Format method: runtime error: invalid memory address or nil pointer dereference))
.resourceGroupName("string")
.description("string")
.transformName("string")
.build());
transform_resource = azure_native.media.Transform("transformResource",
account_name=string,
outputs=[{
preset: {
odataType: #Microsoft.Media.AudioAnalyzerPreset,
audioLanguage: string,
experimentalOptions: {
string: string,
},
mode: string,
},
onError: string,
relativePriority: string,
}],
resource_group_name=string,
description=string,
transform_name=string)
const transformResource = new azure_native.media.Transform("transformResource", {
accountName: "string",
outputs: [{
preset: {
odataType: "#Microsoft.Media.AudioAnalyzerPreset",
audioLanguage: "string",
experimentalOptions: {
string: "string",
},
mode: "string",
},
onError: "string",
relativePriority: "string",
}],
resourceGroupName: "string",
description: "string",
transformName: "string",
});
type: azure-native:media:Transform
properties:
accountName: string
description: string
outputs:
- onError: string
preset:
audioLanguage: string
experimentalOptions:
string: string
mode: string
odataType: '#Microsoft.Media.AudioAnalyzerPreset'
relativePriority: string
resourceGroupName: string
transformName: string
Transform Resource Properties
To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.
Inputs
The Transform resource accepts the following input properties:
- Account
Name string - The Media Services account name.
- Outputs
List<Pulumi.
Azure Native. Media. Inputs. Transform Output> - An array of one or more TransformOutputs that the Transform should generate.
- Resource
Group stringName - The name of the resource group within the Azure subscription.
- Description string
- An optional verbose description of the Transform.
- Transform
Name string - The Transform name.
- Account
Name string - The Media Services account name.
- Outputs
[]Transform
Output Type Args - An array of one or more TransformOutputs that the Transform should generate.
- Resource
Group stringName - The name of the resource group within the Azure subscription.
- Description string
- An optional verbose description of the Transform.
- Transform
Name string - The Transform name.
- account
Name String - The Media Services account name.
- outputs
List<Transform
Output> - An array of one or more TransformOutputs that the Transform should generate.
- resource
Group StringName - The name of the resource group within the Azure subscription.
- description String
- An optional verbose description of the Transform.
- transform
Name String - The Transform name.
- account
Name string - The Media Services account name.
- outputs
Transform
Output[] - An array of one or more TransformOutputs that the Transform should generate.
- resource
Group stringName - The name of the resource group within the Azure subscription.
- description string
- An optional verbose description of the Transform.
- transform
Name string - The Transform name.
- account_
name str - The Media Services account name.
- outputs
Sequence[Transform
Output Args] - An array of one or more TransformOutputs that the Transform should generate.
- resource_
group_ strname - The name of the resource group within the Azure subscription.
- description str
- An optional verbose description of the Transform.
- transform_
name str - The Transform name.
- account
Name String - The Media Services account name.
- outputs List<Property Map>
- An array of one or more TransformOutputs that the Transform should generate.
- resource
Group StringName - The name of the resource group within the Azure subscription.
- description String
- An optional verbose description of the Transform.
- transform
Name String - The Transform name.
Outputs
All input properties are implicitly available as output properties. Additionally, the Transform resource produces the following output properties:
- Created string
- The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
- Id string
- The provider-assigned unique ID for this managed resource.
- Last
Modified string - The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
- Name string
- The name of the resource
- System
Data Pulumi.Azure Native. Media. Outputs. System Data Response - The system metadata relating to this resource.
- Type string
- The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"
- Created string
- The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
- Id string
- The provider-assigned unique ID for this managed resource.
- Last
Modified string - The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
- Name string
- The name of the resource
- System
Data SystemData Response - The system metadata relating to this resource.
- Type string
- The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"
- created String
- The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
- id String
- The provider-assigned unique ID for this managed resource.
- last
Modified String - The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
- name String
- The name of the resource
- system
Data SystemData Response - The system metadata relating to this resource.
- type String
- The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"
- created string
- The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
- id string
- The provider-assigned unique ID for this managed resource.
- last
Modified string - The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
- name string
- The name of the resource
- system
Data SystemData Response - The system metadata relating to this resource.
- type string
- The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"
- created str
- The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
- id str
- The provider-assigned unique ID for this managed resource.
- last_
modified str - The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
- name str
- The name of the resource
- system_
data SystemData Response - The system metadata relating to this resource.
- type str
- The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"
- created String
- The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
- id String
- The provider-assigned unique ID for this managed resource.
- last
Modified String - The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
- name String
- The name of the resource
- system
Data Property Map - The system metadata relating to this resource.
- type String
- The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"
Supporting Types
AacAudio, AacAudioArgs
- Bitrate int
- The bitrate, in bits per second, of the output encoded audio.
- Channels int
- The number of channels in the audio.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Profile
string | Pulumi.
Azure Native. Media. Aac Audio Profile - The encoding profile to be used when encoding audio with AAC.
- Sampling
Rate int - The sampling rate to use for encoding in hertz.
- Bitrate int
- The bitrate, in bits per second, of the output encoded audio.
- Channels int
- The number of channels in the audio.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Profile
string | Aac
Audio Profile - The encoding profile to be used when encoding audio with AAC.
- Sampling
Rate int - The sampling rate to use for encoding in hertz.
- bitrate Integer
- The bitrate, in bits per second, of the output encoded audio.
- channels Integer
- The number of channels in the audio.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- profile
String | Aac
Audio Profile - The encoding profile to be used when encoding audio with AAC.
- sampling
Rate Integer - The sampling rate to use for encoding in hertz.
- bitrate number
- The bitrate, in bits per second, of the output encoded audio.
- channels number
- The number of channels in the audio.
- label string
- An optional label for the codec. The label can be used to control muxing behavior.
- profile
string | Aac
Audio Profile - The encoding profile to be used when encoding audio with AAC.
- sampling
Rate number - The sampling rate to use for encoding in hertz.
- bitrate int
- The bitrate, in bits per second, of the output encoded audio.
- channels int
- The number of channels in the audio.
- label str
- An optional label for the codec. The label can be used to control muxing behavior.
- profile
str | Aac
Audio Profile - The encoding profile to be used when encoding audio with AAC.
- sampling_
rate int - The sampling rate to use for encoding in hertz.
- bitrate Number
- The bitrate, in bits per second, of the output encoded audio.
- channels Number
- The number of channels in the audio.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- profile
String | "Aac
Lc" | "He Aac V1" | "He Aac V2" - The encoding profile to be used when encoding audio with AAC.
- sampling
Rate Number - The sampling rate to use for encoding in hertz.
AacAudioProfile, AacAudioProfileArgs
- Aac
Lc - AacLcSpecifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
- He
Aac V1 - HeAacV1Specifies that the output audio is to be encoded into HE-AAC v1 profile.
- He
Aac V2 - HeAacV2Specifies that the output audio is to be encoded into HE-AAC v2 profile.
- Aac
Audio Profile Aac Lc - AacLcSpecifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
- Aac
Audio Profile He Aac V1 - HeAacV1Specifies that the output audio is to be encoded into HE-AAC v1 profile.
- Aac
Audio Profile He Aac V2 - HeAacV2Specifies that the output audio is to be encoded into HE-AAC v2 profile.
- Aac
Lc - AacLcSpecifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
- He
Aac V1 - HeAacV1Specifies that the output audio is to be encoded into HE-AAC v1 profile.
- He
Aac V2 - HeAacV2Specifies that the output audio is to be encoded into HE-AAC v2 profile.
- Aac
Lc - AacLcSpecifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
- He
Aac V1 - HeAacV1Specifies that the output audio is to be encoded into HE-AAC v1 profile.
- He
Aac V2 - HeAacV2Specifies that the output audio is to be encoded into HE-AAC v2 profile.
- AAC_LC
- AacLcSpecifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
- HE_AAC_V1
- HeAacV1Specifies that the output audio is to be encoded into HE-AAC v1 profile.
- HE_AAC_V2
- HeAacV2Specifies that the output audio is to be encoded into HE-AAC v2 profile.
- "Aac
Lc" - AacLcSpecifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
- "He
Aac V1" - HeAacV1Specifies that the output audio is to be encoded into HE-AAC v1 profile.
- "He
Aac V2" - HeAacV2Specifies that the output audio is to be encoded into HE-AAC v2 profile.
AacAudioResponse, AacAudioResponseArgs
- Bitrate int
- The bitrate, in bits per second, of the output encoded audio.
- Channels int
- The number of channels in the audio.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Profile string
- The encoding profile to be used when encoding audio with AAC.
- Sampling
Rate int - The sampling rate to use for encoding in hertz.
- Bitrate int
- The bitrate, in bits per second, of the output encoded audio.
- Channels int
- The number of channels in the audio.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Profile string
- The encoding profile to be used when encoding audio with AAC.
- Sampling
Rate int - The sampling rate to use for encoding in hertz.
- bitrate Integer
- The bitrate, in bits per second, of the output encoded audio.
- channels Integer
- The number of channels in the audio.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- profile String
- The encoding profile to be used when encoding audio with AAC.
- sampling
Rate Integer - The sampling rate to use for encoding in hertz.
- bitrate number
- The bitrate, in bits per second, of the output encoded audio.
- channels number
- The number of channels in the audio.
- label string
- An optional label for the codec. The label can be used to control muxing behavior.
- profile string
- The encoding profile to be used when encoding audio with AAC.
- sampling
Rate number - The sampling rate to use for encoding in hertz.
- bitrate int
- The bitrate, in bits per second, of the output encoded audio.
- channels int
- The number of channels in the audio.
- label str
- An optional label for the codec. The label can be used to control muxing behavior.
- profile str
- The encoding profile to be used when encoding audio with AAC.
- sampling_
rate int - The sampling rate to use for encoding in hertz.
- bitrate Number
- The bitrate, in bits per second, of the output encoded audio.
- channels Number
- The number of channels in the audio.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- profile String
- The encoding profile to be used when encoding audio with AAC.
- sampling
Rate Number - The sampling rate to use for encoding in hertz.
AnalysisResolution, AnalysisResolutionArgs
- Source
Resolution - SourceResolution
- Standard
Definition - StandardDefinition
- Analysis
Resolution Source Resolution - SourceResolution
- Analysis
Resolution Standard Definition - StandardDefinition
- Source
Resolution - SourceResolution
- Standard
Definition - StandardDefinition
- Source
Resolution - SourceResolution
- Standard
Definition - StandardDefinition
- SOURCE_RESOLUTION
- SourceResolution
- STANDARD_DEFINITION
- StandardDefinition
- "Source
Resolution" - SourceResolution
- "Standard
Definition" - StandardDefinition
Audio, AudioArgs
- Bitrate int
- The bitrate, in bits per second, of the output encoded audio.
- Channels int
- The number of channels in the audio.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Sampling
Rate int - The sampling rate to use for encoding in hertz.
- Bitrate int
- The bitrate, in bits per second, of the output encoded audio.
- Channels int
- The number of channels in the audio.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Sampling
Rate int - The sampling rate to use for encoding in hertz.
- bitrate Integer
- The bitrate, in bits per second, of the output encoded audio.
- channels Integer
- The number of channels in the audio.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- sampling
Rate Integer - The sampling rate to use for encoding in hertz.
- bitrate number
- The bitrate, in bits per second, of the output encoded audio.
- channels number
- The number of channels in the audio.
- label string
- An optional label for the codec. The label can be used to control muxing behavior.
- sampling
Rate number - The sampling rate to use for encoding in hertz.
- bitrate int
- The bitrate, in bits per second, of the output encoded audio.
- channels int
- The number of channels in the audio.
- label str
- An optional label for the codec. The label can be used to control muxing behavior.
- sampling_
rate int - The sampling rate to use for encoding in hertz.
- bitrate Number
- The bitrate, in bits per second, of the output encoded audio.
- channels Number
- The number of channels in the audio.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- sampling
Rate Number - The sampling rate to use for encoding in hertz.
AudioAnalysisMode, AudioAnalysisModeArgs
- Standard
- StandardPerforms all operations included in the Basic mode, additionally performing language detection and speaker diarization.
- Basic
- BasicThis mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.
- Audio
Analysis Mode Standard - StandardPerforms all operations included in the Basic mode, additionally performing language detection and speaker diarization.
- Audio
Analysis Mode Basic - BasicThis mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.
- Standard
- StandardPerforms all operations included in the Basic mode, additionally performing language detection and speaker diarization.
- Basic
- BasicThis mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.
- Standard
- StandardPerforms all operations included in the Basic mode, additionally performing language detection and speaker diarization.
- Basic
- BasicThis mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.
- STANDARD
- StandardPerforms all operations included in the Basic mode, additionally performing language detection and speaker diarization.
- BASIC
- BasicThis mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.
- "Standard"
- StandardPerforms all operations included in the Basic mode, additionally performing language detection and speaker diarization.
- "Basic"
- BasicThis mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.
AudioAnalyzerPreset, AudioAnalyzerPresetArgs
- Audio
Language string - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- Experimental
Options Dictionary<string, string> - Dictionary containing key value pairs for parameters not exposed in the preset itself
- Mode
string | Pulumi.
Azure Native. Media. Audio Analysis Mode - Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- Audio
Language string - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- Experimental
Options map[string]string - Dictionary containing key value pairs for parameters not exposed in the preset itself
- Mode
string | Audio
Analysis Mode - Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio
Language String - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental
Options Map<String,String> - Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode
String | Audio
Analysis Mode - Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio
Language string - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental
Options {[key: string]: string} - Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode
string | Audio
Analysis Mode - Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio_
language str - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental_
options Mapping[str, str] - Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode
str | Audio
Analysis Mode - Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio
Language String - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental
Options Map<String> - Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode String | "Standard" | "Basic"
- Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
AudioAnalyzerPresetResponse, AudioAnalyzerPresetResponseArgs
- Audio
Language string - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- Experimental
Options Dictionary<string, string> - Dictionary containing key value pairs for parameters not exposed in the preset itself
- Mode string
- Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- Audio
Language string - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- Experimental
Options map[string]string - Dictionary containing key value pairs for parameters not exposed in the preset itself
- Mode string
- Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio
Language String - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental
Options Map<String,String> - Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode String
- Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio
Language string - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental
Options {[key: string]: string} - Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode string
- Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio_
language str - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental_
options Mapping[str, str] - Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode str
- Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio
Language String - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental
Options Map<String> - Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode String
- Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
AudioOverlay, AudioOverlayArgs
- Input
Label string - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- Audio
Gain doubleLevel - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- End string
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- Fade
In stringDuration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- Fade
Out stringDuration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- Start string
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- Input
Label string - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- Audio
Gain float64Level - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- End string
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- Fade
In stringDuration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- Fade
Out stringDuration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- Start string
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input
Label String - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio
Gain DoubleLevel - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- end String
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade
In StringDuration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade
Out StringDuration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- start String
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input
Label string - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio
Gain numberLevel - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- end string
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade
In stringDuration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade
Out stringDuration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- start string
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input_
label str - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio_
gain_ floatlevel - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- end str
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade_
in_ strduration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade_
out_ strduration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- start str
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input
Label String - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio
Gain NumberLevel - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- end String
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade
In StringDuration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade
Out StringDuration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- start String
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
AudioOverlayResponse, AudioOverlayResponseArgs
- Input
Label string - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- Audio
Gain doubleLevel - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- End string
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- Fade
In stringDuration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- Fade
Out stringDuration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- Start string
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- Input
Label string - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- Audio
Gain float64Level - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- End string
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- Fade
In stringDuration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- Fade
Out stringDuration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- Start string
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input
Label String - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio
Gain DoubleLevel - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- end String
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade
In StringDuration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade
Out StringDuration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- start String
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input
Label string - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio
Gain numberLevel - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- end string
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade
In stringDuration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade
Out stringDuration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- start string
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input_
label str - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio_
gain_ floatlevel - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- end str
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade_
in_ strduration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade_
out_ strduration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- start str
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input
Label String - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio
Gain NumberLevel - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- end String
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade
In StringDuration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade
Out StringDuration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- start String
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
AudioResponse, AudioResponseArgs
- Bitrate int
- The bitrate, in bits per second, of the output encoded audio.
- Channels int
- The number of channels in the audio.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Sampling
Rate int - The sampling rate to use for encoding in hertz.
- Bitrate int
- The bitrate, in bits per second, of the output encoded audio.
- Channels int
- The number of channels in the audio.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Sampling
Rate int - The sampling rate to use for encoding in hertz.
- bitrate Integer
- The bitrate, in bits per second, of the output encoded audio.
- channels Integer
- The number of channels in the audio.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- sampling
Rate Integer - The sampling rate to use for encoding in hertz.
- bitrate number
- The bitrate, in bits per second, of the output encoded audio.
- channels number
- The number of channels in the audio.
- label string
- An optional label for the codec. The label can be used to control muxing behavior.
- sampling
Rate number - The sampling rate to use for encoding in hertz.
- bitrate int
- The bitrate, in bits per second, of the output encoded audio.
- channels int
- The number of channels in the audio.
- label str
- An optional label for the codec. The label can be used to control muxing behavior.
- sampling_
rate int - The sampling rate to use for encoding in hertz.
- bitrate Number
- The bitrate, in bits per second, of the output encoded audio.
- channels Number
- The number of channels in the audio.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- sampling
Rate Number - The sampling rate to use for encoding in hertz.
BlurType, BlurTypeArgs
- Box
- BoxBox: debug filter, bounding box only
- Low
- LowLow: box-car blur filter
- Med
- MedMed: Gaussian blur filter
- High
- HighHigh: Confuse blur filter
- Black
- BlackBlack: Black out filter
- Blur
Type Box - BoxBox: debug filter, bounding box only
- Blur
Type Low - LowLow: box-car blur filter
- Blur
Type Med - MedMed: Gaussian blur filter
- Blur
Type High - HighHigh: Confuse blur filter
- Blur
Type Black - BlackBlack: Black out filter
- Box
- BoxBox: debug filter, bounding box only
- Low
- LowLow: box-car blur filter
- Med
- MedMed: Gaussian blur filter
- High
- HighHigh: Confuse blur filter
- Black
- BlackBlack: Black out filter
- Box
- BoxBox: debug filter, bounding box only
- Low
- LowLow: box-car blur filter
- Med
- MedMed: Gaussian blur filter
- High
- HighHigh: Confuse blur filter
- Black
- BlackBlack: Black out filter
- BOX
- BoxBox: debug filter, bounding box only
- LOW
- LowLow: box-car blur filter
- MED
- MedMed: Gaussian blur filter
- HIGH
- HighHigh: Confuse blur filter
- BLACK
- BlackBlack: Black out filter
- "Box"
- BoxBox: debug filter, bounding box only
- "Low"
- LowLow: box-car blur filter
- "Med"
- MedMed: Gaussian blur filter
- "High"
- HighHigh: Confuse blur filter
- "Black"
- BlackBlack: Black out filter
BuiltInStandardEncoderPreset, BuiltInStandardEncoderPresetArgs
- Preset
Name string | Pulumi.Azure Native. Media. Encoder Named Preset - The built-in preset to be used for encoding videos.
- Preset
Name string | EncoderNamed Preset - The built-in preset to be used for encoding videos.
- preset
Name String | EncoderNamed Preset - The built-in preset to be used for encoding videos.
- preset
Name string | EncoderNamed Preset - The built-in preset to be used for encoding videos.
- preset_
name str | EncoderNamed Preset - The built-in preset to be used for encoding videos.
- preset
Name String | "H264SingleBitrate SD" | "H264Single Bitrate720p" | "H264Single Bitrate1080p" | "Adaptive Streaming" | "AACGood Quality Audio" | "Content Aware Encoding Experimental" | "Content Aware Encoding" | "Copy All Bitrate Non Interleaved" | "H264Multiple Bitrate1080p" | "H264Multiple Bitrate720p" | "H264Multiple Bitrate SD" | "H265Content Aware Encoding" | "H265Adaptive Streaming" | "H265Single Bitrate720p" | "H265Single Bitrate1080p" | "H265Single Bitrate4K" - The built-in preset to be used for encoding videos.
BuiltInStandardEncoderPresetResponse, BuiltInStandardEncoderPresetResponseArgs
- Preset
Name string - The built-in preset to be used for encoding videos.
- Preset
Name string - The built-in preset to be used for encoding videos.
- preset
Name String - The built-in preset to be used for encoding videos.
- preset
Name string - The built-in preset to be used for encoding videos.
- preset_
name str - The built-in preset to be used for encoding videos.
- preset
Name String - The built-in preset to be used for encoding videos.
CopyAudio, CopyAudioArgs
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- label string
- An optional label for the codec. The label can be used to control muxing behavior.
- label str
- An optional label for the codec. The label can be used to control muxing behavior.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
CopyAudioResponse, CopyAudioResponseArgs
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- label string
- An optional label for the codec. The label can be used to control muxing behavior.
- label str
- An optional label for the codec. The label can be used to control muxing behavior.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
CopyVideo, CopyVideoArgs
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- label string
- An optional label for the codec. The label can be used to control muxing behavior.
- label str
- An optional label for the codec. The label can be used to control muxing behavior.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
CopyVideoResponse, CopyVideoResponseArgs
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- label string
- An optional label for the codec. The label can be used to control muxing behavior.
- label str
- An optional label for the codec. The label can be used to control muxing behavior.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
Deinterlace, DeinterlaceArgs
- Mode
string | Pulumi.
Azure Native. Media. Deinterlace Mode - The deinterlacing mode. Defaults to AutoPixelAdaptive.
- Parity
string | Pulumi.
Azure Native. Media. Deinterlace Parity - The field parity for de-interlacing, defaults to Auto.
- Mode
string | Deinterlace
Mode - The deinterlacing mode. Defaults to AutoPixelAdaptive.
- Parity
string | Deinterlace
Parity - The field parity for de-interlacing, defaults to Auto.
- mode
String | Deinterlace
Mode - The deinterlacing mode. Defaults to AutoPixelAdaptive.
- parity
String | Deinterlace
Parity - The field parity for de-interlacing, defaults to Auto.
- mode
string | Deinterlace
Mode - The deinterlacing mode. Defaults to AutoPixelAdaptive.
- parity
string | Deinterlace
Parity - The field parity for de-interlacing, defaults to Auto.
- mode
str | Deinterlace
Mode - The deinterlacing mode. Defaults to AutoPixelAdaptive.
- parity
str | Deinterlace
Parity - The field parity for de-interlacing, defaults to Auto.
- mode
String | "Off" | "Auto
Pixel Adaptive" - The deinterlacing mode. Defaults to AutoPixelAdaptive.
- parity
String | "Auto" | "Top
Field First" | "Bottom Field First" - The field parity for de-interlacing, defaults to Auto.
DeinterlaceMode, DeinterlaceModeArgs
- Off
- OffDisables de-interlacing of the source video.
- Auto
Pixel Adaptive - AutoPixelAdaptiveApply automatic pixel adaptive de-interlacing on each frame in the input video.
- Deinterlace
Mode Off - OffDisables de-interlacing of the source video.
- Deinterlace
Mode Auto Pixel Adaptive - AutoPixelAdaptiveApply automatic pixel adaptive de-interlacing on each frame in the input video.
- Off
- OffDisables de-interlacing of the source video.
- Auto
Pixel Adaptive - AutoPixelAdaptiveApply automatic pixel adaptive de-interlacing on each frame in the input video.
- Off
- OffDisables de-interlacing of the source video.
- Auto
Pixel Adaptive - AutoPixelAdaptiveApply automatic pixel adaptive de-interlacing on each frame in the input video.
- OFF
- OffDisables de-interlacing of the source video.
- AUTO_PIXEL_ADAPTIVE
- AutoPixelAdaptiveApply automatic pixel adaptive de-interlacing on each frame in the input video.
- "Off"
- OffDisables de-interlacing of the source video.
- "Auto
Pixel Adaptive" - AutoPixelAdaptiveApply automatic pixel adaptive de-interlacing on each frame in the input video.
DeinterlaceParity, DeinterlaceParityArgs
- Auto
- AutoAutomatically detect the order of fields
- Top
Field First - TopFieldFirstApply top field first processing of input video.
- Bottom
Field First - BottomFieldFirstApply bottom field first processing of input video.
- Deinterlace
Parity Auto - AutoAutomatically detect the order of fields
- Deinterlace
Parity Top Field First - TopFieldFirstApply top field first processing of input video.
- Deinterlace
Parity Bottom Field First - BottomFieldFirstApply bottom field first processing of input video.
- Auto
- AutoAutomatically detect the order of fields
- Top
Field First - TopFieldFirstApply top field first processing of input video.
- Bottom
Field First - BottomFieldFirstApply bottom field first processing of input video.
- Auto
- AutoAutomatically detect the order of fields
- Top
Field First - TopFieldFirstApply top field first processing of input video.
- Bottom
Field First - BottomFieldFirstApply bottom field first processing of input video.
- AUTO
- AutoAutomatically detect the order of fields
- TOP_FIELD_FIRST
- TopFieldFirstApply top field first processing of input video.
- BOTTOM_FIELD_FIRST
- BottomFieldFirstApply bottom field first processing of input video.
- "Auto"
- AutoAutomatically detect the order of fields
- "Top
Field First" - TopFieldFirstApply top field first processing of input video.
- "Bottom
Field First" - BottomFieldFirstApply bottom field first processing of input video.
DeinterlaceResponse, DeinterlaceResponseArgs
EncoderNamedPreset, EncoderNamedPresetArgs
- H264Single
Bitrate SD - H264SingleBitrateSDProduces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H264Single
Bitrate720p - H264SingleBitrate720pProduces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H264Single
Bitrate1080p - H264SingleBitrate1080pProduces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- Adaptive
Streaming - AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- AACGood
Quality Audio - AACGoodQualityAudioProduces a single MP4 file containing only stereo audio encoded at 192 kbps.
- Content
Aware Encoding Experimental - ContentAwareEncodingExperimentalExposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
- Content
Aware Encoding - ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- Copy
All Bitrate Non Interleaved - CopyAllBitrateNonInterleavedCopy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
- H264Multiple
Bitrate1080p - H264MultipleBitrate1080pProduces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
- H264Multiple
Bitrate720p - H264MultipleBitrate720pProduces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
- H264Multiple
Bitrate SD - H264MultipleBitrateSDProduces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
- H265Content
Aware Encoding - H265ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- H265Adaptive
Streaming - H265AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- H265Single
Bitrate720p - H265SingleBitrate720pProduces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H265Single
Bitrate1080p - H265SingleBitrate1080pProduces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H265Single
Bitrate4K - H265SingleBitrate4KProduces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- Encoder
Named Preset H264Single Bitrate SD - H264SingleBitrateSDProduces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- Encoder
Named Preset H264Single Bitrate720p - H264SingleBitrate720pProduces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- Encoder
Named Preset H264Single Bitrate1080p - H264SingleBitrate1080pProduces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- Encoder
Named Preset Adaptive Streaming - AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- Encoder
Named Preset AACGood Quality Audio - AACGoodQualityAudioProduces a single MP4 file containing only stereo audio encoded at 192 kbps.
- Encoder
Named Preset Content Aware Encoding Experimental - ContentAwareEncodingExperimentalExposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
- Encoder
Named Preset Content Aware Encoding - ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- Encoder
Named Preset Copy All Bitrate Non Interleaved - CopyAllBitrateNonInterleavedCopy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
- Encoder
Named Preset H264Multiple Bitrate1080p - H264MultipleBitrate1080pProduces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
- Encoder
Named Preset H264Multiple Bitrate720p - H264MultipleBitrate720pProduces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
- Encoder
Named Preset H264Multiple Bitrate SD - H264MultipleBitrateSDProduces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
- Encoder
Named Preset H265Content Aware Encoding - H265ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- Encoder
Named Preset H265Adaptive Streaming - H265AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- Encoder
Named Preset H265Single Bitrate720p - H265SingleBitrate720pProduces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- Encoder
Named Preset H265Single Bitrate1080p - H265SingleBitrate1080pProduces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- Encoder
Named Preset H265Single Bitrate4K - H265SingleBitrate4KProduces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H264Single
Bitrate SD - H264SingleBitrateSDProduces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H264Single
Bitrate720p - H264SingleBitrate720pProduces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H264Single
Bitrate1080p - H264SingleBitrate1080pProduces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- Adaptive
Streaming - AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- AACGood
Quality Audio - AACGoodQualityAudioProduces a single MP4 file containing only stereo audio encoded at 192 kbps.
- Content
Aware Encoding Experimental - ContentAwareEncodingExperimentalExposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
- Content
Aware Encoding - ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- Copy
All Bitrate Non Interleaved - CopyAllBitrateNonInterleavedCopy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
- H264Multiple
Bitrate1080p - H264MultipleBitrate1080pProduces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
- H264Multiple
Bitrate720p - H264MultipleBitrate720pProduces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
- H264Multiple
Bitrate SD - H264MultipleBitrateSDProduces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
- H265Content
Aware Encoding - H265ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- H265Adaptive
Streaming - H265AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- H265Single
Bitrate720p - H265SingleBitrate720pProduces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H265Single
Bitrate1080p - H265SingleBitrate1080pProduces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H265Single
Bitrate4K - H265SingleBitrate4KProduces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H264Single
Bitrate SD - H264SingleBitrateSDProduces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H264Single
Bitrate720p - H264SingleBitrate720pProduces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H264Single
Bitrate1080p - H264SingleBitrate1080pProduces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- Adaptive
Streaming - AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- AACGood
Quality Audio - AACGoodQualityAudioProduces a single MP4 file containing only stereo audio encoded at 192 kbps.
- Content
Aware Encoding Experimental - ContentAwareEncodingExperimentalExposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
- Content
Aware Encoding - ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- Copy
All Bitrate Non Interleaved - CopyAllBitrateNonInterleavedCopy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
- H264Multiple
Bitrate1080p - H264MultipleBitrate1080pProduces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
- H264Multiple
Bitrate720p - H264MultipleBitrate720pProduces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
- H264Multiple
Bitrate SD - H264MultipleBitrateSDProduces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
- H265Content
Aware Encoding - H265ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- H265Adaptive
Streaming - H265AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- H265Single
Bitrate720p - H265SingleBitrate720pProduces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H265Single
Bitrate1080p - H265SingleBitrate1080pProduces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H265Single
Bitrate4K - H265SingleBitrate4KProduces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H264_SINGLE_BITRATE_SD
- H264SingleBitrateSDProduces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H264_SINGLE_BITRATE720P
- H264SingleBitrate720pProduces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H264_SINGLE_BITRATE1080P
- H264SingleBitrate1080pProduces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- ADAPTIVE_STREAMING
- AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- AAC_GOOD_QUALITY_AUDIO
- AACGoodQualityAudioProduces a single MP4 file containing only stereo audio encoded at 192 kbps.
- CONTENT_AWARE_ENCODING_EXPERIMENTAL
- ContentAwareEncodingExperimentalExposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
- CONTENT_AWARE_ENCODING
- ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- COPY_ALL_BITRATE_NON_INTERLEAVED
- CopyAllBitrateNonInterleavedCopy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
- H264_MULTIPLE_BITRATE1080P
- H264MultipleBitrate1080pProduces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
- H264_MULTIPLE_BITRATE720P
- H264MultipleBitrate720pProduces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
- H264_MULTIPLE_BITRATE_SD
- H264MultipleBitrateSDProduces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
- H265_CONTENT_AWARE_ENCODING
- H265ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- H265_ADAPTIVE_STREAMING
- H265AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- H265_SINGLE_BITRATE720P
- H265SingleBitrate720pProduces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H265_SINGLE_BITRATE1080P
- H265SingleBitrate1080pProduces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- H265_SINGLE_BITRATE4_K
- H265SingleBitrate4KProduces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- "H264Single
Bitrate SD" - H264SingleBitrateSDProduces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- "H264Single
Bitrate720p" - H264SingleBitrate720pProduces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- "H264Single
Bitrate1080p" - H264SingleBitrate1080pProduces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- "Adaptive
Streaming" - AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- "AACGood
Quality Audio" - AACGoodQualityAudioProduces a single MP4 file containing only stereo audio encoded at 192 kbps.
- "Content
Aware Encoding Experimental" - ContentAwareEncodingExperimentalExposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
- "Content
Aware Encoding" - ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- "Copy
All Bitrate Non Interleaved" - CopyAllBitrateNonInterleavedCopy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
- "H264Multiple
Bitrate1080p" - H264MultipleBitrate1080pProduces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
- "H264Multiple
Bitrate720p" - H264MultipleBitrate720pProduces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
- "H264Multiple
Bitrate SD" - H264MultipleBitrateSDProduces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
- "H265Content
Aware Encoding" - H265ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
- "H265Adaptive
Streaming" - H265AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
- "H265Single
Bitrate720p" - H265SingleBitrate720pProduces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- "H265Single
Bitrate1080p" - H265SingleBitrate1080pProduces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
- "H265Single
Bitrate4K" - H265SingleBitrate4KProduces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
EntropyMode, EntropyModeArgs
- Cabac
- CabacContext Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
- Cavlc
- CavlcContext Adaptive Variable Length Coder (CAVLC) entropy encoding.
- Entropy
Mode Cabac - CabacContext Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
- Entropy
Mode Cavlc - CavlcContext Adaptive Variable Length Coder (CAVLC) entropy encoding.
- Cabac
- CabacContext Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
- Cavlc
- CavlcContext Adaptive Variable Length Coder (CAVLC) entropy encoding.
- Cabac
- CabacContext Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
- Cavlc
- CavlcContext Adaptive Variable Length Coder (CAVLC) entropy encoding.
- CABAC
- CabacContext Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
- CAVLC
- CavlcContext Adaptive Variable Length Coder (CAVLC) entropy encoding.
- "Cabac"
- CabacContext Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
- "Cavlc"
- CavlcContext Adaptive Variable Length Coder (CAVLC) entropy encoding.
FaceDetectorPreset, FaceDetectorPresetArgs
- Blur
Type string | Pulumi.Azure Native. Media. Blur Type - Blur type
- Experimental
Options Dictionary<string, string> - Dictionary containing key value pairs for parameters not exposed in the preset itself
- Mode
string | Pulumi.
Azure Native. Media. Face Redactor Mode - This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- Resolution
string | Pulumi.
Azure Native. Media. Analysis Resolution - Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
- Blur
Type string | BlurType - Blur type
- Experimental
Options map[string]string - Dictionary containing key value pairs for parameters not exposed in the preset itself
- Mode
string | Face
Redactor Mode - This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- Resolution
string | Analysis
Resolution - Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
- blur
Type String | BlurType - Blur type
- experimental
Options Map<String,String> - Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode
String | Face
Redactor Mode - This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- resolution
String | Analysis
Resolution - Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
- blur
Type string | BlurType - Blur type
- experimental
Options {[key: string]: string} - Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode
string | Face
Redactor Mode - This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- resolution
string | Analysis
Resolution - Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
- blur_
type str | BlurType - Blur type
- experimental_
options Mapping[str, str] - Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode
str | Face
Redactor Mode - This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- resolution
str | Analysis
Resolution - Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
- blur
Type String | "Box" | "Low" | "Med" | "High" | "Black" - Blur type
- experimental
Options Map<String> - Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode String | "Analyze" | "Redact" | "Combined"
- This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- resolution
String | "Source
Resolution" | "Standard Definition" - Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
FaceDetectorPresetResponse, FaceDetectorPresetResponseArgs
- Blur
Type string - Blur type
- Experimental
Options Dictionary<string, string> - Dictionary containing key value pairs for parameters not exposed in the preset itself
- Mode string
- This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- Resolution string
- Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
- Blur
Type string - Blur type
- Experimental
Options map[string]string - Dictionary containing key value pairs for parameters not exposed in the preset itself
- Mode string
- This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- Resolution string
- Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
- blur
Type String - Blur type
- experimental
Options Map<String,String> - Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode String
- This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- resolution String
- Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
- blur
Type string - Blur type
- experimental
Options {[key: string]: string} - Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode string
- This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- resolution string
- Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
- blur_
type str - Blur type
- experimental_
options Mapping[str, str] - Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode str
- This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- resolution str
- Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
- blur
Type String - Blur type
- experimental
Options Map<String> - Dictionary containing key value pairs for parameters not exposed in the preset itself
- mode String
- This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
- resolution String
- Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
FaceRedactorMode, FaceRedactorModeArgs
- Analyze
- AnalyzeAnalyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
- Redact
- RedactRedact mode consumes the metadata file from Analyze mode and redacts the faces found.
- Combined
- CombinedCombined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.
- Face
Redactor Mode Analyze - AnalyzeAnalyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
- Face
Redactor Mode Redact - RedactRedact mode consumes the metadata file from Analyze mode and redacts the faces found.
- Face
Redactor Mode Combined - CombinedCombined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.
- Analyze
- AnalyzeAnalyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
- Redact
- RedactRedact mode consumes the metadata file from Analyze mode and redacts the faces found.
- Combined
- CombinedCombined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.
- Analyze
- AnalyzeAnalyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
- Redact
- RedactRedact mode consumes the metadata file from Analyze mode and redacts the faces found.
- Combined
- CombinedCombined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.
- ANALYZE
- AnalyzeAnalyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
- REDACT
- RedactRedact mode consumes the metadata file from Analyze mode and redacts the faces found.
- COMBINED
- CombinedCombined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.
- "Analyze"
- AnalyzeAnalyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
- "Redact"
- RedactRedact mode consumes the metadata file from Analyze mode and redacts the faces found.
- "Combined"
- CombinedCombined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.
Filters, FiltersArgs
- Crop
Pulumi.
Azure Native. Media. Inputs. Rectangle - The parameters for the rectangular window with which to crop the input video.
- Deinterlace
Pulumi.
Azure Native. Media. Inputs. Deinterlace - The de-interlacing settings.
- Overlays
List<Union<Pulumi.
Azure Native. Media. Inputs. Audio Overlay, Pulumi. Azure Native. Media. Inputs. Video Overlay>> - The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- Rotation
string | Pulumi.
Azure Native. Media. Rotation - The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
- Crop Rectangle
- The parameters for the rectangular window with which to crop the input video.
- Deinterlace Deinterlace
- The de-interlacing settings.
- Overlays []interface{}
- The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- Rotation string | Rotation
- The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
- crop Rectangle
- The parameters for the rectangular window with which to crop the input video.
- deinterlace Deinterlace
- The de-interlacing settings.
- overlays
List<Either<Audio
Overlay,Video Overlay>> - The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- rotation String | Rotation
- The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
- crop Rectangle
- The parameters for the rectangular window with which to crop the input video.
- deinterlace Deinterlace
- The de-interlacing settings.
- overlays
(Audio
Overlay | Video Overlay)[] - The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- rotation string | Rotation
- The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
- crop Rectangle
- The parameters for the rectangular window with which to crop the input video.
- deinterlace Deinterlace
- The de-interlacing settings.
- overlays
Sequence[Union[Audio
Overlay, Video Overlay]] - The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- rotation str | Rotation
- The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
- crop Property Map
- The parameters for the rectangular window with which to crop the input video.
- deinterlace Property Map
- The de-interlacing settings.
- overlays List<Property Map | Property Map>
- The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- rotation String | "Auto" | "None" | "Rotate0" | "Rotate90" | "Rotate180" | "Rotate270"
- The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
FiltersResponse, FiltersResponseArgs
- Crop
Pulumi.
Azure Native. Media. Inputs. Rectangle Response - The parameters for the rectangular window with which to crop the input video.
- Deinterlace
Pulumi.
Azure Native. Media. Inputs. Deinterlace Response - The de-interlacing settings.
- Overlays
List<Union<Pulumi.
Azure Native. Media. Inputs. Audio Overlay Response, Pulumi. Azure Native. Media. Inputs. Video Overlay Response>> - The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- Rotation string
- The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
- Crop
Rectangle
Response - The parameters for the rectangular window with which to crop the input video.
- Deinterlace
Deinterlace
Response - The de-interlacing settings.
- Overlays []interface{}
- The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- Rotation string
- The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
- crop
Rectangle
Response - The parameters for the rectangular window with which to crop the input video.
- deinterlace
Deinterlace
Response - The de-interlacing settings.
- overlays
List<Either<Audio
Overlay Response,Video Overlay Response>> - The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- rotation String
- The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
- crop
Rectangle
Response - The parameters for the rectangular window with which to crop the input video.
- deinterlace
Deinterlace
Response - The de-interlacing settings.
- overlays
(Audio
Overlay Response | Video Overlay Response)[] - The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- rotation string
- The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
- crop
Rectangle
Response - The parameters for the rectangular window with which to crop the input video.
- deinterlace
Deinterlace
Response - The de-interlacing settings.
- overlays
Sequence[Union[Audio
Overlay Response, Video Overlay Response]] - The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- rotation str
- The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
- crop Property Map
- The parameters for the rectangular window with which to crop the input video.
- deinterlace Property Map
- The de-interlacing settings.
- overlays List<Property Map | Property Map>
- The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
- rotation String
- The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
H264Complexity, H264ComplexityArgs
- Speed
- SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- Balanced
- BalancedTells the encoder to use settings that achieve a balance between speed and quality.
- Quality
- QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
- H264Complexity
Speed - SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- H264Complexity
Balanced - BalancedTells the encoder to use settings that achieve a balance between speed and quality.
- H264Complexity
Quality - QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
- Speed
- SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- Balanced
- BalancedTells the encoder to use settings that achieve a balance between speed and quality.
- Quality
- QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
- Speed
- SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- Balanced
- BalancedTells the encoder to use settings that achieve a balance between speed and quality.
- Quality
- QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
- SPEED
- SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- BALANCED
- BalancedTells the encoder to use settings that achieve a balance between speed and quality.
- QUALITY
- QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
- "Speed"
- SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- "Balanced"
- BalancedTells the encoder to use settings that achieve a balance between speed and quality.
- "Quality"
- QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
H264Layer, H264LayerArgs
- Bitrate int
- The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- Adaptive
BFrame bool - Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- BFrames int
- The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- Buffer
Window string - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- Entropy
Mode string | Pulumi.Azure Native. Media. Entropy Mode - The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- Frame
Rate string - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- Height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Level string
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- Max
Bitrate int - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- Profile
string | Pulumi.
Azure Native. Media. H264Video Profile - We currently support Baseline, Main, High, High422, High444. Default is Auto.
- Reference
Frames int - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- Slices int
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- Width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- Bitrate int
- The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- Adaptive
BFrame bool - Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- BFrames int
- The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- Buffer
Window string - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- Entropy
Mode string | EntropyMode - The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- Frame
Rate string - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- Height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Level string
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- Max
Bitrate int - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- Profile
string | H264Video
Profile - We currently support Baseline, Main, High, High422, High444. Default is Auto.
- Reference
Frames int - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- Slices int
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- Width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate Integer
- The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- adaptive
BFrame Boolean - Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames Integer - The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window String - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- entropy
Mode String | EntropyMode - The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- frame
Rate String - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height String
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level String
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate Integer - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile
String | H264Video
Profile - We currently support Baseline, Main, High, High422, High444. Default is Auto.
- reference
Frames Integer - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices Integer
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width String
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate number
- The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- adaptive
BFrame boolean - Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames number - The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window string - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- entropy
Mode string | EntropyMode - The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- frame
Rate string - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level string
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate number - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile
string | H264Video
Profile - We currently support Baseline, Main, High, High422, High444. Default is Auto.
- reference
Frames number - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices number
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate int
- The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- adaptive_
b_ boolframe - Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b_
frames int - The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer_
window str - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- entropy_
mode str | EntropyMode - The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- frame_
rate str - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height str
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label str
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level str
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max_
bitrate int - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile
str | H264Video
Profile - We currently support Baseline, Main, High, High422, High444. Default is Auto.
- reference_
frames int - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices int
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width str
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate Number
- The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- adaptive
BFrame Boolean - Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames Number - The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window String - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- entropy
Mode String | "Cabac" | "Cavlc" - The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- frame
Rate String - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height String
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level String
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate Number - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile String | "Auto" | "Baseline" | "Main" | "High" | "High422" | "High444"
- We currently support Baseline, Main, High, High422, High444. Default is Auto.
- reference
Frames Number - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices Number
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width String
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
H264LayerResponse, H264LayerResponseArgs
- Bitrate int
- The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- Adaptive
BFrame bool - Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- BFrames int
- The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- Buffer
Window string - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- Entropy
Mode string - The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- Frame
Rate string - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- Height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Level string
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- Max
Bitrate int - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- Profile string
- We currently support Baseline, Main, High, High422, High444. Default is Auto.
- Reference
Frames int - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- Slices int
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- Width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- Bitrate int
- The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- Adaptive
BFrame bool - Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- BFrames int
- The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- Buffer
Window string - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- Entropy
Mode string - The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- Frame
Rate string - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- Height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Level string
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- Max
Bitrate int - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- Profile string
- We currently support Baseline, Main, High, High422, High444. Default is Auto.
- Reference
Frames int - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- Slices int
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- Width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate Integer
- The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- adaptive
BFrame Boolean - Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames Integer - The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window String - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- entropy
Mode String - The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- frame
Rate String - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height String
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level String
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate Integer - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile String
- We currently support Baseline, Main, High, High422, High444. Default is Auto.
- reference
Frames Integer - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices Integer
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width String
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate number
- The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- adaptive
BFrame boolean - Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames number - The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window string - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- entropy
Mode string - The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- frame
Rate string - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level string
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate number - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile string
- We currently support Baseline, Main, High, High422, High444. Default is Auto.
- reference
Frames number - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices number
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate int
- The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- adaptive_
b_ boolframe - Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b_
frames int - The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer_
window str - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- entropy_
mode str - The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- frame_
rate str - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height str
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label str
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level str
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max_
bitrate int - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile str
- We currently support Baseline, Main, High, High422, High444. Default is Auto.
- reference_
frames int - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices int
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width str
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate Number
- The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
- adaptive
BFrame Boolean - Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames Number - The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window String - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- entropy
Mode String - The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
- frame
Rate String - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height String
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level String
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate Number - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile String
- We currently support Baseline, Main, High, High422, High444. Default is Auto.
- reference
Frames Number - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices Number
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width String
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
H264Video, H264VideoArgs
- Complexity
string | Pulumi.
Azure Native. Media. H264Complexity - Tells the encoder how to choose its encoding settings. The default value is Balanced.
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Layers
List<Pulumi.
Azure Native. Media. Inputs. H264Layer> - The collection of output H.264 layers to be produced by the encoder.
- Scene
Change boolDetection - Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- Stretch
Mode string | Pulumi.Azure Native. Media. Stretch Mode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string | Pulumi.Azure Native. Media. Video Sync Mode - The Video Sync Mode
- Complexity string | H264Complexity
- Tells the encoder how to choose its encoding settings. The default value is Balanced.
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Layers []H264Layer
- The collection of output H.264 layers to be produced by the encoder.
- Scene
Change boolDetection - Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- Stretch
Mode string | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string | VideoSync Mode - The Video Sync Mode
- complexity String | H264Complexity
- Tells the encoder how to choose its encoding settings. The default value is Balanced.
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- layers List<H264Layer>
- The collection of output H.264 layers to be produced by the encoder.
- scene
Change BooleanDetection - Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode String | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String | VideoSync Mode - The Video Sync Mode
- complexity string | H264Complexity
- Tells the encoder how to choose its encoding settings. The default value is Balanced.
- key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label string
- An optional label for the codec. The label can be used to control muxing behavior.
- layers H264Layer[]
- The collection of output H.264 layers to be produced by the encoder.
- scene
Change booleanDetection - Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode string | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode string | VideoSync Mode - The Video Sync Mode
- complexity str | H264Complexity
- Tells the encoder how to choose its encoding settings. The default value is Balanced.
- key_
frame_ strinterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label str
- An optional label for the codec. The label can be used to control muxing behavior.
- layers Sequence[H264Layer]
- The collection of output H.264 layers to be produced by the encoder.
- scene_
change_ booldetection - Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch_
mode str | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync_
mode str | VideoSync Mode - The Video Sync Mode
- complexity String | "Speed" | "Balanced" | "Quality"
- Tells the encoder how to choose its encoding settings. The default value is Balanced.
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- layers List<Property Map>
- The collection of output H.264 layers to be produced by the encoder.
- scene
Change BooleanDetection - Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode String | "None" | "AutoSize" | "Auto Fit" - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String | "Auto" | "Passthrough" | "Cfr" | "Vfr" - The Video Sync Mode
H264VideoProfile, H264VideoProfileArgs
- Auto
- AutoTells the encoder to automatically determine the appropriate H.264 profile.
- Baseline
- BaselineBaseline profile
- Main
- MainMain profile
- High
- HighHigh profile.
- High422
- High422High 4:2:2 profile.
- High444
- High444High 4:4:4 predictive profile.
- H264Video
Profile Auto - AutoTells the encoder to automatically determine the appropriate H.264 profile.
- H264Video
Profile Baseline - BaselineBaseline profile
- H264Video
Profile Main - MainMain profile
- H264Video
Profile High - HighHigh profile.
- H264Video
Profile High422 - High422High 4:2:2 profile.
- H264Video
Profile High444 - High444High 4:4:4 predictive profile.
- Auto
- AutoTells the encoder to automatically determine the appropriate H.264 profile.
- Baseline
- BaselineBaseline profile
- Main
- MainMain profile
- High
- HighHigh profile.
- High422
- High422High 4:2:2 profile.
- High444
- High444High 4:4:4 predictive profile.
- Auto
- AutoTells the encoder to automatically determine the appropriate H.264 profile.
- Baseline
- BaselineBaseline profile
- Main
- MainMain profile
- High
- HighHigh profile.
- High422
- High422High 4:2:2 profile.
- High444
- High444High 4:4:4 predictive profile.
- AUTO
- AutoTells the encoder to automatically determine the appropriate H.264 profile.
- BASELINE
- BaselineBaseline profile
- MAIN
- MainMain profile
- HIGH
- HighHigh profile.
- HIGH422
- High422High 4:2:2 profile.
- HIGH444
- High444High 4:4:4 predictive profile.
- "Auto"
- AutoTells the encoder to automatically determine the appropriate H.264 profile.
- "Baseline"
- BaselineBaseline profile
- "Main"
- MainMain profile
- "High"
- HighHigh profile.
- "High422"
- High422High 4:2:2 profile.
- "High444"
- High444High 4:4:4 predictive profile.
H264VideoResponse, H264VideoResponseArgs
- Complexity string
- Tells the encoder how to choose its encoding settings. The default value is Balanced.
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Layers
List<Pulumi.
Azure Native. Media. Inputs. H264Layer Response> - The collection of output H.264 layers to be produced by the encoder.
- Scene
Change boolDetection - Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- Stretch
Mode string - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string - The Video Sync Mode
- Complexity string
- Tells the encoder how to choose its encoding settings. The default value is Balanced.
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Layers
[]H264Layer
Response - The collection of output H.264 layers to be produced by the encoder.
- Scene
Change boolDetection - Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- Stretch
Mode string - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string - The Video Sync Mode
- complexity String
- Tells the encoder how to choose its encoding settings. The default value is Balanced.
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- layers
List<H264Layer
Response> - The collection of output H.264 layers to be produced by the encoder.
- scene
Change BooleanDetection - Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode String - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String - The Video Sync Mode
- complexity string
- Tells the encoder how to choose its encoding settings. The default value is Balanced.
- key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label string
- An optional label for the codec. The label can be used to control muxing behavior.
- layers
H264Layer
Response[] - The collection of output H.264 layers to be produced by the encoder.
- scene
Change booleanDetection - Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode string - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode string - The Video Sync Mode
- complexity str
- Tells the encoder how to choose its encoding settings. The default value is Balanced.
- key_
frame_ strinterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label str
- An optional label for the codec. The label can be used to control muxing behavior.
- layers
Sequence[H264Layer
Response] - The collection of output H.264 layers to be produced by the encoder.
- scene_
change_ booldetection - Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch_
mode str - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync_
mode str - The Video Sync Mode
- complexity String
- Tells the encoder how to choose its encoding settings. The default value is Balanced.
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- layers List<Property Map>
- The collection of output H.264 layers to be produced by the encoder.
- scene
Change BooleanDetection - Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode String - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String - The Video Sync Mode
H265Complexity, H265ComplexityArgs
- Speed
- SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- Balanced
- BalancedTells the encoder to use settings that achieve a balance between speed and quality.
- Quality
- QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
- H265Complexity
Speed - SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- H265Complexity
Balanced - BalancedTells the encoder to use settings that achieve a balance between speed and quality.
- H265Complexity
Quality - QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
- Speed
- SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- Balanced
- BalancedTells the encoder to use settings that achieve a balance between speed and quality.
- Quality
- QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
- Speed
- SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- Balanced
- BalancedTells the encoder to use settings that achieve a balance between speed and quality.
- Quality
- QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
- SPEED
- SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- BALANCED
- BalancedTells the encoder to use settings that achieve a balance between speed and quality.
- QUALITY
- QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
- "Speed"
- SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
- "Balanced"
- BalancedTells the encoder to use settings that achieve a balance between speed and quality.
- "Quality"
- QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
H265Layer, H265LayerArgs
- Bitrate int
- The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- Adaptive
BFrame bool - Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- BFrames int
- The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- Buffer
Window string - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- Frame
Rate string - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- Height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Level string
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- Max
Bitrate int - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- Profile
string | Pulumi.
Azure Native. Media. H265Video Profile - We currently support Main. Default is Auto.
- Reference
Frames int - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- Slices int
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- Width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- Bitrate int
- The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- Adaptive
BFrame bool - Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- BFrames int
- The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- Buffer
Window string - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- Frame
Rate string - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- Height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Level string
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- Max
Bitrate int - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- Profile
string | H265Video
Profile - We currently support Main. Default is Auto.
- Reference
Frames int - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- Slices int
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- Width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate Integer
- The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- adaptive
BFrame Boolean - Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames Integer - The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window String - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- frame
Rate String - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height String
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level String
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate Integer - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile
String | H265Video
Profile - We currently support Main. Default is Auto.
- reference
Frames Integer - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices Integer
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width String
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate number
- The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- adaptive
BFrame boolean - Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames number - The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window string - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- frame
Rate string - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level string
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate number - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile
string | H265Video
Profile - We currently support Main. Default is Auto.
- reference
Frames number - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices number
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate int
- The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- adaptive_
b_ boolframe - Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b_
frames int - The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer_
window str - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- frame_
rate str - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height str
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label str
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level str
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max_
bitrate int - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile
str | H265Video
Profile - We currently support Main. Default is Auto.
- reference_
frames int - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices int
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width str
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate Number
- The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- adaptive
BFrame Boolean - Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames Number - The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window String - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- frame
Rate String - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height String
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level String
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate Number - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile String | "Auto" | "Main"
- We currently support Main. Default is Auto.
- reference
Frames Number - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices Number
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width String
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
H265LayerResponse, H265LayerResponseArgs
- Bitrate int
- The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- Adaptive
BFrame bool - Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- BFrames int
- The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- Buffer
Window string - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- Frame
Rate string - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- Height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Level string
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- Max
Bitrate int - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- Profile string
- We currently support Main. Default is Auto.
- Reference
Frames int - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- Slices int
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- Width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- Bitrate int
- The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- Adaptive
BFrame bool - Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- BFrames int
- The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- Buffer
Window string - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- Frame
Rate string - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- Height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Level string
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- Max
Bitrate int - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- Profile string
- We currently support Main. Default is Auto.
- Reference
Frames int - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- Slices int
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- Width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate Integer
- The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- adaptive
BFrame Boolean - Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames Integer - The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window String - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- frame
Rate String - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height String
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level String
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate Integer - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile String
- We currently support Main. Default is Auto.
- reference
Frames Integer - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices Integer
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width String
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate number
- The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- adaptive
BFrame boolean - Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames number - The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window string - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- frame
Rate string - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level string
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate number - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile string
- We currently support Main. Default is Auto.
- reference
Frames number - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices number
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate int
- The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- adaptive_
b_ boolframe - Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b_
frames int - The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer_
window str - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- frame_
rate str - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height str
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label str
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level str
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max_
bitrate int - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile str
- We currently support Main. Default is Auto.
- reference_
frames int - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices int
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width str
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- bitrate Number
- The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
- adaptive
BFrame Boolean - Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
- b
Frames Number - The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
- buffer
Window String - The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
- frame
Rate String - The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
- height String
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- level String
- We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
- max
Bitrate Number - The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
- profile String
- We currently support Main. Default is Auto.
- reference
Frames Number - The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
- slices Number
- The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
- width String
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
H265Video, H265VideoArgs
- Complexity
string | Pulumi.
Azure Native. Media. H265Complexity - Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Layers
List<Pulumi.
Azure Native. Media. Inputs. H265Layer> - The collection of output H.265 layers to be produced by the encoder.
- Scene
Change boolDetection - Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- Stretch
Mode string | Pulumi.Azure Native. Media. Stretch Mode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string | Pulumi.Azure Native. Media. Video Sync Mode - The Video Sync Mode
- Complexity string | H265Complexity
- Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Layers []H265Layer
- The collection of output H.265 layers to be produced by the encoder.
- Scene
Change boolDetection - Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- Stretch
Mode string | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string | VideoSync Mode - The Video Sync Mode
- complexity String | H265Complexity
- Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- layers List<H265Layer>
- The collection of output H.265 layers to be produced by the encoder.
- scene
Change BooleanDetection - Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode String | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String | VideoSync Mode - The Video Sync Mode
- complexity string | H265Complexity
- Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label string
- An optional label for the codec. The label can be used to control muxing behavior.
- layers H265Layer[]
- The collection of output H.265 layers to be produced by the encoder.
- scene
Change booleanDetection - Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode string | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode string | VideoSync Mode - The Video Sync Mode
- complexity str | H265Complexity
- Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- key_
frame_ strinterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label str
- An optional label for the codec. The label can be used to control muxing behavior.
- layers Sequence[H265Layer]
- The collection of output H.265 layers to be produced by the encoder.
- scene_
change_ booldetection - Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch_
mode str | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync_
mode str | VideoSync Mode - The Video Sync Mode
- complexity String | "Speed" | "Balanced" | "Quality"
- Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- layers List<Property Map>
- The collection of output H.265 layers to be produced by the encoder.
- scene
Change BooleanDetection - Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode String | "None" | "AutoSize" | "Auto Fit" - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String | "Auto" | "Passthrough" | "Cfr" | "Vfr" - The Video Sync Mode
H265VideoProfile, H265VideoProfileArgs
- Auto
- AutoTells the encoder to automatically determine the appropriate H.265 profile.
- Main
- MainMain profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)
- H265Video
Profile Auto - AutoTells the encoder to automatically determine the appropriate H.265 profile.
- H265Video
Profile Main - MainMain profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)
- Auto
- AutoTells the encoder to automatically determine the appropriate H.265 profile.
- Main
- MainMain profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)
- Auto
- AutoTells the encoder to automatically determine the appropriate H.265 profile.
- Main
- MainMain profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)
- AUTO
- AutoTells the encoder to automatically determine the appropriate H.265 profile.
- MAIN
- MainMain profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)
- "Auto"
- AutoTells the encoder to automatically determine the appropriate H.265 profile.
- "Main"
- MainMain profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)
H265VideoResponse, H265VideoResponseArgs
- Complexity string
- Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Layers
List<Pulumi.
Azure Native. Media. Inputs. H265Layer Response> - The collection of output H.265 layers to be produced by the encoder.
- Scene
Change boolDetection - Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- Stretch
Mode string - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string - The Video Sync Mode
- Complexity string
- Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Layers
[]H265Layer
Response - The collection of output H.265 layers to be produced by the encoder.
- Scene
Change boolDetection - Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- Stretch
Mode string - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string - The Video Sync Mode
- complexity String
- Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- layers
List<H265Layer
Response> - The collection of output H.265 layers to be produced by the encoder.
- scene
Change BooleanDetection - Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode String - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String - The Video Sync Mode
- complexity string
- Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label string
- An optional label for the codec. The label can be used to control muxing behavior.
- layers
H265Layer
Response[] - The collection of output H.265 layers to be produced by the encoder.
- scene
Change booleanDetection - Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode string - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode string - The Video Sync Mode
- complexity str
- Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- key_
frame_ strinterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label str
- An optional label for the codec. The label can be used to control muxing behavior.
- layers
Sequence[H265Layer
Response] - The collection of output H.265 layers to be produced by the encoder.
- scene_
change_ booldetection - Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch_
mode str - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync_
mode str - The Video Sync Mode
- complexity String
- Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- layers List<Property Map>
- The collection of output H.265 layers to be produced by the encoder.
- scene
Change BooleanDetection - Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
- stretch
Mode String - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String - The Video Sync Mode
Image, ImageArgs
- Start string
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Range string
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Step string
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string | Pulumi.Azure Native. Media. Stretch Mode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string | Pulumi.Azure Native. Media. Video Sync Mode - The Video Sync Mode
- Start string
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Range string
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Step string
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string | VideoSync Mode - The Video Sync Mode
- start String
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- range String
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step String
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String | VideoSync Mode - The Video Sync Mode
- start string
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label string
- An optional label for the codec. The label can be used to control muxing behavior.
- range string
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step string
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode string | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode string | VideoSync Mode - The Video Sync Mode
- start str
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key_
frame_ strinterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label str
- An optional label for the codec. The label can be used to control muxing behavior.
- range str
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step str
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch_
mode str | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync_
mode str | VideoSync Mode - The Video Sync Mode
- start String
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- range String
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step String
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String | "None" | "AutoSize" | "Auto Fit" - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String | "Auto" | "Passthrough" | "Cfr" | "Vfr" - The Video Sync Mode
ImageFormat, ImageFormatArgs
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename_
pattern str - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
ImageFormatResponse, ImageFormatResponseArgs
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename_
pattern str - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
ImageResponse, ImageResponseArgs
- Start string
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Range string
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Step string
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string - The Video Sync Mode
- Start string
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Range string
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Step string
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string - The Video Sync Mode
- start String
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- range String
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step String
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String - The Video Sync Mode
- start string
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label string
- An optional label for the codec. The label can be used to control muxing behavior.
- range string
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step string
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode string - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode string - The Video Sync Mode
- start str
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key_
frame_ strinterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label str
- An optional label for the codec. The label can be used to control muxing behavior.
- range str
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step str
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch_
mode str - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync_
mode str - The Video Sync Mode
- start String
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- range String
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step String
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String - The Video Sync Mode
InsightsType, InsightsTypeArgs
- Audio
Insights Only - AudioInsightsOnlyGenerate audio only insights. Ignore video even if present. Fails if no audio is present.
- Video
Insights Only - VideoInsightsOnlyGenerate video only insights. Ignore audio if present. Fails if no video is present.
- All
Insights - AllInsightsGenerate both audio and video insights. Fails if either audio or video Insights fail.
- Insights
Type Audio Insights Only - AudioInsightsOnlyGenerate audio only insights. Ignore video even if present. Fails if no audio is present.
- Insights
Type Video Insights Only - VideoInsightsOnlyGenerate video only insights. Ignore audio if present. Fails if no video is present.
- Insights
Type All Insights - AllInsightsGenerate both audio and video insights. Fails if either audio or video Insights fail.
- Audio
Insights Only - AudioInsightsOnlyGenerate audio only insights. Ignore video even if present. Fails if no audio is present.
- Video
Insights Only - VideoInsightsOnlyGenerate video only insights. Ignore audio if present. Fails if no video is present.
- All
Insights - AllInsightsGenerate both audio and video insights. Fails if either audio or video Insights fail.
- Audio
Insights Only - AudioInsightsOnlyGenerate audio only insights. Ignore video even if present. Fails if no audio is present.
- Video
Insights Only - VideoInsightsOnlyGenerate video only insights. Ignore audio if present. Fails if no video is present.
- All
Insights - AllInsightsGenerate both audio and video insights. Fails if either audio or video Insights fail.
- AUDIO_INSIGHTS_ONLY
- AudioInsightsOnlyGenerate audio only insights. Ignore video even if present. Fails if no audio is present.
- VIDEO_INSIGHTS_ONLY
- VideoInsightsOnlyGenerate video only insights. Ignore audio if present. Fails if no video is present.
- ALL_INSIGHTS
- AllInsightsGenerate both audio and video insights. Fails if either audio or video Insights fail.
- "Audio
Insights Only" - AudioInsightsOnlyGenerate audio only insights. Ignore video even if present. Fails if no audio is present.
- "Video
Insights Only" - VideoInsightsOnlyGenerate video only insights. Ignore audio if present. Fails if no video is present.
- "All
Insights" - AllInsightsGenerate both audio and video insights. Fails if either audio or video Insights fail.
JpgFormat, JpgFormatArgs
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename_
pattern str - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
JpgFormatResponse, JpgFormatResponseArgs
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename_
pattern str - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
JpgImage, JpgImageArgs
- Start string
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Layers
List<Pulumi.
Azure Native. Media. Inputs. Jpg Layer> - A collection of output JPEG image layers to be produced by the encoder.
- Range string
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Sprite
Column int - Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- Step string
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string | Pulumi.Azure Native. Media. Stretch Mode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string | Pulumi.Azure Native. Media. Video Sync Mode - The Video Sync Mode
- Start string
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Layers
[]Jpg
Layer - A collection of output JPEG image layers to be produced by the encoder.
- Range string
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Sprite
Column int - Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- Step string
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string | VideoSync Mode - The Video Sync Mode
- start String
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- layers
List<Jpg
Layer> - A collection of output JPEG image layers to be produced by the encoder.
- range String
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- sprite
Column Integer - Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- step String
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String | VideoSync Mode - The Video Sync Mode
- start string
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label string
- An optional label for the codec. The label can be used to control muxing behavior.
- layers
Jpg
Layer[] - A collection of output JPEG image layers to be produced by the encoder.
- range string
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- sprite
Column number - Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- step string
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode string | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode string | VideoSync Mode - The Video Sync Mode
- start str
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key_
frame_ strinterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label str
- An optional label for the codec. The label can be used to control muxing behavior.
- layers
Sequence[Jpg
Layer] - A collection of output JPEG image layers to be produced by the encoder.
- range str
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- sprite_
column int - Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- step str
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch_
mode str | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync_
mode str | VideoSync Mode - The Video Sync Mode
- start String
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- layers List<Property Map>
- A collection of output JPEG image layers to be produced by the encoder.
- range String
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- sprite
Column Number - Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- step String
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String | "None" | "AutoSize" | "Auto Fit" - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String | "Auto" | "Passthrough" | "Cfr" | "Vfr" - The Video Sync Mode
JpgImageResponse, JpgImageResponseArgs
- Start string
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Layers
List<Pulumi.
Azure Native. Media. Inputs. Jpg Layer Response> - A collection of output JPEG image layers to be produced by the encoder.
- Range string
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Sprite
Column int - Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- Step string
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string - The Video Sync Mode
- Start string
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Layers
[]Jpg
Layer Response - A collection of output JPEG image layers to be produced by the encoder.
- Range string
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Sprite
Column int - Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- Step string
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string - The Video Sync Mode
- start String
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- layers
List<Jpg
Layer Response> - A collection of output JPEG image layers to be produced by the encoder.
- range String
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- sprite
Column Integer - Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- step String
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String - The Video Sync Mode
- start string
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label string
- An optional label for the codec. The label can be used to control muxing behavior.
- layers
Jpg
Layer Response[] - A collection of output JPEG image layers to be produced by the encoder.
- range string
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- sprite
Column number - Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- step string
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode string - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode string - The Video Sync Mode
- start str
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key_
frame_ strinterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label str
- An optional label for the codec. The label can be used to control muxing behavior.
- layers
Sequence[Jpg
Layer Response] - A collection of output JPEG image layers to be produced by the encoder.
- range str
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- sprite_
column int - Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- step str
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch_
mode str - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync_
mode str - The Video Sync Mode
- start String
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- layers List<Property Map>
- A collection of output JPEG image layers to be produced by the encoder.
- range String
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- sprite
Column Number - Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
- step String
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String - The Video Sync Mode
JpgLayer, JpgLayerArgs
- Height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Quality int
- The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- Width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- Height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Quality int
- The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- Width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height String
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- quality Integer
- The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- width String
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- quality number
- The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height str
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label str
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- quality int
- The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- width str
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height String
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- quality Number
- The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- width String
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
JpgLayerResponse, JpgLayerResponseArgs
- Height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Quality int
- The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- Width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- Height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Quality int
- The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- Width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height String
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- quality Integer
- The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- width String
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- quality number
- The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height str
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label str
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- quality int
- The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- width str
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height String
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- quality Number
- The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
- width String
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
Mp4Format, Mp4FormatArgs
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Output
Files List<Pulumi.Azure Native. Media. Inputs. Output File> - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Output
Files []OutputFile - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files List<OutputFile> - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files OutputFile[] - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename_
pattern str - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output_
files Sequence[OutputFile] - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files List<Property Map> - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
Mp4FormatResponse, Mp4FormatResponseArgs
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Output
Files List<Pulumi.Azure Native. Media. Inputs. Output File Response> - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Output
Files []OutputFile Response - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files List<OutputFile Response> - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files OutputFile Response[] - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename_
pattern str - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output_
files Sequence[OutputFile Response] - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files List<Property Map> - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
MultiBitrateFormat, MultiBitrateFormatArgs
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Output
Files List<Pulumi.Azure Native. Media. Inputs. Output File> - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Output
Files []OutputFile - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files List<OutputFile> - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files OutputFile[] - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename_
pattern str - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output_
files Sequence[OutputFile] - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files List<Property Map> - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
MultiBitrateFormatResponse, MultiBitrateFormatResponseArgs
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Output
Files List<Pulumi.Azure Native. Media. Inputs. Output File Response> - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Output
Files []OutputFile Response - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files List<OutputFile Response> - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files OutputFile Response[] - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename_
pattern str - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output_
files Sequence[OutputFile Response] - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files List<Property Map> - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
OnErrorType, OnErrorTypeArgs
- Stop
Processing Job - StopProcessingJobTells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
- Continue
Job - ContinueJobTells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.
- On
Error Type Stop Processing Job - StopProcessingJobTells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
- On
Error Type Continue Job - ContinueJobTells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.
- Stop
Processing Job - StopProcessingJobTells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
- Continue
Job - ContinueJobTells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.
- Stop
Processing Job - StopProcessingJobTells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
- Continue
Job - ContinueJobTells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.
- STOP_PROCESSING_JOB
- StopProcessingJobTells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
- CONTINUE_JOB
- ContinueJobTells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.
- "Stop
Processing Job" - StopProcessingJobTells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
- "Continue
Job" - ContinueJobTells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.
OutputFile, OutputFileArgs
- Labels List<string>
- The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
- Labels []string
- The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
- labels List<String>
- The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
- labels string[]
- The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
- labels Sequence[str]
- The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
- labels List<String>
- The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
OutputFileResponse, OutputFileResponseArgs
- Labels List<string>
- The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
- Labels []string
- The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
- labels List<String>
- The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
- labels string[]
- The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
- labels Sequence[str]
- The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
- labels List<String>
- The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
PngFormat, PngFormatArgs
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename_
pattern str - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
PngFormatResponse, PngFormatResponseArgs
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename_
pattern str - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
PngImage, PngImageArgs
- Start string
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Layers
List<Pulumi.
Azure Native. Media. Inputs. Png Layer> - A collection of output PNG image layers to be produced by the encoder.
- Range string
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Step string
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string | Pulumi.Azure Native. Media. Stretch Mode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string | Pulumi.Azure Native. Media. Video Sync Mode - The Video Sync Mode
- Start string
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Layers
[]Png
Layer - A collection of output PNG image layers to be produced by the encoder.
- Range string
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Step string
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string | VideoSync Mode - The Video Sync Mode
- start String
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- layers
List<Png
Layer> - A collection of output PNG image layers to be produced by the encoder.
- range String
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step String
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String | VideoSync Mode - The Video Sync Mode
- start string
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label string
- An optional label for the codec. The label can be used to control muxing behavior.
- layers
Png
Layer[] - A collection of output PNG image layers to be produced by the encoder.
- range string
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step string
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode string | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode string | VideoSync Mode - The Video Sync Mode
- start str
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key_
frame_ strinterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label str
- An optional label for the codec. The label can be used to control muxing behavior.
- layers
Sequence[Png
Layer] - A collection of output PNG image layers to be produced by the encoder.
- range str
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step str
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch_
mode str | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync_
mode str | VideoSync Mode - The Video Sync Mode
- start String
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- layers List<Property Map>
- A collection of output PNG image layers to be produced by the encoder.
- range String
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step String
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String | "None" | "AutoSize" | "Auto Fit" - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String | "Auto" | "Passthrough" | "Cfr" | "Vfr" - The Video Sync Mode
PngImageResponse, PngImageResponseArgs
- Start string
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Layers
List<Pulumi.
Azure Native. Media. Inputs. Png Layer Response> - A collection of output PNG image layers to be produced by the encoder.
- Range string
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Step string
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string - The Video Sync Mode
- Start string
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Layers
[]Png
Layer Response - A collection of output PNG image layers to be produced by the encoder.
- Range string
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- Step string
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- Stretch
Mode string - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string - The Video Sync Mode
- start String
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- layers
List<Png
Layer Response> - A collection of output PNG image layers to be produced by the encoder.
- range String
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step String
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String - The Video Sync Mode
- start string
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label string
- An optional label for the codec. The label can be used to control muxing behavior.
- layers
Png
Layer Response[] - A collection of output PNG image layers to be produced by the encoder.
- range string
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step string
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode string - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode string - The Video Sync Mode
- start str
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key_
frame_ strinterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label str
- An optional label for the codec. The label can be used to control muxing behavior.
- layers
Sequence[Png
Layer Response] - A collection of output PNG image layers to be produced by the encoder.
- range str
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step str
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch_
mode str - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync_
mode str - The Video Sync Mode
- start String
- The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- layers List<Property Map>
- A collection of output PNG image layers to be produced by the encoder.
- range String
- The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
- step String
- The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
- stretch
Mode String - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String - The Video Sync Mode
PngLayer, PngLayerArgs
- Height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- Height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height String
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- width String
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height str
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label str
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- width str
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height String
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- width String
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
PngLayerResponse, PngLayerResponseArgs
- Height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- Height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- Label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- Width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height String
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- width String
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height string
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label string
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- width string
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height str
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label str
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- width str
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
- height String
- The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
- label String
- The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
- width String
- The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
Priority, PriorityArgs
- Low
- LowUsed for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
- Normal
- NormalUsed for TransformOutputs that can be generated at Normal priority.
- High
- HighUsed for TransformOutputs that should take precedence over others.
- Priority
Low - LowUsed for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
- Priority
Normal - NormalUsed for TransformOutputs that can be generated at Normal priority.
- Priority
High - HighUsed for TransformOutputs that should take precedence over others.
- Low
- LowUsed for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
- Normal
- NormalUsed for TransformOutputs that can be generated at Normal priority.
- High
- HighUsed for TransformOutputs that should take precedence over others.
- Low
- LowUsed for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
- Normal
- NormalUsed for TransformOutputs that can be generated at Normal priority.
- High
- HighUsed for TransformOutputs that should take precedence over others.
- LOW
- LowUsed for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
- NORMAL
- NormalUsed for TransformOutputs that can be generated at Normal priority.
- HIGH
- HighUsed for TransformOutputs that should take precedence over others.
- "Low"
- LowUsed for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
- "Normal"
- NormalUsed for TransformOutputs that can be generated at Normal priority.
- "High"
- HighUsed for TransformOutputs that should take precedence over others.
Rectangle, RectangleArgs
- Height string
- The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Left string
- The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Top string
- The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Width string
- The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Height string
- The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Left string
- The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Top string
- The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Width string
- The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- height String
- The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- left String
- The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- top String
- The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- width String
- The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- height string
- The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- left string
- The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- top string
- The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- width string
- The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- height str
- The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- left str
- The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- top str
- The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- width str
- The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- height String
- The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- left String
- The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- top String
- The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- width String
- The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
RectangleResponse, RectangleResponseArgs
- Height string
- The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Left string
- The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Top string
- The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Width string
- The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Height string
- The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Left string
- The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Top string
- The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- Width string
- The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- height String
- The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- left String
- The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- top String
- The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- width String
- The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- height string
- The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- left string
- The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- top string
- The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- width string
- The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- height str
- The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- left str
- The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- top str
- The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- width str
- The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- height String
- The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- left String
- The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- top String
- The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
- width String
- The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
Rotation, RotationArgs
- Auto
- AutoAutomatically detect and rotate as needed.
- None
- NoneDo not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
- Rotate0
- Rotate0Do not rotate the video but remove any metadata about the rotation.
- Rotate90
- Rotate90Rotate 90 degrees clockwise.
- Rotate180
- Rotate180Rotate 180 degrees clockwise.
- Rotate270
- Rotate270Rotate 270 degrees clockwise.
- Rotation
Auto - AutoAutomatically detect and rotate as needed.
- Rotation
None - NoneDo not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
- Rotation
Rotate0 - Rotate0Do not rotate the video but remove any metadata about the rotation.
- Rotation
Rotate90 - Rotate90Rotate 90 degrees clockwise.
- Rotation
Rotate180 - Rotate180Rotate 180 degrees clockwise.
- Rotation
Rotate270 - Rotate270Rotate 270 degrees clockwise.
- Auto
- AutoAutomatically detect and rotate as needed.
- None
- NoneDo not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
- Rotate0
- Rotate0Do not rotate the video but remove any metadata about the rotation.
- Rotate90
- Rotate90Rotate 90 degrees clockwise.
- Rotate180
- Rotate180Rotate 180 degrees clockwise.
- Rotate270
- Rotate270Rotate 270 degrees clockwise.
- Auto
- AutoAutomatically detect and rotate as needed.
- None
- NoneDo not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
- Rotate0
- Rotate0Do not rotate the video but remove any metadata about the rotation.
- Rotate90
- Rotate90Rotate 90 degrees clockwise.
- Rotate180
- Rotate180Rotate 180 degrees clockwise.
- Rotate270
- Rotate270Rotate 270 degrees clockwise.
- AUTO
- AutoAutomatically detect and rotate as needed.
- NONE
- NoneDo not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
- ROTATE0
- Rotate0Do not rotate the video but remove any metadata about the rotation.
- ROTATE90
- Rotate90Rotate 90 degrees clockwise.
- ROTATE180
- Rotate180Rotate 180 degrees clockwise.
- ROTATE270
- Rotate270Rotate 270 degrees clockwise.
- "Auto"
- AutoAutomatically detect and rotate as needed.
- "None"
- NoneDo not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
- "Rotate0"
- Rotate0Do not rotate the video but remove any metadata about the rotation.
- "Rotate90"
- Rotate90Rotate 90 degrees clockwise.
- "Rotate180"
- Rotate180Rotate 180 degrees clockwise.
- "Rotate270"
- Rotate270Rotate 270 degrees clockwise.
StandardEncoderPreset, StandardEncoderPresetArgs
- Codecs List<object>
- The list of codecs to be used when encoding the input video.
- Formats List<object>
- The list of outputs to be produced by the encoder.
- Filters
Pulumi.
Azure Native. Media. Inputs. Filters - One or more filtering operations that are applied to the input media before encoding.
- codecs
(Aac
Audio | Audio | Copy Audio | Copy Video | H264Video | H265Video | Image | Jpg Image | Png Image | Video)[] - The list of codecs to be used when encoding the input video.
- formats
(Image
Format | Jpg Format | Mp4Format | Multi Bitrate Format | Png Format | Transport Stream Format)[] - The list of outputs to be produced by the encoder.
- filters Filters
- One or more filtering operations that are applied to the input media before encoding.
- codecs
Sequence[Union[Aac
Audio, Audio, Copy Audio, Copy Video, H264Video, H265Video, Image, Jpg Image, Png Image, Video]] - The list of codecs to be used when encoding the input video.
- formats
Sequence[Union[Image
Format, Jpg Format, Mp4Format, Multi Bitrate Format, Png Format, Transport Stream Format]] - The list of outputs to be produced by the encoder.
- filters Filters
- One or more filtering operations that are applied to the input media before encoding.
- codecs List<Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map>
- The list of codecs to be used when encoding the input video.
- formats List<Property Map | Property Map | Property Map | Property Map | Property Map | Property Map>
- The list of outputs to be produced by the encoder.
- filters Property Map
- One or more filtering operations that are applied to the input media before encoding.
StandardEncoderPresetResponse, StandardEncoderPresetResponseArgs
- Codecs List<object>
- The list of codecs to be used when encoding the input video.
- Formats List<object>
- The list of outputs to be produced by the encoder.
- Filters
Pulumi.
Azure Native. Media. Inputs. Filters Response - One or more filtering operations that are applied to the input media before encoding.
- Codecs []interface{}
- The list of codecs to be used when encoding the input video.
- Formats []interface{}
- The list of outputs to be produced by the encoder.
- Filters
Filters
Response - One or more filtering operations that are applied to the input media before encoding.
- codecs List<Object>
- The list of codecs to be used when encoding the input video.
- formats List<Object>
- The list of outputs to be produced by the encoder.
- filters
Filters
Response - One or more filtering operations that are applied to the input media before encoding.
- codecs
(Aac
Audio Response | Audio Response | Copy Audio Response | Copy Video Response | H264Video Response | H265Video Response | Image Response | Jpg Image Response | Png Image Response | Video Response)[] - The list of codecs to be used when encoding the input video.
- formats
(Image
Format Response | Jpg Format Response | Mp4Format Response | Multi Bitrate Format Response | Png Format Response | Transport Stream Format Response)[] - The list of outputs to be produced by the encoder.
- filters
Filters
Response - One or more filtering operations that are applied to the input media before encoding.
- codecs
Sequence[Union[Aac
Audio Response, Audio Response, Copy Audio Response, Copy Video Response, H264Video Response, H265Video Response, Image Response, Jpg Image Response, Png Image Response, Video Response]] - The list of codecs to be used when encoding the input video.
- formats
Sequence[Union[Image
Format Response, Jpg Format Response, Mp4Format Response, Multi Bitrate Format Response, Png Format Response, Transport Stream Format Response]] - The list of outputs to be produced by the encoder.
- filters
Filters
Response - One or more filtering operations that are applied to the input media before encoding.
- codecs List<Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map>
- The list of codecs to be used when encoding the input video.
- formats List<Property Map | Property Map | Property Map | Property Map | Property Map | Property Map>
- The list of outputs to be produced by the encoder.
- filters Property Map
- One or more filtering operations that are applied to the input media before encoding.
StretchMode, StretchModeArgs
- None
- NoneStrictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
- Auto
Size - AutoSizeOverride the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
- Auto
Fit - AutoFitPad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.
- Stretch
Mode None - NoneStrictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
- Stretch
Mode Auto Size - AutoSizeOverride the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
- Stretch
Mode Auto Fit - AutoFitPad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.
- None
- NoneStrictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
- Auto
Size - AutoSizeOverride the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
- Auto
Fit - AutoFitPad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.
- None
- NoneStrictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
- Auto
Size - AutoSizeOverride the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
- Auto
Fit - AutoFitPad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.
- NONE
- NoneStrictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
- AUTO_SIZE
- AutoSizeOverride the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
- AUTO_FIT
- AutoFitPad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.
- "None"
- NoneStrictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
- "Auto
Size" - AutoSizeOverride the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
- "Auto
Fit" - AutoFitPad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.
SystemDataResponse, SystemDataResponseArgs
- Created
At string - The timestamp of resource creation (UTC).
- Created
By string - The identity that created the resource.
- Created
By stringType - The type of identity that created the resource.
- Last
Modified stringAt - The timestamp of resource last modification (UTC)
- Last
Modified stringBy - The identity that last modified the resource.
- Last
Modified stringBy Type - The type of identity that last modified the resource.
- Created
At string - The timestamp of resource creation (UTC).
- Created
By string - The identity that created the resource.
- Created
By stringType - The type of identity that created the resource.
- Last
Modified stringAt - The timestamp of resource last modification (UTC)
- Last
Modified stringBy - The identity that last modified the resource.
- Last
Modified stringBy Type - The type of identity that last modified the resource.
- created
At String - The timestamp of resource creation (UTC).
- created
By String - The identity that created the resource.
- created
By StringType - The type of identity that created the resource.
- last
Modified StringAt - The timestamp of resource last modification (UTC)
- last
Modified StringBy - The identity that last modified the resource.
- last
Modified StringBy Type - The type of identity that last modified the resource.
- created
At string - The timestamp of resource creation (UTC).
- created
By string - The identity that created the resource.
- created
By stringType - The type of identity that created the resource.
- last
Modified stringAt - The timestamp of resource last modification (UTC)
- last
Modified stringBy - The identity that last modified the resource.
- last
Modified stringBy Type - The type of identity that last modified the resource.
- created_
at str - The timestamp of resource creation (UTC).
- created_
by str - The identity that created the resource.
- created_
by_ strtype - The type of identity that created the resource.
- last_
modified_ strat - The timestamp of resource last modification (UTC)
- last_
modified_ strby - The identity that last modified the resource.
- last_
modified_ strby_ type - The type of identity that last modified the resource.
- created
At String - The timestamp of resource creation (UTC).
- created
By String - The identity that created the resource.
- created
By StringType - The type of identity that created the resource.
- last
Modified StringAt - The timestamp of resource last modification (UTC)
- last
Modified StringBy - The identity that last modified the resource.
- last
Modified StringBy Type - The type of identity that last modified the resource.
TransformOutput, TransformOutputArgs
- Preset
Pulumi.
Azure | Pulumi.Native. Media. Inputs. Audio Analyzer Preset Azure | Pulumi.Native. Media. Inputs. Built In Standard Encoder Preset Azure | Pulumi.Native. Media. Inputs. Face Detector Preset Azure | Pulumi.Native. Media. Inputs. Standard Encoder Preset Azure Native. Media. Inputs. Video Analyzer Preset - Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- On
Error string | Pulumi.Azure Native. Media. On Error Type - A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- Relative
Priority string | Pulumi.Azure Native. Media. Priority - Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
- Preset
Audio
Analyzer | BuiltPreset In | FaceStandard Encoder Preset Detector | StandardPreset Encoder | VideoPreset Analyzer Preset - Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- On
Error string | OnError Type - A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- Relative
Priority string | Priority - Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
- preset
Audio
Analyzer | BuiltPreset In | FaceStandard Encoder Preset Detector | StandardPreset Encoder | VideoPreset Analyzer Preset - Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- on
Error String | OnError Type - A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- relative
Priority String | Priority - Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
- preset
Audio
Analyzer | BuiltPreset In | FaceStandard Encoder Preset Detector | StandardPreset Encoder | VideoPreset Analyzer Preset - Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- on
Error string | OnError Type - A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- relative
Priority string | Priority - Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
- preset
Audio
Analyzer | BuiltPreset In | FaceStandard Encoder Preset Detector | StandardPreset Encoder | VideoPreset Analyzer Preset - Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- on_
error str | OnError Type - A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- relative_
priority str | Priority - Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
- preset Property Map | Property Map | Property Map | Property Map | Property Map
- Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- on
Error String | "StopProcessing Job" | "Continue Job" - A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- relative
Priority String | "Low" | "Normal" | "High" - Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
TransformOutputResponse, TransformOutputResponseArgs
- Preset
Pulumi.
Azure | Pulumi.Native. Media. Inputs. Audio Analyzer Preset Response Azure | Pulumi.Native. Media. Inputs. Built In Standard Encoder Preset Response Azure | Pulumi.Native. Media. Inputs. Face Detector Preset Response Azure | Pulumi.Native. Media. Inputs. Standard Encoder Preset Response Azure Native. Media. Inputs. Video Analyzer Preset Response - Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- On
Error string - A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- Relative
Priority string - Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
- Preset
Audio
Analyzer | BuiltPreset Response In | FaceStandard Encoder Preset Response Detector | StandardPreset Response Encoder | VideoPreset Response Analyzer Preset Response - Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- On
Error string - A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- Relative
Priority string - Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
- preset
Audio
Analyzer | BuiltPreset Response In | FaceStandard Encoder Preset Response Detector | StandardPreset Response Encoder | VideoPreset Response Analyzer Preset Response - Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- on
Error String - A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- relative
Priority String - Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
- preset
Audio
Analyzer | BuiltPreset Response In | FaceStandard Encoder Preset Response Detector | StandardPreset Response Encoder | VideoPreset Response Analyzer Preset Response - Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- on
Error string - A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- relative
Priority string - Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
- preset
Audio
Analyzer | BuiltPreset Response In | FaceStandard Encoder Preset Response Detector | StandardPreset Response Encoder | VideoPreset Response Analyzer Preset Response - Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- on_
error str - A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- relative_
priority str - Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
- preset Property Map | Property Map | Property Map | Property Map | Property Map
- Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
- on
Error String - A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
- relative
Priority String - Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
TransportStreamFormat, TransportStreamFormatArgs
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Output
Files List<Pulumi.Azure Native. Media. Inputs. Output File> - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Output
Files []OutputFile - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files List<OutputFile> - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files OutputFile[] - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename_
pattern str - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output_
files Sequence[OutputFile] - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files List<Property Map> - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
TransportStreamFormatResponse, TransportStreamFormatResponseArgs
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Output
Files List<Pulumi.Azure Native. Media. Inputs. Output File Response> - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- Filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- Output
Files []OutputFile Response - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files List<OutputFile Response> - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern string - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files OutputFile Response[] - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename_
pattern str - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output_
files Sequence[OutputFile Response] - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
- filename
Pattern String - The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
- output
Files List<Property Map> - The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
Video, VideoArgs
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Stretch
Mode string | Pulumi.Azure Native. Media. Stretch Mode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string | Pulumi.Azure Native. Media. Video Sync Mode - The Video Sync Mode
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Stretch
Mode string | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string | VideoSync Mode - The Video Sync Mode
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- stretch
Mode String | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String | VideoSync Mode - The Video Sync Mode
- key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label string
- An optional label for the codec. The label can be used to control muxing behavior.
- stretch
Mode string | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode string | VideoSync Mode - The Video Sync Mode
- key_
frame_ strinterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label str
- An optional label for the codec. The label can be used to control muxing behavior.
- stretch_
mode str | StretchMode - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync_
mode str | VideoSync Mode - The Video Sync Mode
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- stretch
Mode String | "None" | "AutoSize" | "Auto Fit" - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String | "Auto" | "Passthrough" | "Cfr" | "Vfr" - The Video Sync Mode
VideoAnalyzerPreset, VideoAnalyzerPresetArgs
- Audio
Language string - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- Experimental
Options Dictionary<string, string> - Dictionary containing key value pairs for parameters not exposed in the preset itself
- Insights
To string | Pulumi.Extract Azure Native. Media. Insights Type - Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
- Mode
string | Pulumi.
Azure Native. Media. Audio Analysis Mode - Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- Audio
Language string - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- Experimental
Options map[string]string - Dictionary containing key value pairs for parameters not exposed in the preset itself
- Insights
To string | InsightsExtract Type - Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
- Mode
string | Audio
Analysis Mode - Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio
Language String - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental
Options Map<String,String> - Dictionary containing key value pairs for parameters not exposed in the preset itself
- insights
To String | InsightsExtract Type - Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
- mode
String | Audio
Analysis Mode - Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio
Language string - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental
Options {[key: string]: string} - Dictionary containing key value pairs for parameters not exposed in the preset itself
- insights
To string | InsightsExtract Type - Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
- mode
string | Audio
Analysis Mode - Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio_
language str - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental_
options Mapping[str, str] - Dictionary containing key value pairs for parameters not exposed in the preset itself
- insights_
to_ str | Insightsextract Type - Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
- mode
str | Audio
Analysis Mode - Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio
Language String - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental
Options Map<String> - Dictionary containing key value pairs for parameters not exposed in the preset itself
- insights
To String | "AudioExtract Insights Only" | "Video Insights Only" | "All Insights" - Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
- mode String | "Standard" | "Basic"
- Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
VideoAnalyzerPresetResponse, VideoAnalyzerPresetResponseArgs
- Audio
Language string - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- Experimental
Options Dictionary<string, string> - Dictionary containing key value pairs for parameters not exposed in the preset itself
- Insights
To stringExtract - Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
- Mode string
- Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- Audio
Language string - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- Experimental
Options map[string]string - Dictionary containing key value pairs for parameters not exposed in the preset itself
- Insights
To stringExtract - Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
- Mode string
- Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio
Language String - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental
Options Map<String,String> - Dictionary containing key value pairs for parameters not exposed in the preset itself
- insights
To StringExtract - Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
- mode String
- Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio
Language string - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental
Options {[key: string]: string} - Dictionary containing key value pairs for parameters not exposed in the preset itself
- insights
To stringExtract - Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
- mode string
- Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio_
language str - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental_
options Mapping[str, str] - Dictionary containing key value pairs for parameters not exposed in the preset itself
- insights_
to_ strextract - Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
- mode str
- Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
- audio
Language String - The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
- experimental
Options Map<String> - Dictionary containing key value pairs for parameters not exposed in the preset itself
- insights
To StringExtract - Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
- mode String
- Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
VideoOverlay, VideoOverlayArgs
- Input
Label string - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- Audio
Gain doubleLevel - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- Crop
Rectangle Pulumi.Azure Native. Media. Inputs. Rectangle - An optional rectangular window used to crop the overlay image or video.
- End string
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- Fade
In stringDuration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- Fade
Out stringDuration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- Opacity double
- The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
- Position
Pulumi.
Azure Native. Media. Inputs. Rectangle - The location in the input video where the overlay is applied.
- Start string
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- Input
Label string - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- Audio
Gain float64Level - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- Crop
Rectangle Rectangle - An optional rectangular window used to crop the overlay image or video.
- End string
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- Fade
In stringDuration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- Fade
Out stringDuration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- Opacity float64
- The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
- Position Rectangle
- The location in the input video where the overlay is applied.
- Start string
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input
Label String - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio
Gain DoubleLevel - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- crop
Rectangle Rectangle - An optional rectangular window used to crop the overlay image or video.
- end String
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade
In StringDuration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade
Out StringDuration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- opacity Double
- The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
- position Rectangle
- The location in the input video where the overlay is applied.
- start String
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input
Label string - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio
Gain numberLevel - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- crop
Rectangle Rectangle - An optional rectangular window used to crop the overlay image or video.
- end string
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade
In stringDuration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade
Out stringDuration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- opacity number
- The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
- position Rectangle
- The location in the input video where the overlay is applied.
- start string
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input_
label str - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio_
gain_ floatlevel - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- crop_
rectangle Rectangle - An optional rectangular window used to crop the overlay image or video.
- end str
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade_
in_ strduration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade_
out_ strduration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- opacity float
- The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
- position Rectangle
- The location in the input video where the overlay is applied.
- start str
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input
Label String - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio
Gain NumberLevel - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- crop
Rectangle Property Map - An optional rectangular window used to crop the overlay image or video.
- end String
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade
In StringDuration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade
Out StringDuration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- opacity Number
- The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
- position Property Map
- The location in the input video where the overlay is applied.
- start String
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
VideoOverlayResponse, VideoOverlayResponseArgs
- Input
Label string - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- Audio
Gain doubleLevel - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- Crop
Rectangle Pulumi.Azure Native. Media. Inputs. Rectangle Response - An optional rectangular window used to crop the overlay image or video.
- End string
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- Fade
In stringDuration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- Fade
Out stringDuration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- Opacity double
- The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
- Position
Pulumi.
Azure Native. Media. Inputs. Rectangle Response - The location in the input video where the overlay is applied.
- Start string
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- Input
Label string - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- Audio
Gain float64Level - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- Crop
Rectangle RectangleResponse - An optional rectangular window used to crop the overlay image or video.
- End string
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- Fade
In stringDuration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- Fade
Out stringDuration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- Opacity float64
- The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
- Position
Rectangle
Response - The location in the input video where the overlay is applied.
- Start string
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input
Label String - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio
Gain DoubleLevel - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- crop
Rectangle RectangleResponse - An optional rectangular window used to crop the overlay image or video.
- end String
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade
In StringDuration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade
Out StringDuration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- opacity Double
- The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
- position
Rectangle
Response - The location in the input video where the overlay is applied.
- start String
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input
Label string - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio
Gain numberLevel - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- crop
Rectangle RectangleResponse - An optional rectangular window used to crop the overlay image or video.
- end string
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade
In stringDuration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade
Out stringDuration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- opacity number
- The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
- position
Rectangle
Response - The location in the input video where the overlay is applied.
- start string
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input_
label str - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio_
gain_ floatlevel - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- crop_
rectangle RectangleResponse - An optional rectangular window used to crop the overlay image or video.
- end str
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade_
in_ strduration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade_
out_ strduration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- opacity float
- The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
- position
Rectangle
Response - The location in the input video where the overlay is applied.
- start str
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
- input
Label String - The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
- audio
Gain NumberLevel - The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
- crop
Rectangle Property Map - An optional rectangular window used to crop the overlay image or video.
- end String
- The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
- fade
In StringDuration - The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
- fade
Out StringDuration - The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
- opacity Number
- The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
- position Property Map
- The location in the input video where the overlay is applied.
- start String
- The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
VideoResponse, VideoResponseArgs
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Stretch
Mode string - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string - The Video Sync Mode
- Key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- Label string
- An optional label for the codec. The label can be used to control muxing behavior.
- Stretch
Mode string - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- Sync
Mode string - The Video Sync Mode
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- stretch
Mode String - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String - The Video Sync Mode
- key
Frame stringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label string
- An optional label for the codec. The label can be used to control muxing behavior.
- stretch
Mode string - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode string - The Video Sync Mode
- key_
frame_ strinterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label str
- An optional label for the codec. The label can be used to control muxing behavior.
- stretch_
mode str - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync_
mode str - The Video Sync Mode
- key
Frame StringInterval - The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
- label String
- An optional label for the codec. The label can be used to control muxing behavior.
- stretch
Mode String - The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
- sync
Mode String - The Video Sync Mode
VideoSyncMode, VideoSyncModeArgs
- Auto
- AutoThis is the default method. Chooses between Cfr and Vfr depending on muxer capabilities. For output format MP4, the default mode is Cfr.
- Passthrough
- PassthroughThe presentation timestamps on frames are passed through from the input file to the output file writer. Recommended when the input source has variable frame rate, and are attempting to produce multiple layers for adaptive streaming in the output which have aligned GOP boundaries. Note: if two or more frames in the input have duplicate timestamps, then the output will also have the same behavior
- Cfr
- CfrInput frames will be repeated and/or dropped as needed to achieve exactly the requested constant frame rate. Recommended when the output frame rate is explicitly set at a specified value
- Vfr
- VfrSimilar to the Passthrough mode, but if the input has frames that have duplicate timestamps, then only one frame is passed through to the output, and others are dropped. Recommended when the number of output frames is expected to be equal to the number of input frames. For example, the output is used to calculate a quality metric like PSNR against the input
- Video
Sync Mode Auto - AutoThis is the default method. Chooses between Cfr and Vfr depending on muxer capabilities. For output format MP4, the default mode is Cfr.
- Video
Sync Mode Passthrough - PassthroughThe presentation timestamps on frames are passed through from the input file to the output file writer. Recommended when the input source has variable frame rate, and are attempting to produce multiple layers for adaptive streaming in the output which have aligned GOP boundaries. Note: if two or more frames in the input have duplicate timestamps, then the output will also have the same behavior
- Video
Sync Mode Cfr - CfrInput frames will be repeated and/or dropped as needed to achieve exactly the requested constant frame rate. Recommended when the output frame rate is explicitly set at a specified value
- Video
Sync Mode Vfr - VfrSimilar to the Passthrough mode, but if the input has frames that have duplicate timestamps, then only one frame is passed through to the output, and others are dropped. Recommended when the number of output frames is expected to be equal to the number of input frames. For example, the output is used to calculate a quality metric like PSNR against the input
- Auto
- AutoThis is the default method. Chooses between Cfr and Vfr depending on muxer capabilities. For output format MP4, the default mode is Cfr.
- Passthrough
- PassthroughThe presentation timestamps on frames are passed through from the input file to the output file writer. Recommended when the input source has variable frame rate, and are attempting to produce multiple layers for adaptive streaming in the output which have aligned GOP boundaries. Note: if two or more frames in the input have duplicate timestamps, then the output will also have the same behavior
- Cfr
- CfrInput frames will be repeated and/or dropped as needed to achieve exactly the requested constant frame rate. Recommended when the output frame rate is explicitly set at a specified value
- Vfr
- VfrSimilar to the Passthrough mode, but if the input has frames that have duplicate timestamps, then only one frame is passed through to the output, and others are dropped. Recommended when the number of output frames is expected to be equal to the number of input frames. For example, the output is used to calculate a quality metric like PSNR against the input
- Auto
- AutoThis is the default method. Chooses between Cfr and Vfr depending on muxer capabilities. For output format MP4, the default mode is Cfr.
- Passthrough
- PassthroughThe presentation timestamps on frames are passed through from the input file to the output file writer. Recommended when the input source has variable frame rate, and are attempting to produce multiple layers for adaptive streaming in the output which have aligned GOP boundaries. Note: if two or more frames in the input have duplicate timestamps, then the output will also have the same behavior
- Cfr
- CfrInput frames will be repeated and/or dropped as needed to achieve exactly the requested constant frame rate. Recommended when the output frame rate is explicitly set at a specified value
- Vfr
- VfrSimilar to the Passthrough mode, but if the input has frames that have duplicate timestamps, then only one frame is passed through to the output, and others are dropped. Recommended when the number of output frames is expected to be equal to the number of input frames. For example, the output is used to calculate a quality metric like PSNR against the input
- AUTO
- AutoThis is the default method. Chooses between Cfr and Vfr depending on muxer capabilities. For output format MP4, the default mode is Cfr.
- PASSTHROUGH
- PassthroughThe presentation timestamps on frames are passed through from the input file to the output file writer. Recommended when the input source has variable frame rate, and are attempting to produce multiple layers for adaptive streaming in the output which have aligned GOP boundaries. Note: if two or more frames in the input have duplicate timestamps, then the output will also have the same behavior
- CFR
- CfrInput frames will be repeated and/or dropped as needed to achieve exactly the requested constant frame rate. Recommended when the output frame rate is explicitly set at a specified value
- VFR
- VfrSimilar to the Passthrough mode, but if the input has frames that have duplicate timestamps, then only one frame is passed through to the output, and others are dropped. Recommended when the number of output frames is expected to be equal to the number of input frames. For example, the output is used to calculate a quality metric like PSNR against the input
- "Auto"
- AutoThis is the default method. Chooses between Cfr and Vfr depending on muxer capabilities. For output format MP4, the default mode is Cfr.
- "Passthrough"
- PassthroughThe presentation timestamps on frames are passed through from the input file to the output file writer. Recommended when the input source has variable frame rate, and are attempting to produce multiple layers for adaptive streaming in the output which have aligned GOP boundaries. Note: if two or more frames in the input have duplicate timestamps, then the output will also have the same behavior
- "Cfr"
- CfrInput frames will be repeated and/or dropped as needed to achieve exactly the requested constant frame rate. Recommended when the output frame rate is explicitly set at a specified value
- "Vfr"
- VfrSimilar to the Passthrough mode, but if the input has frames that have duplicate timestamps, then only one frame is passed through to the output, and others are dropped. Recommended when the number of output frames is expected to be equal to the number of input frames. For example, the output is used to calculate a quality metric like PSNR against the input
Import
An existing resource can be imported using its type token, name, and identifier, e.g.
$ pulumi import azure-native:media:Transform createdTransform /subscriptions/00000000-0000-0000-0000-000000000000/resourceGroups/contosoresources/providers/Microsoft.Media/mediaservices/contosomedia/transforms/createdTransform
To learn more about importing existing cloud resources, see Importing resources.
Package Details
- Repository
- azure-native-v1 pulumi/pulumi-azure-native
- License
- Apache-2.0