Class H265Settings
- All Implemented Interfaces:
Serializable
,SdkPojo
,ToCopyableBuilder<H265Settings.Builder,
H265Settings>
- See Also:
-
Nested Class Summary
-
Method Summary
Modifier and TypeMethodDescriptionfinal H265AdaptiveQuantization
When you set Adaptive Quantization to Auto, or leave blank, MediaConvert automatically applies quantization to improve the video quality of your output.final String
When you set Adaptive Quantization to Auto, or leave blank, MediaConvert automatically applies quantization to improve the video quality of your output.Enables Alternate Transfer Function SEI message for outputs using Hybrid Log Gamma (HLG) Electro-Optical Transfer Function (EOTF).final String
Enables Alternate Transfer Function SEI message for outputs using Hybrid Log Gamma (HLG) Electro-Optical Transfer Function (EOTF).final BandwidthReductionFilter
The Bandwidth reduction filter increases the video quality of your output relative to its bitrate.final Integer
bitrate()
Specify the average bitrate in bits per second.static H265Settings.Builder
builder()
final H265CodecLevel
H.265 Level.final String
H.265 Level.final H265CodecProfile
Represents the Profile and Tier, per the HEVC (H.265) specification.final String
Represents the Profile and Tier, per the HEVC (H.265) specification.final H265DynamicSubGop
Specify whether to allow the number of B-frames in your output GOP structure to vary or not depending on your input video content.final String
Specify whether to allow the number of B-frames in your output GOP structure to vary or not depending on your input video content.final H265EndOfStreamMarkers
Optionally include or suppress markers at the end of your output that signal the end of the video stream.final String
Optionally include or suppress markers at the end of your output that signal the end of the video stream.final boolean
final boolean
equalsBySdkFields
(Object obj) Indicates whether some other object is "equal to" this one by SDK fields.Enable this setting to have the encoder reduce I-frame pop.final String
Enable this setting to have the encoder reduce I-frame pop.final H265FramerateControl
Use the Framerate setting to specify the frame rate for this output.final String
Use the Framerate setting to specify the frame rate for this output.Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate.final String
Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate.final Integer
When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction.final Integer
When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction.final <T> Optional
<T> getValueForField
(String fieldName, Class<T> clazz) final H265GopBReference
Specify whether to allow B-frames to be referenced by other frame types.final String
Specify whether to allow B-frames to be referenced by other frame types.final Integer
Specify the relative frequency of open to closed GOPs in this output.final Double
gopSize()
Use this setting only when you set GOP mode control to Specified, frames or Specified, seconds.final H265GopSizeUnits
Specify how the transcoder determines GOP size for this output.final String
Specify how the transcoder determines GOP size for this output.final int
hashCode()
final Integer
If your downstream systems have strict buffer requirements: Specify the minimum percentage of the HRD buffer that's available at the end of each encoded video segment.final Integer
Percentage of the buffer that should initially be filled (HRD buffer model).final Integer
Size of buffer (HRD buffer model) in bits.final H265InterlaceMode
Choose the scan line type for the output.final String
Choose the scan line type for the output.final Integer
Maximum bitrate in bits/second.final Integer
Use this setting only when you also enable Scene change detection.final Integer
Specify the number of B-frames between reference frames in this output.final Integer
Number of reference frames to use.final H265ParControl
Optional.final String
Optional.final Integer
Required when you set Pixel aspect ratio to SPECIFIED.final Integer
Required when you set Pixel aspect ratio to SPECIFIED.final H265QualityTuningLevel
Optional.final String
Optional.final H265QvbrSettings
Settings for quality-defined variable bitrate encoding with the H.265 codec.final H265RateControlMode
Use this setting to specify whether this output has a variable bitrate (VBR), constant bitrate (CBR) or quality-defined variable bitrate (QVBR).final String
Use this setting to specify whether this output has a variable bitrate (VBR), constant bitrate (CBR) or quality-defined variable bitrate (QVBR).Specify Sample Adaptive Offset (SAO) filter strength.final String
Specify Sample Adaptive Offset (SAO) filter strength.Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate.final String
Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate.final H265SceneChangeDetect
Enable this setting to insert I-frames at scene changes that the service automatically detects.final String
Enable this setting to insert I-frames at scene changes that the service automatically detects.static Class
<? extends H265Settings.Builder> final Integer
slices()
Number of slices per picture.final H265SlowPal
slowPal()
Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps).final String
Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps).Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity.final String
Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity.final H265Telecine
telecine()
This field applies only if the Streams > Advanced > Framerate field is set to 29.970.final String
This field applies only if the Streams > Advanced > Framerate field is set to 29.970.Keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity.final String
Keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity.final H265TemporalIds
Enables temporal layer identifiers in the encoded bitstream.final String
Enables temporal layer identifiers in the encoded bitstream.final H265Tiles
tiles()
Enable use of tiles, allowing horizontal as well as vertical subdivision of the encoded pictures.final String
Enable use of tiles, allowing horizontal as well as vertical subdivision of the encoded pictures.Take this object and create a builder that contains all of the current property values of this object.final String
toString()
Returns a string representation of this object.Inserts timecode for each frame as 4 bytes of an unregistered SEI message.final String
Inserts timecode for each frame as 4 bytes of an unregistered SEI message.If the location of parameter set NAL units doesn't matter in your workflow, ignore this setting.final String
If the location of parameter set NAL units doesn't matter in your workflow, ignore this setting.Methods inherited from interface software.amazon.awssdk.utils.builder.ToCopyableBuilder
copy
-
Method Details
-
adaptiveQuantization
When you set Adaptive Quantization to Auto, or leave blank, MediaConvert automatically applies quantization to improve the video quality of your output. Set Adaptive Quantization to Low, Medium, High, Higher, or Max to manually control the strength of the quantization filter. When you do, you can specify a value for Spatial Adaptive Quantization, Temporal Adaptive Quantization, and Flicker Adaptive Quantization, to further control the quantization filter. Set Adaptive Quantization to Off to apply no quantization to your output.If the service returns an enum value that is not available in the current SDK version,
adaptiveQuantization
will returnH265AdaptiveQuantization.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromadaptiveQuantizationAsString()
.- Returns:
- When you set Adaptive Quantization to Auto, or leave blank, MediaConvert automatically applies quantization to improve the video quality of your output. Set Adaptive Quantization to Low, Medium, High, Higher, or Max to manually control the strength of the quantization filter. When you do, you can specify a value for Spatial Adaptive Quantization, Temporal Adaptive Quantization, and Flicker Adaptive Quantization, to further control the quantization filter. Set Adaptive Quantization to Off to apply no quantization to your output.
- See Also:
-
adaptiveQuantizationAsString
When you set Adaptive Quantization to Auto, or leave blank, MediaConvert automatically applies quantization to improve the video quality of your output. Set Adaptive Quantization to Low, Medium, High, Higher, or Max to manually control the strength of the quantization filter. When you do, you can specify a value for Spatial Adaptive Quantization, Temporal Adaptive Quantization, and Flicker Adaptive Quantization, to further control the quantization filter. Set Adaptive Quantization to Off to apply no quantization to your output.If the service returns an enum value that is not available in the current SDK version,
adaptiveQuantization
will returnH265AdaptiveQuantization.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromadaptiveQuantizationAsString()
.- Returns:
- When you set Adaptive Quantization to Auto, or leave blank, MediaConvert automatically applies quantization to improve the video quality of your output. Set Adaptive Quantization to Low, Medium, High, Higher, or Max to manually control the strength of the quantization filter. When you do, you can specify a value for Spatial Adaptive Quantization, Temporal Adaptive Quantization, and Flicker Adaptive Quantization, to further control the quantization filter. Set Adaptive Quantization to Off to apply no quantization to your output.
- See Also:
-
alternateTransferFunctionSei
Enables Alternate Transfer Function SEI message for outputs using Hybrid Log Gamma (HLG) Electro-Optical Transfer Function (EOTF).If the service returns an enum value that is not available in the current SDK version,
alternateTransferFunctionSei
will returnH265AlternateTransferFunctionSei.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromalternateTransferFunctionSeiAsString()
.- Returns:
- Enables Alternate Transfer Function SEI message for outputs using Hybrid Log Gamma (HLG) Electro-Optical Transfer Function (EOTF).
- See Also:
-
alternateTransferFunctionSeiAsString
Enables Alternate Transfer Function SEI message for outputs using Hybrid Log Gamma (HLG) Electro-Optical Transfer Function (EOTF).If the service returns an enum value that is not available in the current SDK version,
alternateTransferFunctionSei
will returnH265AlternateTransferFunctionSei.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromalternateTransferFunctionSeiAsString()
.- Returns:
- Enables Alternate Transfer Function SEI message for outputs using Hybrid Log Gamma (HLG) Electro-Optical Transfer Function (EOTF).
- See Also:
-
bandwidthReductionFilter
The Bandwidth reduction filter increases the video quality of your output relative to its bitrate. Use to lower the bitrate of your constant quality QVBR output, with little or no perceptual decrease in quality. Or, use to increase the video quality of outputs with other rate control modes relative to the bitrate that you specify. Bandwidth reduction increases further when your input is low quality or noisy. Outputs that use this feature incur pro-tier pricing. When you include Bandwidth reduction filter, you cannot include the Noise reducer preprocessor.- Returns:
- The Bandwidth reduction filter increases the video quality of your output relative to its bitrate. Use to lower the bitrate of your constant quality QVBR output, with little or no perceptual decrease in quality. Or, use to increase the video quality of outputs with other rate control modes relative to the bitrate that you specify. Bandwidth reduction increases further when your input is low quality or noisy. Outputs that use this feature incur pro-tier pricing. When you include Bandwidth reduction filter, you cannot include the Noise reducer preprocessor.
-
bitrate
Specify the average bitrate in bits per second. Required for VBR and CBR. For MS Smooth outputs, bitrates must be unique when rounded down to the nearest multiple of 1000.- Returns:
- Specify the average bitrate in bits per second. Required for VBR and CBR. For MS Smooth outputs, bitrates must be unique when rounded down to the nearest multiple of 1000.
-
codecLevel
H.265 Level.If the service returns an enum value that is not available in the current SDK version,
codecLevel
will returnH265CodecLevel.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromcodecLevelAsString()
.- Returns:
- H.265 Level.
- See Also:
-
codecLevelAsString
H.265 Level.If the service returns an enum value that is not available in the current SDK version,
codecLevel
will returnH265CodecLevel.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromcodecLevelAsString()
.- Returns:
- H.265 Level.
- See Also:
-
codecProfile
Represents the Profile and Tier, per the HEVC (H.265) specification. Selections are grouped as [Profile] / [Tier], so "Main/High" represents Main Profile with High Tier. 4:2:2 profiles are only available with the HEVC 4:2:2 License.If the service returns an enum value that is not available in the current SDK version,
codecProfile
will returnH265CodecProfile.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromcodecProfileAsString()
.- Returns:
- Represents the Profile and Tier, per the HEVC (H.265) specification. Selections are grouped as [Profile] / [Tier], so "Main/High" represents Main Profile with High Tier. 4:2:2 profiles are only available with the HEVC 4:2:2 License.
- See Also:
-
codecProfileAsString
Represents the Profile and Tier, per the HEVC (H.265) specification. Selections are grouped as [Profile] / [Tier], so "Main/High" represents Main Profile with High Tier. 4:2:2 profiles are only available with the HEVC 4:2:2 License.If the service returns an enum value that is not available in the current SDK version,
codecProfile
will returnH265CodecProfile.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromcodecProfileAsString()
.- Returns:
- Represents the Profile and Tier, per the HEVC (H.265) specification. Selections are grouped as [Profile] / [Tier], so "Main/High" represents Main Profile with High Tier. 4:2:2 profiles are only available with the HEVC 4:2:2 License.
- See Also:
-
dynamicSubGop
Specify whether to allow the number of B-frames in your output GOP structure to vary or not depending on your input video content. To improve the subjective video quality of your output that has high-motion content: Leave blank or keep the default value Adaptive. MediaConvert will use fewer B-frames for high-motion video content than low-motion content. The maximum number of B- frames is limited by the value that you choose for B-frames between reference frames. To use the same number B-frames for all types of content: Choose Static.If the service returns an enum value that is not available in the current SDK version,
dynamicSubGop
will returnH265DynamicSubGop.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromdynamicSubGopAsString()
.- Returns:
- Specify whether to allow the number of B-frames in your output GOP structure to vary or not depending on your input video content. To improve the subjective video quality of your output that has high-motion content: Leave blank or keep the default value Adaptive. MediaConvert will use fewer B-frames for high-motion video content than low-motion content. The maximum number of B- frames is limited by the value that you choose for B-frames between reference frames. To use the same number B-frames for all types of content: Choose Static.
- See Also:
-
dynamicSubGopAsString
Specify whether to allow the number of B-frames in your output GOP structure to vary or not depending on your input video content. To improve the subjective video quality of your output that has high-motion content: Leave blank or keep the default value Adaptive. MediaConvert will use fewer B-frames for high-motion video content than low-motion content. The maximum number of B- frames is limited by the value that you choose for B-frames between reference frames. To use the same number B-frames for all types of content: Choose Static.If the service returns an enum value that is not available in the current SDK version,
dynamicSubGop
will returnH265DynamicSubGop.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromdynamicSubGopAsString()
.- Returns:
- Specify whether to allow the number of B-frames in your output GOP structure to vary or not depending on your input video content. To improve the subjective video quality of your output that has high-motion content: Leave blank or keep the default value Adaptive. MediaConvert will use fewer B-frames for high-motion video content than low-motion content. The maximum number of B- frames is limited by the value that you choose for B-frames between reference frames. To use the same number B-frames for all types of content: Choose Static.
- See Also:
-
endOfStreamMarkers
Optionally include or suppress markers at the end of your output that signal the end of the video stream. To include end of stream markers: Leave blank or keep the default value, Include. To not include end of stream markers: Choose Suppress. This is useful when your output will be inserted into another stream.If the service returns an enum value that is not available in the current SDK version,
endOfStreamMarkers
will returnH265EndOfStreamMarkers.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromendOfStreamMarkersAsString()
.- Returns:
- Optionally include or suppress markers at the end of your output that signal the end of the video stream. To include end of stream markers: Leave blank or keep the default value, Include. To not include end of stream markers: Choose Suppress. This is useful when your output will be inserted into another stream.
- See Also:
-
endOfStreamMarkersAsString
Optionally include or suppress markers at the end of your output that signal the end of the video stream. To include end of stream markers: Leave blank or keep the default value, Include. To not include end of stream markers: Choose Suppress. This is useful when your output will be inserted into another stream.If the service returns an enum value that is not available in the current SDK version,
endOfStreamMarkers
will returnH265EndOfStreamMarkers.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromendOfStreamMarkersAsString()
.- Returns:
- Optionally include or suppress markers at the end of your output that signal the end of the video stream. To include end of stream markers: Leave blank or keep the default value, Include. To not include end of stream markers: Choose Suppress. This is useful when your output will be inserted into another stream.
- See Also:
-
flickerAdaptiveQuantization
Enable this setting to have the encoder reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. This setting is disabled by default. Related setting: In addition to enabling this setting, you must also set adaptiveQuantization to a value other than Off.If the service returns an enum value that is not available in the current SDK version,
flickerAdaptiveQuantization
will returnH265FlickerAdaptiveQuantization.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromflickerAdaptiveQuantizationAsString()
.- Returns:
- Enable this setting to have the encoder reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. This setting is disabled by default. Related setting: In addition to enabling this setting, you must also set adaptiveQuantization to a value other than Off.
- See Also:
-
flickerAdaptiveQuantizationAsString
Enable this setting to have the encoder reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. This setting is disabled by default. Related setting: In addition to enabling this setting, you must also set adaptiveQuantization to a value other than Off.If the service returns an enum value that is not available in the current SDK version,
flickerAdaptiveQuantization
will returnH265FlickerAdaptiveQuantization.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromflickerAdaptiveQuantizationAsString()
.- Returns:
- Enable this setting to have the encoder reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. This setting is disabled by default. Related setting: In addition to enabling this setting, you must also set adaptiveQuantization to a value other than Off.
- See Also:
-
framerateControl
Use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.If the service returns an enum value that is not available in the current SDK version,
framerateControl
will returnH265FramerateControl.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromframerateControlAsString()
.- Returns:
- Use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.
- See Also:
-
framerateControlAsString
Use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.If the service returns an enum value that is not available in the current SDK version,
framerateControl
will returnH265FramerateControl.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromframerateControlAsString()
.- Returns:
- Use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.
- See Also:
-
framerateConversionAlgorithm
Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.If the service returns an enum value that is not available in the current SDK version,
framerateConversionAlgorithm
will returnH265FramerateConversionAlgorithm.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromframerateConversionAlgorithmAsString()
.- Returns:
- Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
- See Also:
-
framerateConversionAlgorithmAsString
Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.If the service returns an enum value that is not available in the current SDK version,
framerateConversionAlgorithm
will returnH265FramerateConversionAlgorithm.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromframerateConversionAlgorithmAsString()
.- Returns:
- Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
- See Also:
-
framerateDenominator
When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.- Returns:
- When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
-
framerateNumerator
When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.- Returns:
- When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
-
gopBReference
Specify whether to allow B-frames to be referenced by other frame types. To use reference B-frames when your GOP structure has 1 or more B-frames: Leave blank or keep the default value Enabled. We recommend that you choose Enabled to help improve the video quality of your output relative to its bitrate. To not use reference B-frames: Choose Disabled.If the service returns an enum value that is not available in the current SDK version,
gopBReference
will returnH265GopBReference.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromgopBReferenceAsString()
.- Returns:
- Specify whether to allow B-frames to be referenced by other frame types. To use reference B-frames when your GOP structure has 1 or more B-frames: Leave blank or keep the default value Enabled. We recommend that you choose Enabled to help improve the video quality of your output relative to its bitrate. To not use reference B-frames: Choose Disabled.
- See Also:
-
gopBReferenceAsString
Specify whether to allow B-frames to be referenced by other frame types. To use reference B-frames when your GOP structure has 1 or more B-frames: Leave blank or keep the default value Enabled. We recommend that you choose Enabled to help improve the video quality of your output relative to its bitrate. To not use reference B-frames: Choose Disabled.If the service returns an enum value that is not available in the current SDK version,
gopBReference
will returnH265GopBReference.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromgopBReferenceAsString()
.- Returns:
- Specify whether to allow B-frames to be referenced by other frame types. To use reference B-frames when your GOP structure has 1 or more B-frames: Leave blank or keep the default value Enabled. We recommend that you choose Enabled to help improve the video quality of your output relative to its bitrate. To not use reference B-frames: Choose Disabled.
- See Also:
-
gopClosedCadence
Specify the relative frequency of open to closed GOPs in this output. For example, if you want to allow four open GOPs and then require a closed GOP, set this value to 5. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, do this by keeping the default empty value. If you do explicitly specify a value, for segmented outputs, don't set this value to 0.- Returns:
- Specify the relative frequency of open to closed GOPs in this output. For example, if you want to allow four open GOPs and then require a closed GOP, set this value to 5. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, do this by keeping the default empty value. If you do explicitly specify a value, for segmented outputs, don't set this value to 0.
-
gopSize
Use this setting only when you set GOP mode control to Specified, frames or Specified, seconds. Specify the GOP length using a whole number of frames or a decimal value of seconds. MediaConvert will interpret this value as frames or seconds depending on the value you choose for GOP mode control. If you want to allow MediaConvert to automatically determine GOP size, leave GOP size blank and set GOP mode control to Auto. If your output group specifies HLS, DASH, or CMAF, leave GOP size blank and set GOP mode control to Auto in each output in your output group.- Returns:
- Use this setting only when you set GOP mode control to Specified, frames or Specified, seconds. Specify the GOP length using a whole number of frames or a decimal value of seconds. MediaConvert will interpret this value as frames or seconds depending on the value you choose for GOP mode control. If you want to allow MediaConvert to automatically determine GOP size, leave GOP size blank and set GOP mode control to Auto. If your output group specifies HLS, DASH, or CMAF, leave GOP size blank and set GOP mode control to Auto in each output in your output group.
-
gopSizeUnits
Specify how the transcoder determines GOP size for this output. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, choose Auto and and leave GOP size blank. By default, if you don't specify GOP mode control, MediaConvert will use automatic behavior. If your output group specifies HLS, DASH, or CMAF, set GOP mode control to Auto and leave GOP size blank in each output in your output group. To explicitly specify the GOP length, choose Specified, frames or Specified, seconds and then provide the GOP length in the related setting GOP size.If the service returns an enum value that is not available in the current SDK version,
gopSizeUnits
will returnH265GopSizeUnits.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromgopSizeUnitsAsString()
.- Returns:
- Specify how the transcoder determines GOP size for this output. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, choose Auto and and leave GOP size blank. By default, if you don't specify GOP mode control, MediaConvert will use automatic behavior. If your output group specifies HLS, DASH, or CMAF, set GOP mode control to Auto and leave GOP size blank in each output in your output group. To explicitly specify the GOP length, choose Specified, frames or Specified, seconds and then provide the GOP length in the related setting GOP size.
- See Also:
-
gopSizeUnitsAsString
Specify how the transcoder determines GOP size for this output. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, choose Auto and and leave GOP size blank. By default, if you don't specify GOP mode control, MediaConvert will use automatic behavior. If your output group specifies HLS, DASH, or CMAF, set GOP mode control to Auto and leave GOP size blank in each output in your output group. To explicitly specify the GOP length, choose Specified, frames or Specified, seconds and then provide the GOP length in the related setting GOP size.If the service returns an enum value that is not available in the current SDK version,
gopSizeUnits
will returnH265GopSizeUnits.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromgopSizeUnitsAsString()
.- Returns:
- Specify how the transcoder determines GOP size for this output. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, choose Auto and and leave GOP size blank. By default, if you don't specify GOP mode control, MediaConvert will use automatic behavior. If your output group specifies HLS, DASH, or CMAF, set GOP mode control to Auto and leave GOP size blank in each output in your output group. To explicitly specify the GOP length, choose Specified, frames or Specified, seconds and then provide the GOP length in the related setting GOP size.
- See Also:
-
hrdBufferFinalFillPercentage
If your downstream systems have strict buffer requirements: Specify the minimum percentage of the HRD buffer that's available at the end of each encoded video segment. For the best video quality: Set to 0 or leave blank to automatically determine the final buffer fill percentage.- Returns:
- If your downstream systems have strict buffer requirements: Specify the minimum percentage of the HRD buffer that's available at the end of each encoded video segment. For the best video quality: Set to 0 or leave blank to automatically determine the final buffer fill percentage.
-
hrdBufferInitialFillPercentage
Percentage of the buffer that should initially be filled (HRD buffer model).- Returns:
- Percentage of the buffer that should initially be filled (HRD buffer model).
-
hrdBufferSize
Size of buffer (HRD buffer model) in bits. For example, enter five megabits as 5000000.- Returns:
- Size of buffer (HRD buffer model) in bits. For example, enter five megabits as 5000000.
-
interlaceMode
Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.If the service returns an enum value that is not available in the current SDK version,
interlaceMode
will returnH265InterlaceMode.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available frominterlaceModeAsString()
.- Returns:
- Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.
- See Also:
-
interlaceModeAsString
Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.If the service returns an enum value that is not available in the current SDK version,
interlaceMode
will returnH265InterlaceMode.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available frominterlaceModeAsString()
.- Returns:
- Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.
- See Also:
-
maxBitrate
Maximum bitrate in bits/second. For example, enter five megabits per second as 5000000. Required when Rate control mode is QVBR.- Returns:
- Maximum bitrate in bits/second. For example, enter five megabits per second as 5000000. Required when Rate control mode is QVBR.
-
minIInterval
Use this setting only when you also enable Scene change detection. This setting determines how the encoder manages the spacing between I-frames that it inserts as part of the I-frame cadence and the I-frames that it inserts for Scene change detection. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, do this by keeping the default empty value. When you explicitly specify a value for this setting, the encoder determines whether to skip a cadence-driven I-frame by the value you set. For example, if you set Min I interval to 5 and a cadence-driven I-frame would fall within 5 frames of a scene-change I-frame, then the encoder skips the cadence-driven I-frame. In this way, one GOP is shrunk slightly and one GOP is stretched slightly. When the cadence-driven I-frames are farther from the scene-change I-frame than the value you set, then the encoder leaves all I-frames in place and the GOPs surrounding the scene change are smaller than the usual cadence GOPs.- Returns:
- Use this setting only when you also enable Scene change detection. This setting determines how the encoder manages the spacing between I-frames that it inserts as part of the I-frame cadence and the I-frames that it inserts for Scene change detection. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, do this by keeping the default empty value. When you explicitly specify a value for this setting, the encoder determines whether to skip a cadence-driven I-frame by the value you set. For example, if you set Min I interval to 5 and a cadence-driven I-frame would fall within 5 frames of a scene-change I-frame, then the encoder skips the cadence-driven I-frame. In this way, one GOP is shrunk slightly and one GOP is stretched slightly. When the cadence-driven I-frames are farther from the scene-change I-frame than the value you set, then the encoder leaves all I-frames in place and the GOPs surrounding the scene change are smaller than the usual cadence GOPs.
-
numberBFramesBetweenReferenceFrames
Specify the number of B-frames between reference frames in this output. For the best video quality: Leave blank. MediaConvert automatically determines the number of B-frames to use based on the characteristics of your input video. To manually specify the number of B-frames between reference frames: Enter an integer from 0 to 7.- Returns:
- Specify the number of B-frames between reference frames in this output. For the best video quality: Leave blank. MediaConvert automatically determines the number of B-frames to use based on the characteristics of your input video. To manually specify the number of B-frames between reference frames: Enter an integer from 0 to 7.
-
numberReferenceFrames
Number of reference frames to use. The encoder may use more than requested if using B-frames and/or interlaced encoding.- Returns:
- Number of reference frames to use. The encoder may use more than requested if using B-frames and/or interlaced encoding.
-
parControl
Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.If the service returns an enum value that is not available in the current SDK version,
parControl
will returnH265ParControl.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromparControlAsString()
.- Returns:
- Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.
- See Also:
-
parControlAsString
Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.If the service returns an enum value that is not available in the current SDK version,
parControl
will returnH265ParControl.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromparControlAsString()
.- Returns:
- Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.
- See Also:
-
parDenominator
Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33.- Returns:
- Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33.
-
parNumerator
Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40.- Returns:
- Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40.
-
qualityTuningLevel
Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.If the service returns an enum value that is not available in the current SDK version,
qualityTuningLevel
will returnH265QualityTuningLevel.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromqualityTuningLevelAsString()
.- Returns:
- Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.
- See Also:
-
qualityTuningLevelAsString
Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.If the service returns an enum value that is not available in the current SDK version,
qualityTuningLevel
will returnH265QualityTuningLevel.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromqualityTuningLevelAsString()
.- Returns:
- Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.
- See Also:
-
qvbrSettings
Settings for quality-defined variable bitrate encoding with the H.265 codec. Use these settings only when you set QVBR for Rate control mode.- Returns:
- Settings for quality-defined variable bitrate encoding with the H.265 codec. Use these settings only when you set QVBR for Rate control mode.
-
rateControlMode
Use this setting to specify whether this output has a variable bitrate (VBR), constant bitrate (CBR) or quality-defined variable bitrate (QVBR).If the service returns an enum value that is not available in the current SDK version,
rateControlMode
will returnH265RateControlMode.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromrateControlModeAsString()
.- Returns:
- Use this setting to specify whether this output has a variable bitrate (VBR), constant bitrate (CBR) or quality-defined variable bitrate (QVBR).
- See Also:
-
rateControlModeAsString
Use this setting to specify whether this output has a variable bitrate (VBR), constant bitrate (CBR) or quality-defined variable bitrate (QVBR).If the service returns an enum value that is not available in the current SDK version,
rateControlMode
will returnH265RateControlMode.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromrateControlModeAsString()
.- Returns:
- Use this setting to specify whether this output has a variable bitrate (VBR), constant bitrate (CBR) or quality-defined variable bitrate (QVBR).
- See Also:
-
sampleAdaptiveOffsetFilterMode
Specify Sample Adaptive Offset (SAO) filter strength. Adaptive mode dynamically selects best strength based on contentIf the service returns an enum value that is not available in the current SDK version,
sampleAdaptiveOffsetFilterMode
will returnH265SampleAdaptiveOffsetFilterMode.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromsampleAdaptiveOffsetFilterModeAsString()
.- Returns:
- Specify Sample Adaptive Offset (SAO) filter strength. Adaptive mode dynamically selects best strength based on content
- See Also:
-
sampleAdaptiveOffsetFilterModeAsString
Specify Sample Adaptive Offset (SAO) filter strength. Adaptive mode dynamically selects best strength based on contentIf the service returns an enum value that is not available in the current SDK version,
sampleAdaptiveOffsetFilterMode
will returnH265SampleAdaptiveOffsetFilterMode.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromsampleAdaptiveOffsetFilterModeAsString()
.- Returns:
- Specify Sample Adaptive Offset (SAO) filter strength. Adaptive mode dynamically selects best strength based on content
- See Also:
-
scanTypeConversionMode
Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.If the service returns an enum value that is not available in the current SDK version,
scanTypeConversionMode
will returnH265ScanTypeConversionMode.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromscanTypeConversionModeAsString()
.- Returns:
- Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.
- See Also:
-
scanTypeConversionModeAsString
Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.If the service returns an enum value that is not available in the current SDK version,
scanTypeConversionMode
will returnH265ScanTypeConversionMode.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromscanTypeConversionModeAsString()
.- Returns:
- Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.
- See Also:
-
sceneChangeDetect
Enable this setting to insert I-frames at scene changes that the service automatically detects. This improves video quality and is enabled by default. If this output uses QVBR, choose Transition detection for further video quality improvement. For more information about QVBR, see https://docs.aws.amazon.com/console/mediaconvert/cbr-vbr-qvbr.If the service returns an enum value that is not available in the current SDK version,
sceneChangeDetect
will returnH265SceneChangeDetect.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromsceneChangeDetectAsString()
.- Returns:
- Enable this setting to insert I-frames at scene changes that the service automatically detects. This improves video quality and is enabled by default. If this output uses QVBR, choose Transition detection for further video quality improvement. For more information about QVBR, see https://docs.aws.amazon.com/console/mediaconvert/cbr-vbr-qvbr.
- See Also:
-
sceneChangeDetectAsString
Enable this setting to insert I-frames at scene changes that the service automatically detects. This improves video quality and is enabled by default. If this output uses QVBR, choose Transition detection for further video quality improvement. For more information about QVBR, see https://docs.aws.amazon.com/console/mediaconvert/cbr-vbr-qvbr.If the service returns an enum value that is not available in the current SDK version,
sceneChangeDetect
will returnH265SceneChangeDetect.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromsceneChangeDetectAsString()
.- Returns:
- Enable this setting to insert I-frames at scene changes that the service automatically detects. This improves video quality and is enabled by default. If this output uses QVBR, choose Transition detection for further video quality improvement. For more information about QVBR, see https://docs.aws.amazon.com/console/mediaconvert/cbr-vbr-qvbr.
- See Also:
-
slices
Number of slices per picture. Must be less than or equal to the number of macroblock rows for progressive pictures, and less than or equal to half the number of macroblock rows for interlaced pictures.- Returns:
- Number of slices per picture. Must be less than or equal to the number of macroblock rows for progressive pictures, and less than or equal to half the number of macroblock rows for interlaced pictures.
-
slowPal
Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25.If the service returns an enum value that is not available in the current SDK version,
slowPal
will returnH265SlowPal.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromslowPalAsString()
.- Returns:
- Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25.
- See Also:
-
slowPalAsString
Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25.If the service returns an enum value that is not available in the current SDK version,
slowPal
will returnH265SlowPal.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromslowPalAsString()
.- Returns:
- Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25.
- See Also:
-
spatialAdaptiveQuantization
Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.If the service returns an enum value that is not available in the current SDK version,
spatialAdaptiveQuantization
will returnH265SpatialAdaptiveQuantization.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromspatialAdaptiveQuantizationAsString()
.- Returns:
- Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.
- See Also:
-
spatialAdaptiveQuantizationAsString
Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.If the service returns an enum value that is not available in the current SDK version,
spatialAdaptiveQuantization
will returnH265SpatialAdaptiveQuantization.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromspatialAdaptiveQuantizationAsString()
.- Returns:
- Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.
- See Also:
-
telecine
This field applies only if the Streams > Advanced > Framerate field is set to 29.970. This field works with the Streams > Advanced > Preprocessors > Deinterlacer field and the Streams > Advanced > Interlaced Mode field to identify the scan type for the output: Progressive, Interlaced, Hard Telecine or Soft Telecine. - Hard: produces 29.97i output from 23.976 input. - Soft: produces 23.976; the player converts this output to 29.97i.If the service returns an enum value that is not available in the current SDK version,
telecine
will returnH265Telecine.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromtelecineAsString()
.- Returns:
- This field applies only if the Streams > Advanced > Framerate field is set to 29.970. This field works with the Streams > Advanced > Preprocessors > Deinterlacer field and the Streams > Advanced > Interlaced Mode field to identify the scan type for the output: Progressive, Interlaced, Hard Telecine or Soft Telecine. - Hard: produces 29.97i output from 23.976 input. - Soft: produces 23.976; the player converts this output to 29.97i.
- See Also:
-
telecineAsString
This field applies only if the Streams > Advanced > Framerate field is set to 29.970. This field works with the Streams > Advanced > Preprocessors > Deinterlacer field and the Streams > Advanced > Interlaced Mode field to identify the scan type for the output: Progressive, Interlaced, Hard Telecine or Soft Telecine. - Hard: produces 29.97i output from 23.976 input. - Soft: produces 23.976; the player converts this output to 29.97i.If the service returns an enum value that is not available in the current SDK version,
telecine
will returnH265Telecine.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromtelecineAsString()
.- Returns:
- This field applies only if the Streams > Advanced > Framerate field is set to 29.970. This field works with the Streams > Advanced > Preprocessors > Deinterlacer field and the Streams > Advanced > Interlaced Mode field to identify the scan type for the output: Progressive, Interlaced, Hard Telecine or Soft Telecine. - Hard: produces 29.97i output from 23.976 input. - Soft: produces 23.976; the player converts this output to 29.97i.
- See Also:
-
temporalAdaptiveQuantization
Keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization.If the service returns an enum value that is not available in the current SDK version,
temporalAdaptiveQuantization
will returnH265TemporalAdaptiveQuantization.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromtemporalAdaptiveQuantizationAsString()
.- Returns:
- Keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization.
- See Also:
-
temporalAdaptiveQuantizationAsString
Keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization.If the service returns an enum value that is not available in the current SDK version,
temporalAdaptiveQuantization
will returnH265TemporalAdaptiveQuantization.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromtemporalAdaptiveQuantizationAsString()
.- Returns:
- Keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization.
- See Also:
-
temporalIds
Enables temporal layer identifiers in the encoded bitstream. Up to 3 layers are supported depending on GOP structure: I- and P-frames form one layer, reference B-frames can form a second layer and non-reference b-frames can form a third layer. Decoders can optionally decode only the lower temporal layers to generate a lower frame rate output. For example, given a bitstream with temporal IDs and with b-frames = 1 (i.e. IbPbPb display order), a decoder could decode all the frames for full frame rate output or only the I and P frames (lowest temporal layer) for a half frame rate output.If the service returns an enum value that is not available in the current SDK version,
temporalIds
will returnH265TemporalIds.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromtemporalIdsAsString()
.- Returns:
- Enables temporal layer identifiers in the encoded bitstream. Up to 3 layers are supported depending on GOP structure: I- and P-frames form one layer, reference B-frames can form a second layer and non-reference b-frames can form a third layer. Decoders can optionally decode only the lower temporal layers to generate a lower frame rate output. For example, given a bitstream with temporal IDs and with b-frames = 1 (i.e. IbPbPb display order), a decoder could decode all the frames for full frame rate output or only the I and P frames (lowest temporal layer) for a half frame rate output.
- See Also:
-
temporalIdsAsString
Enables temporal layer identifiers in the encoded bitstream. Up to 3 layers are supported depending on GOP structure: I- and P-frames form one layer, reference B-frames can form a second layer and non-reference b-frames can form a third layer. Decoders can optionally decode only the lower temporal layers to generate a lower frame rate output. For example, given a bitstream with temporal IDs and with b-frames = 1 (i.e. IbPbPb display order), a decoder could decode all the frames for full frame rate output or only the I and P frames (lowest temporal layer) for a half frame rate output.If the service returns an enum value that is not available in the current SDK version,
temporalIds
will returnH265TemporalIds.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromtemporalIdsAsString()
.- Returns:
- Enables temporal layer identifiers in the encoded bitstream. Up to 3 layers are supported depending on GOP structure: I- and P-frames form one layer, reference B-frames can form a second layer and non-reference b-frames can form a third layer. Decoders can optionally decode only the lower temporal layers to generate a lower frame rate output. For example, given a bitstream with temporal IDs and with b-frames = 1 (i.e. IbPbPb display order), a decoder could decode all the frames for full frame rate output or only the I and P frames (lowest temporal layer) for a half frame rate output.
- See Also:
-
tiles
Enable use of tiles, allowing horizontal as well as vertical subdivision of the encoded pictures.If the service returns an enum value that is not available in the current SDK version,
tiles
will returnH265Tiles.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromtilesAsString()
.- Returns:
- Enable use of tiles, allowing horizontal as well as vertical subdivision of the encoded pictures.
- See Also:
-
tilesAsString
Enable use of tiles, allowing horizontal as well as vertical subdivision of the encoded pictures.If the service returns an enum value that is not available in the current SDK version,
tiles
will returnH265Tiles.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromtilesAsString()
.- Returns:
- Enable use of tiles, allowing horizontal as well as vertical subdivision of the encoded pictures.
- See Also:
-
unregisteredSeiTimecode
Inserts timecode for each frame as 4 bytes of an unregistered SEI message.If the service returns an enum value that is not available in the current SDK version,
unregisteredSeiTimecode
will returnH265UnregisteredSeiTimecode.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromunregisteredSeiTimecodeAsString()
.- Returns:
- Inserts timecode for each frame as 4 bytes of an unregistered SEI message.
- See Also:
-
unregisteredSeiTimecodeAsString
Inserts timecode for each frame as 4 bytes of an unregistered SEI message.If the service returns an enum value that is not available in the current SDK version,
unregisteredSeiTimecode
will returnH265UnregisteredSeiTimecode.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromunregisteredSeiTimecodeAsString()
.- Returns:
- Inserts timecode for each frame as 4 bytes of an unregistered SEI message.
- See Also:
-
writeMp4PackagingType
If the location of parameter set NAL units doesn't matter in your workflow, ignore this setting. Use this setting only with CMAF or DASH outputs, or with standalone file outputs in an MPEG-4 container (MP4 outputs). Choose HVC1 to mark your output as HVC1. This makes your output compliant with the following specification: ISO IECJTC1 SC29 N13798 Text ISO/IEC FDIS 14496-15 3rd Edition. For these outputs, the service stores parameter set NAL units in the sample headers but not in the samples directly. For MP4 outputs, when you choose HVC1, your output video might not work properly with some downstream systems and video players. The service defaults to marking your output as HEV1. For these outputs, the service writes parameter set NAL units directly into the samples.If the service returns an enum value that is not available in the current SDK version,
writeMp4PackagingType
will returnH265WriteMp4PackagingType.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromwriteMp4PackagingTypeAsString()
.- Returns:
- If the location of parameter set NAL units doesn't matter in your workflow, ignore this setting. Use this setting only with CMAF or DASH outputs, or with standalone file outputs in an MPEG-4 container (MP4 outputs). Choose HVC1 to mark your output as HVC1. This makes your output compliant with the following specification: ISO IECJTC1 SC29 N13798 Text ISO/IEC FDIS 14496-15 3rd Edition. For these outputs, the service stores parameter set NAL units in the sample headers but not in the samples directly. For MP4 outputs, when you choose HVC1, your output video might not work properly with some downstream systems and video players. The service defaults to marking your output as HEV1. For these outputs, the service writes parameter set NAL units directly into the samples.
- See Also:
-
writeMp4PackagingTypeAsString
If the location of parameter set NAL units doesn't matter in your workflow, ignore this setting. Use this setting only with CMAF or DASH outputs, or with standalone file outputs in an MPEG-4 container (MP4 outputs). Choose HVC1 to mark your output as HVC1. This makes your output compliant with the following specification: ISO IECJTC1 SC29 N13798 Text ISO/IEC FDIS 14496-15 3rd Edition. For these outputs, the service stores parameter set NAL units in the sample headers but not in the samples directly. For MP4 outputs, when you choose HVC1, your output video might not work properly with some downstream systems and video players. The service defaults to marking your output as HEV1. For these outputs, the service writes parameter set NAL units directly into the samples.If the service returns an enum value that is not available in the current SDK version,
writeMp4PackagingType
will returnH265WriteMp4PackagingType.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromwriteMp4PackagingTypeAsString()
.- Returns:
- If the location of parameter set NAL units doesn't matter in your workflow, ignore this setting. Use this setting only with CMAF or DASH outputs, or with standalone file outputs in an MPEG-4 container (MP4 outputs). Choose HVC1 to mark your output as HVC1. This makes your output compliant with the following specification: ISO IECJTC1 SC29 N13798 Text ISO/IEC FDIS 14496-15 3rd Edition. For these outputs, the service stores parameter set NAL units in the sample headers but not in the samples directly. For MP4 outputs, when you choose HVC1, your output video might not work properly with some downstream systems and video players. The service defaults to marking your output as HEV1. For these outputs, the service writes parameter set NAL units directly into the samples.
- See Also:
-
toBuilder
Description copied from interface:ToCopyableBuilder
Take this object and create a builder that contains all of the current property values of this object.- Specified by:
toBuilder
in interfaceToCopyableBuilder<H265Settings.Builder,
H265Settings> - Returns:
- a builder for type T
-
builder
-
serializableBuilderClass
-
hashCode
public final int hashCode() -
equals
-
equalsBySdkFields
Description copied from interface:SdkPojo
Indicates whether some other object is "equal to" this one by SDK fields. An SDK field is a modeled, non-inherited field in anSdkPojo
class, and is generated based on a service model.If an
SdkPojo
class does not have any inherited fields,equalsBySdkFields
andequals
are essentially the same.- Specified by:
equalsBySdkFields
in interfaceSdkPojo
- Parameters:
obj
- the object to be compared with- Returns:
- true if the other object equals to this object by sdk fields, false otherwise.
-
toString
Returns a string representation of this object. This is useful for testing and debugging. Sensitive data will be redacted from this string using a placeholder value. -
getValueForField
-
sdkFields
-