Class Mpeg2Settings
- All Implemented Interfaces:
Serializable
,SdkPojo
,ToCopyableBuilder<Mpeg2Settings.Builder,
Mpeg2Settings>
- See Also:
-
Nested Class Summary
-
Method Summary
Modifier and TypeMethodDescriptionSpecify the strength of any adaptive quantization filters that you enable.final String
Specify the strength of any adaptive quantization filters that you enable.final Integer
bitrate()
Specify the average bitrate in bits per second.static Mpeg2Settings.Builder
builder()
final Mpeg2CodecLevel
Use Level to set the MPEG-2 level for the video output.final String
Use Level to set the MPEG-2 level for the video output.final Mpeg2CodecProfile
Use Profile to set the MPEG-2 profile for the video output.final String
Use Profile to set the MPEG-2 profile for the video output.final Mpeg2DynamicSubGop
Choose Adaptive to improve subjective video quality for high-motion content.final String
Choose Adaptive to improve subjective video quality for high-motion content.final boolean
final boolean
equalsBySdkFields
(Object obj) Indicates whether some other object is "equal to" this one by SDK fields.final Mpeg2FramerateControl
If you are using the console, use the Framerate setting to specify the frame rate for this output.final String
If you are using the console, use the Framerate setting to specify the frame rate for this output.Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate.final String
Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate.final Integer
When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction.final Integer
When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction.final <T> Optional
<T> getValueForField
(String fieldName, Class<T> clazz) final Integer
Specify the relative frequency of open to closed GOPs in this output.final Double
gopSize()
Specify the interval between keyframes, in seconds or frames, for this output.final Mpeg2GopSizeUnits
Specify the units for GOP size.final String
Specify the units for GOP size.final int
hashCode()
final Integer
If your downstream systems have strict buffer requirements: Specify the minimum percentage of the HRD buffer that's available at the end of each encoded video segment.final Integer
Percentage of the buffer that should initially be filled (HRD buffer model).final Integer
Size of buffer (HRD buffer model) in bits.final Mpeg2InterlaceMode
Choose the scan line type for the output.final String
Choose the scan line type for the output.final Mpeg2IntraDcPrecision
Use Intra DC precision to set quantization precision for intra-block DC coefficients.final String
Use Intra DC precision to set quantization precision for intra-block DC coefficients.final Integer
Maximum bitrate in bits/second.final Integer
Use this setting only when you also enable Scene change detection.final Integer
Specify the number of B-frames that MediaConvert puts between reference frames in this output.final Mpeg2ParControl
Optional.final String
Optional.final Integer
Required when you set Pixel aspect ratio to SPECIFIED.final Integer
Required when you set Pixel aspect ratio to SPECIFIED.final Mpeg2QualityTuningLevel
Optional.final String
Optional.final Mpeg2RateControlMode
Use Rate control mode to specify whether the bitrate is variable (vbr) or constant (cbr).final String
Use Rate control mode to specify whether the bitrate is variable (vbr) or constant (cbr).Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate.final String
Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate.final Mpeg2SceneChangeDetect
Enable this setting to insert I-frames at scene changes that the service automatically detects.final String
Enable this setting to insert I-frames at scene changes that the service automatically detects.static Class
<? extends Mpeg2Settings.Builder> final Mpeg2SlowPal
slowPal()
Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps).final String
Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps).final Integer
softness()
Ignore this setting unless you need to comply with a specification that requires a specific value.Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity.final String
Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity.final Mpeg2Syntax
syntax()
Specify whether this output's video uses the D10 syntax.final String
Specify whether this output's video uses the D10 syntax.final Mpeg2Telecine
telecine()
When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard or soft telecine to create a smoother picture.final String
When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard or soft telecine to create a smoother picture.Keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity.final String
Keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity.Take this object and create a builder that contains all of the current property values of this object.final String
toString()
Returns a string representation of this object.Methods inherited from interface software.amazon.awssdk.utils.builder.ToCopyableBuilder
copy
-
Method Details
-
adaptiveQuantization
Specify the strength of any adaptive quantization filters that you enable. The value that you choose here applies to the following settings: Spatial adaptive quantization, and Temporal adaptive quantization.If the service returns an enum value that is not available in the current SDK version,
adaptiveQuantization
will returnMpeg2AdaptiveQuantization.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromadaptiveQuantizationAsString()
.- Returns:
- Specify the strength of any adaptive quantization filters that you enable. The value that you choose here applies to the following settings: Spatial adaptive quantization, and Temporal adaptive quantization.
- See Also:
-
adaptiveQuantizationAsString
Specify the strength of any adaptive quantization filters that you enable. The value that you choose here applies to the following settings: Spatial adaptive quantization, and Temporal adaptive quantization.If the service returns an enum value that is not available in the current SDK version,
adaptiveQuantization
will returnMpeg2AdaptiveQuantization.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromadaptiveQuantizationAsString()
.- Returns:
- Specify the strength of any adaptive quantization filters that you enable. The value that you choose here applies to the following settings: Spatial adaptive quantization, and Temporal adaptive quantization.
- See Also:
-
bitrate
Specify the average bitrate in bits per second. Required for VBR and CBR. For MS Smooth outputs, bitrates must be unique when rounded down to the nearest multiple of 1000.- Returns:
- Specify the average bitrate in bits per second. Required for VBR and CBR. For MS Smooth outputs, bitrates must be unique when rounded down to the nearest multiple of 1000.
-
codecLevel
Use Level to set the MPEG-2 level for the video output.If the service returns an enum value that is not available in the current SDK version,
codecLevel
will returnMpeg2CodecLevel.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromcodecLevelAsString()
.- Returns:
- Use Level to set the MPEG-2 level for the video output.
- See Also:
-
codecLevelAsString
Use Level to set the MPEG-2 level for the video output.If the service returns an enum value that is not available in the current SDK version,
codecLevel
will returnMpeg2CodecLevel.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromcodecLevelAsString()
.- Returns:
- Use Level to set the MPEG-2 level for the video output.
- See Also:
-
codecProfile
Use Profile to set the MPEG-2 profile for the video output.If the service returns an enum value that is not available in the current SDK version,
codecProfile
will returnMpeg2CodecProfile.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromcodecProfileAsString()
.- Returns:
- Use Profile to set the MPEG-2 profile for the video output.
- See Also:
-
codecProfileAsString
Use Profile to set the MPEG-2 profile for the video output.If the service returns an enum value that is not available in the current SDK version,
codecProfile
will returnMpeg2CodecProfile.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromcodecProfileAsString()
.- Returns:
- Use Profile to set the MPEG-2 profile for the video output.
- See Also:
-
dynamicSubGop
Choose Adaptive to improve subjective video quality for high-motion content. This will cause the service to use fewer B-frames (which infer information based on other frames) for high-motion portions of the video and more B-frames for low-motion portions. The maximum number of B-frames is limited by the value you provide for the setting B frames between reference frames.If the service returns an enum value that is not available in the current SDK version,
dynamicSubGop
will returnMpeg2DynamicSubGop.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromdynamicSubGopAsString()
.- Returns:
- Choose Adaptive to improve subjective video quality for high-motion content. This will cause the service to use fewer B-frames (which infer information based on other frames) for high-motion portions of the video and more B-frames for low-motion portions. The maximum number of B-frames is limited by the value you provide for the setting B frames between reference frames.
- See Also:
-
dynamicSubGopAsString
Choose Adaptive to improve subjective video quality for high-motion content. This will cause the service to use fewer B-frames (which infer information based on other frames) for high-motion portions of the video and more B-frames for low-motion portions. The maximum number of B-frames is limited by the value you provide for the setting B frames between reference frames.If the service returns an enum value that is not available in the current SDK version,
dynamicSubGop
will returnMpeg2DynamicSubGop.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromdynamicSubGopAsString()
.- Returns:
- Choose Adaptive to improve subjective video quality for high-motion content. This will cause the service to use fewer B-frames (which infer information based on other frames) for high-motion portions of the video and more B-frames for low-motion portions. The maximum number of B-frames is limited by the value you provide for the setting B frames between reference frames.
- See Also:
-
framerateControl
If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.If the service returns an enum value that is not available in the current SDK version,
framerateControl
will returnMpeg2FramerateControl.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromframerateControlAsString()
.- Returns:
- If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.
- See Also:
-
framerateControlAsString
If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.If the service returns an enum value that is not available in the current SDK version,
framerateControl
will returnMpeg2FramerateControl.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromframerateControlAsString()
.- Returns:
- If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.
- See Also:
-
framerateConversionAlgorithm
Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.If the service returns an enum value that is not available in the current SDK version,
framerateConversionAlgorithm
will returnMpeg2FramerateConversionAlgorithm.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromframerateConversionAlgorithmAsString()
.- Returns:
- Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
- See Also:
-
framerateConversionAlgorithmAsString
Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.If the service returns an enum value that is not available in the current SDK version,
framerateConversionAlgorithm
will returnMpeg2FramerateConversionAlgorithm.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromframerateConversionAlgorithmAsString()
.- Returns:
- Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
- See Also:
-
framerateDenominator
When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.- Returns:
- When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
-
framerateNumerator
When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.- Returns:
- When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
-
gopClosedCadence
Specify the relative frequency of open to closed GOPs in this output. For example, if you want to allow four open GOPs and then require a closed GOP, set this value to 5. When you create a streaming output, we recommend that you keep the default value, 1, so that players starting mid-stream receive an IDR frame as quickly as possible. Don't set this value to 0; that would break output segmenting.- Returns:
- Specify the relative frequency of open to closed GOPs in this output. For example, if you want to allow four open GOPs and then require a closed GOP, set this value to 5. When you create a streaming output, we recommend that you keep the default value, 1, so that players starting mid-stream receive an IDR frame as quickly as possible. Don't set this value to 0; that would break output segmenting.
-
gopSize
Specify the interval between keyframes, in seconds or frames, for this output. Default: 12 Related settings: When you specify the GOP size in seconds, set GOP mode control to Specified, seconds. The default value for GOP mode control is Frames.- Returns:
- Specify the interval between keyframes, in seconds or frames, for this output. Default: 12 Related settings: When you specify the GOP size in seconds, set GOP mode control to Specified, seconds. The default value for GOP mode control is Frames.
-
gopSizeUnits
Specify the units for GOP size. If you don't specify a value here, by default the encoder measures GOP size in frames.If the service returns an enum value that is not available in the current SDK version,
gopSizeUnits
will returnMpeg2GopSizeUnits.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromgopSizeUnitsAsString()
.- Returns:
- Specify the units for GOP size. If you don't specify a value here, by default the encoder measures GOP size in frames.
- See Also:
-
gopSizeUnitsAsString
Specify the units for GOP size. If you don't specify a value here, by default the encoder measures GOP size in frames.If the service returns an enum value that is not available in the current SDK version,
gopSizeUnits
will returnMpeg2GopSizeUnits.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromgopSizeUnitsAsString()
.- Returns:
- Specify the units for GOP size. If you don't specify a value here, by default the encoder measures GOP size in frames.
- See Also:
-
hrdBufferFinalFillPercentage
If your downstream systems have strict buffer requirements: Specify the minimum percentage of the HRD buffer that's available at the end of each encoded video segment. For the best video quality: Set to 0 or leave blank to automatically determine the final buffer fill percentage.- Returns:
- If your downstream systems have strict buffer requirements: Specify the minimum percentage of the HRD buffer that's available at the end of each encoded video segment. For the best video quality: Set to 0 or leave blank to automatically determine the final buffer fill percentage.
-
hrdBufferInitialFillPercentage
Percentage of the buffer that should initially be filled (HRD buffer model).- Returns:
- Percentage of the buffer that should initially be filled (HRD buffer model).
-
hrdBufferSize
Size of buffer (HRD buffer model) in bits. For example, enter five megabits as 5000000.- Returns:
- Size of buffer (HRD buffer model) in bits. For example, enter five megabits as 5000000.
-
interlaceMode
Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.If the service returns an enum value that is not available in the current SDK version,
interlaceMode
will returnMpeg2InterlaceMode.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available frominterlaceModeAsString()
.- Returns:
- Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.
- See Also:
-
interlaceModeAsString
Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.If the service returns an enum value that is not available in the current SDK version,
interlaceMode
will returnMpeg2InterlaceMode.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available frominterlaceModeAsString()
.- Returns:
- Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.
- See Also:
-
intraDcPrecision
Use Intra DC precision to set quantization precision for intra-block DC coefficients. If you choose the value auto, the service will automatically select the precision based on the per-frame compression ratio.If the service returns an enum value that is not available in the current SDK version,
intraDcPrecision
will returnMpeg2IntraDcPrecision.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromintraDcPrecisionAsString()
.- Returns:
- Use Intra DC precision to set quantization precision for intra-block DC coefficients. If you choose the value auto, the service will automatically select the precision based on the per-frame compression ratio.
- See Also:
-
intraDcPrecisionAsString
Use Intra DC precision to set quantization precision for intra-block DC coefficients. If you choose the value auto, the service will automatically select the precision based on the per-frame compression ratio.If the service returns an enum value that is not available in the current SDK version,
intraDcPrecision
will returnMpeg2IntraDcPrecision.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromintraDcPrecisionAsString()
.- Returns:
- Use Intra DC precision to set quantization precision for intra-block DC coefficients. If you choose the value auto, the service will automatically select the precision based on the per-frame compression ratio.
- See Also:
-
maxBitrate
Maximum bitrate in bits/second. For example, enter five megabits per second as 5000000.- Returns:
- Maximum bitrate in bits/second. For example, enter five megabits per second as 5000000.
-
minIInterval
Use this setting only when you also enable Scene change detection. This setting determines how the encoder manages the spacing between I-frames that it inserts as part of the I-frame cadence and the I-frames that it inserts for Scene change detection. When you specify a value for this setting, the encoder determines whether to skip a cadence-driven I-frame by the value you set. For example, if you set Min I interval to 5 and a cadence-driven I-frame would fall within 5 frames of a scene-change I-frame, then the encoder skips the cadence-driven I-frame. In this way, one GOP is shrunk slightly and one GOP is stretched slightly. When the cadence-driven I-frames are farther from the scene-change I-frame than the value you set, then the encoder leaves all I-frames in place and the GOPs surrounding the scene change are smaller than the usual cadence GOPs.- Returns:
- Use this setting only when you also enable Scene change detection. This setting determines how the encoder manages the spacing between I-frames that it inserts as part of the I-frame cadence and the I-frames that it inserts for Scene change detection. When you specify a value for this setting, the encoder determines whether to skip a cadence-driven I-frame by the value you set. For example, if you set Min I interval to 5 and a cadence-driven I-frame would fall within 5 frames of a scene-change I-frame, then the encoder skips the cadence-driven I-frame. In this way, one GOP is shrunk slightly and one GOP is stretched slightly. When the cadence-driven I-frames are farther from the scene-change I-frame than the value you set, then the encoder leaves all I-frames in place and the GOPs surrounding the scene change are smaller than the usual cadence GOPs.
-
numberBFramesBetweenReferenceFrames
Specify the number of B-frames that MediaConvert puts between reference frames in this output. Valid values are whole numbers from 0 through 7. When you don't specify a value, MediaConvert defaults to 2.- Returns:
- Specify the number of B-frames that MediaConvert puts between reference frames in this output. Valid values are whole numbers from 0 through 7. When you don't specify a value, MediaConvert defaults to 2.
-
parControl
Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.If the service returns an enum value that is not available in the current SDK version,
parControl
will returnMpeg2ParControl.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromparControlAsString()
.- Returns:
- Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.
- See Also:
-
parControlAsString
Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.If the service returns an enum value that is not available in the current SDK version,
parControl
will returnMpeg2ParControl.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromparControlAsString()
.- Returns:
- Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.
- See Also:
-
parDenominator
Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33.- Returns:
- Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33.
-
parNumerator
Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40.- Returns:
- Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40.
-
qualityTuningLevel
Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.If the service returns an enum value that is not available in the current SDK version,
qualityTuningLevel
will returnMpeg2QualityTuningLevel.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromqualityTuningLevelAsString()
.- Returns:
- Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.
- See Also:
-
qualityTuningLevelAsString
Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.If the service returns an enum value that is not available in the current SDK version,
qualityTuningLevel
will returnMpeg2QualityTuningLevel.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromqualityTuningLevelAsString()
.- Returns:
- Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.
- See Also:
-
rateControlMode
Use Rate control mode to specify whether the bitrate is variable (vbr) or constant (cbr).If the service returns an enum value that is not available in the current SDK version,
rateControlMode
will returnMpeg2RateControlMode.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromrateControlModeAsString()
.- Returns:
- Use Rate control mode to specify whether the bitrate is variable (vbr) or constant (cbr).
- See Also:
-
rateControlModeAsString
Use Rate control mode to specify whether the bitrate is variable (vbr) or constant (cbr).If the service returns an enum value that is not available in the current SDK version,
rateControlMode
will returnMpeg2RateControlMode.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromrateControlModeAsString()
.- Returns:
- Use Rate control mode to specify whether the bitrate is variable (vbr) or constant (cbr).
- See Also:
-
scanTypeConversionMode
Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.If the service returns an enum value that is not available in the current SDK version,
scanTypeConversionMode
will returnMpeg2ScanTypeConversionMode.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromscanTypeConversionModeAsString()
.- Returns:
- Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.
- See Also:
-
scanTypeConversionModeAsString
Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.If the service returns an enum value that is not available in the current SDK version,
scanTypeConversionMode
will returnMpeg2ScanTypeConversionMode.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromscanTypeConversionModeAsString()
.- Returns:
- Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.
- See Also:
-
sceneChangeDetect
Enable this setting to insert I-frames at scene changes that the service automatically detects. This improves video quality and is enabled by default.If the service returns an enum value that is not available in the current SDK version,
sceneChangeDetect
will returnMpeg2SceneChangeDetect.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromsceneChangeDetectAsString()
.- Returns:
- Enable this setting to insert I-frames at scene changes that the service automatically detects. This improves video quality and is enabled by default.
- See Also:
-
sceneChangeDetectAsString
Enable this setting to insert I-frames at scene changes that the service automatically detects. This improves video quality and is enabled by default.If the service returns an enum value that is not available in the current SDK version,
sceneChangeDetect
will returnMpeg2SceneChangeDetect.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromsceneChangeDetectAsString()
.- Returns:
- Enable this setting to insert I-frames at scene changes that the service automatically detects. This improves video quality and is enabled by default.
- See Also:
-
slowPal
Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25.If the service returns an enum value that is not available in the current SDK version,
slowPal
will returnMpeg2SlowPal.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromslowPalAsString()
.- Returns:
- Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25.
- See Also:
-
slowPalAsString
Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25.If the service returns an enum value that is not available in the current SDK version,
slowPal
will returnMpeg2SlowPal.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromslowPalAsString()
.- Returns:
- Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25.
- See Also:
-
softness
Ignore this setting unless you need to comply with a specification that requires a specific value. If you don't have a specification requirement, we recommend that you adjust the softness of your output by using a lower value for the setting Sharpness or by enabling a noise reducer filter. The Softness setting specifies the quantization matrices that the encoder uses. Keep the default value, 0, to use the AWS Elemental default matrices. Choose a value from 17 to 128 to use planar interpolation. Increasing values from 17 to 128 result in increasing reduction of high-frequency data. The value 128 results in the softest video.- Returns:
- Ignore this setting unless you need to comply with a specification that requires a specific value. If you don't have a specification requirement, we recommend that you adjust the softness of your output by using a lower value for the setting Sharpness or by enabling a noise reducer filter. The Softness setting specifies the quantization matrices that the encoder uses. Keep the default value, 0, to use the AWS Elemental default matrices. Choose a value from 17 to 128 to use planar interpolation. Increasing values from 17 to 128 result in increasing reduction of high-frequency data. The value 128 results in the softest video.
-
spatialAdaptiveQuantization
Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.If the service returns an enum value that is not available in the current SDK version,
spatialAdaptiveQuantization
will returnMpeg2SpatialAdaptiveQuantization.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromspatialAdaptiveQuantizationAsString()
.- Returns:
- Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.
- See Also:
-
spatialAdaptiveQuantizationAsString
Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.If the service returns an enum value that is not available in the current SDK version,
spatialAdaptiveQuantization
will returnMpeg2SpatialAdaptiveQuantization.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromspatialAdaptiveQuantizationAsString()
.- Returns:
- Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.
- See Also:
-
syntax
Specify whether this output's video uses the D10 syntax. Keep the default value to not use the syntax. Related settings: When you choose D10 for your MXF profile, you must also set this value to D10.If the service returns an enum value that is not available in the current SDK version,
syntax
will returnMpeg2Syntax.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromsyntaxAsString()
.- Returns:
- Specify whether this output's video uses the D10 syntax. Keep the default value to not use the syntax. Related settings: When you choose D10 for your MXF profile, you must also set this value to D10.
- See Also:
-
syntaxAsString
Specify whether this output's video uses the D10 syntax. Keep the default value to not use the syntax. Related settings: When you choose D10 for your MXF profile, you must also set this value to D10.If the service returns an enum value that is not available in the current SDK version,
syntax
will returnMpeg2Syntax.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromsyntaxAsString()
.- Returns:
- Specify whether this output's video uses the D10 syntax. Keep the default value to not use the syntax. Related settings: When you choose D10 for your MXF profile, you must also set this value to D10.
- See Also:
-
telecine
When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard or soft telecine to create a smoother picture. Hard telecine produces a 29.97i output. Soft telecine produces an output with a 23.976 output that signals to the video player device to do the conversion during play back. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.If the service returns an enum value that is not available in the current SDK version,
telecine
will returnMpeg2Telecine.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromtelecineAsString()
.- Returns:
- When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard or soft telecine to create a smoother picture. Hard telecine produces a 29.97i output. Soft telecine produces an output with a 23.976 output that signals to the video player device to do the conversion during play back. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.
- See Also:
-
telecineAsString
When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard or soft telecine to create a smoother picture. Hard telecine produces a 29.97i output. Soft telecine produces an output with a 23.976 output that signals to the video player device to do the conversion during play back. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.If the service returns an enum value that is not available in the current SDK version,
telecine
will returnMpeg2Telecine.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromtelecineAsString()
.- Returns:
- When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard or soft telecine to create a smoother picture. Hard telecine produces a 29.97i output. Soft telecine produces an output with a 23.976 output that signals to the video player device to do the conversion during play back. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.
- See Also:
-
temporalAdaptiveQuantization
Keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization.If the service returns an enum value that is not available in the current SDK version,
temporalAdaptiveQuantization
will returnMpeg2TemporalAdaptiveQuantization.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromtemporalAdaptiveQuantizationAsString()
.- Returns:
- Keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization.
- See Also:
-
temporalAdaptiveQuantizationAsString
Keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization.If the service returns an enum value that is not available in the current SDK version,
temporalAdaptiveQuantization
will returnMpeg2TemporalAdaptiveQuantization.UNKNOWN_TO_SDK_VERSION
. The raw value returned by the service is available fromtemporalAdaptiveQuantizationAsString()
.- Returns:
- Keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization.
- See Also:
-
toBuilder
Description copied from interface:ToCopyableBuilder
Take this object and create a builder that contains all of the current property values of this object.- Specified by:
toBuilder
in interfaceToCopyableBuilder<Mpeg2Settings.Builder,
Mpeg2Settings> - Returns:
- a builder for type T
-
builder
-
serializableBuilderClass
-
hashCode
public final int hashCode() -
equals
-
equalsBySdkFields
Description copied from interface:SdkPojo
Indicates whether some other object is "equal to" this one by SDK fields. An SDK field is a modeled, non-inherited field in anSdkPojo
class, and is generated based on a service model.If an
SdkPojo
class does not have any inherited fields,equalsBySdkFields
andequals
are essentially the same.- Specified by:
equalsBySdkFields
in interfaceSdkPojo
- Parameters:
obj
- the object to be compared with- Returns:
- true if the other object equals to this object by sdk fields, false otherwise.
-
toString
Returns a string representation of this object. This is useful for testing and debugging. Sensitive data will be redacted from this string using a placeholder value. -
getValueForField
-
sdkFields
-