Class CfnModelPackage.TransformInputProperty
Describes the input source of a transform job and the way the transform job consumes it.
Inheritance
Implements
Namespace: Amazon.CDK.AWS.Sagemaker
Assembly: Amazon.CDK.AWS.Sagemaker.dll
Syntax (csharp)
public class TransformInputProperty : Object, CfnModelPackage.ITransformInputProperty
Syntax (vb)
Public Class TransformInputProperty
Inherits Object
Implements CfnModelPackage.ITransformInputProperty
Remarks
ExampleMetadata: fixture=_generated
Examples
// The code below shows an example of how to instantiate this type.
// The values are placeholders you should change.
using Amazon.CDK.AWS.Sagemaker;
var transformInputProperty = new TransformInputProperty {
DataSource = new DataSourceProperty {
S3DataSource = new S3DataSourceProperty {
S3DataType = "s3DataType",
S3Uri = "s3Uri"
}
},
// the properties below are optional
CompressionType = "compressionType",
ContentType = "contentType",
SplitType = "splitType"
};
Synopsis
Constructors
TransformInputProperty() |
Properties
CompressionType | If your transform data is compressed, specify the compression type. |
ContentType | The multipurpose internet mail extension (MIME) type of the data. |
DataSource | Describes the location of the channel data, which is, the S3 location of the input data that the model can consume. |
SplitType | The method to use to split the transform job's data files into smaller batches. |
Constructors
TransformInputProperty()
public TransformInputProperty()
Properties
CompressionType
If your transform data is compressed, specify the compression type.
public string CompressionType { get; set; }
Property Value
System.String
Remarks
Amazon SageMaker automatically decompresses the data for the transform job accordingly. The default value is None
.
ContentType
The multipurpose internet mail extension (MIME) type of the data.
public string ContentType { get; set; }
Property Value
System.String
Remarks
Amazon SageMaker uses the MIME type with each http call to transfer data to the transform job.
DataSource
Describes the location of the channel data, which is, the S3 location of the input data that the model can consume.
public object DataSource { get; set; }
Property Value
System.Object
Remarks
SplitType
The method to use to split the transform job's data files into smaller batches.
public string SplitType { get; set; }
Property Value
System.String
Remarks
Splitting is necessary when the total size of each object is too large to fit in a single request. You can also use data splitting to improve performance by processing multiple concurrent mini-batches. The default value for SplitType
is None
, which indicates that input data files are not split, and request payloads contain the entire contents of an input object. Set the value of this parameter to Line
to split records on a newline character boundary. SplitType
also supports a number of record-oriented binary data formats. Currently, the supported record formats are:
When splitting is enabled, the size of a mini-batch depends on the values of the BatchStrategy
and MaxPayloadInMB
parameters. When the value of BatchStrategy
is MultiRecord
, Amazon SageMaker sends the maximum number of records in each request, up to the MaxPayloadInMB
limit. If the value of BatchStrategy
is SingleRecord
, Amazon SageMaker sends individual records in each request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting is applied to a binary data format, padding is removed if the value of <code>BatchStrategy</code> is set to <code>SingleRecord</code> . Padding is not removed if the value of <code>BatchStrategy</code> is set to <code>MultiRecord</code> .
For more information about RecordIO
, see Create a Dataset Using RecordIO in the MXNet documentation. For more information about TFRecord
, see Consuming TFRecord data in the TensorFlow documentation.