Class CfnDeliveryStream.ParquetSerDeProperty
A serializer to use for converting data to the Parquet format before storing it in Amazon S3.
Inheritance
Implements
Namespace: Amazon.CDK.AWS.KinesisFirehose
Assembly: Amazon.CDK.AWS.KinesisFirehose.dll
Syntax (csharp)
public class ParquetSerDeProperty : Object, CfnDeliveryStream.IParquetSerDeProperty
Syntax (vb)
Public Class ParquetSerDeProperty
Inherits Object
Implements CfnDeliveryStream.IParquetSerDeProperty
Remarks
For more information, see Apache Parquet .
ExampleMetadata: fixture=_generated
Examples
// The code below shows an example of how to instantiate this type.
// The values are placeholders you should change.
using Amazon.CDK.AWS.KinesisFirehose;
var parquetSerDeProperty = new ParquetSerDeProperty {
BlockSizeBytes = 123,
Compression = "compression",
EnableDictionaryCompression = false,
MaxPaddingBytes = 123,
PageSizeBytes = 123,
WriterVersion = "writerVersion"
};
Synopsis
Constructors
ParquetSerDeProperty() |
Properties
BlockSizeBytes | The Hadoop Distributed File System (HDFS) block size. |
Compression | The compression code to use over data blocks. |
EnableDictionaryCompression | Indicates whether to enable dictionary compression. |
MaxPaddingBytes | The maximum amount of padding to apply. |
PageSizeBytes | The Parquet page size. |
WriterVersion | Indicates the version of row format to output. |
Constructors
ParquetSerDeProperty()
public ParquetSerDeProperty()
Properties
BlockSizeBytes
The Hadoop Distributed File System (HDFS) block size.
public Nullable<double> BlockSizeBytes { get; set; }
Property Value
System.Nullable<System.Double>
Remarks
This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the minimum is 64 MiB. Kinesis Data Firehose uses this value for padding calculations.
Compression
The compression code to use over data blocks.
public string Compression { get; set; }
Property Value
System.String
Remarks
The possible values are UNCOMPRESSED
, SNAPPY
, and GZIP
, with the default being SNAPPY
. Use SNAPPY
for higher decompression speed. Use GZIP
if the compression ratio is more important than speed.
EnableDictionaryCompression
Indicates whether to enable dictionary compression.
public object EnableDictionaryCompression { get; set; }
Property Value
System.Object
Remarks
MaxPaddingBytes
The maximum amount of padding to apply.
public Nullable<double> MaxPaddingBytes { get; set; }
Property Value
System.Nullable<System.Double>
Remarks
This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 0.
PageSizeBytes
The Parquet page size.
public Nullable<double> PageSizeBytes { get; set; }
Property Value
System.Nullable<System.Double>
Remarks
Column chunks are divided into pages. A page is conceptually an indivisible unit (in terms of compression and encoding). The minimum value is 64 KiB and the default is 1 MiB.
WriterVersion
Indicates the version of row format to output.
public string WriterVersion { get; set; }
Property Value
System.String
Remarks
The possible values are V1
and V2
. The default is V1
.