Skip to end of metadata
Go to start of metadata

You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 4 Next »

In the Advanced tab you can configure additional properties for optimizing the performance of the Parquet. Depending on selected ....



SettingDescription

Compression Codec

The compression algorithm used to compress pages when encoding. Valid choices are Uncompressed, Snappy, Gzip, Lzo, Brotli, Lz4, and Zstd.

Block Size

The Block Size is the size of a row group buffered in memory. Block size limits the memory usage when writing. Larger values will improve I/O when reading but consume more memory when writing. Default Block Size is set to 134217728.

Page SizeThe page is the smallest unit that must be read fully to access a single record. When reading, each page can be decompressed independently. If this value is too small, the compression will deteriorate. Default is 1048576.
Dictionary Page SizeThere is one dictionary page per column per row group when dictionary encoding is used. Dictionary page size works like the page size but for dictionary encoding. Default is 1048576
Enable DictionaryA boolean that indicates whether to include the dictionary compression strategy in the generated parquet document. Enabling dictionary allows for building a dictionary of values encountered in columns.
ValidatingBoolean that enables schema validation when writing.
Writer VersionParquet format version to use when writing. 1.0 ensures compatibility with older readers. Default is v1. The option is to set v2
  • No labels