Package org.apache.iceberg.parquet
Class Parquet.ReadBuilder
java.lang.Object
org.apache.iceberg.parquet.Parquet.ReadBuilder
- All Implemented Interfaces:
InternalData.ReadBuilder
- Enclosing class:
Parquet
-
Method Summary
Modifier and TypeMethodDescription<D> CloseableIterable<D> build()Build the configured reader.callInit()Deprecated.caseSensitive(boolean newCaseSensitive) createBatchedReaderFunc(Function<org.apache.parquet.schema.MessageType, VectorizedReader<?>> func) createReaderFunc(BiFunction<Schema, org.apache.parquet.schema.MessageType, ParquetValueReader<?>> newReaderFunction) createReaderFunc(Function<org.apache.parquet.schema.MessageType, ParquetValueReader<?>> newReaderFunction) filter(Expression newFilter) filterRecords(boolean newFilterRecords) Set the projection schema.readSupport(org.apache.parquet.hadoop.api.ReadSupport<?> newFilterSupport) Deprecated.will be removed in 2.0.0; usecreateReaderFunc(Function)insteadrecordsPerBatch(int numRowsPerBatch) Reuse container classes, like structs, lists, and maps.setCustomType(int fieldId, Class<? extends StructLike> structClass) Set a custom class for in-memory objects at the given field ID.setRootType(Class<? extends StructLike> rootClass) Set a custom class for in-memory objects at the schema root.split(long newStart, long newLength) Restricts the read to the given range: [start, start + length).withAADPrefix(ByteBuffer aadPrefix) withFileEncryptionKey(ByteBuffer encryptionKey) withNameMapping(NameMapping newNameMapping)
-
Method Details
-
split
Restricts the read to the given range: [start, start + length).- Specified by:
splitin interfaceInternalData.ReadBuilder- Parameters:
newStart- the start position for this readnewLength- the length of the range this read should scan- Returns:
- this builder for method chaining
-
project
Description copied from interface:InternalData.ReadBuilderSet the projection schema.- Specified by:
projectin interfaceInternalData.ReadBuilder
-
caseInsensitive
-
caseSensitive
-
filterRecords
-
filter
-
readSupport
@Deprecated public Parquet.ReadBuilder readSupport(org.apache.parquet.hadoop.api.ReadSupport<?> newFilterSupport) Deprecated.will be removed in 2.0.0; usecreateReaderFunc(Function)instead -
createReaderFunc
public Parquet.ReadBuilder createReaderFunc(Function<org.apache.parquet.schema.MessageType, ParquetValueReader<?>> newReaderFunction) -
createReaderFunc
public Parquet.ReadBuilder createReaderFunc(BiFunction<Schema, org.apache.parquet.schema.MessageType, ParquetValueReader<?>> newReaderFunction) -
createBatchedReaderFunc
public Parquet.ReadBuilder createBatchedReaderFunc(Function<org.apache.parquet.schema.MessageType, VectorizedReader<?>> func) -
set
-
callInit
Deprecated.will be removed in 2.0.0; usecreateReaderFunc(Function)instead -
reuseContainers
Description copied from interface:InternalData.ReadBuilderReuse container classes, like structs, lists, and maps.- Specified by:
reuseContainersin interfaceInternalData.ReadBuilder
-
recordsPerBatch
-
withNameMapping
-
setRootType
Description copied from interface:InternalData.ReadBuilderSet a custom class for in-memory objects at the schema root.- Specified by:
setRootTypein interfaceInternalData.ReadBuilder
-
setCustomType
Description copied from interface:InternalData.ReadBuilderSet a custom class for in-memory objects at the given field ID.- Specified by:
setCustomTypein interfaceInternalData.ReadBuilder
-
withFileEncryptionKey
-
withAADPrefix
-
build
Description copied from interface:InternalData.ReadBuilderBuild the configured reader.- Specified by:
buildin interfaceInternalData.ReadBuilder
-
createReaderFunc(Function)instead