All Classes Interface Summary Class Summary Enum Summary Exception Summary
Class |
Description |
Accessor<T> |
|
Accessors |
|
Action<R> |
An action performed on a table.
|
Actions |
|
AllDataFilesTable |
A Table implementation that exposes a table's valid data files as rows.
|
AllDataFilesTable.AllDataFilesTableScan |
|
AllEntriesTable |
A Table implementation that exposes a table's manifest entries as rows, for both delete and data files.
|
AllManifestsTable |
A Table implementation that exposes a table's valid manifest files as rows.
|
AllManifestsTable.AllManifestsTableScan |
|
AlreadyExistsException |
Exception raised when attempting to create a table that already exists.
|
And |
|
AppendFiles |
API for appending new files in a table.
|
ArrowAllocation |
|
ArrowSchemaUtil |
|
ArrowVectorAccessor |
|
ArrowVectorAccessors |
|
Avro |
|
Avro.ReadBuilder |
|
Avro.WriteBuilder |
|
AvroIterable<D> |
|
AvroSchemaUtil |
|
AvroSchemaVisitor<T> |
|
AvroSchemaWithTypeVisitor<T> |
|
AvroWithSparkSchemaVisitor<T> |
|
BaseColumnIterator |
|
BaseCombinedScanTask |
|
BaseMetastoreCatalog |
|
BaseMetastoreTableOperations |
|
BaseOverwriteFiles |
|
BasePageIterator |
|
BasePageIterator.IntIterator |
|
BaseParquetReaders<T> |
|
BaseParquetWriter<T> |
|
BaseReplacePartitions |
|
BaseRewriteManifests |
|
BaseRowDelta |
|
BaseTable |
Base Table implementation.
|
BaseVectorizedParquetValuesReader |
A values reader for Parquet's run-length encoded data that reads column data in batches instead of one value at a
time.
|
BinaryUtil |
|
Binder |
Rewrites expressions by replacing unbound named references with references to
fields in a struct schema.
|
BinPacking |
|
BinPacking.ListPacker<T> |
|
BinPacking.PackingIterable<T> |
|
Bound<T> |
Represents a bound value expression.
|
BoundLiteralPredicate<T> |
|
BoundPredicate<T> |
|
BoundReference<T> |
|
BoundSetPredicate<T> |
|
BoundTerm<T> |
Represents a bound term.
|
BoundTransform<S,T> |
A transform expression.
|
BoundUnaryPredicate<T> |
|
ByteBuffers |
|
CachingCatalog |
|
Catalog |
A Catalog API for table create, drop, and load operations.
|
CharSequenceSet |
|
CharSequenceWrapper |
Wrapper class to adapt CharSequence for use in maps and sets.
|
CheckCompatibility |
|
CherrypickAncestorCommitException |
This exception occurs when one cherrypicks an ancestor or when the picked snapshot is already linked to
a published ancestor.
|
ClientPool<C,E extends java.lang.Exception> |
|
ClientPool.Action<R,C,E extends java.lang.Exception> |
|
CloseableGroup |
|
CloseableIterable<T> |
|
CloseableIterable.ConcatCloseableIterable<E> |
|
CloseableIterator<T> |
|
ColumnarBatchReader |
VectorizedReader that returns Spark's ColumnarBatch to support Spark's vectorized read path.
|
ColumnIterator<T> |
|
ColumnWriter<T> |
|
CombinedScanTask |
A scan task made of several ranges from files.
|
CommitFailedException |
Exception raised when a commit fails because of out of date metadata.
|
Comparators |
|
ContentFile<F> |
|
Conversions |
|
CreateSnapshotEvent |
|
DataFile |
Interface for data files listed in a table manifest.
|
DataFiles |
|
DataFiles.Builder |
|
DataFilesTable |
A Table implementation that exposes a table's data files as rows.
|
DataFilesTable.FilesTableScan |
|
DataOperations |
Data operations that produce snapshots.
|
DataReader<T> |
|
DataTableScan |
|
DataTask |
A task that returns data as rows instead of where to read data.
|
DataWriter<T> |
|
DateTimeUtil |
|
DelegatingInputStream |
|
DelegatingOutputStream |
|
DeleteFile |
Interface for delete files listed in a table delete manifest.
|
DeleteFiles |
API for deleting files from a table.
|
DuplicateWAPCommitException |
This exception occurs when the WAP workflow detects a duplicate wap commit.
|
DynClasses |
|
DynClasses.Builder |
|
DynConstructors |
Copied from parquet-common
|
DynConstructors.Builder |
|
DynConstructors.Ctor<C> |
|
DynFields |
|
DynFields.BoundField<T> |
|
DynFields.Builder |
|
DynFields.StaticField<T> |
|
DynFields.UnboundField<T> |
Convenience wrapper class around Field .
|
DynMethods |
Copied from parquet-common
|
DynMethods.BoundMethod |
|
DynMethods.Builder |
|
DynMethods.StaticMethod |
|
DynMethods.UnboundMethod |
Convenience wrapper class around Method .
|
EncryptedFiles |
|
EncryptedInputFile |
Thin wrapper around an InputFile instance that is encrypted.
|
EncryptedOutputFile |
Thin wrapper around a OutputFile that is encrypting bytes written to the underlying
file system, via an encryption key that is symbolized by the enclosed
EncryptionKeyMetadata .
|
EncryptionKeyMetadata |
Light typedef over a ByteBuffer that indicates that the given bytes represent metadata about
an encrypted data file's encryption key.
|
EncryptionKeyMetadatas |
|
EncryptionManager |
Module for encrypting and decrypting table data files.
|
Evaluator |
|
Exceptions |
|
ExceptionUtil |
|
ExpireSnapshots |
|
Expression |
Represents a boolean expression tree.
|
Expression.Operation |
|
Expressions |
|
ExpressionVisitors |
|
ExpressionVisitors.BoundExpressionVisitor<R> |
Deprecated.
|
ExpressionVisitors.BoundVisitor<R> |
|
ExpressionVisitors.ExpressionVisitor<R> |
|
False |
|
FileAppender<D> |
|
FileContent |
Content type stored in a file, one of DATA, POSITION_DELETES, or EQUALITY_DELETES.
|
FileFormat |
Enum of supported file formats.
|
FileIO |
Pluggable module for reading, writing, and deleting files.
|
Files |
|
FileScanTask |
A scan task over a range of a single file.
|
FindFiles |
|
FindFiles.Builder |
|
FlinkAvroReader |
|
FlinkAvroWriter |
|
FlinkParquetReaders |
|
FlinkParquetWriters |
|
FlinkSchemaUtil |
|
FlinkTypeToType |
|
FlinkTypeVisitor<T> |
|
GenericManifestFile |
|
GenericManifestFile.CopyBuilder |
|
GenericOrcReader |
|
GenericOrcReaders |
|
GenericOrcWriter |
|
GenericParquetReaders |
|
GenericParquetWriter |
|
GenericPartitionFieldSummary |
|
GenericRecord |
|
GuavaClasses |
|
HadoopCatalog |
HadoopCatalog provides a way to use table names like db.table to work with path-based tables under a common
location.
|
HadoopFileIO |
|
HadoopInputFile |
InputFile implementation using the Hadoop FileSystem API.
|
HadoopOutputFile |
OutputFile implementation using the Hadoop FileSystem API.
|
HadoopTableOperations |
TableOperations implementation for file systems that support atomic rename.
|
HadoopTables |
Implementation of Iceberg tables that uses the Hadoop FileSystem
to store metadata and manifests.
|
HasTableOperations |
Used to expose a table's TableOperations.
|
HiddenPathFilter |
A PathFilter that filters out hidden paths.
|
HistoryEntry |
Table history entry.
|
HistoryTable |
A Table implementation that exposes a table's history as rows.
|
HiveCatalog |
|
HiveCatalogs |
|
HiveClientPool |
|
HiveTableOperations |
TODO we should be able to extract some more commonalities to BaseMetastoreTableOperations to
avoid code duplication between this class and Metacat Tables.
|
HiveTypeConverter |
|
IcebergArrowColumnVector |
Implementation of Spark's ColumnVector interface.
|
IcebergArrowVectors |
The general way of getting a value at an index in the Arrow vector
is by calling get(index).
|
IcebergArrowVectors.DecimalArrowVector |
Extension of Arrow's @DecimalVector .
|
IcebergArrowVectors.VarcharArrowVector |
Extension of Arrow's @VarCharVector .
|
IcebergBinaryObjectInspector |
|
IcebergDateObjectInspector |
|
IcebergDecimalObjectInspector |
|
IcebergDecoder<D> |
|
IcebergEncoder<D> |
|
IcebergGenerics |
|
IcebergGenerics.ScanBuilder |
|
IcebergInputFormat<T> |
Generic Mrv2 InputFormat API for Iceberg.
|
IcebergObjectInspector |
|
IcebergPigInputFormat<T> |
|
IcebergRecordObjectInspector |
|
IcebergSerDe |
|
IcebergSource |
|
IcebergStorage |
|
IcebergTimestampObjectInspector |
|
IcebergWritable |
Wraps an Iceberg Record in a Writable which Hive can use in the SerDe.
|
IdentityPartitionConverters |
|
InclusiveMetricsEvaluator |
|
IndexByName |
|
IndexParents |
|
InputFile |
|
InputFormatConfig |
|
InputFormatConfig.ConfigBuilder |
|
InputFormatConfig.InMemoryDataModel |
|
JsonUtil |
|
Listener<E> |
A listener interface that can receive notifications.
|
Listeners |
Static registration and notification for listeners.
|
Literal<T> |
Represents a literal fixed value in an expression predicate
|
LocationProvider |
Interface for providing data file locations to write tasks.
|
LocationProviders |
|
LogicalMap |
|
ManageSnapshots |
API for managing snapshots.
|
ManifestContent |
Content type stored in a manifest file, either DATA or DELETES.
|
ManifestEntriesTable |
A Table implementation that exposes a table's manifest entries as rows, for both delete and data files.
|
ManifestEvaluator |
|
ManifestFile |
Represents a manifest file that can be scanned to find data files in a table.
|
ManifestFile.PartitionFieldSummary |
Summarizes the values of one partition field stored in a manifest file.
|
ManifestFiles |
|
ManifestFileUtil |
|
ManifestReader<F extends ContentFile<F>> |
Base reader for data and delete manifest files.
|
ManifestReader.FileType |
|
ManifestsTable |
A Table implementation that exposes a table's manifest files as rows.
|
ManifestWriter<F extends ContentFile<F>> |
Writer for manifest files.
|
MappedField |
An immutable mapping between a field ID and a set of names.
|
MappedFields |
|
MappingUtil |
|
MetadataTableType |
|
Metrics |
Iceberg file format metrics.
|
MetricsConfig |
|
MetricsModes |
This class defines different metrics modes, which allow users to control the collection of
value_counts, null_value_counts, lower_bounds, upper_bounds for different columns in metadata.
|
MetricsModes.Counts |
Under this mode, only value_counts, null_value_counts are persisted.
|
MetricsModes.Full |
Under this mode, value_counts, null_value_counts and full lower_bounds, upper_bounds are persisted.
|
MetricsModes.MetricsMode |
|
MetricsModes.None |
Under this mode, value_counts, null_value_counts, lower_bounds, upper_bounds are not persisted.
|
MetricsModes.Truncate |
Under this mode, value_counts, null_value_counts and truncated lower_bounds, upper_bounds are persisted.
|
NamedReference<T> |
|
NameMapping |
Represents a mapping from external schema names to Iceberg type IDs.
|
NameMappingParser |
Parses external name mappings from a JSON representation.
|
Namespace |
|
NamespaceNotEmptyException |
Exception raised when attempting to drop a namespace that is not empty.
|
NoSuchIcebergTableException |
NoSuchTableException thrown when a table is found but it is not an Iceberg table.
|
NoSuchNamespaceException |
Exception raised when attempting to load a namespace that does not exist.
|
NoSuchTableException |
Exception raised when attempting to load a table that does not exist.
|
Not |
|
NotFoundException |
Exception raised when attempting to read a file that does not exist.
|
NullabilityHolder |
Instances of this class simply track whether a value at an index is null.
|
NullValuesColumnVector |
|
Or |
|
ORC |
|
ORC.ReadBuilder |
|
ORC.WriteBuilder |
|
OrcMetrics |
|
OrcRowReader<T> |
Used for implementing ORC row readers.
|
ORCSchemaUtil |
Utilities for mapping Iceberg to ORC schemas.
|
ORCSchemaUtil.BinaryType |
|
ORCSchemaUtil.LongType |
|
OrcSchemaVisitor<T> |
Generic visitor of an ORC Schema.
|
OrcSchemaWithTypeVisitor<T> |
|
OrcValueReader<T> |
|
OrcValueReaders |
|
OrcValueReaders.StructReader<T> |
|
OrcValueWriter<T> |
Write data value of a schema.
|
OutputFile |
|
OverwriteFiles |
API for overwriting files in a table.
|
Pair<X,Y> |
|
ParallelIterable<T> |
|
Parquet |
|
Parquet.ReadBuilder |
|
Parquet.WriteBuilder |
|
ParquetAvroReader |
|
ParquetAvroValueReaders |
|
ParquetAvroValueReaders.TimeMillisReader |
|
ParquetAvroValueReaders.TimestampMillisReader |
|
ParquetAvroWriter |
|
ParquetDictionaryRowGroupFilter |
|
ParquetIterable<T> |
|
ParquetMetricsRowGroupFilter |
|
ParquetReader<T> |
|
ParquetSchemaUtil |
|
ParquetSchemaUtil.HasIds |
|
ParquetTypeVisitor<T> |
|
ParquetUtil |
|
ParquetValueReader<T> |
|
ParquetValueReaders |
|
ParquetValueReaders.BinaryAsDecimalReader |
|
ParquetValueReaders.BytesReader |
|
ParquetValueReaders.FloatAsDoubleReader |
|
ParquetValueReaders.IntAsLongReader |
|
ParquetValueReaders.IntegerAsDecimalReader |
|
ParquetValueReaders.ListReader<E> |
|
ParquetValueReaders.LongAsDecimalReader |
|
ParquetValueReaders.MapReader<K,V> |
|
ParquetValueReaders.PrimitiveReader<T> |
|
ParquetValueReaders.RepeatedKeyValueReader<M,I,K,V> |
|
ParquetValueReaders.RepeatedReader<T,I,E> |
|
ParquetValueReaders.ReusableEntry<K,V> |
|
ParquetValueReaders.StringReader |
|
ParquetValueReaders.StructReader<T,I> |
|
ParquetValueReaders.UnboxedReader<T> |
|
ParquetValueWriter<T> |
|
ParquetValueWriters |
|
ParquetValueWriters.PrimitiveWriter<T> |
|
ParquetValueWriters.RepeatedKeyValueWriter<M,K,V> |
|
ParquetValueWriters.RepeatedWriter<L,E> |
|
ParquetValueWriters.StructWriter<S> |
|
ParquetWithSparkSchemaVisitor<T> |
Visitor for traversing a Parquet type with a companion Spark type.
|
ParquetWriteAdapter<D> |
|
PartitionField |
|
PartitionSpec |
Represents how to produce partition data for a table.
|
PartitionSpec.Builder |
|
PartitionSpecParser |
|
PartitionSpecVisitor<T> |
|
PartitionsTable |
A Table implementation that exposes a table's partitions as rows.
|
PartitionUtil |
|
PendingUpdate<T> |
API for table metadata changes.
|
PigParquetReader |
|
PlaintextEncryptionManager |
|
PositionOutputStream |
|
Predicate<T,C extends Term> |
|
ProjectionDatumReader<D> |
|
Projections |
Utils to project expressions on rows to expressions on partitions.
|
Projections.ProjectionEvaluator |
A class that projects expressions for a table's data rows into expressions on the table's
partition values, for a table's partition spec .
|
PropertyUtil |
|
PruneColumnsWithoutReordering |
|
PruneColumnsWithReordering |
|
Record |
|
Reference<T> |
|
RemoveIds |
|
RemoveIds |
|
RemoveOrphanFilesAction |
An action that removes orphan metadata and data files by listing a given location and comparing
the actual files in that location with data and metadata files referenced by all valid snapshots.
|
ReplacePartitions |
Not recommended: API for overwriting files in a table by partition.
|
ResidualEvaluator |
|
RewriteDataFilesAction |
|
RewriteDataFilesActionResult |
|
RewriteFiles |
API for replacing files in a table.
|
RewriteManifests |
API for rewriting manifests for a table.
|
RewriteManifestsAction |
An action that rewrites manifests in a distributed manner and co-locates metadata for partitions.
|
RewriteManifestsActionResult |
|
Rollback |
API for rolling table data back to the state at an older table snapshot .
|
RollbackStagedTable |
An implementation of StagedTable that mimics the behavior of Spark's non-atomic CTAS and RTAS.
|
RowDataRewriter |
|
RowDelta |
API for encoding row-level changes to a table.
|
RuntimeIOException |
Deprecated.
|
RuntimeMetaException |
Exception used to wrap MetaException as a RuntimeException and add context.
|
ScanEvent |
Event sent to listeners when a table scan is planned.
|
ScanSummary |
|
ScanSummary.Builder |
|
ScanSummary.PartitionMetrics |
|
ScanTask |
A scan task.
|
Schema |
The schema of a data table.
|
SchemaParser |
|
SchemaUtil |
|
SeekableInputStream |
SeekableInputStream is an interface with the methods needed to read data from a file or
Hadoop data stream.
|
SerializableConfiguration |
Wraps a Configuration object in a Serializable layer.
|
SerializableSupplier<T> |
|
SerializationUtil |
|
SetLocation |
|
Snapshot |
A snapshot of the data in a table at a point in time.
|
SnapshotManager |
|
SnapshotParser |
|
SnapshotsTable |
A Table implementation that exposes a table's known snapshots as rows.
|
SnapshotSummary |
|
SnapshotSummary.Builder |
|
SnapshotUpdate<ThisT> |
API for table changes that produce snapshots.
|
SnapshotUpdateAction<ThisT,R> |
|
SnapshotUtil |
|
Spark3Util |
|
Spark3Util.DescribeSchemaVisitor |
|
SparkAvroReader |
|
SparkAvroWriter |
|
SparkCatalog |
A Spark TableCatalog implementation that wraps an Iceberg Catalog .
|
SparkDataFile |
|
SparkExceptionUtil |
|
SparkFilters |
|
SparkOrcReader |
Converts the OrcIterator, which returns ORC's VectorizedRowBatch to a
set of Spark's UnsafeRows.
|
SparkOrcWriter |
This class acts as an adaptor from an OrcFileAppender to a
FileAppender<InternalRow>.
|
SparkParquetReaders |
|
SparkParquetWriters |
|
SparkScanBuilder |
|
SparkSchemaUtil |
Helper methods for working with Spark/Hive metadata.
|
SparkSessionCatalog<T extends org.apache.spark.sql.connector.catalog.TableCatalog & org.apache.spark.sql.connector.catalog.SupportsNamespaces> |
A Spark catalog that can also load non-Iceberg tables.
|
SparkStreamingWrite |
|
SparkStructLike |
|
SparkTable |
|
SparkTableUtil |
Java version of the original SparkTableUtil.scala
https://github.com/apache/iceberg/blob/apache-iceberg-0.8.0-incubating/spark/src/main/scala/org/apache/iceberg/spark/SparkTableUtil.scala
|
SparkTableUtil.SparkPartition |
Class representing a table partition.
|
SparkUtil |
|
SparkValueConverter |
A utility class that converts Spark values to Iceberg's internal representation.
|
SparkValueReaders |
|
SparkValueWriters |
|
StagedSparkTable |
|
StrictMetricsEvaluator |
|
StructLike |
Interface for accessing data by position in a schema.
|
StructLikeWrapper |
Wrapper to adapt StructLike for use in maps and sets by implementing equals and hashCode.
|
SupportsNamespaces |
Catalog methods for working with namespaces.
|
SystemProperties |
Configuration properties that are controlled by Java system properties.
|
Table |
Represents a table.
|
TableIdentifier |
Identifies a table in iceberg catalog.
|
TableMetadata |
Metadata for a table.
|
TableMetadata.MetadataLogEntry |
|
TableMetadata.SnapshotLogEntry |
|
TableMetadataParser |
|
TableMetadataParser.Codec |
|
TableOperations |
SPI interface to abstract table metadata access and updates.
|
TableProperties |
|
Tables |
Generic interface for creating and loading a table implementation.
|
TableScan |
API for configuring a table scan.
|
TableScanUtil |
|
Tasks |
|
Tasks.Builder<I> |
|
Tasks.FailureTask<I,E extends java.lang.Exception> |
|
Tasks.Task<I,E extends java.lang.Exception> |
|
Tasks.UnrecoverableException |
|
Term |
An expression that evaluates to a value.
|
ThreadPools |
|
Transaction |
A transaction for performing multiple updates to a table.
|
Transactions |
|
Transform<S,T> |
A transform function used for partitioning.
|
Transforms |
Factory methods for transforms.
|
TripleWriter<T> |
|
True |
|
Type |
|
Type.NestedType |
|
Type.PrimitiveType |
|
Type.TypeID |
|
Types |
|
Types.BinaryType |
|
Types.BooleanType |
|
Types.DateType |
|
Types.DecimalType |
|
Types.DoubleType |
|
Types.FixedType |
|
Types.FloatType |
|
Types.IntegerType |
|
Types.ListType |
|
Types.LongType |
|
Types.MapType |
|
Types.NestedField |
|
Types.StringType |
|
Types.StructType |
|
Types.TimestampType |
|
Types.TimeType |
|
Types.UUIDType |
|
TypeToMessageType |
|
TypeUtil |
|
TypeUtil.CustomOrderSchemaVisitor<T> |
|
TypeUtil.NextID |
Interface for passing a function that assigns column IDs.
|
TypeUtil.SchemaVisitor<T> |
|
TypeWithSchemaVisitor<T> |
Visitor for traversing a Parquet type with a companion Iceberg type.
|
Unbound<T,B> |
Represents an unbound expression node.
|
UnboundPredicate<T> |
|
UnboundTerm<T> |
Represents an unbound term.
|
UnboundTransform<S,T> |
|
UnicodeUtil |
|
UnknownTransform<S,T> |
|
UpdateLocation |
API for setting a table's base location.
|
UpdateProperties |
API for updating table properties.
|
UpdateSchema |
API for schema evolution.
|
Util |
|
UUIDConversion |
|
ValidationException |
Exception raised when validation checks fail.
|
ValueReader<T> |
|
ValueReaders |
|
ValueReaders.StructReader<S> |
|
ValuesAsBytesReader |
Implements a ValuesReader specifically to read given number of bytes from the underlying ByteBufferInputStream .
|
ValueWriter<D> |
|
ValueWriters |
|
ValueWriters.StructWriter<S> |
|
VectorHolder |
Container class for holding the Arrow vector storing a batch of values along with other state needed for reading
values out of it.
|
VectorizedArrowReader |
|
VectorizedColumnIterator |
Vectorized version of the ColumnIterator that reads column values in data pages of a column in a row group in a
batched fashion.
|
VectorizedDictionaryEncodedParquetValuesReader |
This decoder reads Parquet dictionary encoded data in a vectorized fashion.
|
VectorizedPageIterator |
|
VectorizedParquetDefinitionLevelReader |
|
VectorizedParquetReader<T> |
|
VectorizedReader<T> |
Interface for vectorized Iceberg readers.
|
VectorizedRowBatchIterator |
An adaptor so that the ORC RecordReader can be used as an Iterator.
|
VectorizedSparkParquetReaders |
|
WapUtil |
|