All Classes Interface Summary Class Summary Enum Summary Exception Summary
| Class |
Description |
| Accessor<T> |
|
| Accessors |
|
| Action<R> |
An action performed on a table.
|
| Actions |
|
| AllDataFilesTable |
A Table implementation that exposes a table's valid data files as rows.
|
| AllDataFilesTable.AllDataFilesTableScan |
|
| AllEntriesTable |
A Table implementation that exposes a table's manifest entries as rows, for both delete and data files.
|
| AllManifestsTable |
A Table implementation that exposes a table's valid manifest files as rows.
|
| AllManifestsTable.AllManifestsTableScan |
|
| AlreadyExistsException |
Exception raised when attempting to create a table that already exists.
|
| And |
|
| AppendFiles |
API for appending new files in a table.
|
| ArrowAllocation |
|
| ArrowSchemaUtil |
|
| ArrowVectorAccessor |
|
| ArrowVectorAccessors |
|
| Avro |
|
| Avro.ReadBuilder |
|
| Avro.WriteBuilder |
|
| AvroIterable<D> |
|
| AvroSchemaUtil |
|
| AvroSchemaVisitor<T> |
|
| AvroSchemaWithTypeVisitor<T> |
|
| AvroWithSparkSchemaVisitor<T> |
|
| BaseColumnIterator |
|
| BaseCombinedScanTask |
|
| BaseMetastoreCatalog |
|
| BaseMetastoreTableOperations |
|
| BaseOverwriteFiles |
|
| BasePageIterator |
|
| BasePageIterator.IntIterator |
|
| BaseParquetReaders<T> |
|
| BaseParquetWriter<T> |
|
| BaseReplacePartitions |
|
| BaseRewriteManifests |
|
| BaseRowDelta |
|
| BaseTable |
Base Table implementation.
|
| BaseVectorizedParquetValuesReader |
A values reader for Parquet's run-length encoded data that reads column data in batches instead of one value at a
time.
|
| BinaryUtil |
|
| Binder |
Rewrites expressions by replacing unbound named references with references to
fields in a struct schema.
|
| BinPacking |
|
| BinPacking.ListPacker<T> |
|
| BinPacking.PackingIterable<T> |
|
| Bound<T> |
Represents a bound value expression.
|
| BoundLiteralPredicate<T> |
|
| BoundPredicate<T> |
|
| BoundReference<T> |
|
| BoundSetPredicate<T> |
|
| BoundTerm<T> |
Represents a bound term.
|
| BoundTransform<S,T> |
A transform expression.
|
| BoundUnaryPredicate<T> |
|
| ByteBuffers |
|
| CachingCatalog |
|
| Catalog |
A Catalog API for table create, drop, and load operations.
|
| CharSequenceSet |
|
| CharSequenceWrapper |
Wrapper class to adapt CharSequence for use in maps and sets.
|
| CheckCompatibility |
|
| CherrypickAncestorCommitException |
This exception occurs when one cherrypicks an ancestor or when the picked snapshot is already linked to
a published ancestor.
|
| ClientPool<C,E extends java.lang.Exception> |
|
| ClientPool.Action<R,C,E extends java.lang.Exception> |
|
| CloseableGroup |
|
| CloseableIterable<T> |
|
| CloseableIterable.ConcatCloseableIterable<E> |
|
| CloseableIterator<T> |
|
| ColumnarBatchReader |
VectorizedReader that returns Spark's ColumnarBatch to support Spark's vectorized read path.
|
| ColumnIterator<T> |
|
| ColumnWriter<T> |
|
| CombinedScanTask |
A scan task made of several ranges from files.
|
| CommitFailedException |
Exception raised when a commit fails because of out of date metadata.
|
| Comparators |
|
| ContentFile<F> |
|
| Conversions |
|
| CreateSnapshotEvent |
|
| DataFile |
Interface for data files listed in a table manifest.
|
| DataFiles |
|
| DataFiles.Builder |
|
| DataFilesTable |
A Table implementation that exposes a table's data files as rows.
|
| DataFilesTable.FilesTableScan |
|
| DataOperations |
Data operations that produce snapshots.
|
| DataReader<T> |
|
| DataTableScan |
|
| DataTask |
A task that returns data as rows instead of where to read data.
|
| DataWriter<T> |
|
| DateTimeUtil |
|
| DelegatingInputStream |
|
| DelegatingOutputStream |
|
| DeleteFile |
Interface for delete files listed in a table delete manifest.
|
| DeleteFiles |
API for deleting files from a table.
|
| DuplicateWAPCommitException |
This exception occurs when the WAP workflow detects a duplicate wap commit.
|
| DynClasses |
|
| DynClasses.Builder |
|
| DynConstructors |
Copied from parquet-common
|
| DynConstructors.Builder |
|
| DynConstructors.Ctor<C> |
|
| DynFields |
|
| DynFields.BoundField<T> |
|
| DynFields.Builder |
|
| DynFields.StaticField<T> |
|
| DynFields.UnboundField<T> |
Convenience wrapper class around Field.
|
| DynMethods |
Copied from parquet-common
|
| DynMethods.BoundMethod |
|
| DynMethods.Builder |
|
| DynMethods.StaticMethod |
|
| DynMethods.UnboundMethod |
Convenience wrapper class around Method.
|
| EncryptedFiles |
|
| EncryptedInputFile |
Thin wrapper around an InputFile instance that is encrypted.
|
| EncryptedOutputFile |
Thin wrapper around a OutputFile that is encrypting bytes written to the underlying
file system, via an encryption key that is symbolized by the enclosed
EncryptionKeyMetadata.
|
| EncryptionKeyMetadata |
Light typedef over a ByteBuffer that indicates that the given bytes represent metadata about
an encrypted data file's encryption key.
|
| EncryptionKeyMetadatas |
|
| EncryptionManager |
Module for encrypting and decrypting table data files.
|
| Evaluator |
|
| Exceptions |
|
| ExceptionUtil |
|
| ExpireSnapshots |
|
| Expression |
Represents a boolean expression tree.
|
| Expression.Operation |
|
| Expressions |
|
| ExpressionVisitors |
|
| ExpressionVisitors.BoundExpressionVisitor<R> |
Deprecated.
|
| ExpressionVisitors.BoundVisitor<R> |
|
| ExpressionVisitors.ExpressionVisitor<R> |
|
| False |
|
| FileAppender<D> |
|
| FileContent |
Content type stored in a file, one of DATA, POSITION_DELETES, or EQUALITY_DELETES.
|
| FileFormat |
Enum of supported file formats.
|
| FileIO |
Pluggable module for reading, writing, and deleting files.
|
| Files |
|
| FileScanTask |
A scan task over a range of a single file.
|
| FindFiles |
|
| FindFiles.Builder |
|
| FlinkAvroReader |
|
| FlinkAvroWriter |
|
| FlinkParquetReaders |
|
| FlinkParquetWriters |
|
| FlinkSchemaUtil |
|
| FlinkTypeToType |
|
| FlinkTypeVisitor<T> |
|
| GenericManifestFile |
|
| GenericManifestFile.CopyBuilder |
|
| GenericOrcReader |
|
| GenericOrcReaders |
|
| GenericOrcWriter |
|
| GenericParquetReaders |
|
| GenericParquetWriter |
|
| GenericPartitionFieldSummary |
|
| GenericRecord |
|
| GuavaClasses |
|
| HadoopCatalog |
HadoopCatalog provides a way to use table names like db.table to work with path-based tables under a common
location.
|
| HadoopFileIO |
|
| HadoopInputFile |
InputFile implementation using the Hadoop FileSystem API.
|
| HadoopOutputFile |
OutputFile implementation using the Hadoop FileSystem API.
|
| HadoopTableOperations |
TableOperations implementation for file systems that support atomic rename.
|
| HadoopTables |
Implementation of Iceberg tables that uses the Hadoop FileSystem
to store metadata and manifests.
|
| HasTableOperations |
Used to expose a table's TableOperations.
|
| HiddenPathFilter |
A PathFilter that filters out hidden paths.
|
| HistoryEntry |
Table history entry.
|
| HistoryTable |
A Table implementation that exposes a table's history as rows.
|
| HiveCatalog |
|
| HiveCatalogs |
|
| HiveClientPool |
|
| HiveTableOperations |
TODO we should be able to extract some more commonalities to BaseMetastoreTableOperations to
avoid code duplication between this class and Metacat Tables.
|
| HiveTypeConverter |
|
| IcebergArrowColumnVector |
Implementation of Spark's ColumnVector interface.
|
| IcebergArrowVectors |
The general way of getting a value at an index in the Arrow vector
is by calling get(index).
|
| IcebergArrowVectors.DecimalArrowVector |
Extension of Arrow's @DecimalVector.
|
| IcebergArrowVectors.VarcharArrowVector |
Extension of Arrow's @VarCharVector.
|
| IcebergBinaryObjectInspector |
|
| IcebergDateObjectInspector |
|
| IcebergDecimalObjectInspector |
|
| IcebergDecoder<D> |
|
| IcebergEncoder<D> |
|
| IcebergGenerics |
|
| IcebergGenerics.ScanBuilder |
|
| IcebergInputFormat<T> |
Generic Mrv2 InputFormat API for Iceberg.
|
| IcebergObjectInspector |
|
| IcebergPigInputFormat<T> |
|
| IcebergRecordObjectInspector |
|
| IcebergSerDe |
|
| IcebergSource |
|
| IcebergStorage |
|
| IcebergTimestampObjectInspector |
|
| IcebergWritable |
Wraps an Iceberg Record in a Writable which Hive can use in the SerDe.
|
| IdentityPartitionConverters |
|
| InclusiveMetricsEvaluator |
|
| IndexByName |
|
| IndexParents |
|
| InputFile |
|
| InputFormatConfig |
|
| InputFormatConfig.ConfigBuilder |
|
| InputFormatConfig.InMemoryDataModel |
|
| JsonUtil |
|
| Listener<E> |
A listener interface that can receive notifications.
|
| Listeners |
Static registration and notification for listeners.
|
| Literal<T> |
Represents a literal fixed value in an expression predicate
|
| LocationProvider |
Interface for providing data file locations to write tasks.
|
| LocationProviders |
|
| LogicalMap |
|
| ManageSnapshots |
API for managing snapshots.
|
| ManifestContent |
Content type stored in a manifest file, either DATA or DELETES.
|
| ManifestEntriesTable |
A Table implementation that exposes a table's manifest entries as rows, for both delete and data files.
|
| ManifestEvaluator |
|
| ManifestFile |
Represents a manifest file that can be scanned to find data files in a table.
|
| ManifestFile.PartitionFieldSummary |
Summarizes the values of one partition field stored in a manifest file.
|
| ManifestFiles |
|
| ManifestFileUtil |
|
| ManifestReader<F extends ContentFile<F>> |
Base reader for data and delete manifest files.
|
| ManifestReader.FileType |
|
| ManifestsTable |
A Table implementation that exposes a table's manifest files as rows.
|
| ManifestWriter<F extends ContentFile<F>> |
Writer for manifest files.
|
| MappedField |
An immutable mapping between a field ID and a set of names.
|
| MappedFields |
|
| MappingUtil |
|
| MetadataTableType |
|
| Metrics |
Iceberg file format metrics.
|
| MetricsConfig |
|
| MetricsModes |
This class defines different metrics modes, which allow users to control the collection of
value_counts, null_value_counts, lower_bounds, upper_bounds for different columns in metadata.
|
| MetricsModes.Counts |
Under this mode, only value_counts, null_value_counts are persisted.
|
| MetricsModes.Full |
Under this mode, value_counts, null_value_counts and full lower_bounds, upper_bounds are persisted.
|
| MetricsModes.MetricsMode |
|
| MetricsModes.None |
Under this mode, value_counts, null_value_counts, lower_bounds, upper_bounds are not persisted.
|
| MetricsModes.Truncate |
Under this mode, value_counts, null_value_counts and truncated lower_bounds, upper_bounds are persisted.
|
| NamedReference<T> |
|
| NameMapping |
Represents a mapping from external schema names to Iceberg type IDs.
|
| NameMappingParser |
Parses external name mappings from a JSON representation.
|
| Namespace |
|
| NamespaceNotEmptyException |
Exception raised when attempting to drop a namespace that is not empty.
|
| NoSuchIcebergTableException |
NoSuchTableException thrown when a table is found but it is not an Iceberg table.
|
| NoSuchNamespaceException |
Exception raised when attempting to load a namespace that does not exist.
|
| NoSuchTableException |
Exception raised when attempting to load a table that does not exist.
|
| Not |
|
| NotFoundException |
Exception raised when attempting to read a file that does not exist.
|
| NullabilityHolder |
Instances of this class simply track whether a value at an index is null.
|
| NullValuesColumnVector |
|
| Or |
|
| ORC |
|
| ORC.ReadBuilder |
|
| ORC.WriteBuilder |
|
| OrcMetrics |
|
| OrcRowReader<T> |
Used for implementing ORC row readers.
|
| ORCSchemaUtil |
Utilities for mapping Iceberg to ORC schemas.
|
| ORCSchemaUtil.BinaryType |
|
| ORCSchemaUtil.LongType |
|
| OrcSchemaVisitor<T> |
Generic visitor of an ORC Schema.
|
| OrcSchemaWithTypeVisitor<T> |
|
| OrcValueReader<T> |
|
| OrcValueReaders |
|
| OrcValueReaders.StructReader<T> |
|
| OrcValueWriter<T> |
Write data value of a schema.
|
| OutputFile |
|
| OverwriteFiles |
API for overwriting files in a table.
|
| Pair<X,Y> |
|
| ParallelIterable<T> |
|
| Parquet |
|
| Parquet.ReadBuilder |
|
| Parquet.WriteBuilder |
|
| ParquetAvroReader |
|
| ParquetAvroValueReaders |
|
| ParquetAvroValueReaders.TimeMillisReader |
|
| ParquetAvroValueReaders.TimestampMillisReader |
|
| ParquetAvroWriter |
|
| ParquetDictionaryRowGroupFilter |
|
| ParquetIterable<T> |
|
| ParquetMetricsRowGroupFilter |
|
| ParquetReader<T> |
|
| ParquetSchemaUtil |
|
| ParquetSchemaUtil.HasIds |
|
| ParquetTypeVisitor<T> |
|
| ParquetUtil |
|
| ParquetValueReader<T> |
|
| ParquetValueReaders |
|
| ParquetValueReaders.BinaryAsDecimalReader |
|
| ParquetValueReaders.BytesReader |
|
| ParquetValueReaders.FloatAsDoubleReader |
|
| ParquetValueReaders.IntAsLongReader |
|
| ParquetValueReaders.IntegerAsDecimalReader |
|
| ParquetValueReaders.ListReader<E> |
|
| ParquetValueReaders.LongAsDecimalReader |
|
| ParquetValueReaders.MapReader<K,V> |
|
| ParquetValueReaders.PrimitiveReader<T> |
|
| ParquetValueReaders.RepeatedKeyValueReader<M,I,K,V> |
|
| ParquetValueReaders.RepeatedReader<T,I,E> |
|
| ParquetValueReaders.ReusableEntry<K,V> |
|
| ParquetValueReaders.StringReader |
|
| ParquetValueReaders.StructReader<T,I> |
|
| ParquetValueReaders.UnboxedReader<T> |
|
| ParquetValueWriter<T> |
|
| ParquetValueWriters |
|
| ParquetValueWriters.PrimitiveWriter<T> |
|
| ParquetValueWriters.RepeatedKeyValueWriter<M,K,V> |
|
| ParquetValueWriters.RepeatedWriter<L,E> |
|
| ParquetValueWriters.StructWriter<S> |
|
| ParquetWithSparkSchemaVisitor<T> |
Visitor for traversing a Parquet type with a companion Spark type.
|
| ParquetWriteAdapter<D> |
|
| PartitionField |
|
| PartitionSpec |
Represents how to produce partition data for a table.
|
| PartitionSpec.Builder |
|
| PartitionSpecParser |
|
| PartitionSpecVisitor<T> |
|
| PartitionsTable |
A Table implementation that exposes a table's partitions as rows.
|
| PartitionUtil |
|
| PendingUpdate<T> |
API for table metadata changes.
|
| PigParquetReader |
|
| PlaintextEncryptionManager |
|
| PositionOutputStream |
|
| Predicate<T,C extends Term> |
|
| ProjectionDatumReader<D> |
|
| Projections |
Utils to project expressions on rows to expressions on partitions.
|
| Projections.ProjectionEvaluator |
A class that projects expressions for a table's data rows into expressions on the table's
partition values, for a table's partition spec.
|
| PropertyUtil |
|
| PruneColumnsWithoutReordering |
|
| PruneColumnsWithReordering |
|
| Record |
|
| Reference<T> |
|
| RemoveIds |
|
| RemoveIds |
|
| RemoveOrphanFilesAction |
An action that removes orphan metadata and data files by listing a given location and comparing
the actual files in that location with data and metadata files referenced by all valid snapshots.
|
| ReplacePartitions |
Not recommended: API for overwriting files in a table by partition.
|
| ResidualEvaluator |
|
| RewriteDataFilesAction |
|
| RewriteDataFilesActionResult |
|
| RewriteFiles |
API for replacing files in a table.
|
| RewriteManifests |
API for rewriting manifests for a table.
|
| RewriteManifestsAction |
An action that rewrites manifests in a distributed manner and co-locates metadata for partitions.
|
| RewriteManifestsActionResult |
|
| Rollback |
API for rolling table data back to the state at an older table snapshot.
|
| RollbackStagedTable |
An implementation of StagedTable that mimics the behavior of Spark's non-atomic CTAS and RTAS.
|
| RowDataRewriter |
|
| RowDelta |
API for encoding row-level changes to a table.
|
| RuntimeIOException |
Deprecated.
|
| RuntimeMetaException |
Exception used to wrap MetaException as a RuntimeException and add context.
|
| ScanEvent |
Event sent to listeners when a table scan is planned.
|
| ScanSummary |
|
| ScanSummary.Builder |
|
| ScanSummary.PartitionMetrics |
|
| ScanTask |
A scan task.
|
| Schema |
The schema of a data table.
|
| SchemaParser |
|
| SchemaUtil |
|
| SeekableInputStream |
SeekableInputStream is an interface with the methods needed to read data from a file or
Hadoop data stream.
|
| SerializableConfiguration |
Wraps a Configuration object in a Serializable layer.
|
| SerializableSupplier<T> |
|
| SerializationUtil |
|
| SetLocation |
|
| Snapshot |
A snapshot of the data in a table at a point in time.
|
| SnapshotManager |
|
| SnapshotParser |
|
| SnapshotsTable |
A Table implementation that exposes a table's known snapshots as rows.
|
| SnapshotSummary |
|
| SnapshotSummary.Builder |
|
| SnapshotUpdate<ThisT> |
API for table changes that produce snapshots.
|
| SnapshotUpdateAction<ThisT,R> |
|
| SnapshotUtil |
|
| Spark3Util |
|
| Spark3Util.DescribeSchemaVisitor |
|
| SparkAvroReader |
|
| SparkAvroWriter |
|
| SparkCatalog |
A Spark TableCatalog implementation that wraps an Iceberg Catalog.
|
| SparkDataFile |
|
| SparkExceptionUtil |
|
| SparkFilters |
|
| SparkOrcReader |
Converts the OrcIterator, which returns ORC's VectorizedRowBatch to a
set of Spark's UnsafeRows.
|
| SparkOrcWriter |
This class acts as an adaptor from an OrcFileAppender to a
FileAppender<InternalRow>.
|
| SparkParquetReaders |
|
| SparkParquetWriters |
|
| SparkScanBuilder |
|
| SparkSchemaUtil |
Helper methods for working with Spark/Hive metadata.
|
| SparkSessionCatalog<T extends org.apache.spark.sql.connector.catalog.TableCatalog & org.apache.spark.sql.connector.catalog.SupportsNamespaces> |
A Spark catalog that can also load non-Iceberg tables.
|
| SparkStreamingWrite |
|
| SparkStructLike |
|
| SparkTable |
|
| SparkTableUtil |
Java version of the original SparkTableUtil.scala
https://github.com/apache/iceberg/blob/apache-iceberg-0.8.0-incubating/spark/src/main/scala/org/apache/iceberg/spark/SparkTableUtil.scala
|
| SparkTableUtil.SparkPartition |
Class representing a table partition.
|
| SparkUtil |
|
| SparkValueConverter |
A utility class that converts Spark values to Iceberg's internal representation.
|
| SparkValueReaders |
|
| SparkValueWriters |
|
| StagedSparkTable |
|
| StrictMetricsEvaluator |
|
| StructLike |
Interface for accessing data by position in a schema.
|
| StructLikeWrapper |
Wrapper to adapt StructLike for use in maps and sets by implementing equals and hashCode.
|
| SupportsNamespaces |
Catalog methods for working with namespaces.
|
| SystemProperties |
Configuration properties that are controlled by Java system properties.
|
| Table |
Represents a table.
|
| TableIdentifier |
Identifies a table in iceberg catalog.
|
| TableMetadata |
Metadata for a table.
|
| TableMetadata.MetadataLogEntry |
|
| TableMetadata.SnapshotLogEntry |
|
| TableMetadataParser |
|
| TableMetadataParser.Codec |
|
| TableOperations |
SPI interface to abstract table metadata access and updates.
|
| TableProperties |
|
| Tables |
Generic interface for creating and loading a table implementation.
|
| TableScan |
API for configuring a table scan.
|
| TableScanUtil |
|
| Tasks |
|
| Tasks.Builder<I> |
|
| Tasks.FailureTask<I,E extends java.lang.Exception> |
|
| Tasks.Task<I,E extends java.lang.Exception> |
|
| Tasks.UnrecoverableException |
|
| Term |
An expression that evaluates to a value.
|
| ThreadPools |
|
| Transaction |
A transaction for performing multiple updates to a table.
|
| Transactions |
|
| Transform<S,T> |
A transform function used for partitioning.
|
| Transforms |
Factory methods for transforms.
|
| TripleWriter<T> |
|
| True |
|
| Type |
|
| Type.NestedType |
|
| Type.PrimitiveType |
|
| Type.TypeID |
|
| Types |
|
| Types.BinaryType |
|
| Types.BooleanType |
|
| Types.DateType |
|
| Types.DecimalType |
|
| Types.DoubleType |
|
| Types.FixedType |
|
| Types.FloatType |
|
| Types.IntegerType |
|
| Types.ListType |
|
| Types.LongType |
|
| Types.MapType |
|
| Types.NestedField |
|
| Types.StringType |
|
| Types.StructType |
|
| Types.TimestampType |
|
| Types.TimeType |
|
| Types.UUIDType |
|
| TypeToMessageType |
|
| TypeUtil |
|
| TypeUtil.CustomOrderSchemaVisitor<T> |
|
| TypeUtil.NextID |
Interface for passing a function that assigns column IDs.
|
| TypeUtil.SchemaVisitor<T> |
|
| TypeWithSchemaVisitor<T> |
Visitor for traversing a Parquet type with a companion Iceberg type.
|
| Unbound<T,B> |
Represents an unbound expression node.
|
| UnboundPredicate<T> |
|
| UnboundTerm<T> |
Represents an unbound term.
|
| UnboundTransform<S,T> |
|
| UnicodeUtil |
|
| UnknownTransform<S,T> |
|
| UpdateLocation |
API for setting a table's base location.
|
| UpdateProperties |
API for updating table properties.
|
| UpdateSchema |
API for schema evolution.
|
| Util |
|
| UUIDConversion |
|
| ValidationException |
Exception raised when validation checks fail.
|
| ValueReader<T> |
|
| ValueReaders |
|
| ValueReaders.StructReader<S> |
|
| ValuesAsBytesReader |
Implements a ValuesReader specifically to read given number of bytes from the underlying ByteBufferInputStream.
|
| ValueWriter<D> |
|
| ValueWriters |
|
| ValueWriters.StructWriter<S> |
|
| VectorHolder |
Container class for holding the Arrow vector storing a batch of values along with other state needed for reading
values out of it.
|
| VectorizedArrowReader |
|
| VectorizedColumnIterator |
Vectorized version of the ColumnIterator that reads column values in data pages of a column in a row group in a
batched fashion.
|
| VectorizedDictionaryEncodedParquetValuesReader |
This decoder reads Parquet dictionary encoded data in a vectorized fashion.
|
| VectorizedPageIterator |
|
| VectorizedParquetDefinitionLevelReader |
|
| VectorizedParquetReader<T> |
|
| VectorizedReader<T> |
Interface for vectorized Iceberg readers.
|
| VectorizedRowBatchIterator |
An adaptor so that the ORC RecordReader can be used as an Iterator.
|
| VectorizedSparkParquetReaders |
|
| WapUtil |
|