Package org.apache.iceberg
package org.apache.iceberg
- 
ClassDescriptionAccessor<T>Position2Accessor and Position3Accessor here is an optimization.A scan task for inserts generated by adding a data file to the table.ATableimplementation that exposes a table's valid data files as rows.ATableimplementation that exposes its valid delete files as rows.ATableimplementation that exposes a table's manifest entries as rows, for both delete and data files.ATableimplementation that exposes its valid files as rows.ATableimplementation that exposes a table's valid manifest files as rows.API for appending new files in a table.Base class for metadata tables.BaseScanTaskGroup<T extends ScanTask>BaseTableimplementation.API for configuring a batch scan.A metadata about a statistics or indices blob.Class that wraps an Iceberg Catalog to cache tables.An enum representing possible operations in a changelog.A changelog scan task.ClientPool<C,E extends Exception> ClientPool.Action<R,C, E extends Exception> ClientPoolImpl<C,E extends Exception> A scan task made of several ranges from files.ContentFile<F>Superinterface ofDataFileandDeleteFilethat exposes common methods.ContentScanTask<F extends ContentFile<F>>A scan task over a range of bytes in a content file.Interface for data files listed in a table manifest.ATableimplementation that exposes a table's data files as rows.Data operations that produce snapshots.A task that returns data asrowsinstead of where to read data.A scan task for deletes generated by removing a data file from the table.A scan task for deletes generated by adding delete files to the table.Interface for delete files listed in a table delete manifest.API for deleting files from a table.ATableimplementation that exposes a table's delete files as rows.Enum of supported write distribution mode, it defines the write behavior of batch or streaming job:Iceberg internally tracked field level metrics, used by Parquet and ORC writers only.API for removing oldsnapshotsfrom a table.FieldMetrics<T>Iceberg internally tracked field level metrics.Content type stored in a file, one of DATA, POSITION_DELETES, or EQUALITY_DELETES.Enum of supported file formats.A scan task over a range of bytes in a single data file.ATableimplementation that exposes a table's files as rows.Iceberg internally tracked field level metrics, used by Parquet and ORC writers only.Used to expose a table's TableOperations.Table history entry.ATableimplementation that exposes a table's history as rows.Loads iceberg-version.properties with build information.API for configuring an incremental table scan for appends only snapshotsAPI for configuring a scan for table changes.IncrementalScan<ThisT,T extends ScanTask, G extends ScanTaskGroup<T>> API for configuring an incremental scan.An isolation level in a table.An interface for locking, used to ensure commit isolation.API for managing snapshots.Content type stored in a manifest file, either DATA or DELETES.ATableimplementation that exposes a table's manifest entries as rows, for both delete and data files.Represents a manifest file that can be scanned to find files in a table.Summarizes the values of one partition field stored in a manifest file.ManifestReader<F extends ContentFile<F>>Base reader for data and delete manifest files.ATableimplementation that exposes a table's manifest files as rows.ManifestWriter<F extends ContentFile<F>>Writer for manifest files.MergeableScanTask<ThisT>A scan task that can be potentially merged with other scan tasks.Represents a change to table or view metadata.Iceberg file format metrics.This class defines different metrics modes, which allow users to control the collection of value_counts, null_value_counts, nan_value_counts, lower_bounds, upper_bounds for different columns in metadata.Under this mode, only value_counts, null_value_counts, nan_value_counts are persisted.Under this mode, value_counts, null_value_counts, nan_value_counts and full lower_bounds, upper_bounds are persisted.A metrics calculation mode.Under this mode, value_counts, null_value_counts, nan_value_counts, lower_bounds, upper_bounds are not persisted.Under this mode, value_counts, null_value_counts, nan_value_counts and truncated lower_bounds, upper_bounds are persisted.A struct of readable metric values for a primitive columnFixed definition of a readable metric column, ie a mapping of a raw metric to a readable metricA struct, consisting of allMetricsUtil.ReadableColMetricsStructfor all primitive columns of the tableAPI for overwriting files in a table.Represents a single field in aPartitionSpec.A struct of partition values.A scan task for data within a particular partitionRepresents how to produce partition data for a table.Used to create validpartition specs.ATableimplementation that exposes a table's partitions as rows.Represents a partition statistics file that can be used to read table data more efficiently.Computes, writes and reads thePartitionStatisticsFile.Deprecated.API for table metadata changes.AScanTaskfor position delete filesATableimplementation whoseScanprovidesPositionDeletesScanTask, for reading of position delete files.ATableimplementation that exposes a table's known snapshot references as rows.API for overwriting files in a table by partition.API for replacing table sort order with a newly created order.API for replacing files in a table.Enum of supported rewrite job order, it defines the order in which the file groups should be written.API for rewriting manifests for a table.Utilities for Rewrite table path action.Class providing engine-specific methods to read and write position delete files.Rewrite result.RollingManifestWriter<F extends ContentFile<F>>As opposed toManifestWriter, a rolling writer could produce multiple manifest files.API for encoding row-level changes to a table.Iceberg supports two ways to modify records in a table: copy-on-write and merge-on-read.Scan<ThisT,T extends ScanTask, G extends ScanTaskGroup<T>> Scan objects are immutable and can be shared between threads.A scan task.ScanTaskGroup<T extends ScanTask>A scan task that may include partial input files, multiple input files or both.The schema of a data table.A read-only serializable table that can be sent to other nodes in a cluster.A snapshot of the data in a table at a point in time.SnapshotScan<ThisT,T extends ScanTask, G extends ScanTaskGroup<T>> This is a common base class to share code between different BaseScan implementations that handle scans of a particular snapshot.ATableimplementation that exposes a table's known snapshots as rows.SnapshotUpdate<ThisT>API for table changes that produce snapshots.A field in aSortOrder.A struct of flattened sort field values.A sort order that defines how data and delete files should be ordered in a table.A builder used to create validsort orders.Methods for building a sort order.A batch data scan that can utilize Spark cluster resources for planning.SplittableScanTask<ThisT>A scan task that can be split into smaller scan tasks.TableOperations implementation that provides access to metadata for a Table at some point in time, using a table metadata location.Represents a statistics file in the Puffin format, that can be used to read table data more efficiently.Deleteimplementation that avoids loading full manifests in memory.Interface for accessing data by position in a schema.Configuration properties that are controlled by Java system properties or environmental variable.Deprecated.UseSystemConfigsinstead; will be removed in 2.0.0Represents a table.Metadata for a table.SPI interface to abstract table metadata access and updates.Generic interface for creating and loading a table implementation.API for configuring a table scan.A transaction for performing multiple updates to a table.API for setting a table's or view's base location.API for partition spec evolution.API for updating partition statistics files in a table.API for updating table properties.Represents a requirement for aMetadataUpdateAPI for schema evolution.API for updating statistics files in a table.
PartitionStatsHandlerdirectly