org.apache.hadoop.hive.ql.exec.vector.VectorizationContext vContext
org.apache.hadoop.hive.ql.exec.vector.VectorizationContext vContext
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression conditionEvaluator
org.apache.hadoop.hive.ql.plan.VectorGroupByDesc vectorDesc
org.apache.hadoop.hive.ql.exec.vector.expressions.aggregates.VectorAggregateExpression[] aggregators
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression[] keyExpressions
int outputKeyLength
boolean isVectorOutput
org.apache.hadoop.hive.ql.exec.vector.VectorizationContext vOutContext
float memoryThreshold
org.apache.hadoop.hive.ql.exec.vector.VectorizationContext vOutContext
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression[] keyExpressions
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression[] bigTableFilterExpressions
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression[] bigTableValueExpressions
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpressionWriter[] rowWriters
org.apache.hadoop.hive.ql.exec.vector.VectorizationContext vContext
org.apache.hadoop.hive.serde2.Deserializer currentPartDeserializer
org.apache.hadoop.hive.serde2.objectinspector.StructObjectInspector currentPartRawRowObjectInspector
org.apache.hadoop.hive.ql.exec.vector.VectorAssignRow currentVectorAssign
org.apache.hadoop.hive.ql.exec.vector.VectorizationContext vContext
org.apache.hadoop.hive.ql.plan.VectorSelectDesc vectorDesc
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression[] vExpressions
int[] projectedOutputColumns
org.apache.hadoop.hive.ql.exec.vector.VectorizationContext vOutContext
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression[] bigTableValueExpressions
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression[] bigTableFilterExpressions
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression[] keyExpressions
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpressionWriter[] keyOutputWriters
org.apache.hadoop.hive.ql.exec.vector.VectorizationContext vOutContext
org.apache.hadoop.hive.ql.exec.vector.VectorizationContext vContext
org.apache.hadoop.hive.ql.exec.vector.VectorizationContext vContext
FileScanTask[] tasks
DeleteFile[] deletes
final java.lang.Object writeReplace()
PartitionSpec spec
SortOrder sortOrder
BaseTable table
java.lang.String name
java.util.UUID uuid
StructLike groupingKey
java.lang.Object[] tasks
java.lang.Object writeReplace()
TableOperations ops
java.lang.String name
MetricsReporter reporter
java.lang.Object writeReplace()
int[] fromProjectionPos
InputFile file
java.lang.String manifestPath
java.lang.Long length
int specId
ManifestContent content
long sequenceNumber
long minSequenceNumber
java.lang.Long snapshotId
java.lang.Integer addedFilesCount
java.lang.Integer existingFilesCount
java.lang.Integer deletedFilesCount
java.lang.Long addedRowsCount
java.lang.Long existingRowsCount
java.lang.Long deletedRowsCount
ManifestFile.PartitionFieldSummary[] partitions
byte[] keyMetadata
int[] fromProjectionPos
boolean containsNull
java.lang.Boolean containsNaN
byte[] lowerBound
byte[] upperBound
java.lang.String partition
UnboundPartitionSpec spec
Schema schema
int lastColumnId
Snapshot snapshot
UnboundSortOrder sortOrder
ViewVersion viewVersion
java.lang.String uuid
long snapshotId
java.util.Set<E> removed
long snapshotId
java.lang.String refName
long snapshotId
int schemaId
int versionId
int specId
int sortOrderId
java.lang.String location
PartitionStatisticsFile partitionStatisticsFile
java.util.Map<K,V> updated
java.lang.String refName
java.lang.Long snapshotId
org.apache.iceberg.SnapshotRefType type
java.lang.Integer minSnapshotsToKeep
java.lang.Long maxSnapshotAgeMs
java.lang.Long maxRefAgeMs
long snapshotId
StatisticsFile statisticsFile
int formatVersion
private void readObject(java.io.ObjectInputStream in) throws java.io.IOException, java.lang.ClassNotFoundException
java.io.IOException
- On serialization errorjava.lang.ClassNotFoundException
- If the class is not foundprivate void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException
java.io.IOException
- On serialization errorjava.lang.Long rowCount
java.util.Map<K,V> columnSizes
java.util.Map<K,V> valueCounts
java.util.Map<K,V> nullValueCounts
java.util.Map<K,V> nanValueCounts
java.util.Map<K,V> lowerBounds
java.util.Map<K,V> upperBounds
java.util.Map<K,V> columnModes
MetricsModes.MetricsMode defaultMode
int length
Types.StructType partitionType
int size
java.lang.Object[] data
java.lang.String stringSchema
PartitionSpec spec
Schema inputSchema
Schema schema
int specId
PartitionField[] fields
int lastAssignedFieldId
Schema schema
boolean unpartitionedTable
Schema schema
int defaultSpecId
java.util.Map<K,V> specs
Types.StructType struct
int schemaId
int[] identifierFieldIds
int highestFieldId
java.lang.String name
java.lang.String location
java.lang.String metadataFileLocation
java.util.Map<K,V> properties
java.lang.String schemaAsJson
int defaultSpecId
java.util.Map<K,V> specAsJsonMap
java.lang.String sortOrderAsJson
FileIO io
EncryptionManager encryption
java.util.Map<K,V> refs
java.util.UUID uuid
MetadataTableType type
java.lang.String baseTableName
long snapshotId
org.apache.iceberg.SnapshotRefType type
java.lang.Integer minSnapshotsToKeep
java.lang.Long maxSnapshotAgeMs
java.lang.Long maxRefAgeMs
Transform<S,T> transform
int sourceId
SortDirection direction
NullOrder nullOrder
java.lang.String metadataFileLocation
int formatVersion
java.lang.String uuid
java.lang.String location
long lastSequenceNumber
long lastUpdatedMillis
int lastColumnId
int currentSchemaId
java.util.List<E> schemas
int defaultSpecId
java.util.List<E> specs
int lastAssignedPartitionId
int defaultSortOrderId
java.util.List<E> sortOrders
java.util.Map<K,V> properties
long currentSnapshotId
java.util.Map<K,V> schemasById
java.util.Map<K,V> specsById
java.util.Map<K,V> sortOrdersById
java.util.List<E> snapshotLog
java.util.List<E> previousFiles
java.util.List<E> statisticsFiles
java.util.List<E> partitionStatisticsFiles
java.util.List<E> changes
SerializableSupplier<T> snapshotsSupplier
java.util.List<E> snapshots
java.util.Map<K,V> snapshotsById
java.util.Map<K,V> refs
boolean snapshotsLoaded
long timestampMillis
long snapshotId
java.lang.String ossEndpoint
java.lang.String accessKeyId
java.lang.String accessKeySecret
java.lang.String securityToken
java.lang.String ossStagingDirectory
SerializableSupplier<T> oss
AliyunProperties aliyunProperties
MetricsContext metrics
java.util.concurrent.atomic.AtomicBoolean isResourceClosed
AwsProperties awsProperties
HttpClientProperties httpClientProperties
S3FileIOProperties s3FileIOProperties
java.lang.String roleSessionName
java.lang.String clientRegion
java.lang.String clientCredentialsProvider
java.util.Map<K,V> clientCredentialsProviderProperties
java.util.Set<E> stsClientAssumeRoleTags
java.lang.String clientAssumeRoleArn
java.lang.String clientAssumeRoleExternalId
int clientAssumeRoleTimeoutSec
java.lang.String clientAssumeRoleRegion
java.lang.String clientAssumeRoleSessionName
java.lang.String clientCredentialsProvider
java.util.Map<K,V> clientCredentialsProviderProperties
java.lang.String glueEndpoint
java.lang.String glueCatalogId
boolean glueCatalogSkipArchive
boolean glueCatalogSkipNameValidation
boolean glueLakeFormationEnabled
java.lang.String dynamoDbTableName
java.lang.String dynamoDbEndpoint
java.lang.String restSigningRegion
java.lang.String restSigningName
java.lang.String restAccessKeyId
java.lang.String restSecretAccessKey
java.lang.String restSessionToken
java.lang.String httpClientType
java.util.Map<K,V> httpClientProperties
java.lang.String dbName
java.lang.String tableName
java.lang.String glueCatalogId
java.lang.String glueAccountId
java.lang.String credential
SerializableSupplier<T> s3
S3FileIOProperties s3FileIOProperties
SerializableMap<K,V> properties
MetricsContext metrics
java.util.concurrent.atomic.AtomicBoolean isResourceClosed
java.lang.String sseType
java.lang.String sseKey
java.lang.String sseMd5
java.lang.String accessKeyId
java.lang.String secretAccessKey
java.lang.String sessionToken
boolean isS3AccessGrantsEnabled
boolean isS3AccessGrantsFallbackToIamEnabled
int multipartUploadThreads
int multiPartSize
int deleteBatchSize
double multipartThresholdFactor
java.lang.String stagingDirectory
software.amazon.awssdk.services.s3.model.ObjectCannedACL acl
boolean isChecksumEnabled
java.util.Set<E> writeTags
boolean isWriteTableTagEnabled
boolean isWriteNamespaceTagEnabled
java.util.Set<E> deleteTags
int deleteThreads
boolean isDeleteEnabled
java.util.Map<K,V> bucketToAccessPointMapping
boolean isPreloadClientEnabled
boolean isDualStackEnabled
boolean isPathStyleAccess
boolean isUseArnRegionEnabled
boolean isAccelerationEnabled
java.lang.String endpoint
boolean isRemoteSigningEnabled
java.lang.String writeStorageClass
java.util.Map<K,V> allProperties
java.util.Map<K,V> adlsSasTokens
java.util.Map<K,V> adlsConnectionStrings
com.azure.storage.common.StorageSharedKeyCredential namedKeyCreds
java.lang.Integer adlsReadBlockSize
java.lang.Long adlsWriteBlockSize
AzureProperties azureProperties
MetricsContext metrics
SerializableMap<K,V> properties
java.lang.String ecsS3Endpoint
java.lang.String ecsS3AccessKeyId
java.lang.String ecsS3SecretAccessKey
SerializableSupplier<T> s3
DellProperties dellProperties
DellClientFactory dellClientFactory
java.util.concurrent.atomic.AtomicBoolean isResourceClosed
MetricsContext metrics
FileIO io
EncryptionManager em
java.lang.String tableKeyId
int dataKeyLength
Expression.Operation op
Term term
Expression left
Expression right
Types.NestedField field
Accessor<T> accessor
java.lang.String name
java.util.Set<E> literalSet
BoundReference<T> ref
Transform<S,T> transform
SerializableFunction<S,T> func
int fieldId
Types.NestedField field
Expression expr
java.lang.Object writeReplace() throws java.io.ObjectStreamException
java.io.ObjectStreamException
int fieldId
Type.PrimitiveType type
java.util.Comparator<T> comparator
int fieldId
Type.PrimitiveType type
java.util.Comparator<T> comparator
java.lang.String name
Expression child
Expression left
Expression right
Expression.Operation op
Term term
PartitionSpec spec
Expression expr
boolean caseSensitive
java.lang.Object writeReplace() throws java.io.ObjectStreamException
java.io.ObjectStreamException
java.util.List<E> literals
NamedReference<T> ref
Transform<S,T> transform
NamedReference<T>[] refs
SerializableConfiguration hadoopConf
java.util.Map<K,V> properties
java.lang.String name
java.lang.String impl
java.lang.String catalogName
SerializableConfiguration hadoopConf
java.lang.String warehouseLocation
java.util.Map<K,V> properties
java.lang.String catalogName
SerializableConfiguration hadoopConf
java.lang.String uri
java.lang.String warehouse
int clientPoolSize
java.util.Map<K,V> properties
java.lang.String catalogName
SerializableConfiguration hadoopConf
java.util.Map<K,V> properties
org.apache.flink.table.types.logical.RowType rowType
Evaluator evaluator
Types.StructType struct
RowDataWrapper wrapper
CatalogLoader catalogLoader
java.lang.String identifier
java.lang.String location
SerializableConfiguration hadoopConf
org.apache.flink.formats.avro.AvroToRowDataConverters.AvroToRowDataConverter converter
Schema schema
org.apache.flink.table.types.logical.RowType flinkSchema
java.util.Map<K,V> props
PartitionSpec spec
int[] equalityFieldIds
Schema eqDeleteRowSchema
Schema posDeleteRowSchema
Table table
org.apache.flink.table.types.logical.RowType eqDeleteFlinkSchema
org.apache.flink.table.types.logical.RowType posDeleteFlinkSchema
java.util.function.Supplier<T> tableSupplier
Schema schema
org.apache.flink.table.types.logical.RowType flinkSchema
PartitionSpec spec
long targetFileSizeBytes
FileFormat format
java.util.List<E> equalityFieldIds
boolean upsert
FileAppenderFactory<T> appenderFactory
java.lang.String operatorName
org.apache.flink.api.common.typeutils.TypeSerializer<T> statisticsSerializer
RowDataFileScanTaskReader rowDataReader
RowDataToAvroGenericRecordConverter converter
Schema readSchema
TableLoader tableLoader
FileIO io
EncryptionManager encryption
ScanContext context
FileScanTaskReader<T> rowDataReader
CombinedScanTask task
TableLoader tableLoader
ScanContext scanContext
ReaderFunction<T> readerFunction
SplitAssignerFactory assignerFactory
SerializableComparator<T> splitComparator
SerializableRecordEmitter<T> emitter
java.lang.String tableName
Schema tableSchema
Schema projectedSchema
java.lang.String nameMapping
boolean caseSensitive
FlinkSourceFilter rowFilter
TaskWriter<T> writer
int subTaskId
int attemptId
Schema schema
java.lang.String nameMapping
FileIO io
boolean caseSensitive
EncryptionManager encryptionManager
TaskWriterFactory<T> taskWriterFactory
RowDataFileScanTaskReader rowDataReader
boolean caseSensitive
boolean exposeLocality
java.lang.Long snapshotId
java.lang.String branch
java.lang.String tag
StreamingStartingStrategy startingStrategy
java.lang.Long startSnapshotId
java.lang.Long startSnapshotTimestamp
java.lang.Long endSnapshotId
java.lang.Long asOfTimestamp
java.lang.String startTag
java.lang.String endTag
java.lang.Long splitSize
java.lang.Integer splitLookback
java.lang.Long splitOpenFileCost
boolean isStreaming
java.time.Duration monitorInterval
java.lang.String nameMapping
Schema schema
java.util.List<E> filters
long limit
boolean includeColumnStats
java.util.Collection<E> includeStatsForColumns
java.lang.Integer planParallelism
int maxPlanningSnapshotCount
int maxAllowedPlanningFailures
java.lang.String watermarkColumn
java.util.concurrent.TimeUnit watermarkColumnTimeUnit
TableLoader tableLoader
ScanContext scanContext
boolean isRunning
long lastSnapshotId
org.apache.flink.api.common.operators.MailboxExecutor executor
FlinkInputFormat format
SerializableComparator<T> comparator
org.apache.iceberg.flink.source.enumerator.IcebergEnumeratorPosition lastEnumeratedPosition
java.util.Collection<E> pendingSplits
int[] enumerationSplitCountHistory
java.lang.String tableName
Schema readSchema
FileIO io
EncryptionManager encryption
RowDataFileScanTaskReader rowDataReader
int eventTimeFieldId
java.lang.String eventTimeFieldName
java.util.concurrent.TimeUnit timeUnit
DataIteratorBatcher<T> batcher
Schema readSchema
FileIO io
EncryptionManager encryption
Schema tableSchema
Schema readSchema
java.lang.String nameMapping
boolean caseSensitive
FileIO io
EncryptionManager encryption
java.util.List<E> filters
CombinedScanTask task
int fileOffset
long recordOffset
java.util.Collection<E> finishedSplitIds
java.lang.String requesterHostname
java.lang.String projectId
java.lang.String clientLibToken
java.lang.String serviceHost
java.lang.String gcsDecryptionKey
java.lang.String gcsEncryptionKey
java.lang.String gcsUserProject
java.lang.Integer gcsChannelReadChunkSize
java.lang.Integer gcsChannelWriteChunkSize
boolean gcsNoAuth
java.lang.String gcsOAuth2Token
java.util.Date gcsOAuth2TokenExpiresAt
int gcsDeleteBatchSize
SerializableSupplier<T> storageSupplier
GCPProperties gcpProperties
MetricsContext metrics
java.util.concurrent.atomic.AtomicBoolean isResourceClosed
SerializableMap<K,V> properties
SerializableSupplier<T> hadoopConf
SerializableMap<K,V> properties
java.lang.String scheme
private void readObject(java.io.ObjectInputStream in) throws java.lang.ClassNotFoundException, java.io.IOException
java.lang.ClassNotFoundException
java.io.IOException
private void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException
java.io.IOException
boolean closed
int numberFailedObjects
java.util.Map<K,V> ioInstances
java.util.concurrent.atomic.AtomicBoolean isClosed
SerializableMap<K,V> properties
SerializableSupplier<T> hadoopConf
DataFile[] dataFiles
DeleteFile[] deleteFiles
java.lang.CharSequence[] referencedDataFiles
java.util.Set<E> names
java.lang.Integer id
MappedFields nestedMapping
java.util.List<E> fields
MappedFields mapping
java.util.Map<K,V> values
java.lang.String uri
java.lang.String format
java.lang.String path
java.lang.Long length
java.lang.Integer partitionSpecId
java.lang.Long addedSnapshotId
java.lang.Integer content
java.lang.Long sequenceNumber
java.util.Set<E> set
org.apache.spark.sql.types.DataType sqlType
int precision
int scale
org.apache.spark.sql.types.DataType sqlType
org.apache.spark.sql.types.DataType sqlType
int precision
int scale
DeleteFile[] taskFiles
DataFile[] dataFiles
DeleteFile[] deleteFiles
java.lang.CharSequence[] referencedDataFiles
DataFile[] taskFiles
java.lang.Object writeReplace() throws java.io.ObjectStreamException
java.io.ObjectStreamException
java.lang.Object writeReplace() throws java.io.ObjectStreamException
java.io.ObjectStreamException
java.lang.Object writeReplace() throws java.io.ObjectStreamException
java.io.ObjectStreamException
java.lang.String transform
java.lang.Object writeReplace() throws java.io.ObjectStreamException
java.io.ObjectStreamException
int scale
int precision
int length
Types.NestedField elementField
Types.NestedField keyField
Types.NestedField valueField
boolean isOptional
int id
java.lang.String name
Type type
java.lang.String doc
Types.NestedField[] fields
boolean adjustToUTC
java.util.Map<K,V> wrapperMap
java.util.Set<E> wrapperSet
java.lang.CharSequence wrapped
org.apache.avro.Schema schema
java.lang.Object first
java.lang.Object second
java.util.Map<K,V> copiedMap
ViewOperations ops
java.lang.String name
scala.collection.Seq<A> catalogAndNamespace
int nestedViewDepth
int maxNestedViewDepth
scala.collection.mutable.Map<K,V> relationCache
scala.collection.Seq<A> referredTempViewNames
scala.collection.mutable.Set<A> referredTempFunctionNames
scala.Option<A> outerPlan
private java.lang.Object readResolve()
java.lang.ThreadLocal<T> value
scala.Enumeration.Value BOTH
scala.Enumeration.Value SOURCE
scala.Enumeration.Value TARGET
long timestamp
private java.lang.Object readResolve()
java.lang.String version
private java.lang.Object readResolve()
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.expressions.Expression identifierExpr
scala.Function1<T1,R> exprBuilder
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
int ordinal
org.apache.spark.sql.types.DataType dataType
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
java.lang.String viewName
java.lang.String colName
int ordinal
int expectedNumCandidates
scala.Option<A> viewDDL
boolean bitmap$0
private java.lang.Object readResolve()
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.expressions.Expression child
scala.collection.Seq<A> names
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
java.lang.String name
boolean resolved
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan child
scala.collection.immutable.Map<K,V> args
private java.lang.Object readResolve()
java.lang.String message
scala.Option<A> cause
private java.lang.Object readResolve()
private java.lang.Object readResolve()
private java.lang.Object readResolve()
private java.lang.Object readResolve()
private java.lang.Object readResolve()
org.apache.spark.sql.catalyst.expressions.Expression identifierExpr
scala.Function1<T1,R> planBuilder
scala.collection.Seq<A> nodePatterns
boolean resolved
boolean bitmap$0
private java.lang.Object readResolve()
int pos
boolean resolved
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan child
org.apache.spark.sql.catalyst.expressions.Expression[] args
private java.lang.Object readResolve()
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan relation
scala.Option<A> timestamp
scala.Option<A> version
scala.collection.Seq<A> nodePatterns
boolean resolved
boolean bitmap$0
private java.lang.Object readResolve()
java.lang.Class<T> cls
scala.collection.Seq<A> outputAttrIds
private java.lang.Object readResolve()
private java.lang.Object readResolve()
scala.collection.Seq<A> org$apache$spark$sql$catalyst$analysis$ColumnResolutionHelper$$literalFunctions
scala.collection.Seq<A> path
org.apache.spark.sql.types.StructField field
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.TableChange.ColumnPosition position
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.CatalogPlugin catalog
org.apache.spark.sql.connector.catalog.Identifier identifier
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.CatalogPlugin catalog
scala.collection.Seq<A> namespace
private java.lang.Object readResolve()
java.lang.String name
org.apache.spark.sql.connector.catalog.functions.UnboundFunction func
private java.lang.Object readResolve()
scala.collection.Seq<A> names
org.apache.spark.sql.catalyst.InternalRow ident
scala.Option<A> location
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.FunctionCatalog catalog
org.apache.spark.sql.connector.catalog.Identifier identifier
org.apache.spark.sql.connector.catalog.functions.UnboundFunction func
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.CatalogPlugin catalog
org.apache.spark.sql.connector.catalog.Identifier identifier
org.apache.spark.sql.types.StructType viewSchema
private java.lang.Object readResolve()
scala.collection.Seq<A> expressions
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.TableCatalog catalog
org.apache.spark.sql.connector.catalog.Identifier identifier
org.apache.spark.sql.connector.catalog.Table table
scala.collection.Seq<A> outputAttributes
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.Identifier identifier
org.apache.spark.sql.types.StructType viewSchema
private java.lang.Object readResolve()
org.apache.spark.sql.catalyst.expressions.Alias alias
int index
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.CatalogManager catalogManager
org.apache.spark.sql.SparkSession spark
org.apache.spark.sql.connector.catalog.LookupCatalog.CatalogAndMultipartIdentifier$ CatalogAndMultipartIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog.SessionCatalogAndIdentifier$ SessionCatalogAndIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog.NonSessionCatalogAndIdentifier$ NonSessionCatalogAndIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog.CatalogAndNamespace$ CatalogAndNamespace$module
org.apache.spark.sql.connector.catalog.LookupCatalog.CatalogAndIdentifier$ CatalogAndIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog.AsTableIdentifier$ AsTableIdentifier$module
boolean bitmap$0
private java.lang.Object readResolve()
private java.lang.Object readResolve()
scala.collection.Seq<A> org$apache$spark$sql$catalyst$analysis$ColumnResolutionHelper$$literalFunctions
org.apache.spark.sql.connector.catalog.CatalogManager catalogManager
org.apache.spark.sql.SparkSession spark
org.apache.spark.sql.connector.catalog.LookupCatalog.CatalogAndMultipartIdentifier$ CatalogAndMultipartIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog.SessionCatalogAndIdentifier$ SessionCatalogAndIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog.NonSessionCatalogAndIdentifier$ NonSessionCatalogAndIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog.CatalogAndNamespace$ CatalogAndNamespace$module
org.apache.spark.sql.connector.catalog.LookupCatalog.CatalogAndIdentifier$ CatalogAndIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog.AsTableIdentifier$ AsTableIdentifier$module
boolean bitmap$0
private java.lang.Object readResolve()
org.apache.spark.sql.connector.catalog.CatalogManager catalogManager
org.apache.spark.sql.catalyst.analysis.RewriteViewCommands.ResolvedIdent$ ResolvedIdent$module
org.apache.spark.sql.catalyst.analysis.RewriteViewCommands.ResolvedView$ ResolvedView$module
org.apache.spark.sql.SparkSession spark
org.apache.spark.sql.connector.catalog.LookupCatalog.CatalogAndMultipartIdentifier$ CatalogAndMultipartIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog.SessionCatalogAndIdentifier$ SessionCatalogAndIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog.NonSessionCatalogAndIdentifier$ NonSessionCatalogAndIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog.CatalogAndNamespace$ CatalogAndNamespace$module
org.apache.spark.sql.connector.catalog.LookupCatalog.CatalogAndIdentifier$ CatalogAndIdentifier$module
org.apache.spark.sql.connector.catalog.LookupCatalog.AsTableIdentifier$ AsTableIdentifier$module
boolean bitmap$0
private java.lang.Object readResolve()
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.expressions.Expression canonicalized
org.apache.spark.sql.catalyst.expressions.Expression child
scala.collection.Seq<A> nameParts
boolean hasTried
scala.collection.Seq<A> nodePatterns
byte bitmap$0
private java.lang.Object readResolve()
private java.lang.Object readResolve()
java.lang.String errorSubClass
scala.collection.immutable.Map<K,V> messageParameters
private java.lang.Object readResolve()
java.lang.String errorSubClass
scala.collection.immutable.Map<K,V> messageParameters
private java.lang.Object readResolve()
java.lang.String message
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.expressions.Expression child
scala.Option<A> aliasFunc
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> nameParts
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.expressions.Expression deserializer
scala.collection.Seq<A> inputAttributes
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.expressions.Expression child
org.apache.spark.sql.catalyst.expressions.Expression extraction
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> name
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.connector.catalog.TableChange.ColumnPosition position
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.Seq<A> nameParts
scala.collection.Seq<A> arguments
boolean isDistinct
scala.Option<A> filter
boolean ignoreNulls
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
scala.collection.Seq<A> multipartIdentifier
java.lang.String commandName
boolean requirePersistent
scala.Option<A> funcTypeMismatchHint
scala.Option<A> possibleQualifiedName
scala.collection.Seq<A> nodePatterns
boolean resolved
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
org.apache.spark.sql.catalyst.FunctionIdentifier name
scala.collection.Seq<A> children
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
org.apache.spark.sql.catalyst.expressions.Expression havingCondition
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan child
scala.collection.Seq<A> nodePatterns
boolean resolved
org.apache.spark.sql.catalyst.expressions.ExpressionSet validConstraints
byte bitmap$0
private java.lang.Object readResolve()
scala.collection.Seq<A> nameParts
boolean allowTemp
boolean resolved
boolean bitmap$0
private java.lang.Object readResolve()
boolean expressionsResolved
scala.collection.Seq<A> names
scala.collection.Seq<A> rows
boolean resolved
byte bitmap$0
private java.lang.Object readResolve()
scala.collection.Seq<A> multipartIdentifier
boolean resolved
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
int ordinal
scala.collection.Seq<A> nodePatterns
boolean bitmap$0
private java.lang.Object readResolve()
boolean resolved
scala.collection.immutable.Map<K,V> spec
scala.Option<A> location
boolean bitmap$0
private java.lang.Object readResolve()
java.lang.String regexPattern
scala.Option<A> table
boolean caseSensitive
private java.lang.Object readResolve()
scala.collection.Seq<A> multipartIdentifier
org.apache.spark.sql.util.CaseInsensitiveStringMap options
boolean isStreaming
scala.collection.Seq<A> nodePatterns
boolean resolved
boolean bitmap$0
private java.lang.Object readResolve()
private java.lang.Object readResolve()
boolean resolved
boolean bitmap$0
scala.Option<A> target
private java.lang.Object readResolve()
scala.collection.Seq<A> outputColumnNames
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan child
scala.collection.Seq<A> nodePatterns
boolean resolved
org.apache.spark.sql.catalyst.expressions.ExpressionSet validConstraints
byte bitmap$0
private java.lang.Object readResolve()
scala.collection.Seq<A> multipartIdentifier
java.lang.String commandName
scala.Option<A> relationTypeMismatchHint
boolean resolved
boolean bitmap$0
private java.lang.Object readResolve()
scala.collection.Seq<A> multipartIdentifier
java.lang.String commandName
boolean allowTempView
boolean resolved
boolean bitmap$0
private java.lang.Object readResolve()
scala.collection.Seq<A> name
scala.collection.Seq<A> functionArgs
scala.collection.Seq<A> nodePatterns
boolean resolved
boolean bitmap$0
private java.lang.Object readResolve()
scala.collection.Seq<A> name
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan child
scala.collection.Seq<A> outputNames
scala.collection.Seq<A> nodePatterns
boolean resolved
org.apache.spark.sql.catalyst.expressions.ExpressionSet validConstraints
byte bitmap$0
private java.lang.Object readResolve()
scala.collection.Seq<A> multipartIdentifier
java.lang.String commandName
boolean allowTemp
scala.Option<A> relationTypeMismatchHint
boolean resolved
boolean bitmap$0
private java.lang.Object readResolve()
private java.lang.Object readResolve()
scala.Option<A> command
org.apache.spark.sql.catalyst.trees.Origin start
org.apache.spark.sql.catalyst.trees.Origin stop
private java.lang.Object readResolve()