org.apache.flink.api.common.typeutils.TypeSerializer<T> typeSerializer
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
com.google.protobuf.UnknownFieldSet unknownFields
int bitField0_
long sum_
byte memoizedIsInitialized
int memoizedSerializedSize
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
com.google.protobuf.UnknownFieldSet unknownFields
int bitField0_
int numHashFunctions_
List<E> bitset_
com.google.protobuf.ByteString utf8Bitset_
byte memoizedIsInitialized
int memoizedSerializedSize
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
com.google.protobuf.UnknownFieldSet unknownFields
int bitField0_
org.apache.orc.OrcProto.ColumnEncoding.Kind kind_
int dictionarySize_
int bloomEncoding_
byte memoizedIsInitialized
int memoizedSerializedSize
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
com.google.protobuf.UnknownFieldSet unknownFields
int bitField0_
long numberOfValues_
org.apache.orc.OrcProto.IntegerStatistics intStatistics_
org.apache.orc.OrcProto.DoubleStatistics doubleStatistics_
org.apache.orc.OrcProto.StringStatistics stringStatistics_
org.apache.orc.OrcProto.BucketStatistics bucketStatistics_
org.apache.orc.OrcProto.DecimalStatistics decimalStatistics_
org.apache.orc.OrcProto.DateStatistics dateStatistics_
org.apache.orc.OrcProto.BinaryStatistics binaryStatistics_
org.apache.orc.OrcProto.TimestampStatistics timestampStatistics_
boolean hasNull_
long bytesOnDisk_
byte memoizedIsInitialized
int memoizedSerializedSize
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
com.google.protobuf.UnknownFieldSet unknownFields
int bitField0_
int minimum_
int maximum_
byte memoizedIsInitialized
int memoizedSerializedSize
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
com.google.protobuf.UnknownFieldSet unknownFields
int bitField0_
double minimum_
double maximum_
double sum_
byte memoizedIsInitialized
int memoizedSerializedSize
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
com.google.protobuf.UnknownFieldSet unknownFields
int bitField0_
org.apache.orc.OrcProto.PostScript postscript_
org.apache.orc.OrcProto.Footer footer_
long fileLength_
long postscriptLength_
byte memoizedIsInitialized
int memoizedSerializedSize
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
com.google.protobuf.UnknownFieldSet unknownFields
int bitField0_
long headerLength_
long contentLength_
List<E> stripes_
List<E> types_
List<E> metadata_
long numberOfRows_
List<E> statistics_
int rowIndexStride_
int writer_
byte memoizedIsInitialized
int memoizedSerializedSize
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
com.google.protobuf.UnknownFieldSet unknownFields
int bitField0_
long minimum_
long maximum_
long sum_
byte memoizedIsInitialized
int memoizedSerializedSize
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
com.google.protobuf.UnknownFieldSet unknownFields
int bitField0_
long footerLength_
org.apache.orc.OrcProto.CompressionKind compression_
long compressionBlockSize_
List<E> version_
int versionMemoizedSerializedSize
long metadataLength_
int writerVersion_
Object magic_
byte memoizedIsInitialized
int memoizedSerializedSize
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
com.google.protobuf.UnknownFieldSet unknownFields
int bitField0_
List<E> positions_
int positionsMemoizedSerializedSize
org.apache.orc.OrcProto.ColumnStatistics statistics_
byte memoizedIsInitialized
int memoizedSerializedSize
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
com.google.protobuf.UnknownFieldSet unknownFields
int bitField0_
org.apache.orc.OrcProto.Stream.Kind kind_
int column_
long length_
byte memoizedIsInitialized
int memoizedSerializedSize
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
com.google.protobuf.UnknownFieldSet unknownFields
int bitField0_
long offset_
long indexLength_
long dataLength_
long footerLength_
long numberOfRows_
byte memoizedIsInitialized
int memoizedSerializedSize
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
com.google.protobuf.UnknownFieldSet unknownFields
int bitField0_
long minimum_
long maximum_
long minimumUtc_
long maximumUtc_
byte memoizedIsInitialized
int memoizedSerializedSize
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
com.google.protobuf.UnknownFieldSet unknownFields
int bitField0_
org.apache.orc.OrcProto.Type.Kind kind_
List<E> subtypes_
int subtypesMemoizedSerializedSize
com.google.protobuf.LazyStringList fieldNames_
int maximumLength_
int precision_
int scale_
byte memoizedIsInitialized
int memoizedSerializedSize
protected Object writeReplace() throws ObjectStreamException
ObjectStreamException
com.google.protobuf.UnknownFieldSet unknownFields
int bitField0_
Object name_
com.google.protobuf.ByteString value_
byte memoizedIsInitialized
int memoizedSerializedSize
org.apache.hadoop.fs.Path path
String versionString
org.apache.orc.OrcProto.PostScript postscript
Options options
Identifier identifier
String column
Identifier identifier
String column
String database
String database
String database
Identifier identifier
Map<K,V> partitionSpec
Identifier identifier
Identifier identifier
Identifier identifier
Identifier identifier
String compress
int zstdLevel
FileStoreTable table
Table table
int arity
int nullBitsSizeInBytes
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
int precision
int scale
long longVal
BigDecimal decimalVal
Object array
int size
boolean isPrimitiveArray
long millisecond
int nanoOfMillisecond
int arity
int nullBitsSizeInBytes
long millisecond
int nanoOfMillisecond
ColumnVector data
int offset
int numElements
ColumnVector keyColumnVector
ColumnVector valueColumnVector
int offset
int numElements
RowKind rowKind
VectorizedColumnBatch vectorizedColumnBatch
int rowId
int numRows
ColumnVector[] columns
boolean[] isNull
HeapIntVector dictionaryIds
int len
long[] offsets
long[] lengths
int size
ColumnVector child
boolean[] vector
int[] start
int[] length
byte[] buffer
int elementsAppended
int capacity
byte[] vector
double[] vector
float[] vector
int[] vector
long[] vector
long[] offsets
long[] lengths
int size
ColumnVector keys
ColumnVector values
WritableColumnVector[] fields
short[] vector
long[] milliseconds
int[] nanoOfMilliseconds
boolean noNulls
Dictionary dictionary
int numFields
int fixedLengthPartSize
int precision
int scale
DataType eleType
Serializer<T> eleSer
InternalArray.ElementGetter elementGetter
DataType keyType
DataType valueType
Serializer<T> keySerializer
Serializer<T> valueSerializer
InternalArray.ElementGetter keyGetter
InternalArray.ElementGetter valueGetter
BinaryRowSerializer binarySerializer
DataType[] types
Serializer<T>[] fieldSerializers
InternalRow.FieldGetter[] fieldGetters
BinaryWriter.ValueSetter[] valueSetters
Serializer<T> elementSerializer
Serializer<T> originalSerializer
InternalRow.FieldGetter[] getters
org.apache.paimon.data.serializer.RowCompactedSerializer.FieldWriter[] writers
org.apache.paimon.data.serializer.RowCompactedSerializer.FieldReader[] readers
RowType rowType
org.apache.paimon.data.serializer.RowCompactedSerializer.RowWriter rowWriter
org.apache.paimon.data.serializer.RowCompactedSerializer.RowReader rowReader
int precision
org.apache.flink.core.fs.Path path
String columnName
Expression expression
String pattern
SerializableSupplier<T> converterSupplier
org.apache.paimon.shade.jackson2.com.fasterxml.jackson.databind.JsonNode root
TypeMapping typeMapping
List<E> computedColumns
org.apache.avro.generic.GenericRecord keyRecord
org.apache.avro.generic.GenericRecord valueRecord
String topic
String schemaRegistryUrl
ConfluentAvroDeserializationSchema avroDeserializer
org.apache.paimon.shade.jackson2.com.fasterxml.jackson.databind.ObjectMapper objectMapper
org.apache.paimon.shade.jackson2.com.fasterxml.jackson.databind.ObjectMapper objectMapper
java.time.ZoneId serverTimeZone
List<E> computedColumns
TypeMapping typeMapping
boolean isDebeziumSchemaCommentsEnabled
DebeziumEvent root
String currentTable
String databaseName
CdcMetadataConverter[] metadataConverters
Set<E> nonPkTables
org.apache.paimon.shade.jackson2.com.fasterxml.jackson.databind.ObjectMapper objectMapper
java.time.ZoneId serverTimeZone
List<E> computedColumns
TypeMapping typeMapping
DebeziumEvent root
String currentTable
String databaseName
CdcMetadataConverter[] metadataConverters
String topic
String schemaRegistryUrl
ConfluentAvroDeserializationSchema avroDeserializer
org.apache.paimon.shade.jackson2.com.fasterxml.jackson.databind.ObjectMapper objectMapper
org.apache.paimon.shade.jackson2.com.fasterxml.jackson.databind.ObjectMapper objectMapper
CdcTimestampExtractor timestampExtractor
long currentMaxTimestamp
int numChannels
FileStoreTable table
FileStoreTable table
org.apache.flink.api.common.typeinfo.TypeInformation<T> producedType
int fieldCount
int[] primaryKey
org.apache.flink.api.common.serialization.DeserializationSchema<T> primaryKeyDeserializer
org.apache.flink.api.common.serialization.DeserializationSchema<T> valueDeserializer
org.apache.flink.table.data.RowData.FieldGetter[] keyFieldGetters
int[][] projectFields
String topic
org.apache.flink.api.common.serialization.SerializationSchema<T> primaryKeySerializer
org.apache.flink.api.common.serialization.SerializationSchema<T> valueSerializer
CoreOptions.LogChangelogMode changelogMode
String topic
Properties properties
org.apache.flink.api.common.serialization.SerializationSchema<T> primaryKeySerializer
org.apache.flink.api.common.serialization.SerializationSchema<T> valueSerializer
CoreOptions.LogConsistency consistency
CoreOptions.LogChangelogMode changelogMode
String topic
Properties properties
org.apache.flink.table.types.DataType physicalType
int[] primaryKey
org.apache.flink.api.common.serialization.DeserializationSchema<T> primaryKeyDeserializer
org.apache.flink.api.common.serialization.DeserializationSchema<T> valueDeserializer
int[][] projectFields
CoreOptions.LogConsistency consistency
CoreOptions.StartupMode scanMode
Long timestampMills
LogSinkFunction.WriteCallback writeCallback
NewLookupFunction function
int threadNumber
Table table
java.time.Duration refreshInterval
Comparator<T> comparator
java.time.LocalDateTime lastRefresh
BinaryRow partition
Table table
DynamicPartitionLoader partitionLoader
List<E> projectFields
List<E> joinKeys
Predicate predicate
RefreshBlacklist refreshBlacklist
org.apache.flink.table.functions.FunctionContext functionContext
Filter<T> cacheRowFilter
LookupFileStoreTable.LookupStreamScanMode lookupScanMode
FileStoreLookupFunction function
FileStoreLookupFunction function
Integer parallelism
ServiceManager serviceManager
Table table
Table table
long monitorInterval
boolean isRunning
int totalRangeNum
RowType rowType
boolean isSortBySize
int numSample
FileStoreTable table
String commitUser
String commitUser
Catalog.Loader catalogLoader
CommitterOperator<CommitT,GlobalCommitT> commitOperator
SerializableSupplier<T> snapshotManagerFactory
SerializableSupplier<T> tagManagerFactory
SerializableSupplier<T> tagDeletionFactory
SerializableSupplier<T> callbacksSupplier
NavigableSet<E> identifiersForTags
java.time.Duration tagTimeRetained
CommitterOperator<CommitT,GlobalCommitT> commitOperator
FileStoreTable table
Catalog.Loader catalogLoader
boolean ignorePreviousFiles
Options options
Deque<E> inputs
boolean streamingCheckpointEnabled
CommitterOperator.endInput
.
Remaining data will be committed in CommitterOperator.notifyCheckpointComplete(long)
. If
checkpoint is not enabled we need to commit remaining data in CommitterOperator.endInput
.boolean forceSingleParallelism
String initialCommitUser
NavigableMap<K,V> committablesPerCheckpoint
Committer.Factory<CommitT,GlobalCommitT> committerFactory
CommittableStateManager<GlobalCommitT> committableStateManager
Committer<CommitT,GlobalCommitT> committer
Long endInputWatermark
LogSinkFunction logSinkFunction
FileStoreTable table
boolean ignorePreviousFiles
ChannelComputer<T> channelComputer
String initialCommitUser
FileStoreTable table
Integer numAssigners
SerializableFunction<T,R> extractorFunction
boolean overwrite
TableSchema schema
boolean ignoreDelete
StoreSinkWrite.Provider storeSinkWriteProvider
org.apache.flink.streaming.api.environment.CheckpointConfig checkpointConfig
boolean isStreaming
boolean ignorePreviousFiles
String initialCommitUser
Catalog.Loader catalogLoader
Catalog catalog
Map<K,V> tables
Map<K,V> writes
String commitUser
Options options
boolean endOfInput
SerializableSupplier<T> committableSerializer
org.apache.flink.api.common.state.ListState<T> streamingCommitterState
TableSchema schema
Integer numAssigners
TableSchema schema
boolean hasLogSink
LogSinkFunction logSinkFunction
long currentWatermark
InternalTimerService
.TableSchema schema
FileStoreTable table
StoreSinkWrite.Provider storeSinkWriteProvider
String initialCommitUser
FileStoreTable table
StoreSinkWrite.Provider storeSinkWriteProvider
String initialCommitUser
FileStoreTable table
FileStoreTable table
LogSinkFunction logSinkFunction
Integer parallelism
long retrySleepMillis
EventParser.Factory<T> parserFactory
String database
Catalog.Loader catalogLoader
TableSchema schema
Catalog.Loader catalogLoader
Map<K,V> channelComputers
EventParser.Factory<T> parserFactory
EventParser.Factory<T> parserFactory
StoreSinkWrite.WithWriteBufferProvider storeSinkWriteProvider
String initialCommitUser
Catalog.Loader catalogLoader
MemoryPoolFactory memoryPoolFactory
Catalog catalog
Map<K,V> tables
StoreSinkWriteState state
Map<K,V> writes
String commitUser
ExecutorService compactExecutor
long retrySleepMillis
Integer parallelism
boolean isOverwrite
Catalog.Loader catalogLoader
double commitCpuCores
MemorySize commitHeapMemory
boolean commitChaining
String commitUser
SchemaManager schemaManager
Identifier identifier
Catalog.Loader catalogLoader
Catalog catalog
boolean allowUpperCase
GlobalIndexAssigner assigner
IndexBootstrap bootstrap
SerializableFunction<T,R> converter
org.apache.flink.api.common.typeutils.TypeSerializer<T> keyPartSerializer
org.apache.flink.api.common.typeutils.TypeSerializer<T> rowSerializer
RowType keyType
RowType rowType
long maxMemory
int pageSize
int arity
int spillSortMaxNumFiles
CompressOptions spillCompression
int sinkParallelism
MemorySize maxDiskSize
boolean sequenceOrder
FileStoreTable table
FileStoreTable table
FileStoreTable table
boolean streaming
long scanInterval
Predicate filter
boolean isRunning
Map<K,V> options
BucketMode bucketMode
ReadBuilder readBuilder
Long limit
LogSourceProvider provider
long from
long to
org.apache.flink.api.common.eventtime.WatermarkStrategy<T> strategy
long lastConsumeSnapshotId
FileStoreTable table
Predicate partitionPredicate
int splitBatchSize
FlinkConnectorOptions.SplitAssignMode splitAssignMode
DynamicPartitionFilteringInfo dynamicPartitionFilteringInfo
long checkpointId
DataSplit dataSplit
MultiTableScanBase<T> tableScan
long monitorInterval
Catalog.Loader catalogLoader
Pattern includingPattern
Pattern excludingPattern
Pattern databasePattern
boolean isStreaming
long monitorInterval
MultiTableScanBase<T> tableScan
ReadBuilder readBuilder
long monitorInterval
boolean emitSnapshotWatermark
boolean isRunning
Catalog.Loader catalogLoader
boolean isStreaming
java.time.Duration partitionIdleTime
Catalog.Loader catalogLoader
java.time.Duration partitionIdleTime
ReadBuilder readBuilder
InternalRowSerializer internalRowSerializer
InternalTypeSerializer<T> serializer
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
OrcFilters.Predicate[] preds
OrcFilters.Predicate pred
OrcFilters.Predicate[] preds
org.apache.orc.TypeDescription schema
RowType rowType
org.apache.hadoop.conf.Configuration conf
URI uri
Options options
SerializableConfiguration hadoopConf
private void readObject(ObjectInputStream in) throws ClassNotFoundException, IOException
ClassNotFoundException
IOException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
SerializableHiveConf hiveConf
String clientClassName
Options options
Identifier identifier
TableSchema schema
SerializableHiveConf hiveConf
String clientClassName
Options options
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
byte[] serializedConf
InternalRowSerializer partSerializer
InternalMapSerializer nullValueCountsSerializer
InternalMapSerializer lowerBoundsSerializer
InternalMapSerializer upperBoundsSerializer
IcebergDataFileMetaSerializer fileSerializer
IcebergPartitionSummarySerializer partitionSummarySerializer
InternalRowSerializer rowSerializer
InternalRowSerializer rowSerializer
byte[] buffer
int end
int position
DataFileMetaSerializer dataFileMetaSerializer
int version
boolean distinct
InternalArray.ElementGetter elementGetter
java.util.function.BiFunction<T,U,R> equaliser
boolean initialized
boolean initialized
FieldAggregator aggregator
String delimiter
InternalArray.ElementGetter keyGetter
InternalArray.ElementGetter valueGetter
int nestedFields
Projection keyProjection
RecordEqualiser elementEqualiser
RoaringBitmap32 roaringBitmapAcc
RoaringBitmap32 roaringBitmapInput
RoaringBitmap64 roaringBitmapAcc
RoaringBitmap64 roaringBitmapInput
org.apache.paimon.metrics.DescriptiveStatisticsHistogram.CircularDoubleArray descriptiveStatistics
org.apache.paimon.metrics.DescriptiveStatisticsHistogramStatistics.CommonMetricsSnapshot statisticsSummary
ThreadPoolExecutor executor
List<E> deleteFiles
Set<E> candidateDeletes
CatalogLockFactory lockFactory
CatalogLockContext lockContext
Identifier tablePath
FileStoreTable table
FileIO fileIO
long olderThanMillis
SerializableConsumer<T> fileCleaner
int partitionKeysNum
Path location
long bytes
Options hadoopOptions
CompoundPredicate.Function function
List<E> children
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
LeafFunction function
DataType type
int fieldIndex
String fieldName
PrivilegeChecker privilegeChecker
Identifier identifier
Options hadoopOptions
String fieldName
DataType dataType
String description
SchemaChange.Move move
String fieldName
String fieldName
String referenceFieldName
SchemaChange.Move.MoveType type
String key
String[] fieldNames
boolean newNullability
SchemaChange.Move move
String comment
int version
long id
List<E> fields
int highestFieldId
List<E> partitionKeys
List<E> primaryKeys
List<E> bucketKeys
int numBucket
Map<K,V> options
String comment
long timeMillis
HilbertIndexer.HProcessFunction process
int fieldIndex
ZIndexer.ZProcessFunction process
int byteSize
int fieldIndex
int varTypeSize
byte[] nullVarBytes
private Object readResolve()
private Object readResolve()
private Object readResolve()
private Object readResolve()
scala.Option<A> filters
private Object readResolve()
private Object readResolve()
long value
private Object readResolve()
scala.collection.Seq<A> inputPartitions
ReadBuilder readBuilder
scala.collection.Seq<A> metadataColumns
private Object readResolve()
scala.collection.Seq<A> splits
int bucket
private Object readResolve()
OptionalLong nullCount
Optional<T> min
Optional<T> max
OptionalLong distinctCount
OptionalLong avgLen
OptionalLong maxLen
private Object readResolve()
private Object readResolve()
org.apache.spark.sql.catalyst.InternalRow[] rows
org.apache.spark.sql.types.StructType readSchema
Table table
org.apache.spark.sql.sources.Filter[] filters
private Object readResolve()
private Object readResolve()
long value
private Object readResolve()
scala.Function1<T1,R> readFunc
org.apache.paimon.spark.PaimonInputPartition partition
SparkInternalRow row
scala.collection.Seq<A> metadataColumns
scala.collection.Iterator<A> splits
org.apache.paimon.spark.PaimonRecordReaderIterator currentRecordReader
boolean advanced
InternalRow currentRow
private Object readResolve()
IOManager ioManager
SparkInternalRow row
ReadBuilder readBuilder
scala.collection.Seq<A> metadataColumns
byte bitmap$0
private Object readResolve()
private Object readResolve()
long value
private Object readResolve()
RecordReader<T> reader
scala.collection.Seq<A> metadataColumns
Split split
Path lastFilePath
boolean isFileRecordIterator
RecordReader.RecordIterator<T> currentIterator
boolean advanced
InternalRow currentResult
boolean needMetadata
GenericRow metadataRow
JoinedRow joinedRow
private Object readResolve()
private Object readResolve()
long value
private Object readResolve()
Table table
org.apache.spark.sql.types.StructType requiredSchema
scala.collection.Seq<A> filters
scala.collection.Seq<A> reservedFilters
scala.Option<A> pushDownLimit
boolean bucketedScanDisabled
private Object readResolve()
private Object readResolve()
long value
private Object readResolve()
private Object readResolve()
long value
private Object readResolve()
Table table
DataSplit[] dataSplits
org.apache.spark.sql.types.StructType requiredSchema
scala.collection.Seq<A> filters
CoreOptions coreOptions
org.apache.spark.sql.SparkSession org$apache$paimon$spark$ScanHelper$$spark
boolean org$apache$paimon$spark$ScanHelper$$deletionVectors
long org$apache$paimon$spark$ScanHelper$$openCostInBytes
int org$apache$paimon$spark$ScanHelper$$leafNodeDefaultParallelism
RowType tableRowType
org.apache.spark.sql.types.StructType tableSchema
scala.Tuple2<T1,T2> org$apache$paimon$spark$ColumnPruningAndPushDown$$x$2
RowType readTableRowType
org.apache.spark.sql.types.StructField[] metadataFields
ReadBuilder readBuilder
byte bitmap$0
private Object readResolve()
private Object readResolve()
long value
private Object readResolve()
long rowCount
long scannedTotalSize
Statistics paimonStats
boolean paimonStatsEnabled
org.apache.paimon.spark.PaimonBaseScan scan
byte bitmap$0
private Object readResolve()
private Object readResolve()
scala.collection.Seq<A> splits
private Object readResolve()
DataType elementType
InternalArray array
RowType rowType
InternalRow row
org.apache.spark.sql.types.StructType schema
Table table
RowType org$apache$paimon$spark$PaimonPartitionManagement$$partitionRowType
org.apache.spark.sql.types.StructType partitionSchema
byte bitmap$0
private Object readResolve()
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
org.apache.spark.sql.expressions.UserDefinedFunction hilbertCurveUDF
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
int numCols
int inputCol
int totalOutputBytes
int varTypeSize
int maxOutputSize
org.apache.spark.sql.expressions.UserDefinedFunction interleaveUDF
Identifier identifier
Lock.Factory lockFactory
MetastoreClient.Factory metastoreClientFactory
LineageMetaFactory lineageMetaFactory
FileStoreTable wrapped
FileStoreTable fallback
Identifier identifier
RowType rowType
List<E> partitionKeys
String location
FormatTable.Format format
Map<K,V> options
String comment
InnerTable origin
DataSplit[] splits
InnerTable table
String commitUser
Map<K,V> staticPartition
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
TableSchema schema
InnerTable table
String commitUser
private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException
IOException
ClassNotFoundException
private void writeObject(ObjectOutputStream out) throws IOException
IOException
long snapshotId
BinaryRow partition
int bucket
String bucketPath
List<E> beforeFiles
List<E> beforeDeletionFiles
List<E> dataFiles
List<E> dataDeletionFiles
boolean isStreaming
boolean rawConvertible
String path
long offset
long length
FileStoreTable wrapped
FileStoreTable storeTable
Options catalogOptions
FileStoreTable wrapped
boolean isContinuous
String databaseName
FileStoreTable wrapped
FileStoreTable storeTable
FileStoreTable dataTable
FileStoreTable storeTable
FileStoreTable wrapped
FileIO fileIO
Path location
String branch
FileStoreTable dataTable
FileStoreTable dataTable
LineageMetaFactory lineageMetaFactory
Options options
DataType elementType
int length
int length
boolean isNullable
DataTypeRoot typeRoot
int precision
int scale
int precision
DataType elementType
List<E> fields
InternalRow.FieldGetter[] fieldGetters
int precision
int precision
int length
int length
int keyArity
int valueArity
long value
InternalRowSerializer rowSerializer
int i1
int i2
InternalRow.FieldGetter[] fieldGetters
Map<K,V> projectMapping
InternalRow.FieldGetter[] projectGetters
RowType rowType
InternalRow.FieldGetter[] fieldGetters
org.apache.parquet.filter2.predicate.Operators.UserDefined<T extends Comparable<T>,U extends org.apache.parquet.filter2.predicate.UserDefinedPredicate<T>> udp
org.apache.parquet.filter2.predicate.FilterPredicate predicate
org.apache.parquet.filter2.predicate.Operators.Column<T extends Comparable<T>> column
Set<E> values
org.apache.parquet.filter2.predicate.Operators.Column<T extends Comparable<T>> column
org.apache.parquet.filter2.predicate.UserDefinedPredicate<T extends Comparable<T>> udpInstance
Copyright © 2023–2024 The Apache Software Foundation. All rights reserved.