Modifier and Type | Method and Description |
---|---|
Path |
CoreOptions.path() |
static Path |
CoreOptions.path(Map<String,String> options) |
static Path |
CoreOptions.path(Options options) |
Modifier and Type | Method and Description |
---|---|
static Changelog |
Changelog.fromPath(FileIO fileIO,
Path path) |
static Snapshot |
Snapshot.fromPath(FileIO fileIO,
Path path) |
static Snapshot |
Snapshot.tryFromPath(FileIO fileIO,
Path path) |
Modifier and Type | Method and Description |
---|---|
void |
FileStore.setManifestCache(SegmentsCache<Path> manifestCache) |
void |
FileStore.setSnapshotCache(org.apache.paimon.shade.caffeine2.com.github.benmanes.caffeine.cache.Cache<Path,Snapshot> cache) |
Modifier and Type | Field and Description |
---|---|
protected SegmentsCache<Path> |
CachingCatalog.manifestCache |
Modifier and Type | Method and Description |
---|---|
Path |
DelegateCatalog.getTableLocation(Identifier identifier) |
Path |
Catalog.getTableLocation(Identifier identifier)
Get the table location in this catalog.
|
Path |
AbstractCatalog.getTableLocation(Identifier identifier) |
Path |
AbstractCatalog.newDatabasePath(String database)
Get warehouse path for specified database.
|
static Path |
AbstractCatalog.newDatabasePath(String warehouse,
String database) |
static Path |
AbstractCatalog.newTableLocation(String warehouse,
Identifier identifier) |
static Path |
CatalogUtils.path(String warehouse,
String database,
String table) |
static Path |
CatalogFactory.warehouse(CatalogContext context) |
Modifier and Type | Method and Description |
---|---|
Map<String,Map<String,Path>> |
AbstractCatalog.allTablePaths() |
Modifier and Type | Method and Description |
---|---|
Catalog |
FileSystemCatalogFactory.create(FileIO fileIO,
Path warehouse,
CatalogContext context) |
default Catalog |
CatalogFactory.create(FileIO fileIO,
Path warehouse,
CatalogContext context) |
static CatalogContext |
CatalogContext.create(Path warehouse) |
static String |
CatalogUtils.database(Path path) |
protected List<String> |
AbstractCatalog.listDatabasesInFileSystem(Path warehouse) |
protected List<String> |
AbstractCatalog.listTablesInFileSystem(Path databasePath) |
static String |
CatalogUtils.table(Path path) |
protected boolean |
AbstractCatalog.tableExistsInFileSystem(Path tablePath,
String branchName) |
Optional<TableSchema> |
AbstractCatalog.tableSchemaInFileSystem(Path tablePath,
String branchName) |
Constructor and Description |
---|
FileSystemCatalog(FileIO fileIO,
Path warehouse) |
FileSystemCatalog(FileIO fileIO,
Path warehouse,
Options options) |
Modifier and Type | Method and Description |
---|---|
static Optional<Consumer> |
Consumer.fromPath(FileIO fileIO,
Path path) |
Constructor and Description |
---|
ConsumerManager(FileIO fileIO,
Path tablePath) |
ConsumerManager(FileIO fileIO,
Path tablePath,
String branchName) |
Modifier and Type | Method and Description |
---|---|
Path |
ColumnarRowIterator.filePath() |
Constructor and Description |
---|
ColumnarRowIterator(Path filePath,
ColumnarRow row,
Runnable recycler) |
Modifier and Type | Method and Description |
---|---|
Path |
ApplyDeletionFileRecordIterator.filePath() |
Constructor and Description |
---|
FileIndexPredicate(Path path,
FileIO fileIO,
RowType fileRowType) |
Modifier and Type | Method and Description |
---|---|
Path |
ApplyBitmapIndexFileRecordIterator.filePath() |
Modifier and Type | Method and Description |
---|---|
boolean |
FlinkFileIO.delete(Path path,
boolean recursive) |
boolean |
FlinkFileIO.exists(Path path) |
FileStatus |
FlinkFileIO.getFileStatus(Path path) |
FileStatus[] |
FlinkFileIO.listStatus(Path path) |
FileIO |
FlinkFileIOLoader.load(Path path) |
boolean |
FlinkFileIO.mkdirs(Path path) |
SeekableInputStream |
FlinkFileIO.newInputStream(Path path) |
PositionOutputStream |
FlinkFileIO.newOutputStream(Path path,
boolean overwrite) |
boolean |
FlinkFileIO.rename(Path src,
Path dst) |
Constructor and Description |
---|
FlinkFileIO(Path path) |
Modifier and Type | Method and Description |
---|---|
static List<Path> |
PickFilesUtil.getUsedFilesForLatestSnapshot(FileStoreTable table) |
Modifier and Type | Method and Description |
---|---|
static long |
FlinkOrphanFilesClean.executeDatabaseOrphanFiles(org.apache.flink.streaming.api.environment.StreamExecutionEnvironment env,
Catalog catalog,
long olderThanMillis,
SerializableConsumer<Path> fileCleaner,
Integer parallelism,
String databaseName,
String tableName) |
Constructor and Description |
---|
FlinkOrphanFilesClean(FileStoreTable table,
long olderThanMillis,
SerializableConsumer<Path> fileCleaner,
Integer parallelism) |
Modifier and Type | Method and Description |
---|---|
Path |
FormatReaderContext.filePath() |
Path |
FormatReaderFactory.Context.filePath() |
Modifier and Type | Method and Description |
---|---|
SimpleColStats[] |
SimpleStatsExtractor.extract(FileIO fileIO,
Path path) |
Pair<SimpleColStats[],SimpleStatsExtractor.FileInfo> |
SimpleStatsExtractor.extractWithFileInfo(FileIO fileIO,
Path path) |
Constructor and Description |
---|
FormatReaderContext(FileIO fileIO,
Path file,
long fileSize) |
FormatReaderContext(FileIO fileIO,
Path file,
long fileSize,
FileIndexResult fileIndexResult) |
OrcFormatReaderContext(FileIO fileIO,
Path filePath,
long fileSize,
int poolSize) |
Modifier and Type | Method and Description |
---|---|
SimpleColStats[] |
AvroSimpleStatsExtractor.extract(FileIO fileIO,
Path path) |
Pair<SimpleColStats[],SimpleStatsExtractor.FileInfo> |
AvroSimpleStatsExtractor.extractWithFileInfo(FileIO fileIO,
Path path) |
Modifier and Type | Method and Description |
---|---|
static org.apache.orc.Reader |
OrcReaderFactory.createReader(org.apache.hadoop.conf.Configuration conf,
FileIO fileIO,
Path path,
FileIndexResult fileIndexResult) |
org.apache.paimon.format.orc.OrcReaderFactory.OrcReaderBatch |
OrcReaderFactory.createReaderBatch(Path filePath,
org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch orcBatch,
Pool.Recycler<org.apache.paimon.format.orc.OrcReaderFactory.OrcReaderBatch> recycler)
Creates the
OrcReaderBatch structure, which is responsible for holding the data
structures that hold the batch data (column vectors, row arrays, ...) and the batch
conversion from the ORC representation to the result format. |
Modifier and Type | Method and Description |
---|---|
SimpleColStats[] |
OrcSimpleStatsExtractor.extract(FileIO fileIO,
Path path) |
Pair<SimpleColStats[],SimpleStatsExtractor.FileInfo> |
OrcSimpleStatsExtractor.extractWithFileInfo(FileIO fileIO,
Path path) |
Modifier and Type | Method and Description |
---|---|
Path |
ParquetInputFile.getPath() |
Modifier and Type | Method and Description |
---|---|
SimpleColStats[] |
ParquetSimpleStatsExtractor.extract(FileIO fileIO,
Path path) |
static Pair<Map<String,org.apache.parquet.column.statistics.Statistics<?>>,SimpleStatsExtractor.FileInfo> |
ParquetUtil.extractColumnStats(FileIO fileIO,
Path path)
Extract stats from specified Parquet files path.
|
Pair<SimpleColStats[],SimpleStatsExtractor.FileInfo> |
ParquetSimpleStatsExtractor.extractWithFileInfo(FileIO fileIO,
Path path) |
static ParquetInputFile |
ParquetInputFile.fromPath(FileIO fileIO,
Path path) |
static ParquetFileReader |
ParquetUtil.getParquetReader(FileIO fileIO,
Path path,
FileIndexResult fileIndexResult)
Generate
ParquetFileReader instance to read the Parquet files at the given path. |
Modifier and Type | Method and Description |
---|---|
Path |
Path.createTempPath()
Create a temporary path (to be used as a copy) for this path.
|
Path |
Path.getParent()
Returns the parent of a path or null if at root.
|
Path |
FileStatus.getPath()
Returns the corresponding Path to the FileStatus.
|
Modifier and Type | Method and Description |
---|---|
static FileIOLoader |
FileIOUtils.checkAccess(FileIOLoader fileIO,
Path path,
CatalogContext config) |
default void |
FileIO.checkOrMkdirs(Path path) |
int |
Path.compareTo(Path that) |
default void |
FileIO.copyFile(Path sourcePath,
Path targetPath,
boolean overwrite)
Copy content of one file into another.
|
default void |
FileIO.copyFiles(Path sourceDirectory,
Path targetDirectory,
boolean overwrite)
Copy all files in sourceDirectory to directory targetDirectory.
|
protected abstract FileIO |
PluginFileIO.createFileIO(Path path) |
boolean |
PluginFileIO.delete(Path path,
boolean recursive) |
boolean |
FileIO.delete(Path path,
boolean recursive)
Delete a file.
|
default void |
FileIO.deleteDirectoryQuietly(Path directory) |
default void |
FileIO.deleteQuietly(Path file) |
boolean |
PluginFileIO.exists(Path path) |
boolean |
FileIO.exists(Path path)
Check if exists.
|
static FileIO |
FileIO.get(Path path,
CatalogContext config)
Returns a reference to the
FileIO instance for accessing the file system identified
by the given path. |
default long |
FileIO.getFileSize(Path path) |
FileStatus |
PluginFileIO.getFileStatus(Path path) |
FileStatus |
FileIO.getFileStatus(Path path)
Return a file status object that represents the path.
|
default boolean |
FileIO.isDir(Path path) |
default FileStatus[] |
FileIO.listDirectories(Path path)
List the statuses of the directories in the given path if the path is a directory.
|
FileStatus[] |
PluginFileIO.listStatus(Path path) |
FileStatus[] |
FileIO.listStatus(Path path)
List the statuses of the files/directories in the given path if the path is a directory.
|
FileIO |
FileIOLoader.load(Path path) |
boolean |
PluginFileIO.mkdirs(Path path) |
boolean |
FileIO.mkdirs(Path path)
Make the given file and all non-existent parents into directories.
|
SeekableInputStream |
PluginFileIO.newInputStream(Path path) |
SeekableInputStream |
FileIO.newInputStream(Path path)
Opens an SeekableInputStream at the indicated Path.
|
PositionOutputStream |
PluginFileIO.newOutputStream(Path path,
boolean overwrite) |
PositionOutputStream |
FileIO.newOutputStream(Path path,
boolean overwrite)
Opens an PositionOutputStream at the indicated Path.
|
default void |
FileIO.overwriteFileUtf8(Path path,
String content)
Overwrite file by content atomically, different
FileIO s have different atomic
implementations. |
default String |
FileIO.readFileUtf8(Path path)
Read file to UTF_8 decoding.
|
default Optional<String> |
FileIO.readOverwrittenFileUtf8(Path path)
Read file from
FileIO.overwriteFileUtf8(org.apache.paimon.fs.Path, java.lang.String) file. |
boolean |
PluginFileIO.rename(Path src,
Path dst) |
boolean |
FileIO.rename(Path src,
Path dst)
Renames the file/directory src to dst.
|
default boolean |
FileIO.tryToWriteAtomic(Path path,
String content)
Write content to one file atomically, initially writes to temp hidden file and only renames
to the target file once temp file is closed.
|
default void |
FileIO.writeFile(Path path,
String content,
boolean overwrite) |
Modifier and Type | Method and Description |
---|---|
default void |
FileIO.deleteFilesQuietly(List<Path> files) |
Constructor and Description |
---|
Path(Path parent,
Path child)
Create a new Path based on the child path resolved against the parent path.
|
Path(Path parent,
String child)
Create a new Path based on the child path resolved against the parent path.
|
Path(String parent,
Path child)
Create a new Path based on the child path resolved against the parent path.
|
Modifier and Type | Method and Description |
---|---|
boolean |
HadoopFileIO.delete(Path path,
boolean recursive) |
boolean |
HadoopFileIO.exists(Path path) |
FileStatus |
HadoopFileIO.getFileStatus(Path path) |
FileStatus[] |
HadoopFileIO.listStatus(Path path) |
HadoopFileIO |
HadoopViewFsFileIOLoader.load(Path path) |
HadoopFileIO |
HadoopFileIOLoader.load(Path path) |
boolean |
HadoopFileIO.mkdirs(Path path) |
SeekableInputStream |
HadoopFileIO.newInputStream(Path path) |
PositionOutputStream |
HadoopFileIO.newOutputStream(Path path,
boolean overwrite) |
void |
HadoopFileIO.overwriteFileUtf8(Path path,
String content) |
boolean |
HadoopFileIO.rename(Path src,
Path dst) |
void |
HadoopFileIO.setFileSystem(Path path,
org.apache.hadoop.fs.FileSystem fs) |
boolean |
HadoopFileIO.tryAtomicOverwriteViaRename(Path dst,
String content) |
Modifier and Type | Method and Description |
---|---|
void |
LocalFileIO.copyFile(Path sourcePath,
Path targetPath,
boolean overwrite) |
boolean |
LocalFileIO.delete(Path path,
boolean recursive) |
boolean |
LocalFileIO.exists(Path path) |
FileStatus |
LocalFileIO.getFileStatus(Path path) |
FileStatus[] |
LocalFileIO.listStatus(Path path) |
LocalFileIO |
LocalFileIOLoader.load(Path path) |
boolean |
LocalFileIO.mkdirs(Path path) |
SeekableInputStream |
LocalFileIO.newInputStream(Path path) |
PositionOutputStream |
LocalFileIO.newOutputStream(Path path,
boolean overwrite) |
boolean |
LocalFileIO.rename(Path src,
Path dst) |
File |
LocalFileIO.toFile(Path path)
Converts the given Path to a File for this file system.
|
Modifier and Type | Method and Description |
---|---|
Path |
HiveCatalog.getTableLocation(Identifier identifier) |
Modifier and Type | Method and Description |
---|---|
void |
StorageLocationHelper.createPathIfRequired(Path dbPath,
FileIO fileIO) |
void |
LocationHelper.createPathIfRequired(Path dbPath,
FileIO fileIO) |
void |
TBPropertiesLocationHelper.createPathIfRequired(Path path,
FileIO fileIO) |
void |
StorageLocationHelper.dropPathIfRequired(Path path,
FileIO fileIO) |
void |
LocationHelper.dropPathIfRequired(Path path,
FileIO fileIO) |
void |
TBPropertiesLocationHelper.dropPathIfRequired(Path path,
FileIO fileIO) |
void |
StorageLocationHelper.specifyDatabaseLocation(Path path,
org.apache.hadoop.hive.metastore.api.Database database) |
void |
LocationHelper.specifyDatabaseLocation(Path path,
org.apache.hadoop.hive.metastore.api.Database database) |
void |
TBPropertiesLocationHelper.specifyDatabaseLocation(Path path,
org.apache.hadoop.hive.metastore.api.Database database) |
Modifier and Type | Method and Description |
---|---|
HiveMigrator.MigrateTask |
HiveMigrator.importUnPartitionedTableTask(FileIO fileIO,
org.apache.hadoop.hive.metastore.api.Table sourceTable,
FileStoreTable paimonTable,
Map<Path,Path> rollback) |
HiveMigrator.MigrateTask |
HiveMigrator.importUnPartitionedTableTask(FileIO fileIO,
org.apache.hadoop.hive.metastore.api.Table sourceTable,
FileStoreTable paimonTable,
Map<Path,Path> rollback) |
Constructor and Description |
---|
MigrateTask(FileIO fileIO,
String format,
String location,
FileStoreTable paimonTable,
BinaryRow partitionRow,
Path newDir,
Map<Path,Path> rollback) |
Constructor and Description |
---|
MigrateTask(FileIO fileIO,
String format,
String location,
FileStoreTable paimonTable,
BinaryRow partitionRow,
Path newDir,
Map<Path,Path> rollback) |
MigrateTask(FileIO fileIO,
String format,
String location,
FileStoreTable paimonTable,
BinaryRow partitionRow,
Path newDir,
Map<Path,Path> rollback) |
Modifier and Type | Method and Description |
---|---|
Path |
IcebergPathFactory.metadataDirectory() |
Path |
IcebergPathFactory.newManifestFile() |
Path |
IcebergPathFactory.newManifestListFile() |
Path |
IcebergPathFactory.toManifestFilePath(String manifestFileName) |
Path |
IcebergPathFactory.toManifestListPath(String manifestListName) |
Path |
IcebergPathFactory.toMetadataPath(long snapshotId) |
Modifier and Type | Method and Description |
---|---|
java.util.stream.Stream<Path> |
IcebergPathFactory.getAllMetadataPathBefore(FileIO fileIO,
long snapshotId) |
Modifier and Type | Method and Description |
---|---|
void |
IcebergHiveMetadataCommitter.commitMetadata(Path newMetadataPath,
Path baseMetadataPath) |
void |
IcebergMetadataCommitter.commitMetadata(Path newMetadataPath,
Path baseMetadataPath) |
Constructor and Description |
---|
IcebergPathFactory(Path metadataDirectory) |
Modifier and Type | Method and Description |
---|---|
static IcebergMetadata |
IcebergMetadata.fromPath(FileIO fileIO,
Path path) |
Modifier and Type | Method and Description |
---|---|
Path |
IndexFileHandler.filePath(IndexFileMeta file) |
Path |
HashIndexFile.path(String fileName) |
Modifier and Type | Field and Description |
---|---|
protected Path |
SingleFileWriter.path |
Modifier and Type | Method and Description |
---|---|
static Path |
DataFilePathFactory.createNewFileIndexFilePath(Path filePath) |
static Path |
DataFilePathFactory.dataFileToFileIndexPath(Path dataFilePath) |
Path |
DataFilePathFactory.newChangelogPath() |
Path |
KeyValueFileWriterFactory.newChangelogPath(int level) |
Path |
DataFilePathFactory.newPath() |
Path |
SingleFileWriter.path() |
Path |
DataFilePathFactory.toPath(String fileName) |
Modifier and Type | Method and Description |
---|---|
List<Path> |
DataFileMeta.collectFiles(DataFilePathFactory pathFactory) |
Modifier and Type | Method and Description |
---|---|
static DataFileIndexWriter |
DataFileIndexWriter.create(FileIO fileIO,
Path path,
RowType rowType,
FileIndexOptions fileIndexOptions) |
static DataFileIndexWriter |
DataFileIndexWriter.create(FileIO fileIO,
Path path,
RowType rowType,
FileIndexOptions fileIndexOptions,
Map<String,String> colNameMapping) |
static Path |
DataFilePathFactory.createNewFileIndexFilePath(Path filePath) |
static Path |
DataFilePathFactory.dataFileToFileIndexPath(Path dataFilePath) |
Constructor and Description |
---|
DataFileIndexWriter(FileIO fileIO,
Path path,
RowType rowType,
FileIndexOptions fileIndexOptions,
Map<String,String> colNameMapping) |
DataFilePathFactory(Path parent,
String formatIdentifier,
String dataFilePrefix,
String changelogFilePrefix,
boolean fileSuffixIncludeCompression,
String fileCompression) |
KeyValueDataFileWriter(FileIO fileIO,
FormatWriterFactory factory,
Path path,
java.util.function.Function<KeyValue,InternalRow> converter,
RowType keyType,
RowType valueType,
SimpleStatsExtractor simpleStatsExtractor,
long schemaId,
int level,
String compression,
CoreOptions options,
FileSource fileSource,
FileIndexOptions fileIndexOptions) |
RowDataFileWriter(FileIO fileIO,
FormatWriterFactory factory,
Path path,
RowType writeSchema,
SimpleStatsExtractor simpleStatsExtractor,
long schemaId,
LongCounter seqNumCounter,
String fileCompression,
SimpleColStatsCollector.Factory[] statsCollectors,
FileIndexOptions fileIndexOptions,
FileSource fileSource,
boolean asyncFileWrite,
boolean statsDenseStore) |
SingleFileWriter(FileIO fileIO,
FormatWriterFactory factory,
Path path,
java.util.function.Function<T,InternalRow> converter,
String compression,
boolean asyncWrite) |
StatsCollectingSingleFileWriter(FileIO fileIO,
FormatWriterFactory factory,
Path path,
java.util.function.Function<T,InternalRow> converter,
RowType writeSchema,
SimpleStatsExtractor simpleStatsExtractor,
String compression,
SimpleColStatsCollector.Factory[] statsCollectors,
boolean asyncWrite) |
Modifier and Type | Method and Description |
---|---|
Catalog |
JdbcCatalogFactory.create(FileIO fileIO,
Path warehouse,
CatalogContext context) |
Constructor and Description |
---|
Factory(FileIO fileIO,
FileFormat fileFormat,
String compression,
FileStorePathFactory pathFactory,
SegmentsCache<Path> cache) |
Factory(FileIO fileIO,
FileFormat fileFormat,
String compression,
FileStorePathFactory pathFactory,
SegmentsCache<Path> cache) |
Factory(FileIO fileIO,
SchemaManager schemaManager,
RowType partitionType,
FileFormat fileFormat,
String compression,
FileStorePathFactory pathFactory,
long suggestedFileSize,
SegmentsCache<Path> cache) |
Modifier and Type | Method and Description |
---|---|
static List<DataFileMeta> |
FileMetaUtils.construct(FileIO fileIO,
String format,
String location,
Table paimonTable,
java.util.function.Predicate<FileStatus> filter,
Path dir,
Map<Path,Path> rollback) |
Modifier and Type | Method and Description |
---|---|
static List<DataFileMeta> |
FileMetaUtils.construct(FileIO fileIO,
String format,
String location,
Table paimonTable,
java.util.function.Predicate<FileStatus> filter,
Path dir,
Map<Path,Path> rollback) |
static List<DataFileMeta> |
FileMetaUtils.construct(FileIO fileIO,
String format,
String location,
Table paimonTable,
java.util.function.Predicate<FileStatus> filter,
Path dir,
Map<Path,Path> rollback) |
Modifier and Type | Field and Description |
---|---|
protected Path |
OrphanFilesClean.location |
Modifier and Type | Field and Description |
---|---|
protected SerializableConsumer<Path> |
OrphanFilesClean.fileCleaner |
Modifier and Type | Method and Description |
---|---|
List<Path> |
LocalOrphanFilesClean.clean() |
static SerializableConsumer<Path> |
OrphanFilesClean.createFileCleaner(Catalog catalog,
Boolean dryRun) |
protected List<Path> |
OrphanFilesClean.listPaimonFileDirs()
List directories that contains data files and manifest files.
|
Modifier and Type | Method and Description |
---|---|
protected List<FileStatus> |
OrphanFilesClean.tryBestListingDirs(Path dir)
If failed to list directory, just return an empty result because it's OK to not delete them.
|
Modifier and Type | Method and Description |
---|---|
protected void |
OrphanFilesClean.cleanSnapshotDir(List<String> branches,
java.util.function.Consumer<Path> deletedFileConsumer) |
static List<LocalOrphanFilesClean> |
LocalOrphanFilesClean.createOrphanFilesCleans(Catalog catalog,
String databaseName,
String tableName,
long olderThanMillis,
SerializableConsumer<Path> fileCleaner,
Integer parallelism) |
protected void |
FileDeletionBase.doCleanUnusedDataFile(Map<Path,Pair<ManifestEntry,List<Path>>> dataFileToDelete,
java.util.function.Predicate<ManifestEntry> skipper) |
protected void |
FileDeletionBase.doCleanUnusedDataFile(Map<Path,Pair<ManifestEntry,List<Path>>> dataFileToDelete,
java.util.function.Predicate<ManifestEntry> skipper) |
static long |
LocalOrphanFilesClean.executeDatabaseOrphanFiles(Catalog catalog,
String databaseName,
String tableName,
long olderThanMillis,
SerializableConsumer<Path> fileCleaner,
Integer parallelism) |
protected void |
FileDeletionBase.getDataFileToDelete(Map<Path,Pair<ManifestEntry,List<Path>>> dataFileToDelete,
List<ManifestEntry> dataFileEntries) |
protected void |
FileDeletionBase.getDataFileToDelete(Map<Path,Pair<ManifestEntry,List<Path>>> dataFileToDelete,
List<ManifestEntry> dataFileEntries) |
Constructor and Description |
---|
LocalOrphanFilesClean(FileStoreTable table,
long olderThanMillis,
SerializableConsumer<Path> fileCleaner) |
OrphanFilesClean(FileStoreTable table,
long olderThanMillis,
SerializableConsumer<Path> fileCleaner) |
Modifier and Type | Method and Description |
---|---|
boolean |
HadoopCompliantFileIO.delete(Path path,
boolean recursive) |
boolean |
HadoopCompliantFileIO.exists(Path path) |
FileStatus |
HadoopCompliantFileIO.getFileStatus(Path path) |
FileStatus[] |
HadoopCompliantFileIO.listStatus(Path path) |
FileIO |
OSSLoader.load(Path path) |
boolean |
HadoopCompliantFileIO.mkdirs(Path path) |
SeekableInputStream |
HadoopCompliantFileIO.newInputStream(Path path) |
PositionOutputStream |
HadoopCompliantFileIO.newOutputStream(Path path,
boolean overwrite) |
boolean |
HadoopCompliantFileIO.rename(Path src,
Path dst) |
Modifier and Type | Method and Description |
---|---|
static SuccessFile |
SuccessFileMarkDoneAction.safelyFromPath(FileIO fileIO,
Path path) |
Constructor and Description |
---|
SuccessFileMarkDoneAction(FileIO fileIO,
Path tablePath) |
Modifier and Type | Method and Description |
---|---|
static SuccessFile |
SuccessFile.fromPath(FileIO fileIO,
Path path) |
static SuccessFile |
SuccessFile.safelyFromPath(FileIO fileIO,
Path path) |
Modifier and Type | Method and Description |
---|---|
void |
PrivilegedFileStore.setManifestCache(SegmentsCache<Path> manifestCache) |
void |
PrivilegedFileStore.setSnapshotCache(org.apache.paimon.shade.caffeine2.com.github.benmanes.caffeine.cache.Cache<Path,Snapshot> cache) |
Modifier and Type | Method and Description |
---|---|
Path |
FileRecordIterator.filePath() |
Modifier and Type | Method and Description |
---|---|
boolean |
HadoopCompliantFileIO.delete(Path path,
boolean recursive) |
boolean |
HadoopCompliantFileIO.exists(Path path) |
FileStatus |
HadoopCompliantFileIO.getFileStatus(Path path) |
FileStatus[] |
HadoopCompliantFileIO.listStatus(Path path) |
FileIO |
S3Loader.load(Path path) |
boolean |
HadoopCompliantFileIO.mkdirs(Path path) |
SeekableInputStream |
HadoopCompliantFileIO.newInputStream(Path path) |
PositionOutputStream |
HadoopCompliantFileIO.newOutputStream(Path path,
boolean overwrite) |
boolean |
HadoopCompliantFileIO.rename(Path src,
Path dst) |
Modifier and Type | Method and Description |
---|---|
Path |
SchemaManager.schemaDirectory() |
Path |
SchemaManager.toSchemaPath(long schemaId) |
Modifier and Type | Method and Description |
---|---|
List<Path> |
SchemaManager.schemaPaths(java.util.function.Predicate<Long> predicate) |
Modifier and Type | Method and Description |
---|---|
static TableSchema |
TableSchema.fromPath(FileIO fileIO,
Path path) |
static TableSchema |
TableSchema.tryFromPath(FileIO fileIO,
Path path) |
Constructor and Description |
---|
SchemaManager(FileIO fileIO,
Path tableRoot) |
SchemaManager(FileIO fileIO,
Path tableRoot,
String branch)
Specify the default branch for data writing.
|
Modifier and Type | Method and Description |
---|---|
Path |
ServiceManager.tablePath() |
Constructor and Description |
---|
ServiceManager(FileIO fileIO,
Path tablePath) |
Modifier and Type | Method and Description |
---|---|
static Statistics |
Statistics.fromPath(FileIO fileIO,
Path path) |
Modifier and Type | Method and Description |
---|---|
Path |
DataTable.location() |
Path |
DelegatedFileStoreTable.location() |
Modifier and Type | Method and Description |
---|---|
static FileStoreTable |
FileStoreTableFactory.create(FileIO fileIO,
Path path) |
static FileStoreTable |
FileStoreTableFactory.create(FileIO fileIO,
Path tablePath,
TableSchema tableSchema) |
static FileStoreTable |
FileStoreTableFactory.create(FileIO fileIO,
Path tablePath,
TableSchema tableSchema,
CatalogEnvironment catalogEnvironment) |
static FileStoreTable |
FileStoreTableFactory.create(FileIO fileIO,
Path tablePath,
TableSchema tableSchema,
Options dynamicOptions,
CatalogEnvironment catalogEnvironment) |
static FileStoreTable |
FileStoreTableFactory.createWithoutFallbackBranch(FileIO fileIO,
Path tablePath,
TableSchema tableSchema,
Options dynamicOptions,
CatalogEnvironment catalogEnvironment) |
Modifier and Type | Method and Description |
---|---|
void |
FileStoreTable.setManifestCache(SegmentsCache<Path> manifestCache) |
void |
DelegatedFileStoreTable.setManifestCache(SegmentsCache<Path> manifestCache) |
void |
FileStoreTable.setSnapshotCache(org.apache.paimon.shade.caffeine2.com.github.benmanes.caffeine.cache.Cache<Path,Snapshot> cache) |
void |
DelegatedFileStoreTable.setSnapshotCache(org.apache.paimon.shade.caffeine2.com.github.benmanes.caffeine.cache.Cache<Path,Snapshot> cache) |
Modifier and Type | Method and Description |
---|---|
Path |
FileMonitorTable.location() |
Path |
CompactBucketsTable.location() |
Path |
AuditLogTable.location() |
Path |
ReadOptimizedTable.location() |
Modifier and Type | Method and Description |
---|---|
static Table |
SystemTableLoader.loadGlobal(String tableName,
FileIO fileIO,
java.util.function.Supplier<Map<String,Map<String,Path>>> allTablePaths,
Options catalogOptions,
LineageMetaFactory lineageMetaFactory) |
protected static Map<String,Map<String,Map<String,String>>> |
AllTableOptionsTable.options(FileIO fileIO,
Map<String,Map<String,Path>> allTablePaths) |
Constructor and Description |
---|
AllTableOptionsTable(FileIO fileIO,
Map<String,Map<String,Path>> allTablePaths) |
Modifier and Type | Method and Description |
---|---|
static Tag |
Tag.fromPath(FileIO fileIO,
Path path) |
static Tag |
Tag.tryFromPath(FileIO fileIO,
Path path) |
Modifier and Type | Method and Description |
---|---|
Path |
BranchManager.branchDirectory()
Return the root Directory of branch.
|
Path |
BranchManager.branchPath(String branchName)
Return the path of a branch.
|
Path |
FileStorePathFactory.bucketPath(BinaryRow partition,
int bucket) |
Path |
SnapshotManager.changelogDirectory() |
Path |
IteratorResultIterator.filePath() |
Path |
SnapshotManager.longLivedChangelogPath(long snapshotId) |
Path |
FileStorePathFactory.newManifestFile() |
Path |
FileStorePathFactory.newManifestList() |
Path |
PathFactory.newPath() |
Path |
FileStorePathFactory.relativePartitionAndBucketPath(BinaryRow partition,
int bucket) |
Path |
FileStorePathFactory.root() |
Path |
SnapshotManager.snapshotDirectory() |
Path |
SnapshotManager.snapshotPath(long snapshotId) |
Path |
SnapshotManager.tablePath() |
Path |
TagManager.tagDirectory()
Return the root Directory of tags.
|
Path |
TagManager.tagPath(String tagName)
Return the path of a tag.
|
Path |
FileStorePathFactory.toManifestFilePath(String manifestFileName) |
Path |
FileStorePathFactory.toManifestListPath(String manifestListName) |
Path |
PathFactory.toPath(String fileName) |
Modifier and Type | Method and Description |
---|---|
List<Path> |
FileStorePathFactory.getHierarchicalPartitionPath(BinaryRow partition) |
List<Path> |
SnapshotManager.snapshotPaths(java.util.function.Predicate<Long> predicate) |
List<Path> |
TagManager.tagPaths(java.util.function.Predicate<Path> predicate) |
List<Path> |
SnapshotManager.tryGetNonChangelogFiles(java.util.function.Predicate<FileStatus> fileStatusFilter) |
List<Path> |
SnapshotManager.tryGetNonSnapshotFiles(java.util.function.Predicate<FileStatus> fileStatusFilter)
Try to get non snapshot files.
|
Modifier and Type | Method and Description |
---|---|
static String |
BranchManager.branchPath(Path tablePath,
String branch)
Return the path string of a branch.
|
static void |
FileUtils.checkExists(FileIO fileIO,
Path file) |
static RecordReader<InternalRow> |
FileUtils.createFormatReader(FileIO fileIO,
FormatReaderFactory format,
Path file,
Long fileSize) |
static LinkedHashMap<String,String> |
PartitionPathUtils.extractPartitionSpecFromPath(Path currPath)
Make partition spec from path.
|
boolean |
BranchManager.fileExists(Path path)
Check if path exists.
|
static java.util.stream.Stream<String> |
FileUtils.listOriginalVersionedFiles(FileIO fileIO,
Path dir,
String prefix)
List original versioned files for the directory.
|
static java.util.stream.Stream<FileStatus> |
FileUtils.listVersionedDirectories(FileIO fileIO,
Path dir,
String prefix)
List versioned directories for the directory.
|
static java.util.stream.Stream<Long> |
FileUtils.listVersionedFiles(FileIO fileIO,
Path dir,
String prefix)
List versioned files for the directory.
|
static java.util.stream.Stream<FileStatus> |
FileUtils.listVersionedFileStatus(FileIO fileIO,
Path dir,
String prefix)
List versioned file status for the directory.
|
Long |
SnapshotManager.readHint(String fileName,
Path dir) |
static IntIterator |
IntFileUtils.readInts(FileIO fileIO,
Path path) |
static void |
IntFileUtils.writeInts(FileIO fileIO,
Path path,
IntIterator input) |
Modifier and Type | Method and Description |
---|---|
List<Path> |
TagManager.tagPaths(java.util.function.Predicate<Path> predicate) |
Constructor and Description |
---|
BranchManager(FileIO fileIO,
Path path,
SnapshotManager snapshotManager,
TagManager tagManager,
SchemaManager schemaManager) |
FileStorePathFactory(Path root,
RowType partitionType,
String defaultPartValue,
String formatIdentifier,
String dataFilePrefix,
String changelogFilePrefix,
boolean legacyPartitionName,
boolean fileSuffixIncludeCompression,
String fileCompression) |
IteratorResultIterator(IteratorWithException<InternalRow,IOException> records,
Runnable recycler,
Path filePath,
long pos) |
SnapshotManager(FileIO fileIO,
Path tablePath) |
SnapshotManager(FileIO fileIO,
Path tablePath,
String branchName)
Specify the default branch for data writing.
|
SnapshotManager(FileIO fileIO,
Path tablePath,
String branchName,
org.apache.paimon.shade.caffeine2.com.github.benmanes.caffeine.cache.Cache<Path,Snapshot> cache) |
TagManager(FileIO fileIO,
Path tablePath) |
TagManager(FileIO fileIO,
Path tablePath,
String branch)
Specify the default branch for data writing.
|
Constructor and Description |
---|
ObjectsFile(FileIO fileIO,
ObjectSerializer<T> serializer,
RowType formatType,
FormatReaderFactory readerFactory,
FormatWriterFactory writerFactory,
String compression,
PathFactory pathFactory,
SegmentsCache<Path> cache) |
SnapshotManager(FileIO fileIO,
Path tablePath,
String branchName,
org.apache.paimon.shade.caffeine2.com.github.benmanes.caffeine.cache.Cache<Path,Snapshot> cache) |
Copyright © 2023–2024 The Apache Software Foundation. All rights reserved.