From 07b82517bff638de258e79284afe0831162504f5 Mon Sep 17 00:00:00 2001 From: Mehul Batra Date: Fri, 19 Sep 2025 02:29:58 +0530 Subject: [PATCH 1/4] added javadoc for public modules --- pom.xml | 131 + tools/refresh_javadoc.sh | 62 + .../0.8-SNAPSHOT/allclasses-index.html | 1627 +++ .../javadoc/0.8-SNAPSHOT/allclasses.html | 268 + .../0.8-SNAPSHOT/allpackages-index.html | 350 + website/static/javadoc/0.8-SNAPSHOT/argfile | 1401 ++ .../javadoc/0.8-SNAPSHOT/constant-values.html | 572 + .../javadoc/0.8-SNAPSHOT/deprecated-list.html | 175 + .../static/javadoc/0.8-SNAPSHOT/element-list | 47 + .../0.8-SNAPSHOT/index-files/index-1.html | 369 + .../0.8-SNAPSHOT/index-files/index-10.html | 157 + .../0.8-SNAPSHOT/index-files/index-11.html | 187 + .../0.8-SNAPSHOT/index-files/index-12.html | 524 + .../0.8-SNAPSHOT/index-files/index-13.html | 195 + .../0.8-SNAPSHOT/index-files/index-14.html | 219 + .../0.8-SNAPSHOT/index-files/index-15.html | 340 + .../0.8-SNAPSHOT/index-files/index-16.html | 314 + .../0.8-SNAPSHOT/index-files/index-17.html | 153 + .../0.8-SNAPSHOT/index-files/index-18.html | 333 + .../0.8-SNAPSHOT/index-files/index-19.html | 536 + .../0.8-SNAPSHOT/index-files/index-2.html | 248 + .../0.8-SNAPSHOT/index-files/index-20.html | 547 + .../0.8-SNAPSHOT/index-files/index-21.html | 222 + .../0.8-SNAPSHOT/index-files/index-22.html | 222 + .../0.8-SNAPSHOT/index-files/index-23.html | 232 + .../0.8-SNAPSHOT/index-files/index-3.html | 561 + .../0.8-SNAPSHOT/index-files/index-4.html | 334 + .../0.8-SNAPSHOT/index-files/index-5.html | 236 + .../0.8-SNAPSHOT/index-files/index-6.html | 548 + .../0.8-SNAPSHOT/index-files/index-7.html | 659 + .../0.8-SNAPSHOT/index-files/index-8.html | 238 + .../0.8-SNAPSHOT/index-files/index-9.html | 327 + .../static/javadoc/0.8-SNAPSHOT/index.html | 346 + .../static/javadoc/0.8-SNAPSHOT/javadoc.sh | 1 + .../0.8-SNAPSHOT/jquery-ui.overrides.css | 35 + .../jquery/external/jquery/jquery.js | 10872 +++++++++++++++ .../0.8-SNAPSHOT/jquery/jquery-3.7.1.min.js | 2 + .../0.8-SNAPSHOT/jquery/jquery-ui.min.css | 6 + .../0.8-SNAPSHOT/jquery/jquery-ui.min.js | 6 + .../jquery/jszip-utils/dist/jszip-utils-ie.js | 56 + .../jszip-utils/dist/jszip-utils-ie.min.js | 10 + .../jquery/jszip-utils/dist/jszip-utils.js | 118 + .../jszip-utils/dist/jszip-utils.min.js | 10 + .../0.8-SNAPSHOT/jquery/jszip/dist/jszip.js | 11370 ++++++++++++++++ .../jquery/jszip/dist/jszip.min.js | 13 + .../legal/ADDITIONAL_LICENSE_INFO | 37 + .../0.8-SNAPSHOT/legal/ASSEMBLY_EXCEPTION | 27 + .../static/javadoc/0.8-SNAPSHOT/legal/LICENSE | 347 + .../javadoc/0.8-SNAPSHOT/legal/jquery.md | 26 + .../javadoc/0.8-SNAPSHOT/legal/jqueryUI.md | 49 + .../javadoc/0.8-SNAPSHOT/legal/jszip.md | 653 + .../static/javadoc/0.8-SNAPSHOT/legal/pako.md | 45 + .../0.8-SNAPSHOT/member-search-index.js | 1 + .../0.8-SNAPSHOT/member-search-index.zip | Bin 0 -> 20501 bytes website/static/javadoc/0.8-SNAPSHOT/options | 35 + .../org/apache/fluss/client/Connection.html | 340 + .../fluss/client/ConnectionFactory.html | 314 + .../apache/fluss/client/FlussConnection.html | 435 + .../org/apache/fluss/client/admin/Admin.html | 1040 ++ .../fluss/client/admin/CreateAclsResult.html | 385 + .../admin/DropAclsResult.FilterResult.html | 292 + .../admin/DropAclsResult.FilterResults.html | 274 + .../fluss/client/admin/DropAclsResult.html | 352 + .../apache/fluss/client/admin/FlussAdmin.html | 1173 ++ .../fluss/client/admin/ListOffsetsResult.html | 326 + .../client/admin/OffsetSpec.EarliestSpec.html | 312 + .../client/admin/OffsetSpec.LatestSpec.html | 312 + .../admin/OffsetSpec.TimestampSpec.html | 350 + .../apache/fluss/client/admin/OffsetSpec.html | 399 + .../fluss/client/admin/class-use/Admin.html | 413 + .../admin/class-use/CreateAclsResult.html | 199 + .../DropAclsResult.FilterResult.html | 194 + .../DropAclsResult.FilterResults.html | 195 + .../admin/class-use/DropAclsResult.html | 199 + .../client/admin/class-use/FlussAdmin.html | 148 + .../admin/class-use/ListOffsetsResult.html | 221 + .../class-use/OffsetSpec.EarliestSpec.html | 148 + .../class-use/OffsetSpec.LatestSpec.html | 148 + .../class-use/OffsetSpec.TimestampSpec.html | 148 + .../client/admin/class-use/OffsetSpec.html | 281 + .../fluss/client/admin/package-summary.html | 241 + .../fluss/client/admin/package-tree.html | 185 + .../fluss/client/admin/package-use.html | 414 + .../fluss/client/class-use/Connection.html | 246 + .../client/class-use/ConnectionFactory.html | 148 + .../client/class-use/FlussConnection.html | 217 + .../client/lookup/AbstractLookupBatch.html | 412 + .../client/lookup/AbstractLookupQuery.html | 356 + .../apache/fluss/client/lookup/Lookup.html | 354 + .../fluss/client/lookup/LookupBatch.html | 370 + .../fluss/client/lookup/LookupClient.html | 397 + .../fluss/client/lookup/LookupQuery.html | 303 + .../fluss/client/lookup/LookupResult.html | 395 + .../fluss/client/lookup/LookupType.html | 380 + .../apache/fluss/client/lookup/Lookuper.html | 269 + .../fluss/client/lookup/PrefixLookup.html | 312 + .../client/lookup/PrefixLookupBatch.html | 368 + .../client/lookup/PrefixLookupQuery.html | 303 + .../fluss/client/lookup/TableLookup.html | 377 + .../lookup/class-use/AbstractLookupBatch.html | 148 + .../lookup/class-use/AbstractLookupQuery.html | 248 + .../fluss/client/lookup/class-use/Lookup.html | 255 + .../client/lookup/class-use/LookupBatch.html | 193 + .../client/lookup/class-use/LookupClient.html | 218 + .../client/lookup/class-use/LookupQuery.html | 207 + .../client/lookup/class-use/LookupResult.html | 194 + .../client/lookup/class-use/LookupType.html | 246 + .../client/lookup/class-use/Lookuper.html | 200 + .../client/lookup/class-use/PrefixLookup.html | 148 + .../lookup/class-use/PrefixLookupBatch.html | 193 + .../lookup/class-use/PrefixLookupQuery.html | 207 + .../client/lookup/class-use/TableLookup.html | 148 + .../fluss/client/lookup/package-summary.html | 265 + .../fluss/client/lookup/package-tree.html | 196 + .../fluss/client/lookup/package-use.html | 328 + .../client/metadata/KvSnapshotMetadata.html | 354 + .../fluss/client/metadata/KvSnapshots.html | 385 + .../fluss/client/metadata/LakeSnapshot.html | 363 + .../client/metadata/MetadataUpdater.html | 721 + .../class-use/KvSnapshotMetadata.html | 227 + .../metadata/class-use/KvSnapshots.html | 239 + .../metadata/class-use/LakeSnapshot.html | 225 + .../metadata/class-use/MetadataUpdater.html | 398 + .../client/metadata/package-summary.html | 186 + .../fluss/client/metadata/package-tree.html | 166 + .../fluss/client/metadata/package-use.html | 359 + .../client/metrics/ScannerMetricGroup.html | 494 + .../client/metrics/WriterMetricGroup.html | 460 + .../metrics/class-use/ScannerMetricGroup.html | 203 + .../metrics/class-use/WriterMetricGroup.html | 197 + .../fluss/client/metrics/package-summary.html | 174 + .../fluss/client/metrics/package-tree.html | 168 + .../fluss/client/metrics/package-use.html | 208 + .../apache/fluss/client/package-summary.html | 191 + .../org/apache/fluss/client/package-tree.html | 174 + .../org/apache/fluss/client/package-use.html | 254 + .../apache/fluss/client/table/FlussTable.html | 451 + .../org/apache/fluss/client/table/Table.html | 358 + .../client/table/class-use/FlussTable.html | 148 + .../fluss/client/table/class-use/Table.html | 306 + .../client/table/getter/PartitionGetter.html | 309 + .../getter/class-use/PartitionGetter.html | 197 + .../client/table/getter/package-summary.html | 168 + .../client/table/getter/package-tree.html | 163 + .../client/table/getter/package-use.html | 185 + .../fluss/client/table/package-summary.html | 185 + .../fluss/client/table/package-tree.html | 173 + .../fluss/client/table/package-use.html | 277 + .../table/scanner/RemoteFileDownloader.html | 435 + .../fluss/client/table/scanner/Scan.html | 374 + .../client/table/scanner/ScanRecord.html | 447 + .../fluss/client/table/scanner/TableScan.html | 449 + .../table/scanner/batch/BatchScanUtils.html | 367 + .../table/scanner/batch/BatchScanner.html | 305 + .../scanner/batch/KvSnapshotBatchScanner.html | 423 + .../scanner/batch/LimitBatchScanner.html | 365 + .../batch/class-use/BatchScanUtils.html | 148 + .../scanner/batch/class-use/BatchScanner.html | 358 + .../class-use/KvSnapshotBatchScanner.html | 148 + .../batch/class-use/LimitBatchScanner.html | 148 + .../table/scanner/batch/package-summary.html | 197 + .../table/scanner/batch/package-tree.html | 179 + .../table/scanner/batch/package-use.html | 254 + .../class-use/RemoteFileDownloader.html | 268 + .../client/table/scanner/class-use/Scan.html | 274 + .../table/scanner/class-use/ScanRecord.html | 351 + .../table/scanner/class-use/TableScan.html | 148 + .../table/scanner/log/LogFetchBuffer.html | 340 + .../table/scanner/log/LogFetchCollector.html | 329 + .../client/table/scanner/log/LogFetcher.html | 423 + .../client/table/scanner/log/LogScan.html | 334 + .../client/table/scanner/log/LogScanner.html | 503 + .../table/scanner/log/LogScannerImpl.html | 502 + .../table/scanner/log/LogScannerStatus.html | 271 + .../scanner/log/RemoteLogDownloadFuture.html | 351 + .../scanner/log/RemoteLogDownloader.html | 378 + .../client/table/scanner/log/ScanRecords.html | 445 + .../scanner/log/class-use/LogFetchBuffer.html | 194 + .../log/class-use/LogFetchCollector.html | 148 + .../scanner/log/class-use/LogFetcher.html | 148 + .../table/scanner/log/class-use/LogScan.html | 194 + .../scanner/log/class-use/LogScanner.html | 227 + .../scanner/log/class-use/LogScannerImpl.html | 148 + .../log/class-use/LogScannerStatus.html | 203 + .../class-use/RemoteLogDownloadFuture.html | 195 + .../log/class-use/RemoteLogDownloader.html | 148 + .../scanner/log/class-use/ScanRecords.html | 214 + .../table/scanner/log/package-summary.html | 235 + .../table/scanner/log/package-tree.html | 181 + .../client/table/scanner/log/package-use.html | 239 + .../client/table/scanner/package-summary.html | 198 + .../client/table/scanner/package-tree.html | 171 + .../client/table/scanner/package-use.html | 332 + .../table/writer/AbstractTableWriter.html | 443 + .../fluss/client/table/writer/Append.html | 265 + .../client/table/writer/AppendResult.html | 274 + .../client/table/writer/AppendWriter.html | 277 + .../client/table/writer/DeleteResult.html | 274 + .../client/table/writer/TableAppend.html | 324 + .../client/table/writer/TableUpsert.html | 384 + .../client/table/writer/TableWriter.html | 275 + .../fluss/client/table/writer/Upsert.html | 322 + .../client/table/writer/UpsertResult.html | 274 + .../client/table/writer/UpsertWrite.html | 352 + .../client/table/writer/UpsertWriter.html | 302 + .../writer/class-use/AbstractTableWriter.html | 148 + .../client/table/writer/class-use/Append.html | 228 + .../table/writer/class-use/AppendResult.html | 194 + .../table/writer/class-use/AppendWriter.html | 199 + .../table/writer/class-use/DeleteResult.html | 195 + .../table/writer/class-use/TableAppend.html | 148 + .../table/writer/class-use/TableUpsert.html | 148 + .../table/writer/class-use/TableWriter.html | 218 + .../client/table/writer/class-use/Upsert.html | 260 + .../table/writer/class-use/UpsertResult.html | 194 + .../table/writer/class-use/UpsertWrite.html | 194 + .../table/writer/class-use/UpsertWriter.html | 200 + .../client/table/writer/package-summary.html | 246 + .../client/table/writer/package-tree.html | 182 + .../client/table/writer/package-use.html | 264 + .../token/DefaultSecurityTokenManager.html | 360 + .../token/DefaultSecurityTokenProvider.html | 325 + .../client/token/SecurityTokenManager.html | 280 + .../client/token/SecurityTokenProvider.html | 265 + .../DefaultSecurityTokenManager.html | 148 + .../DefaultSecurityTokenProvider.html | 148 + .../token/class-use/SecurityTokenManager.html | 194 + .../class-use/SecurityTokenProvider.html | 208 + .../fluss/client/token/package-summary.html | 197 + .../fluss/client/token/package-tree.html | 171 + .../fluss/client/token/package-use.html | 191 + .../client/utils/ClientRpcMessageUtils.html | 556 + .../fluss/client/utils/ClientUtils.html | 343 + .../fluss/client/utils/MetadataUtils.html | 420 + .../class-use/ClientRpcMessageUtils.html | 148 + .../client/utils/class-use/ClientUtils.html | 148 + .../client/utils/class-use/MetadataUtils.html | 148 + .../fluss/client/utils/package-summary.html | 181 + .../fluss/client/utils/package-tree.html | 165 + .../fluss/client/utils/package-use.html | 148 + .../client/write/ArrowLogWriteBatch.html | 597 + .../client/write/DynamicPartitionCreator.html | 314 + .../write/DynamicWriteBatchSizeEstimator.html | 329 + .../client/write/HashBucketAssigner.html | 370 + .../client/write/IdempotenceBucketEntry.html | 224 + .../client/write/IdempotenceBucketMap.html | 270 + .../client/write/IdempotenceManager.html | 276 + .../client/write/IndexedLogWriteBatch.html | 597 + .../fluss/client/write/KvWriteBatch.html | 617 + .../fluss/client/write/ReadyWriteBatch.html | 325 + .../RecordAccumulator.ReadyCheckResult.html | 347 + .../RecordAccumulator.RecordAppendResult.html | 350 + .../fluss/client/write/RecordAccumulator.html | 541 + .../write/RoundRobinBucketAssigner.html | 355 + .../org/apache/fluss/client/write/Sender.html | 402 + .../client/write/StickyBucketAssigner.html | 356 + .../write/WriteBatch.RequestFuture.html | 336 + .../apache/fluss/client/write/WriteBatch.html | 763 ++ .../fluss/client/write/WriteCallback.html | 255 + .../fluss/client/write/WriteFormat.html | 393 + .../fluss/client/write/WriteRecord.html | 461 + .../fluss/client/write/WriterClient.html | 419 + .../write/class-use/ArrowLogWriteBatch.html | 148 + .../class-use/DynamicPartitionCreator.html | 148 + .../DynamicWriteBatchSizeEstimator.html | 148 + .../write/class-use/HashBucketAssigner.html | 148 + .../class-use/IdempotenceBucketEntry.html | 148 + .../write/class-use/IdempotenceBucketMap.html | 148 + .../write/class-use/IdempotenceManager.html | 197 + .../write/class-use/IndexedLogWriteBatch.html | 148 + .../client/write/class-use/KvWriteBatch.html | 148 + .../write/class-use/ReadyWriteBatch.html | 249 + .../RecordAccumulator.ReadyCheckResult.html | 194 + .../RecordAccumulator.RecordAppendResult.html | 198 + .../write/class-use/RecordAccumulator.html | 197 + .../class-use/RoundRobinBucketAssigner.html | 148 + .../fluss/client/write/class-use/Sender.html | 148 + .../write/class-use/StickyBucketAssigner.html | 148 + .../class-use/WriteBatch.RequestFuture.html | 192 + .../client/write/class-use/WriteBatch.html | 254 + .../client/write/class-use/WriteCallback.html | 248 + .../client/write/class-use/WriteFormat.html | 207 + .../client/write/class-use/WriteRecord.html | 308 + .../client/write/class-use/WriterClient.html | 245 + .../fluss/client/write/package-summary.html | 319 + .../fluss/client/write/package-tree.html | 205 + .../fluss/client/write/package-use.html | 317 + ...FlinkConnectorOptions.ScanStartupMode.html | 446 + .../fluss/flink/FlinkConnectorOptions.html | 476 + .../flink/adapter/CatalogTableAdapter.html | 315 + .../SingleThreadFetcherManagerAdapter.html | 320 + ...hreadMultiplexSourceReaderBaseAdapter.html | 334 + .../class-use/CatalogTableAdapter.html | 148 + .../SingleThreadFetcherManagerAdapter.html | 194 + ...hreadMultiplexSourceReaderBaseAdapter.html | 194 + .../fluss/flink/adapter/package-summary.html | 181 + .../fluss/flink/adapter/package-tree.html | 181 + .../fluss/flink/adapter/package-use.html | 209 + .../fluss/flink/catalog/FlinkCatalog.html | 1463 ++ .../flink/catalog/FlinkCatalogFactory.html | 446 + .../flink/catalog/FlinkCatalogOptions.html | 271 + .../flink/catalog/FlinkTableFactory.html | 412 + .../flink/catalog/class-use/FlinkCatalog.html | 192 + .../class-use/FlinkCatalogFactory.html | 148 + .../class-use/FlinkCatalogOptions.html | 148 + .../catalog/class-use/FlinkTableFactory.html | 148 + .../fluss/flink/catalog/package-summary.html | 186 + .../fluss/flink/catalog/package-tree.html | 170 + .../fluss/flink/catalog/package-use.html | 185 + ...FlinkConnectorOptions.ScanStartupMode.html | 243 + .../class-use/FlinkConnectorOptions.html | 148 + .../apache/fluss/flink/lake/LakeCatalog.html | 318 + .../flink/lake/LakeRecordRecordEmitter.html | 311 + .../fluss/flink/lake/LakeSplitGenerator.html | 330 + .../flink/lake/LakeSplitReaderGenerator.html | 329 + .../fluss/flink/lake/LakeSplitSerializer.html | 340 + .../flink/lake/LakeSplitStateInitializer.html | 307 + .../fluss/flink/lake/LakeTableFactory.html | 309 + .../flink/lake/class-use/LakeCatalog.html | 148 + .../class-use/LakeRecordRecordEmitter.html | 148 + .../lake/class-use/LakeSplitGenerator.html | 148 + .../class-use/LakeSplitReaderGenerator.html | 148 + .../lake/class-use/LakeSplitSerializer.html | 148 + .../class-use/LakeSplitStateInitializer.html | 148 + .../lake/class-use/LakeTableFactory.html | 148 + .../fluss/flink/lake/package-summary.html | 204 + .../apache/fluss/flink/lake/package-tree.html | 169 + .../apache/fluss/flink/lake/package-use.html | 148 + .../IndexedLakeSplitRecordIterator.html | 393 + .../fluss/flink/lake/reader/KeyValueRow.html | 339 + .../LakeSnapshotAndLogSplitScanner.html | 363 + .../lake/reader/LakeSnapshotScanner.html | 358 + .../SeekableLakeSnapshotSplitScanner.html | 373 + .../IndexedLakeSplitRecordIterator.html | 148 + .../lake/reader/class-use/KeyValueRow.html | 148 + .../LakeSnapshotAndLogSplitScanner.html | 148 + .../reader/class-use/LakeSnapshotScanner.html | 148 + .../SeekableLakeSnapshotSplitScanner.html | 148 + .../flink/lake/reader/package-summary.html | 193 + .../fluss/flink/lake/reader/package-tree.html | 167 + .../fluss/flink/lake/reader/package-use.html | 148 + .../split/LakeSnapshotAndFlussLogSplit.html | 624 + .../flink/lake/split/LakeSnapshotSplit.html | 501 + .../LakeSnapshotAndFlussLogSplit.html | 258 + .../split/class-use/LakeSnapshotSplit.html | 215 + .../flink/lake/split/package-summary.html | 174 + .../fluss/flink/lake/split/package-tree.html | 168 + .../fluss/flink/lake/split/package-use.html | 243 + .../LakeSnapshotAndFlussLogSplitState.html | 365 + .../lake/state/LakeSnapshotSplitState.html | 337 + .../LakeSnapshotAndFlussLogSplitState.html | 148 + .../class-use/LakeSnapshotSplitState.html | 148 + .../flink/lake/state/package-summary.html | 174 + .../fluss/flink/lake/state/package-tree.html | 168 + .../fluss/flink/lake/state/package-use.html | 148 + .../fluss/flink/metrics/FlinkCounter.html | 395 + .../fluss/flink/metrics/FlinkGauge.html | 323 + .../fluss/flink/metrics/FlinkHistogram.html | 359 + .../fluss/flink/metrics/FlinkMeter.html | 377 + .../flink/metrics/FlinkMetricRegistry.html | 477 + .../flink/metrics/class-use/FlinkCounter.html | 148 + .../flink/metrics/class-use/FlinkGauge.html | 148 + .../metrics/class-use/FlinkHistogram.html | 148 + .../flink/metrics/class-use/FlinkMeter.html | 148 + .../class-use/FlinkMetricRegistry.html | 192 + .../fluss/flink/metrics/package-summary.html | 193 + .../fluss/flink/metrics/package-tree.html | 167 + .../fluss/flink/metrics/package-use.html | 186 + .../apache/fluss/flink/package-summary.html | 185 + .../org/apache/fluss/flink/package-tree.html | 177 + .../org/apache/fluss/flink/package-use.html | 208 + .../flink/procedure/AbstractAclProcedure.html | 419 + .../flink/procedure/AddAclProcedure.html | 397 + .../flink/procedure/DropAclProcedure.html | 401 + .../flink/procedure/ListAclProcedure.html | 401 + .../fluss/flink/procedure/ProcedureBase.html | 358 + .../flink/procedure/ProcedureManager.html | 323 + .../class-use/AbstractAclProcedure.html | 208 + .../procedure/class-use/AddAclProcedure.html | 148 + .../procedure/class-use/DropAclProcedure.html | 148 + .../procedure/class-use/ListAclProcedure.html | 148 + .../procedure/class-use/ProcedureBase.html | 215 + .../procedure/class-use/ProcedureManager.html | 148 + .../flink/procedure/package-summary.html | 198 + .../fluss/flink/procedure/package-tree.html | 174 + .../fluss/flink/procedure/package-use.html | 191 + .../fluss/flink/row/FlinkAsFlussRow.html | 643 + .../apache/fluss/flink/row/OperationType.html | 432 + .../org/apache/fluss/flink/row/RowWithOp.html | 406 + .../flink/row/class-use/FlinkAsFlussRow.html | 192 + .../flink/row/class-use/OperationType.html | 225 + .../fluss/flink/row/class-use/RowWithOp.html | 201 + .../fluss/flink/row/package-summary.html | 192 + .../apache/fluss/flink/row/package-tree.html | 178 + .../apache/fluss/flink/row/package-use.html | 215 + .../fluss/flink/sink/ChannelComputer.html | 315 + .../sink/FlinkRowDataChannelComputer.html | 369 + .../flink/sink/FlinkStreamPartitioner.html | 454 + .../fluss/flink/sink/FlinkTableSink.html | 541 + .../apache/fluss/flink/sink/FlussSink.html | 376 + .../fluss/flink/sink/FlussSinkBuilder.html | 455 + .../flink/sink/class-use/ChannelComputer.html | 224 + .../FlinkRowDataChannelComputer.html | 148 + .../class-use/FlinkStreamPartitioner.html | 148 + .../flink/sink/class-use/FlinkTableSink.html | 148 + .../fluss/flink/sink/class-use/FlussSink.html | 194 + .../sink/class-use/FlussSinkBuilder.html | 244 + .../fluss/flink/sink/package-summary.html | 210 + .../apache/fluss/flink/sink/package-tree.html | 181 + .../apache/fluss/flink/sink/package-use.html | 198 + ...alizationSchema.InitializationContext.html | 272 + .../serializer/FlussSerializationSchema.html | 331 + .../RowDataSerializationSchema.html | 384 + .../serializer/SerializerInitContextImpl.html | 322 + ...alizationSchema.InitializationContext.html | 218 + .../class-use/FlussSerializationSchema.html | 294 + .../class-use/RowDataSerializationSchema.html | 148 + .../class-use/SerializerInitContextImpl.html | 148 + .../sink/serializer/package-summary.html | 197 + .../flink/sink/serializer/package-tree.html | 175 + .../flink/sink/serializer/package-use.html | 237 + .../flink/sink/writer/AppendSinkWriter.html | 399 + .../flink/sink/writer/FlinkSinkWriter.html | 593 + .../flink/sink/writer/UpsertSinkWriter.html | 404 + .../writer/class-use/AppendSinkWriter.html | 148 + .../writer/class-use/FlinkSinkWriter.html | 201 + .../writer/class-use/UpsertSinkWriter.html | 148 + .../flink/sink/writer/package-summary.html | 180 + .../fluss/flink/sink/writer/package-tree.html | 168 + .../fluss/flink/sink/writer/package-use.html | 185 + .../fluss/flink/source/FlinkSource.html | 550 + .../fluss/flink/source/FlinkTableSource.html | 704 + .../fluss/flink/source/FlussSource.html | 338 + .../flink/source/FlussSourceBuilder.html | 547 + .../flink/source/class-use/FlinkSource.html | 194 + .../source/class-use/FlinkTableSource.html | 148 + .../flink/source/class-use/FlussSource.html | 194 + .../source/class-use/FlussSourceBuilder.html | 250 + .../DeserializerInitContextImpl.html | 380 + ...alizationSchema.InitializationContext.html | 322 + .../FlussDeserializationSchema.html | 357 + .../JsonStringDeserializationSchema.html | 435 + .../RowDataDeserializationSchema.html | 414 + .../DeserializerInitContextImpl.html | 148 + ...alizationSchema.InitializationContext.html | 225 + .../class-use/FlussDeserializationSchema.html | 292 + .../JsonStringDeserializationSchema.html | 148 + .../RowDataDeserializationSchema.html | 148 + .../source/deserializer/package-summary.html | 205 + .../source/deserializer/package-tree.html | 176 + .../source/deserializer/package-use.html | 240 + .../source/emitter/FlinkRecordEmitter.html | 327 + .../emitter/class-use/FlinkRecordEmitter.html | 198 + .../flink/source/emitter/package-summary.html | 168 + .../flink/source/emitter/package-tree.html | 163 + .../flink/source/emitter/package-use.html | 185 + .../enumerator/FlinkSourceEnumerator.html | 580 + .../class-use/FlinkSourceEnumerator.html | 148 + .../BucketOffsetsRetrieverImpl.html | 365 + .../initializer/LatestOffsetsInitializer.html | 357 + .../NoStoppingOffsetsInitializer.html | 356 + ...etsInitializer.BucketOffsetsRetriever.html | 299 + .../initializer/OffsetsInitializer.html | 410 + .../SnapshotOffsetsInitializer.html | 357 + .../TimestampOffsetsInitializer.html | 356 + .../class-use/BucketOffsetsRetrieverImpl.html | 148 + .../class-use/LatestOffsetsInitializer.html | 148 + .../NoStoppingOffsetsInitializer.html | 148 + ...etsInitializer.BucketOffsetsRetriever.html | 273 + .../class-use/OffsetsInitializer.html | 431 + .../class-use/SnapshotOffsetsInitializer.html | 148 + .../TimestampOffsetsInitializer.html | 148 + .../initializer/package-summary.html | 217 + .../enumerator/initializer/package-tree.html | 178 + .../enumerator/initializer/package-use.html | 268 + .../source/enumerator/package-summary.html | 168 + .../flink/source/enumerator/package-tree.html | 163 + .../flink/source/enumerator/package-use.html | 148 + .../PartitionBucketsUnsubscribedEvent.html | 371 + .../source/event/PartitionsRemovedEvent.html | 337 + .../PartitionBucketsUnsubscribedEvent.html | 148 + .../class-use/PartitionsRemovedEvent.html | 148 + .../flink/source/event/package-summary.html | 175 + .../flink/source/event/package-tree.html | 164 + .../fluss/flink/source/event/package-use.html | 148 + .../lookup/FlinkAsyncLookupFunction.html | 417 + .../source/lookup/FlinkLookupFunction.html | 418 + .../LookupNormalizer.RemainingFilter.html | 271 + .../flink/source/lookup/LookupNormalizer.html | 418 + .../class-use/FlinkAsyncLookupFunction.html | 148 + .../lookup/class-use/FlinkLookupFunction.html | 148 + .../LookupNormalizer.RemainingFilter.html | 192 + .../lookup/class-use/LookupNormalizer.html | 234 + .../flink/source/lookup/package-summary.html | 187 + .../flink/source/lookup/package-tree.html | 186 + .../flink/source/lookup/package-use.html | 192 + .../metrics/FlinkSourceReaderMetrics.html | 523 + .../class-use/FlinkSourceReaderMetrics.html | 242 + .../flink/source/metrics/package-summary.html | 168 + .../flink/source/metrics/package-tree.html | 163 + .../flink/source/metrics/package-use.html | 185 + .../fluss/flink/source/package-summary.html | 186 + .../fluss/flink/source/package-tree.html | 169 + .../fluss/flink/source/package-use.html | 197 + .../source/reader/BoundedSplitReader.html | 350 + .../reader/FlinkRecordsWithSplitIds.html | 444 + .../source/reader/FlinkSourceReader.html | 443 + .../source/reader/FlinkSourceSplitReader.html | 411 + .../source/reader/MutableRecordAndPos.html | 359 + .../flink/source/reader/RecordAndPos.html | 544 + .../flink/source/reader/SplitScanner.html | 269 + .../reader/class-use/BoundedSplitReader.html | 192 + .../class-use/FlinkRecordsWithSplitIds.html | 192 + .../reader/class-use/FlinkSourceReader.html | 148 + .../class-use/FlinkSourceSplitReader.html | 148 + .../reader/class-use/MutableRecordAndPos.html | 148 + .../source/reader/class-use/RecordAndPos.html | 359 + .../source/reader/class-use/SplitScanner.html | 148 + .../fetcher/FlinkSourceFetcherManager.html | 373 + .../class-use/FlinkSourceFetcherManager.html | 148 + .../reader/fetcher/package-summary.html | 168 + .../source/reader/fetcher/package-tree.html | 175 + .../source/reader/fetcher/package-use.html | 148 + .../flink/source/reader/package-summary.html | 217 + .../flink/source/reader/package-tree.html | 193 + .../flink/source/reader/package-use.html | 274 + .../source/split/HybridSnapshotLogSplit.html | 476 + .../split/HybridSnapshotLogSplitState.html | 368 + .../fluss/flink/source/split/LogSplit.html | 501 + .../flink/source/split/LogSplitState.html | 354 + .../flink/source/split/SnapshotSplit.html | 496 + .../flink/source/split/SourceSplitBase.html | 582 + .../source/split/SourceSplitSerializer.html | 362 + .../flink/source/split/SourceSplitState.html | 435 + .../class-use/HybridSnapshotLogSplit.html | 212 + .../HybridSnapshotLogSplitState.html | 194 + .../source/split/class-use/LogSplit.html | 212 + .../source/split/class-use/LogSplitState.html | 194 + .../source/split/class-use/SnapshotSplit.html | 195 + .../split/class-use/SourceSplitBase.html | 733 + .../class-use/SourceSplitSerializer.html | 148 + .../split/class-use/SourceSplitState.html | 364 + .../flink/source/split/package-summary.html | 211 + .../flink/source/split/package-tree.html | 179 + .../fluss/flink/source/split/package-use.html | 447 + .../FlussSourceEnumeratorStateSerializer.html | 362 + .../source/state/SourceEnumeratorState.html | 395 + .../FlussSourceEnumeratorStateSerializer.html | 148 + .../class-use/SourceEnumeratorState.html | 287 + .../flink/source/state/package-summary.html | 174 + .../flink/source/state/package-tree.html | 164 + .../fluss/flink/source/state/package-use.html | 231 + .../flink/tiering/LakeTieringJobBuilder.html | 292 + .../class-use/LakeTieringJobBuilder.html | 195 + .../tiering/committer/CommittableMessage.html | 318 + .../committer/CommittableMessageTypeInfo.html | 485 + .../FlussTableLakeSnapshotCommitter.html | 354 + .../committer/TieringCommitOperator.html | 499 + .../TieringCommitOperatorFactory.html | 376 + .../TieringCommitterInitContext.html | 316 + .../class-use/CommittableMessage.html | 254 + .../class-use/CommittableMessageTypeInfo.html | 148 + .../FlussTableLakeSnapshotCommitter.html | 148 + .../class-use/TieringCommitOperator.html | 148 + .../TieringCommitOperatorFactory.html | 148 + .../TieringCommitterInitContext.html | 148 + .../tiering/committer/package-summary.html | 199 + .../flink/tiering/committer/package-tree.html | 180 + .../flink/tiering/committer/package-use.html | 185 + .../tiering/event/FailedTieringEvent.html | 332 + .../tiering/event/FinishedTieringEvent.html | 316 + .../tiering/event/TieringRestoreEvent.html | 278 + .../event/class-use/FailedTieringEvent.html | 148 + .../event/class-use/FinishedTieringEvent.html | 148 + .../event/class-use/TieringRestoreEvent.html | 148 + .../flink/tiering/event/package-summary.html | 180 + .../flink/tiering/event/package-tree.html | 165 + .../flink/tiering/event/package-use.html | 148 + .../fluss/flink/tiering/package-summary.html | 168 + .../fluss/flink/tiering/package-tree.html | 163 + .../fluss/flink/tiering/package-use.html | 185 + .../source/TableBucketWriteResult.html | 405 + .../source/TableBucketWriteResultEmitter.html | 320 + .../TableBucketWriteResultSerializer.html | 362 + .../TableBucketWriteResultTypeInfo.html | 485 + .../tiering/source/TieringSource.Builder.html | 327 + .../flink/tiering/source/TieringSource.html | 514 + .../tiering/source/TieringSourceOptions.html | 329 + .../tiering/source/TieringSourceReader.html | 446 + .../tiering/source/TieringSplitReader.html | 385 + .../source/TieringWriterInitContext.html | 378 + .../class-use/TableBucketWriteResult.html | 293 + .../TableBucketWriteResultEmitter.html | 148 + .../TableBucketWriteResultSerializer.html | 148 + .../TableBucketWriteResultTypeInfo.html | 148 + .../class-use/TieringSource.Builder.html | 192 + .../source/class-use/TieringSource.html | 192 + .../class-use/TieringSourceOptions.html | 148 + .../source/class-use/TieringSourceReader.html | 148 + .../source/class-use/TieringSplitReader.html | 148 + .../class-use/TieringWriterInitContext.html | 148 + .../enumerator/TieringSourceEnumerator.html | 468 + .../class-use/TieringSourceEnumerator.html | 148 + .../source/enumerator/package-summary.html | 169 + .../source/enumerator/package-tree.html | 163 + .../source/enumerator/package-use.html | 148 + .../flink/tiering/source/package-summary.html | 224 + .../flink/tiering/source/package-tree.html | 184 + .../flink/tiering/source/package-use.html | 224 + .../tiering/source/split/TieringLogSplit.html | 473 + .../source/split/TieringSnapshotSplit.html | 474 + .../tiering/source/split/TieringSplit.html | 660 + .../source/split/TieringSplitGenerator.html | 312 + .../source/split/TieringSplitSerializer.html | 404 + .../split/class-use/TieringLogSplit.html | 199 + .../split/class-use/TieringSnapshotSplit.html | 199 + .../source/split/class-use/TieringSplit.html | 452 + .../class-use/TieringSplitGenerator.html | 148 + .../class-use/TieringSplitSerializer.html | 192 + .../tiering/source/split/package-summary.html | 192 + .../tiering/source/split/package-tree.html | 170 + .../tiering/source/split/package-use.html | 272 + .../state/TieringSourceEnumeratorState.html | 329 + ...ieringSourceEnumeratorStateSerializer.html | 404 + .../source/state/TieringSplitState.html | 353 + .../TieringSourceEnumeratorState.html | 287 + ...ieringSourceEnumeratorStateSerializer.html | 192 + .../state/class-use/TieringSplitState.html | 230 + .../tiering/source/state/package-summary.html | 180 + .../tiering/source/state/package-tree.html | 165 + .../tiering/source/state/package-use.html | 243 + .../flink/utils/CatalogExceptionUtils.html | 407 + .../flink/utils/CatalogPropertiesUtils.html | 379 + .../fluss/flink/utils/DataLakeUtils.html | 321 + ...kConnectorOptionsUtils.StartupOptions.html | 329 + .../utils/FlinkConnectorOptionsUtils.html | 424 + .../fluss/flink/utils/FlinkConversions.html | 422 + ...Converter.FlussSerializationConverter.html | 275 + .../utils/FlinkRowToFlussRowConverter.html | 371 + .../apache/fluss/flink/utils/FlinkUtils.html | 317 + ...nverter.FlussDeserializationConverter.html | 275 + .../utils/FlussRowToFlinkRowConverter.html | 381 + ...sonConverters.FlussRowToJsonConverter.html | 269 + .../flink/utils/FlussRowToJsonConverters.html | 332 + .../fluss/flink/utils/LakeSourceUtils.html | 356 + .../fluss/flink/utils/PojoToRowConverter.html | 349 + .../flink/utils/PushdownUtils.FieldEqual.html | 340 + .../utils/PushdownUtils.ValueConversion.html | 383 + .../fluss/flink/utils/PushdownUtils.html | 434 + .../fluss/flink/utils/TimestampFormat.html | 390 + .../class-use/CatalogExceptionUtils.html | 148 + .../class-use/CatalogPropertiesUtils.html | 148 + .../flink/utils/class-use/DataLakeUtils.html | 148 + ...kConnectorOptionsUtils.StartupOptions.html | 231 + .../class-use/FlinkConnectorOptionsUtils.html | 148 + .../utils/class-use/FlinkConversions.html | 148 + ...Converter.FlussSerializationConverter.html | 148 + .../FlinkRowToFlussRowConverter.html | 204 + .../flink/utils/class-use/FlinkUtils.html | 148 + ...nverter.FlussDeserializationConverter.html | 214 + .../FlussRowToFlinkRowConverter.html | 148 + ...sonConverters.FlussRowToJsonConverter.html | 192 + .../class-use/FlussRowToJsonConverters.html | 148 + .../utils/class-use/LakeSourceUtils.html | 148 + .../utils/class-use/PojoToRowConverter.html | 148 + .../class-use/PushdownUtils.FieldEqual.html | 292 + .../PushdownUtils.ValueConversion.html | 223 + .../flink/utils/class-use/PushdownUtils.html | 148 + .../utils/class-use/TimestampFormat.html | 215 + .../fluss/flink/utils/package-summary.html | 300 + .../fluss/flink/utils/package-tree.html | 203 + .../apache/fluss/flink/utils/package-use.html | 276 + .../0.8-SNAPSHOT/overview-summary.html | 23 + .../javadoc/0.8-SNAPSHOT/overview-tree.html | 581 + .../0.8-SNAPSHOT/package-search-index.js | 1 + .../0.8-SNAPSHOT/package-search-index.zip | Bin 0 -> 489 bytes website/static/javadoc/0.8-SNAPSHOT/packages | 185 + .../javadoc/0.8-SNAPSHOT/resources/glass.png | Bin 0 -> 499 bytes .../javadoc/0.8-SNAPSHOT/resources/x.png | Bin 0 -> 394 bytes website/static/javadoc/0.8-SNAPSHOT/script.js | 149 + website/static/javadoc/0.8-SNAPSHOT/search.js | 326 + .../javadoc/0.8-SNAPSHOT/serialized-form.html | 899 ++ .../javadoc/0.8-SNAPSHOT/stylesheet.css | 910 ++ .../javadoc/0.8-SNAPSHOT/type-search-index.js | 1 + .../0.8-SNAPSHOT/type-search-index.zip | Bin 0 -> 2356 bytes website/static/javadoc/index.html | 3 + 686 files changed, 207361 insertions(+) create mode 100755 tools/refresh_javadoc.sh create mode 100644 website/static/javadoc/0.8-SNAPSHOT/allclasses-index.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/allclasses.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/allpackages-index.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/argfile create mode 100644 website/static/javadoc/0.8-SNAPSHOT/constant-values.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/deprecated-list.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/element-list create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-1.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-10.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-11.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-12.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-13.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-14.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-15.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-16.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-17.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-18.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-19.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-2.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-20.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-21.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-22.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-23.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-3.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-4.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-5.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-6.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-7.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-8.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-9.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/index.html create mode 100755 website/static/javadoc/0.8-SNAPSHOT/javadoc.sh create mode 100644 website/static/javadoc/0.8-SNAPSHOT/jquery-ui.overrides.css create mode 100644 website/static/javadoc/0.8-SNAPSHOT/jquery/external/jquery/jquery.js create mode 100644 website/static/javadoc/0.8-SNAPSHOT/jquery/jquery-3.7.1.min.js create mode 100644 website/static/javadoc/0.8-SNAPSHOT/jquery/jquery-ui.min.css create mode 100644 website/static/javadoc/0.8-SNAPSHOT/jquery/jquery-ui.min.js create mode 100644 website/static/javadoc/0.8-SNAPSHOT/jquery/jszip-utils/dist/jszip-utils-ie.js create mode 100644 website/static/javadoc/0.8-SNAPSHOT/jquery/jszip-utils/dist/jszip-utils-ie.min.js create mode 100644 website/static/javadoc/0.8-SNAPSHOT/jquery/jszip-utils/dist/jszip-utils.js create mode 100644 website/static/javadoc/0.8-SNAPSHOT/jquery/jszip-utils/dist/jszip-utils.min.js create mode 100644 website/static/javadoc/0.8-SNAPSHOT/jquery/jszip/dist/jszip.js create mode 100644 website/static/javadoc/0.8-SNAPSHOT/jquery/jszip/dist/jszip.min.js create mode 100644 website/static/javadoc/0.8-SNAPSHOT/legal/ADDITIONAL_LICENSE_INFO create mode 100644 website/static/javadoc/0.8-SNAPSHOT/legal/ASSEMBLY_EXCEPTION create mode 100644 website/static/javadoc/0.8-SNAPSHOT/legal/LICENSE create mode 100644 website/static/javadoc/0.8-SNAPSHOT/legal/jquery.md create mode 100644 website/static/javadoc/0.8-SNAPSHOT/legal/jqueryUI.md create mode 100644 website/static/javadoc/0.8-SNAPSHOT/legal/jszip.md create mode 100644 website/static/javadoc/0.8-SNAPSHOT/legal/pako.md create mode 100644 website/static/javadoc/0.8-SNAPSHOT/member-search-index.js create mode 100644 website/static/javadoc/0.8-SNAPSHOT/member-search-index.zip create mode 100644 website/static/javadoc/0.8-SNAPSHOT/options create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/Connection.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/ConnectionFactory.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/FlussConnection.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/Admin.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/CreateAclsResult.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.FilterResult.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.FilterResults.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/FlussAdmin.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/ListOffsetsResult.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.EarliestSpec.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.LatestSpec.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.TimestampSpec.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/Admin.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/CreateAclsResult.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.FilterResult.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.FilterResults.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/FlussAdmin.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/ListOffsetsResult.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.EarliestSpec.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.LatestSpec.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.TimestampSpec.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/Connection.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/ConnectionFactory.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/FlussConnection.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/AbstractLookupBatch.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/AbstractLookupQuery.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/Lookup.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupBatch.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupClient.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupQuery.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupResult.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupType.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/Lookuper.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookup.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookupBatch.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookupQuery.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/TableLookup.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/AbstractLookupBatch.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/AbstractLookupQuery.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/Lookup.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupBatch.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupClient.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupQuery.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupResult.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupType.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/Lookuper.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookup.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookupBatch.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookupQuery.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/TableLookup.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/KvSnapshotMetadata.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/KvSnapshots.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/LakeSnapshot.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/MetadataUpdater.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/KvSnapshotMetadata.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/KvSnapshots.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/LakeSnapshot.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/MetadataUpdater.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/ScannerMetricGroup.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/WriterMetricGroup.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/class-use/ScannerMetricGroup.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/class-use/WriterMetricGroup.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/FlussTable.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/Table.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/class-use/FlussTable.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/class-use/Table.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/PartitionGetter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/class-use/PartitionGetter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/RemoteFileDownloader.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/Scan.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/ScanRecord.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/TableScan.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/BatchScanUtils.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/BatchScanner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/KvSnapshotBatchScanner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/LimitBatchScanner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/BatchScanUtils.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/BatchScanner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/KvSnapshotBatchScanner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/LimitBatchScanner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/RemoteFileDownloader.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/Scan.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/ScanRecord.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/TableScan.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetchBuffer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetchCollector.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetcher.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScan.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScanner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScannerImpl.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScannerStatus.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/RemoteLogDownloadFuture.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/RemoteLogDownloader.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/ScanRecords.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetchBuffer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetchCollector.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetcher.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScan.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScanner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScannerImpl.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScannerStatus.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/RemoteLogDownloadFuture.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/RemoteLogDownloader.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/ScanRecords.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AbstractTableWriter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/Append.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AppendResult.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AppendWriter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/DeleteResult.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableAppend.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableUpsert.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableWriter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/Upsert.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertResult.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertWrite.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertWriter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AbstractTableWriter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/Append.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AppendResult.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AppendWriter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/DeleteResult.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableAppend.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableUpsert.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableWriter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/Upsert.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertResult.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertWrite.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertWriter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/DefaultSecurityTokenManager.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/DefaultSecurityTokenProvider.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/SecurityTokenManager.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/SecurityTokenProvider.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/DefaultSecurityTokenManager.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/DefaultSecurityTokenProvider.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/SecurityTokenManager.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/SecurityTokenProvider.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/ClientRpcMessageUtils.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/ClientUtils.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/MetadataUtils.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/ClientRpcMessageUtils.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/ClientUtils.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/MetadataUtils.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/ArrowLogWriteBatch.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/DynamicPartitionCreator.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/DynamicWriteBatchSizeEstimator.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/HashBucketAssigner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceBucketEntry.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceBucketMap.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceManager.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IndexedLogWriteBatch.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/KvWriteBatch.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/ReadyWriteBatch.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.ReadyCheckResult.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.RecordAppendResult.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RoundRobinBucketAssigner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/Sender.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/StickyBucketAssigner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteBatch.RequestFuture.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteBatch.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteCallback.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteFormat.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteRecord.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriterClient.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/ArrowLogWriteBatch.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/DynamicPartitionCreator.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/DynamicWriteBatchSizeEstimator.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/HashBucketAssigner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceBucketEntry.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceBucketMap.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceManager.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IndexedLogWriteBatch.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/KvWriteBatch.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/ReadyWriteBatch.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.ReadyCheckResult.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.RecordAppendResult.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RoundRobinBucketAssigner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/Sender.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/StickyBucketAssigner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteBatch.RequestFuture.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteBatch.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteCallback.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteFormat.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteRecord.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriterClient.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/FlinkConnectorOptions.ScanStartupMode.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/FlinkConnectorOptions.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/CatalogTableAdapter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/SingleThreadFetcherManagerAdapter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/SingleThreadMultiplexSourceReaderBaseAdapter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/CatalogTableAdapter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/SingleThreadFetcherManagerAdapter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/SingleThreadMultiplexSourceReaderBaseAdapter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalog.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalogFactory.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalogOptions.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkTableFactory.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalog.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalogFactory.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalogOptions.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkTableFactory.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/class-use/FlinkConnectorOptions.ScanStartupMode.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/class-use/FlinkConnectorOptions.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeCatalog.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeRecordRecordEmitter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitGenerator.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitReaderGenerator.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitSerializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitStateInitializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeTableFactory.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeCatalog.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeRecordRecordEmitter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitGenerator.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitReaderGenerator.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitSerializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitStateInitializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeTableFactory.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/IndexedLakeSplitRecordIterator.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/KeyValueRow.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/LakeSnapshotAndLogSplitScanner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/LakeSnapshotScanner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/SeekableLakeSnapshotSplitScanner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/IndexedLakeSplitRecordIterator.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/KeyValueRow.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/LakeSnapshotAndLogSplitScanner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/LakeSnapshotScanner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/SeekableLakeSnapshotSplitScanner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/LakeSnapshotAndFlussLogSplit.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/LakeSnapshotSplit.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/class-use/LakeSnapshotAndFlussLogSplit.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/class-use/LakeSnapshotSplit.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/LakeSnapshotAndFlussLogSplitState.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/LakeSnapshotSplitState.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/class-use/LakeSnapshotAndFlussLogSplitState.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/class-use/LakeSnapshotSplitState.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkCounter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkGauge.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkHistogram.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkMeter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkMetricRegistry.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkCounter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkGauge.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkHistogram.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkMeter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkMetricRegistry.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/AbstractAclProcedure.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/AddAclProcedure.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/DropAclProcedure.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ListAclProcedure.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ProcedureBase.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ProcedureManager.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/AbstractAclProcedure.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/AddAclProcedure.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/DropAclProcedure.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ListAclProcedure.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ProcedureBase.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ProcedureManager.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/FlinkAsFlussRow.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/OperationType.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/RowWithOp.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/FlinkAsFlussRow.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/OperationType.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/RowWithOp.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/ChannelComputer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkRowDataChannelComputer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkStreamPartitioner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkTableSink.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlussSink.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlussSinkBuilder.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/ChannelComputer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkRowDataChannelComputer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkStreamPartitioner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkTableSink.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlussSink.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlussSinkBuilder.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/FlussSerializationSchema.InitializationContext.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/FlussSerializationSchema.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/RowDataSerializationSchema.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/SerializerInitContextImpl.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/FlussSerializationSchema.InitializationContext.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/FlussSerializationSchema.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/RowDataSerializationSchema.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/SerializerInitContextImpl.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/AppendSinkWriter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/FlinkSinkWriter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/UpsertSinkWriter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/AppendSinkWriter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/FlinkSinkWriter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/UpsertSinkWriter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlinkSource.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlinkTableSource.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlussSource.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlussSourceBuilder.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlinkSource.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlinkTableSource.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlussSource.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlussSourceBuilder.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/DeserializerInitContextImpl.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/FlussDeserializationSchema.InitializationContext.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/FlussDeserializationSchema.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/JsonStringDeserializationSchema.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/RowDataDeserializationSchema.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/DeserializerInitContextImpl.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/FlussDeserializationSchema.InitializationContext.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/FlussDeserializationSchema.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/JsonStringDeserializationSchema.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/RowDataDeserializationSchema.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/FlinkRecordEmitter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/class-use/FlinkRecordEmitter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/FlinkSourceEnumerator.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/class-use/FlinkSourceEnumerator.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/BucketOffsetsRetrieverImpl.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/LatestOffsetsInitializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/NoStoppingOffsetsInitializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/OffsetsInitializer.BucketOffsetsRetriever.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/OffsetsInitializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/SnapshotOffsetsInitializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/TimestampOffsetsInitializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/BucketOffsetsRetrieverImpl.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/LatestOffsetsInitializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/NoStoppingOffsetsInitializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/OffsetsInitializer.BucketOffsetsRetriever.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/OffsetsInitializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/SnapshotOffsetsInitializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/TimestampOffsetsInitializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/PartitionBucketsUnsubscribedEvent.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/PartitionsRemovedEvent.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/class-use/PartitionBucketsUnsubscribedEvent.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/class-use/PartitionsRemovedEvent.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/FlinkAsyncLookupFunction.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/FlinkLookupFunction.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/LookupNormalizer.RemainingFilter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/LookupNormalizer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/FlinkAsyncLookupFunction.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/FlinkLookupFunction.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/LookupNormalizer.RemainingFilter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/LookupNormalizer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/FlinkSourceReaderMetrics.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/class-use/FlinkSourceReaderMetrics.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/BoundedSplitReader.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkRecordsWithSplitIds.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkSourceReader.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkSourceSplitReader.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/MutableRecordAndPos.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/RecordAndPos.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/SplitScanner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/BoundedSplitReader.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkRecordsWithSplitIds.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkSourceReader.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkSourceSplitReader.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/MutableRecordAndPos.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/RecordAndPos.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/SplitScanner.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/FlinkSourceFetcherManager.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/class-use/FlinkSourceFetcherManager.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/HybridSnapshotLogSplit.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/HybridSnapshotLogSplitState.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/LogSplit.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/LogSplitState.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SnapshotSplit.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitBase.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitSerializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitState.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/HybridSnapshotLogSplit.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/HybridSnapshotLogSplitState.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/LogSplit.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/LogSplitState.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SnapshotSplit.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitBase.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitSerializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitState.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/FlussSourceEnumeratorStateSerializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/SourceEnumeratorState.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/class-use/FlussSourceEnumeratorStateSerializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/class-use/SourceEnumeratorState.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/LakeTieringJobBuilder.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/class-use/LakeTieringJobBuilder.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/CommittableMessage.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/CommittableMessageTypeInfo.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/FlussTableLakeSnapshotCommitter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitOperator.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitOperatorFactory.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitterInitContext.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/CommittableMessage.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/CommittableMessageTypeInfo.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/FlussTableLakeSnapshotCommitter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitOperator.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitOperatorFactory.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitterInitContext.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/FailedTieringEvent.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/FinishedTieringEvent.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/TieringRestoreEvent.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/FailedTieringEvent.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/FinishedTieringEvent.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/TieringRestoreEvent.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResult.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultEmitter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultSerializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultTypeInfo.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSource.Builder.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSource.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSourceOptions.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSourceReader.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSplitReader.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringWriterInitContext.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResult.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultEmitter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultSerializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultTypeInfo.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSource.Builder.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSource.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSourceOptions.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSourceReader.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSplitReader.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringWriterInitContext.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/TieringSourceEnumerator.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/class-use/TieringSourceEnumerator.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringLogSplit.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSnapshotSplit.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplit.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplitGenerator.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplitSerializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringLogSplit.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSnapshotSplit.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplit.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplitGenerator.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplitSerializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSourceEnumeratorState.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSourceEnumeratorStateSerializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSplitState.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSourceEnumeratorState.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSourceEnumeratorStateSerializer.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSplitState.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/CatalogExceptionUtils.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/CatalogPropertiesUtils.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/DataLakeUtils.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConnectorOptionsUtils.StartupOptions.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConnectorOptionsUtils.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConversions.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkRowToFlussRowConverter.FlussSerializationConverter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkRowToFlussRowConverter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkUtils.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToFlinkRowConverter.FlussDeserializationConverter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToFlinkRowConverter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToJsonConverters.FlussRowToJsonConverter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToJsonConverters.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/LakeSourceUtils.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PojoToRowConverter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.FieldEqual.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.ValueConversion.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/TimestampFormat.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/CatalogExceptionUtils.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/CatalogPropertiesUtils.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/DataLakeUtils.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConnectorOptionsUtils.StartupOptions.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConnectorOptionsUtils.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConversions.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkRowToFlussRowConverter.FlussSerializationConverter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkRowToFlussRowConverter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkUtils.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToFlinkRowConverter.FlussDeserializationConverter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToFlinkRowConverter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToJsonConverters.FlussRowToJsonConverter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToJsonConverters.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/LakeSourceUtils.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PojoToRowConverter.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.FieldEqual.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.ValueConversion.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/TimestampFormat.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-use.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/overview-summary.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/overview-tree.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/package-search-index.js create mode 100644 website/static/javadoc/0.8-SNAPSHOT/package-search-index.zip create mode 100644 website/static/javadoc/0.8-SNAPSHOT/packages create mode 100644 website/static/javadoc/0.8-SNAPSHOT/resources/glass.png create mode 100644 website/static/javadoc/0.8-SNAPSHOT/resources/x.png create mode 100644 website/static/javadoc/0.8-SNAPSHOT/script.js create mode 100644 website/static/javadoc/0.8-SNAPSHOT/search.js create mode 100644 website/static/javadoc/0.8-SNAPSHOT/serialized-form.html create mode 100644 website/static/javadoc/0.8-SNAPSHOT/stylesheet.css create mode 100644 website/static/javadoc/0.8-SNAPSHOT/type-search-index.js create mode 100644 website/static/javadoc/0.8-SNAPSHOT/type-search-index.zip create mode 100644 website/static/javadoc/index.html diff --git a/pom.xml b/pom.xml index 9ae2a7527a..32b95848fb 100644 --- a/pom.xml +++ b/pom.xml @@ -162,6 +162,11 @@ -Djdk.reflect.useDirectMethodHandle=false -Dio.netty.tryReflectionSetAccessible=true + + false + false + true + 3.6.3 @@ -531,6 +536,57 @@ + + javadoc-aggregate + + + javadoc.aggregate + + + + + fluss-client + fluss-flink + + + + + org.apache.maven.plugins + maven-javadoc-plugin + ${javadoc.version} + + + 11 + UTF-8 + Apache Fluss ${project.version} API + Apache Fluss ${project.version} API + protected + true + true + true + true + true + none + false + false + true + false + + -J-Xmx2g + -J--illegal-access=permit + + *.internal.*:*.impl.*:*.test.*:*.tests.*:*.generated.*:*.benchmark.*:*.protogen.* + + + + https://docs.oracle.com/en/java/javase/11/docs/api/ + https://nightlies.apache.org/flink/flink-docs-master/api/java/ + + + + + + @@ -945,6 +1001,81 @@ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + diff --git a/tools/refresh_javadoc.sh b/tools/refresh_javadoc.sh new file mode 100755 index 0000000000..3610a8c78b --- /dev/null +++ b/tools/refresh_javadoc.sh @@ -0,0 +1,62 @@ +#!/bin/bash + +# Licensed to the Apache Software Foundation (ASF) under one or more +# contributor license agreements. See the NOTICE file distributed with +# this work for additional information regarding copyright ownership. +# The ASF licenses this file to You under the Apache License, Version 2.0 +# (the "License"); you may not use this file except in compliance with +# the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +set -e + +# Parse arguments +while [[ $# -gt 0 ]]; do + case $1 in + -v|--version) VERSION_OVERRIDE="$2"; shift 2 ;; + -h|--help) echo "Usage: $0 [-v VERSION] [-h|--help]"; exit 0 ;; + *) echo "Unknown option: $1"; exit 1 ;; + esac +done + +# Find project root and get version +cd "$(dirname "$0")" +while [[ ! -f "pom.xml" && "$(pwd)" != "/" ]]; do cd ..; done +[[ ! -f "pom.xml" ]] && { echo "Error: pom.xml not found"; exit 1; } + +VERSION=${VERSION_OVERRIDE:-$(mvn help:evaluate -Dexpression=project.version -q -DforceStdout)} +SHORT_VERSION=$(echo "$VERSION" | cut -d. -f1-2) +OUTPUT_DIR="website/static/javadoc/$SHORT_VERSION" + +echo "Generating Javadoc $VERSION" + +# Setup output +rm -rf "$OUTPUT_DIR" +mkdir -p "$OUTPUT_DIR" + +# Generate javadoc +export MAVEN_OPTS="--add-exports=java.rmi/sun.rmi.registry=ALL-UNNAMED --add-opens=java.base/java.lang=ALL-UNNAMED" + +if mvn javadoc:aggregate -Pjavadoc-aggregate -Djavadoc.aggregate=true -q 2>/dev/null; then + [[ -d "target/site/apidocs" ]] && cp -r target/site/apidocs/* "$OUTPUT_DIR/" || { echo "No javadoc generated"; exit 1; } +else + echo "Javadoc generation failed"; exit 1 +fi + +# Create redirect +mkdir -p "website/static/javadoc" +cat > "website/static/javadoc/index.html" << EOF + + +Apache Fluss ${SHORT_VERSION} API +EOF + +echo "Javadoc generated: $OUTPUT_DIR" +echo "URL: http://localhost:3000/javadoc/$SHORT_VERSION/" \ No newline at end of file diff --git a/website/static/javadoc/0.8-SNAPSHOT/allclasses-index.html b/website/static/javadoc/0.8-SNAPSHOT/allclasses-index.html new file mode 100644 index 0000000000..4789b53dd7 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/allclasses-index.html @@ -0,0 +1,1627 @@ + + + + + +All Classes (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

All Classes

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/allclasses.html b/website/static/javadoc/0.8-SNAPSHOT/allclasses.html new file mode 100644 index 0000000000..8594df6d9e --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/allclasses.html @@ -0,0 +1,268 @@ + + + + + +All Classes (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + +
+

All Classes

+
+ +
+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/allpackages-index.html b/website/static/javadoc/0.8-SNAPSHOT/allpackages-index.html new file mode 100644 index 0000000000..0a66089a38 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/allpackages-index.html @@ -0,0 +1,350 @@ + + + + + +All Packages (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

All Packages

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/argfile b/website/static/javadoc/0.8-SNAPSHOT/argfile new file mode 100644 index 0000000000..d90e6ba99d --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/argfile @@ -0,0 +1,1401 @@ +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/HistogramStatistics.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/NoOpCounter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/MetricView.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/Gauge.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/reporter/ScheduledMetricReporter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/reporter/ReporterSetup.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/reporter/MetricReporter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/reporter/MetricReporterPlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/CharacterFilter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/Metric.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/DescriptiveStatisticsHistogramStatistics.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/DescriptiveStatisticsHistogram.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/utils/MetricGroupUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/groups/AbstractMetricGroup.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/groups/GenericValueMetricGroup.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/groups/GenericKeyMetricGroup.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/groups/ProxyMetricGroup.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/groups/FrontMetricGroup.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/groups/ReporterScopedSettings.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/groups/GenericMetricGroup.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/groups/MetricGroup.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/Counter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/SimpleCounter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/Meter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/MetricType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/registry/MetricRegistry.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/registry/MetricRegistryImpl.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/registry/NOPMetricRegistry.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/ThreadSafeSimpleCounter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/MetricViewUpdater.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/Histogram.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/MetricNames.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/MeterView.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/rocksdb/RocksDBHandle.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/rocksdb/RocksDBOperationUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/rocksdb/RocksIteratorWrapper.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/cluster/TabletServerInfo.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/cluster/ServerType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/cluster/Cluster.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/cluster/ServerNode.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/cluster/Endpoint.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/cluster/BucketLocation.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/PredicateBuilder.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/LessThan.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/Contains.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/CompareUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/NotIn.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/EndsWith.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/LeafPredicate.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/CompoundPredicate.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/NullFalseLeafBinaryFunction.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/IsNotNull.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/FieldRef.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/In.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/PredicateVisitor.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/UnsupportedExpression.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/Predicate.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/LessOrEqual.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/And.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/NotEqual.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/IsNull.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/PartitionPredicateVisitor.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/LeafFunction.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/GreaterOrEqual.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/FunctionVisitor.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/GreaterThan.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/StartsWith.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/Equal.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/Or.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/LeafUnaryFunction.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/TinyIntType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/DataType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/LocalZonedTimestampType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/DoubleType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/StringType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/DataTypeVisitor.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/BinaryType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/DecimalType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/RowType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/CharType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/MapType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/DataTypeDefaultVisitor.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/DataTypeFamily.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/DataTypeChecks.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/BytesType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/DataTypeRoot.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/DateType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/BigIntType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/SmallIntType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/IntType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/DataTypes.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/TimeType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/TimestampType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/FloatType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/DataField.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/BooleanType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/ArrayType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/FileLogProjection.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/TimestampAndOffset.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/KvRecord.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/ChangeType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/GenericRecord.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/FileLogInputStream.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/DefaultValueRecord.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/IndexedLogRecord.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/KvRecordReadContext.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/bytesview/MultiBytesView.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/bytesview/BytesView.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/bytesview/MemorySegmentBytesView.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/bytesview/FlussFileRegion.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/bytesview/FileRegionBytesView.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/bytesview/ByteBufBytesView.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/LogRecord.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/MemoryLogRecords.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/ChangeTypeVectorWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/ValueRecordReadContext.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/DefaultLogRecordBatch.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/LogRecords.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/LogRecordBatchFormat.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/KvRecordBatch.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/FileLogRecords.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/MemoryLogRecordsIndexedBuilder.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/FileChannelChunk.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/DefaultValueRecordBatch.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/BytesViewLogRecords.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/MemoryLogRecordsArrowBuilder.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/ValueRecord.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/DefaultKvRecord.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/MemorySegmentLogInputStream.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/KvRecordBatchBuilder.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/DefaultKvRecordBatch.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/LogInputStream.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/LogRecordReadContext.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/LogRecordBatchIterator.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/send/FileRegionSend.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/send/MultiSend.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/send/ByteBufSend.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/send/WritableOutput.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/send/Send.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/send/SendWritableOutput.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/send/ByteBufWritableOutput.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/ChangeTypeVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/ValueRecordBatch.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/LogRecordBatch.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/OutputView.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/MemorySegmentPool.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/MemorySegmentInputView.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/MemorySegmentOutputView.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/AbstractPagedOutputView.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/MemorySegmentWritable.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/LazyMemorySegmentPool.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/MemoryUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/ManagedPagedOutputView.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/MemorySegment.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/UnmanagedPagedOutputView.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/PreAllocatedPagedOutputView.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/InputView.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/MemorySize.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/StructuredOptionsSplitter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/GlobalConfiguration.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/ConfigBuilder.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/Configuration.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/AutoPartitionTimeUnit.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/TableConfig.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/FallbackKey.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/ReadableConfig.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/ConfigOption.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/ConfigurationUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/FlussConfigUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/ConfigOptions.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/Password.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/AuthenticationFactory.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/plain/PlainSaslServer.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/plain/PlainLoginModule.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/plain/PlainAuthenticateCallback.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/plain/PlainSaslServerProvider.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/plain/PlainServerCallbackHandler.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/authenticator/SaslServerAuthenticator.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/authenticator/SaslClientAuthenticator.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/authenticator/SaslAuthenticationPlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/jaas/DefaultLogin.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/jaas/SaslClientCallbackHandler.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/jaas/JaasConfig.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/jaas/JaasContext.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/jaas/AuthenticateCallbackHandler.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/jaas/Login.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/jaas/LoginManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/jaas/SaslServerFactory.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/AuthenticationPlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/ClientAuthenticationPlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/ServerAuthenticator.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/PlainTextAuthenticationPlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/ClientAuthenticator.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/ServerAuthenticationPlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/acl/PermissionType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/acl/OperationType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/acl/Resource.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/acl/ResourceFilter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/acl/AclBinding.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/acl/AccessControlEntry.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/acl/AccessControlEntryFilter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/acl/FlussPrincipal.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/acl/ResourceType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/acl/AclBindingFilter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/plugin/PluginUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/plugin/PluginLoader.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/plugin/PluginManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/plugin/Plugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/plugin/PluginDescriptor.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/plugin/DefaultPluginManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/plugin/PluginConfig.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/plugin/DirectoryBasedPluginFinder.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/plugin/PluginFinder.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/annotation/Internal.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/annotation/VisibleForTesting.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/annotation/PublicEvolving.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/annotation/PublicStable.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/annotation/PublicUnstable.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/Preconditions.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/CloseableRegistry.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/AbstractAutoCloseableRegistry.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/IOUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/StringUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/PartitionUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/ExecutorUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/ByteBufferReadableChannel.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/types/Tuple.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/types/Either.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/types/Tuple2.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/OperatingSystem.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/InstantiationUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/TypeUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/UnionIterator.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/MapUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/FileUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/SerializedValue.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/ArrayUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/CollectionUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/crc/Checksums.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/crc/Crc32C.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/crc/PureJavaCrc32C.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/crc/Java.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/MathUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/VarLengthUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/NetUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/CloseableIterator.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/function/FunctionUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/function/SupplierWithException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/function/QuadConsumer.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/function/ThrowingConsumer.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/function/RunnableWithException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/function/ThrowingRunnable.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/function/CheckedSupplier.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/function/BiFunction.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/function/FunctionWithException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/ExponentialBackoff.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/FatalExitExceptionHandler.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/FlussPaths.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/ArrowUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/PagedMemorySegmentWritableChannel.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/DateTimeUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/PropertiesUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/SerializableObject.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/json/TablePartitionJsonSerde.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/json/BucketOffsetJsonSerde.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/json/JsonSerdeUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/json/JsonDeserializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/json/DatabaseDescriptorJsonSerde.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/json/JsonSerializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/json/ColumnJsonSerde.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/json/SchemaJsonSerde.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/json/TableDescriptorJsonSerde.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/json/DataTypeJsonSerde.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/UnsafeUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/DecimalUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/FileLock.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/TimeUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/WrappingProxy.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/BytesUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/ExceptionUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/Projection.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/AutoCloseableAsync.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/concurrent/Scheduler.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/concurrent/FlussScheduler.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/concurrent/ExecutorThreadFactory.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/concurrent/ShutdownableThread.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/concurrent/FutureUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/concurrent/LockUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/concurrent/ThreadUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/concurrent/Executors.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/BinaryStringUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/TemporaryClassLoaderContext.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/log/FairBucketStatusMap.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/log/ByteBufferUnmapper.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/AbstractIterator.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/MurmurHashUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/AutoPartitionStrategy.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/WrappingProxyUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/ProtoCodecUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/MemorySegmentWritableChannel.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/EncodingUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/clock/SystemClock.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/clock/ManualClock.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/clock/Clock.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/CopyOnWriteMap.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/classloading/ComponentClassLoader.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/BinarySegmentUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/DoubleColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/FloatColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/DecimalColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/ColumnarRow.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/ColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/BytesColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/TimestampLtzColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/BooleanColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/TimestampNtzColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/VectorizedColumnBatch.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/ShortColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/ByteColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/IntColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/LongColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/InternalRow.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/compacted/CompactedRowDeserializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/compacted/CompactedRowReader.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/compacted/CompactedRowWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/compacted/CompactedRow.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/compacted/CompactedKeyWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowDoubleWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowFieldWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowTimestampLtzWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowBinaryWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowSmallIntWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowTimeWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowIntWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowDateWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowVarCharWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowBigIntWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowBooleanWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowDecimalWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowFloatWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowTimestampNtzWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowVarBinaryWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowTinyIntWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowTinyIntColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowDateColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowBinaryColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowTimeColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowVarCharColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowTimestampNtzColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowFloatColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowDoubleColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowVarBinaryColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowIntColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowDecimalColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowSmallIntColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowBigIntColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowTimestampLtzColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowBooleanColumnVector.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/ArrowWriterProvider.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/ArrowWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/ArrowWriterPool.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/ArrowReader.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/GenericRow.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/Decimal.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/MemoryAwareGetters.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/BinaryString.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/encode/IndexedRowEncoder.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/encode/ValueEncoder.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/encode/paimon/PaimonBinaryRowWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/encode/paimon/PaimonKeyEncoder.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/encode/KeyEncoder.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/encode/RowEncoder.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/encode/iceberg/IcebergBinaryRowWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/encode/iceberg/IcebergKeyEncoder.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/encode/ValueDecoder.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/encode/CompactedKeyEncoder.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/encode/CompactedRowEncoder.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/indexed/IndexedRowReader.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/indexed/IndexedRowWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/indexed/IndexedRow.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/decode/CompactedRowDecoder.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/decode/IndexedRowDecoder.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/decode/RowDecoder.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/ProjectedRow.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/BinaryRow.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/TimestampNtz.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/TimestampLtz.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/NullAwareGetters.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/BinarySection.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/lakestorage/LakeCatalog.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/lakestorage/LakeStoragePlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/lakestorage/PluginLakeStorageWrapper.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/lakestorage/LakeStorage.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/lakestorage/LakeStoragePluginSetUp.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/serializer/SimpleVersionedSerializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/source/Planner.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/source/SortedRecordReader.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/source/LakeSplit.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/source/LakeSource.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/source/RecordReader.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/writer/LakeWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/writer/WriterInitContext.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/writer/LakeTieringFactory.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/writer/SupportsRecordBatchWrite.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/committer/LakeCommitter.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/committer/CommitterInitContext.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/committer/BucketOffset.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/committer/CommittedLakeSnapshot.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/batch/ArrowRecordBatch.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/batch/RecordBatch.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/compression/FlussLZ4BlockOutputStream.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/compression/ZstdArrowCompressionCodec.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/compression/Lz4ArrowCompressionCodec.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/compression/ArrowCompressionInfo.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/compression/FlussLZ4BlockInputStream.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/compression/ArrowCompressionFactory.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/compression/ArrowCompressionRatioEstimator.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/compression/ArrowCompressionType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/bucketing/PaimonBucketingFunction.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/bucketing/FlussBucketingFunction.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/bucketing/IcebergBucketingFunction.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/bucketing/BucketingFunction.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/rpc/messages/ErrorMessage.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/rpc/messages/ApiMessage.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/token/SecurityTokenReceiver.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/token/ObtainedSecurityToken.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/token/Credentials.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/token/CredentialsJsonSerde.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/FileStatus.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/WrappingProxyCloseable.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/FSDataInputStreamWrapper.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/FileSystem.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/UnsupportedFileSystemSchemeException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/FileSystemSafetyNet.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/FSDataInputStream.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/ClosingFSDataInputStream.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/FsPathAndFileName.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/utils/FileDownloadUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/utils/FileDownloadSpec.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/SafetyNetCloseableRegistry.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/local/LocalFileSystem.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/local/LocalSecurityTokenReceiver.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/local/LocalFileStatus.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/local/LocalFileSystemPlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/local/LocalDataInputStream.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/local/LocalDataOutputStream.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/ClosingFSDataOutputStream.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/FSDataOutputStreamWrapper.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/FsPath.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/PluginFileSystemWrapper.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/FileSystemPlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/FSDataOutputStream.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/SafetyNetWrapperFileSystem.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/ApiException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/FlussException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidMetadataException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/BufferExhaustedException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/DatabaseNotExistException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/TableNotExistException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/DisconnectException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidColumnProjectionException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidServerRackInfoException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/IllegalConfigurationException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/LeaderNotAvailableException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/FetchException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/PartitionAlreadyExistsException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/SecurityTokenException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/FlussRuntimeException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/RecordTooLargeException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/DatabaseAlreadyExistException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/TooManyPartitionsException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/UnknownServerException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/KvSnapshotNotExistException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidTargetColumnException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidTableException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/LakeStorageNotConfiguredException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/StorageException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/OutOfOrderSequenceException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidRecordException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/KvStorageException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/RemoteResourceNotFoundException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/IndexOffsetOverflowException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/CorruptSnapshotException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/LakeTableSnapshotNotExistException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/LakeTableAlreadyExistException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/NotLeaderOrFollowerException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/StaleMetadataException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/TooManyBucketsException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidDatabaseException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/PartitionNotExistException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/CorruptRecordException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/UnknownTableOrBucketException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/TableAlreadyExistException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidPartitionException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/RetriableAuthenticationException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidConfigException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidUpdateVersionException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/FencedLeaderEpochException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/AuthorizationException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/CorruptMessageException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/TableNotPartitionedException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidTimestampException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidCoordinatorException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidRequiredAcksException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/UnknownWriterIdException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/LogOffsetOutOfRangeException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/DatabaseNotEmptyException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/TimeoutException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/LogSegmentOffsetOverflowException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/WakeupException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidBucketsException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/AuthenticationException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/RemoteStorageException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/NonPrimaryKeyTableException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidOffsetException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/OperationNotAttemptedException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/LogStorageException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/UnsupportedVersionException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidReplicationFactorException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/NetworkException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/TraversableOnceException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/NotEnoughReplicasException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/DuplicateSequenceException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/LogRecordAppendFailedException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/NotEnoughReplicasAfterAppendException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/RetriableException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/SecurityDisabledException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/FencedTieringEpochException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/SchemaNotExistException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/ValidationException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/PartitionInfo.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/TableException.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/TablePath.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/ResolvedPartitionSpec.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/TableBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/DataLakeFormat.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/MergeEngineType.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/Schema.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/TableBucketReplica.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/TablePartition.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/LogFormat.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/SchemaInfo.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/DatabaseDescriptor.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/TableInfo.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/PartitionSpec.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/DatabaseInfo.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/KvFormat.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/TableDescriptor.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/PhysicalTablePath.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/remote/RemoteLogFetchInfo.java' +'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/remote/RemoteLogSegment.java' +'/Users/mehulbatra/Documents/fluss/fluss-metrics/fluss-metrics-prometheus/src/main/java/org/apache/fluss/metrics/prometheus/PrometheusReporter.java' +'/Users/mehulbatra/Documents/fluss/fluss-metrics/fluss-metrics-prometheus/src/main/java/org/apache/fluss/metrics/prometheus/PrometheusReporterPlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/ReadyWriteBatch.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/IdempotenceBucketEntry.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/WriteRecord.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/BucketAssigner.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/IdempotenceManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/RoundRobinBucketAssigner.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/StickyBucketAssigner.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/ArrowLogWriteBatch.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/IdempotenceBucketMap.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/IncompleteBatches.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/HashBucketAssigner.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/WriteBatch.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/Sender.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/DynamicWriteBatchSizeEstimator.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/StaticBucketAssigner.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/RecordAccumulator.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/KvWriteBatch.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/DynamicPartitionCreator.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/WriterClient.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/WriteCallback.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/IndexedLogWriteBatch.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/WriteFormat.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/DynamicBucketAssigner.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/token/SecurityTokenProvider.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/token/DefaultSecurityTokenManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/token/DefaultSecurityTokenProvider.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/token/SecurityTokenManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/token/SecurityTokenReceiverRepository.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/metrics/WriterMetricGroup.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/metrics/ScannerMetricGroup.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/FlussConnection.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/admin/Admin.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/admin/FlussAdmin.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/admin/ListOffsetsResult.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/admin/OffsetSpec.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/admin/DropAclsResult.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/admin/CreateAclsResult.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/PrefixLookupBatch.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/LookupType.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/LookupSender.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/AbstractLookupQuery.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/Lookup.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/Lookuper.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/LookupQueue.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/LookupBatch.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/PrefixLookup.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/LookupResult.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/LookupQuery.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/PrimaryKeyLookuper.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/LookupClient.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/PrefixLookupQuery.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/TableLookup.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/PrefixKeyLookuper.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/AbstractLookupBatch.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/utils/ClientRpcMessageUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/utils/ClientUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/utils/MetadataUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/getter/PartitionGetter.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/FlussTable.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/Table.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/AppendResult.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/AbstractTableWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/UpsertWriterImpl.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/TableUpsert.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/Append.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/UpsertWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/TableAppend.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/TableWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/AppendWriterImpl.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/Upsert.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/AppendWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/DeleteResult.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/UpsertWrite.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/UpsertResult.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/Scan.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/RemoteFileDownloader.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/batch/KvSnapshotBatchScanner.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/batch/SnapshotFilesReader.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/batch/BatchScanUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/batch/LimitBatchScanner.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/batch/BatchScanner.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/CompletedPendingFetch.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/LogFetcher.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/RemoteCompletedFetch.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/RemoteLogDownloader.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/DefaultCompletedFetch.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/PendingFetch.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/LogFetchCollector.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/LogScannerImpl.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/LogScanner.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/CompletedFetch.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/LogScan.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/LogFetchBuffer.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/RemotePendingFetch.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/LogScannerStatus.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/ScanRecords.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/RemoteLogDownloadFuture.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/BucketScanStatus.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/TableScan.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/ScanRecord.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/Connection.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/ConnectionFactory.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/metadata/KvSnapshots.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/metadata/KvSnapshotMetadata.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/metadata/LakeSnapshot.java' +'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/metadata/MetadataUpdater.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/metrics/ConnectionMetricGroup.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/metrics/ClientMetricGroup.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/RpcGateway.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/RpcClient.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/util/CommonRpcMessageUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/GatewayClientProxy.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/entity/ProduceLogResultForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/entity/LookupResultForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/entity/ListOffsetsResultForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/entity/PutKvResultForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/entity/LimitScanResultForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/entity/FetchLogResultForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/entity/ResultForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/entity/PrefixLookupResultForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/entity/WriteResultForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/RpcGatewayService.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/protocol/ResponseType.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/protocol/MessageCodec.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/protocol/ApiMethod.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/protocol/RPC.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/protocol/ApiKeys.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/protocol/ApiError.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/protocol/NetworkProtocolPlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/protocol/Errors.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/protocol/ApiManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/protocol/RequestType.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/NettyLogger.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/FlussRequestHandler.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/RequestProcessor.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/FlussRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/RequestsMetrics.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/RequestHandler.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/ShutdownRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/RequestProcessorPool.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/NettyServerHandler.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/RpcRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/RequestChannel.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/ServerChannelInitializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/FlussProtocolPlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/NettyServer.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/Session.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/NettyChannelInitializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/NettyMetrics.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/NettyUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/client/ClientChannelInitializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/client/NettyClient.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/client/NettyClientHandler.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/client/ClientHandlerCallback.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/client/ServerConnection.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/client/ServerApiVersions.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/gateway/CoordinatorGateway.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/gateway/TabletServerGateway.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/gateway/AdminReadOnlyGateway.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/gateway/AdminGateway.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/RpcServer.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CreatePartitionResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/DropTableResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/DatabaseExistsRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbCommitLakeTableSnapshotRespForTable.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/DropTableRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PrefixLookupRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbStopReplicaRespForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbRemotePathAndLocalFile.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/DropAclsResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbFetchLogReqForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetTableSchemaResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/DropDatabaseRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbLakeTableOffsetForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbProduceLogRespForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/AdjustIsrResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbPartitionSpec.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ListAclsRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/StopReplicaResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/DropAclsRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbLookupReqForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbAdjustIsrRespForTable.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbPutKvReqForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbNotifyLeaderAndIsrReqForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbBucketMetadata.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetTableSchemaRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbCreateAclRespInfo.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbFetchLogReqForTable.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/NotifyRemoteLogOffsetsResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbHeartbeatRespForTable.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ListPartitionInfosRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/TableExistsRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbValueList.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PutKvResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbKvSnapshot.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbAclFilter.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ListTablesResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/NotifyKvSnapshotOffsetResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbTableBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/NotifyRemoteLogOffsetsRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CreateTableResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CreateAclsRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ListOffsetsResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/NotifyKvSnapshotOffsetRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetKvSnapshotMetadataResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/LimitScanResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetLatestKvSnapshotsRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/NotifyLeaderAndIsrResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/LookupResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbProduceLogReqForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/AdjustIsrRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetDatabaseInfoResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbLakeSnapshotForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/DropPartitionResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbLakeTableSnapshotInfo.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbPutKvRespForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/NotifyLakeTableOffsetRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/NotifyLeaderAndIsrRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/DatabaseExistsResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbRemoteLogSegment.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ApiVersionsRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CommitKvSnapshotResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbPhysicalTablePath.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbApiVersion.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ListPartitionInfosResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CreatePartitionRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbRemoteLogFetchInfo.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CommitKvSnapshotRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbPartitionInfo.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ProduceLogRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbNotifyLeaderAndIsrRespForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/MetadataResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ListOffsetsRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ListTablesRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetLatestKvSnapshotsResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbDropAclsFilterResult.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbLookupRespForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CommitLakeTableSnapshotRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbFetchLogRespForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CreateDatabaseRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CreateTableRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbKeyValue.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbLakeTieringTableInfo.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CommitRemoteLogManifestRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/AuthenticateRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbHeartbeatReqForTable.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/NotifyLakeTableOffsetResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbListOffsetsRespForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetLatestLakeSnapshotResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbAdjustIsrRespForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PrefixLookupResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetKvSnapshotMetadataRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CreateAclsResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/UpdateMetadataResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbTablePath.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/InitWriterRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ApiVersionsResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/DropPartitionRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/DropDatabaseResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/LakeTieringHeartbeatResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/FetchLogRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbAclInfo.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/LakeTieringHeartbeatRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbValue.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbNotifyLakeTableOffsetReqForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbServerNode.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CommitRemoteLogManifestResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetFileSystemSecurityTokenRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/LimitScanRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ListAclsResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ErrorResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/FetchLogResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/LookupRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetTableInfoRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbPrefixLookupRespForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbAdjustIsrReqForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/UpdateMetadataRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbTableMetadata.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetFileSystemSecurityTokenResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbDropAclsMatchingAcl.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbPrefixLookupReqForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbFetchLogRespForTable.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbPartitionMetadata.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ListDatabasesResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbStopReplicaReqForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CommitLakeTableSnapshotResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbAdjustIsrReqForTable.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/TableExistsResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ProduceLogResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ListDatabasesRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetLatestLakeSnapshotRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetDatabaseInfoRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PutKvRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CreateDatabaseResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/StopReplicaRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/InitWriterResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/MetadataRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/AuthenticateResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetTableInfoResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-dist/src/main/java/org/apache/fluss/dist/DummyClass.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-hadoop/src/main/java/org/apache/fluss/fs/hdfs/HadoopSecurityTokenReceiver.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-hadoop/src/main/java/org/apache/fluss/fs/hdfs/HadoopDataInputStream.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-hadoop/src/main/java/org/apache/fluss/fs/hdfs/HadoopFileSystem.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-hadoop/src/main/java/org/apache/fluss/fs/hdfs/utils/HadoopUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-hadoop/src/main/java/org/apache/fluss/fs/hdfs/HadoopFileStatus.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-hadoop/src/main/java/org/apache/fluss/fs/hdfs/HadoopFsPlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-hadoop/src/main/java/org/apache/fluss/fs/hdfs/HadoopDataOutputStream.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-oss/src/main/java/org/apache/fluss/fs/oss/token/OSSSecurityTokenReceiver.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-oss/src/main/java/org/apache/fluss/fs/oss/token/OSSSecurityTokenProvider.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-oss/src/main/java/org/apache/fluss/fs/oss/token/DynamicTemporaryOssCredentialsProvider.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-oss/src/main/java/org/apache/fluss/fs/oss/OSSFileSystem.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-oss/src/main/java/org/apache/fluss/fs/oss/OSSFileSystemPlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-s3/src/main/java/org/apache/fluss/fs/s3/token/S3ADelegationTokenReceiver.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-s3/src/main/java/org/apache/fluss/fs/s3/token/S3DelegationTokenProvider.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-s3/src/main/java/org/apache/fluss/fs/s3/token/S3DelegationTokenReceiver.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-s3/src/main/java/org/apache/fluss/fs/s3/token/DynamicTemporaryAWSCredentialsProvider.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-s3/src/main/java/org/apache/fluss/fs/s3/S3FileSystemPlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-s3/src/main/java/org/apache/fluss/fs/s3/S3FileSystem.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-s3/src/main/java/org/apache/fluss/fs/s3/S3AFileSystemPlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-s3/src/main/java/com/amazonaws/services/s3/model/transform/XmlResponsesSaxParser.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-gs/src/main/java/org/apache/fluss/fs/gs/GSFileSystemPlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-gs/src/main/java/org/apache/fluss/fs/gs/GSFileSystem.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-obs/src/main/java/org/apache/fluss/fs/obs/token/DynamicTemporaryOBSCredentialsProvider.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-obs/src/main/java/org/apache/fluss/fs/obs/token/OBSSecurityTokenReceiver.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-obs/src/main/java/org/apache/fluss/fs/obs/token/OBSSecurityTokenProvider.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-obs/src/main/java/org/apache/fluss/fs/obs/OBSFileSystemPlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-obs/src/main/java/org/apache/fluss/fs/obs/OBSFileSystem.java' +'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/junit/parameterized/Parameter.java' +'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/junit/parameterized/ParameterizedTestExtension.java' +'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/junit/parameterized/Parameters.java' +'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/AllCallbackWrapper.java' +'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/FlussCompletableFutureAssert.java' +'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/CheckedThread.java' +'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/CustomExtension.java' +'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/OneShotLatch.java' +'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/TestLoggerExtension.java' +'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/ManuallyTriggeredScheduledExecutorService.java' +'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/ClassLoaderUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/ContextClassLoaderExtension.java' +'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/ScheduledTask.java' +'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/CommonTestUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/TestExecutorExtension.java' +'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/FlussAssertions.java' +'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/EachCallbackWrapper.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metrics/group/TableMetricGroup.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metrics/group/TabletServerMetricGroup.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metrics/group/CoordinatorMetricGroup.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metrics/group/BucketMetricGroup.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metrics/ServerMetricUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/tablet/TabletServer.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/tablet/TabletService.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/ZkSequenceIDCounter.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/CuratorFrameworkWithUnhandledErrorListener.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/ZooKeeperClient.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/ZkAsyncResponse.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/ZkAsyncRequest.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/LakeTableSnapshot.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/CoordinatorAddress.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/BucketSnapshotJsonSerde.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/PartitionAssignment.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/DatabaseRegistrationJsonSerde.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/DatabaseRegistration.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/TabletServerRegistration.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/TableAssignmentJsonSerde.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/TabletServerRegistrationJsonSerde.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/PartitionAssignmentJsonSerde.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/RemoteLogManifestHandle.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/TableAssignment.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/RemoteLogManifestHandleJsonSerde.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/BucketAssignment.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/LakeTableSnapshotJsonSerde.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/TableRegistrationJsonSerde.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/BucketSnapshot.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/LeaderAndIsr.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/ResourceAcl.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/CoordinatorAddressJsonSerde.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/ResourceAclJsonSerde.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/TableRegistration.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/ZkData.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/LeaderAndIsrJsonSerde.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/ZooKeeperUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/LakeTableTieringManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/CoordinatorEventProcessor.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/CoordinatorServer.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/CoordinatorContext.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/MetadataManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/TableManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/CompletedSnapshotStoreManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/statemachine/BaseState.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/statemachine/BucketState.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/statemachine/ReplicaState.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/statemachine/ReplicaLeaderElectionAlgorithms.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/statemachine/ReplicaStateMachine.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/statemachine/TableBucketStateMachine.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/CoordinatorChannelManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/CoordinatorRequestBatch.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/AutoPartitionManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/RemoteStorageCleaner.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/CoordinatorService.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/AdjustIsrReceivedEvent.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/CommitKvSnapshotEvent.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/CoordinatorEventManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/DeadTabletServerEvent.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/DropTableEvent.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/watcher/TableChangeWatcher.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/watcher/TabletServerChangeWatcher.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/FencedCoordinatorEvent.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/AccessContextEvent.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/NotifyLeaderAndIsrResponseReceivedEvent.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/NotifyKvSnapshotOffsetEvent.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/EventManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/DeleteReplicaResponseReceivedEvent.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/CreateTableEvent.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/CommitRemoteLogManifestEvent.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/ShutdownEventThreadEvent.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/CreatePartitionEvent.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/EventProcessor.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/DropPartitionEvent.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/CoordinatorEvent.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/CommitLakeTableSnapshotEvent.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/NewTabletServerEvent.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/rocksdb/RocksDBKv.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/rocksdb/RocksDBWriteBatchWrapper.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/rocksdb/RocksDBKvBuilder.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/rocksdb/RocksDBResourceContainer.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/SnapshotRunner.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/KvSnapshotDownloadSpec.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/CompletedSnapshotHandleStore.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/CompletedSnapshotHandle.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/SharedKvFileRegistryKey.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/SnapshotLocation.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/ZooKeeperCompletedSnapshotHandleStore.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/CompletedSnapshot.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/KvSnapshotDataTransfer.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/PeriodicSnapshotManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/DefaultCompletedKvSnapshotCommitter.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/SnapshotsCleaner.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/KvSnapshotDataUploader.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/KvSnapshotHandle.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/AsyncSnapshotCallable.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/SnapshotFileScope.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/SnapshotResources.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/DefaultSnapshotContext.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/KvSnapshotDataDownloader.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/SnapshotResult.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/RocksIncrementalSnapshot.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/CompletedKvSnapshotCommitter.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/CompletedSnapshotStore.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/KvFileHandleAndLocalPath.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/SnapshotContext.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/CompletedSnapshotJsonSerde.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/SnapshotResultSupplier.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/KvFileHandle.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/KvTabletSnapshotTarget.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/PlaceholderKvFileHandler.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/SharedKvFileRegistry.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/prewrite/KvPreWriteBuffer.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/KvTablet.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/wal/WalBuilder.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/wal/IndexWalBuilder.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/wal/ArrowWalBuilder.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/KvManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/partialupdate/PartialUpdater.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/partialupdate/PartialUpdaterCache.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/KvBatchWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/KvRecoverHelper.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/KvSnapshotResource.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/rowmerger/FirstRowRowMerger.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/rowmerger/VersionedRowMerger.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/rowmerger/DefaultRowMerger.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/rowmerger/RowMerger.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/TabletManagerBase.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/NotifyKvSnapshotOffsetData.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/NotifyLeaderAndIsrData.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/FetchReqInfo.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/LakeBucketOffset.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/CommitRemoteLogManifestData.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/DeleteReplicaResultForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/NotifyRemoteLogOffsetsData.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/StopReplicaData.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/AdjustIsrResultForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/CommitKvSnapshotData.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/NotifyLakeTableOffsetData.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/NotifyLeaderAndIsrResultForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/BatchRegisterLeadAndIsr.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/StopReplicaResultForBucket.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/RegisterTableBucketLeadAndIsrInfo.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/CommitLakeTableSnapshotData.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/LakeTieringTableInfo.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/ServerBase.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/ResourceGuard.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/ConfigurationParserUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/timer/TimerTaskList.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/timer/TimerTaskEntry.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/timer/Timer.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/timer/TimerTask.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/timer/TimingWheel.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/timer/DefaultTimer.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/ServerRpcMessageUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/ShutdownHookUtil.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/FatalErrorHandler.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/TableDescriptorValidation.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/TableAssignmentUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/RpcGatewayManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/SnapshotUtil.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/LakeStorageUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/cli/ParserResultFactory.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/cli/ServerConfigurationParserFactory.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/cli/ServerConfiguration.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/cli/CommandLineOptions.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/cli/CommandLineParser.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/authorizer/AclCreateResult.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/authorizer/AuthorizationPlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/authorizer/AclDeleteResult.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/authorizer/AuthorizerLoader.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/authorizer/ZkNodeChangeNotificationWatcher.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/authorizer/DefaultAuthorizer.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/authorizer/Authorizer.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/authorizer/DefaultAuthorizationPlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/authorizer/AbstractAuthorizer.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/authorizer/Action.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/OffsetIndex.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/StorageAction.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LocalLog.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/checkpoint/OffsetCheckpointFile.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/checkpoint/CheckpointFile.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LogAppendInfo.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LogSegments.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LazyIndex.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/WriterStateEntry.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LogManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/FetchIsolation.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LogSegment.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/FetchParams.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/SnapshotFile.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LogTablet.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/IndexEntry.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/TimestampOffset.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/WriterStateManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/FetchDataInfo.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/ListOffsetsParam.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LogReadInfo.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LogLoader.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/AbstractIndex.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/IndexSearchType.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/RollParams.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/WriterAppendInfo.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/TimeIndex.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LoadedLogOffsets.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LogOffsetMetadata.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/OffsetPosition.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LogOffsetSnapshot.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/remote/RemoteLogManifestJsonSerde.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/remote/RemoteLogIndexCache.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/remote/DefaultRemoteLogStorage.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/remote/LogSegmentFiles.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/remote/RemoteLogManifest.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/remote/RemoteLogTablet.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/remote/RemoteLogManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/remote/RemoteLogStorage.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/remote/FsRemoteLogOutputStream.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/remote/LogTieringTask.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/IsrState.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/fetcher/ReplicaFetcherManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/fetcher/ReplicaFetcherThread.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/fetcher/InitialFetchStatus.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/fetcher/RemoteLeaderEndpoint.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/fetcher/FetchLogContext.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/fetcher/BucketFetchStatus.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/fetcher/LeaderEndpoint.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/fetcher/DelayedItem.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/ReplicaManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/Replica.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/FollowerReplica.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/AdjustIsrManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/delay/DelayedFetchLog.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/delay/DelayedWrite.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/delay/DelayedOperationKey.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/delay/DelayedOperation.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/delay/DelayedTableBucketKey.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/delay/DelayedOperationManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/SequenceIDCounter.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/exception/FlussParseException.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/exception/KvBuildingException.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/exception/EndpointNotAvailableException.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/exception/CorruptIndexException.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metadata/ServerInfo.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metadata/TableMetadata.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metadata/ServerMetadataCache.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metadata/CoordinatorMetadataCache.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metadata/ServerMetadataSnapshot.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metadata/PartitionMetadata.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metadata/BucketMetadata.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metadata/TabletServerMetadataCache.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metadata/ClusterMetadata.java' +'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/RpcServiceBase.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/metrics/FlinkMeter.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/metrics/FlinkMetricRegistry.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/metrics/FlinkCounter.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/metrics/FlinkHistogram.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/metrics/FlinkGauge.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/catalog/FlinkCatalogFactory.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/catalog/FlinkCatalog.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/catalog/FlinkTableFactory.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/catalog/FlinkCatalogOptions.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/reader/MutableRecordAndPos.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/reader/FlinkSourceSplitReader.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/reader/fetcher/FlinkSourceFetcherManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/reader/FlinkRecordsWithSplitIds.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/reader/BoundedSplitReader.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/reader/SplitScanner.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/reader/RecordAndPos.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/reader/FlinkSourceReader.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/split/HybridSnapshotLogSplitState.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/split/SourceSplitBase.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/split/SourceSplitSerializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/split/LogSplitState.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/split/SnapshotSplit.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/split/LogSplit.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/split/SourceSplitState.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/split/HybridSnapshotLogSplit.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/metrics/FlinkSourceReaderMetrics.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/FlussSource.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/deserializer/JsonStringDeserializationSchema.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/deserializer/FlussDeserializationSchema.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/deserializer/DeserializerInitContextImpl.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/deserializer/RowDataDeserializationSchema.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/emitter/FlinkRecordEmitter.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/lookup/LookupNormalizer.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/lookup/FlinkLookupFunction.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/lookup/FlinkAsyncLookupFunction.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/state/FlussSourceEnumeratorStateSerializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/state/SourceEnumeratorState.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/FlinkSource.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/FlinkTableSource.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/event/PartitionsRemovedEvent.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/event/PartitionBucketsUnsubscribedEvent.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/enumerator/initializer/LatestOffsetsInitializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/enumerator/initializer/BucketOffsetsRetrieverImpl.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/enumerator/initializer/TimestampOffsetsInitializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/enumerator/initializer/SnapshotOffsetsInitializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/enumerator/initializer/NoStoppingOffsetsInitializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/enumerator/initializer/OffsetsInitializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/enumerator/initializer/EarliestOffsetsInitializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/enumerator/FlinkSourceEnumerator.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/FlussSourceBuilder.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/split/TieringSplitGenerator.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/split/TieringSplitSerializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/split/TieringSplit.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/split/TieringLogSplit.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/split/TieringSnapshotSplit.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/TieringSourceOptions.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/TableBucketWriteResultTypeInfo.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/TieringSource.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/TieringSourceReader.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/TableBucketWriteResultEmitter.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/TableBucketWriteResult.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/state/TieringSourceEnumeratorState.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/state/TieringSplitState.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/state/TieringSourceEnumeratorStateSerializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/TieringSplitReader.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/TieringWriterInitContext.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/TableBucketWriteResultSerializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/enumerator/TieringSourceEnumerator.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/LakeTieringJobBuilder.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/committer/FlussTableLakeSnapshot.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/committer/CommittableMessageTypeInfo.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/committer/TieringCommitOperator.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/committer/CommittableMessage.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/committer/TieringCommitterInitContext.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/committer/TieringCommitOperatorFactory.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/committer/FlussTableLakeSnapshotCommitter.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/event/FinishedTieringEvent.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/event/TieringRestoreEvent.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/event/FailedTieringEvent.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/FlussSink.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/FlinkStreamPartitioner.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/serializer/RowDataSerializationSchema.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/serializer/SerializerInitContextImpl.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/serializer/FlussSerializationSchema.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/FlinkTableSink.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/writer/AppendSinkWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/writer/FlinkSinkWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/writer/UpsertSinkWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/FlinkRowDataChannelComputer.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/ChannelComputer.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/FlussSinkBuilder.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/FlinkSink.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/FlinkTypeToFlussType.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/PojoToRowConverter.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/FlinkConversions.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/TimestampFormat.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/CatalogExceptionUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/FlussRowToJsonConverters.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/CatalogPropertiesUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/FlussRowToFlinkRowConverter.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/LakeSourceUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/FlussTypeToFlinkType.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/PushdownUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/FlinkConnectorOptionsUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/DataLakeUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/FlinkUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/FlinkRowToFlussRowConverter.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/TimeFormats.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/FlinkConnectorOptions.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/adapter/SingleThreadFetcherManagerAdapter.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/adapter/CatalogTableAdapter.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/adapter/SingleThreadMultiplexSourceReaderBaseAdapter.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/procedure/ProcedureManager.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/procedure/ListAclProcedure.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/procedure/AddAclProcedure.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/procedure/AbstractAclProcedure.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/procedure/ProcedureBase.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/procedure/DropAclProcedure.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/row/OperationType.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/row/RowWithOp.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/row/FlinkAsFlussRow.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/reader/KeyValueRow.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/reader/LakeSnapshotAndLogSplitScanner.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/reader/SortMergeReader.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/reader/LakeSnapshotScanner.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/reader/SeekableLakeSnapshotSplitScanner.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/reader/IndexedLakeSplitRecordIterator.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/split/LakeSnapshotAndFlussLogSplit.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/split/LakeSnapshotSplit.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/LakeTableFactory.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/LakeCatalog.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/LakeSplitStateInitializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/LakeSplitReaderGenerator.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/state/LakeSnapshotSplitState.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/state/LakeSnapshotAndFlussLogSplitState.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/LakeSplitGenerator.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/LakeSplitSerializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/LakeRecordRecordEmitter.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-1.20/src/main/java/org/apache/fluss/flink/DummyClass120.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-1.19/src/main/java/org/apache/fluss/flink/adapter/CatalogTableAdapter.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-1.18/src/main/java/org/apache/fluss/flink/adapter/SingleThreadFetcherManagerAdapter.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-1.18/src/main/java/org/apache/fluss/flink/adapter/CatalogTableAdapter.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-1.18/src/main/java/org/apache/fluss/flink/adapter/SingleThreadMultiplexSourceReaderBaseAdapter.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-1.18/src/main/java/org/apache/flink/api/connector/sink2/WriterInitContext.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-2.1/src/main/java/org/apache/flink/api/connector/sink2/Sink.java' +'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-tiering/src/main/java/org/apache/fluss/flink/tiering/FlussLakeTieringEntrypoint.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/PaimonLakeCatalog.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/PaimonLakeStorage.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/source/PaimonLakeSource.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/source/PaimonSplit.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/source/PaimonSplitSerializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/source/PaimonSortedRecordReader.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/source/PaimonRecordReader.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/source/FlussRowAsPaimonRow.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/source/PaimonSplitPlanner.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonWriteResult.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonCatalogProvider.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonCommittable.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/append/AppendOnlyWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonLakeWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonCommittableSerializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/mergetree/MergeTreeWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonWriteResultSerializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/RecordWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonLakeCommitter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonLakeTieringFactory.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/FlussRecordAsPaimonRow.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/utils/PaimonRowAsFlussRow.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/utils/FlussToPaimonPredicateConverter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/utils/PaimonConversions.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/PaimonLakeStoragePlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/FlussDataTypeToPaimonDataType.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/IcebergLakeCatalog.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/IcebergLakeStoragePlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/IcebergLakeStorage.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/source/IcebergSplitSerializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/source/IcebergSplit.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/source/IcebergRecordAsFlussRow.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/source/IcebergLakeSource.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/source/IcebergSplitPlanner.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/source/IcebergRecordReader.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/IcebergWriteResult.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/IcebergCommittable.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/FlussRecordAsIcebergRecord.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/IcebergCommittableSerializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/IcebergWriteResultSerializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/IcebergCatalogProvider.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/IcebergLakeCommitter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/IcebergLakeTieringFactory.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/writer/GenericRecordAppendOnlyWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/writer/DeltaTaskWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/writer/TaskWriterFactory.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/writer/AppendOnlyTaskWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/writer/GenericRecordDeltaWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/RecordWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/IcebergLakeWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/utils/IcebergConversions.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/utils/IcebergCatalogUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/maintenance/IcebergRewriteDataFiles.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/maintenance/RewriteDataFileResult.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/conf/IcebergConfiguration.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/conf/HadoopUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/conf/HadoopConfSerde.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/FlussDataTypeToIcebergDataType.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/iceberg/data/IcebergGenericReader.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/LanceLakeStorage.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/LanceLakeCatalog.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowDoubleWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowFieldWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowTimestampLtzWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowBinaryWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowSmallIntWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowTimeWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowIntWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowDateWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowVarCharWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowBigIntWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowBooleanWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowDecimalWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowFloatWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowTimestampNtzWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowVarBinaryWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowTinyIntWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/tiering/LanceWriteResultSerializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/tiering/LanceWriteResult.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/tiering/LanceCommittable.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/tiering/LanceLakeTieringFactory.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/tiering/LanceCommittableSerializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/tiering/LanceLakeWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/tiering/LanceLakeCommitter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/tiering/ArrowWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/tiering/LanceArrowWriter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/utils/LanceArrowUtils.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/utils/LanceDatasetAdapter.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/LanceLakeStoragePlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/LanceConfig.java' +'/Users/mehulbatra/Documents/fluss/fluss-kafka/src/main/java/org/apache/fluss/kafka/KafkaRequestHandler.java' +'/Users/mehulbatra/Documents/fluss/fluss-kafka/src/main/java/org/apache/fluss/kafka/KafkaChannelInitializer.java' +'/Users/mehulbatra/Documents/fluss/fluss-kafka/src/main/java/org/apache/fluss/kafka/KafkaCommandDecoder.java' +'/Users/mehulbatra/Documents/fluss/fluss-kafka/src/main/java/org/apache/fluss/kafka/KafkaProtocolPlugin.java' +'/Users/mehulbatra/Documents/fluss/fluss-kafka/src/main/java/org/apache/fluss/kafka/KafkaRequest.java' +'/Users/mehulbatra/Documents/fluss/tools/ci/fluss-ci-tools/src/main/java/org/apache/fluss/tools/ci/utils/deploy/DeployParser.java' +'/Users/mehulbatra/Documents/fluss/tools/ci/fluss-ci-tools/src/main/java/org/apache/fluss/tools/ci/utils/dependency/DependencyParser.java' +'/Users/mehulbatra/Documents/fluss/tools/ci/fluss-ci-tools/src/main/java/org/apache/fluss/tools/ci/utils/shared/DependencyTree.java' +'/Users/mehulbatra/Documents/fluss/tools/ci/fluss-ci-tools/src/main/java/org/apache/fluss/tools/ci/utils/shared/ParserUtils.java' +'/Users/mehulbatra/Documents/fluss/tools/ci/fluss-ci-tools/src/main/java/org/apache/fluss/tools/ci/utils/shared/Dependency.java' +'/Users/mehulbatra/Documents/fluss/tools/ci/fluss-ci-tools/src/main/java/org/apache/fluss/tools/ci/utils/notice/NoticeContents.java' +'/Users/mehulbatra/Documents/fluss/tools/ci/fluss-ci-tools/src/main/java/org/apache/fluss/tools/ci/utils/notice/NoticeParser.java' +'/Users/mehulbatra/Documents/fluss/tools/ci/fluss-ci-tools/src/main/java/org/apache/fluss/tools/ci/utils/shade/ShadeParser.java' +'/Users/mehulbatra/Documents/fluss/tools/ci/fluss-ci-tools/src/main/java/org/apache/fluss/tools/ci/licensecheck/LicenseChecker.java' +'/Users/mehulbatra/Documents/fluss/tools/ci/fluss-ci-tools/src/main/java/org/apache/fluss/tools/ci/licensecheck/JarFileChecker.java' +'/Users/mehulbatra/Documents/fluss/tools/ci/fluss-ci-tools/src/main/java/org/apache/fluss/tools/ci/licensecheck/NoticeFileChecker.java' diff --git a/website/static/javadoc/0.8-SNAPSHOT/constant-values.html b/website/static/javadoc/0.8-SNAPSHOT/constant-values.html new file mode 100644 index 0000000000..7b69325fe7 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/constant-values.html @@ -0,0 +1,572 @@ + + + + + +Constant Field Values (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Constant Field Values

+
+

Contents

+ +
+
+
+ + +
+

org.apache.*

+ + +
    +
  • + + + + + + + + + + + + + + +
    org.apache.fluss.client.table.scanner.log.LogScanner 
    Modifier and TypeConstant FieldValue
    + +public static final longEARLIEST_OFFSET-2L
    +
  • +
+ + + + + + + + + +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/deprecated-list.html b/website/static/javadoc/0.8-SNAPSHOT/deprecated-list.html new file mode 100644 index 0000000000..3876b76067 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/deprecated-list.html @@ -0,0 +1,175 @@ + + + + + +Deprecated List (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Deprecated API

+

Contents

+ +
+
+ + + +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/element-list b/website/static/javadoc/0.8-SNAPSHOT/element-list new file mode 100644 index 0000000000..f5b65b0576 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/element-list @@ -0,0 +1,47 @@ +org.apache.fluss.client +org.apache.fluss.client.admin +org.apache.fluss.client.lookup +org.apache.fluss.client.metadata +org.apache.fluss.client.metrics +org.apache.fluss.client.table +org.apache.fluss.client.table.getter +org.apache.fluss.client.table.scanner +org.apache.fluss.client.table.scanner.batch +org.apache.fluss.client.table.scanner.log +org.apache.fluss.client.table.writer +org.apache.fluss.client.token +org.apache.fluss.client.utils +org.apache.fluss.client.write +org.apache.fluss.flink +org.apache.fluss.flink.adapter +org.apache.fluss.flink.catalog +org.apache.fluss.flink.lake +org.apache.fluss.flink.lake.reader +org.apache.fluss.flink.lake.split +org.apache.fluss.flink.lake.state +org.apache.fluss.flink.metrics +org.apache.fluss.flink.procedure +org.apache.fluss.flink.row +org.apache.fluss.flink.sink +org.apache.fluss.flink.sink.serializer +org.apache.fluss.flink.sink.writer +org.apache.fluss.flink.source +org.apache.fluss.flink.source.deserializer +org.apache.fluss.flink.source.emitter +org.apache.fluss.flink.source.enumerator +org.apache.fluss.flink.source.enumerator.initializer +org.apache.fluss.flink.source.event +org.apache.fluss.flink.source.lookup +org.apache.fluss.flink.source.metrics +org.apache.fluss.flink.source.reader +org.apache.fluss.flink.source.reader.fetcher +org.apache.fluss.flink.source.split +org.apache.fluss.flink.source.state +org.apache.fluss.flink.tiering +org.apache.fluss.flink.tiering.committer +org.apache.fluss.flink.tiering.event +org.apache.fluss.flink.tiering.source +org.apache.fluss.flink.tiering.source.enumerator +org.apache.fluss.flink.tiering.source.split +org.apache.fluss.flink.tiering.source.state +org.apache.fluss.flink.utils diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-1.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-1.html new file mode 100644 index 0000000000..272a9558bd --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-1.html @@ -0,0 +1,369 @@ + + + + + +A-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages + + +

A

+
+
abort(Exception) - Method in class org.apache.fluss.client.write.WriteBatch
+
+
Abort the batch and complete the future and callbacks.
+
+
abortBatches(Exception) - Method in class org.apache.fluss.client.write.RecordAccumulator
+
+
Abort all incomplete batches (whether they have been sent or not).
+
+
abortIfBatchFull() - Method in class org.apache.fluss.client.write.RoundRobinBucketAssigner
+
 
+
abortIfBatchFull() - Method in class org.apache.fluss.client.write.StickyBucketAssigner
+
 
+
abortRecordAppends() - Method in class org.apache.fluss.client.write.ArrowLogWriteBatch
+
 
+
abortRecordAppends() - Method in class org.apache.fluss.client.write.IndexedLogWriteBatch
+
 
+
abortRecordAppends() - Method in class org.apache.fluss.client.write.KvWriteBatch
+
 
+
abortRecordAppends() - Method in class org.apache.fluss.client.write.WriteBatch
+
 
+
abortRecordForNewBatch - Variable in class org.apache.fluss.client.write.RecordAccumulator.RecordAppendResult
+
+
Whether this record was abort because the new batch created in record accumulator.
+
+
AbstractAclProcedure - Class in org.apache.fluss.flink.procedure
+
+
Procedure to operate acl.
+
+
AbstractAclProcedure() - Constructor for class org.apache.fluss.flink.procedure.AbstractAclProcedure
+
 
+
AbstractLookupBatch<T> - Class in org.apache.fluss.client.lookup
+
+
An abstract lookup batch.
+
+
AbstractLookupBatch(TableBucket) - Constructor for class org.apache.fluss.client.lookup.AbstractLookupBatch
+
 
+
AbstractLookupQuery<T> - Class in org.apache.fluss.client.lookup
+
+
Abstract Class to represent a lookup operation.
+
+
AbstractLookupQuery(TableBucket, byte[]) - Constructor for class org.apache.fluss.client.lookup.AbstractLookupQuery
+
 
+
AbstractTableWriter - Class in org.apache.fluss.client.table.writer
+
+
A base class for AppendWriter and UpsertWriter to write data to table.
+
+
AbstractTableWriter(TablePath, TableInfo, WriterClient) - Constructor for class org.apache.fluss.client.table.writer.AbstractTableWriter
+
 
+
aclOperation(Resource, PermissionType, FlussPrincipal, OperationType, String) - Method in class org.apache.fluss.flink.procedure.AbstractAclProcedure
+
 
+
aclOperation(Resource, PermissionType, FlussPrincipal, OperationType, String) - Method in class org.apache.fluss.flink.procedure.AddAclProcedure
+
 
+
aclOperation(Resource, PermissionType, FlussPrincipal, OperationType, String) - Method in class org.apache.fluss.flink.procedure.DropAclProcedure
+
 
+
aclOperation(Resource, PermissionType, FlussPrincipal, OperationType, String) - Method in class org.apache.fluss.flink.procedure.ListAclProcedure
+
 
+
AddAclProcedure - Class in org.apache.fluss.flink.procedure
+
+
Procedure to add acl.
+
+
AddAclProcedure() - Constructor for class org.apache.fluss.flink.procedure.AddAclProcedure
+
 
+
addLookup(AbstractLookupQuery<T>) - Method in class org.apache.fluss.client.lookup.AbstractLookupBatch
+
 
+
addLookup(LookupQuery) - Method in class org.apache.fluss.client.lookup.LookupBatch
+
 
+
addLookup(PrefixLookupQuery) - Method in class org.apache.fluss.client.lookup.PrefixLookupBatch
+
 
+
addReader(int) - Method in class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
+
 
+
addReader(int) - Method in class org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator
+
 
+
addSplit(SourceSplitBase, Queue<SourceSplitBase>) - Method in class org.apache.fluss.flink.lake.LakeSplitReaderGenerator
+
 
+
addSplitsBack(List<SourceSplitBase>, int) - Method in class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
+
 
+
addSplitsBack(List<TieringSplit>, int) - Method in class org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator
+
 
+
admin - Variable in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
admin - Variable in class org.apache.fluss.flink.procedure.ProcedureBase
+
 
+
Admin - Interface in org.apache.fluss.client.admin
+
+
The administrative client for Fluss, which supports managing and inspecting tables, servers, + configurations and ACLs.
+
+
all() - Method in class org.apache.fluss.client.admin.CreateAclsResult
+
 
+
all() - Method in class org.apache.fluss.client.admin.DropAclsResult
+
+
Return a future which succeeds only if all the ACLs deletions succeed, and which contains all + the deleted ACLs.
+
+
all() - Method in class org.apache.fluss.client.admin.ListOffsetsResult
+
 
+
alterDatabase(String, CatalogDatabase, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
alterFunction(ObjectPath, CatalogFunction, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
alterPartition(ObjectPath, CatalogPartitionSpec, CatalogPartition, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
alterPartitionColumnStatistics(ObjectPath, CatalogPartitionSpec, CatalogColumnStatistics, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
alterPartitionStatistics(ObjectPath, CatalogPartitionSpec, CatalogTableStatistics, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
alterTable(ObjectPath, CatalogBaseTable, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
alterTableColumnStatistics(ObjectPath, CatalogColumnStatistics, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
alterTableStatistics(ObjectPath, CatalogTableStatistics, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
append(WriteRecord, WriteCallback, Cluster, int, boolean) - Method in class org.apache.fluss.client.write.RecordAccumulator
+
+
Add a record to the accumulator, return to append result.
+
+
append(InternalRow) - Method in interface org.apache.fluss.client.table.writer.AppendWriter
+
+
Append row into a Log Table.
+
+
Append - Interface in org.apache.fluss.client.table.writer
+
+
Used to configure and create a AppendWriter to write data to a Log Table.
+
+
APPEND - org.apache.fluss.flink.row.OperationType
+
+
Represents an append-only (insert) operation.
+
+
AppendResult - Class in org.apache.fluss.client.table.writer
+
+
The result of appending a record (AppendWriter.append(InternalRow)).
+
+
AppendResult() - Constructor for class org.apache.fluss.client.table.writer.AppendResult
+
 
+
AppendSinkWriter<InputT> - Class in org.apache.fluss.flink.sink.writer
+
+
An append only sink writer for fluss log table.
+
+
AppendSinkWriter(TablePath, Configuration, RowType, MailboxExecutor, FlussSerializationSchema<InputT>) - Constructor for class org.apache.fluss.flink.sink.writer.AppendSinkWriter
+
 
+
AppendWriter - Interface in org.apache.fluss.client.table.writer
+
+
The writer to write data to the log table.
+
+
applyAggregates(List<int[]>, List<AggregateExpression>, DataType) - Method in class org.apache.fluss.flink.source.FlinkTableSource
+
 
+
applyDeleteFilters(List<ResolvedExpression>) - Method in class org.apache.fluss.flink.sink.FlinkTableSink
+
 
+
applyFilters(List<ResolvedExpression>) - Method in class org.apache.fluss.flink.source.FlinkTableSource
+
 
+
applyLimit(long) - Method in class org.apache.fluss.flink.source.FlinkTableSource
+
 
+
applyProjection(int[][], DataType) - Method in class org.apache.fluss.flink.source.FlinkTableSource
+
 
+
applyRowLevelDelete(RowLevelModificationScanContext) - Method in class org.apache.fluss.flink.sink.FlinkTableSink
+
 
+
applyRowLevelModificationScan(SupportsRowLevelModificationScan.RowLevelModificationType, RowLevelModificationScanContext) - Method in class org.apache.fluss.flink.source.FlinkTableSource
+
 
+
applyRowLevelUpdate(List<Column>, RowLevelModificationScanContext) - Method in class org.apache.fluss.flink.sink.FlinkTableSink
+
 
+
applyStaticPartition(Map<String, String>) - Method in class org.apache.fluss.flink.sink.FlinkTableSink
+
 
+
ARROW_LOG - org.apache.fluss.client.write.WriteFormat
+
 
+
ArrowLogWriteBatch - Class in org.apache.fluss.client.write
+
+
A batch of log records managed in ARROW format that is or will be sent to server by ProduceLogRequest.
+
+
ArrowLogWriteBatch(int, PhysicalTablePath, int, ArrowWriter, AbstractPagedOutputView, long) - Constructor for class org.apache.fluss.client.write.ArrowLogWriteBatch
+
 
+
asHybridSnapshotLogSplit() - Method in class org.apache.fluss.flink.source.split.SourceSplitBase
+
+
Casts this split into a HybridSnapshotLogSplit.
+
+
asHybridSnapshotLogSplitState() - Method in class org.apache.fluss.flink.source.split.SourceSplitState
+
+
Casts this split state into a HybridSnapshotLogSplitState.
+
+
asLogSplit() - Method in class org.apache.fluss.flink.source.split.SourceSplitBase
+
+
Casts this split into a LogSplit.
+
+
asLogSplitState() - Method in class org.apache.fluss.flink.source.split.SourceSplitState
+
+
Casts this split state into a LogSplitState.
+
+
assignBucket(byte[]) - Method in class org.apache.fluss.client.write.HashBucketAssigner
+
 
+
assignBucket(Cluster) - Method in class org.apache.fluss.client.write.RoundRobinBucketAssigner
+
 
+
assignBucket(Cluster) - Method in class org.apache.fluss.client.write.StickyBucketAssigner
+
 
+
asSummaryString() - Method in class org.apache.fluss.flink.sink.FlinkTableSink
+
 
+
asSummaryString() - Method in class org.apache.fluss.flink.source.FlinkTableSource
+
 
+
asTieringLogSplit() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
+
+
Casts this split into a TieringLogSplit.
+
+
asTieringSnapshotSplit() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
+
+
Casts this split into a TieringSnapshotSplit.
+
+
asyncLookup(RowData) - Method in class org.apache.fluss.flink.source.lookup.FlinkAsyncLookupFunction
+
+
The invoke entry point of lookup function.
+
+
await() - Method in class org.apache.fluss.client.write.WriteBatch.RequestFuture
+
+
Await the completion of this request.
+
+
awaitFlushCompletion() - Method in class org.apache.fluss.client.write.RecordAccumulator
+
+
Mark all buckets as ready to send and block until to send is complete.
+
+
awaitNotEmpty(long) - Method in class org.apache.fluss.client.table.scanner.log.LogFetcher
+
 
+
+A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-10.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-10.html new file mode 100644 index 0000000000..3ef1223edc --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-10.html @@ -0,0 +1,157 @@ + + + + + +J-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages + + +

J

+
+
JsonStringDeserializationSchema - Class in org.apache.fluss.flink.source.deserializer
+
+
A deserialization schema that converts LogRecord objects to JSON strings.
+
+
JsonStringDeserializationSchema() - Constructor for class org.apache.fluss.flink.source.deserializer.JsonStringDeserializationSchema
+
 
+
+A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-11.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-11.html new file mode 100644 index 0000000000..9e79166a86 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-11.html @@ -0,0 +1,187 @@ + + + + + +K-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages + + +

K

+
+
key() - Method in class org.apache.fluss.client.lookup.AbstractLookupQuery
+
 
+
keyRow() - Method in class org.apache.fluss.flink.lake.reader.KeyValueRow
+
 
+
KeyValueRow - Class in org.apache.fluss.flink.lake.reader
+
+
An InternalRow with the key part.
+
+
KeyValueRow(int[], InternalRow, boolean) - Constructor for class org.apache.fluss.flink.lake.reader.KeyValueRow
+
 
+
KV - org.apache.fluss.client.write.WriteFormat
+
 
+
KvSnapshotBatchScanner - Class in org.apache.fluss.client.table.scanner.batch
+
+
A scanner to scan the snapshot data of a kv bucket.
+
+
KvSnapshotBatchScanner(RowType, TableBucket, List<FsPathAndFileName>, int[], String, KvFormat, RemoteFileDownloader) - Constructor for class org.apache.fluss.client.table.scanner.batch.KvSnapshotBatchScanner
+
 
+
KvSnapshotMetadata - Class in org.apache.fluss.client.metadata
+
+
A class to represent the snapshot info of a kv tablet (bucket).
+
+
KvSnapshotMetadata(List<FsPathAndFileName>, long) - Constructor for class org.apache.fluss.client.metadata.KvSnapshotMetadata
+
 
+
KvSnapshots - Class in org.apache.fluss.client.metadata
+
+
A class representing the kv snapshots of a table or a partition.
+
+
KvSnapshots(long, Long, Map<Integer, Long>, Map<Integer, Long>) - Constructor for class org.apache.fluss.client.metadata.KvSnapshots
+
 
+
KvWriteBatch - Class in org.apache.fluss.client.write
+
+
A batch of kv records that is or will be sent to server by PutKvRequest.
+
+
KvWriteBatch(int, PhysicalTablePath, int, KvFormat, int, AbstractPagedOutputView, int[], long) - Constructor for class org.apache.fluss.client.write.KvWriteBatch
+
 
+
+A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-12.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-12.html new file mode 100644 index 0000000000..585244194a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-12.html @@ -0,0 +1,524 @@ + + + + + +L-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages + + +

L

+
+
LAKE_SNAPSHOT_FLUSS_LOG_SPLIT_KIND - Static variable in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
+
 
+
LAKE_SNAPSHOT_SPLIT_KIND - Static variable in class org.apache.fluss.flink.lake.split.LakeSnapshotSplit
+
 
+
LAKE_TABLE_SPLITTER - Static variable in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
LakeCatalog - Class in org.apache.fluss.flink.lake
+
+
A lake catalog to delegate the operations on lake table.
+
+
LakeCatalog(String, Map<String, String>, ClassLoader) - Constructor for class org.apache.fluss.flink.lake.LakeCatalog
+
 
+
LakeRecordRecordEmitter<OUT> - Class in org.apache.fluss.flink.lake
+
+
The emitter to emit record from lake split.
+
+
LakeRecordRecordEmitter(BiConsumer<ScanRecord, SourceOutput<OUT>>) - Constructor for class org.apache.fluss.flink.lake.LakeRecordRecordEmitter
+
 
+
LakeSnapshot - Class in org.apache.fluss.client.metadata
+
+
A class representing the lake snapshot information of a table.
+
+
LakeSnapshot(long, Map<TableBucket, Long>, Map<Long, String>) - Constructor for class org.apache.fluss.client.metadata.LakeSnapshot
+
 
+
LakeSnapshotAndFlussLogSplit - Class in org.apache.fluss.flink.lake.split
+
+
A split mixing Lake snapshot and Fluss log.
+
+
LakeSnapshotAndFlussLogSplit(TableBucket, String, List<LakeSplit>, long, long) - Constructor for class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
+
 
+
LakeSnapshotAndFlussLogSplit(TableBucket, String, List<LakeSplit>, long, long, long, int, boolean) - Constructor for class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
+
 
+
LakeSnapshotAndFlussLogSplitState - Class in org.apache.fluss.flink.lake.state
+
+ +
+
LakeSnapshotAndFlussLogSplitState(LakeSnapshotAndFlussLogSplit) - Constructor for class org.apache.fluss.flink.lake.state.LakeSnapshotAndFlussLogSplitState
+
 
+
LakeSnapshotAndLogSplitScanner - Class in org.apache.fluss.flink.lake.reader
+
+
A scanner to merge the lakehouse's snapshot and change log.
+
+
LakeSnapshotAndLogSplitScanner(Table, LakeSource<LakeSplit>, LakeSnapshotAndFlussLogSplit, int[]) - Constructor for class org.apache.fluss.flink.lake.reader.LakeSnapshotAndLogSplitScanner
+
 
+
LakeSnapshotScanner - Class in org.apache.fluss.flink.lake.reader
+
+
A scanner for reading lake split LakeSnapshotSplit.
+
+
LakeSnapshotScanner(LakeSource<LakeSplit>, LakeSnapshotSplit) - Constructor for class org.apache.fluss.flink.lake.reader.LakeSnapshotScanner
+
 
+
LakeSnapshotSplit - Class in org.apache.fluss.flink.lake.split
+
+
A split for reading a snapshot of lake.
+
+
LakeSnapshotSplit(TableBucket, String, LakeSplit, int) - Constructor for class org.apache.fluss.flink.lake.split.LakeSnapshotSplit
+
 
+
LakeSnapshotSplit(TableBucket, String, LakeSplit, int, long) - Constructor for class org.apache.fluss.flink.lake.split.LakeSnapshotSplit
+
 
+
LakeSnapshotSplitState - Class in org.apache.fluss.flink.lake.state
+
+
The state of LakeSnapshotSplit.
+
+
LakeSnapshotSplitState(LakeSnapshotSplit) - Constructor for class org.apache.fluss.flink.lake.state.LakeSnapshotSplitState
+
 
+
LakeSourceUtils - Class in org.apache.fluss.flink.utils
+
+
Utils for create lake source.
+
+
LakeSourceUtils() - Constructor for class org.apache.fluss.flink.utils.LakeSourceUtils
+
 
+
LakeSplitGenerator - Class in org.apache.fluss.flink.lake
+
+
A generator for lake splits.
+
+
LakeSplitGenerator(TableInfo, Admin, LakeSource<LakeSplit>, OffsetsInitializer.BucketOffsetsRetriever, OffsetsInitializer, int, Supplier<Set<PartitionInfo>>) - Constructor for class org.apache.fluss.flink.lake.LakeSplitGenerator
+
 
+
LakeSplitReaderGenerator - Class in org.apache.fluss.flink.lake
+
+
A generator to generate reader for lake split.
+
+
LakeSplitReaderGenerator(Table, int[], LakeSource<LakeSplit>) - Constructor for class org.apache.fluss.flink.lake.LakeSplitReaderGenerator
+
 
+
LakeSplitSerializer - Class in org.apache.fluss.flink.lake
+
+
A serializer for lake split.
+
+
LakeSplitSerializer(SimpleVersionedSerializer<LakeSplit>) - Constructor for class org.apache.fluss.flink.lake.LakeSplitSerializer
+
 
+
LakeSplitStateInitializer - Class in org.apache.fluss.flink.lake
+
+
The state initializer for lake split.
+
+
LakeSplitStateInitializer() - Constructor for class org.apache.fluss.flink.lake.LakeSplitStateInitializer
+
 
+
LakeTableFactory - Class in org.apache.fluss.flink.lake
+
+
A factory to create DynamicTableSource for lake table.
+
+
LakeTableFactory() - Constructor for class org.apache.fluss.flink.lake.LakeTableFactory
+
 
+
LakeTieringJobBuilder - Class in org.apache.fluss.flink.tiering
+
+
The builder to build Flink lake tiering job.
+
+
latest() - Static method in interface org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer
+
+
Get an OffsetsInitializer which initializes the offsets to the latest offsets of each + bucket.
+
+
LATEST - org.apache.fluss.flink.FlinkConnectorOptions.ScanStartupMode
+
 
+
latestOffsets(String, Collection<Integer>) - Method in class org.apache.fluss.flink.source.enumerator.initializer.BucketOffsetsRetrieverImpl
+
 
+
latestOffsets(String, Collection<Integer>) - Method in interface org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer.BucketOffsetsRetriever
+
 
+
LatestOffsetsInitializer - Class in org.apache.fluss.flink.source.enumerator.initializer
+
+
An implementation of OffsetsInitializer to initialize the offsets based on a + latest-offset.
+
+
LatestOffsetsInitializer() - Constructor for class org.apache.fluss.flink.source.enumerator.initializer.LatestOffsetsInitializer
+
 
+
LatestSpec() - Constructor for class org.apache.fluss.client.admin.OffsetSpec.LatestSpec
+
 
+
leaderFor(TableBucket) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
+
 
+
limit(int) - Method in interface org.apache.fluss.client.table.scanner.Scan
+
+
Returns a new scan from this that will read the given limited row number.
+
+
limit(int) - Method in class org.apache.fluss.client.table.scanner.TableScan
+
 
+
LimitBatchScanner - Class in org.apache.fluss.client.table.scanner.batch
+
+
A BatchScanner implementation that scans a limited number of records from a table.
+
+
LimitBatchScanner(TableInfo, TableBucket, MetadataUpdater, int[], int) - Constructor for class org.apache.fluss.client.table.scanner.batch.LimitBatchScanner
+
 
+
limitScan(TablePath, Configuration, RowType, int[], long) - Static method in class org.apache.fluss.flink.utils.PushdownUtils
+
 
+
LIST_EARLIEST_OFFSET - Static variable in class org.apache.fluss.client.admin.OffsetSpec
+
 
+
LIST_LATEST_OFFSET - Static variable in class org.apache.fluss.client.admin.OffsetSpec
+
 
+
LIST_OFFSET_FROM_TIMESTAMP - Static variable in class org.apache.fluss.client.admin.OffsetSpec
+
 
+
ListAclProcedure - Class in org.apache.fluss.flink.procedure
+
+
Procedure to list acl.
+
+
ListAclProcedure() - Constructor for class org.apache.fluss.flink.procedure.ListAclProcedure
+
 
+
listAcls(AclBindingFilter) - Method in interface org.apache.fluss.client.admin.Admin
+
+
Retrieves ACL entries filtered by principal for the specified resource.
+
+
listAcls(AclBindingFilter) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
listDatabases() - Method in interface org.apache.fluss.client.admin.Admin
+
+
List all databases in fluss cluster asynchronously.
+
+
listDatabases() - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
listDatabases() - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
listFunctions(String) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
listOffsets(TablePath, String, Collection<Integer>, OffsetSpec) - Method in interface org.apache.fluss.client.admin.Admin
+
+
List offset for the specified buckets.
+
+
listOffsets(TablePath, String, Collection<Integer>, OffsetSpec) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
listOffsets(TablePath, Collection<Integer>, OffsetSpec) - Method in interface org.apache.fluss.client.admin.Admin
+
+
List offset for the specified buckets.
+
+
listOffsets(TablePath, Collection<Integer>, OffsetSpec) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
ListOffsetsResult - Class in org.apache.fluss.client.admin
+
+
Result of list offsets request.
+
+
ListOffsetsResult(Map<Integer, CompletableFuture<Long>>) - Constructor for class org.apache.fluss.client.admin.ListOffsetsResult
+
 
+
listPartitionInfos(TablePath) - Method in interface org.apache.fluss.client.admin.Admin
+
+
List all partitions in the given table in fluss cluster asynchronously.
+
+
listPartitionInfos(TablePath) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
listPartitionInfos(TablePath, PartitionSpec) - Method in interface org.apache.fluss.client.admin.Admin
+
+
List all partitions in fluss cluster that are under the given table and the given partial + PartitionSpec asynchronously.
+
+
listPartitionInfos(TablePath, PartitionSpec) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
listPartitions(ObjectPath) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
listPartitions(ObjectPath, CatalogPartitionSpec) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
listPartitionsByFilter(ObjectPath, List<Expression>) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
listProcedures() - Static method in class org.apache.fluss.flink.procedure.ProcedureManager
+
 
+
listProcedures(String) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
listTables(String) - Method in interface org.apache.fluss.client.admin.Admin
+
+
List all tables in the given database in fluss cluster asynchronously.
+
+
listTables(String) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
listTables(String) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
listViews(String) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
LOG - Static variable in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
+
 
+
LOG - Static variable in class org.apache.fluss.flink.source.FlinkTableSource
+
 
+
LOG - Static variable in class org.apache.fluss.flink.utils.LakeSourceUtils
+
 
+
LOG_SPLIT_FLAG - Static variable in class org.apache.fluss.flink.source.split.SourceSplitBase
+
 
+
logEndOffset() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResult
+
 
+
LogFetchBuffer - Class in org.apache.fluss.client.table.scanner.log
+
+
LogFetchBuffer buffers up the results from the tablet server + responses as they are received.
+
+
LogFetchBuffer() - Constructor for class org.apache.fluss.client.table.scanner.log.LogFetchBuffer
+
 
+
LogFetchCollector - Class in org.apache.fluss.client.table.scanner.log
+
+
LogFetchCollector operates at the LogRecordBatch level, as that is what is stored + in the LogFetchBuffer.
+
+
LogFetchCollector(TablePath, LogScannerStatus, Configuration, MetadataUpdater) - Constructor for class org.apache.fluss.client.table.scanner.log.LogFetchCollector
+
 
+
LogFetcher - Class in org.apache.fluss.client.table.scanner.log
+
+
fetcher to fetch log.
+
+
LogFetcher(TableInfo, Projection, LogScannerStatus, Configuration, MetadataUpdater, ScannerMetricGroup, RemoteFileDownloader) - Constructor for class org.apache.fluss.client.table.scanner.log.LogFetcher
+
 
+
logOffset() - Method in class org.apache.fluss.client.table.scanner.ScanRecord
+
+
The position of this record in the corresponding fluss table bucket.
+
+
LogScan - Class in org.apache.fluss.client.table.scanner.log
+
+
Used to describe the operation to scan log data by LogScanner to a table.
+
+
LogScan() - Constructor for class org.apache.fluss.client.table.scanner.log.LogScan
+
 
+
LogScanner - Interface in org.apache.fluss.client.table.scanner.log
+
+
The scanner is used to scan log data of specify table from Fluss.
+
+
LogScannerImpl - Class in org.apache.fluss.client.table.scanner.log
+
+
The default impl of LogScanner.
+
+
LogScannerImpl(Configuration, TableInfo, MetadataUpdater, ClientMetricGroup, RemoteFileDownloader, int[]) - Constructor for class org.apache.fluss.client.table.scanner.log.LogScannerImpl
+
 
+
LogScannerStatus - Class in org.apache.fluss.client.table.scanner.log
+
+
The status of a LogScanner.
+
+
LogScannerStatus() - Constructor for class org.apache.fluss.client.table.scanner.log.LogScannerStatus
+
 
+
LogSplit - Class in org.apache.fluss.flink.source.split
+
+
The split for log.
+
+
LogSplit(TableBucket, String, long) - Constructor for class org.apache.fluss.flink.source.split.LogSplit
+
 
+
LogSplit(TableBucket, String, long, long) - Constructor for class org.apache.fluss.flink.source.split.LogSplit
+
 
+
LogSplitState - Class in org.apache.fluss.flink.source.split
+
+
The state of LogSplit.
+
+
LogSplitState(LogSplit) - Constructor for class org.apache.fluss.flink.source.split.LogSplitState
+
 
+
lookup(RowData) - Method in class org.apache.fluss.flink.source.lookup.FlinkLookupFunction
+
+
The invoke entry point of lookup function.
+
+
lookup(TableBucket, byte[]) - Method in class org.apache.fluss.client.lookup.LookupClient
+
 
+
lookup(InternalRow) - Method in interface org.apache.fluss.client.lookup.Lookuper
+
+
Lookups certain row from the given lookup key.
+
+
Lookup - Interface in org.apache.fluss.client.lookup
+
+
Used to configure and create a Lookuper to lookup rows of a primary key table.
+
+
LOOKUP - org.apache.fluss.client.lookup.LookupType
+
 
+
LOOKUP_ASYNC - Static variable in class org.apache.fluss.flink.FlinkConnectorOptions
+
 
+
LOOKUP_THREAD_PREFIX - Static variable in class org.apache.fluss.client.lookup.LookupClient
+
 
+
LookupBatch - Class in org.apache.fluss.client.lookup
+
+
A batch that contains the lookup operations that send to same tablet bucket together.
+
+
LookupBatch(TableBucket) - Constructor for class org.apache.fluss.client.lookup.LookupBatch
+
 
+
lookupBy(String...) - Method in interface org.apache.fluss.client.lookup.Lookup
+
 
+
lookupBy(List<String>) - Method in interface org.apache.fluss.client.lookup.Lookup
+
+
Returns a new Lookup instance with the given lookup columns.
+
+
lookupBy(List<String>) - Method in class org.apache.fluss.client.lookup.TableLookup
+
 
+
LookupClient - Class in org.apache.fluss.client.lookup
+
+
A client that lookups value of keys from server.
+
+
LookupClient(Configuration, MetadataUpdater) - Constructor for class org.apache.fluss.client.lookup.LookupClient
+
 
+
Lookuper - Interface in org.apache.fluss.client.lookup
+
+
The lookup-er is used to lookup row of a primary key table by primary key or prefix key.
+
+
LookupNormalizer - Class in org.apache.fluss.flink.source.lookup
+
+
A utility class to normalize the lookup key row to match the Fluss key fields order and drop the + lookup result that doesn't match remaining conditions.
+
+
LookupNormalizer.RemainingFilter - Class in org.apache.fluss.flink.source.lookup
+
+
A filter to check if the lookup result matches the remaining conditions.
+
+
LookupQuery - Class in org.apache.fluss.client.lookup
+
+
Class to represent a Lookup operation, it contains the table bucket that the key should lookup + from, the bytes of the key, and a future for the lookup operation.
+
+
LookupResult - Class in org.apache.fluss.client.lookup
+
+ +
+
LookupResult(List<InternalRow>) - Constructor for class org.apache.fluss.client.lookup.LookupResult
+
 
+
LookupResult(InternalRow) - Constructor for class org.apache.fluss.client.lookup.LookupResult
+
 
+
lookups - Variable in class org.apache.fluss.client.lookup.AbstractLookupBatch
+
 
+
lookups() - Method in class org.apache.fluss.client.lookup.AbstractLookupBatch
+
 
+
lookups() - Method in class org.apache.fluss.client.lookup.LookupBatch
+
 
+
lookups() - Method in class org.apache.fluss.client.lookup.PrefixLookupBatch
+
 
+
lookupType() - Method in class org.apache.fluss.client.lookup.AbstractLookupQuery
+
 
+
lookupType() - Method in class org.apache.fluss.client.lookup.LookupQuery
+
 
+
lookupType() - Method in class org.apache.fluss.client.lookup.PrefixLookupQuery
+
 
+
LookupType - Enum in org.apache.fluss.client.lookup
+
+
Enum to represent the type of lookup operation.
+
+
+A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-13.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-13.html new file mode 100644 index 0000000000..8ec98c2ae9 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-13.html @@ -0,0 +1,195 @@ + + + + + +M-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages + + +

M

+
+
makeCreatePartitionRequest(TablePath, PartitionSpec, boolean) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
+
 
+
makeDropPartitionRequest(TablePath, PartitionSpec, boolean) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
+
 
+
makeListOffsetsRequest(long, Long, List<Integer>, OffsetSpec) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
+
 
+
makeLookupRequest(long, Collection<LookupBatch>) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
+
 
+
makeMetadataRequest(Set<TablePath>, Collection<PhysicalTablePath>, Collection<Long>) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
+
 
+
makePbPartitionSpec(PartitionSpec) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
+
 
+
makePrefixLookupRequest(long, Collection<PrefixLookupBatch>) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
+
 
+
makeProduceLogRequest(long, int, int, List<ReadyWriteBatch>) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
+
 
+
makePutKvRequest(long, int, int, List<ReadyWriteBatch>) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
+
 
+
markEvent() - Method in class org.apache.fluss.flink.metrics.FlinkMeter
+
 
+
markEvent(long) - Method in class org.apache.fluss.flink.metrics.FlinkMeter
+
 
+
MetadataUpdater - Class in org.apache.fluss.client.metadata
+
+
The updater to initialize and update client metadata.
+
+
MetadataUpdater(Configuration, RpcClient) - Constructor for class org.apache.fluss.client.metadata.MetadataUpdater
+
 
+
MetadataUpdater(RpcClient, Cluster) - Constructor for class org.apache.fluss.client.metadata.MetadataUpdater
+
 
+
MetadataUtils - Class in org.apache.fluss.client.utils
+
+
Utils for metadata for client.
+
+
MetadataUtils() - Constructor for class org.apache.fluss.client.utils.MetadataUtils
+
 
+
metricGroup - Variable in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
+
 
+
MutableRecordAndPos - Class in org.apache.fluss.flink.source.reader
+
+
A mutable version of the RecordAndPos.
+
+
MutableRecordAndPos() - Constructor for class org.apache.fluss.flink.source.reader.MutableRecordAndPos
+
 
+
+A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-14.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-14.html new file mode 100644 index 0000000000..804dcc3a73 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-14.html @@ -0,0 +1,219 @@ + + + + + +N-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages + + +

N

+
+
newAppend() - Method in class org.apache.fluss.client.table.FlussTable
+
 
+
newAppend() - Method in interface org.apache.fluss.client.table.Table
+
+
Creates a new Append to build a AppendWriter to append data to this table + (requires to be a Log Table).
+
+
newBatchCreated - Variable in class org.apache.fluss.client.write.RecordAccumulator.RecordAppendResult
+
 
+
newBuilder(StreamExecutionEnvironment, Configuration, Configuration, String) - Static method in class org.apache.fluss.flink.tiering.LakeTieringJobBuilder
+
 
+
newCoordinatorServerClient() - Method in class org.apache.fluss.client.metadata.MetadataUpdater
+
 
+
newLookup() - Method in class org.apache.fluss.client.table.FlussTable
+
 
+
newLookup() - Method in interface org.apache.fluss.client.table.Table
+
+
Creates a new Lookup for this table to configure and create a Lookuper to + lookup data for this table by primary key or a prefix of primary key.
+
+
newRandomTabletServerClient() - Method in class org.apache.fluss.client.metadata.MetadataUpdater
+
 
+
newScan() - Method in class org.apache.fluss.client.table.FlussTable
+
 
+
newScan() - Method in interface org.apache.fluss.client.table.Table
+
+
Creates a new Scan for this table to configure and create a scanner to scan data for + this table.
+
+
newTabletServerClientForNode(int) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
+
 
+
newUpsert() - Method in class org.apache.fluss.client.table.FlussTable
+
 
+
newUpsert() - Method in interface org.apache.fluss.client.table.Table
+
+
Creates a new Upsert to build a UpsertWriter to upsert and delete data to + this table (requires to be a Primary Key Table).
+
+
next() - Method in class org.apache.fluss.flink.lake.reader.IndexedLakeSplitRecordIterator
+
 
+
nextReadyCheckDelayMs - Variable in class org.apache.fluss.client.write.RecordAccumulator.ReadyCheckResult
+
 
+
nextRecordFromSplit() - Method in class org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds
+
 
+
nextSplit() - Method in class org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds
+
 
+
NO_DATA_AVAILABLE - Static variable in class org.apache.fluss.client.table.scanner.batch.KvSnapshotBatchScanner
+
 
+
NO_READ_RECORDS_COUNT - Static variable in class org.apache.fluss.flink.source.reader.RecordAndPos
+
 
+
NO_STOPPING_OFFSET - Static variable in class org.apache.fluss.flink.source.split.LogSplit
+
 
+
nonPhysicalColumnsCount(Map<String, String>, List<String>) - Static method in class org.apache.fluss.flink.utils.CatalogPropertiesUtils
+
 
+
normalizeLookupKey(RowData) - Method in class org.apache.fluss.flink.source.lookup.LookupNormalizer
+
+
Normalize the lookup key row to match the request key and the key fields order.
+
+
NoStoppingOffsetsInitializer - Class in org.apache.fluss.flink.source.enumerator.initializer
+
+
An implementation of OffsetsInitializer which does not initialize anything.
+
+
NoStoppingOffsetsInitializer() - Constructor for class org.apache.fluss.flink.source.enumerator.initializer.NoStoppingOffsetsInitializer
+
 
+
numberOfSplits - Variable in class org.apache.fluss.flink.tiering.source.split.TieringSplit
+
 
+
numberOfWriteResults() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResult
+
 
+
+A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-15.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-15.html new file mode 100644 index 0000000000..4069f8b68f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-15.html @@ -0,0 +1,340 @@ + + + + + +O-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages + + +

O

+
+
obtainSecurityToken() - Method in class org.apache.fluss.client.token.DefaultSecurityTokenProvider
+
 
+
obtainSecurityToken() - Method in interface org.apache.fluss.client.token.SecurityTokenProvider
+
+
Obtain security token.
+
+
obtainSecurityTokensAndGetNextRenewal(AtomicReference<ObtainedSecurityToken>) - Method in class org.apache.fluss.client.token.DefaultSecurityTokenManager
+
 
+
of(SerializableSupplier<SimpleVersionedSerializer<Committable>>) - Static method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
+
 
+
of(SerializableSupplier<SimpleVersionedSerializer<WriteResult>>) - Static method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
+
 
+
offsetsFromTimestamp(String, Collection<Integer>, long) - Method in class org.apache.fluss.flink.source.enumerator.initializer.BucketOffsetsRetrieverImpl
+
 
+
offsetsFromTimestamp(String, Collection<Integer>, long) - Method in interface org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer.BucketOffsetsRetriever
+
 
+
offsetsInitializer - Variable in class org.apache.fluss.flink.source.FlinkSource
+
 
+
OffsetsInitializer - Interface in org.apache.fluss.flink.source.enumerator.initializer
+
+
An interface for users to specify the starting offset of a SourceSplitBase.
+
+
OffsetsInitializer.BucketOffsetsRetriever - Interface in org.apache.fluss.flink.source.enumerator.initializer
+
+
An interface that provides necessary information to the OffsetsInitializer to get the + initial offsets of the fluss buckets.
+
+
OffsetSpec - Class in org.apache.fluss.client.admin
+
+
Offset spec.
+
+
OffsetSpec() - Constructor for class org.apache.fluss.client.admin.OffsetSpec
+
 
+
OffsetSpec.EarliestSpec - Class in org.apache.fluss.client.admin
+
+
Earliest offset spec.
+
+
OffsetSpec.LatestSpec - Class in org.apache.fluss.client.admin
+
+
latest offset spec.
+
+
OffsetSpec.TimestampSpec - Class in org.apache.fluss.client.admin
+
+
timestamp offset spec.
+
+
onComplete(Runnable) - Method in class org.apache.fluss.client.table.scanner.log.RemoteLogDownloadFuture
+
 
+
onCompletion(Exception) - Method in interface org.apache.fluss.client.write.WriteCallback
+
 
+
onNewBatch(Cluster, int) - Method in class org.apache.fluss.client.write.RoundRobinBucketAssigner
+
 
+
onNewBatch(Cluster, int) - Method in class org.apache.fluss.client.write.StickyBucketAssigner
+
 
+
onSplitFinished(Map<String, SourceSplitState>) - Method in class org.apache.fluss.flink.source.reader.FlinkSourceReader
+
 
+
onSplitFinished(Map<String, TieringSplitState>) - Method in class org.apache.fluss.flink.tiering.source.TieringSourceReader
+
 
+
open() - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
open() - Method in class org.apache.fluss.flink.tiering.committer.FlussTableLakeSnapshotCommitter
+
 
+
open() - Method in class org.apache.fluss.flink.tiering.committer.TieringCommitOperator
+
 
+
open(FunctionContext) - Method in class org.apache.fluss.flink.source.lookup.FlinkAsyncLookupFunction
+
 
+
open(FunctionContext) - Method in class org.apache.fluss.flink.source.lookup.FlinkLookupFunction
+
 
+
open(FlussSerializationSchema.InitializationContext) - Method in interface org.apache.fluss.flink.sink.serializer.FlussSerializationSchema
+
+
Initialization method for the schema.
+
+
open(FlussSerializationSchema.InitializationContext) - Method in class org.apache.fluss.flink.sink.serializer.RowDataSerializationSchema
+
+
Initializes the schema and its internal converter.
+
+
open(FlussDeserializationSchema.InitializationContext) - Method in interface org.apache.fluss.flink.source.deserializer.FlussDeserializationSchema
+
+
Initialization method for the schema.
+
+
open(FlussDeserializationSchema.InitializationContext) - Method in class org.apache.fluss.flink.source.deserializer.JsonStringDeserializationSchema
+
+
Initializes the JSON serialization mechanism.
+
+
open(FlussDeserializationSchema.InitializationContext) - Method in class org.apache.fluss.flink.source.deserializer.RowDataDeserializationSchema
+
+
Initializes the deserialization schema.
+
+
OperationType - Enum in org.apache.fluss.flink.row
+
+
Enumeration of row operation types used in Fluss-Flink data processing.
+
+
optionalOptions() - Method in class org.apache.fluss.flink.catalog.FlinkCatalogFactory
+
 
+
optionalOptions() - Method in class org.apache.fluss.flink.catalog.FlinkTableFactory
+
 
+
org.apache.fluss.client - package org.apache.fluss.client
+
 
+
org.apache.fluss.client.admin - package org.apache.fluss.client.admin
+
 
+
org.apache.fluss.client.lookup - package org.apache.fluss.client.lookup
+
 
+
org.apache.fluss.client.metadata - package org.apache.fluss.client.metadata
+
 
+
org.apache.fluss.client.metrics - package org.apache.fluss.client.metrics
+
 
+
org.apache.fluss.client.table - package org.apache.fluss.client.table
+
 
+
org.apache.fluss.client.table.getter - package org.apache.fluss.client.table.getter
+
 
+
org.apache.fluss.client.table.scanner - package org.apache.fluss.client.table.scanner
+
 
+
org.apache.fluss.client.table.scanner.batch - package org.apache.fluss.client.table.scanner.batch
+
 
+
org.apache.fluss.client.table.scanner.log - package org.apache.fluss.client.table.scanner.log
+
 
+
org.apache.fluss.client.table.writer - package org.apache.fluss.client.table.writer
+
 
+
org.apache.fluss.client.token - package org.apache.fluss.client.token
+
 
+
org.apache.fluss.client.utils - package org.apache.fluss.client.utils
+
 
+
org.apache.fluss.client.write - package org.apache.fluss.client.write
+
 
+
org.apache.fluss.flink - package org.apache.fluss.flink
+
 
+
org.apache.fluss.flink.adapter - package org.apache.fluss.flink.adapter
+
 
+
org.apache.fluss.flink.catalog - package org.apache.fluss.flink.catalog
+
 
+
org.apache.fluss.flink.lake - package org.apache.fluss.flink.lake
+
 
+
org.apache.fluss.flink.lake.reader - package org.apache.fluss.flink.lake.reader
+
 
+
org.apache.fluss.flink.lake.split - package org.apache.fluss.flink.lake.split
+
 
+
org.apache.fluss.flink.lake.state - package org.apache.fluss.flink.lake.state
+
 
+
org.apache.fluss.flink.metrics - package org.apache.fluss.flink.metrics
+
 
+
org.apache.fluss.flink.procedure - package org.apache.fluss.flink.procedure
+
 
+
org.apache.fluss.flink.row - package org.apache.fluss.flink.row
+
 
+
org.apache.fluss.flink.sink - package org.apache.fluss.flink.sink
+
 
+
org.apache.fluss.flink.sink.serializer - package org.apache.fluss.flink.sink.serializer
+
 
+
org.apache.fluss.flink.sink.writer - package org.apache.fluss.flink.sink.writer
+
 
+
org.apache.fluss.flink.source - package org.apache.fluss.flink.source
+
 
+
org.apache.fluss.flink.source.deserializer - package org.apache.fluss.flink.source.deserializer
+
 
+
org.apache.fluss.flink.source.emitter - package org.apache.fluss.flink.source.emitter
+
 
+
org.apache.fluss.flink.source.enumerator - package org.apache.fluss.flink.source.enumerator
+
 
+
org.apache.fluss.flink.source.enumerator.initializer - package org.apache.fluss.flink.source.enumerator.initializer
+
 
+
org.apache.fluss.flink.source.event - package org.apache.fluss.flink.source.event
+
 
+
org.apache.fluss.flink.source.lookup - package org.apache.fluss.flink.source.lookup
+
 
+
org.apache.fluss.flink.source.metrics - package org.apache.fluss.flink.source.metrics
+
 
+
org.apache.fluss.flink.source.reader - package org.apache.fluss.flink.source.reader
+
 
+
org.apache.fluss.flink.source.reader.fetcher - package org.apache.fluss.flink.source.reader.fetcher
+
 
+
org.apache.fluss.flink.source.split - package org.apache.fluss.flink.source.split
+
 
+
org.apache.fluss.flink.source.state - package org.apache.fluss.flink.source.state
+
 
+
org.apache.fluss.flink.tiering - package org.apache.fluss.flink.tiering
+
 
+
org.apache.fluss.flink.tiering.committer - package org.apache.fluss.flink.tiering.committer
+
 
+
org.apache.fluss.flink.tiering.event - package org.apache.fluss.flink.tiering.event
+
 
+
org.apache.fluss.flink.tiering.source - package org.apache.fluss.flink.tiering.source
+
 
+
org.apache.fluss.flink.tiering.source.enumerator - package org.apache.fluss.flink.tiering.source.enumerator
+
 
+
org.apache.fluss.flink.tiering.source.split - package org.apache.fluss.flink.tiering.source.split
+
 
+
org.apache.fluss.flink.tiering.source.state - package org.apache.fluss.flink.tiering.source.state
+
 
+
org.apache.fluss.flink.utils - package org.apache.fluss.flink.utils
+
 
+
+A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-16.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-16.html new file mode 100644 index 0000000000..a8757d8f7b --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-16.html @@ -0,0 +1,314 @@ + + + + + +P-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages + + +

P

+
+
parseAndValidateAddresses(List<String>) - Static method in class org.apache.fluss.client.utils.ClientUtils
+
 
+
parseTimestamp(String, String, ZoneId) - Static method in class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils
+
+
Parses timestamp String to Long.
+
+
partialUpdate(int[]) - Method in class org.apache.fluss.client.table.writer.TableUpsert
+
 
+
partialUpdate(int[]) - Method in interface org.apache.fluss.client.table.writer.Upsert
+
+
Apply partial update columns and returns a new Upsert instance.
+
+
partialUpdate(String...) - Method in class org.apache.fluss.client.table.writer.TableUpsert
+
 
+
partialUpdate(String...) - Method in interface org.apache.fluss.client.table.writer.Upsert
+
 
+
partition() - Method in class org.apache.fluss.flink.tiering.source.TieringWriterInitContext
+
 
+
partition(DataStream<T>, ChannelComputer<T>, Integer) - Static method in class org.apache.fluss.flink.sink.FlinkStreamPartitioner
+
 
+
PARTITION_GROUP - Static variable in class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
+
 
+
PartitionBucketsUnsubscribedEvent - Class in org.apache.fluss.flink.source.event
+
+
An event send from reader to enumerator to indicate the splits of the partition buckets have been + removed by the reader.
+
+
PartitionBucketsUnsubscribedEvent(Collection<TableBucket>) - Constructor for class org.apache.fluss.flink.source.event.PartitionBucketsUnsubscribedEvent
+
 
+
partitionExists(ObjectPath, CatalogPartitionSpec) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
PartitionGetter - Class in org.apache.fluss.client.table.getter
+
+
A getter to get partition name from a row.
+
+
PartitionGetter(RowType, List<String>) - Constructor for class org.apache.fluss.client.table.getter.PartitionGetter
+
 
+
partitionName - Variable in class org.apache.fluss.flink.source.split.SourceSplitBase
+
 
+
partitionName - Variable in class org.apache.fluss.flink.tiering.source.split.TieringSplit
+
 
+
partitionName() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResult
+
 
+
PartitionsRemovedEvent - Class in org.apache.fluss.flink.source.event
+
+
A source event to represent partitions is removed to send from enumerator to reader.
+
+
PartitionsRemovedEvent(Map<Long, String>) - Constructor for class org.apache.fluss.flink.source.event.PartitionsRemovedEvent
+
 
+
physicalTablePath() - Method in class org.apache.fluss.client.write.WriteBatch
+
 
+
PojoToRowConverter<T> - Class in org.apache.fluss.flink.utils
+
+
Utility class for converting Java POJOs to Fluss's InternalRow format.
+
+
PojoToRowConverter(Class<T>, RowType) - Constructor for class org.apache.fluss.flink.utils.PojoToRowConverter
+
+
Creates a new converter for the specified POJO class and row type.
+
+
poll(Duration) - Method in interface org.apache.fluss.client.table.scanner.log.LogScanner
+
+
Poll log data from tablet server.
+
+
poll(Duration) - Method in class org.apache.fluss.client.table.scanner.log.LogScannerImpl
+
 
+
poll(Duration) - Method in interface org.apache.fluss.flink.source.reader.SplitScanner
+
 
+
POLL_TIERING_TABLE_INTERVAL - Static variable in class org.apache.fluss.flink.tiering.source.TieringSourceOptions
+
 
+
pollBatch(Duration) - Method in interface org.apache.fluss.client.table.scanner.batch.BatchScanner
+
+
Poll one batch records.
+
+
pollBatch(Duration) - Method in class org.apache.fluss.client.table.scanner.batch.KvSnapshotBatchScanner
+
+
Fetch data from snapshot.
+
+
pollBatch(Duration) - Method in class org.apache.fluss.client.table.scanner.batch.LimitBatchScanner
+
 
+
pollBatch(Duration) - Method in class org.apache.fluss.flink.lake.reader.LakeSnapshotAndLogSplitScanner
+
 
+
pollBatch(Duration) - Method in class org.apache.fluss.flink.lake.reader.LakeSnapshotScanner
+
 
+
pollBatch(Duration) - Method in class org.apache.fluss.flink.lake.reader.SeekableLakeSnapshotSplitScanner
+
 
+
pooledMemorySegments() - Method in class org.apache.fluss.client.write.ArrowLogWriteBatch
+
 
+
pooledMemorySegments() - Method in class org.apache.fluss.client.write.IndexedLogWriteBatch
+
 
+
pooledMemorySegments() - Method in class org.apache.fluss.client.write.KvWriteBatch
+
 
+
pooledMemorySegments() - Method in class org.apache.fluss.client.write.WriteBatch
+
+
get pooled memory segments to de-allocate.
+
+
PREFIX_LOOKUP - org.apache.fluss.client.lookup.LookupType
+
 
+
prefixLookup(TableBucket, byte[]) - Method in class org.apache.fluss.client.lookup.LookupClient
+
 
+
PrefixLookup - Class in org.apache.fluss.client.lookup
+
+
Used to describe the operation to prefix lookup by PrefixLookuper to a primary key table.
+
+
PrefixLookup(List<String>) - Constructor for class org.apache.fluss.client.lookup.PrefixLookup
+
 
+
PrefixLookupBatch - Class in org.apache.fluss.client.lookup
+
+
A batch that contains the prefix lookup operations that send to same destination and some table + together.
+
+
PrefixLookupBatch(TableBucket) - Constructor for class org.apache.fluss.client.lookup.PrefixLookupBatch
+
 
+
PrefixLookupQuery - Class in org.apache.fluss.client.lookup
+
+
Class to represent a prefix lookup operation, it contains the table id, bucketNums and related + CompletableFuture.
+
+
ProcedureBase - Class in org.apache.fluss.flink.procedure
+
+
Base implementation for flink Procedure.
+
+
ProcedureBase() - Constructor for class org.apache.fluss.flink.procedure.ProcedureBase
+
 
+
ProcedureManager - Class in org.apache.fluss.flink.procedure
+
+
ProcedureUtil to load procedure.
+
+
ProcedureManager() - Constructor for class org.apache.fluss.flink.procedure.ProcedureManager
+
 
+
processElement(StreamRecord<TableBucketWriteResult<WriteResult>>) - Method in class org.apache.fluss.flink.tiering.committer.TieringCommitOperator
+
 
+
project(int[]) - Method in interface org.apache.fluss.client.table.scanner.Scan
+
+
Returns a new scan from this that will read the given data columns.
+
+
project(int[]) - Method in class org.apache.fluss.client.table.scanner.TableScan
+
 
+
project(List<String>) - Method in interface org.apache.fluss.client.table.scanner.Scan
+
+
Returns a new scan from this that will read the given data columns.
+
+
project(List<String>) - Method in class org.apache.fluss.client.table.scanner.TableScan
+
 
+
projectRowType(RowType, int[]) - Static method in class org.apache.fluss.flink.utils.FlinkUtils
+
+
Returns projected RowType by given projection indexes over original RowType.
+
+
PushdownUtils - Class in org.apache.fluss.flink.utils
+
+
Utilities for pushdown abilities.
+
+
PushdownUtils() - Constructor for class org.apache.fluss.flink.utils.PushdownUtils
+
 
+
PushdownUtils.FieldEqual - Class in org.apache.fluss.flink.utils
+
+
A structure represents a source field equal literal expression.
+
+
PushdownUtils.ValueConversion - Enum in org.apache.fluss.flink.utils
+
+
The value conversion type between Flink internal value and Fluss internal value.
+
+
putVariables(Map<String, String>) - Method in class org.apache.fluss.client.metrics.ScannerMetricGroup
+
 
+
+A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-17.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-17.html new file mode 100644 index 0000000000..6810314938 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-17.html @@ -0,0 +1,153 @@ + + + + + +Q-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages + + +

Q

+
+
querySingleRow(GenericRowData, TablePath, Configuration, RowType, int[], int, int[]) - Static method in class org.apache.fluss.flink.utils.PushdownUtils
+
 
+
+A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-18.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-18.html new file mode 100644 index 0000000000..12c8a5a49d --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-18.html @@ -0,0 +1,333 @@ + + + + + +R-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages + + +

R

+
+
readBatch() - Method in class org.apache.fluss.flink.source.reader.BoundedSplitReader
+
+
Read next batch of data.
+
+
READER_METRIC_GROUP - Static variable in class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
+
 
+
readRecordsCount - Variable in class org.apache.fluss.flink.source.reader.RecordAndPos
+
 
+
readRecordsCount() - Method in class org.apache.fluss.flink.source.reader.RecordAndPos
+
 
+
ready(Cluster) - Method in class org.apache.fluss.client.write.RecordAccumulator
+
+
Get a list of nodes whose buckets are ready to be sent.
+
+
ReadyCheckResult(Set<Integer>, long, Set<PhysicalTablePath>) - Constructor for class org.apache.fluss.client.write.RecordAccumulator.ReadyCheckResult
+
 
+
readyNodes - Variable in class org.apache.fluss.client.write.RecordAccumulator.ReadyCheckResult
+
 
+
ReadyWriteBatch - Class in org.apache.fluss.client.write
+
+
The WriteBatch already ready in sender.
+
+
ReadyWriteBatch(TableBucket, WriteBatch) - Constructor for class org.apache.fluss.client.write.ReadyWriteBatch
+
 
+
record() - Method in class org.apache.fluss.flink.source.reader.RecordAndPos
+
 
+
RecordAccumulator - Class in org.apache.fluss.client.write
+
+
This class act as a queue that accumulates records into WriteBatch instances to be sent + to tablet servers.
+
+
RecordAccumulator.ReadyCheckResult - Class in org.apache.fluss.client.write
+
+
The set of nodes that have at leader one complete record batch in the accumulator.
+
+
RecordAccumulator.RecordAppendResult - Class in org.apache.fluss.client.write
+
+
Metadata about a record just appended to the record accumulator.
+
+
RecordAndPos - Class in org.apache.fluss.flink.source.reader
+
+
A record wrapping a Fluss ScanRecord, the readRecordsCount when the record is + from reading snapshot, the code currentSplitIndex where the record is from when the Flink + source split read by split reader contains multiple splits(splittable unit), like LakeSnapshotAndFlussLogSplit contains multiple LakeSplits.
+
+
RecordAndPos(ScanRecord) - Constructor for class org.apache.fluss.flink.source.reader.RecordAndPos
+
 
+
RecordAndPos(ScanRecord, long) - Constructor for class org.apache.fluss.flink.source.reader.RecordAndPos
+
 
+
RecordAndPos(ScanRecord, long, int) - Constructor for class org.apache.fluss.flink.source.reader.RecordAndPos
+
 
+
RecordAppendResult(boolean, boolean, boolean) - Constructor for class org.apache.fluss.client.write.RecordAccumulator.RecordAppendResult
+
 
+
recordCount - Variable in class org.apache.fluss.client.write.WriteBatch
+
 
+
recordCurrentOffset(TableBucket, long) - Method in class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
+
+
Update current consuming offset of the given TableBucket.
+
+
recordPerBatch() - Method in class org.apache.fluss.client.metrics.WriterMetricGroup
+
 
+
recordPollEnd(long) - Method in class org.apache.fluss.client.metrics.ScannerMetricGroup
+
 
+
recordPollStart(long) - Method in class org.apache.fluss.client.metrics.ScannerMetricGroup
+
 
+
records(TableBucket) - Method in class org.apache.fluss.client.table.scanner.log.ScanRecords
+
+
Get just the records for the given bucketId.
+
+
recordsRetryTotal() - Method in class org.apache.fluss.client.metrics.WriterMetricGroup
+
 
+
recordsSendTotal() - Method in class org.apache.fluss.client.metrics.WriterMetricGroup
+
 
+
recordsToSkip - Variable in class org.apache.fluss.flink.source.split.SnapshotSplit
+
+
The records to skip when reading the snapshot.
+
+
recordsToSkip() - Method in class org.apache.fluss.flink.source.split.SnapshotSplit
+
 
+
recycle() - Method in class org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds
+
 
+
reEnqueue(ReadyWriteBatch) - Method in class org.apache.fluss.client.write.RecordAccumulator
+
 
+
register(Metric, String, AbstractMetricGroup) - Method in class org.apache.fluss.flink.metrics.FlinkMetricRegistry
+
 
+
registerTableBucket(TableBucket) - Method in class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
+
 
+
remoteFetchBytes() - Method in class org.apache.fluss.client.metrics.ScannerMetricGroup
+
 
+
remoteFetchErrorCount() - Method in class org.apache.fluss.client.metrics.ScannerMetricGroup
+
 
+
remoteFetchRequestCount() - Method in class org.apache.fluss.client.metrics.ScannerMetricGroup
+
 
+
RemoteFileDownloader - Class in org.apache.fluss.client.table.scanner
+
+
The downloader that has a IO thread pool to download the remote files (like kv snapshots files, + log segment files).
+
+
RemoteFileDownloader(int) - Constructor for class org.apache.fluss.client.table.scanner.RemoteFileDownloader
+
 
+
RemoteLogDownloader - Class in org.apache.fluss.client.table.scanner.log
+
+
Downloader to read remote log files to local disk.
+
+
RemoteLogDownloader(TablePath, Configuration, RemoteFileDownloader, ScannerMetricGroup) - Constructor for class org.apache.fluss.client.table.scanner.log.RemoteLogDownloader
+
 
+
RemoteLogDownloadFuture - Class in org.apache.fluss.client.table.scanner.log
+
+
Represents the future of a remote log download request.
+
+
RemoteLogDownloadFuture(CompletableFuture<File>, Runnable) - Constructor for class org.apache.fluss.client.table.scanner.log.RemoteLogDownloadFuture
+
 
+
removePartitions(Map<Long, String>) - Method in class org.apache.fluss.flink.source.reader.FlinkSourceSplitReader
+
 
+
removePartitions(Map<Long, String>, Consumer<Set<TableBucket>>) - Method in class org.apache.fluss.flink.source.reader.fetcher.FlinkSourceFetcherManager
+
 
+
renameTable(ObjectPath, String, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
reopened - Variable in class org.apache.fluss.client.write.WriteBatch
+
 
+
replace(RowData) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
+
 
+
reportRecordEventTime(long) - Method in class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
+
 
+
RequestFuture() - Constructor for class org.apache.fluss.client.write.WriteBatch.RequestFuture
+
 
+
requestRemoteLog(FsPath, RemoteLogSegment) - Method in class org.apache.fluss.client.table.scanner.log.RemoteLogDownloader
+
+
Request to fetch remote log segment to local.
+
+
requiredOptions() - Method in class org.apache.fluss.flink.catalog.FlinkCatalogFactory
+
 
+
requiredOptions() - Method in class org.apache.fluss.flink.catalog.FlinkTableFactory
+
 
+
resetWriterState(long, int) - Method in class org.apache.fluss.client.write.ArrowLogWriteBatch
+
 
+
resetWriterState(long, int) - Method in class org.apache.fluss.client.write.IndexedLogWriteBatch
+
 
+
resetWriterState(long, int) - Method in class org.apache.fluss.client.write.KvWriteBatch
+
 
+
resetWriterState(long, int) - Method in class org.apache.fluss.client.write.WriteBatch
+
 
+
restoreEnumerator(SplitEnumeratorContext<SourceSplitBase>, SourceEnumeratorState) - Method in class org.apache.fluss.flink.source.FlinkSource
+
 
+
restoreEnumerator(SplitEnumeratorContext<TieringSplit>, TieringSourceEnumeratorState) - Method in class org.apache.fluss.flink.tiering.source.TieringSource
+
 
+
RoundRobinBucketAssigner - Class in org.apache.fluss.client.write
+
+
The bucket assigner use round-robin strategy.
+
+
RoundRobinBucketAssigner(PhysicalTablePath) - Constructor for class org.apache.fluss.client.write.RoundRobinBucketAssigner
+
 
+
RowDataDeserializationSchema - Class in org.apache.fluss.flink.source.deserializer
+
+
A deserialization schema that converts LogRecord objects to Flink's RowData + format.
+
+
RowDataDeserializationSchema() - Constructor for class org.apache.fluss.flink.source.deserializer.RowDataDeserializationSchema
+
 
+
RowDataSerializationSchema - Class in org.apache.fluss.flink.sink.serializer
+
+
Default implementation of RowDataConverter for RowData.
+
+
RowDataSerializationSchema(boolean, boolean) - Constructor for class org.apache.fluss.flink.sink.serializer.RowDataSerializationSchema
+
+
Constructs a new RowSerializationSchema.
+
+
RowWithOp - Class in org.apache.fluss.flink.row
+
+
A wrapper class that associates an InternalRow with an OperationType for use in + Fluss-Flink data processing.
+
+
RowWithOp(InternalRow, OperationType) - Constructor for class org.apache.fluss.flink.row.RowWithOp
+
+
Constructs a RowWithOp with the specified internal row and operation type.
+
+
run() - Method in class org.apache.fluss.client.write.Sender
+
 
+
runOnce() - Method in class org.apache.fluss.client.write.Sender
+
+
Run a single iteration of sending.
+
+
+A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-19.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-19.html new file mode 100644 index 0000000000..1eb01fce2f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-19.html @@ -0,0 +1,536 @@ + + + + + +S-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages + + +

S

+
+
Scan - Interface in org.apache.fluss.client.table.scanner
+
+
Used to configure and create a scanner to scan data for a table.
+
+
SCAN_PARTITION_DISCOVERY_INTERVAL - Static variable in class org.apache.fluss.flink.FlinkConnectorOptions
+
 
+
SCAN_STARTUP_MODE - Static variable in class org.apache.fluss.flink.FlinkConnectorOptions
+
 
+
SCAN_STARTUP_TIMESTAMP - Static variable in class org.apache.fluss.flink.FlinkConnectorOptions
+
 
+
ScannerMetricGroup - Class in org.apache.fluss.client.metrics
+
+
The metric group for scanner, including LogScanner and BatchScanner.
+
+
ScannerMetricGroup(ClientMetricGroup, TablePath) - Constructor for class org.apache.fluss.client.metrics.ScannerMetricGroup
+
 
+
scanPartitionDiscoveryIntervalMs - Variable in class org.apache.fluss.flink.source.FlinkSource
+
 
+
scanRecord - Variable in class org.apache.fluss.flink.source.reader.RecordAndPos
+
 
+
ScanRecord - Class in org.apache.fluss.client.table.scanner
+
+
one scan record.
+
+
ScanRecord(long, long, ChangeType, InternalRow) - Constructor for class org.apache.fluss.client.table.scanner.ScanRecord
+
 
+
ScanRecord(InternalRow) - Constructor for class org.apache.fluss.client.table.scanner.ScanRecord
+
 
+
ScanRecords - Class in org.apache.fluss.client.table.scanner.log
+
+
A container that holds the list ScanRecord per bucket for a particular table.
+
+
ScanRecords(Map<TableBucket, List<ScanRecord>>) - Constructor for class org.apache.fluss.client.table.scanner.log.ScanRecords
+
 
+
SecurityTokenManager - Interface in org.apache.fluss.client.token
+
+
API for security token manager.
+
+
SecurityTokenProvider - Interface in org.apache.fluss.client.token
+
+
Security token provider API.
+
+
SeekableLakeSnapshotSplitScanner - Class in org.apache.fluss.flink.lake.reader
+
+
A scanner that supports seeking to a specific LakeSplit and reading from that point.
+
+
SeekableLakeSnapshotSplitScanner(LakeSource<LakeSplit>, List<LakeSplit>, int) - Constructor for class org.apache.fluss.flink.lake.reader.SeekableLakeSnapshotSplitScanner
+
 
+
select(int, int) - Static method in interface org.apache.fluss.flink.sink.ChannelComputer
+
 
+
select(String, int, int) - Static method in interface org.apache.fluss.flink.sink.ChannelComputer
+
 
+
selectChannel(SerializationDelegate<StreamRecord<T>>) - Method in class org.apache.fluss.flink.sink.FlinkStreamPartitioner
+
 
+
selectRowCount - Variable in class org.apache.fluss.flink.source.FlinkTableSource
+
 
+
send(WriteRecord) - Method in class org.apache.fluss.client.table.writer.AbstractTableWriter
+
 
+
send(WriteRecord, WriteCallback) - Method in class org.apache.fluss.client.write.WriterClient
+
+
Asynchronously send a record to a table and invoke the provided callback when to send has + been acknowledged.
+
+
Sender - Class in org.apache.fluss.client.write
+
+
This background thread handles the sending of produce requests to the tablet server.
+
+
Sender(RecordAccumulator, int, int, short, int, MetadataUpdater, IdempotenceManager, WriterMetricGroup) - Constructor for class org.apache.fluss.client.write.Sender
+
 
+
SENDER_THREAD_PREFIX - Static variable in class org.apache.fluss.client.write.WriterClient
+
 
+
sendFetches() - Method in class org.apache.fluss.client.table.scanner.log.LogFetcher
+
+
Set up a fetch request for any node that we have assigned buckets for which doesn't already + have an in-flight fetch or pending fetch data.
+
+
sendMetadataRequestAndRebuildCluster(Cluster, RpcClient, Set<TablePath>, Collection<PhysicalTablePath>, Collection<Long>) - Static method in class org.apache.fluss.client.utils.MetadataUtils
+
+
Partial update cluster, means we will rebuild the cluster by sending metadata request to + request the input tables/partitions in physicalTablePaths, after that add those + tables/partitions into cluster.
+
+
sendMetadataRequestAndRebuildCluster(AdminReadOnlyGateway, boolean, Cluster, Set<TablePath>, Collection<PhysicalTablePath>, Collection<Long>) - Static method in class org.apache.fluss.client.utils.MetadataUtils
+
+
maybe partial update cluster.
+
+
sendMetadataRequestAndRebuildCluster(AdminReadOnlyGateway, Set<TablePath>) - Static method in class org.apache.fluss.client.utils.MetadataUtils
+
+
full update cluster, means we will rebuild the cluster by clearing all cached table in + cluster, and then send metadata request to request the input tables in tablePaths, after that + add those table into cluster.
+
+
sequenceHasBeenReset() - Method in class org.apache.fluss.client.write.WriteBatch
+
 
+
serialize(Object) - Method in interface org.apache.fluss.flink.utils.FlinkRowToFlussRowConverter.FlussSerializationConverter
+
+
Deprecated.
+
Convert a Flink field object of RowData to the Fluss's internal data structure + object.
+
+
serialize(DataOutputSerializer, SourceSplitBase) - Method in class org.apache.fluss.flink.lake.LakeSplitSerializer
+
 
+
serialize(RowData) - Method in class org.apache.fluss.flink.sink.serializer.RowDataSerializationSchema
+
+
Serializes a Flink RowData into a Fluss RowWithOp containing an InternalRow and the corresponding OperationType.
+
+
serialize(SourceSplitBase) - Method in class org.apache.fluss.flink.source.split.SourceSplitSerializer
+
 
+
serialize(SourceEnumeratorState) - Method in class org.apache.fluss.flink.source.state.FlussSourceEnumeratorStateSerializer
+
 
+
serialize(TieringSplit) - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplitSerializer
+
 
+
serialize(TieringSourceEnumeratorState) - Method in class org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorStateSerializer
+
 
+
serialize(TableBucketWriteResult<WriteResult>) - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultSerializer
+
 
+
serialize(T) - Method in interface org.apache.fluss.flink.sink.serializer.FlussSerializationSchema
+
+
Serializes an object of type T into a RowWithOp.
+
+
serializeComputedColumns(List<Column>) - Static method in class org.apache.fluss.flink.utils.CatalogPropertiesUtils
+
 
+
serializeComputedColumns(Map<String, String>, List<Column>) - Static method in class org.apache.fluss.flink.utils.CatalogPropertiesUtils
+
 
+
SerializerInitContextImpl - Class in org.apache.fluss.flink.sink.serializer
+
+
Contextual information that can be used during initialization of FlussSerializationSchema.
+
+
SerializerInitContextImpl(RowType) - Constructor for class org.apache.fluss.flink.sink.serializer.SerializerInitContextImpl
+
 
+
serializeWatermarkSpecs(Map<String, String>, List<WatermarkSpec>) - Static method in class org.apache.fluss.flink.utils.CatalogPropertiesUtils
+
 
+
setBatchQueueTimeMs(long) - Method in class org.apache.fluss.client.metrics.WriterMetricGroup
+
 
+
setBootstrapServers(String) - Method in class org.apache.fluss.flink.sink.FlussSinkBuilder
+
+
Set the bootstrap server for the sink.
+
+
setBootstrapServers(String) - Method in class org.apache.fluss.flink.source.FlussSourceBuilder
+
+
Sets the bootstrap servers for the Fluss source connection.
+
+
setCurrentLakeSplitIndex(int) - Method in class org.apache.fluss.flink.lake.state.LakeSnapshotAndFlussLogSplitState
+
 
+
setDatabase(String) - Method in class org.apache.fluss.flink.sink.FlussSinkBuilder
+
+
Set the database for the sink.
+
+
setDatabase(String) - Method in class org.apache.fluss.flink.source.FlussSourceBuilder
+
+
Sets the database name for the Fluss source.
+
+
setDeserializationSchema(FlussDeserializationSchema<OUT>) - Method in class org.apache.fluss.flink.source.FlussSourceBuilder
+
+
Sets the deserialization schema for converting Fluss records to output records.
+
+
setFlussConfig(Configuration) - Method in class org.apache.fluss.flink.source.FlussSourceBuilder
+
+
Sets custom Fluss configuration properties for the source connector.
+
+
setNextLogOffset(long) - Method in class org.apache.fluss.flink.lake.state.LakeSnapshotAndFlussLogSplitState
+
 
+
setNextOffset(long) - Method in class org.apache.fluss.flink.source.split.HybridSnapshotLogSplitState
+
 
+
setNextOffset(long) - Method in class org.apache.fluss.flink.source.split.LogSplitState
+
 
+
setOption(String, String) - Method in class org.apache.fluss.flink.sink.FlussSinkBuilder
+
+
Set a configuration option.
+
+
setOptions(Map<String, String>) - Method in class org.apache.fluss.flink.sink.FlussSinkBuilder
+
+
Set multiple configuration options.
+
+
setProjectedFields(String...) - Method in class org.apache.fluss.flink.source.FlussSourceBuilder
+
+
Sets the projected fields for this source using field names.
+
+
setRecord(ScanRecord, long) - Method in class org.apache.fluss.flink.source.reader.MutableRecordAndPos
+
 
+
setRecord(ScanRecord, long, int) - Method in class org.apache.fluss.flink.source.reader.MutableRecordAndPos
+
 
+
setRecordsToSkip(long) - Method in class org.apache.fluss.flink.lake.state.LakeSnapshotAndFlussLogSplitState
+
 
+
setRecordsToSkip(long) - Method in class org.apache.fluss.flink.lake.state.LakeSnapshotSplitState
+
 
+
setRecordsToSkip(long) - Method in class org.apache.fluss.flink.source.split.HybridSnapshotLogSplitState
+
 
+
setScanPartitionDiscoveryIntervalMs(long) - Method in class org.apache.fluss.flink.source.FlussSourceBuilder
+
+
Sets the scan partition discovery interval in milliseconds.
+
+
setSendLatencyInMs(long) - Method in class org.apache.fluss.client.metrics.WriterMetricGroup
+
 
+
setSerializationSchema(FlussSerializationSchema<InputT>) - Method in class org.apache.fluss.flink.sink.FlussSinkBuilder
+
+
Set a FlussSerializationSchema.
+
+
setShuffleByBucketId(boolean) - Method in class org.apache.fluss.flink.sink.FlussSinkBuilder
+
+
Set shuffle by bucket id.
+
+
setStartingOffsets(OffsetsInitializer) - Method in class org.apache.fluss.flink.source.FlussSourceBuilder
+
+
Sets the starting offsets strategy for the Fluss source.
+
+
setTable(String) - Method in class org.apache.fluss.flink.sink.FlussSinkBuilder
+
+
Set the table name for the sink.
+
+
setTable(String) - Method in class org.apache.fluss.flink.source.FlussSourceBuilder
+
+
Sets the table name for the Fluss source.
+
+
setTableBucketStoppingOffset(TableBucket, long) - Method in class org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds
+
 
+
setup(int) - Method in interface org.apache.fluss.flink.sink.ChannelComputer
+
 
+
setup(int) - Method in class org.apache.fluss.flink.sink.FlinkRowDataChannelComputer
+
 
+
setup(int) - Method in class org.apache.fluss.flink.sink.FlinkStreamPartitioner
+
 
+
setup(StreamTask<?, ?>, StreamConfig, Output<StreamRecord<CommittableMessage<Committable>>>) - Method in class org.apache.fluss.flink.tiering.committer.TieringCommitOperator
+
 
+
setWriterState(long, int) - Method in class org.apache.fluss.client.write.ArrowLogWriteBatch
+
 
+
setWriterState(long, int) - Method in class org.apache.fluss.client.write.IndexedLogWriteBatch
+
 
+
setWriterState(long, int) - Method in class org.apache.fluss.client.write.KvWriteBatch
+
 
+
setWriterState(long, int) - Method in class org.apache.fluss.client.write.WriteBatch
+
 
+
SingleThreadFetcherManagerAdapter<E,​SplitT extends org.apache.flink.api.connector.source.SourceSplit> - Class in org.apache.fluss.flink.adapter
+
+
Adapter for SingleThreadFetcherManager.TODO: remove it until not supported in flink 1.18.
+
+
SingleThreadFetcherManagerAdapter(FutureCompletingBlockingQueue<RecordsWithSplitIds<E>>, Supplier<SplitReader<E, SplitT>>, Configuration, Consumer<Collection<String>>) - Constructor for class org.apache.fluss.flink.adapter.SingleThreadFetcherManagerAdapter
+
 
+
SingleThreadMultiplexSourceReaderBaseAdapter<E,​T,​SplitT extends org.apache.flink.api.connector.source.SourceSplit,​SplitStateT> - Class in org.apache.fluss.flink.adapter
+
+
Adapter for SingleThreadMultiplexSourceReaderBase.TODO: remove it until not supported in + flink 1.18.
+
+
SingleThreadMultiplexSourceReaderBaseAdapter(FutureCompletingBlockingQueue<RecordsWithSplitIds<E>>, SingleThreadFetcherManager<E, SplitT>, RecordEmitter<E, T, SplitStateT>, Configuration, SourceReaderContext) - Constructor for class org.apache.fluss.flink.adapter.SingleThreadMultiplexSourceReaderBaseAdapter
+
 
+
SINK_BUCKET_SHUFFLE - Static variable in class org.apache.fluss.flink.FlinkConnectorOptions
+
 
+
SINK_IGNORE_DELETE - Static variable in class org.apache.fluss.flink.FlinkConnectorOptions
+
 
+
snapshotId - Variable in class org.apache.fluss.flink.source.split.SnapshotSplit
+
+
The snapshot id.
+
+
SnapshotOffsetsInitializer - Class in org.apache.fluss.flink.source.enumerator.initializer
+
+
An implementation of OffsetsInitializer which initializes the offsets.
+
+
SnapshotOffsetsInitializer() - Constructor for class org.apache.fluss.flink.source.enumerator.initializer.SnapshotOffsetsInitializer
+
 
+
SnapshotSplit - Class in org.apache.fluss.flink.source.split
+
+
The split for snapshot.
+
+
SnapshotSplit(TableBucket, String, long) - Constructor for class org.apache.fluss.flink.source.split.SnapshotSplit
+
 
+
SnapshotSplit(TableBucket, String, long, long) - Constructor for class org.apache.fluss.flink.source.split.SnapshotSplit
+
 
+
snapshotState(long) - Method in class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
+
 
+
snapshotState(long) - Method in class org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator
+
 
+
snapshotState(long) - Method in class org.apache.fluss.flink.tiering.source.TieringSourceReader
+
 
+
SourceEnumeratorState - Class in org.apache.fluss.flink.source.state
+
+
A checkpoint of the current state of the containing the buckets that is already assigned.
+
+
SourceEnumeratorState(Set<TableBucket>, Map<Long, String>, List<SourceSplitBase>) - Constructor for class org.apache.fluss.flink.source.state.SourceEnumeratorState
+
 
+
SourceSplitBase - Class in org.apache.fluss.flink.source.split
+
+
A base source split for SnapshotSplit and LogSplit.
+
+
SourceSplitBase(TableBucket, String) - Constructor for class org.apache.fluss.flink.source.split.SourceSplitBase
+
 
+
SourceSplitSerializer - Class in org.apache.fluss.flink.source.split
+
+
A serializer for the SourceSplitBase.
+
+
SourceSplitSerializer(LakeSource<LakeSplit>) - Constructor for class org.apache.fluss.flink.source.split.SourceSplitSerializer
+
 
+
SourceSplitState - Class in org.apache.fluss.flink.source.split
+
+
State of the reader, essentially a mutable version of the SourceSplitBase.
+
+
SourceSplitState(SourceSplitBase) - Constructor for class org.apache.fluss.flink.source.split.SourceSplitState
+
 
+
split - Variable in class org.apache.fluss.flink.source.split.SourceSplitState
+
 
+
splitId() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
+
 
+
splitId() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotSplit
+
 
+
splitId() - Method in class org.apache.fluss.flink.source.split.HybridSnapshotLogSplit
+
 
+
splitId() - Method in class org.apache.fluss.flink.source.split.LogSplit
+
 
+
splitId() - Method in class org.apache.fluss.flink.tiering.source.split.TieringLogSplit
+
 
+
splitId() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSnapshotSplit
+
 
+
splitKind() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
+
 
+
splitKind() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotSplit
+
 
+
splitKind() - Method in class org.apache.fluss.flink.source.split.LogSplit
+
 
+
splitKind() - Method in class org.apache.fluss.flink.source.split.SourceSplitBase
+
 
+
splitKind() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
+
 
+
SplitScanner - Interface in org.apache.fluss.flink.source.reader
+
+
A scanner to scan split to records .
+
+
SQL - org.apache.fluss.flink.utils.TimestampFormat
+
+
Options to specify TIMESTAMP/TIMESTAMP_WITH_LOCAL_ZONE format.
+
+
start() - Method in class org.apache.fluss.client.table.scanner.log.RemoteLogDownloader
+
 
+
start() - Method in class org.apache.fluss.client.token.DefaultSecurityTokenManager
+
 
+
start() - Method in interface org.apache.fluss.client.token.SecurityTokenManager
+
+
Creates a re-occurring task which obtains new tokens.
+
+
start() - Method in class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
+
 
+
start() - Method in class org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator
+
 
+
start() - Method in class org.apache.fluss.flink.tiering.source.TieringSourceReader
+
 
+
startupMode - Variable in class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils.StartupOptions
+
 
+
StartupOptions() - Constructor for class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils.StartupOptions
+
 
+
startupTimestampMs - Variable in class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils.StartupOptions
+
 
+
StickyBucketAssigner - Class in org.apache.fluss.client.write
+
+
The bucket assigner with sticky strategy.
+
+
StickyBucketAssigner(PhysicalTablePath) - Constructor for class org.apache.fluss.client.write.StickyBucketAssigner
+
 
+
stop() - Method in class org.apache.fluss.client.token.DefaultSecurityTokenManager
+
+
Stops re-occurring token obtain task.
+
+
stop() - Method in interface org.apache.fluss.client.token.SecurityTokenManager
+
+
Stops re-occurring token obtain task.
+
+
subscribe(int, long) - Method in interface org.apache.fluss.client.table.scanner.log.LogScanner
+
+
Subscribe to the given table bucket in given offset dynamically.
+
+
subscribe(int, long) - Method in class org.apache.fluss.client.table.scanner.log.LogScannerImpl
+
 
+
subscribe(long, int, long) - Method in interface org.apache.fluss.client.table.scanner.log.LogScanner
+
+
Subscribe to the given partitioned table bucket in given offset dynamically.
+
+
subscribe(long, int, long) - Method in class org.apache.fluss.client.table.scanner.log.LogScannerImpl
+
 
+
subscribeFromBeginning(int) - Method in interface org.apache.fluss.client.table.scanner.log.LogScanner
+
+
Subscribe to the given table buckets from beginning dynamically.
+
+
subscribeFromBeginning(long, int) - Method in interface org.apache.fluss.client.table.scanner.log.LogScanner
+
+
Subscribe to the given partitioned table bucket from beginning dynamically.
+
+
supportsNestedProjection() - Method in class org.apache.fluss.flink.source.FlinkTableSource
+
 
+
+A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-2.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-2.html new file mode 100644 index 0000000000..425fda4652 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-2.html @@ -0,0 +1,248 @@ + + + + + +B-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages + + +

B

+
+
batchIsFull - Variable in class org.apache.fluss.client.write.RecordAccumulator.RecordAppendResult
+
 
+
BatchScanner - Interface in org.apache.fluss.client.table.scanner.batch
+
+
The scanner that reads records form a table in a batch fashion.
+
+
BatchScanUtils - Class in org.apache.fluss.client.table.scanner.batch
+
+
Utility class for batch scan.
+
+
BatchScanUtils() - Constructor for class org.apache.fluss.client.table.scanner.batch.BatchScanUtils
+
 
+
batchSequence() - Method in class org.apache.fluss.client.write.ArrowLogWriteBatch
+
 
+
batchSequence() - Method in class org.apache.fluss.client.write.IndexedLogWriteBatch
+
 
+
batchSequence() - Method in class org.apache.fluss.client.write.KvWriteBatch
+
 
+
batchSequence() - Method in class org.apache.fluss.client.write.WriteBatch
+
 
+
beginFlush() - Method in class org.apache.fluss.client.write.RecordAccumulator
+
+
Initiate the flushing of data from the accumulator...this makes all requests immediately + ready.
+
+
binding() - Method in class org.apache.fluss.client.admin.DropAclsResult.FilterResult
+
+
Return the deleted ACL binding or null if there was an error.
+
+
BOOTSTRAP_SERVERS - Static variable in class org.apache.fluss.flink.FlinkConnectorOptions
+
 
+
bootstrapServers - Variable in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
BoundedSplitReader - Class in org.apache.fluss.flink.source.reader
+
+
A bounded reader to reading Fluss's bounded split (e.g., SnapshotSplit) into RecordAndPoss.
+
+
BoundedSplitReader(BatchScanner, long) - Constructor for class org.apache.fluss.flink.source.reader.BoundedSplitReader
+
 
+
BUCKET_GROUP - Static variable in class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
+
 
+
BUCKET_KEY - Static variable in class org.apache.fluss.flink.FlinkConnectorOptions
+
 
+
BUCKET_NUMBER - Static variable in class org.apache.fluss.flink.FlinkConnectorOptions
+
 
+
bucketId() - Method in class org.apache.fluss.client.write.WriteBatch
+
 
+
BucketOffsetsRetrieverImpl - Class in org.apache.fluss.flink.source.enumerator.initializer
+
+
The default implementation for offsets retriever.
+
+
BucketOffsetsRetrieverImpl(Admin, TablePath) - Constructor for class org.apache.fluss.flink.source.enumerator.initializer.BucketOffsetsRetrieverImpl
+
 
+
bucketResult(int) - Method in class org.apache.fluss.client.admin.ListOffsetsResult
+
 
+
buckets() - Method in class org.apache.fluss.client.table.scanner.log.ScanRecords
+
+
Get the bucket ids which have records contained in this record set.
+
+
build() - Method in class org.apache.fluss.client.write.ArrowLogWriteBatch
+
 
+
build() - Method in class org.apache.fluss.client.write.IndexedLogWriteBatch
+
 
+
build() - Method in class org.apache.fluss.client.write.KvWriteBatch
+
 
+
build() - Method in class org.apache.fluss.client.write.WriteBatch
+
+
Gets the memory segment bytes view of the batch.
+
+
build() - Method in class org.apache.fluss.flink.sink.FlussSinkBuilder
+
+
Build the FlussSink.
+
+
build() - Method in class org.apache.fluss.flink.source.FlussSourceBuilder
+
+
Builds and returns a new FlussSource instance with the configured properties.
+
+
build() - Method in class org.apache.fluss.flink.tiering.LakeTieringJobBuilder
+
 
+
build() - Method in class org.apache.fluss.flink.tiering.source.TieringSource.Builder
+
 
+
builder() - Static method in class org.apache.fluss.flink.sink.FlussSink
+
+
Creates a new FlussSinkBuilder instance for building a FlussSink.
+
+
builder() - Static method in class org.apache.fluss.flink.source.FlussSource
+
+
Get a FlussSourceBuilder to build a FlussSource.
+
+
Builder(Configuration, LakeTieringFactory<WriteResult, ?>) - Constructor for class org.apache.fluss.flink.tiering.source.TieringSource.Builder
+
 
+
bytesPerBatch() - Method in class org.apache.fluss.client.metrics.WriterMetricGroup
+
 
+
bytesPerRequest() - Method in class org.apache.fluss.client.metrics.ScannerMetricGroup
+
 
+
bytesSendTotal() - Method in class org.apache.fluss.client.metrics.WriterMetricGroup
+
 
+
+A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-20.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-20.html new file mode 100644 index 0000000000..03b3777a58 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-20.html @@ -0,0 +1,547 @@ + + + + + +T-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages + + +

T

+
+
table - Variable in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
+
 
+
Table - Interface in org.apache.fluss.client.table
+
+
Used to communicate with a single Fluss table.
+
+
TABLE_OPTIONS - Static variable in class org.apache.fluss.flink.FlinkConnectorOptions
+
 
+
TableAppend - Class in org.apache.fluss.client.table.writer
+
+
API for configuring and creating AppendWriter.
+
+
TableAppend(TablePath, TableInfo, WriterClient) - Constructor for class org.apache.fluss.client.table.writer.TableAppend
+
 
+
tableBucket - Variable in class org.apache.fluss.flink.source.split.SourceSplitBase
+
 
+
tableBucket - Variable in class org.apache.fluss.flink.tiering.source.split.TieringSplit
+
 
+
tableBucket() - Method in class org.apache.fluss.client.lookup.AbstractLookupBatch
+
 
+
tableBucket() - Method in class org.apache.fluss.client.lookup.AbstractLookupQuery
+
 
+
tableBucket() - Method in class org.apache.fluss.client.lookup.LookupBatch
+
 
+
tableBucket() - Method in class org.apache.fluss.client.lookup.PrefixLookupBatch
+
 
+
tableBucket() - Method in class org.apache.fluss.client.write.ReadyWriteBatch
+
 
+
tableBucket() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResult
+
 
+
tableBucket() - Method in class org.apache.fluss.flink.tiering.source.TieringWriterInitContext
+
 
+
TableBucketWriteResult<WriteResult> - Class in org.apache.fluss.flink.tiering.source
+
+
This class contains the TableBucketWriteResult of LakeWriter, the table path and the bucket + that the write result is for, the end log offset of tiering, the total number of write results in + one round of tiering.
+
+
TableBucketWriteResult(TablePath, TableBucket, String, WriteResult, long, int) - Constructor for class org.apache.fluss.flink.tiering.source.TableBucketWriteResult
+
 
+
TableBucketWriteResultEmitter<WriteResult> - Class in org.apache.fluss.flink.tiering.source
+
+
The emitter to emit TableBucketWriteResult to downstream LakeCommitter.
+
+
TableBucketWriteResultEmitter() - Constructor for class org.apache.fluss.flink.tiering.source.TableBucketWriteResultEmitter
+
 
+
TableBucketWriteResultSerializer<WriteResult> - Class in org.apache.fluss.flink.tiering.source
+
+
The serializer for TableBucketWriteResult.
+
+
TableBucketWriteResultSerializer(SimpleVersionedSerializer<WriteResult>) - Constructor for class org.apache.fluss.flink.tiering.source.TableBucketWriteResultSerializer
+
 
+
TableBucketWriteResultTypeInfo<WriteResult> - Class in org.apache.fluss.flink.tiering.source
+
+
A TypeInformation for TableBucketWriteResult .
+
+
tableExists(ObjectPath) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
tableExists(TablePath) - Method in interface org.apache.fluss.client.admin.Admin
+
+
Get whether table exists asynchronously.
+
+
tableExists(TablePath) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
tableInfo() - Method in class org.apache.fluss.flink.tiering.source.TieringWriterInitContext
+
 
+
TableLookup - Class in org.apache.fluss.client.lookup
+
+
API for configuring and creating Lookuper.
+
+
TableLookup(TableInfo, MetadataUpdater, LookupClient) - Constructor for class org.apache.fluss.client.lookup.TableLookup
+
 
+
tablePath - Variable in class org.apache.fluss.client.table.writer.AbstractTableWriter
+
 
+
tablePath - Variable in class org.apache.fluss.flink.tiering.source.split.TieringSplit
+
 
+
tablePath() - Method in class org.apache.fluss.flink.tiering.committer.TieringCommitterInitContext
+
 
+
tablePath() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResult
+
 
+
tablePath() - Method in class org.apache.fluss.flink.tiering.source.TieringWriterInitContext
+
 
+
tableRowType - Variable in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
+
 
+
TableScan - Class in org.apache.fluss.client.table.scanner
+
+
API for configuring and creating LogScanner and BatchScanner.
+
+
TableScan(FlussConnection, TableInfo) - Constructor for class org.apache.fluss.client.table.scanner.TableScan
+
 
+
TableUpsert - Class in org.apache.fluss.client.table.writer
+
+
API for configuring and creating UpsertWriter.
+
+
TableUpsert(TablePath, TableInfo, WriterClient) - Constructor for class org.apache.fluss.client.table.writer.TableUpsert
+
 
+
TableWriter - Interface in org.apache.fluss.client.table.writer
+
+
A writer to write data to a Fluss table.
+
+
targetColumnIndexes - Variable in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
+
 
+
TIERING_LOG_SPLIT_FLAG - Static variable in class org.apache.fluss.flink.tiering.source.split.TieringSplit
+
 
+
TIERING_SNAPSHOT_SPLIT_FLAG - Static variable in class org.apache.fluss.flink.tiering.source.split.TieringSplit
+
 
+
TIERING_SOURCE_OPERATOR_UID - Static variable in class org.apache.fluss.flink.tiering.source.TieringSource
+
 
+
TIERING_SOURCE_TRANSFORMATION_UID - Static variable in class org.apache.fluss.flink.tiering.source.TieringSource
+
 
+
TieringCommitOperator<WriteResult,​Committable> - Class in org.apache.fluss.flink.tiering.committer
+
+
A Flink operator to aggregate TieringCommitOperators by table to TieringCommitOperator which will + then be committed to lake & Fluss cluster.
+
+
TieringCommitOperator(StreamOperatorParameters<CommittableMessage<Committable>>, Configuration, LakeTieringFactory<WriteResult, Committable>) - Constructor for class org.apache.fluss.flink.tiering.committer.TieringCommitOperator
+
 
+
TieringCommitOperatorFactory<WriteResult,​Committable> - Class in org.apache.fluss.flink.tiering.committer
+
+
The factory to create TieringCommitOperator.
+
+
TieringCommitOperatorFactory(Configuration, LakeTieringFactory<WriteResult, Committable>) - Constructor for class org.apache.fluss.flink.tiering.committer.TieringCommitOperatorFactory
+
 
+
TieringCommitterInitContext - Class in org.apache.fluss.flink.tiering.committer
+
+
The CommitterInitContext implementation for LakeCommitter.
+
+
TieringCommitterInitContext(TablePath) - Constructor for class org.apache.fluss.flink.tiering.committer.TieringCommitterInitContext
+
 
+
TieringLogSplit - Class in org.apache.fluss.flink.tiering.source.split
+
+
The table split for tiering service.
+
+
TieringLogSplit(TablePath, TableBucket, String, long, long) - Constructor for class org.apache.fluss.flink.tiering.source.split.TieringLogSplit
+
 
+
TieringLogSplit(TablePath, TableBucket, String, long, long, int) - Constructor for class org.apache.fluss.flink.tiering.source.split.TieringLogSplit
+
 
+
TieringRestoreEvent - Class in org.apache.fluss.flink.tiering.event
+
+
SourceEvent used to represent tiering is restoring.
+
+
TieringRestoreEvent() - Constructor for class org.apache.fluss.flink.tiering.event.TieringRestoreEvent
+
 
+
TieringSnapshotSplit - Class in org.apache.fluss.flink.tiering.source.split
+
+
The table split for tiering service.
+
+
TieringSnapshotSplit(TablePath, TableBucket, String, long, long) - Constructor for class org.apache.fluss.flink.tiering.source.split.TieringSnapshotSplit
+
 
+
TieringSnapshotSplit(TablePath, TableBucket, String, long, long, int) - Constructor for class org.apache.fluss.flink.tiering.source.split.TieringSnapshotSplit
+
 
+
TieringSource<WriteResult> - Class in org.apache.fluss.flink.tiering.source
+
+
The flink source implementation for tiering data from Fluss to downstream lake.
+
+
TieringSource(Configuration, LakeTieringFactory<WriteResult, ?>, long) - Constructor for class org.apache.fluss.flink.tiering.source.TieringSource
+
 
+
TieringSource.Builder<WriteResult> - Class in org.apache.fluss.flink.tiering.source
+
+
Builder for TieringSource.
+
+
TieringSourceEnumerator - Class in org.apache.fluss.flink.tiering.source.enumerator
+
+
An implementation of SplitEnumerator used to request TieringSplit from Fluss + Cluster.
+
+
TieringSourceEnumerator(Configuration, SplitEnumeratorContext<TieringSplit>, long) - Constructor for class org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator
+
 
+
TieringSourceEnumeratorState - Class in org.apache.fluss.flink.tiering.source.state
+
+
The marker class of stateless component TieringSourceEnumerator.
+
+
TieringSourceEnumeratorState() - Constructor for class org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorState
+
 
+
TieringSourceEnumeratorStateSerializer - Class in org.apache.fluss.flink.tiering.source.state
+
+ +
+
TieringSourceEnumeratorStateSerializer() - Constructor for class org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorStateSerializer
+
 
+
TieringSourceOptions - Class in org.apache.fluss.flink.tiering.source
+
+
Configuration options for the TieringSource.
+
+
TieringSourceOptions() - Constructor for class org.apache.fluss.flink.tiering.source.TieringSourceOptions
+
 
+
TieringSourceReader<WriteResult> - Class in org.apache.fluss.flink.tiering.source
+
+
A SourceReader that read records from Fluss and write to lake.
+
+
TieringSourceReader(SourceReaderContext, Configuration, LakeTieringFactory<WriteResult, ?>) - Constructor for class org.apache.fluss.flink.tiering.source.TieringSourceReader
+
 
+
tieringSplit - Variable in class org.apache.fluss.flink.tiering.source.state.TieringSplitState
+
 
+
TieringSplit - Class in org.apache.fluss.flink.tiering.source.split
+
+
The base table split for tiering service.
+
+
TieringSplit(TablePath, TableBucket, String, int) - Constructor for class org.apache.fluss.flink.tiering.source.split.TieringSplit
+
 
+
TieringSplitGenerator - Class in org.apache.fluss.flink.tiering.source.split
+
+
A generator for lake splits.
+
+
TieringSplitGenerator(Admin) - Constructor for class org.apache.fluss.flink.tiering.source.split.TieringSplitGenerator
+
 
+
TieringSplitReader<WriteResult> - Class in org.apache.fluss.flink.tiering.source
+
+
The SplitReader implementation which will read Fluss and write to lake.
+
+
TieringSplitReader(Configuration, LakeTieringFactory<WriteResult, ?>) - Constructor for class org.apache.fluss.flink.tiering.source.TieringSplitReader
+
 
+
TieringSplitSerializer - Class in org.apache.fluss.flink.tiering.source.split
+
+
A serializer for the TieringSplit.
+
+
TieringSplitSerializer() - Constructor for class org.apache.fluss.flink.tiering.source.split.TieringSplitSerializer
+
 
+
TieringSplitState - Class in org.apache.fluss.flink.tiering.source.state
+
+
The state of a TieringSplit.
+
+
TieringSplitState(TieringSplit) - Constructor for class org.apache.fluss.flink.tiering.source.state.TieringSplitState
+
 
+
TieringWriterInitContext - Class in org.apache.fluss.flink.tiering.source
+
+
The implementation of WriterInitContext.
+
+
TieringWriterInitContext(TablePath, TableBucket, String, TableInfo) - Constructor for class org.apache.fluss.flink.tiering.source.TieringWriterInitContext
+
 
+
timestamp() - Method in class org.apache.fluss.client.table.scanner.ScanRecord
+
 
+
timestamp(long) - Static method in interface org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer
+
+
Get an OffsetsInitializer which initializes the offsets in each bucket so that the + initialized offset is the offset of the first record batch whose commit timestamp is greater + than or equals the given timestamp (milliseconds).
+
+
TIMESTAMP - org.apache.fluss.flink.FlinkConnectorOptions.ScanStartupMode
+
 
+
TimestampFormat - Enum in org.apache.fluss.flink.utils
+
+
Timestamp format Enums.
+
+
TimestampOffsetsInitializer - Class in org.apache.fluss.flink.source.enumerator.initializer
+
+
An implementation of OffsetsInitializer to initialize the offsets based on a timestamp.
+
+
TimestampOffsetsInitializer(long) - Constructor for class org.apache.fluss.flink.source.enumerator.initializer.TimestampOffsetsInitializer
+
 
+
TimestampSpec(long) - Constructor for class org.apache.fluss.client.admin.OffsetSpec.TimestampSpec
+
 
+
toBucketOffsetsProperty(Map<TableBucket, Long>, Map<Long, String>, List<String>) - Static method in class org.apache.fluss.flink.tiering.committer.TieringCommitOperator
+
 
+
toCatalogTable(Schema, String, List<String>, Map<String, String>) - Static method in class org.apache.fluss.flink.adapter.CatalogTableAdapter
+
 
+
toFlinkOption(ConfigOption<T>) - Static method in class org.apache.fluss.flink.utils.FlinkConversions
+
+
Convert Fluss's ConfigOption to Flink's ConfigOption.
+
+
toFlinkOptions(Collection<ConfigOption<?>>) - Static method in class org.apache.fluss.flink.utils.FlinkConversions
+
+
Convert Fluss's ConfigOptions to Flink's ConfigOptions.
+
+
toFlinkRowData(LogRecord) - Method in class org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter
+
 
+
toFlinkRowData(InternalRow) - Method in class org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter
+
 
+
toFlinkRowKind(ChangeType) - Static method in class org.apache.fluss.flink.utils.FlinkConversions
+
 
+
toFlinkRowType(RowType) - Static method in class org.apache.fluss.flink.utils.FlinkConversions
+
+
Convert Fluss's RowType to Flink's RowType.
+
+
toFlinkTable(TableInfo) - Static method in class org.apache.fluss.flink.utils.FlinkConversions
+
+
Convert Fluss's table to Flink's table.
+
+
toFlinkType(DataType) - Static method in class org.apache.fluss.flink.utils.FlinkConversions
+
+
Convert Fluss's type to Flink's type.
+
+
toFlussDatabase(CatalogDatabase) - Static method in class org.apache.fluss.flink.utils.FlinkConversions
+
+
Convert Flink's table to Fluss's database.
+
+
toFlussRowType(RowType) - Static method in class org.apache.fluss.flink.utils.FlinkConversions
+
+
Convert Flink's RowType to Fluss' RowType.
+
+
toFlussTable(ResolvedCatalogTable) - Static method in class org.apache.fluss.flink.utils.FlinkConversions
+
+
Convert Flink's table to Fluss's table.
+
+
toFlussType(DataType) - Static method in class org.apache.fluss.flink.utils.FlinkConversions
+
+
Convert Flink's physical type to Fluss' type.
+
+
toFsPathAndFileName(List<PbRemotePathAndLocalFile>) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
+
 
+
toInternalRow(RowData) - Method in class org.apache.fluss.flink.utils.FlinkRowToFlussRowConverter
+
+
Deprecated.
+
toKeyValueMap(List<PbKeyValue>) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
+
 
+
toKvSnapshotMetadata(GetKvSnapshotMetadataResponse) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
+
 
+
toKvSnapshots(GetLatestKvSnapshotsResponse) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
+
 
+
toLakeTableSnapshotInfo(GetLatestLakeSnapshotResponse) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
+
 
+
toList(CloseableIterator<T>) - Static method in class org.apache.fluss.client.table.scanner.batch.BatchScanUtils
+
+
Collect all rows from the iterator and close it.
+
+
toPartitionInfos(ListPartitionInfosResponse) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
+
 
+
toSecurityToken(GetFileSystemSecurityTokenResponse) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
+
 
+
toSourceSplit() - Method in class org.apache.fluss.flink.lake.state.LakeSnapshotAndFlussLogSplitState
+
 
+
toSourceSplit() - Method in class org.apache.fluss.flink.lake.state.LakeSnapshotSplitState
+
 
+
toSourceSplit() - Method in class org.apache.fluss.flink.source.split.HybridSnapshotLogSplitState
+
 
+
toSourceSplit() - Method in class org.apache.fluss.flink.source.split.LogSplitState
+
 
+
toSourceSplit() - Method in class org.apache.fluss.flink.source.split.SourceSplitState
+
 
+
toSourceSplit() - Method in class org.apache.fluss.flink.tiering.source.state.TieringSplitState
+
 
+
toSplitId(String, TableBucket) - Static method in class org.apache.fluss.flink.source.split.SourceSplitBase
+
 
+
toSplitId(String, TableBucket) - Static method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
+
 
+
toSplitType(String, SourceSplitState) - Method in class org.apache.fluss.flink.source.reader.FlinkSourceReader
+
 
+
toSplitType(String, TieringSplitState) - Method in class org.apache.fluss.flink.tiering.source.TieringSourceReader
+
 
+
toString() - Method in class org.apache.fluss.client.lookup.LookupResult
+
 
+
toString() - Method in class org.apache.fluss.client.metadata.KvSnapshotMetadata
+
 
+
toString() - Method in class org.apache.fluss.client.metadata.LakeSnapshot
+
 
+
toString() - Method in class org.apache.fluss.client.table.scanner.ScanRecord
+
 
+
toString() - Method in enum org.apache.fluss.flink.FlinkConnectorOptions.ScanStartupMode
+
 
+
toString() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
+
 
+
toString() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotSplit
+
 
+
toString() - Method in class org.apache.fluss.flink.sink.FlinkRowDataChannelComputer
+
 
+
toString() - Method in class org.apache.fluss.flink.sink.FlinkStreamPartitioner
+
 
+
toString() - Method in class org.apache.fluss.flink.source.event.PartitionBucketsUnsubscribedEvent
+
 
+
toString() - Method in class org.apache.fluss.flink.source.event.PartitionsRemovedEvent
+
 
+
toString() - Method in class org.apache.fluss.flink.source.reader.RecordAndPos
+
 
+
toString() - Method in class org.apache.fluss.flink.source.split.HybridSnapshotLogSplit
+
 
+
toString() - Method in class org.apache.fluss.flink.source.split.LogSplit
+
 
+
toString() - Method in class org.apache.fluss.flink.source.split.SnapshotSplit
+
 
+
toString() - Method in class org.apache.fluss.flink.source.state.SourceEnumeratorState
+
 
+
toString() - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
+
 
+
toString() - Method in class org.apache.fluss.flink.tiering.source.split.TieringLogSplit
+
 
+
toString() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSnapshotSplit
+
 
+
toString() - Method in class org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorState
+
 
+
toString() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
+
 
+
toTablePath(ObjectPath) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
transferAllToDirectory(List<FsPathAndFileName>, Path, CloseableRegistry) - Method in class org.apache.fluss.client.table.scanner.RemoteFileDownloader
+
 
+
tryAppend(WriteRecord, WriteCallback) - Method in class org.apache.fluss.client.write.ArrowLogWriteBatch
+
 
+
tryAppend(WriteRecord, WriteCallback) - Method in class org.apache.fluss.client.write.IndexedLogWriteBatch
+
 
+
tryAppend(WriteRecord, WriteCallback) - Method in class org.apache.fluss.client.write.KvWriteBatch
+
 
+
tryAppend(WriteRecord, WriteCallback) - Method in class org.apache.fluss.client.write.WriteBatch
+
+
try to append one write record to the record batch.
+
+
+A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-21.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-21.html new file mode 100644 index 0000000000..e8b47bdece --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-21.html @@ -0,0 +1,222 @@ + + + + + +U-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages + + +

U

+
+
UNINITIALIZED - Static variable in class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
+
 
+
UNKNOWN_NUMBER_OF_SPLITS - Static variable in class org.apache.fluss.flink.tiering.source.split.TieringSplit
+
 
+
unknownLeaderTables - Variable in class org.apache.fluss.client.write.RecordAccumulator.ReadyCheckResult
+
 
+
unregister(Metric, String, AbstractMetricGroup) - Method in class org.apache.fluss.flink.metrics.FlinkMetricRegistry
+
 
+
unsubscribe(long, int) - Method in interface org.apache.fluss.client.table.scanner.log.LogScanner
+
+
Unsubscribe from the given bucket of given partition dynamically.
+
+
unsubscribe(long, int) - Method in class org.apache.fluss.client.table.scanner.log.LogScannerImpl
+
 
+
update(long) - Method in class org.apache.fluss.flink.metrics.FlinkHistogram
+
 
+
updateEstimation(PhysicalTablePath, int) - Method in class org.apache.fluss.client.write.DynamicWriteBatchSizeEstimator
+
 
+
updateFetchLatency(long) - Method in class org.apache.fluss.client.metrics.ScannerMetricGroup
+
 
+
updateMetadata(Set<TablePath>, Collection<PhysicalTablePath>, Collection<Long>) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
+
 
+
updatePhysicalTableMetadata(Set<PhysicalTablePath>) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
+
+
Update the table or partition metadata info.
+
+
updateTableOrPartitionMetadata(TablePath, Long) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
+
 
+
updateWithCurrentLakeSplitIndex(int) - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
+
 
+
updateWithLakeSplitFinished(boolean) - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
+
 
+
updateWithRecordsToSkip(long) - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
+
 
+
updateWithStartingOffset(long) - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
+
 
+
upsert(InternalRow) - Method in interface org.apache.fluss.client.table.writer.UpsertWriter
+
+
Inserts row into Fluss table if they do not already exist, or updates them if they do exist.
+
+
Upsert - Interface in org.apache.fluss.client.table.writer
+
+
Used to configure and create UpsertWriter to upsert and delete data to a Primary Key + Table.
+
+
UPSERT - org.apache.fluss.flink.row.OperationType
+
+
Represents an upsert operation (update or insert).
+
+
UpsertResult - Class in org.apache.fluss.client.table.writer
+
+
The result of upserting a record (UpsertWriter.upsert(InternalRow)).
+
+
UpsertResult() - Constructor for class org.apache.fluss.client.table.writer.UpsertResult
+
 
+
UpsertSinkWriter<InputT> - Class in org.apache.fluss.flink.sink.writer
+
+
An upsert sink writer or fluss primary key table.
+
+
UpsertSinkWriter(TablePath, Configuration, RowType, int[], MailboxExecutor, FlussSerializationSchema<InputT>) - Constructor for class org.apache.fluss.flink.sink.writer.UpsertSinkWriter
+
 
+
UpsertWrite - Class in org.apache.fluss.client.table.writer
+
+
Used to describe the operation to write data by UpsertWriter to a table.
+
+
UpsertWrite() - Constructor for class org.apache.fluss.client.table.writer.UpsertWrite
+
 
+
UpsertWriter - Interface in org.apache.fluss.client.table.writer
+
+
The writer to write data to the primary key table.
+
+
+A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-22.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-22.html new file mode 100644 index 0000000000..fffe6643a3 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-22.html @@ -0,0 +1,222 @@ + + + + + +V-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages + + +

V

+
+
validateAndCreateLookupNormalizer(int[][], int[], int[], int[], RowType, int[]) - Static method in class org.apache.fluss.flink.source.lookup.LookupNormalizer
+
+
Validate the lookup key indexes and primary keys, and create a LookupNormalizer.
+
+
validateTableSourceOptions(ReadableConfig) - Static method in class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils
+
 
+
valueOf(String) - Static method in enum org.apache.fluss.client.lookup.LookupType
+
+
Returns the enum constant of this type with the specified name.
+
+
valueOf(String) - Static method in enum org.apache.fluss.client.write.WriteFormat
+
+
Returns the enum constant of this type with the specified name.
+
+
valueOf(String) - Static method in enum org.apache.fluss.flink.FlinkConnectorOptions.ScanStartupMode
+
+
Returns the enum constant of this type with the specified name.
+
+
valueOf(String) - Static method in enum org.apache.fluss.flink.row.OperationType
+
+
Returns the enum constant of this type with the specified name.
+
+
valueOf(String) - Static method in enum org.apache.fluss.flink.utils.PushdownUtils.ValueConversion
+
+
Returns the enum constant of this type with the specified name.
+
+
valueOf(String) - Static method in enum org.apache.fluss.flink.utils.TimestampFormat
+
+
Returns the enum constant of this type with the specified name.
+
+
valueRow() - Method in class org.apache.fluss.flink.lake.reader.KeyValueRow
+
 
+
values() - Method in class org.apache.fluss.client.admin.DropAclsResult.FilterResults
+
+
Return a list of delete ACLs results for a given filter.
+
+
values() - Method in class org.apache.fluss.client.admin.DropAclsResult
+
+
Return a map from acl filters to futures which can be used to check the status of the + deletions by each filter.
+
+
values() - Static method in enum org.apache.fluss.client.lookup.LookupType
+
+
Returns an array containing the constants of this enum type, in +the order they are declared.
+
+
values() - Static method in enum org.apache.fluss.client.write.WriteFormat
+
+
Returns an array containing the constants of this enum type, in +the order they are declared.
+
+
values() - Static method in enum org.apache.fluss.flink.FlinkConnectorOptions.ScanStartupMode
+
+
Returns an array containing the constants of this enum type, in +the order they are declared.
+
+
values() - Static method in enum org.apache.fluss.flink.row.OperationType
+
+
Returns an array containing the constants of this enum type, in +the order they are declared.
+
+
values() - Static method in enum org.apache.fluss.flink.utils.PushdownUtils.ValueConversion
+
+
Returns an array containing the constants of this enum type, in +the order they are declared.
+
+
values() - Static method in enum org.apache.fluss.flink.utils.TimestampFormat
+
+
Returns an array containing the constants of this enum type, in +the order they are declared.
+
+
+A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-23.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-23.html new file mode 100644 index 0000000000..1c5ac6f1d0 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-23.html @@ -0,0 +1,232 @@ + + + + + +W-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages + + +

W

+
+
waitedTimeMs(long) - Method in class org.apache.fluss.client.write.WriteBatch
+
 
+
wakeup() - Method in class org.apache.fluss.client.table.scanner.log.LogFetchBuffer
+
 
+
wakeup() - Method in class org.apache.fluss.client.table.scanner.log.LogFetcher
+
 
+
wakeup() - Method in interface org.apache.fluss.client.table.scanner.log.LogScanner
+
+
Wake up the log scanner in case the fetcher thread in log scanner is blocking in LogScanner.poll(Duration timeout).
+
+
wakeup() - Method in class org.apache.fluss.client.table.scanner.log.LogScannerImpl
+
 
+
wakeUp() - Method in class org.apache.fluss.flink.source.reader.FlinkSourceSplitReader
+
 
+
wakeUp() - Method in class org.apache.fluss.flink.tiering.source.TieringSplitReader
+
 
+
withAdmin(Admin) - Method in class org.apache.fluss.flink.procedure.ProcedureBase
+
 
+
withPartialUpdate(int[]) - Method in class org.apache.fluss.client.table.writer.UpsertWrite
+
+
Returns a new instance of UpsertWrite description with partial update the specified columns.
+
+
withPollTieringTableIntervalMs(long) - Method in class org.apache.fluss.flink.tiering.source.TieringSource.Builder
+
 
+
withProjectedFields(int[]) - Method in class org.apache.fluss.client.table.scanner.log.LogScan
+
+
Returns a new instance of LogScan description with column projection.
+
+
wrapIntoNullableInternalConverter(FlussRowToFlinkRowConverter.FlussDeserializationConverter) - Method in class org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter
+
 
+
write(InputT, SinkWriter.Context) - Method in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
+
 
+
writeBatch() - Method in class org.apache.fluss.client.write.ReadyWriteBatch
+
 
+
WriteBatch - Class in org.apache.fluss.client.write
+
+
The abstract write batch contains write callback object to wait write request feedback.
+
+
WriteBatch(int, PhysicalTablePath, long) - Constructor for class org.apache.fluss.client.write.WriteBatch
+
 
+
WriteBatch.RequestFuture - Class in org.apache.fluss.client.write
+
+
The future for this batch.
+
+
WriteCallback - Interface in org.apache.fluss.client.write
+
+
A callback interface that the user can implement to allow code to execute when the write request + is complete.
+
+
WriteFormat - Enum in org.apache.fluss.client.write
+
+
The format of the write record.
+
+
writerClient - Variable in class org.apache.fluss.client.table.writer.AbstractTableWriter
+
 
+
WriterClient - Class in org.apache.fluss.client.write
+
+
A client that write records to server.
+
+
WriterClient(Configuration, MetadataUpdater, ClientMetricGroup, Admin) - Constructor for class org.apache.fluss.client.write.WriterClient
+
 
+
WriteRecord - Class in org.apache.fluss.client.write
+
+
A record to write to a table.
+
+
writeResult() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResult
+
 
+
writerId() - Method in class org.apache.fluss.client.write.ArrowLogWriteBatch
+
 
+
writerId() - Method in class org.apache.fluss.client.write.IndexedLogWriteBatch
+
 
+
writerId() - Method in class org.apache.fluss.client.write.KvWriteBatch
+
 
+
writerId() - Method in class org.apache.fluss.client.write.WriteBatch
+
 
+
WriterMetricGroup - Class in org.apache.fluss.client.metrics
+
+
Metrics for WriterClient.
+
+
WriterMetricGroup(ClientMetricGroup) - Constructor for class org.apache.fluss.client.metrics.WriterMetricGroup
+
 
+
+A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-3.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-3.html new file mode 100644 index 0000000000..3b152c3dac --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-3.html @@ -0,0 +1,561 @@ + + + + + +C-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages + + +

C

+
+
call(ProcedureContext, String, String, String, String, String) - Method in class org.apache.fluss.flink.procedure.AddAclProcedure
+
 
+
call(ProcedureContext, String, String, String, String, String) - Method in class org.apache.fluss.flink.procedure.DropAclProcedure
+
 
+
call(ProcedureContext, String, String, String, String, String) - Method in class org.apache.fluss.flink.procedure.ListAclProcedure
+
 
+
callbacks - Variable in class org.apache.fluss.client.write.WriteBatch
+
 
+
canEqual(Object) - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
+
 
+
canEqual(Object) - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
+
 
+
CatalogExceptionUtils - Class in org.apache.fluss.flink.utils
+
+
Utility class for catalog exceptions.
+
+
catalogName - Variable in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
CatalogPropertiesUtils - Class in org.apache.fluss.flink.utils
+
+
Utilities for de/serializing Catalog objects into a map of string properties.
+
+
CatalogTableAdapter - Class in org.apache.fluss.flink.adapter
+
+
A adapter for CatalogTable constructor.
+
+
CatalogTableAdapter() - Constructor for class org.apache.fluss.flink.adapter.CatalogTableAdapter
+
 
+
channel(InputT) - Method in class org.apache.fluss.flink.sink.FlinkRowDataChannelComputer
+
 
+
channel(T) - Method in interface org.apache.fluss.flink.sink.ChannelComputer
+
 
+
ChannelComputer<T> - Interface in org.apache.fluss.flink.sink
+
+
A utility class to compute which downstream channel a given record should be sent to before flink + sink.
+
+
checkAndCreatePartitionAsync(PhysicalTablePath) - Method in class org.apache.fluss.client.write.DynamicPartitionCreator
+
 
+
checkAndUpdateMetadata(TablePath, TableBucket) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
+
+
Check the table/partition info for the given table bucket exist in metadata cache, if not, + try to update the metadata cache.
+
+
checkAndUpdatePartitionMetadata(PhysicalTablePath) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
+
+
Check the partition exists in metadata cache, if not, try to update the metadata cache, if + not exist yet, throw exception.
+
+
checkAndUpdatePartitionMetadata(TablePath, Collection<Long>) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
+
+
Check the partitions info for the given partition ids exist in metadata cache, if not, try to + update the metadata cache.
+
+
checkAndUpdateTableMetadata(Set<TablePath>) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
+
 
+
checkAsyncException() - Method in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
+
+
This method should only be invoked in the mailbox thread since the counter is not volatile.
+
+
checkFieldCount(InternalRow) - Method in class org.apache.fluss.client.table.writer.AbstractTableWriter
+
 
+
classLoader - Variable in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
CLIENT_OPTIONS - Static variable in class org.apache.fluss.flink.FlinkConnectorOptions
+
 
+
ClientRpcMessageUtils - Class in org.apache.fluss.client.utils
+
+
Utils for making rpc request/response from inner object or convert inner class to rpc + request/response for client.
+
+
ClientRpcMessageUtils() - Constructor for class org.apache.fluss.client.utils.ClientRpcMessageUtils
+
 
+
ClientUtils - Class in org.apache.fluss.client.utils
+
+
Utils for Fluss Client.
+
+
close() - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
close() - Method in interface org.apache.fluss.client.Connection
+
+
Close the connection and release all resources.
+
+
close() - Method in class org.apache.fluss.client.FlussConnection
+
 
+
close() - Method in class org.apache.fluss.client.table.FlussTable
+
 
+
close() - Method in interface org.apache.fluss.client.table.scanner.batch.BatchScanner
+
+
Closes the scanner and should release all resources.
+
+
close() - Method in class org.apache.fluss.client.table.scanner.batch.KvSnapshotBatchScanner
+
 
+
close() - Method in class org.apache.fluss.client.table.scanner.batch.LimitBatchScanner
+
 
+
close() - Method in class org.apache.fluss.client.table.scanner.log.LogFetchBuffer
+
 
+
close() - Method in class org.apache.fluss.client.table.scanner.log.LogFetcher
+
 
+
close() - Method in class org.apache.fluss.client.table.scanner.log.LogScannerImpl
+
 
+
close() - Method in class org.apache.fluss.client.table.scanner.log.RemoteLogDownloader
+
 
+
close() - Method in class org.apache.fluss.client.table.scanner.RemoteFileDownloader
+
 
+
close() - Method in class org.apache.fluss.client.write.ArrowLogWriteBatch
+
 
+
close() - Method in class org.apache.fluss.client.write.IndexedLogWriteBatch
+
 
+
close() - Method in class org.apache.fluss.client.write.KvWriteBatch
+
 
+
close() - Method in class org.apache.fluss.client.write.RecordAccumulator
+
+
Close this accumulator and force all the record buffers to be drained.
+
+
close() - Method in class org.apache.fluss.client.write.WriteBatch
+
+
close the batch to not append new records.
+
+
close() - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
close() - Method in class org.apache.fluss.flink.lake.reader.IndexedLakeSplitRecordIterator
+
 
+
close() - Method in class org.apache.fluss.flink.lake.reader.LakeSnapshotAndLogSplitScanner
+
 
+
close() - Method in class org.apache.fluss.flink.lake.reader.LakeSnapshotScanner
+
 
+
close() - Method in class org.apache.fluss.flink.lake.reader.SeekableLakeSnapshotSplitScanner
+
 
+
close() - Method in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
+
 
+
close() - Method in class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
+
 
+
close() - Method in class org.apache.fluss.flink.source.lookup.FlinkAsyncLookupFunction
+
 
+
close() - Method in class org.apache.fluss.flink.source.lookup.FlinkLookupFunction
+
 
+
close() - Method in class org.apache.fluss.flink.source.reader.BoundedSplitReader
+
 
+
close() - Method in class org.apache.fluss.flink.source.reader.FlinkSourceSplitReader
+
 
+
close() - Method in class org.apache.fluss.flink.tiering.committer.FlussTableLakeSnapshotCommitter
+
 
+
close() - Method in class org.apache.fluss.flink.tiering.committer.TieringCommitOperator
+
 
+
close() - Method in class org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator
+
 
+
close() - Method in class org.apache.fluss.flink.tiering.source.TieringSplitReader
+
 
+
close() - Method in class org.apache.fluss.flink.utils.FlinkRowToFlussRowConverter
+
+
Deprecated.
+
close(Duration) - Method in class org.apache.fluss.client.lookup.LookupClient
+
 
+
close(Duration) - Method in class org.apache.fluss.client.write.WriterClient
+
 
+
closeAsync() - Method in class org.apache.fluss.flink.metrics.FlinkMetricRegistry
+
 
+
cluster - Variable in class org.apache.fluss.client.metadata.MetadataUpdater
+
 
+
collectAllRows(List<BatchScanner>) - Static method in class org.apache.fluss.client.table.scanner.batch.BatchScanUtils
+
+
Collect all rows from the scanners.
+
+
collectFetch() - Method in class org.apache.fluss.client.table.scanner.log.LogFetcher
+
 
+
collectFetch(LogFetchBuffer) - Method in class org.apache.fluss.client.table.scanner.log.LogFetchCollector
+
+
Return the fetched log records, empty the record buffer and update the consumed position.
+
+
collectLimitedRows(List<BatchScanner>, int) - Static method in class org.apache.fluss.client.table.scanner.batch.BatchScanUtils
+
+
Collect rows from the scanners until the number of rows reaches the limit.
+
+
collectRows(BatchScanner) - Static method in class org.apache.fluss.client.table.scanner.batch.BatchScanUtils
+
+
Collect all rows from the scanner.
+
+
columnKey(int) - Static method in class org.apache.fluss.flink.utils.CatalogPropertiesUtils
+
 
+
commit(long, CommittedLakeSnapshot) - Method in class org.apache.fluss.flink.tiering.committer.FlussTableLakeSnapshotCommitter
+
 
+
committable() - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessage
+
 
+
CommittableMessage<Committable> - Class in org.apache.fluss.flink.tiering.committer
+
+
A class wrapping CommittableMessage to commit to lake.
+
+
CommittableMessage(Committable) - Constructor for class org.apache.fluss.flink.tiering.committer.CommittableMessage
+
 
+
CommittableMessageTypeInfo<Committable> - Class in org.apache.fluss.flink.tiering.committer
+
+
A TypeInformation for CommittableMessage.
+
+
complete() - Method in class org.apache.fluss.client.write.WriteBatch
+
+
Complete the batch successfully.
+
+
complete(List<byte[]>) - Method in class org.apache.fluss.client.lookup.LookupBatch
+
+
Complete the lookup operations using given values .
+
+
complete(List<List<byte[]>>) - Method in class org.apache.fluss.client.lookup.PrefixLookupBatch
+
 
+
complete(List<PbCreateAclRespInfo>) - Method in class org.apache.fluss.client.admin.CreateAclsResult
+
+
Completes individual futures based on RPC response information.
+
+
complete(List<PbDropAclsFilterResult>) - Method in class org.apache.fluss.client.admin.DropAclsResult
+
 
+
complete(List<T>) - Method in class org.apache.fluss.client.lookup.AbstractLookupBatch
+
+
Complete the lookup operations using given values .
+
+
completeExceptionally(Exception) - Method in class org.apache.fluss.client.lookup.AbstractLookupBatch
+
+
Complete the get operations with given exception.
+
+
completeExceptionally(Exception) - Method in class org.apache.fluss.client.lookup.LookupBatch
+
+
Complete the lookup operations with given exception.
+
+
completeExceptionally(Exception) - Method in class org.apache.fluss.client.lookup.PrefixLookupBatch
+
+
Complete the get operations with given exception.
+
+
completeExceptionally(Exception) - Method in class org.apache.fluss.client.write.WriteBatch
+
+
Complete the batch exceptionally.
+
+
completeExceptionally(Throwable) - Method in class org.apache.fluss.client.admin.CreateAclsResult
+
+
Marks all futures as exceptionally completed with the provided throwable.
+
+
completeExceptionally(Throwable) - Method in class org.apache.fluss.client.admin.DropAclsResult
+
 
+
connection - Variable in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
Connection - Interface in org.apache.fluss.client
+
+
A cluster connection encapsulating lower level individual connections to actual Fluss servers.
+
+
ConnectionFactory - Class in org.apache.fluss.client
+
+
A non-instantiable class that manages creation of Connections.
+
+
convert(ObjectMapper, JsonNode, Object) - Method in interface org.apache.fluss.flink.utils.FlussRowToJsonConverters.FlussRowToJsonConverter
+
 
+
convert(T) - Method in class org.apache.fluss.flink.utils.PojoToRowConverter
+
+
Converts a POJO to a GenericRow object according to the schema.
+
+
copy() - Method in class org.apache.fluss.flink.sink.FlinkStreamPartitioner
+
 
+
copy() - Method in class org.apache.fluss.flink.sink.FlinkTableSink
+
 
+
copy() - Method in class org.apache.fluss.flink.source.FlinkTableSource
+
 
+
copy(int) - Method in class org.apache.fluss.flink.tiering.source.split.TieringLogSplit
+
 
+
copy(int) - Method in class org.apache.fluss.flink.tiering.source.split.TieringSnapshotSplit
+
 
+
copy(int) - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
+
 
+
count() - Method in class org.apache.fluss.client.table.scanner.log.ScanRecords
+
+
The number of records for all buckets.
+
+
countLogTable(TablePath, Configuration) - Static method in class org.apache.fluss.flink.utils.PushdownUtils
+
 
+
create(RowType) - Static method in class org.apache.fluss.flink.utils.FlinkRowToFlussRowConverter
+
+
Deprecated.
+
Create a FlinkRowToFlussRowConverter which will convert to IndexedRow.
+
+
create(RowType, KvFormat) - Static method in class org.apache.fluss.flink.utils.FlinkRowToFlussRowConverter
+
+
Deprecated.
+
Create a FlinkRowToFlussRowConverter according to the given KvFormat.
+
+
createAcls(Collection<AclBinding>) - Method in interface org.apache.fluss.client.admin.Admin
+
+
Creates multiple ACL entries in a single atomic operation.
+
+
createAcls(Collection<AclBinding>) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
CreateAclsResult - Class in org.apache.fluss.client.admin
+
+
Represents the result of a batch ACL operation, managing asynchronous completion of individual + ACL operations.
+
+
CreateAclsResult(Collection<AclBinding>) - Constructor for class org.apache.fluss.client.admin.CreateAclsResult
+
 
+
createBatchScanner(TableBucket) - Method in interface org.apache.fluss.client.table.scanner.Scan
+
+
Creates a BatchScanner to read current data in the given table bucket for this scan.
+
+
createBatchScanner(TableBucket) - Method in class org.apache.fluss.client.table.scanner.TableScan
+
 
+
createBatchScanner(TableBucket, long) - Method in interface org.apache.fluss.client.table.scanner.Scan
+
+
Creates a BatchScanner to read given snapshot data in the given table bucket for this + scan.
+
+
createBatchScanner(TableBucket, long) - Method in class org.apache.fluss.client.table.scanner.TableScan
+
 
+
createCatalog(CatalogFactory.Context) - Method in class org.apache.fluss.flink.catalog.FlinkCatalogFactory
+
 
+
createConnection(Configuration) - Static method in class org.apache.fluss.client.ConnectionFactory
+
+
Creates a new Connection to the Fluss cluster.
+
+
createConnection(Configuration, MetricRegistry) - Static method in class org.apache.fluss.client.ConnectionFactory
+
+
Create a new Connection to the Fluss cluster with registering metrics to the given + metricRegistry.
+
+
createDatabase(String, CatalogDatabase, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
createDatabase(String, DatabaseDescriptor, boolean) - Method in interface org.apache.fluss.client.admin.Admin
+
+
Create a new database asynchronously.
+
+
createDatabase(String, DatabaseDescriptor, boolean) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
createDynamicTableSink(DynamicTableFactory.Context) - Method in class org.apache.fluss.flink.catalog.FlinkTableFactory
+
 
+
createDynamicTableSource(DynamicTableFactory.Context) - Method in class org.apache.fluss.flink.catalog.FlinkTableFactory
+
 
+
createDynamicTableSource(DynamicTableFactory.Context, String) - Method in class org.apache.fluss.flink.lake.LakeTableFactory
+
 
+
createEnumerator(SplitEnumeratorContext<SourceSplitBase>) - Method in class org.apache.fluss.flink.source.FlinkSource
+
 
+
createEnumerator(SplitEnumeratorContext<TieringSplit>) - Method in class org.apache.fluss.flink.tiering.source.TieringSource
+
 
+
createFunction(ObjectPath, CatalogFunction, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
createLakeSource(TablePath, Map<String, String>) - Static method in class org.apache.fluss.flink.utils.LakeSourceUtils
+
+
Return the lake source of the given table.
+
+
createLogScanner() - Method in interface org.apache.fluss.client.table.scanner.Scan
+
+
Creates a LogScanner to continuously read log data for this scan.
+
+
createLogScanner() - Method in class org.apache.fluss.client.table.scanner.TableScan
+
 
+
createLookuper() - Method in interface org.apache.fluss.client.lookup.Lookup
+
+
Creates a Lookuper instance to lookup rows of a primary key table by the specified + lookup columns.
+
+
createLookuper() - Method in class org.apache.fluss.client.lookup.TableLookup
+
 
+
createNullableConverter(DataType) - Method in class org.apache.fluss.flink.utils.FlussRowToJsonConverters
+
 
+
createNullableInternalConverter(DataType) - Method in class org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter
+
+
Create a nullable runtime FlussRowToFlinkRowConverter.FlussDeserializationConverter from given DataType.
+
+
createPartition(ObjectPath, CatalogPartitionSpec, CatalogPartition, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
createPartition(TablePath, PartitionSpec, boolean) - Method in interface org.apache.fluss.client.admin.Admin
+
+
Create a new partition for a partitioned table.
+
+
createPartition(TablePath, PartitionSpec, boolean) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
createPrimaryKeyLookupNormalizer(int[], RowType) - Static method in class org.apache.fluss.flink.source.lookup.LookupNormalizer
+
+
Create a LookupNormalizer for primary key lookup.
+
+
createReader(SourceReaderContext) - Method in class org.apache.fluss.flink.source.FlinkSource
+
 
+
createReader(SourceReaderContext) - Method in class org.apache.fluss.flink.tiering.source.TieringSource
+
 
+
createRemainingFilter(RowData) - Method in class org.apache.fluss.flink.source.lookup.LookupNormalizer
+
 
+
createSerializer(ExecutionConfig) - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
+
 
+
createSerializer(ExecutionConfig) - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
+
 
+
createStreamOperator(StreamOperatorParameters<CommittableMessage<Committable>>) - Method in class org.apache.fluss.flink.tiering.committer.TieringCommitOperatorFactory
+
 
+
createTable(ObjectPath, CatalogBaseTable, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
createTable(TablePath, TableDescriptor, boolean) - Method in interface org.apache.fluss.client.admin.Admin
+
+
Create a new table asynchronously.
+
+
createTable(TablePath, TableDescriptor, boolean) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
createWriter() - Method in interface org.apache.fluss.client.table.writer.Append
+
+
Create a new AppendWriter to write data to a Log Table.
+
+
createWriter() - Method in class org.apache.fluss.client.table.writer.TableAppend
+
 
+
createWriter() - Method in class org.apache.fluss.client.table.writer.TableUpsert
+
 
+
createWriter() - Method in interface org.apache.fluss.client.table.writer.Upsert
+
+
Create a new UpsertWriter with the optional Upsert.partialUpdate(String...) + information to upsert and delete data to a Primary Key Table.
+
+
CURRENT_OFFSET_METRIC_GAUGE - Static variable in class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
+
 
+
currentSplitIndex - Variable in class org.apache.fluss.flink.source.reader.RecordAndPos
+
 
+
+A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-4.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-4.html new file mode 100644 index 0000000000..16d40b4acc --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-4.html @@ -0,0 +1,334 @@ + + + + + +D-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages + + +

D

+
+
DATA_LAKE_CONFIG_PREFIX - Static variable in class org.apache.fluss.flink.tiering.source.TieringSourceOptions
+
 
+
databaseExists(String) - Method in interface org.apache.fluss.client.admin.Admin
+
+
Get whether database exists asynchronously.
+
+
databaseExists(String) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
databaseExists(String) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
DataLakeUtils - Class in org.apache.fluss.flink.utils
+
+
Utility class for accessing data lake related configurations.
+
+
DataLakeUtils() - Constructor for class org.apache.fluss.flink.utils.DataLakeUtils
+
 
+
deallocate(WriteBatch) - Method in class org.apache.fluss.client.write.RecordAccumulator
+
+
Deallocate the record batch.
+
+
dec() - Method in class org.apache.fluss.flink.metrics.FlinkCounter
+
 
+
dec(long) - Method in class org.apache.fluss.flink.metrics.FlinkCounter
+
 
+
DEFAULT_DATABASE - Static variable in class org.apache.fluss.flink.catalog.FlinkCatalogOptions
+
 
+
DEFAULT_SPLIT_INDEX - Static variable in class org.apache.fluss.flink.source.reader.RecordAndPos
+
 
+
defaultDatabase - Variable in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
DefaultSecurityTokenManager - Class in org.apache.fluss.client.token
+
+
Manager for security tokens to access files in fluss client.
+
+
DefaultSecurityTokenManager(Configuration, SecurityTokenProvider) - Constructor for class org.apache.fluss.client.token.DefaultSecurityTokenManager
+
 
+
DefaultSecurityTokenProvider - Class in org.apache.fluss.client.token
+
+
A default implementation of SecurityTokenProvider to get token from server.
+
+
DefaultSecurityTokenProvider(AdminReadOnlyGateway) - Constructor for class org.apache.fluss.client.token.DefaultSecurityTokenProvider
+
 
+
delete(InternalRow) - Method in interface org.apache.fluss.client.table.writer.UpsertWriter
+
+
Delete certain row by the input row in Fluss table, the input row must contain the primary + key.
+
+
DELETE - org.apache.fluss.flink.row.OperationType
+
+
Represents a delete operation.
+
+
DeleteResult - Class in org.apache.fluss.client.table.writer
+
+
The result of deleting a record (UpsertWriter.delete(InternalRow).
+
+
DeleteResult() - Constructor for class org.apache.fluss.client.table.writer.DeleteResult
+
 
+
deleteSingleRow(GenericRow, TablePath, Configuration) - Static method in class org.apache.fluss.flink.utils.PushdownUtils
+
 
+
deserialize(byte, TableBucket, String, DataInputDeserializer) - Method in class org.apache.fluss.flink.lake.LakeSplitSerializer
+
 
+
deserialize(int, byte[]) - Method in class org.apache.fluss.flink.source.split.SourceSplitSerializer
+
 
+
deserialize(int, byte[]) - Method in class org.apache.fluss.flink.source.state.FlussSourceEnumeratorStateSerializer
+
 
+
deserialize(int, byte[]) - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplitSerializer
+
 
+
deserialize(int, byte[]) - Method in class org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorStateSerializer
+
 
+
deserialize(int, byte[]) - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultSerializer
+
 
+
deserialize(Object) - Method in interface org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter.FlussDeserializationConverter
+
+
Convert a Fluss field object of InternalRow to the Flink's internal data + structure object.
+
+
deserialize(LogRecord) - Method in interface org.apache.fluss.flink.source.deserializer.FlussDeserializationSchema
+
+
Deserializes a LogRecord into an object of type T.
+
+
deserialize(LogRecord) - Method in class org.apache.fluss.flink.source.deserializer.JsonStringDeserializationSchema
+
+
Deserializes a LogRecord into a JSON String.
+
+
deserialize(LogRecord) - Method in class org.apache.fluss.flink.source.deserializer.RowDataDeserializationSchema
+
+
Deserializes a LogRecord into a Flink RowData object.
+
+
deserializeComputedColumn(Map<String, String>, int, Schema.Builder) - Static method in class org.apache.fluss.flink.utils.CatalogPropertiesUtils
+
 
+
deserializeOptions(Map<String, String>) - Static method in class org.apache.fluss.flink.utils.CatalogPropertiesUtils
+
 
+
DeserializerInitContextImpl - Class in org.apache.fluss.flink.source.deserializer
+
+
Contextual information that can be used during initialization of FlussDeserializationSchema.
+
+
DeserializerInitContextImpl(MetricGroup, UserCodeClassLoader, RowType) - Constructor for class org.apache.fluss.flink.source.deserializer.DeserializerInitContextImpl
+
 
+
deserializeWatermark(Map<String, String>, Schema.Builder) - Static method in class org.apache.fluss.flink.utils.CatalogPropertiesUtils
+
 
+
done() - Method in class org.apache.fluss.client.write.WriteBatch.RequestFuture
+
+
Mark this request as complete and unblock any threads waiting on its completion.
+
+
downloadFile(Path, FsPath) - Method in class org.apache.fluss.client.table.scanner.RemoteFileDownloader
+
+
Copies the file from a remote file path to the given target file path, returns the number of + downloaded bytes.
+
+
downloadFileAsync(FsPathAndFileName, Path) - Method in class org.apache.fluss.client.table.scanner.RemoteFileDownloader
+
+
Downloads the file from the given remote file path to the target directory asynchronously, + returns a Future object of the number of downloaded bytes.
+
+
downloadThreadPool - Variable in class org.apache.fluss.client.table.scanner.RemoteFileDownloader
+
 
+
drain(Cluster, Set<Integer>, int) - Method in class org.apache.fluss.client.write.RecordAccumulator
+
+
Drain all the data for the given nodes and collate them into a list of batches that will fit + within the specified size on a per-node basis.
+
+
DropAclProcedure - Class in org.apache.fluss.flink.procedure
+
+
Procedure to drop acl.
+
+
DropAclProcedure() - Constructor for class org.apache.fluss.flink.procedure.DropAclProcedure
+
 
+
dropAcls(Collection<AclBindingFilter>) - Method in interface org.apache.fluss.client.admin.Admin
+
+
Removes multiple ACL entries in a single atomic operation.
+
+
dropAcls(Collection<AclBindingFilter>) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
DropAclsResult - Class in org.apache.fluss.client.admin
+
+
Represents the result of a drop ACLs operation.
+
+
DropAclsResult.FilterResult - Class in org.apache.fluss.client.admin
+
+
A class containing either the deleted ACL binding or an exception if the delete failed.
+
+
DropAclsResult.FilterResults - Class in org.apache.fluss.client.admin
+
+
A class containing the results of the delete ACLs operation.
+
+
dropDatabase(String, boolean, boolean) - Method in interface org.apache.fluss.client.admin.Admin
+
+
Drop the database with the given name asynchronously.
+
+
dropDatabase(String, boolean, boolean) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
dropDatabase(String, boolean, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
dropFunction(ObjectPath, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
dropPartition(ObjectPath, CatalogPartitionSpec, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
dropPartition(TablePath, PartitionSpec, boolean) - Method in interface org.apache.fluss.client.admin.Admin
+
+
Drop a partition from a partitioned table.
+
+
dropPartition(TablePath, PartitionSpec, boolean) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
dropTable(ObjectPath, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
dropTable(TablePath, boolean) - Method in interface org.apache.fluss.client.admin.Admin
+
+
Drop the table with the given table path asynchronously.
+
+
dropTable(TablePath, boolean) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
DynamicPartitionCreator - Class in org.apache.fluss.client.write
+
+
A creator to create partition when dynamic partition create enable for table.
+
+
DynamicPartitionCreator(MetadataUpdater, Admin, boolean, Consumer<Throwable>) - Constructor for class org.apache.fluss.client.write.DynamicPartitionCreator
+
 
+
DynamicWriteBatchSizeEstimator - Class in org.apache.fluss.client.write
+
+
An estimator to estimate the buffer usage of a writeBatch.
+
+
DynamicWriteBatchSizeEstimator(boolean, int, int) - Constructor for class org.apache.fluss.client.write.DynamicWriteBatchSizeEstimator
+
 
+
+A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-5.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-5.html new file mode 100644 index 0000000000..7cff70d6a2 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-5.html @@ -0,0 +1,236 @@ + + + + + +E-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages + + +

E

+
+
earliest() - Static method in interface org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer
+
+
Get an OffsetsInitializer which initializes the offsets to the earliest available + offsets of each bucket.
+
+
EARLIEST - org.apache.fluss.flink.FlinkConnectorOptions.ScanStartupMode
+
 
+
EARLIEST_OFFSET - Static variable in interface org.apache.fluss.client.table.scanner.log.LogScanner
+
+
The earliest offset to fetch from.
+
+
earliestOffsets(String, Collection<Integer>) - Method in class org.apache.fluss.flink.source.enumerator.initializer.BucketOffsetsRetrieverImpl
+
 
+
earliestOffsets(String, Collection<Integer>) - Method in interface org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer.BucketOffsetsRetriever
+
 
+
EarliestSpec() - Constructor for class org.apache.fluss.client.admin.OffsetSpec.EarliestSpec
+
 
+
emitRecord(RecordAndPos, SourceOutput<OUT>, SourceSplitState) - Method in class org.apache.fluss.flink.source.emitter.FlinkRecordEmitter
+
 
+
emitRecord(SourceSplitState, SourceOutput<OUT>, RecordAndPos) - Method in class org.apache.fluss.flink.lake.LakeRecordRecordEmitter
+
 
+
emitRecord(TableBucketWriteResult<WriteResult>, SourceOutput<TableBucketWriteResult<WriteResult>>, TieringSplitState) - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultEmitter
+
 
+
EMPTY - Static variable in class org.apache.fluss.client.table.scanner.log.ScanRecords
+
 
+
emptyRecords(FlinkSourceReaderMetrics) - Static method in class org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds
+
 
+
equals(Object) - Method in class org.apache.fluss.client.lookup.LookupResult
+
 
+
equals(Object) - Method in class org.apache.fluss.client.table.scanner.ScanRecord
+
 
+
equals(Object) - Method in class org.apache.fluss.flink.row.RowWithOp
+
+
Indicates whether some other object is "equal to" this one.
+
+
equals(Object) - Method in class org.apache.fluss.flink.source.event.PartitionBucketsUnsubscribedEvent
+
 
+
equals(Object) - Method in class org.apache.fluss.flink.source.reader.RecordAndPos
+
 
+
equals(Object) - Method in class org.apache.fluss.flink.source.split.HybridSnapshotLogSplit
+
 
+
equals(Object) - Method in class org.apache.fluss.flink.source.split.LogSplit
+
 
+
equals(Object) - Method in class org.apache.fluss.flink.source.split.SnapshotSplit
+
 
+
equals(Object) - Method in class org.apache.fluss.flink.source.split.SourceSplitBase
+
 
+
equals(Object) - Method in class org.apache.fluss.flink.source.state.SourceEnumeratorState
+
 
+
equals(Object) - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
+
 
+
equals(Object) - Method in class org.apache.fluss.flink.tiering.source.split.TieringLogSplit
+
 
+
equals(Object) - Method in class org.apache.fluss.flink.tiering.source.split.TieringSnapshotSplit
+
 
+
equals(Object) - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
+
 
+
equals(Object) - Method in class org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorState
+
 
+
equals(Object) - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
+
 
+
equalValue - Variable in class org.apache.fluss.flink.utils.PushdownUtils.FieldEqual
+
 
+
estimatedSizeInBytes() - Method in class org.apache.fluss.client.write.ArrowLogWriteBatch
+
 
+
estimatedSizeInBytes() - Method in class org.apache.fluss.client.write.IndexedLogWriteBatch
+
 
+
estimatedSizeInBytes() - Method in class org.apache.fluss.client.write.KvWriteBatch
+
 
+
estimatedSizeInBytes() - Method in class org.apache.fluss.client.write.WriteBatch
+
+
Get an estimate of the number of bytes written to the underlying buffer.
+
+
exception() - Method in class org.apache.fluss.client.admin.DropAclsResult.FilterResult
+
+
Return an exception if the ACL delete was not successful or null if it was.
+
+
executeDeletion() - Method in class org.apache.fluss.flink.sink.FlinkTableSink
+
 
+
extractFieldEquals(List<ResolvedExpression>, Map<Integer, LogicalType>, List<ResolvedExpression>, List<ResolvedExpression>, PushdownUtils.ValueConversion) - Static method in class org.apache.fluss.flink.utils.PushdownUtils
+
+
Extract field equality information from expressions.
+
+
extractLakeCatalogProperties(Configuration) - Static method in class org.apache.fluss.flink.utils.DataLakeUtils
+
 
+
+A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-6.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-6.html new file mode 100644 index 0000000000..5f272d66d3 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-6.html @@ -0,0 +1,548 @@ + + + + + +F-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages + + +

F

+
+
factoryIdentifier() - Method in class org.apache.fluss.flink.catalog.FlinkCatalogFactory
+
 
+
factoryIdentifier() - Method in class org.apache.fluss.flink.catalog.FlinkTableFactory
+
 
+
FailedTieringEvent - Class in org.apache.fluss.flink.tiering.event
+
+
SourceEvent used to represent a Fluss table is failed during tiering.
+
+
FailedTieringEvent(long, String) - Constructor for class org.apache.fluss.flink.tiering.event.FailedTieringEvent
+
 
+
failReason() - Method in class org.apache.fluss.flink.tiering.event.FailedTieringEvent
+
 
+
fetch() - Method in class org.apache.fluss.flink.source.reader.FlinkSourceSplitReader
+
 
+
fetch() - Method in class org.apache.fluss.flink.tiering.source.TieringSplitReader
+
 
+
fetchRequestCount() - Method in class org.apache.fluss.client.metrics.ScannerMetricGroup
+
 
+
fieldCount - Variable in class org.apache.fluss.client.table.writer.AbstractTableWriter
+
 
+
FieldEqual(int, Object) - Constructor for class org.apache.fluss.flink.utils.PushdownUtils.FieldEqual
+
 
+
fieldIndex - Variable in class org.apache.fluss.flink.utils.PushdownUtils.FieldEqual
+
 
+
finishedSplits() - Method in class org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds
+
 
+
FinishedTieringEvent - Class in org.apache.fluss.flink.tiering.event
+
+
SourceEvent used to represent a Fluss table has been tiered finished.
+
+
FinishedTieringEvent(long) - Constructor for class org.apache.fluss.flink.tiering.event.FinishedTieringEvent
+
 
+
FLINK_INTERNAL_VALUE - org.apache.fluss.flink.utils.PushdownUtils.ValueConversion
+
 
+
FlinkAsFlussRow - Class in org.apache.fluss.flink.row
+
+
Wraps a Flink RowData as a Fluss InternalRow.
+
+
FlinkAsFlussRow() - Constructor for class org.apache.fluss.flink.row.FlinkAsFlussRow
+
 
+
FlinkAsyncLookupFunction - Class in org.apache.fluss.flink.source.lookup
+
+
A flink async lookup function for fluss.
+
+
FlinkAsyncLookupFunction(Configuration, TablePath, RowType, int, LookupNormalizer, int[]) - Constructor for class org.apache.fluss.flink.source.lookup.FlinkAsyncLookupFunction
+
 
+
FlinkCatalog - Class in org.apache.fluss.flink.catalog
+
+
A Flink Catalog for fluss.
+
+
FlinkCatalog(String, String, String, ClassLoader, Map<String, String>) - Constructor for class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
FlinkCatalogFactory - Class in org.apache.fluss.flink.catalog
+
+
Factory for FlinkCatalog.
+
+
FlinkCatalogFactory() - Constructor for class org.apache.fluss.flink.catalog.FlinkCatalogFactory
+
 
+
FlinkCatalogOptions - Class in org.apache.fluss.flink.catalog
+
+
Options for flink catalog.
+
+
FlinkConnectorOptions - Class in org.apache.fluss.flink
+
+
Options for flink connector.
+
+
FlinkConnectorOptions() - Constructor for class org.apache.fluss.flink.FlinkConnectorOptions
+
 
+
FlinkConnectorOptions.ScanStartupMode - Enum in org.apache.fluss.flink
+
+
Startup mode for the fluss scanner, see FlinkConnectorOptions.SCAN_STARTUP_MODE.
+
+
FlinkConnectorOptionsUtils - Class in org.apache.fluss.flink.utils
+
+
Utility class for FlinkConnectorOptions.
+
+
FlinkConnectorOptionsUtils() - Constructor for class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils
+
 
+
FlinkConnectorOptionsUtils.StartupOptions - Class in org.apache.fluss.flink.utils
+
+
Fluss startup options.
+
+
FlinkConversions - Class in org.apache.fluss.flink.utils
+
+
Utils for conversion between Flink and Fluss.
+
+
FlinkCounter - Class in org.apache.fluss.flink.metrics
+
+
An implementation of Flink's Counter which wraps Fluss's Counter.
+
+
FlinkCounter(Counter) - Constructor for class org.apache.fluss.flink.metrics.FlinkCounter
+
 
+
FlinkGauge<T> - Class in org.apache.fluss.flink.metrics
+
+
An implementation of Flink's Gauge which wraps Fluss's Gauge.
+
+
FlinkGauge(Gauge<T>) - Constructor for class org.apache.fluss.flink.metrics.FlinkGauge
+
 
+
FlinkHistogram - Class in org.apache.fluss.flink.metrics
+
+
An implementation of Flink's Histogram which wraps Fluss's Histogram.
+
+
FlinkHistogram(Histogram) - Constructor for class org.apache.fluss.flink.metrics.FlinkHistogram
+
 
+
FlinkLookupFunction - Class in org.apache.fluss.flink.source.lookup
+
+
A flink lookup function for fluss.
+
+
FlinkLookupFunction(Configuration, TablePath, RowType, int, LookupNormalizer, int[]) - Constructor for class org.apache.fluss.flink.source.lookup.FlinkLookupFunction
+
 
+
FlinkMeter - Class in org.apache.fluss.flink.metrics
+
+
An implementation of Flink's Meter which wraps Fluss's Meter.
+
+
FlinkMeter(Meter) - Constructor for class org.apache.fluss.flink.metrics.FlinkMeter
+
 
+
flinkMetricRegistry - Variable in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
+
 
+
FlinkMetricRegistry - Class in org.apache.fluss.flink.metrics
+
+
An implementation of MetricRegistry which registers all metrics into Flink's metric + system.
+
+
FlinkMetricRegistry(MetricGroup) - Constructor for class org.apache.fluss.flink.metrics.FlinkMetricRegistry
+
 
+
FlinkMetricRegistry(MetricGroup, Set<String>) - Constructor for class org.apache.fluss.flink.metrics.FlinkMetricRegistry
+
 
+
FlinkRecordEmitter<OUT> - Class in org.apache.fluss.flink.source.emitter
+
+
The RecordEmitter implementation for FlinkSourceReader.
+
+
FlinkRecordEmitter(FlussDeserializationSchema<OUT>) - Constructor for class org.apache.fluss.flink.source.emitter.FlinkRecordEmitter
+
 
+
FlinkRecordsWithSplitIds - Class in org.apache.fluss.flink.source.reader
+
+
An implementation of RecordsWithSplitIds which contains records from multiple splits.
+
+
FlinkRecordsWithSplitIds(String, TableBucket, CloseableIterator<RecordAndPos>, FlinkSourceReaderMetrics) - Constructor for class org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds
+
 
+
FlinkRecordsWithSplitIds(Map<String, CloseableIterator<RecordAndPos>>, Iterator<String>, Iterator<TableBucket>, Set<String>, FlinkSourceReaderMetrics) - Constructor for class org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds
+
 
+
FlinkRecordsWithSplitIds(Set<String>, FlinkSourceReaderMetrics) - Constructor for class org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds
+
 
+
FlinkRowDataChannelComputer<InputT> - Class in org.apache.fluss.flink.sink
+
+
ChannelComputer for flink RowData.
+
+
FlinkRowDataChannelComputer(RowType, List<String>, List<String>, DataLakeFormat, int, FlussSerializationSchema<InputT>) - Constructor for class org.apache.fluss.flink.sink.FlinkRowDataChannelComputer
+
 
+
FlinkRowToFlussRowConverter - Class in org.apache.fluss.flink.utils
+
+
Deprecated. +
This class is deprecated and will be removed in future versions. It is recommended to + use the FlinkAsFlussRow to wraps Flink's RowData as Fluss's InternalRow.
+
+
+
FlinkRowToFlussRowConverter.FlussSerializationConverter - Interface in org.apache.fluss.flink.utils
+
+
Deprecated.
+
Runtime converter to convert field in Flink's RowData to Fluss's InternalRow + type object.
+
+
FlinkSinkWriter<InputT> - Class in org.apache.fluss.flink.sink.writer
+
+
Base class for Flink SinkWriter implementations in Fluss.
+
+
FlinkSinkWriter(TablePath, Configuration, RowType, int[], MailboxExecutor, FlussSerializationSchema<InputT>) - Constructor for class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
+
 
+
FlinkSinkWriter(TablePath, Configuration, RowType, MailboxExecutor, FlussSerializationSchema<InputT>) - Constructor for class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
+
 
+
FlinkSource<OUT> - Class in org.apache.fluss.flink.source
+
+
Flink source for Fluss.
+
+
FlinkSource(Configuration, TablePath, boolean, boolean, RowType, int[], OffsetsInitializer, long, FlussDeserializationSchema<OUT>, boolean, List<PushdownUtils.FieldEqual>) - Constructor for class org.apache.fluss.flink.source.FlinkSource
+
 
+
FlinkSource(Configuration, TablePath, boolean, boolean, RowType, int[], OffsetsInitializer, long, FlussDeserializationSchema<OUT>, boolean, List<PushdownUtils.FieldEqual>, LakeSource<LakeSplit>) - Constructor for class org.apache.fluss.flink.source.FlinkSource
+
 
+
FlinkSourceEnumerator - Class in org.apache.fluss.flink.source.enumerator
+
+
An implementation of SplitEnumerator for the data of Fluss.
+
+
FlinkSourceEnumerator(TablePath, Configuration, boolean, boolean, SplitEnumeratorContext<SourceSplitBase>, Set<TableBucket>, Map<Long, String>, List<SourceSplitBase>, OffsetsInitializer, long, boolean, List<PushdownUtils.FieldEqual>, LakeSource<LakeSplit>) - Constructor for class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
+
 
+
FlinkSourceEnumerator(TablePath, Configuration, boolean, boolean, SplitEnumeratorContext<SourceSplitBase>, OffsetsInitializer, long, boolean, List<PushdownUtils.FieldEqual>) - Constructor for class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
+
 
+
FlinkSourceEnumerator(TablePath, Configuration, boolean, boolean, SplitEnumeratorContext<SourceSplitBase>, OffsetsInitializer, long, boolean, List<PushdownUtils.FieldEqual>, LakeSource<LakeSplit>) - Constructor for class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
+
 
+
FlinkSourceFetcherManager - Class in org.apache.fluss.flink.source.reader.fetcher
+
+
The SplitFetcherManager for Fluss source.
+
+
FlinkSourceFetcherManager(FutureCompletingBlockingQueue<RecordsWithSplitIds<RecordAndPos>>, Supplier<SplitReader<RecordAndPos, SourceSplitBase>>, Consumer<Collection<String>>) - Constructor for class org.apache.fluss.flink.source.reader.fetcher.FlinkSourceFetcherManager
+
+
Creates a new SplitFetcherManager with a single I/O threads.
+
+
FlinkSourceReader<OUT> - Class in org.apache.fluss.flink.source.reader
+
+
The source reader for Fluss.
+
+
FlinkSourceReader(FutureCompletingBlockingQueue<RecordsWithSplitIds<RecordAndPos>>, Configuration, TablePath, RowType, SourceReaderContext, int[], FlinkSourceReaderMetrics, FlinkRecordEmitter<OUT>, LakeSource<LakeSplit>) - Constructor for class org.apache.fluss.flink.source.reader.FlinkSourceReader
+
 
+
FlinkSourceReaderMetrics - Class in org.apache.fluss.flink.source.metrics
+
+
A collection class for handling metrics in FlinkSourceReader of Fluss.
+
+
FlinkSourceReaderMetrics(SourceReaderMetricGroup) - Constructor for class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
+
 
+
FlinkSourceSplitReader - Class in org.apache.fluss.flink.source.reader
+
+
An implementation of SplitReader for reading splits into RecordAndPos.
+
+
FlinkSourceSplitReader(Configuration, TablePath, RowType, int[], FlinkSourceReaderMetrics, LakeSource<LakeSplit>) - Constructor for class org.apache.fluss.flink.source.reader.FlinkSourceSplitReader
+
 
+
FlinkStreamPartitioner<T> - Class in org.apache.fluss.flink.sink
+
+
A StreamPartitioner which wraps a ChannelComputer.
+
+
FlinkStreamPartitioner(ChannelComputer<T>) - Constructor for class org.apache.fluss.flink.sink.FlinkStreamPartitioner
+
 
+
FlinkTableFactory - Class in org.apache.fluss.flink.catalog
+
+
Factory to create table source and table sink for Fluss.
+
+
FlinkTableFactory() - Constructor for class org.apache.fluss.flink.catalog.FlinkTableFactory
+
 
+
FlinkTableSink - Class in org.apache.fluss.flink.sink
+
+
A Flink DynamicTableSink.
+
+
FlinkTableSink(TablePath, Configuration, RowType, int[], List<String>, boolean, MergeEngineType, DataLakeFormat, boolean, int, List<String>, boolean) - Constructor for class org.apache.fluss.flink.sink.FlinkTableSink
+
 
+
FlinkTableSource - Class in org.apache.fluss.flink.source
+
+
Flink table source to scan Fluss data.
+
+
FlinkTableSource(TablePath, Configuration, RowType, int[], int[], int[], boolean, FlinkConnectorOptionsUtils.StartupOptions, int, boolean, LookupCache, long, boolean, MergeEngineType, Map<String, String>) - Constructor for class org.apache.fluss.flink.source.FlinkTableSource
+
 
+
FlinkUtils - Class in org.apache.fluss.flink.utils
+
+
Utils for Flink classes.
+
+
FlinkUtils() - Constructor for class org.apache.fluss.flink.utils.FlinkUtils
+
 
+
flush() - Method in class org.apache.fluss.client.table.writer.AbstractTableWriter
+
+
Flush data written that have not yet been sent to the server, forcing the client to send the + requests to server and blocks on the completion of the requests associated with these + records.
+
+
flush() - Method in interface org.apache.fluss.client.table.writer.TableWriter
+
+
Flush data written that have not yet been sent to the server, forcing the client to send the + requests to server and blocks on the completion of the requests associated with these + records.
+
+
flush() - Method in class org.apache.fluss.client.write.WriterClient
+
+
Invoking this method makes all buffered records immediately available to send (even if + linger.ms is greater than 0) and blocks on the completion of the requests associated + with these records.
+
+
flush(boolean) - Method in class org.apache.fluss.flink.sink.writer.AppendSinkWriter
+
 
+
flush(boolean) - Method in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
+
 
+
flush(boolean) - Method in class org.apache.fluss.flink.sink.writer.UpsertSinkWriter
+
 
+
FLUSS_GROUP_NAME - Static variable in class org.apache.fluss.flink.metrics.FlinkMetricRegistry
+
 
+
FLUSS_INTERNAL_VALUE - org.apache.fluss.flink.utils.PushdownUtils.ValueConversion
+
 
+
FLUSS_METRIC_GROUP - Static variable in class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
+
 
+
FlussAdmin - Class in org.apache.fluss.client.admin
+
+
The default implementation of Admin.
+
+
FlussAdmin(RpcClient, MetadataUpdater) - Constructor for class org.apache.fluss.client.admin.FlussAdmin
+
 
+
FlussConnection - Class in org.apache.fluss.client
+
+
A connection to Fluss cluster, and holds the client session resources.
+
+
FlussDeserializationSchema<T> - Interface in org.apache.fluss.flink.source.deserializer
+
+
Interface for deserialization schema used to deserialize LogRecord objects into specific + data types.
+
+
FlussDeserializationSchema.InitializationContext - Interface in org.apache.fluss.flink.source.deserializer
+
+
A contextual information provided for FlussDeserializationSchema.open(InitializationContext) method.
+
+
FlussRowToFlinkRowConverter - Class in org.apache.fluss.flink.utils
+
+
A converter to convert Fluss's InternalRow to Flink's RowData.
+
+
FlussRowToFlinkRowConverter(RowType) - Constructor for class org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter
+
 
+
FlussRowToFlinkRowConverter.FlussDeserializationConverter - Interface in org.apache.fluss.flink.utils
+
+
Runtime converter to convert field in Fluss's InternalRow to Flink's RowData + type object.
+
+
FlussRowToJsonConverters - Class in org.apache.fluss.flink.utils
+
+
A converter to convert Fluss's InternalRow to JsonNode.
+
+
FlussRowToJsonConverters(TimestampFormat) - Constructor for class org.apache.fluss.flink.utils.FlussRowToJsonConverters
+
 
+
FlussRowToJsonConverters.FlussRowToJsonConverter - Interface in org.apache.fluss.flink.utils
+
+
Runtime converter that converts objects of Fluss data structures to corresponding JsonNodes.
+
+
FlussSerializationSchema<T> - Interface in org.apache.fluss.flink.sink.serializer
+
+
A serialization schema for Fluss.
+
+
FlussSerializationSchema.InitializationContext - Interface in org.apache.fluss.flink.sink.serializer
+
+
A contextual information provided for FlussSerializationSchema.open(InitializationContext) method.
+
+
FlussSink<InputT> - Class in org.apache.fluss.flink.sink
+
+
FlussSink is a specialized Flink sink for writing data to Fluss.
+
+
FlussSinkBuilder<InputT> - Class in org.apache.fluss.flink.sink
+
+
Builder for creating and configuring Fluss sink connectors for Apache Flink.
+
+
FlussSinkBuilder() - Constructor for class org.apache.fluss.flink.sink.FlussSinkBuilder
+
 
+
FlussSource<OUT> - Class in org.apache.fluss.flink.source
+
+
A Flink DataStream source implementation for reading data from Fluss tables.
+
+
FlussSourceBuilder<OUT> - Class in org.apache.fluss.flink.source
+
+
Builder class for creating FlussSource instances.
+
+
FlussSourceBuilder() - Constructor for class org.apache.fluss.flink.source.FlussSourceBuilder
+
 
+
FlussSourceEnumeratorStateSerializer - Class in org.apache.fluss.flink.source.state
+
+
A serializer for SourceEnumeratorState.
+
+
FlussSourceEnumeratorStateSerializer(LakeSource<LakeSplit>) - Constructor for class org.apache.fluss.flink.source.state.FlussSourceEnumeratorStateSerializer
+
 
+
FlussTable - Class in org.apache.fluss.client.table
+
+
The base impl of Table.
+
+
FlussTable(FlussConnection, TablePath, TableInfo) - Constructor for class org.apache.fluss.client.table.FlussTable
+
 
+
FlussTableLakeSnapshotCommitter - Class in org.apache.fluss.flink.tiering.committer
+
+
Committer to commit FlussTableLakeSnapshot of lake to Fluss.
+
+
FlussTableLakeSnapshotCommitter(Configuration) - Constructor for class org.apache.fluss.flink.tiering.committer.FlussTableLakeSnapshotCommitter
+
 
+
forArrowAppend(PhysicalTablePath, InternalRow, byte[]) - Static method in class org.apache.fluss.client.write.WriteRecord
+
+
Creates a write record for append operation for Arrow format.
+
+
forceClose() - Method in class org.apache.fluss.client.write.Sender
+
+
Closes the sender without sending out any pending messages.
+
+
forDelete(PhysicalTablePath, byte[], byte[], int[]) - Static method in class org.apache.fluss.client.write.WriteRecord
+
+
Create a write record for delete operation and partial-delete update.
+
+
forIndexedAppend(PhysicalTablePath, IndexedRow, byte[]) - Static method in class org.apache.fluss.client.write.WriteRecord
+
+
Create a write record for append operation for indexed format.
+
+
forUpsert(PhysicalTablePath, BinaryRow, byte[], byte[], int[]) - Static method in class org.apache.fluss.client.write.WriteRecord
+
+
Create a write record for upsert operation and partial-upsert operation.
+
+
fromFlinkDecimal(DecimalData) - Static method in class org.apache.fluss.flink.row.FlinkAsFlussRow
+
 
+
full() - Static method in interface org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer
+
+
Get an OffsetsInitializer which performs a full snapshot on the table upon first + startup, and continue to read log with the offset to the snapshot.
+
+
FULL - org.apache.fluss.flink.FlinkConnectorOptions.ScanStartupMode
+
 
+
functionExists(ObjectPath) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
future() - Method in class org.apache.fluss.client.lookup.AbstractLookupQuery
+
 
+
future() - Method in class org.apache.fluss.client.lookup.LookupQuery
+
 
+
future() - Method in class org.apache.fluss.client.lookup.PrefixLookupQuery
+
 
+
+A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-7.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-7.html new file mode 100644 index 0000000000..da9f8f0acc --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-7.html @@ -0,0 +1,659 @@ + + + + + +G-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages + + +

G

+
+
generateHybridLakeFlussSplits() - Method in class org.apache.fluss.flink.lake.LakeSplitGenerator
+
+
Return A list of hybrid lake snapshot LakeSnapshotSplit, LakeSnapshotAndFlussLogSplit and the corresponding Fluss LogSplit based on the lake + snapshot.
+
+
generateTableSplits(TablePath) - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplitGenerator
+
 
+
getAdmin() - Method in interface org.apache.fluss.client.Connection
+
+
Retrieve a new Admin client to administer a Fluss cluster.
+
+
getAdmin() - Method in class org.apache.fluss.client.FlussConnection
+
 
+
getArity() - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
+
 
+
getArity() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
+
 
+
getAssignedBuckets() - Method in class org.apache.fluss.flink.source.state.SourceEnumeratorState
+
 
+
getAssignedPartitions() - Method in class org.apache.fluss.flink.source.state.SourceEnumeratorState
+
 
+
getBinary(int, int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
+
 
+
getBoolean(int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
+
 
+
getBoundedness() - Method in class org.apache.fluss.flink.source.FlinkSource
+
 
+
getBoundedness() - Method in class org.apache.fluss.flink.tiering.source.TieringSource
+
 
+
getBoundedSplitScanner(SourceSplitBase) - Method in class org.apache.fluss.flink.lake.LakeSplitReaderGenerator
+
 
+
getBucketIds() - Method in class org.apache.fluss.client.metadata.KvSnapshots
+
 
+
getBucketKey() - Method in class org.apache.fluss.client.write.WriteRecord
+
 
+
getBucketKeyIndexes() - Method in class org.apache.fluss.flink.source.FlinkTableSource
+
 
+
getBucketKeyIndexes(ReadableConfig, RowType) - Static method in class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils
+
 
+
getBucketKeys() - Method in class org.apache.fluss.flink.sink.FlinkTableSink
+
 
+
getBucketKeys(ReadableConfig) - Static method in class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils
+
 
+
getBucketLocation(TableBucket) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
+
 
+
getBucketOffsets(String, Collection<Integer>, OffsetsInitializer.BucketOffsetsRetriever) - Method in class org.apache.fluss.flink.source.enumerator.initializer.LatestOffsetsInitializer
+
 
+
getBucketOffsets(String, Collection<Integer>, OffsetsInitializer.BucketOffsetsRetriever) - Method in class org.apache.fluss.flink.source.enumerator.initializer.NoStoppingOffsetsInitializer
+
 
+
getBucketOffsets(String, Collection<Integer>, OffsetsInitializer.BucketOffsetsRetriever) - Method in interface org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer
+
+
Get the initial offsets for the given fluss buckets.
+
+
getBucketOffsets(String, Collection<Integer>, OffsetsInitializer.BucketOffsetsRetriever) - Method in class org.apache.fluss.flink.source.enumerator.initializer.SnapshotOffsetsInitializer
+
+
For table with primary key.
+
+
getBucketOffsets(String, Collection<Integer>, OffsetsInitializer.BucketOffsetsRetriever) - Method in class org.apache.fluss.flink.source.enumerator.initializer.TimestampOffsetsInitializer
+
 
+
getByte(int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
+
 
+
getBytes(int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
+
 
+
getCache() - Method in class org.apache.fluss.flink.source.FlinkTableSource
+
 
+
getChangelogMode() - Method in class org.apache.fluss.flink.source.FlinkTableSource
+
 
+
getChangelogMode(ChangelogMode) - Method in class org.apache.fluss.flink.sink.FlinkTableSink
+
 
+
getChangeType() - Method in class org.apache.fluss.client.table.scanner.ScanRecord
+
 
+
getChar(int, int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
+
 
+
getClientMetricGroup() - Method in class org.apache.fluss.client.FlussConnection
+
 
+
getCluster() - Method in class org.apache.fluss.client.metadata.MetadataUpdater
+
 
+
getConfiguration() - Method in interface org.apache.fluss.client.Connection
+
+
Retrieve the configuration used to create this connection.
+
+
getConfiguration() - Method in class org.apache.fluss.client.FlussConnection
+
 
+
getCoordinatorServer() - Method in class org.apache.fluss.client.metadata.MetadataUpdater
+
 
+
getCount() - Method in class org.apache.fluss.flink.metrics.FlinkCounter
+
 
+
getCount() - Method in class org.apache.fluss.flink.metrics.FlinkHistogram
+
 
+
getCount() - Method in class org.apache.fluss.flink.metrics.FlinkMeter
+
 
+
getCurrentLakeSplitIndex() - Method in class org.apache.fluss.flink.lake.reader.IndexedLakeSplitRecordIterator
+
 
+
getCurrentLakeSplitIndex() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
+
 
+
getCurrentSplitIndex() - Method in class org.apache.fluss.flink.source.reader.RecordAndPos
+
 
+
getDatabase(String) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
getDatabaseInfo(String) - Method in interface org.apache.fluss.client.admin.Admin
+
+
Get the database with the given database name asynchronously.
+
+
getDatabaseInfo(String) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
getDatalakeFormat(ReadableConfig) - Static method in class org.apache.fluss.flink.utils.DataLakeUtils
+
 
+
getDecimal(int, int, int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
+
 
+
getDescription() - Method in enum org.apache.fluss.flink.FlinkConnectorOptions.ScanStartupMode
+
 
+
getDouble(int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
+
 
+
getDownstreamSubtaskStateMapper() - Method in class org.apache.fluss.flink.sink.FlinkStreamPartitioner
+
 
+
getEnumeratorCheckpointSerializer() - Method in class org.apache.fluss.flink.source.FlinkSource
+
 
+
getEnumeratorCheckpointSerializer() - Method in class org.apache.fluss.flink.tiering.source.TieringSource
+
 
+
getEstimatedBatchSize(PhysicalTablePath) - Method in class org.apache.fluss.client.write.DynamicWriteBatchSizeEstimator
+
 
+
getEstimatedSizeInBytes() - Method in class org.apache.fluss.client.write.WriteRecord
+
+
Get the estimated size in bytes of the record with batch header.
+
+
getFactory() - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
getFieldCount() - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
+
 
+
getFileLogRecords(int) - Method in class org.apache.fluss.client.table.scanner.log.RemoteLogDownloadFuture
+
 
+
getFloat(int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
+
 
+
getFlussMetric(String) - Method in class org.apache.fluss.flink.metrics.FlinkMetricRegistry
+
+
Exposes the metrics of Fluss metics group for flink.
+
+
getFsPathAndFileName(FsPath, RemoteLogSegment) - Static method in class org.apache.fluss.client.table.scanner.log.RemoteLogDownloader
+
 
+
getFunction(ObjectPath) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
getFutures() - Method in class org.apache.fluss.client.admin.CreateAclsResult
+
+
Gets the map of ACL bindings to their associated futures.
+
+
getGroupName(CharacterFilter) - Method in class org.apache.fluss.client.metrics.ScannerMetricGroup
+
 
+
getGroupName(CharacterFilter) - Method in class org.apache.fluss.client.metrics.WriterMetricGroup
+
 
+
getHost(String) - Static method in class org.apache.fluss.client.utils.ClientUtils
+
+
Extracts the hostname from a "host:port" address string.
+
+
getInt(int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
+
 
+
getKey() - Method in class org.apache.fluss.client.write.WriteRecord
+
 
+
getKvSnapshotMetadata(TableBucket, long) - Method in interface org.apache.fluss.client.admin.Admin
+
+
Get the kv snapshot metadata of the given kv snapshot asynchronously.
+
+
getKvSnapshotMetadata(TableBucket, long) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
getLakeSplit() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotSplit
+
 
+
getLakeSplits() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
+
 
+
getLakeTable(String, String, Configuration) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
getLatestKvSnapshots(TablePath) - Method in interface org.apache.fluss.client.admin.Admin
+
+
Get the latest kv snapshots of the given table asynchronously.
+
+
getLatestKvSnapshots(TablePath) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
getLatestKvSnapshots(TablePath, String) - Method in interface org.apache.fluss.client.admin.Admin
+
+
Get the latest kv snapshots of the given table partition asynchronously.
+
+
getLatestKvSnapshots(TablePath, String) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
getLatestLakeSnapshot(TablePath) - Method in interface org.apache.fluss.client.admin.Admin
+
+
Get table lake snapshot info of the given table asynchronously.
+
+
getLatestLakeSnapshot(TablePath) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
getLocalTimeZone(String) - Static method in class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils
+
 
+
getLogOffset() - Method in class org.apache.fluss.client.metadata.KvSnapshotMetadata
+
 
+
getLogOffset(int) - Method in class org.apache.fluss.client.metadata.KvSnapshots
+
+
Get the log offset to read after the snapshot for this kv tablet (bucket), or empty if there + are no snapshots.
+
+
getLogOffsetOfSnapshot() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSnapshotSplit
+
 
+
getLogStartingOffset() - Method in class org.apache.fluss.flink.source.split.HybridSnapshotLogSplit
+
 
+
getLong(int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
+
 
+
getLookupColumnNames() - Method in class org.apache.fluss.client.lookup.PrefixLookup
+
 
+
getLookupKeyIndexes() - Method in class org.apache.fluss.flink.source.lookup.LookupNormalizer
+
+
Returns the indexes of the normalized lookup keys.
+
+
getLookupRuntimeProvider(LookupTableSource.LookupContext) - Method in class org.apache.fluss.flink.source.FlinkTableSource
+
 
+
getLookupType() - Method in class org.apache.fluss.flink.source.lookup.LookupNormalizer
+
+
Returns the lookup type (primary key lookup, or prefix key lookup) requested from Flink to + Fluss.
+
+
getMetadataUpdater() - Method in class org.apache.fluss.client.FlussConnection
+
 
+
getMetricGroup() - Method in class org.apache.fluss.flink.source.deserializer.DeserializerInitContextImpl
+
 
+
getMetricGroup() - Method in interface org.apache.fluss.flink.source.deserializer.FlussDeserializationSchema.InitializationContext
+
+
Returns the metric group for the parallel subtask of the source that runs this FlussDeserializationSchema.
+
+
getNumberOfSplits() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
+
 
+
getNumberReporters() - Method in class org.apache.fluss.flink.metrics.FlinkMetricRegistry
+
 
+
getOneAvailableTabletServerNode(Cluster) - Static method in class org.apache.fluss.client.utils.MetadataUtils
+
 
+
getOperationType() - Method in class org.apache.fluss.flink.row.RowWithOp
+
+
Returns the operation type associated with this row.
+
+
getOrCreateLookupClient() - Method in class org.apache.fluss.client.FlussConnection
+
 
+
getOrCreateRemoteFileDownloader() - Method in class org.apache.fluss.client.FlussConnection
+
 
+
getOrCreateWriterClient() - Method in class org.apache.fluss.client.FlussConnection
+
 
+
getPartialUpdateColumns() - Method in class org.apache.fluss.client.table.writer.UpsertWrite
+
+
Returns the columns to partial update.
+
+
getPartition(ObjectPath, CatalogPartitionSpec) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
getPartition(InternalRow) - Method in class org.apache.fluss.client.table.getter.PartitionGetter
+
 
+
getPartitionColumnStatistics(ObjectPath, CatalogPartitionSpec) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
getPartitionId() - Method in class org.apache.fluss.client.metadata.KvSnapshots
+
 
+
getPartitionId(PhysicalTablePath) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
+
 
+
getPartitionId(InternalRow, PartitionGetter, TablePath, MetadataUpdater) - Static method in class org.apache.fluss.client.utils.ClientUtils
+
+
Return the id of the partition the row belongs to.
+
+
getPartitionIdOrElseThrow(PhysicalTablePath) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
+
 
+
getPartitionKeyIndexes() - Method in class org.apache.fluss.flink.source.FlinkTableSource
+
 
+
getPartitionName() - Method in class org.apache.fluss.flink.source.split.SourceSplitBase
+
 
+
getPartitionName() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
+
 
+
getPartitionNameById() - Method in class org.apache.fluss.client.metadata.LakeSnapshot
+
 
+
getPartitionStatistics(ObjectPath, CatalogPartitionSpec) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
getPhysicalPath(InternalRow) - Method in class org.apache.fluss.client.table.writer.AbstractTableWriter
+
 
+
getPhysicalTablePath() - Method in class org.apache.fluss.client.write.WriteRecord
+
 
+
getPhysicalTablePathByIds(Collection<Long>, Collection<TablePartition>) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
+
+
Get the table physical paths by table ids and partition ids.
+
+
getPhysicalTablePathsInBatches() - Method in class org.apache.fluss.client.write.RecordAccumulator
+
 
+
getPort(String) - Static method in class org.apache.fluss.client.utils.ClientUtils
+
+
Extracts the port number from a "host:port" address string.
+
+
getPrimaryKeyIndexes() - Method in class org.apache.fluss.flink.source.FlinkTableSource
+
 
+
getProcedure(ObjectPath) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
getProcedure(Admin, ObjectPath) - Static method in class org.apache.fluss.flink.procedure.ProcedureManager
+
 
+
getProducedType() - Method in class org.apache.fluss.flink.source.FlinkSource
+
 
+
getProducedType(RowType) - Method in interface org.apache.fluss.flink.source.deserializer.FlussDeserializationSchema
+
+
Gets the data type (as a TypeInformation) produced by this deserializer.
+
+
getProducedType(RowType) - Method in class org.apache.fluss.flink.source.deserializer.JsonStringDeserializationSchema
+
+
Returns the TypeInformation for the produced String type.
+
+
getProducedType(RowType) - Method in class org.apache.fluss.flink.source.deserializer.RowDataDeserializationSchema
+
+
Returns the TypeInformation for the produced RowData type.
+
+
getProjectedFields() - Method in class org.apache.fluss.client.table.scanner.log.LogScan
+
 
+
getQueueTimeMs() - Method in class org.apache.fluss.client.write.WriteBatch
+
 
+
getRandomTabletServer() - Method in class org.apache.fluss.client.metadata.MetadataUpdater
+
 
+
getRate() - Method in class org.apache.fluss.flink.metrics.FlinkMeter
+
 
+
getRecordCount() - Method in class org.apache.fluss.client.write.WriteBatch
+
 
+
getRecordsToSkip() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
+
 
+
getRecordsToSkip() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotSplit
+
 
+
getRecycleCallback() - Method in class org.apache.fluss.client.table.scanner.log.RemoteLogDownloadFuture
+
 
+
getRemainingHybridLakeFlussSplits() - Method in class org.apache.fluss.flink.source.state.SourceEnumeratorState
+
 
+
getRemovedPartitions() - Method in class org.apache.fluss.flink.source.event.PartitionsRemovedEvent
+
 
+
getRemovedTableBuckets() - Method in class org.apache.fluss.flink.source.event.PartitionBucketsUnsubscribedEvent
+
 
+
getRequestFuture() - Method in class org.apache.fluss.client.write.WriteBatch
+
 
+
getRow() - Method in class org.apache.fluss.client.table.scanner.ScanRecord
+
 
+
getRow() - Method in class org.apache.fluss.client.write.WriteRecord
+
 
+
getRow() - Method in class org.apache.fluss.flink.row.RowWithOp
+
+
Returns the internal row data.
+
+
getRowList() - Method in class org.apache.fluss.client.lookup.LookupResult
+
 
+
getRowSchema() - Method in interface org.apache.fluss.flink.sink.serializer.FlussSerializationSchema.InitializationContext
+
+
Returns the target row schema.
+
+
getRowSchema() - Method in class org.apache.fluss.flink.sink.serializer.SerializerInitContextImpl
+
 
+
getRowSchema() - Method in class org.apache.fluss.flink.source.deserializer.DeserializerInitContextImpl
+
 
+
getRowSchema() - Method in interface org.apache.fluss.flink.source.deserializer.FlussDeserializationSchema.InitializationContext
+
+
Returns the schema of the LogRecord.getRow().
+
+
getRpcClient() - Method in class org.apache.fluss.client.FlussConnection
+
 
+
getScanRuntimeProvider(ScanTableSource.ScanContext) - Method in class org.apache.fluss.flink.source.FlinkTableSource
+
 
+
getSecurityConfigs() - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
getServerNodes() - Method in interface org.apache.fluss.client.admin.Admin
+
+
Get the current server node information.
+
+
getServerNodes() - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
getShort(int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
+
 
+
getSingletonRow() - Method in class org.apache.fluss.client.lookup.LookupResult
+
 
+
getSinkRuntimeProvider(DynamicTableSink.Context) - Method in class org.apache.fluss.flink.sink.FlinkTableSink
+
 
+
getSnapshotFiles() - Method in class org.apache.fluss.client.metadata.KvSnapshotMetadata
+
 
+
getSnapshotId() - Method in class org.apache.fluss.client.metadata.LakeSnapshot
+
 
+
getSnapshotId() - Method in class org.apache.fluss.flink.source.split.SnapshotSplit
+
 
+
getSnapshotId() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSnapshotSplit
+
 
+
getSnapshotId(int) - Method in class org.apache.fluss.client.metadata.KvSnapshots
+
+
Get the latest snapshot id for this kv tablet (bucket), or empty if there are no snapshots.
+
+
getSourceReaderMetricGroup() - Method in class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
+
 
+
getSplitIndex() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotSplit
+
 
+
getSplitOwner(SourceSplitBase) - Method in class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
+
+
Returns the index of the target subtask that a specific split should be assigned to.
+
+
getSplitSerializer() - Method in class org.apache.fluss.flink.source.FlinkSource
+
 
+
getSplitSerializer() - Method in class org.apache.fluss.flink.tiering.source.TieringSource
+
 
+
getStartingOffset() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
+
 
+
getStartingOffset() - Method in class org.apache.fluss.flink.source.split.LogSplit
+
 
+
getStartingOffset() - Method in class org.apache.fluss.flink.tiering.source.split.TieringLogSplit
+
 
+
getStartupOptions(ReadableConfig, ZoneId) - Static method in class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils
+
 
+
getStatistics() - Method in class org.apache.fluss.flink.metrics.FlinkHistogram
+
 
+
getStoppingOffset() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
+
 
+
getStoppingOffset() - Method in class org.apache.fluss.flink.source.split.LogSplit
+
 
+
getStoppingOffset() - Method in class org.apache.fluss.flink.tiering.source.split.TieringLogSplit
+
 
+
getStreamOperatorClass(ClassLoader) - Method in class org.apache.fluss.flink.tiering.committer.TieringCommitOperatorFactory
+
 
+
getString(int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
+
 
+
getTable(ObjectPath) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
getTable(ObjectPath) - Method in class org.apache.fluss.flink.lake.LakeCatalog
+
 
+
getTable(TablePath) - Method in interface org.apache.fluss.client.Connection
+
+
Retrieve a new Table client to operate data in table.
+
+
getTable(TablePath) - Method in class org.apache.fluss.client.FlussConnection
+
 
+
getTableBucket() - Method in class org.apache.fluss.flink.source.split.SourceSplitBase
+
 
+
getTableBucket() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
+
 
+
getTableBucketsOffset() - Method in class org.apache.fluss.client.metadata.LakeSnapshot
+
 
+
getTableColumnStatistics(ObjectPath) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
getTableId() - Method in class org.apache.fluss.client.metadata.KvSnapshots
+
 
+
getTableId() - Method in class org.apache.fluss.flink.tiering.event.FailedTieringEvent
+
 
+
getTableId() - Method in class org.apache.fluss.flink.tiering.event.FinishedTieringEvent
+
 
+
getTableId(TablePath) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
+
 
+
getTableInfo() - Method in class org.apache.fluss.client.table.FlussTable
+
 
+
getTableInfo() - Method in interface org.apache.fluss.client.table.Table
+
+
Get the TableInfo for this table.
+
+
getTableInfo(TablePath) - Method in interface org.apache.fluss.client.admin.Admin
+
+
Get the table with the given table path asynchronously.
+
+
getTableInfo(TablePath) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
getTableInfoOrElseThrow(long) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
+
 
+
getTableInfoOrElseThrow(TablePath) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
+
 
+
getTablePath() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
+
 
+
getTableSchema(TablePath) - Method in interface org.apache.fluss.client.admin.Admin
+
+
Get the latest table schema of the given table asynchronously.
+
+
getTableSchema(TablePath) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
getTableSchema(TablePath, int) - Method in interface org.apache.fluss.client.admin.Admin
+
+
Get the specific table schema of the given table by schema id asynchronously.
+
+
getTableSchema(TablePath, int) - Method in class org.apache.fluss.client.admin.FlussAdmin
+
 
+
getTableStatistics(ObjectPath) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
+
 
+
getTargetColumns() - Method in class org.apache.fluss.client.write.KvWriteBatch
+
 
+
getTargetColumns() - Method in class org.apache.fluss.client.write.WriteRecord
+
 
+
getTimestamp() - Method in class org.apache.fluss.client.admin.OffsetSpec.TimestampSpec
+
 
+
getTimestampLtz(int, int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
+
 
+
getTimestampNtz(int, int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
+
 
+
getTotalFields() - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
+
 
+
getTotalFields() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
+
 
+
getTypeClass() - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
+
 
+
getTypeClass() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
+
 
+
getUserCodeClassLoader() - Method in class org.apache.fluss.flink.source.deserializer.DeserializerInitContextImpl
+
 
+
getUserCodeClassLoader() - Method in interface org.apache.fluss.flink.source.deserializer.FlussDeserializationSchema.InitializationContext
+
+
Gets the UserCodeClassLoader to load classes that are not in system's classpath, + but are part of the jar file of a user job.
+
+
getValue() - Method in class org.apache.fluss.flink.metrics.FlinkGauge
+
 
+
getVersion() - Method in class org.apache.fluss.flink.source.split.SourceSplitSerializer
+
 
+
getVersion() - Method in class org.apache.fluss.flink.source.state.FlussSourceEnumeratorStateSerializer
+
 
+
getVersion() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplitSerializer
+
 
+
getVersion() - Method in class org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorStateSerializer
+
 
+
getVersion() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultSerializer
+
 
+
getWriteFormat() - Method in class org.apache.fluss.client.write.WriteRecord
+
 
+
+A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-8.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-8.html new file mode 100644 index 0000000000..9212594a68 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-8.html @@ -0,0 +1,238 @@ + + + + + +H-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages + + +

H

+
+
handleSourceEvent(int, SourceEvent) - Method in class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
+
 
+
handleSourceEvent(int, SourceEvent) - Method in class org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator
+
 
+
handleSourceEvents(SourceEvent) - Method in class org.apache.fluss.flink.source.reader.FlinkSourceReader
+
 
+
handleSplitRequest(int, String) - Method in class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
+
 
+
handleSplitRequest(int, String) - Method in class org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator
+
 
+
handleSplitsChanges(SplitsChange<SourceSplitBase>) - Method in class org.apache.fluss.flink.source.reader.FlinkSourceSplitReader
+
 
+
handleSplitsChanges(SplitsChange<TieringSplit>) - Method in class org.apache.fluss.flink.tiering.source.TieringSplitReader
+
 
+
hasAvailableFetches() - Method in class org.apache.fluss.client.table.scanner.log.LogFetcher
+
+
Return whether we have any completed fetches that are fetch-able.
+
+
hasBatchSequence() - Method in class org.apache.fluss.client.write.WriteBatch
+
 
+
HashBucketAssigner - Class in org.apache.fluss.client.write
+
+
Hash bucket assigner.
+
+
HashBucketAssigner(int) - Constructor for class org.apache.fluss.client.write.HashBucketAssigner
+
 
+
HashBucketAssigner(int, BucketingFunction) - Constructor for class org.apache.fluss.client.write.HashBucketAssigner
+
 
+
hashCode() - Method in class org.apache.fluss.client.lookup.LookupResult
+
 
+
hashCode() - Method in class org.apache.fluss.client.table.scanner.ScanRecord
+
 
+
hashCode() - Method in class org.apache.fluss.flink.row.RowWithOp
+
+
Returns a hash code value for the object.
+
+
hashCode() - Method in class org.apache.fluss.flink.source.event.PartitionBucketsUnsubscribedEvent
+
 
+
hashCode() - Method in class org.apache.fluss.flink.source.reader.RecordAndPos
+
 
+
hashCode() - Method in class org.apache.fluss.flink.source.split.HybridSnapshotLogSplit
+
 
+
hashCode() - Method in class org.apache.fluss.flink.source.split.LogSplit
+
 
+
hashCode() - Method in class org.apache.fluss.flink.source.split.SnapshotSplit
+
 
+
hashCode() - Method in class org.apache.fluss.flink.source.split.SourceSplitBase
+
 
+
hashCode() - Method in class org.apache.fluss.flink.source.state.SourceEnumeratorState
+
 
+
hashCode() - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
+
 
+
hashCode() - Method in class org.apache.fluss.flink.tiering.source.split.TieringLogSplit
+
 
+
hashCode() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSnapshotSplit
+
 
+
hashCode() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
+
 
+
hashCode() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
+
 
+
hasIncomplete() - Method in class org.apache.fluss.client.write.RecordAccumulator
+
+
Check whether there are any pending batches (whether sent or unsent).
+
+
hasNext() - Method in class org.apache.fluss.flink.lake.reader.IndexedLakeSplitRecordIterator
+
 
+
hasUnDrained() - Method in class org.apache.fluss.client.write.RecordAccumulator
+
+
Check whether there are any batches which haven't been drained.
+
+
HYBRID_SNAPSHOT_SPLIT_FLAG - Static variable in class org.apache.fluss.flink.source.split.SourceSplitBase
+
 
+
HybridSnapshotLogSplit - Class in org.apache.fluss.flink.source.split
+
+
The hybrid split for first reading the snapshot files and then switch to read the cdc log from a + specified offset.
+
+
HybridSnapshotLogSplit(TableBucket, String, long, long) - Constructor for class org.apache.fluss.flink.source.split.HybridSnapshotLogSplit
+
 
+
HybridSnapshotLogSplit(TableBucket, String, long, long, boolean, long) - Constructor for class org.apache.fluss.flink.source.split.HybridSnapshotLogSplit
+
 
+
HybridSnapshotLogSplitState - Class in org.apache.fluss.flink.source.split
+
+ +
+
HybridSnapshotLogSplitState(HybridSnapshotLogSplit) - Constructor for class org.apache.fluss.flink.source.split.HybridSnapshotLogSplitState
+
 
+
+A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-9.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-9.html new file mode 100644 index 0000000000..f3d2fc54cf --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-9.html @@ -0,0 +1,327 @@ + + + + + +I-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages + + +

I

+
+
IdempotenceBucketEntry - Class in org.apache.fluss.client.write
+
+
Entry to store the idempotence information of each table-bucket.
+
+
IdempotenceBucketMap - Class in org.apache.fluss.client.write
+
+
Map to manage IdempotenceBucketEntry of all table-bucket.
+
+
IdempotenceBucketMap() - Constructor for class org.apache.fluss.client.write.IdempotenceBucketMap
+
 
+
IdempotenceManager - Class in org.apache.fluss.client.write
+
+
A class which manages the idempotence in writer keeps the state necessary to ensure idempotent of + writer.
+
+
IdempotenceManager(boolean, int, TabletServerGateway) - Constructor for class org.apache.fluss.client.write.IdempotenceManager
+
 
+
IDENTIFIER - Static variable in class org.apache.fluss.flink.catalog.FlinkCatalogFactory
+
 
+
IGNORE - org.apache.fluss.flink.row.OperationType
+
+
Represents an operation that should be ignored.
+
+
inc() - Method in class org.apache.fluss.flink.metrics.FlinkCounter
+
 
+
inc(long) - Method in class org.apache.fluss.flink.metrics.FlinkCounter
+
 
+
INDEXED_LOG - org.apache.fluss.client.write.WriteFormat
+
 
+
IndexedLakeSplitRecordIterator - Class in org.apache.fluss.flink.lake.reader
+
+
An iterator wrapper that converts LogRecord objects to InternalRow objects while tracking the + current LakeSplit index being processed.
+
+
IndexedLakeSplitRecordIterator(CloseableIterator<LogRecord>, int) - Constructor for class org.apache.fluss.flink.lake.reader.IndexedLakeSplitRecordIterator
+
 
+
IndexedLogWriteBatch - Class in org.apache.fluss.client.write
+
+
A batch of log records managed in INDEXED format that is or will be sent to server by ProduceLogRequest.
+
+
IndexedLogWriteBatch(int, PhysicalTablePath, int, int, AbstractPagedOutputView, long) - Constructor for class org.apache.fluss.client.write.IndexedLogWriteBatch
+
 
+
INITIAL_OFFSET - Static variable in class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
+
 
+
initialize(SinkWriterMetricGroup) - Method in class org.apache.fluss.flink.sink.writer.AppendSinkWriter
+
 
+
initialize(SinkWriterMetricGroup) - Method in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
+
 
+
initialize(SinkWriterMetricGroup) - Method in class org.apache.fluss.flink.sink.writer.UpsertSinkWriter
+
 
+
initializedState(SourceSplitBase) - Static method in class org.apache.fluss.flink.lake.LakeSplitStateInitializer
+
 
+
initializedState(SourceSplitBase) - Method in class org.apache.fluss.flink.source.reader.FlinkSourceReader
+
 
+
initializedState(TieringSplit) - Method in class org.apache.fluss.flink.tiering.source.TieringSourceReader
+
 
+
initiateClose() - Method in class org.apache.fluss.client.write.Sender
+
+
Start closing the sender (won't actually complete until all data is sent out).
+
+
initMetrics() - Method in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
+
 
+
INSTANCE - Static variable in class org.apache.fluss.flink.tiering.source.split.TieringSplitSerializer
+
 
+
INSTANCE - Static variable in class org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorStateSerializer
+
 
+
internalCall(String, String, String, String, String) - Method in class org.apache.fluss.flink.procedure.AbstractAclProcedure
+
+
Invokes the ACL operation as a Flink table procedure.
+
+
invalidPhysicalTableBucketMeta(Set<PhysicalTablePath>) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
+
+
Invalid the bucket metadata for the given physical table paths.
+
+
isBasicType() - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
+
 
+
isBasicType() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
+
 
+
isClosed() - Method in class org.apache.fluss.client.write.ArrowLogWriteBatch
+
 
+
isClosed() - Method in class org.apache.fluss.client.write.IndexedLogWriteBatch
+
 
+
isClosed() - Method in class org.apache.fluss.client.write.KvWriteBatch
+
 
+
isClosed() - Method in class org.apache.fluss.client.write.WriteBatch
+
+
check if the batch is closed.
+
+
isDatabaseAlreadyExist(Throwable) - Static method in class org.apache.fluss.flink.utils.CatalogExceptionUtils
+
 
+
isDatabaseNotEmpty(Throwable) - Static method in class org.apache.fluss.flink.utils.CatalogExceptionUtils
+
 
+
isDatabaseNotExist(Throwable) - Static method in class org.apache.fluss.flink.utils.CatalogExceptionUtils
+
 
+
isDelete() - Method in class org.apache.fluss.flink.lake.reader.KeyValueRow
+
 
+
isDone() - Method in class org.apache.fluss.client.table.scanner.log.RemoteLogDownloadFuture
+
 
+
isEmpty() - Method in class org.apache.fluss.client.table.scanner.log.ScanRecords
+
 
+
isHybridSnapshotLogSplit() - Method in class org.apache.fluss.flink.source.split.SourceSplitBase
+
+
Checks whether this split is a hybrid snapshot and log split.
+
+
isHybridSnapshotLogSplitState() - Method in class org.apache.fluss.flink.source.split.SourceSplitState
+
+
Checks whether this split state is a hybrid snapshot log split state.
+
+
isKeyType() - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
+
 
+
isKeyType() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
+
 
+
isLakeSplit() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
+
 
+
isLakeSplit() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotSplit
+
 
+
isLakeSplit() - Method in class org.apache.fluss.flink.source.split.SourceSplitBase
+
 
+
isLakeSplit() - Method in class org.apache.fluss.flink.source.split.SourceSplitState
+
 
+
isLakeSplitFinished() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
+
 
+
isLakeTableAlreadyExist(Throwable) - Static method in class org.apache.fluss.flink.utils.CatalogExceptionUtils
+
 
+
isLogSplit() - Method in class org.apache.fluss.flink.source.split.SourceSplitBase
+
+
Checks whether this split is a log split.
+
+
isLogSplitState() - Method in class org.apache.fluss.flink.source.split.SourceSplitState
+
+
Checks whether this split state is a log split state.
+
+
isMatch(RowData) - Method in class org.apache.fluss.flink.source.lookup.LookupNormalizer.RemainingFilter
+
 
+
isNullAt(int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
+
 
+
ISO_8601 - org.apache.fluss.flink.utils.TimestampFormat
+
+
Options to specify TIMESTAMP/TIMESTAMP_WITH_LOCAL_ZONE format.
+
+
isPartitionAlreadyExists(Throwable) - Static method in class org.apache.fluss.flink.utils.CatalogExceptionUtils
+
 
+
isPartitionInvalid(Throwable) - Static method in class org.apache.fluss.flink.utils.CatalogExceptionUtils
+
 
+
isPartitionNotExist(Throwable) - Static method in class org.apache.fluss.flink.utils.CatalogExceptionUtils
+
 
+
isPointwise() - Method in class org.apache.fluss.flink.sink.FlinkStreamPartitioner
+
 
+
isRunning() - Method in class org.apache.fluss.client.write.Sender
+
 
+
isSnapshotFinished() - Method in class org.apache.fluss.flink.source.split.HybridSnapshotLogSplit
+
 
+
isStreaming() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
+
 
+
isTableAlreadyExist(Throwable) - Static method in class org.apache.fluss.flink.utils.CatalogExceptionUtils
+
 
+
isTableInvalid(Throwable) - Static method in class org.apache.fluss.flink.utils.CatalogExceptionUtils
+
 
+
isTableNotExist(Throwable) - Static method in class org.apache.fluss.flink.utils.CatalogExceptionUtils
+
 
+
isTableNotPartitioned(Throwable) - Static method in class org.apache.fluss.flink.utils.CatalogExceptionUtils
+
 
+
isTieringLogSplit() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
+
+
Checks whether this split is a log split to tier.
+
+
isTieringSnapshotSplit() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
+
+
Checks whether this split is a primary key table split to tier.
+
+
isTupleType() - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
+
 
+
isTupleType() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
+
 
+
iterator() - Method in class org.apache.fluss.client.table.scanner.log.ScanRecords
+
 
+
+A B C D E F G H I J K L M N O P Q R S T U V W 
All Classes All Packages
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/index.html b/website/static/javadoc/0.8-SNAPSHOT/index.html new file mode 100644 index 0000000000..3009a01bc3 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/index.html @@ -0,0 +1,346 @@ + + + + + +Overview (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Apache Fluss 0.8-SNAPSHOT API

+
+
+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
Packages 
PackageDescription
org.apache.fluss.client 
org.apache.fluss.client.admin 
org.apache.fluss.client.lookup 
org.apache.fluss.client.metadata 
org.apache.fluss.client.metrics 
org.apache.fluss.client.table 
org.apache.fluss.client.table.getter 
org.apache.fluss.client.table.scanner 
org.apache.fluss.client.table.scanner.batch 
org.apache.fluss.client.table.scanner.log 
org.apache.fluss.client.table.writer 
org.apache.fluss.client.token 
org.apache.fluss.client.utils 
org.apache.fluss.client.write 
org.apache.fluss.flink 
org.apache.fluss.flink.adapter 
org.apache.fluss.flink.catalog 
org.apache.fluss.flink.lake 
org.apache.fluss.flink.lake.reader 
org.apache.fluss.flink.lake.split 
org.apache.fluss.flink.lake.state 
org.apache.fluss.flink.metrics 
org.apache.fluss.flink.procedure 
org.apache.fluss.flink.row 
org.apache.fluss.flink.sink 
org.apache.fluss.flink.sink.serializer 
org.apache.fluss.flink.sink.writer 
org.apache.fluss.flink.source 
org.apache.fluss.flink.source.deserializer 
org.apache.fluss.flink.source.emitter 
org.apache.fluss.flink.source.enumerator 
org.apache.fluss.flink.source.enumerator.initializer 
org.apache.fluss.flink.source.event 
org.apache.fluss.flink.source.lookup 
org.apache.fluss.flink.source.metrics 
org.apache.fluss.flink.source.reader 
org.apache.fluss.flink.source.reader.fetcher 
org.apache.fluss.flink.source.split 
org.apache.fluss.flink.source.state 
org.apache.fluss.flink.tiering 
org.apache.fluss.flink.tiering.committer 
org.apache.fluss.flink.tiering.event 
org.apache.fluss.flink.tiering.source 
org.apache.fluss.flink.tiering.source.enumerator 
org.apache.fluss.flink.tiering.source.split 
org.apache.fluss.flink.tiering.source.state 
org.apache.fluss.flink.utils 
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/javadoc.sh b/website/static/javadoc/0.8-SNAPSHOT/javadoc.sh new file mode 100755 index 0000000000..136f87af54 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/javadoc.sh @@ -0,0 +1 @@ +/Users/mehulbatra/.sdkman/candidates/java/11.0.28-zulu/zulu-11.jdk/Contents/Home/bin/javadoc -Xdoclint:none -J-Xmx2g -J--illegal-access=permit @options @packages \ No newline at end of file diff --git a/website/static/javadoc/0.8-SNAPSHOT/jquery-ui.overrides.css b/website/static/javadoc/0.8-SNAPSHOT/jquery-ui.overrides.css new file mode 100644 index 0000000000..facf852c27 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/jquery-ui.overrides.css @@ -0,0 +1,35 @@ +/* + * Copyright (c) 2020, 2022, Oracle and/or its affiliates. All rights reserved. + * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. + * + * This code is free software; you can redistribute it and/or modify it + * under the terms of the GNU General Public License version 2 only, as + * published by the Free Software Foundation. Oracle designates this + * particular file as subject to the "Classpath" exception as provided + * by Oracle in the LICENSE file that accompanied this code. + * + * This code is distributed in the hope that it will be useful, but WITHOUT + * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or + * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License + * version 2 for more details (a copy is included in the LICENSE file that + * accompanied this code). + * + * You should have received a copy of the GNU General Public License version + * 2 along with this work; if not, write to the Free Software Foundation, + * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. + * + * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA + * or visit www.oracle.com if you need additional information or have any + * questions. + */ + +.ui-state-active, +.ui-widget-content .ui-state-active, +.ui-widget-header .ui-state-active, +a.ui-button:active, +.ui-button:active, +.ui-button.ui-state-active:hover { + /* Overrides the color of selection used in jQuery UI */ + background: #F8981D; + border: 1px solid #F8981D; +} diff --git a/website/static/javadoc/0.8-SNAPSHOT/jquery/external/jquery/jquery.js b/website/static/javadoc/0.8-SNAPSHOT/jquery/external/jquery/jquery.js new file mode 100644 index 0000000000..50937333b9 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/jquery/external/jquery/jquery.js @@ -0,0 +1,10872 @@ +/*! + * jQuery JavaScript Library v3.5.1 + * https://jquery.com/ + * + * Includes Sizzle.js + * https://sizzlejs.com/ + * + * Copyright JS Foundation and other contributors + * Released under the MIT license + * https://jquery.org/license + * + * Date: 2020-05-04T22:49Z + */ +( function( global, factory ) { + + "use strict"; + + if ( typeof module === "object" && typeof module.exports === "object" ) { + + // For CommonJS and CommonJS-like environments where a proper `window` + // is present, execute the factory and get jQuery. + // For environments that do not have a `window` with a `document` + // (such as Node.js), expose a factory as module.exports. + // This accentuates the need for the creation of a real `window`. + // e.g. var jQuery = require("jquery")(window); + // See ticket #14549 for more info. + module.exports = global.document ? + factory( global, true ) : + function( w ) { + if ( !w.document ) { + throw new Error( "jQuery requires a window with a document" ); + } + return factory( w ); + }; + } else { + factory( global ); + } + +// Pass this if window is not defined yet +} )( typeof window !== "undefined" ? window : this, function( window, noGlobal ) { + +// Edge <= 12 - 13+, Firefox <=18 - 45+, IE 10 - 11, Safari 5.1 - 9+, iOS 6 - 9.1 +// throw exceptions when non-strict code (e.g., ASP.NET 4.5) accesses strict mode +// arguments.callee.caller (trac-13335). But as of jQuery 3.0 (2016), strict mode should be common +// enough that all such attempts are guarded in a try block. +"use strict"; + +var arr = []; + +var getProto = Object.getPrototypeOf; + +var slice = arr.slice; + +var flat = arr.flat ? function( array ) { + return arr.flat.call( array ); +} : function( array ) { + return arr.concat.apply( [], array ); +}; + + +var push = arr.push; + +var indexOf = arr.indexOf; + +var class2type = {}; + +var toString = class2type.toString; + +var hasOwn = class2type.hasOwnProperty; + +var fnToString = hasOwn.toString; + +var ObjectFunctionString = fnToString.call( Object ); + +var support = {}; + +var isFunction = function isFunction( obj ) { + + // Support: Chrome <=57, Firefox <=52 + // In some browsers, typeof returns "function" for HTML elements + // (i.e., `typeof document.createElement( "object" ) === "function"`). + // We don't want to classify *any* DOM node as a function. + return typeof obj === "function" && typeof obj.nodeType !== "number"; + }; + + +var isWindow = function isWindow( obj ) { + return obj != null && obj === obj.window; + }; + + +var document = window.document; + + + + var preservedScriptAttributes = { + type: true, + src: true, + nonce: true, + noModule: true + }; + + function DOMEval( code, node, doc ) { + doc = doc || document; + + var i, val, + script = doc.createElement( "script" ); + + script.text = code; + if ( node ) { + for ( i in preservedScriptAttributes ) { + + // Support: Firefox 64+, Edge 18+ + // Some browsers don't support the "nonce" property on scripts. + // On the other hand, just using `getAttribute` is not enough as + // the `nonce` attribute is reset to an empty string whenever it + // becomes browsing-context connected. + // See https://github.com/whatwg/html/issues/2369 + // See https://html.spec.whatwg.org/#nonce-attributes + // The `node.getAttribute` check was added for the sake of + // `jQuery.globalEval` so that it can fake a nonce-containing node + // via an object. + val = node[ i ] || node.getAttribute && node.getAttribute( i ); + if ( val ) { + script.setAttribute( i, val ); + } + } + } + doc.head.appendChild( script ).parentNode.removeChild( script ); + } + + +function toType( obj ) { + if ( obj == null ) { + return obj + ""; + } + + // Support: Android <=2.3 only (functionish RegExp) + return typeof obj === "object" || typeof obj === "function" ? + class2type[ toString.call( obj ) ] || "object" : + typeof obj; +} +/* global Symbol */ +// Defining this global in .eslintrc.json would create a danger of using the global +// unguarded in another place, it seems safer to define global only for this module + + + +var + version = "3.5.1", + + // Define a local copy of jQuery + jQuery = function( selector, context ) { + + // The jQuery object is actually just the init constructor 'enhanced' + // Need init if jQuery is called (just allow error to be thrown if not included) + return new jQuery.fn.init( selector, context ); + }; + +jQuery.fn = jQuery.prototype = { + + // The current version of jQuery being used + jquery: version, + + constructor: jQuery, + + // The default length of a jQuery object is 0 + length: 0, + + toArray: function() { + return slice.call( this ); + }, + + // Get the Nth element in the matched element set OR + // Get the whole matched element set as a clean array + get: function( num ) { + + // Return all the elements in a clean array + if ( num == null ) { + return slice.call( this ); + } + + // Return just the one element from the set + return num < 0 ? this[ num + this.length ] : this[ num ]; + }, + + // Take an array of elements and push it onto the stack + // (returning the new matched element set) + pushStack: function( elems ) { + + // Build a new jQuery matched element set + var ret = jQuery.merge( this.constructor(), elems ); + + // Add the old object onto the stack (as a reference) + ret.prevObject = this; + + // Return the newly-formed element set + return ret; + }, + + // Execute a callback for every element in the matched set. + each: function( callback ) { + return jQuery.each( this, callback ); + }, + + map: function( callback ) { + return this.pushStack( jQuery.map( this, function( elem, i ) { + return callback.call( elem, i, elem ); + } ) ); + }, + + slice: function() { + return this.pushStack( slice.apply( this, arguments ) ); + }, + + first: function() { + return this.eq( 0 ); + }, + + last: function() { + return this.eq( -1 ); + }, + + even: function() { + return this.pushStack( jQuery.grep( this, function( _elem, i ) { + return ( i + 1 ) % 2; + } ) ); + }, + + odd: function() { + return this.pushStack( jQuery.grep( this, function( _elem, i ) { + return i % 2; + } ) ); + }, + + eq: function( i ) { + var len = this.length, + j = +i + ( i < 0 ? len : 0 ); + return this.pushStack( j >= 0 && j < len ? [ this[ j ] ] : [] ); + }, + + end: function() { + return this.prevObject || this.constructor(); + }, + + // For internal use only. + // Behaves like an Array's method, not like a jQuery method. + push: push, + sort: arr.sort, + splice: arr.splice +}; + +jQuery.extend = jQuery.fn.extend = function() { + var options, name, src, copy, copyIsArray, clone, + target = arguments[ 0 ] || {}, + i = 1, + length = arguments.length, + deep = false; + + // Handle a deep copy situation + if ( typeof target === "boolean" ) { + deep = target; + + // Skip the boolean and the target + target = arguments[ i ] || {}; + i++; + } + + // Handle case when target is a string or something (possible in deep copy) + if ( typeof target !== "object" && !isFunction( target ) ) { + target = {}; + } + + // Extend jQuery itself if only one argument is passed + if ( i === length ) { + target = this; + i--; + } + + for ( ; i < length; i++ ) { + + // Only deal with non-null/undefined values + if ( ( options = arguments[ i ] ) != null ) { + + // Extend the base object + for ( name in options ) { + copy = options[ name ]; + + // Prevent Object.prototype pollution + // Prevent never-ending loop + if ( name === "__proto__" || target === copy ) { + continue; + } + + // Recurse if we're merging plain objects or arrays + if ( deep && copy && ( jQuery.isPlainObject( copy ) || + ( copyIsArray = Array.isArray( copy ) ) ) ) { + src = target[ name ]; + + // Ensure proper type for the source value + if ( copyIsArray && !Array.isArray( src ) ) { + clone = []; + } else if ( !copyIsArray && !jQuery.isPlainObject( src ) ) { + clone = {}; + } else { + clone = src; + } + copyIsArray = false; + + // Never move original objects, clone them + target[ name ] = jQuery.extend( deep, clone, copy ); + + // Don't bring in undefined values + } else if ( copy !== undefined ) { + target[ name ] = copy; + } + } + } + } + + // Return the modified object + return target; +}; + +jQuery.extend( { + + // Unique for each copy of jQuery on the page + expando: "jQuery" + ( version + Math.random() ).replace( /\D/g, "" ), + + // Assume jQuery is ready without the ready module + isReady: true, + + error: function( msg ) { + throw new Error( msg ); + }, + + noop: function() {}, + + isPlainObject: function( obj ) { + var proto, Ctor; + + // Detect obvious negatives + // Use toString instead of jQuery.type to catch host objects + if ( !obj || toString.call( obj ) !== "[object Object]" ) { + return false; + } + + proto = getProto( obj ); + + // Objects with no prototype (e.g., `Object.create( null )`) are plain + if ( !proto ) { + return true; + } + + // Objects with prototype are plain iff they were constructed by a global Object function + Ctor = hasOwn.call( proto, "constructor" ) && proto.constructor; + return typeof Ctor === "function" && fnToString.call( Ctor ) === ObjectFunctionString; + }, + + isEmptyObject: function( obj ) { + var name; + + for ( name in obj ) { + return false; + } + return true; + }, + + // Evaluates a script in a provided context; falls back to the global one + // if not specified. + globalEval: function( code, options, doc ) { + DOMEval( code, { nonce: options && options.nonce }, doc ); + }, + + each: function( obj, callback ) { + var length, i = 0; + + if ( isArrayLike( obj ) ) { + length = obj.length; + for ( ; i < length; i++ ) { + if ( callback.call( obj[ i ], i, obj[ i ] ) === false ) { + break; + } + } + } else { + for ( i in obj ) { + if ( callback.call( obj[ i ], i, obj[ i ] ) === false ) { + break; + } + } + } + + return obj; + }, + + // results is for internal usage only + makeArray: function( arr, results ) { + var ret = results || []; + + if ( arr != null ) { + if ( isArrayLike( Object( arr ) ) ) { + jQuery.merge( ret, + typeof arr === "string" ? + [ arr ] : arr + ); + } else { + push.call( ret, arr ); + } + } + + return ret; + }, + + inArray: function( elem, arr, i ) { + return arr == null ? -1 : indexOf.call( arr, elem, i ); + }, + + // Support: Android <=4.0 only, PhantomJS 1 only + // push.apply(_, arraylike) throws on ancient WebKit + merge: function( first, second ) { + var len = +second.length, + j = 0, + i = first.length; + + for ( ; j < len; j++ ) { + first[ i++ ] = second[ j ]; + } + + first.length = i; + + return first; + }, + + grep: function( elems, callback, invert ) { + var callbackInverse, + matches = [], + i = 0, + length = elems.length, + callbackExpect = !invert; + + // Go through the array, only saving the items + // that pass the validator function + for ( ; i < length; i++ ) { + callbackInverse = !callback( elems[ i ], i ); + if ( callbackInverse !== callbackExpect ) { + matches.push( elems[ i ] ); + } + } + + return matches; + }, + + // arg is for internal usage only + map: function( elems, callback, arg ) { + var length, value, + i = 0, + ret = []; + + // Go through the array, translating each of the items to their new values + if ( isArrayLike( elems ) ) { + length = elems.length; + for ( ; i < length; i++ ) { + value = callback( elems[ i ], i, arg ); + + if ( value != null ) { + ret.push( value ); + } + } + + // Go through every key on the object, + } else { + for ( i in elems ) { + value = callback( elems[ i ], i, arg ); + + if ( value != null ) { + ret.push( value ); + } + } + } + + // Flatten any nested arrays + return flat( ret ); + }, + + // A global GUID counter for objects + guid: 1, + + // jQuery.support is not used in Core but other projects attach their + // properties to it so it needs to exist. + support: support +} ); + +if ( typeof Symbol === "function" ) { + jQuery.fn[ Symbol.iterator ] = arr[ Symbol.iterator ]; +} + +// Populate the class2type map +jQuery.each( "Boolean Number String Function Array Date RegExp Object Error Symbol".split( " " ), +function( _i, name ) { + class2type[ "[object " + name + "]" ] = name.toLowerCase(); +} ); + +function isArrayLike( obj ) { + + // Support: real iOS 8.2 only (not reproducible in simulator) + // `in` check used to prevent JIT error (gh-2145) + // hasOwn isn't used here due to false negatives + // regarding Nodelist length in IE + var length = !!obj && "length" in obj && obj.length, + type = toType( obj ); + + if ( isFunction( obj ) || isWindow( obj ) ) { + return false; + } + + return type === "array" || length === 0 || + typeof length === "number" && length > 0 && ( length - 1 ) in obj; +} +var Sizzle = +/*! + * Sizzle CSS Selector Engine v2.3.5 + * https://sizzlejs.com/ + * + * Copyright JS Foundation and other contributors + * Released under the MIT license + * https://js.foundation/ + * + * Date: 2020-03-14 + */ +( function( window ) { +var i, + support, + Expr, + getText, + isXML, + tokenize, + compile, + select, + outermostContext, + sortInput, + hasDuplicate, + + // Local document vars + setDocument, + document, + docElem, + documentIsHTML, + rbuggyQSA, + rbuggyMatches, + matches, + contains, + + // Instance-specific data + expando = "sizzle" + 1 * new Date(), + preferredDoc = window.document, + dirruns = 0, + done = 0, + classCache = createCache(), + tokenCache = createCache(), + compilerCache = createCache(), + nonnativeSelectorCache = createCache(), + sortOrder = function( a, b ) { + if ( a === b ) { + hasDuplicate = true; + } + return 0; + }, + + // Instance methods + hasOwn = ( {} ).hasOwnProperty, + arr = [], + pop = arr.pop, + pushNative = arr.push, + push = arr.push, + slice = arr.slice, + + // Use a stripped-down indexOf as it's faster than native + // https://jsperf.com/thor-indexof-vs-for/5 + indexOf = function( list, elem ) { + var i = 0, + len = list.length; + for ( ; i < len; i++ ) { + if ( list[ i ] === elem ) { + return i; + } + } + return -1; + }, + + booleans = "checked|selected|async|autofocus|autoplay|controls|defer|disabled|hidden|" + + "ismap|loop|multiple|open|readonly|required|scoped", + + // Regular expressions + + // http://www.w3.org/TR/css3-selectors/#whitespace + whitespace = "[\\x20\\t\\r\\n\\f]", + + // https://www.w3.org/TR/css-syntax-3/#ident-token-diagram + identifier = "(?:\\\\[\\da-fA-F]{1,6}" + whitespace + + "?|\\\\[^\\r\\n\\f]|[\\w-]|[^\0-\\x7f])+", + + // Attribute selectors: http://www.w3.org/TR/selectors/#attribute-selectors + attributes = "\\[" + whitespace + "*(" + identifier + ")(?:" + whitespace + + + // Operator (capture 2) + "*([*^$|!~]?=)" + whitespace + + + // "Attribute values must be CSS identifiers [capture 5] + // or strings [capture 3 or capture 4]" + "*(?:'((?:\\\\.|[^\\\\'])*)'|\"((?:\\\\.|[^\\\\\"])*)\"|(" + identifier + "))|)" + + whitespace + "*\\]", + + pseudos = ":(" + identifier + ")(?:\\((" + + + // To reduce the number of selectors needing tokenize in the preFilter, prefer arguments: + // 1. quoted (capture 3; capture 4 or capture 5) + "('((?:\\\\.|[^\\\\'])*)'|\"((?:\\\\.|[^\\\\\"])*)\")|" + + + // 2. simple (capture 6) + "((?:\\\\.|[^\\\\()[\\]]|" + attributes + ")*)|" + + + // 3. anything else (capture 2) + ".*" + + ")\\)|)", + + // Leading and non-escaped trailing whitespace, capturing some non-whitespace characters preceding the latter + rwhitespace = new RegExp( whitespace + "+", "g" ), + rtrim = new RegExp( "^" + whitespace + "+|((?:^|[^\\\\])(?:\\\\.)*)" + + whitespace + "+$", "g" ), + + rcomma = new RegExp( "^" + whitespace + "*," + whitespace + "*" ), + rcombinators = new RegExp( "^" + whitespace + "*([>+~]|" + whitespace + ")" + whitespace + + "*" ), + rdescend = new RegExp( whitespace + "|>" ), + + rpseudo = new RegExp( pseudos ), + ridentifier = new RegExp( "^" + identifier + "$" ), + + matchExpr = { + "ID": new RegExp( "^#(" + identifier + ")" ), + "CLASS": new RegExp( "^\\.(" + identifier + ")" ), + "TAG": new RegExp( "^(" + identifier + "|[*])" ), + "ATTR": new RegExp( "^" + attributes ), + "PSEUDO": new RegExp( "^" + pseudos ), + "CHILD": new RegExp( "^:(only|first|last|nth|nth-last)-(child|of-type)(?:\\(" + + whitespace + "*(even|odd|(([+-]|)(\\d*)n|)" + whitespace + "*(?:([+-]|)" + + whitespace + "*(\\d+)|))" + whitespace + "*\\)|)", "i" ), + "bool": new RegExp( "^(?:" + booleans + ")$", "i" ), + + // For use in libraries implementing .is() + // We use this for POS matching in `select` + "needsContext": new RegExp( "^" + whitespace + + "*[>+~]|:(even|odd|eq|gt|lt|nth|first|last)(?:\\(" + whitespace + + "*((?:-\\d)?\\d*)" + whitespace + "*\\)|)(?=[^-]|$)", "i" ) + }, + + rhtml = /HTML$/i, + rinputs = /^(?:input|select|textarea|button)$/i, + rheader = /^h\d$/i, + + rnative = /^[^{]+\{\s*\[native \w/, + + // Easily-parseable/retrievable ID or TAG or CLASS selectors + rquickExpr = /^(?:#([\w-]+)|(\w+)|\.([\w-]+))$/, + + rsibling = /[+~]/, + + // CSS escapes + // http://www.w3.org/TR/CSS21/syndata.html#escaped-characters + runescape = new RegExp( "\\\\[\\da-fA-F]{1,6}" + whitespace + "?|\\\\([^\\r\\n\\f])", "g" ), + funescape = function( escape, nonHex ) { + var high = "0x" + escape.slice( 1 ) - 0x10000; + + return nonHex ? + + // Strip the backslash prefix from a non-hex escape sequence + nonHex : + + // Replace a hexadecimal escape sequence with the encoded Unicode code point + // Support: IE <=11+ + // For values outside the Basic Multilingual Plane (BMP), manually construct a + // surrogate pair + high < 0 ? + String.fromCharCode( high + 0x10000 ) : + String.fromCharCode( high >> 10 | 0xD800, high & 0x3FF | 0xDC00 ); + }, + + // CSS string/identifier serialization + // https://drafts.csswg.org/cssom/#common-serializing-idioms + rcssescape = /([\0-\x1f\x7f]|^-?\d)|^-$|[^\0-\x1f\x7f-\uFFFF\w-]/g, + fcssescape = function( ch, asCodePoint ) { + if ( asCodePoint ) { + + // U+0000 NULL becomes U+FFFD REPLACEMENT CHARACTER + if ( ch === "\0" ) { + return "\uFFFD"; + } + + // Control characters and (dependent upon position) numbers get escaped as code points + return ch.slice( 0, -1 ) + "\\" + + ch.charCodeAt( ch.length - 1 ).toString( 16 ) + " "; + } + + // Other potentially-special ASCII characters get backslash-escaped + return "\\" + ch; + }, + + // Used for iframes + // See setDocument() + // Removing the function wrapper causes a "Permission Denied" + // error in IE + unloadHandler = function() { + setDocument(); + }, + + inDisabledFieldset = addCombinator( + function( elem ) { + return elem.disabled === true && elem.nodeName.toLowerCase() === "fieldset"; + }, + { dir: "parentNode", next: "legend" } + ); + +// Optimize for push.apply( _, NodeList ) +try { + push.apply( + ( arr = slice.call( preferredDoc.childNodes ) ), + preferredDoc.childNodes + ); + + // Support: Android<4.0 + // Detect silently failing push.apply + // eslint-disable-next-line no-unused-expressions + arr[ preferredDoc.childNodes.length ].nodeType; +} catch ( e ) { + push = { apply: arr.length ? + + // Leverage slice if possible + function( target, els ) { + pushNative.apply( target, slice.call( els ) ); + } : + + // Support: IE<9 + // Otherwise append directly + function( target, els ) { + var j = target.length, + i = 0; + + // Can't trust NodeList.length + while ( ( target[ j++ ] = els[ i++ ] ) ) {} + target.length = j - 1; + } + }; +} + +function Sizzle( selector, context, results, seed ) { + var m, i, elem, nid, match, groups, newSelector, + newContext = context && context.ownerDocument, + + // nodeType defaults to 9, since context defaults to document + nodeType = context ? context.nodeType : 9; + + results = results || []; + + // Return early from calls with invalid selector or context + if ( typeof selector !== "string" || !selector || + nodeType !== 1 && nodeType !== 9 && nodeType !== 11 ) { + + return results; + } + + // Try to shortcut find operations (as opposed to filters) in HTML documents + if ( !seed ) { + setDocument( context ); + context = context || document; + + if ( documentIsHTML ) { + + // If the selector is sufficiently simple, try using a "get*By*" DOM method + // (excepting DocumentFragment context, where the methods don't exist) + if ( nodeType !== 11 && ( match = rquickExpr.exec( selector ) ) ) { + + // ID selector + if ( ( m = match[ 1 ] ) ) { + + // Document context + if ( nodeType === 9 ) { + if ( ( elem = context.getElementById( m ) ) ) { + + // Support: IE, Opera, Webkit + // TODO: identify versions + // getElementById can match elements by name instead of ID + if ( elem.id === m ) { + results.push( elem ); + return results; + } + } else { + return results; + } + + // Element context + } else { + + // Support: IE, Opera, Webkit + // TODO: identify versions + // getElementById can match elements by name instead of ID + if ( newContext && ( elem = newContext.getElementById( m ) ) && + contains( context, elem ) && + elem.id === m ) { + + results.push( elem ); + return results; + } + } + + // Type selector + } else if ( match[ 2 ] ) { + push.apply( results, context.getElementsByTagName( selector ) ); + return results; + + // Class selector + } else if ( ( m = match[ 3 ] ) && support.getElementsByClassName && + context.getElementsByClassName ) { + + push.apply( results, context.getElementsByClassName( m ) ); + return results; + } + } + + // Take advantage of querySelectorAll + if ( support.qsa && + !nonnativeSelectorCache[ selector + " " ] && + ( !rbuggyQSA || !rbuggyQSA.test( selector ) ) && + + // Support: IE 8 only + // Exclude object elements + ( nodeType !== 1 || context.nodeName.toLowerCase() !== "object" ) ) { + + newSelector = selector; + newContext = context; + + // qSA considers elements outside a scoping root when evaluating child or + // descendant combinators, which is not what we want. + // In such cases, we work around the behavior by prefixing every selector in the + // list with an ID selector referencing the scope context. + // The technique has to be used as well when a leading combinator is used + // as such selectors are not recognized by querySelectorAll. + // Thanks to Andrew Dupont for this technique. + if ( nodeType === 1 && + ( rdescend.test( selector ) || rcombinators.test( selector ) ) ) { + + // Expand context for sibling selectors + newContext = rsibling.test( selector ) && testContext( context.parentNode ) || + context; + + // We can use :scope instead of the ID hack if the browser + // supports it & if we're not changing the context. + if ( newContext !== context || !support.scope ) { + + // Capture the context ID, setting it first if necessary + if ( ( nid = context.getAttribute( "id" ) ) ) { + nid = nid.replace( rcssescape, fcssescape ); + } else { + context.setAttribute( "id", ( nid = expando ) ); + } + } + + // Prefix every selector in the list + groups = tokenize( selector ); + i = groups.length; + while ( i-- ) { + groups[ i ] = ( nid ? "#" + nid : ":scope" ) + " " + + toSelector( groups[ i ] ); + } + newSelector = groups.join( "," ); + } + + try { + push.apply( results, + newContext.querySelectorAll( newSelector ) + ); + return results; + } catch ( qsaError ) { + nonnativeSelectorCache( selector, true ); + } finally { + if ( nid === expando ) { + context.removeAttribute( "id" ); + } + } + } + } + } + + // All others + return select( selector.replace( rtrim, "$1" ), context, results, seed ); +} + +/** + * Create key-value caches of limited size + * @returns {function(string, object)} Returns the Object data after storing it on itself with + * property name the (space-suffixed) string and (if the cache is larger than Expr.cacheLength) + * deleting the oldest entry + */ +function createCache() { + var keys = []; + + function cache( key, value ) { + + // Use (key + " ") to avoid collision with native prototype properties (see Issue #157) + if ( keys.push( key + " " ) > Expr.cacheLength ) { + + // Only keep the most recent entries + delete cache[ keys.shift() ]; + } + return ( cache[ key + " " ] = value ); + } + return cache; +} + +/** + * Mark a function for special use by Sizzle + * @param {Function} fn The function to mark + */ +function markFunction( fn ) { + fn[ expando ] = true; + return fn; +} + +/** + * Support testing using an element + * @param {Function} fn Passed the created element and returns a boolean result + */ +function assert( fn ) { + var el = document.createElement( "fieldset" ); + + try { + return !!fn( el ); + } catch ( e ) { + return false; + } finally { + + // Remove from its parent by default + if ( el.parentNode ) { + el.parentNode.removeChild( el ); + } + + // release memory in IE + el = null; + } +} + +/** + * Adds the same handler for all of the specified attrs + * @param {String} attrs Pipe-separated list of attributes + * @param {Function} handler The method that will be applied + */ +function addHandle( attrs, handler ) { + var arr = attrs.split( "|" ), + i = arr.length; + + while ( i-- ) { + Expr.attrHandle[ arr[ i ] ] = handler; + } +} + +/** + * Checks document order of two siblings + * @param {Element} a + * @param {Element} b + * @returns {Number} Returns less than 0 if a precedes b, greater than 0 if a follows b + */ +function siblingCheck( a, b ) { + var cur = b && a, + diff = cur && a.nodeType === 1 && b.nodeType === 1 && + a.sourceIndex - b.sourceIndex; + + // Use IE sourceIndex if available on both nodes + if ( diff ) { + return diff; + } + + // Check if b follows a + if ( cur ) { + while ( ( cur = cur.nextSibling ) ) { + if ( cur === b ) { + return -1; + } + } + } + + return a ? 1 : -1; +} + +/** + * Returns a function to use in pseudos for input types + * @param {String} type + */ +function createInputPseudo( type ) { + return function( elem ) { + var name = elem.nodeName.toLowerCase(); + return name === "input" && elem.type === type; + }; +} + +/** + * Returns a function to use in pseudos for buttons + * @param {String} type + */ +function createButtonPseudo( type ) { + return function( elem ) { + var name = elem.nodeName.toLowerCase(); + return ( name === "input" || name === "button" ) && elem.type === type; + }; +} + +/** + * Returns a function to use in pseudos for :enabled/:disabled + * @param {Boolean} disabled true for :disabled; false for :enabled + */ +function createDisabledPseudo( disabled ) { + + // Known :disabled false positives: fieldset[disabled] > legend:nth-of-type(n+2) :can-disable + return function( elem ) { + + // Only certain elements can match :enabled or :disabled + // https://html.spec.whatwg.org/multipage/scripting.html#selector-enabled + // https://html.spec.whatwg.org/multipage/scripting.html#selector-disabled + if ( "form" in elem ) { + + // Check for inherited disabledness on relevant non-disabled elements: + // * listed form-associated elements in a disabled fieldset + // https://html.spec.whatwg.org/multipage/forms.html#category-listed + // https://html.spec.whatwg.org/multipage/forms.html#concept-fe-disabled + // * option elements in a disabled optgroup + // https://html.spec.whatwg.org/multipage/forms.html#concept-option-disabled + // All such elements have a "form" property. + if ( elem.parentNode && elem.disabled === false ) { + + // Option elements defer to a parent optgroup if present + if ( "label" in elem ) { + if ( "label" in elem.parentNode ) { + return elem.parentNode.disabled === disabled; + } else { + return elem.disabled === disabled; + } + } + + // Support: IE 6 - 11 + // Use the isDisabled shortcut property to check for disabled fieldset ancestors + return elem.isDisabled === disabled || + + // Where there is no isDisabled, check manually + /* jshint -W018 */ + elem.isDisabled !== !disabled && + inDisabledFieldset( elem ) === disabled; + } + + return elem.disabled === disabled; + + // Try to winnow out elements that can't be disabled before trusting the disabled property. + // Some victims get caught in our net (label, legend, menu, track), but it shouldn't + // even exist on them, let alone have a boolean value. + } else if ( "label" in elem ) { + return elem.disabled === disabled; + } + + // Remaining elements are neither :enabled nor :disabled + return false; + }; +} + +/** + * Returns a function to use in pseudos for positionals + * @param {Function} fn + */ +function createPositionalPseudo( fn ) { + return markFunction( function( argument ) { + argument = +argument; + return markFunction( function( seed, matches ) { + var j, + matchIndexes = fn( [], seed.length, argument ), + i = matchIndexes.length; + + // Match elements found at the specified indexes + while ( i-- ) { + if ( seed[ ( j = matchIndexes[ i ] ) ] ) { + seed[ j ] = !( matches[ j ] = seed[ j ] ); + } + } + } ); + } ); +} + +/** + * Checks a node for validity as a Sizzle context + * @param {Element|Object=} context + * @returns {Element|Object|Boolean} The input node if acceptable, otherwise a falsy value + */ +function testContext( context ) { + return context && typeof context.getElementsByTagName !== "undefined" && context; +} + +// Expose support vars for convenience +support = Sizzle.support = {}; + +/** + * Detects XML nodes + * @param {Element|Object} elem An element or a document + * @returns {Boolean} True iff elem is a non-HTML XML node + */ +isXML = Sizzle.isXML = function( elem ) { + var namespace = elem.namespaceURI, + docElem = ( elem.ownerDocument || elem ).documentElement; + + // Support: IE <=8 + // Assume HTML when documentElement doesn't yet exist, such as inside loading iframes + // https://bugs.jquery.com/ticket/4833 + return !rhtml.test( namespace || docElem && docElem.nodeName || "HTML" ); +}; + +/** + * Sets document-related variables once based on the current document + * @param {Element|Object} [doc] An element or document object to use to set the document + * @returns {Object} Returns the current document + */ +setDocument = Sizzle.setDocument = function( node ) { + var hasCompare, subWindow, + doc = node ? node.ownerDocument || node : preferredDoc; + + // Return early if doc is invalid or already selected + // Support: IE 11+, Edge 17 - 18+ + // IE/Edge sometimes throw a "Permission denied" error when strict-comparing + // two documents; shallow comparisons work. + // eslint-disable-next-line eqeqeq + if ( doc == document || doc.nodeType !== 9 || !doc.documentElement ) { + return document; + } + + // Update global variables + document = doc; + docElem = document.documentElement; + documentIsHTML = !isXML( document ); + + // Support: IE 9 - 11+, Edge 12 - 18+ + // Accessing iframe documents after unload throws "permission denied" errors (jQuery #13936) + // Support: IE 11+, Edge 17 - 18+ + // IE/Edge sometimes throw a "Permission denied" error when strict-comparing + // two documents; shallow comparisons work. + // eslint-disable-next-line eqeqeq + if ( preferredDoc != document && + ( subWindow = document.defaultView ) && subWindow.top !== subWindow ) { + + // Support: IE 11, Edge + if ( subWindow.addEventListener ) { + subWindow.addEventListener( "unload", unloadHandler, false ); + + // Support: IE 9 - 10 only + } else if ( subWindow.attachEvent ) { + subWindow.attachEvent( "onunload", unloadHandler ); + } + } + + // Support: IE 8 - 11+, Edge 12 - 18+, Chrome <=16 - 25 only, Firefox <=3.6 - 31 only, + // Safari 4 - 5 only, Opera <=11.6 - 12.x only + // IE/Edge & older browsers don't support the :scope pseudo-class. + // Support: Safari 6.0 only + // Safari 6.0 supports :scope but it's an alias of :root there. + support.scope = assert( function( el ) { + docElem.appendChild( el ).appendChild( document.createElement( "div" ) ); + return typeof el.querySelectorAll !== "undefined" && + !el.querySelectorAll( ":scope fieldset div" ).length; + } ); + + /* Attributes + ---------------------------------------------------------------------- */ + + // Support: IE<8 + // Verify that getAttribute really returns attributes and not properties + // (excepting IE8 booleans) + support.attributes = assert( function( el ) { + el.className = "i"; + return !el.getAttribute( "className" ); + } ); + + /* getElement(s)By* + ---------------------------------------------------------------------- */ + + // Check if getElementsByTagName("*") returns only elements + support.getElementsByTagName = assert( function( el ) { + el.appendChild( document.createComment( "" ) ); + return !el.getElementsByTagName( "*" ).length; + } ); + + // Support: IE<9 + support.getElementsByClassName = rnative.test( document.getElementsByClassName ); + + // Support: IE<10 + // Check if getElementById returns elements by name + // The broken getElementById methods don't pick up programmatically-set names, + // so use a roundabout getElementsByName test + support.getById = assert( function( el ) { + docElem.appendChild( el ).id = expando; + return !document.getElementsByName || !document.getElementsByName( expando ).length; + } ); + + // ID filter and find + if ( support.getById ) { + Expr.filter[ "ID" ] = function( id ) { + var attrId = id.replace( runescape, funescape ); + return function( elem ) { + return elem.getAttribute( "id" ) === attrId; + }; + }; + Expr.find[ "ID" ] = function( id, context ) { + if ( typeof context.getElementById !== "undefined" && documentIsHTML ) { + var elem = context.getElementById( id ); + return elem ? [ elem ] : []; + } + }; + } else { + Expr.filter[ "ID" ] = function( id ) { + var attrId = id.replace( runescape, funescape ); + return function( elem ) { + var node = typeof elem.getAttributeNode !== "undefined" && + elem.getAttributeNode( "id" ); + return node && node.value === attrId; + }; + }; + + // Support: IE 6 - 7 only + // getElementById is not reliable as a find shortcut + Expr.find[ "ID" ] = function( id, context ) { + if ( typeof context.getElementById !== "undefined" && documentIsHTML ) { + var node, i, elems, + elem = context.getElementById( id ); + + if ( elem ) { + + // Verify the id attribute + node = elem.getAttributeNode( "id" ); + if ( node && node.value === id ) { + return [ elem ]; + } + + // Fall back on getElementsByName + elems = context.getElementsByName( id ); + i = 0; + while ( ( elem = elems[ i++ ] ) ) { + node = elem.getAttributeNode( "id" ); + if ( node && node.value === id ) { + return [ elem ]; + } + } + } + + return []; + } + }; + } + + // Tag + Expr.find[ "TAG" ] = support.getElementsByTagName ? + function( tag, context ) { + if ( typeof context.getElementsByTagName !== "undefined" ) { + return context.getElementsByTagName( tag ); + + // DocumentFragment nodes don't have gEBTN + } else if ( support.qsa ) { + return context.querySelectorAll( tag ); + } + } : + + function( tag, context ) { + var elem, + tmp = [], + i = 0, + + // By happy coincidence, a (broken) gEBTN appears on DocumentFragment nodes too + results = context.getElementsByTagName( tag ); + + // Filter out possible comments + if ( tag === "*" ) { + while ( ( elem = results[ i++ ] ) ) { + if ( elem.nodeType === 1 ) { + tmp.push( elem ); + } + } + + return tmp; + } + return results; + }; + + // Class + Expr.find[ "CLASS" ] = support.getElementsByClassName && function( className, context ) { + if ( typeof context.getElementsByClassName !== "undefined" && documentIsHTML ) { + return context.getElementsByClassName( className ); + } + }; + + /* QSA/matchesSelector + ---------------------------------------------------------------------- */ + + // QSA and matchesSelector support + + // matchesSelector(:active) reports false when true (IE9/Opera 11.5) + rbuggyMatches = []; + + // qSa(:focus) reports false when true (Chrome 21) + // We allow this because of a bug in IE8/9 that throws an error + // whenever `document.activeElement` is accessed on an iframe + // So, we allow :focus to pass through QSA all the time to avoid the IE error + // See https://bugs.jquery.com/ticket/13378 + rbuggyQSA = []; + + if ( ( support.qsa = rnative.test( document.querySelectorAll ) ) ) { + + // Build QSA regex + // Regex strategy adopted from Diego Perini + assert( function( el ) { + + var input; + + // Select is set to empty string on purpose + // This is to test IE's treatment of not explicitly + // setting a boolean content attribute, + // since its presence should be enough + // https://bugs.jquery.com/ticket/12359 + docElem.appendChild( el ).innerHTML = "" + + ""; + + // Support: IE8, Opera 11-12.16 + // Nothing should be selected when empty strings follow ^= or $= or *= + // The test attribute must be unknown in Opera but "safe" for WinRT + // https://msdn.microsoft.com/en-us/library/ie/hh465388.aspx#attribute_section + if ( el.querySelectorAll( "[msallowcapture^='']" ).length ) { + rbuggyQSA.push( "[*^$]=" + whitespace + "*(?:''|\"\")" ); + } + + // Support: IE8 + // Boolean attributes and "value" are not treated correctly + if ( !el.querySelectorAll( "[selected]" ).length ) { + rbuggyQSA.push( "\\[" + whitespace + "*(?:value|" + booleans + ")" ); + } + + // Support: Chrome<29, Android<4.4, Safari<7.0+, iOS<7.0+, PhantomJS<1.9.8+ + if ( !el.querySelectorAll( "[id~=" + expando + "-]" ).length ) { + rbuggyQSA.push( "~=" ); + } + + // Support: IE 11+, Edge 15 - 18+ + // IE 11/Edge don't find elements on a `[name='']` query in some cases. + // Adding a temporary attribute to the document before the selection works + // around the issue. + // Interestingly, IE 10 & older don't seem to have the issue. + input = document.createElement( "input" ); + input.setAttribute( "name", "" ); + el.appendChild( input ); + if ( !el.querySelectorAll( "[name='']" ).length ) { + rbuggyQSA.push( "\\[" + whitespace + "*name" + whitespace + "*=" + + whitespace + "*(?:''|\"\")" ); + } + + // Webkit/Opera - :checked should return selected option elements + // http://www.w3.org/TR/2011/REC-css3-selectors-20110929/#checked + // IE8 throws error here and will not see later tests + if ( !el.querySelectorAll( ":checked" ).length ) { + rbuggyQSA.push( ":checked" ); + } + + // Support: Safari 8+, iOS 8+ + // https://bugs.webkit.org/show_bug.cgi?id=136851 + // In-page `selector#id sibling-combinator selector` fails + if ( !el.querySelectorAll( "a#" + expando + "+*" ).length ) { + rbuggyQSA.push( ".#.+[+~]" ); + } + + // Support: Firefox <=3.6 - 5 only + // Old Firefox doesn't throw on a badly-escaped identifier. + el.querySelectorAll( "\\\f" ); + rbuggyQSA.push( "[\\r\\n\\f]" ); + } ); + + assert( function( el ) { + el.innerHTML = "" + + ""; + + // Support: Windows 8 Native Apps + // The type and name attributes are restricted during .innerHTML assignment + var input = document.createElement( "input" ); + input.setAttribute( "type", "hidden" ); + el.appendChild( input ).setAttribute( "name", "D" ); + + // Support: IE8 + // Enforce case-sensitivity of name attribute + if ( el.querySelectorAll( "[name=d]" ).length ) { + rbuggyQSA.push( "name" + whitespace + "*[*^$|!~]?=" ); + } + + // FF 3.5 - :enabled/:disabled and hidden elements (hidden elements are still enabled) + // IE8 throws error here and will not see later tests + if ( el.querySelectorAll( ":enabled" ).length !== 2 ) { + rbuggyQSA.push( ":enabled", ":disabled" ); + } + + // Support: IE9-11+ + // IE's :disabled selector does not pick up the children of disabled fieldsets + docElem.appendChild( el ).disabled = true; + if ( el.querySelectorAll( ":disabled" ).length !== 2 ) { + rbuggyQSA.push( ":enabled", ":disabled" ); + } + + // Support: Opera 10 - 11 only + // Opera 10-11 does not throw on post-comma invalid pseudos + el.querySelectorAll( "*,:x" ); + rbuggyQSA.push( ",.*:" ); + } ); + } + + if ( ( support.matchesSelector = rnative.test( ( matches = docElem.matches || + docElem.webkitMatchesSelector || + docElem.mozMatchesSelector || + docElem.oMatchesSelector || + docElem.msMatchesSelector ) ) ) ) { + + assert( function( el ) { + + // Check to see if it's possible to do matchesSelector + // on a disconnected node (IE 9) + support.disconnectedMatch = matches.call( el, "*" ); + + // This should fail with an exception + // Gecko does not error, returns false instead + matches.call( el, "[s!='']:x" ); + rbuggyMatches.push( "!=", pseudos ); + } ); + } + + rbuggyQSA = rbuggyQSA.length && new RegExp( rbuggyQSA.join( "|" ) ); + rbuggyMatches = rbuggyMatches.length && new RegExp( rbuggyMatches.join( "|" ) ); + + /* Contains + ---------------------------------------------------------------------- */ + hasCompare = rnative.test( docElem.compareDocumentPosition ); + + // Element contains another + // Purposefully self-exclusive + // As in, an element does not contain itself + contains = hasCompare || rnative.test( docElem.contains ) ? + function( a, b ) { + var adown = a.nodeType === 9 ? a.documentElement : a, + bup = b && b.parentNode; + return a === bup || !!( bup && bup.nodeType === 1 && ( + adown.contains ? + adown.contains( bup ) : + a.compareDocumentPosition && a.compareDocumentPosition( bup ) & 16 + ) ); + } : + function( a, b ) { + if ( b ) { + while ( ( b = b.parentNode ) ) { + if ( b === a ) { + return true; + } + } + } + return false; + }; + + /* Sorting + ---------------------------------------------------------------------- */ + + // Document order sorting + sortOrder = hasCompare ? + function( a, b ) { + + // Flag for duplicate removal + if ( a === b ) { + hasDuplicate = true; + return 0; + } + + // Sort on method existence if only one input has compareDocumentPosition + var compare = !a.compareDocumentPosition - !b.compareDocumentPosition; + if ( compare ) { + return compare; + } + + // Calculate position if both inputs belong to the same document + // Support: IE 11+, Edge 17 - 18+ + // IE/Edge sometimes throw a "Permission denied" error when strict-comparing + // two documents; shallow comparisons work. + // eslint-disable-next-line eqeqeq + compare = ( a.ownerDocument || a ) == ( b.ownerDocument || b ) ? + a.compareDocumentPosition( b ) : + + // Otherwise we know they are disconnected + 1; + + // Disconnected nodes + if ( compare & 1 || + ( !support.sortDetached && b.compareDocumentPosition( a ) === compare ) ) { + + // Choose the first element that is related to our preferred document + // Support: IE 11+, Edge 17 - 18+ + // IE/Edge sometimes throw a "Permission denied" error when strict-comparing + // two documents; shallow comparisons work. + // eslint-disable-next-line eqeqeq + if ( a == document || a.ownerDocument == preferredDoc && + contains( preferredDoc, a ) ) { + return -1; + } + + // Support: IE 11+, Edge 17 - 18+ + // IE/Edge sometimes throw a "Permission denied" error when strict-comparing + // two documents; shallow comparisons work. + // eslint-disable-next-line eqeqeq + if ( b == document || b.ownerDocument == preferredDoc && + contains( preferredDoc, b ) ) { + return 1; + } + + // Maintain original order + return sortInput ? + ( indexOf( sortInput, a ) - indexOf( sortInput, b ) ) : + 0; + } + + return compare & 4 ? -1 : 1; + } : + function( a, b ) { + + // Exit early if the nodes are identical + if ( a === b ) { + hasDuplicate = true; + return 0; + } + + var cur, + i = 0, + aup = a.parentNode, + bup = b.parentNode, + ap = [ a ], + bp = [ b ]; + + // Parentless nodes are either documents or disconnected + if ( !aup || !bup ) { + + // Support: IE 11+, Edge 17 - 18+ + // IE/Edge sometimes throw a "Permission denied" error when strict-comparing + // two documents; shallow comparisons work. + /* eslint-disable eqeqeq */ + return a == document ? -1 : + b == document ? 1 : + /* eslint-enable eqeqeq */ + aup ? -1 : + bup ? 1 : + sortInput ? + ( indexOf( sortInput, a ) - indexOf( sortInput, b ) ) : + 0; + + // If the nodes are siblings, we can do a quick check + } else if ( aup === bup ) { + return siblingCheck( a, b ); + } + + // Otherwise we need full lists of their ancestors for comparison + cur = a; + while ( ( cur = cur.parentNode ) ) { + ap.unshift( cur ); + } + cur = b; + while ( ( cur = cur.parentNode ) ) { + bp.unshift( cur ); + } + + // Walk down the tree looking for a discrepancy + while ( ap[ i ] === bp[ i ] ) { + i++; + } + + return i ? + + // Do a sibling check if the nodes have a common ancestor + siblingCheck( ap[ i ], bp[ i ] ) : + + // Otherwise nodes in our document sort first + // Support: IE 11+, Edge 17 - 18+ + // IE/Edge sometimes throw a "Permission denied" error when strict-comparing + // two documents; shallow comparisons work. + /* eslint-disable eqeqeq */ + ap[ i ] == preferredDoc ? -1 : + bp[ i ] == preferredDoc ? 1 : + /* eslint-enable eqeqeq */ + 0; + }; + + return document; +}; + +Sizzle.matches = function( expr, elements ) { + return Sizzle( expr, null, null, elements ); +}; + +Sizzle.matchesSelector = function( elem, expr ) { + setDocument( elem ); + + if ( support.matchesSelector && documentIsHTML && + !nonnativeSelectorCache[ expr + " " ] && + ( !rbuggyMatches || !rbuggyMatches.test( expr ) ) && + ( !rbuggyQSA || !rbuggyQSA.test( expr ) ) ) { + + try { + var ret = matches.call( elem, expr ); + + // IE 9's matchesSelector returns false on disconnected nodes + if ( ret || support.disconnectedMatch || + + // As well, disconnected nodes are said to be in a document + // fragment in IE 9 + elem.document && elem.document.nodeType !== 11 ) { + return ret; + } + } catch ( e ) { + nonnativeSelectorCache( expr, true ); + } + } + + return Sizzle( expr, document, null, [ elem ] ).length > 0; +}; + +Sizzle.contains = function( context, elem ) { + + // Set document vars if needed + // Support: IE 11+, Edge 17 - 18+ + // IE/Edge sometimes throw a "Permission denied" error when strict-comparing + // two documents; shallow comparisons work. + // eslint-disable-next-line eqeqeq + if ( ( context.ownerDocument || context ) != document ) { + setDocument( context ); + } + return contains( context, elem ); +}; + +Sizzle.attr = function( elem, name ) { + + // Set document vars if needed + // Support: IE 11+, Edge 17 - 18+ + // IE/Edge sometimes throw a "Permission denied" error when strict-comparing + // two documents; shallow comparisons work. + // eslint-disable-next-line eqeqeq + if ( ( elem.ownerDocument || elem ) != document ) { + setDocument( elem ); + } + + var fn = Expr.attrHandle[ name.toLowerCase() ], + + // Don't get fooled by Object.prototype properties (jQuery #13807) + val = fn && hasOwn.call( Expr.attrHandle, name.toLowerCase() ) ? + fn( elem, name, !documentIsHTML ) : + undefined; + + return val !== undefined ? + val : + support.attributes || !documentIsHTML ? + elem.getAttribute( name ) : + ( val = elem.getAttributeNode( name ) ) && val.specified ? + val.value : + null; +}; + +Sizzle.escape = function( sel ) { + return ( sel + "" ).replace( rcssescape, fcssescape ); +}; + +Sizzle.error = function( msg ) { + throw new Error( "Syntax error, unrecognized expression: " + msg ); +}; + +/** + * Document sorting and removing duplicates + * @param {ArrayLike} results + */ +Sizzle.uniqueSort = function( results ) { + var elem, + duplicates = [], + j = 0, + i = 0; + + // Unless we *know* we can detect duplicates, assume their presence + hasDuplicate = !support.detectDuplicates; + sortInput = !support.sortStable && results.slice( 0 ); + results.sort( sortOrder ); + + if ( hasDuplicate ) { + while ( ( elem = results[ i++ ] ) ) { + if ( elem === results[ i ] ) { + j = duplicates.push( i ); + } + } + while ( j-- ) { + results.splice( duplicates[ j ], 1 ); + } + } + + // Clear input after sorting to release objects + // See https://github.com/jquery/sizzle/pull/225 + sortInput = null; + + return results; +}; + +/** + * Utility function for retrieving the text value of an array of DOM nodes + * @param {Array|Element} elem + */ +getText = Sizzle.getText = function( elem ) { + var node, + ret = "", + i = 0, + nodeType = elem.nodeType; + + if ( !nodeType ) { + + // If no nodeType, this is expected to be an array + while ( ( node = elem[ i++ ] ) ) { + + // Do not traverse comment nodes + ret += getText( node ); + } + } else if ( nodeType === 1 || nodeType === 9 || nodeType === 11 ) { + + // Use textContent for elements + // innerText usage removed for consistency of new lines (jQuery #11153) + if ( typeof elem.textContent === "string" ) { + return elem.textContent; + } else { + + // Traverse its children + for ( elem = elem.firstChild; elem; elem = elem.nextSibling ) { + ret += getText( elem ); + } + } + } else if ( nodeType === 3 || nodeType === 4 ) { + return elem.nodeValue; + } + + // Do not include comment or processing instruction nodes + + return ret; +}; + +Expr = Sizzle.selectors = { + + // Can be adjusted by the user + cacheLength: 50, + + createPseudo: markFunction, + + match: matchExpr, + + attrHandle: {}, + + find: {}, + + relative: { + ">": { dir: "parentNode", first: true }, + " ": { dir: "parentNode" }, + "+": { dir: "previousSibling", first: true }, + "~": { dir: "previousSibling" } + }, + + preFilter: { + "ATTR": function( match ) { + match[ 1 ] = match[ 1 ].replace( runescape, funescape ); + + // Move the given value to match[3] whether quoted or unquoted + match[ 3 ] = ( match[ 3 ] || match[ 4 ] || + match[ 5 ] || "" ).replace( runescape, funescape ); + + if ( match[ 2 ] === "~=" ) { + match[ 3 ] = " " + match[ 3 ] + " "; + } + + return match.slice( 0, 4 ); + }, + + "CHILD": function( match ) { + + /* matches from matchExpr["CHILD"] + 1 type (only|nth|...) + 2 what (child|of-type) + 3 argument (even|odd|\d*|\d*n([+-]\d+)?|...) + 4 xn-component of xn+y argument ([+-]?\d*n|) + 5 sign of xn-component + 6 x of xn-component + 7 sign of y-component + 8 y of y-component + */ + match[ 1 ] = match[ 1 ].toLowerCase(); + + if ( match[ 1 ].slice( 0, 3 ) === "nth" ) { + + // nth-* requires argument + if ( !match[ 3 ] ) { + Sizzle.error( match[ 0 ] ); + } + + // numeric x and y parameters for Expr.filter.CHILD + // remember that false/true cast respectively to 0/1 + match[ 4 ] = +( match[ 4 ] ? + match[ 5 ] + ( match[ 6 ] || 1 ) : + 2 * ( match[ 3 ] === "even" || match[ 3 ] === "odd" ) ); + match[ 5 ] = +( ( match[ 7 ] + match[ 8 ] ) || match[ 3 ] === "odd" ); + + // other types prohibit arguments + } else if ( match[ 3 ] ) { + Sizzle.error( match[ 0 ] ); + } + + return match; + }, + + "PSEUDO": function( match ) { + var excess, + unquoted = !match[ 6 ] && match[ 2 ]; + + if ( matchExpr[ "CHILD" ].test( match[ 0 ] ) ) { + return null; + } + + // Accept quoted arguments as-is + if ( match[ 3 ] ) { + match[ 2 ] = match[ 4 ] || match[ 5 ] || ""; + + // Strip excess characters from unquoted arguments + } else if ( unquoted && rpseudo.test( unquoted ) && + + // Get excess from tokenize (recursively) + ( excess = tokenize( unquoted, true ) ) && + + // advance to the next closing parenthesis + ( excess = unquoted.indexOf( ")", unquoted.length - excess ) - unquoted.length ) ) { + + // excess is a negative index + match[ 0 ] = match[ 0 ].slice( 0, excess ); + match[ 2 ] = unquoted.slice( 0, excess ); + } + + // Return only captures needed by the pseudo filter method (type and argument) + return match.slice( 0, 3 ); + } + }, + + filter: { + + "TAG": function( nodeNameSelector ) { + var nodeName = nodeNameSelector.replace( runescape, funescape ).toLowerCase(); + return nodeNameSelector === "*" ? + function() { + return true; + } : + function( elem ) { + return elem.nodeName && elem.nodeName.toLowerCase() === nodeName; + }; + }, + + "CLASS": function( className ) { + var pattern = classCache[ className + " " ]; + + return pattern || + ( pattern = new RegExp( "(^|" + whitespace + + ")" + className + "(" + whitespace + "|$)" ) ) && classCache( + className, function( elem ) { + return pattern.test( + typeof elem.className === "string" && elem.className || + typeof elem.getAttribute !== "undefined" && + elem.getAttribute( "class" ) || + "" + ); + } ); + }, + + "ATTR": function( name, operator, check ) { + return function( elem ) { + var result = Sizzle.attr( elem, name ); + + if ( result == null ) { + return operator === "!="; + } + if ( !operator ) { + return true; + } + + result += ""; + + /* eslint-disable max-len */ + + return operator === "=" ? result === check : + operator === "!=" ? result !== check : + operator === "^=" ? check && result.indexOf( check ) === 0 : + operator === "*=" ? check && result.indexOf( check ) > -1 : + operator === "$=" ? check && result.slice( -check.length ) === check : + operator === "~=" ? ( " " + result.replace( rwhitespace, " " ) + " " ).indexOf( check ) > -1 : + operator === "|=" ? result === check || result.slice( 0, check.length + 1 ) === check + "-" : + false; + /* eslint-enable max-len */ + + }; + }, + + "CHILD": function( type, what, _argument, first, last ) { + var simple = type.slice( 0, 3 ) !== "nth", + forward = type.slice( -4 ) !== "last", + ofType = what === "of-type"; + + return first === 1 && last === 0 ? + + // Shortcut for :nth-*(n) + function( elem ) { + return !!elem.parentNode; + } : + + function( elem, _context, xml ) { + var cache, uniqueCache, outerCache, node, nodeIndex, start, + dir = simple !== forward ? "nextSibling" : "previousSibling", + parent = elem.parentNode, + name = ofType && elem.nodeName.toLowerCase(), + useCache = !xml && !ofType, + diff = false; + + if ( parent ) { + + // :(first|last|only)-(child|of-type) + if ( simple ) { + while ( dir ) { + node = elem; + while ( ( node = node[ dir ] ) ) { + if ( ofType ? + node.nodeName.toLowerCase() === name : + node.nodeType === 1 ) { + + return false; + } + } + + // Reverse direction for :only-* (if we haven't yet done so) + start = dir = type === "only" && !start && "nextSibling"; + } + return true; + } + + start = [ forward ? parent.firstChild : parent.lastChild ]; + + // non-xml :nth-child(...) stores cache data on `parent` + if ( forward && useCache ) { + + // Seek `elem` from a previously-cached index + + // ...in a gzip-friendly way + node = parent; + outerCache = node[ expando ] || ( node[ expando ] = {} ); + + // Support: IE <9 only + // Defend against cloned attroperties (jQuery gh-1709) + uniqueCache = outerCache[ node.uniqueID ] || + ( outerCache[ node.uniqueID ] = {} ); + + cache = uniqueCache[ type ] || []; + nodeIndex = cache[ 0 ] === dirruns && cache[ 1 ]; + diff = nodeIndex && cache[ 2 ]; + node = nodeIndex && parent.childNodes[ nodeIndex ]; + + while ( ( node = ++nodeIndex && node && node[ dir ] || + + // Fallback to seeking `elem` from the start + ( diff = nodeIndex = 0 ) || start.pop() ) ) { + + // When found, cache indexes on `parent` and break + if ( node.nodeType === 1 && ++diff && node === elem ) { + uniqueCache[ type ] = [ dirruns, nodeIndex, diff ]; + break; + } + } + + } else { + + // Use previously-cached element index if available + if ( useCache ) { + + // ...in a gzip-friendly way + node = elem; + outerCache = node[ expando ] || ( node[ expando ] = {} ); + + // Support: IE <9 only + // Defend against cloned attroperties (jQuery gh-1709) + uniqueCache = outerCache[ node.uniqueID ] || + ( outerCache[ node.uniqueID ] = {} ); + + cache = uniqueCache[ type ] || []; + nodeIndex = cache[ 0 ] === dirruns && cache[ 1 ]; + diff = nodeIndex; + } + + // xml :nth-child(...) + // or :nth-last-child(...) or :nth(-last)?-of-type(...) + if ( diff === false ) { + + // Use the same loop as above to seek `elem` from the start + while ( ( node = ++nodeIndex && node && node[ dir ] || + ( diff = nodeIndex = 0 ) || start.pop() ) ) { + + if ( ( ofType ? + node.nodeName.toLowerCase() === name : + node.nodeType === 1 ) && + ++diff ) { + + // Cache the index of each encountered element + if ( useCache ) { + outerCache = node[ expando ] || + ( node[ expando ] = {} ); + + // Support: IE <9 only + // Defend against cloned attroperties (jQuery gh-1709) + uniqueCache = outerCache[ node.uniqueID ] || + ( outerCache[ node.uniqueID ] = {} ); + + uniqueCache[ type ] = [ dirruns, diff ]; + } + + if ( node === elem ) { + break; + } + } + } + } + } + + // Incorporate the offset, then check against cycle size + diff -= last; + return diff === first || ( diff % first === 0 && diff / first >= 0 ); + } + }; + }, + + "PSEUDO": function( pseudo, argument ) { + + // pseudo-class names are case-insensitive + // http://www.w3.org/TR/selectors/#pseudo-classes + // Prioritize by case sensitivity in case custom pseudos are added with uppercase letters + // Remember that setFilters inherits from pseudos + var args, + fn = Expr.pseudos[ pseudo ] || Expr.setFilters[ pseudo.toLowerCase() ] || + Sizzle.error( "unsupported pseudo: " + pseudo ); + + // The user may use createPseudo to indicate that + // arguments are needed to create the filter function + // just as Sizzle does + if ( fn[ expando ] ) { + return fn( argument ); + } + + // But maintain support for old signatures + if ( fn.length > 1 ) { + args = [ pseudo, pseudo, "", argument ]; + return Expr.setFilters.hasOwnProperty( pseudo.toLowerCase() ) ? + markFunction( function( seed, matches ) { + var idx, + matched = fn( seed, argument ), + i = matched.length; + while ( i-- ) { + idx = indexOf( seed, matched[ i ] ); + seed[ idx ] = !( matches[ idx ] = matched[ i ] ); + } + } ) : + function( elem ) { + return fn( elem, 0, args ); + }; + } + + return fn; + } + }, + + pseudos: { + + // Potentially complex pseudos + "not": markFunction( function( selector ) { + + // Trim the selector passed to compile + // to avoid treating leading and trailing + // spaces as combinators + var input = [], + results = [], + matcher = compile( selector.replace( rtrim, "$1" ) ); + + return matcher[ expando ] ? + markFunction( function( seed, matches, _context, xml ) { + var elem, + unmatched = matcher( seed, null, xml, [] ), + i = seed.length; + + // Match elements unmatched by `matcher` + while ( i-- ) { + if ( ( elem = unmatched[ i ] ) ) { + seed[ i ] = !( matches[ i ] = elem ); + } + } + } ) : + function( elem, _context, xml ) { + input[ 0 ] = elem; + matcher( input, null, xml, results ); + + // Don't keep the element (issue #299) + input[ 0 ] = null; + return !results.pop(); + }; + } ), + + "has": markFunction( function( selector ) { + return function( elem ) { + return Sizzle( selector, elem ).length > 0; + }; + } ), + + "contains": markFunction( function( text ) { + text = text.replace( runescape, funescape ); + return function( elem ) { + return ( elem.textContent || getText( elem ) ).indexOf( text ) > -1; + }; + } ), + + // "Whether an element is represented by a :lang() selector + // is based solely on the element's language value + // being equal to the identifier C, + // or beginning with the identifier C immediately followed by "-". + // The matching of C against the element's language value is performed case-insensitively. + // The identifier C does not have to be a valid language name." + // http://www.w3.org/TR/selectors/#lang-pseudo + "lang": markFunction( function( lang ) { + + // lang value must be a valid identifier + if ( !ridentifier.test( lang || "" ) ) { + Sizzle.error( "unsupported lang: " + lang ); + } + lang = lang.replace( runescape, funescape ).toLowerCase(); + return function( elem ) { + var elemLang; + do { + if ( ( elemLang = documentIsHTML ? + elem.lang : + elem.getAttribute( "xml:lang" ) || elem.getAttribute( "lang" ) ) ) { + + elemLang = elemLang.toLowerCase(); + return elemLang === lang || elemLang.indexOf( lang + "-" ) === 0; + } + } while ( ( elem = elem.parentNode ) && elem.nodeType === 1 ); + return false; + }; + } ), + + // Miscellaneous + "target": function( elem ) { + var hash = window.location && window.location.hash; + return hash && hash.slice( 1 ) === elem.id; + }, + + "root": function( elem ) { + return elem === docElem; + }, + + "focus": function( elem ) { + return elem === document.activeElement && + ( !document.hasFocus || document.hasFocus() ) && + !!( elem.type || elem.href || ~elem.tabIndex ); + }, + + // Boolean properties + "enabled": createDisabledPseudo( false ), + "disabled": createDisabledPseudo( true ), + + "checked": function( elem ) { + + // In CSS3, :checked should return both checked and selected elements + // http://www.w3.org/TR/2011/REC-css3-selectors-20110929/#checked + var nodeName = elem.nodeName.toLowerCase(); + return ( nodeName === "input" && !!elem.checked ) || + ( nodeName === "option" && !!elem.selected ); + }, + + "selected": function( elem ) { + + // Accessing this property makes selected-by-default + // options in Safari work properly + if ( elem.parentNode ) { + // eslint-disable-next-line no-unused-expressions + elem.parentNode.selectedIndex; + } + + return elem.selected === true; + }, + + // Contents + "empty": function( elem ) { + + // http://www.w3.org/TR/selectors/#empty-pseudo + // :empty is negated by element (1) or content nodes (text: 3; cdata: 4; entity ref: 5), + // but not by others (comment: 8; processing instruction: 7; etc.) + // nodeType < 6 works because attributes (2) do not appear as children + for ( elem = elem.firstChild; elem; elem = elem.nextSibling ) { + if ( elem.nodeType < 6 ) { + return false; + } + } + return true; + }, + + "parent": function( elem ) { + return !Expr.pseudos[ "empty" ]( elem ); + }, + + // Element/input types + "header": function( elem ) { + return rheader.test( elem.nodeName ); + }, + + "input": function( elem ) { + return rinputs.test( elem.nodeName ); + }, + + "button": function( elem ) { + var name = elem.nodeName.toLowerCase(); + return name === "input" && elem.type === "button" || name === "button"; + }, + + "text": function( elem ) { + var attr; + return elem.nodeName.toLowerCase() === "input" && + elem.type === "text" && + + // Support: IE<8 + // New HTML5 attribute values (e.g., "search") appear with elem.type === "text" + ( ( attr = elem.getAttribute( "type" ) ) == null || + attr.toLowerCase() === "text" ); + }, + + // Position-in-collection + "first": createPositionalPseudo( function() { + return [ 0 ]; + } ), + + "last": createPositionalPseudo( function( _matchIndexes, length ) { + return [ length - 1 ]; + } ), + + "eq": createPositionalPseudo( function( _matchIndexes, length, argument ) { + return [ argument < 0 ? argument + length : argument ]; + } ), + + "even": createPositionalPseudo( function( matchIndexes, length ) { + var i = 0; + for ( ; i < length; i += 2 ) { + matchIndexes.push( i ); + } + return matchIndexes; + } ), + + "odd": createPositionalPseudo( function( matchIndexes, length ) { + var i = 1; + for ( ; i < length; i += 2 ) { + matchIndexes.push( i ); + } + return matchIndexes; + } ), + + "lt": createPositionalPseudo( function( matchIndexes, length, argument ) { + var i = argument < 0 ? + argument + length : + argument > length ? + length : + argument; + for ( ; --i >= 0; ) { + matchIndexes.push( i ); + } + return matchIndexes; + } ), + + "gt": createPositionalPseudo( function( matchIndexes, length, argument ) { + var i = argument < 0 ? argument + length : argument; + for ( ; ++i < length; ) { + matchIndexes.push( i ); + } + return matchIndexes; + } ) + } +}; + +Expr.pseudos[ "nth" ] = Expr.pseudos[ "eq" ]; + +// Add button/input type pseudos +for ( i in { radio: true, checkbox: true, file: true, password: true, image: true } ) { + Expr.pseudos[ i ] = createInputPseudo( i ); +} +for ( i in { submit: true, reset: true } ) { + Expr.pseudos[ i ] = createButtonPseudo( i ); +} + +// Easy API for creating new setFilters +function setFilters() {} +setFilters.prototype = Expr.filters = Expr.pseudos; +Expr.setFilters = new setFilters(); + +tokenize = Sizzle.tokenize = function( selector, parseOnly ) { + var matched, match, tokens, type, + soFar, groups, preFilters, + cached = tokenCache[ selector + " " ]; + + if ( cached ) { + return parseOnly ? 0 : cached.slice( 0 ); + } + + soFar = selector; + groups = []; + preFilters = Expr.preFilter; + + while ( soFar ) { + + // Comma and first run + if ( !matched || ( match = rcomma.exec( soFar ) ) ) { + if ( match ) { + + // Don't consume trailing commas as valid + soFar = soFar.slice( match[ 0 ].length ) || soFar; + } + groups.push( ( tokens = [] ) ); + } + + matched = false; + + // Combinators + if ( ( match = rcombinators.exec( soFar ) ) ) { + matched = match.shift(); + tokens.push( { + value: matched, + + // Cast descendant combinators to space + type: match[ 0 ].replace( rtrim, " " ) + } ); + soFar = soFar.slice( matched.length ); + } + + // Filters + for ( type in Expr.filter ) { + if ( ( match = matchExpr[ type ].exec( soFar ) ) && ( !preFilters[ type ] || + ( match = preFilters[ type ]( match ) ) ) ) { + matched = match.shift(); + tokens.push( { + value: matched, + type: type, + matches: match + } ); + soFar = soFar.slice( matched.length ); + } + } + + if ( !matched ) { + break; + } + } + + // Return the length of the invalid excess + // if we're just parsing + // Otherwise, throw an error or return tokens + return parseOnly ? + soFar.length : + soFar ? + Sizzle.error( selector ) : + + // Cache the tokens + tokenCache( selector, groups ).slice( 0 ); +}; + +function toSelector( tokens ) { + var i = 0, + len = tokens.length, + selector = ""; + for ( ; i < len; i++ ) { + selector += tokens[ i ].value; + } + return selector; +} + +function addCombinator( matcher, combinator, base ) { + var dir = combinator.dir, + skip = combinator.next, + key = skip || dir, + checkNonElements = base && key === "parentNode", + doneName = done++; + + return combinator.first ? + + // Check against closest ancestor/preceding element + function( elem, context, xml ) { + while ( ( elem = elem[ dir ] ) ) { + if ( elem.nodeType === 1 || checkNonElements ) { + return matcher( elem, context, xml ); + } + } + return false; + } : + + // Check against all ancestor/preceding elements + function( elem, context, xml ) { + var oldCache, uniqueCache, outerCache, + newCache = [ dirruns, doneName ]; + + // We can't set arbitrary data on XML nodes, so they don't benefit from combinator caching + if ( xml ) { + while ( ( elem = elem[ dir ] ) ) { + if ( elem.nodeType === 1 || checkNonElements ) { + if ( matcher( elem, context, xml ) ) { + return true; + } + } + } + } else { + while ( ( elem = elem[ dir ] ) ) { + if ( elem.nodeType === 1 || checkNonElements ) { + outerCache = elem[ expando ] || ( elem[ expando ] = {} ); + + // Support: IE <9 only + // Defend against cloned attroperties (jQuery gh-1709) + uniqueCache = outerCache[ elem.uniqueID ] || + ( outerCache[ elem.uniqueID ] = {} ); + + if ( skip && skip === elem.nodeName.toLowerCase() ) { + elem = elem[ dir ] || elem; + } else if ( ( oldCache = uniqueCache[ key ] ) && + oldCache[ 0 ] === dirruns && oldCache[ 1 ] === doneName ) { + + // Assign to newCache so results back-propagate to previous elements + return ( newCache[ 2 ] = oldCache[ 2 ] ); + } else { + + // Reuse newcache so results back-propagate to previous elements + uniqueCache[ key ] = newCache; + + // A match means we're done; a fail means we have to keep checking + if ( ( newCache[ 2 ] = matcher( elem, context, xml ) ) ) { + return true; + } + } + } + } + } + return false; + }; +} + +function elementMatcher( matchers ) { + return matchers.length > 1 ? + function( elem, context, xml ) { + var i = matchers.length; + while ( i-- ) { + if ( !matchers[ i ]( elem, context, xml ) ) { + return false; + } + } + return true; + } : + matchers[ 0 ]; +} + +function multipleContexts( selector, contexts, results ) { + var i = 0, + len = contexts.length; + for ( ; i < len; i++ ) { + Sizzle( selector, contexts[ i ], results ); + } + return results; +} + +function condense( unmatched, map, filter, context, xml ) { + var elem, + newUnmatched = [], + i = 0, + len = unmatched.length, + mapped = map != null; + + for ( ; i < len; i++ ) { + if ( ( elem = unmatched[ i ] ) ) { + if ( !filter || filter( elem, context, xml ) ) { + newUnmatched.push( elem ); + if ( mapped ) { + map.push( i ); + } + } + } + } + + return newUnmatched; +} + +function setMatcher( preFilter, selector, matcher, postFilter, postFinder, postSelector ) { + if ( postFilter && !postFilter[ expando ] ) { + postFilter = setMatcher( postFilter ); + } + if ( postFinder && !postFinder[ expando ] ) { + postFinder = setMatcher( postFinder, postSelector ); + } + return markFunction( function( seed, results, context, xml ) { + var temp, i, elem, + preMap = [], + postMap = [], + preexisting = results.length, + + // Get initial elements from seed or context + elems = seed || multipleContexts( + selector || "*", + context.nodeType ? [ context ] : context, + [] + ), + + // Prefilter to get matcher input, preserving a map for seed-results synchronization + matcherIn = preFilter && ( seed || !selector ) ? + condense( elems, preMap, preFilter, context, xml ) : + elems, + + matcherOut = matcher ? + + // If we have a postFinder, or filtered seed, or non-seed postFilter or preexisting results, + postFinder || ( seed ? preFilter : preexisting || postFilter ) ? + + // ...intermediate processing is necessary + [] : + + // ...otherwise use results directly + results : + matcherIn; + + // Find primary matches + if ( matcher ) { + matcher( matcherIn, matcherOut, context, xml ); + } + + // Apply postFilter + if ( postFilter ) { + temp = condense( matcherOut, postMap ); + postFilter( temp, [], context, xml ); + + // Un-match failing elements by moving them back to matcherIn + i = temp.length; + while ( i-- ) { + if ( ( elem = temp[ i ] ) ) { + matcherOut[ postMap[ i ] ] = !( matcherIn[ postMap[ i ] ] = elem ); + } + } + } + + if ( seed ) { + if ( postFinder || preFilter ) { + if ( postFinder ) { + + // Get the final matcherOut by condensing this intermediate into postFinder contexts + temp = []; + i = matcherOut.length; + while ( i-- ) { + if ( ( elem = matcherOut[ i ] ) ) { + + // Restore matcherIn since elem is not yet a final match + temp.push( ( matcherIn[ i ] = elem ) ); + } + } + postFinder( null, ( matcherOut = [] ), temp, xml ); + } + + // Move matched elements from seed to results to keep them synchronized + i = matcherOut.length; + while ( i-- ) { + if ( ( elem = matcherOut[ i ] ) && + ( temp = postFinder ? indexOf( seed, elem ) : preMap[ i ] ) > -1 ) { + + seed[ temp ] = !( results[ temp ] = elem ); + } + } + } + + // Add elements to results, through postFinder if defined + } else { + matcherOut = condense( + matcherOut === results ? + matcherOut.splice( preexisting, matcherOut.length ) : + matcherOut + ); + if ( postFinder ) { + postFinder( null, results, matcherOut, xml ); + } else { + push.apply( results, matcherOut ); + } + } + } ); +} + +function matcherFromTokens( tokens ) { + var checkContext, matcher, j, + len = tokens.length, + leadingRelative = Expr.relative[ tokens[ 0 ].type ], + implicitRelative = leadingRelative || Expr.relative[ " " ], + i = leadingRelative ? 1 : 0, + + // The foundational matcher ensures that elements are reachable from top-level context(s) + matchContext = addCombinator( function( elem ) { + return elem === checkContext; + }, implicitRelative, true ), + matchAnyContext = addCombinator( function( elem ) { + return indexOf( checkContext, elem ) > -1; + }, implicitRelative, true ), + matchers = [ function( elem, context, xml ) { + var ret = ( !leadingRelative && ( xml || context !== outermostContext ) ) || ( + ( checkContext = context ).nodeType ? + matchContext( elem, context, xml ) : + matchAnyContext( elem, context, xml ) ); + + // Avoid hanging onto element (issue #299) + checkContext = null; + return ret; + } ]; + + for ( ; i < len; i++ ) { + if ( ( matcher = Expr.relative[ tokens[ i ].type ] ) ) { + matchers = [ addCombinator( elementMatcher( matchers ), matcher ) ]; + } else { + matcher = Expr.filter[ tokens[ i ].type ].apply( null, tokens[ i ].matches ); + + // Return special upon seeing a positional matcher + if ( matcher[ expando ] ) { + + // Find the next relative operator (if any) for proper handling + j = ++i; + for ( ; j < len; j++ ) { + if ( Expr.relative[ tokens[ j ].type ] ) { + break; + } + } + return setMatcher( + i > 1 && elementMatcher( matchers ), + i > 1 && toSelector( + + // If the preceding token was a descendant combinator, insert an implicit any-element `*` + tokens + .slice( 0, i - 1 ) + .concat( { value: tokens[ i - 2 ].type === " " ? "*" : "" } ) + ).replace( rtrim, "$1" ), + matcher, + i < j && matcherFromTokens( tokens.slice( i, j ) ), + j < len && matcherFromTokens( ( tokens = tokens.slice( j ) ) ), + j < len && toSelector( tokens ) + ); + } + matchers.push( matcher ); + } + } + + return elementMatcher( matchers ); +} + +function matcherFromGroupMatchers( elementMatchers, setMatchers ) { + var bySet = setMatchers.length > 0, + byElement = elementMatchers.length > 0, + superMatcher = function( seed, context, xml, results, outermost ) { + var elem, j, matcher, + matchedCount = 0, + i = "0", + unmatched = seed && [], + setMatched = [], + contextBackup = outermostContext, + + // We must always have either seed elements or outermost context + elems = seed || byElement && Expr.find[ "TAG" ]( "*", outermost ), + + // Use integer dirruns iff this is the outermost matcher + dirrunsUnique = ( dirruns += contextBackup == null ? 1 : Math.random() || 0.1 ), + len = elems.length; + + if ( outermost ) { + + // Support: IE 11+, Edge 17 - 18+ + // IE/Edge sometimes throw a "Permission denied" error when strict-comparing + // two documents; shallow comparisons work. + // eslint-disable-next-line eqeqeq + outermostContext = context == document || context || outermost; + } + + // Add elements passing elementMatchers directly to results + // Support: IE<9, Safari + // Tolerate NodeList properties (IE: "length"; Safari: ) matching elements by id + for ( ; i !== len && ( elem = elems[ i ] ) != null; i++ ) { + if ( byElement && elem ) { + j = 0; + + // Support: IE 11+, Edge 17 - 18+ + // IE/Edge sometimes throw a "Permission denied" error when strict-comparing + // two documents; shallow comparisons work. + // eslint-disable-next-line eqeqeq + if ( !context && elem.ownerDocument != document ) { + setDocument( elem ); + xml = !documentIsHTML; + } + while ( ( matcher = elementMatchers[ j++ ] ) ) { + if ( matcher( elem, context || document, xml ) ) { + results.push( elem ); + break; + } + } + if ( outermost ) { + dirruns = dirrunsUnique; + } + } + + // Track unmatched elements for set filters + if ( bySet ) { + + // They will have gone through all possible matchers + if ( ( elem = !matcher && elem ) ) { + matchedCount--; + } + + // Lengthen the array for every element, matched or not + if ( seed ) { + unmatched.push( elem ); + } + } + } + + // `i` is now the count of elements visited above, and adding it to `matchedCount` + // makes the latter nonnegative. + matchedCount += i; + + // Apply set filters to unmatched elements + // NOTE: This can be skipped if there are no unmatched elements (i.e., `matchedCount` + // equals `i`), unless we didn't visit _any_ elements in the above loop because we have + // no element matchers and no seed. + // Incrementing an initially-string "0" `i` allows `i` to remain a string only in that + // case, which will result in a "00" `matchedCount` that differs from `i` but is also + // numerically zero. + if ( bySet && i !== matchedCount ) { + j = 0; + while ( ( matcher = setMatchers[ j++ ] ) ) { + matcher( unmatched, setMatched, context, xml ); + } + + if ( seed ) { + + // Reintegrate element matches to eliminate the need for sorting + if ( matchedCount > 0 ) { + while ( i-- ) { + if ( !( unmatched[ i ] || setMatched[ i ] ) ) { + setMatched[ i ] = pop.call( results ); + } + } + } + + // Discard index placeholder values to get only actual matches + setMatched = condense( setMatched ); + } + + // Add matches to results + push.apply( results, setMatched ); + + // Seedless set matches succeeding multiple successful matchers stipulate sorting + if ( outermost && !seed && setMatched.length > 0 && + ( matchedCount + setMatchers.length ) > 1 ) { + + Sizzle.uniqueSort( results ); + } + } + + // Override manipulation of globals by nested matchers + if ( outermost ) { + dirruns = dirrunsUnique; + outermostContext = contextBackup; + } + + return unmatched; + }; + + return bySet ? + markFunction( superMatcher ) : + superMatcher; +} + +compile = Sizzle.compile = function( selector, match /* Internal Use Only */ ) { + var i, + setMatchers = [], + elementMatchers = [], + cached = compilerCache[ selector + " " ]; + + if ( !cached ) { + + // Generate a function of recursive functions that can be used to check each element + if ( !match ) { + match = tokenize( selector ); + } + i = match.length; + while ( i-- ) { + cached = matcherFromTokens( match[ i ] ); + if ( cached[ expando ] ) { + setMatchers.push( cached ); + } else { + elementMatchers.push( cached ); + } + } + + // Cache the compiled function + cached = compilerCache( + selector, + matcherFromGroupMatchers( elementMatchers, setMatchers ) + ); + + // Save selector and tokenization + cached.selector = selector; + } + return cached; +}; + +/** + * A low-level selection function that works with Sizzle's compiled + * selector functions + * @param {String|Function} selector A selector or a pre-compiled + * selector function built with Sizzle.compile + * @param {Element} context + * @param {Array} [results] + * @param {Array} [seed] A set of elements to match against + */ +select = Sizzle.select = function( selector, context, results, seed ) { + var i, tokens, token, type, find, + compiled = typeof selector === "function" && selector, + match = !seed && tokenize( ( selector = compiled.selector || selector ) ); + + results = results || []; + + // Try to minimize operations if there is only one selector in the list and no seed + // (the latter of which guarantees us context) + if ( match.length === 1 ) { + + // Reduce context if the leading compound selector is an ID + tokens = match[ 0 ] = match[ 0 ].slice( 0 ); + if ( tokens.length > 2 && ( token = tokens[ 0 ] ).type === "ID" && + context.nodeType === 9 && documentIsHTML && Expr.relative[ tokens[ 1 ].type ] ) { + + context = ( Expr.find[ "ID" ]( token.matches[ 0 ] + .replace( runescape, funescape ), context ) || [] )[ 0 ]; + if ( !context ) { + return results; + + // Precompiled matchers will still verify ancestry, so step up a level + } else if ( compiled ) { + context = context.parentNode; + } + + selector = selector.slice( tokens.shift().value.length ); + } + + // Fetch a seed set for right-to-left matching + i = matchExpr[ "needsContext" ].test( selector ) ? 0 : tokens.length; + while ( i-- ) { + token = tokens[ i ]; + + // Abort if we hit a combinator + if ( Expr.relative[ ( type = token.type ) ] ) { + break; + } + if ( ( find = Expr.find[ type ] ) ) { + + // Search, expanding context for leading sibling combinators + if ( ( seed = find( + token.matches[ 0 ].replace( runescape, funescape ), + rsibling.test( tokens[ 0 ].type ) && testContext( context.parentNode ) || + context + ) ) ) { + + // If seed is empty or no tokens remain, we can return early + tokens.splice( i, 1 ); + selector = seed.length && toSelector( tokens ); + if ( !selector ) { + push.apply( results, seed ); + return results; + } + + break; + } + } + } + } + + // Compile and execute a filtering function if one is not provided + // Provide `match` to avoid retokenization if we modified the selector above + ( compiled || compile( selector, match ) )( + seed, + context, + !documentIsHTML, + results, + !context || rsibling.test( selector ) && testContext( context.parentNode ) || context + ); + return results; +}; + +// One-time assignments + +// Sort stability +support.sortStable = expando.split( "" ).sort( sortOrder ).join( "" ) === expando; + +// Support: Chrome 14-35+ +// Always assume duplicates if they aren't passed to the comparison function +support.detectDuplicates = !!hasDuplicate; + +// Initialize against the default document +setDocument(); + +// Support: Webkit<537.32 - Safari 6.0.3/Chrome 25 (fixed in Chrome 27) +// Detached nodes confoundingly follow *each other* +support.sortDetached = assert( function( el ) { + + // Should return 1, but returns 4 (following) + return el.compareDocumentPosition( document.createElement( "fieldset" ) ) & 1; +} ); + +// Support: IE<8 +// Prevent attribute/property "interpolation" +// https://msdn.microsoft.com/en-us/library/ms536429%28VS.85%29.aspx +if ( !assert( function( el ) { + el.innerHTML = ""; + return el.firstChild.getAttribute( "href" ) === "#"; +} ) ) { + addHandle( "type|href|height|width", function( elem, name, isXML ) { + if ( !isXML ) { + return elem.getAttribute( name, name.toLowerCase() === "type" ? 1 : 2 ); + } + } ); +} + +// Support: IE<9 +// Use defaultValue in place of getAttribute("value") +if ( !support.attributes || !assert( function( el ) { + el.innerHTML = ""; + el.firstChild.setAttribute( "value", "" ); + return el.firstChild.getAttribute( "value" ) === ""; +} ) ) { + addHandle( "value", function( elem, _name, isXML ) { + if ( !isXML && elem.nodeName.toLowerCase() === "input" ) { + return elem.defaultValue; + } + } ); +} + +// Support: IE<9 +// Use getAttributeNode to fetch booleans when getAttribute lies +if ( !assert( function( el ) { + return el.getAttribute( "disabled" ) == null; +} ) ) { + addHandle( booleans, function( elem, name, isXML ) { + var val; + if ( !isXML ) { + return elem[ name ] === true ? name.toLowerCase() : + ( val = elem.getAttributeNode( name ) ) && val.specified ? + val.value : + null; + } + } ); +} + +return Sizzle; + +} )( window ); + + + +jQuery.find = Sizzle; +jQuery.expr = Sizzle.selectors; + +// Deprecated +jQuery.expr[ ":" ] = jQuery.expr.pseudos; +jQuery.uniqueSort = jQuery.unique = Sizzle.uniqueSort; +jQuery.text = Sizzle.getText; +jQuery.isXMLDoc = Sizzle.isXML; +jQuery.contains = Sizzle.contains; +jQuery.escapeSelector = Sizzle.escape; + + + + +var dir = function( elem, dir, until ) { + var matched = [], + truncate = until !== undefined; + + while ( ( elem = elem[ dir ] ) && elem.nodeType !== 9 ) { + if ( elem.nodeType === 1 ) { + if ( truncate && jQuery( elem ).is( until ) ) { + break; + } + matched.push( elem ); + } + } + return matched; +}; + + +var siblings = function( n, elem ) { + var matched = []; + + for ( ; n; n = n.nextSibling ) { + if ( n.nodeType === 1 && n !== elem ) { + matched.push( n ); + } + } + + return matched; +}; + + +var rneedsContext = jQuery.expr.match.needsContext; + + + +function nodeName( elem, name ) { + + return elem.nodeName && elem.nodeName.toLowerCase() === name.toLowerCase(); + +}; +var rsingleTag = ( /^<([a-z][^\/\0>:\x20\t\r\n\f]*)[\x20\t\r\n\f]*\/?>(?:<\/\1>|)$/i ); + + + +// Implement the identical functionality for filter and not +function winnow( elements, qualifier, not ) { + if ( isFunction( qualifier ) ) { + return jQuery.grep( elements, function( elem, i ) { + return !!qualifier.call( elem, i, elem ) !== not; + } ); + } + + // Single element + if ( qualifier.nodeType ) { + return jQuery.grep( elements, function( elem ) { + return ( elem === qualifier ) !== not; + } ); + } + + // Arraylike of elements (jQuery, arguments, Array) + if ( typeof qualifier !== "string" ) { + return jQuery.grep( elements, function( elem ) { + return ( indexOf.call( qualifier, elem ) > -1 ) !== not; + } ); + } + + // Filtered directly for both simple and complex selectors + return jQuery.filter( qualifier, elements, not ); +} + +jQuery.filter = function( expr, elems, not ) { + var elem = elems[ 0 ]; + + if ( not ) { + expr = ":not(" + expr + ")"; + } + + if ( elems.length === 1 && elem.nodeType === 1 ) { + return jQuery.find.matchesSelector( elem, expr ) ? [ elem ] : []; + } + + return jQuery.find.matches( expr, jQuery.grep( elems, function( elem ) { + return elem.nodeType === 1; + } ) ); +}; + +jQuery.fn.extend( { + find: function( selector ) { + var i, ret, + len = this.length, + self = this; + + if ( typeof selector !== "string" ) { + return this.pushStack( jQuery( selector ).filter( function() { + for ( i = 0; i < len; i++ ) { + if ( jQuery.contains( self[ i ], this ) ) { + return true; + } + } + } ) ); + } + + ret = this.pushStack( [] ); + + for ( i = 0; i < len; i++ ) { + jQuery.find( selector, self[ i ], ret ); + } + + return len > 1 ? jQuery.uniqueSort( ret ) : ret; + }, + filter: function( selector ) { + return this.pushStack( winnow( this, selector || [], false ) ); + }, + not: function( selector ) { + return this.pushStack( winnow( this, selector || [], true ) ); + }, + is: function( selector ) { + return !!winnow( + this, + + // If this is a positional/relative selector, check membership in the returned set + // so $("p:first").is("p:last") won't return true for a doc with two "p". + typeof selector === "string" && rneedsContext.test( selector ) ? + jQuery( selector ) : + selector || [], + false + ).length; + } +} ); + + +// Initialize a jQuery object + + +// A central reference to the root jQuery(document) +var rootjQuery, + + // A simple way to check for HTML strings + // Prioritize #id over to avoid XSS via location.hash (#9521) + // Strict HTML recognition (#11290: must start with <) + // Shortcut simple #id case for speed + rquickExpr = /^(?:\s*(<[\w\W]+>)[^>]*|#([\w-]+))$/, + + init = jQuery.fn.init = function( selector, context, root ) { + var match, elem; + + // HANDLE: $(""), $(null), $(undefined), $(false) + if ( !selector ) { + return this; + } + + // Method init() accepts an alternate rootjQuery + // so migrate can support jQuery.sub (gh-2101) + root = root || rootjQuery; + + // Handle HTML strings + if ( typeof selector === "string" ) { + if ( selector[ 0 ] === "<" && + selector[ selector.length - 1 ] === ">" && + selector.length >= 3 ) { + + // Assume that strings that start and end with <> are HTML and skip the regex check + match = [ null, selector, null ]; + + } else { + match = rquickExpr.exec( selector ); + } + + // Match html or make sure no context is specified for #id + if ( match && ( match[ 1 ] || !context ) ) { + + // HANDLE: $(html) -> $(array) + if ( match[ 1 ] ) { + context = context instanceof jQuery ? context[ 0 ] : context; + + // Option to run scripts is true for back-compat + // Intentionally let the error be thrown if parseHTML is not present + jQuery.merge( this, jQuery.parseHTML( + match[ 1 ], + context && context.nodeType ? context.ownerDocument || context : document, + true + ) ); + + // HANDLE: $(html, props) + if ( rsingleTag.test( match[ 1 ] ) && jQuery.isPlainObject( context ) ) { + for ( match in context ) { + + // Properties of context are called as methods if possible + if ( isFunction( this[ match ] ) ) { + this[ match ]( context[ match ] ); + + // ...and otherwise set as attributes + } else { + this.attr( match, context[ match ] ); + } + } + } + + return this; + + // HANDLE: $(#id) + } else { + elem = document.getElementById( match[ 2 ] ); + + if ( elem ) { + + // Inject the element directly into the jQuery object + this[ 0 ] = elem; + this.length = 1; + } + return this; + } + + // HANDLE: $(expr, $(...)) + } else if ( !context || context.jquery ) { + return ( context || root ).find( selector ); + + // HANDLE: $(expr, context) + // (which is just equivalent to: $(context).find(expr) + } else { + return this.constructor( context ).find( selector ); + } + + // HANDLE: $(DOMElement) + } else if ( selector.nodeType ) { + this[ 0 ] = selector; + this.length = 1; + return this; + + // HANDLE: $(function) + // Shortcut for document ready + } else if ( isFunction( selector ) ) { + return root.ready !== undefined ? + root.ready( selector ) : + + // Execute immediately if ready is not present + selector( jQuery ); + } + + return jQuery.makeArray( selector, this ); + }; + +// Give the init function the jQuery prototype for later instantiation +init.prototype = jQuery.fn; + +// Initialize central reference +rootjQuery = jQuery( document ); + + +var rparentsprev = /^(?:parents|prev(?:Until|All))/, + + // Methods guaranteed to produce a unique set when starting from a unique set + guaranteedUnique = { + children: true, + contents: true, + next: true, + prev: true + }; + +jQuery.fn.extend( { + has: function( target ) { + var targets = jQuery( target, this ), + l = targets.length; + + return this.filter( function() { + var i = 0; + for ( ; i < l; i++ ) { + if ( jQuery.contains( this, targets[ i ] ) ) { + return true; + } + } + } ); + }, + + closest: function( selectors, context ) { + var cur, + i = 0, + l = this.length, + matched = [], + targets = typeof selectors !== "string" && jQuery( selectors ); + + // Positional selectors never match, since there's no _selection_ context + if ( !rneedsContext.test( selectors ) ) { + for ( ; i < l; i++ ) { + for ( cur = this[ i ]; cur && cur !== context; cur = cur.parentNode ) { + + // Always skip document fragments + if ( cur.nodeType < 11 && ( targets ? + targets.index( cur ) > -1 : + + // Don't pass non-elements to Sizzle + cur.nodeType === 1 && + jQuery.find.matchesSelector( cur, selectors ) ) ) { + + matched.push( cur ); + break; + } + } + } + } + + return this.pushStack( matched.length > 1 ? jQuery.uniqueSort( matched ) : matched ); + }, + + // Determine the position of an element within the set + index: function( elem ) { + + // No argument, return index in parent + if ( !elem ) { + return ( this[ 0 ] && this[ 0 ].parentNode ) ? this.first().prevAll().length : -1; + } + + // Index in selector + if ( typeof elem === "string" ) { + return indexOf.call( jQuery( elem ), this[ 0 ] ); + } + + // Locate the position of the desired element + return indexOf.call( this, + + // If it receives a jQuery object, the first element is used + elem.jquery ? elem[ 0 ] : elem + ); + }, + + add: function( selector, context ) { + return this.pushStack( + jQuery.uniqueSort( + jQuery.merge( this.get(), jQuery( selector, context ) ) + ) + ); + }, + + addBack: function( selector ) { + return this.add( selector == null ? + this.prevObject : this.prevObject.filter( selector ) + ); + } +} ); + +function sibling( cur, dir ) { + while ( ( cur = cur[ dir ] ) && cur.nodeType !== 1 ) {} + return cur; +} + +jQuery.each( { + parent: function( elem ) { + var parent = elem.parentNode; + return parent && parent.nodeType !== 11 ? parent : null; + }, + parents: function( elem ) { + return dir( elem, "parentNode" ); + }, + parentsUntil: function( elem, _i, until ) { + return dir( elem, "parentNode", until ); + }, + next: function( elem ) { + return sibling( elem, "nextSibling" ); + }, + prev: function( elem ) { + return sibling( elem, "previousSibling" ); + }, + nextAll: function( elem ) { + return dir( elem, "nextSibling" ); + }, + prevAll: function( elem ) { + return dir( elem, "previousSibling" ); + }, + nextUntil: function( elem, _i, until ) { + return dir( elem, "nextSibling", until ); + }, + prevUntil: function( elem, _i, until ) { + return dir( elem, "previousSibling", until ); + }, + siblings: function( elem ) { + return siblings( ( elem.parentNode || {} ).firstChild, elem ); + }, + children: function( elem ) { + return siblings( elem.firstChild ); + }, + contents: function( elem ) { + if ( elem.contentDocument != null && + + // Support: IE 11+ + // elements with no `data` attribute has an object + // `contentDocument` with a `null` prototype. + getProto( elem.contentDocument ) ) { + + return elem.contentDocument; + } + + // Support: IE 9 - 11 only, iOS 7 only, Android Browser <=4.3 only + // Treat the template element as a regular one in browsers that + // don't support it. + if ( nodeName( elem, "template" ) ) { + elem = elem.content || elem; + } + + return jQuery.merge( [], elem.childNodes ); + } +}, function( name, fn ) { + jQuery.fn[ name ] = function( until, selector ) { + var matched = jQuery.map( this, fn, until ); + + if ( name.slice( -5 ) !== "Until" ) { + selector = until; + } + + if ( selector && typeof selector === "string" ) { + matched = jQuery.filter( selector, matched ); + } + + if ( this.length > 1 ) { + + // Remove duplicates + if ( !guaranteedUnique[ name ] ) { + jQuery.uniqueSort( matched ); + } + + // Reverse order for parents* and prev-derivatives + if ( rparentsprev.test( name ) ) { + matched.reverse(); + } + } + + return this.pushStack( matched ); + }; +} ); +var rnothtmlwhite = ( /[^\x20\t\r\n\f]+/g ); + + + +// Convert String-formatted options into Object-formatted ones +function createOptions( options ) { + var object = {}; + jQuery.each( options.match( rnothtmlwhite ) || [], function( _, flag ) { + object[ flag ] = true; + } ); + return object; +} + +/* + * Create a callback list using the following parameters: + * + * options: an optional list of space-separated options that will change how + * the callback list behaves or a more traditional option object + * + * By default a callback list will act like an event callback list and can be + * "fired" multiple times. + * + * Possible options: + * + * once: will ensure the callback list can only be fired once (like a Deferred) + * + * memory: will keep track of previous values and will call any callback added + * after the list has been fired right away with the latest "memorized" + * values (like a Deferred) + * + * unique: will ensure a callback can only be added once (no duplicate in the list) + * + * stopOnFalse: interrupt callings when a callback returns false + * + */ +jQuery.Callbacks = function( options ) { + + // Convert options from String-formatted to Object-formatted if needed + // (we check in cache first) + options = typeof options === "string" ? + createOptions( options ) : + jQuery.extend( {}, options ); + + var // Flag to know if list is currently firing + firing, + + // Last fire value for non-forgettable lists + memory, + + // Flag to know if list was already fired + fired, + + // Flag to prevent firing + locked, + + // Actual callback list + list = [], + + // Queue of execution data for repeatable lists + queue = [], + + // Index of currently firing callback (modified by add/remove as needed) + firingIndex = -1, + + // Fire callbacks + fire = function() { + + // Enforce single-firing + locked = locked || options.once; + + // Execute callbacks for all pending executions, + // respecting firingIndex overrides and runtime changes + fired = firing = true; + for ( ; queue.length; firingIndex = -1 ) { + memory = queue.shift(); + while ( ++firingIndex < list.length ) { + + // Run callback and check for early termination + if ( list[ firingIndex ].apply( memory[ 0 ], memory[ 1 ] ) === false && + options.stopOnFalse ) { + + // Jump to end and forget the data so .add doesn't re-fire + firingIndex = list.length; + memory = false; + } + } + } + + // Forget the data if we're done with it + if ( !options.memory ) { + memory = false; + } + + firing = false; + + // Clean up if we're done firing for good + if ( locked ) { + + // Keep an empty list if we have data for future add calls + if ( memory ) { + list = []; + + // Otherwise, this object is spent + } else { + list = ""; + } + } + }, + + // Actual Callbacks object + self = { + + // Add a callback or a collection of callbacks to the list + add: function() { + if ( list ) { + + // If we have memory from a past run, we should fire after adding + if ( memory && !firing ) { + firingIndex = list.length - 1; + queue.push( memory ); + } + + ( function add( args ) { + jQuery.each( args, function( _, arg ) { + if ( isFunction( arg ) ) { + if ( !options.unique || !self.has( arg ) ) { + list.push( arg ); + } + } else if ( arg && arg.length && toType( arg ) !== "string" ) { + + // Inspect recursively + add( arg ); + } + } ); + } )( arguments ); + + if ( memory && !firing ) { + fire(); + } + } + return this; + }, + + // Remove a callback from the list + remove: function() { + jQuery.each( arguments, function( _, arg ) { + var index; + while ( ( index = jQuery.inArray( arg, list, index ) ) > -1 ) { + list.splice( index, 1 ); + + // Handle firing indexes + if ( index <= firingIndex ) { + firingIndex--; + } + } + } ); + return this; + }, + + // Check if a given callback is in the list. + // If no argument is given, return whether or not list has callbacks attached. + has: function( fn ) { + return fn ? + jQuery.inArray( fn, list ) > -1 : + list.length > 0; + }, + + // Remove all callbacks from the list + empty: function() { + if ( list ) { + list = []; + } + return this; + }, + + // Disable .fire and .add + // Abort any current/pending executions + // Clear all callbacks and values + disable: function() { + locked = queue = []; + list = memory = ""; + return this; + }, + disabled: function() { + return !list; + }, + + // Disable .fire + // Also disable .add unless we have memory (since it would have no effect) + // Abort any pending executions + lock: function() { + locked = queue = []; + if ( !memory && !firing ) { + list = memory = ""; + } + return this; + }, + locked: function() { + return !!locked; + }, + + // Call all callbacks with the given context and arguments + fireWith: function( context, args ) { + if ( !locked ) { + args = args || []; + args = [ context, args.slice ? args.slice() : args ]; + queue.push( args ); + if ( !firing ) { + fire(); + } + } + return this; + }, + + // Call all the callbacks with the given arguments + fire: function() { + self.fireWith( this, arguments ); + return this; + }, + + // To know if the callbacks have already been called at least once + fired: function() { + return !!fired; + } + }; + + return self; +}; + + +function Identity( v ) { + return v; +} +function Thrower( ex ) { + throw ex; +} + +function adoptValue( value, resolve, reject, noValue ) { + var method; + + try { + + // Check for promise aspect first to privilege synchronous behavior + if ( value && isFunction( ( method = value.promise ) ) ) { + method.call( value ).done( resolve ).fail( reject ); + + // Other thenables + } else if ( value && isFunction( ( method = value.then ) ) ) { + method.call( value, resolve, reject ); + + // Other non-thenables + } else { + + // Control `resolve` arguments by letting Array#slice cast boolean `noValue` to integer: + // * false: [ value ].slice( 0 ) => resolve( value ) + // * true: [ value ].slice( 1 ) => resolve() + resolve.apply( undefined, [ value ].slice( noValue ) ); + } + + // For Promises/A+, convert exceptions into rejections + // Since jQuery.when doesn't unwrap thenables, we can skip the extra checks appearing in + // Deferred#then to conditionally suppress rejection. + } catch ( value ) { + + // Support: Android 4.0 only + // Strict mode functions invoked without .call/.apply get global-object context + reject.apply( undefined, [ value ] ); + } +} + +jQuery.extend( { + + Deferred: function( func ) { + var tuples = [ + + // action, add listener, callbacks, + // ... .then handlers, argument index, [final state] + [ "notify", "progress", jQuery.Callbacks( "memory" ), + jQuery.Callbacks( "memory" ), 2 ], + [ "resolve", "done", jQuery.Callbacks( "once memory" ), + jQuery.Callbacks( "once memory" ), 0, "resolved" ], + [ "reject", "fail", jQuery.Callbacks( "once memory" ), + jQuery.Callbacks( "once memory" ), 1, "rejected" ] + ], + state = "pending", + promise = { + state: function() { + return state; + }, + always: function() { + deferred.done( arguments ).fail( arguments ); + return this; + }, + "catch": function( fn ) { + return promise.then( null, fn ); + }, + + // Keep pipe for back-compat + pipe: function( /* fnDone, fnFail, fnProgress */ ) { + var fns = arguments; + + return jQuery.Deferred( function( newDefer ) { + jQuery.each( tuples, function( _i, tuple ) { + + // Map tuples (progress, done, fail) to arguments (done, fail, progress) + var fn = isFunction( fns[ tuple[ 4 ] ] ) && fns[ tuple[ 4 ] ]; + + // deferred.progress(function() { bind to newDefer or newDefer.notify }) + // deferred.done(function() { bind to newDefer or newDefer.resolve }) + // deferred.fail(function() { bind to newDefer or newDefer.reject }) + deferred[ tuple[ 1 ] ]( function() { + var returned = fn && fn.apply( this, arguments ); + if ( returned && isFunction( returned.promise ) ) { + returned.promise() + .progress( newDefer.notify ) + .done( newDefer.resolve ) + .fail( newDefer.reject ); + } else { + newDefer[ tuple[ 0 ] + "With" ]( + this, + fn ? [ returned ] : arguments + ); + } + } ); + } ); + fns = null; + } ).promise(); + }, + then: function( onFulfilled, onRejected, onProgress ) { + var maxDepth = 0; + function resolve( depth, deferred, handler, special ) { + return function() { + var that = this, + args = arguments, + mightThrow = function() { + var returned, then; + + // Support: Promises/A+ section 2.3.3.3.3 + // https://promisesaplus.com/#point-59 + // Ignore double-resolution attempts + if ( depth < maxDepth ) { + return; + } + + returned = handler.apply( that, args ); + + // Support: Promises/A+ section 2.3.1 + // https://promisesaplus.com/#point-48 + if ( returned === deferred.promise() ) { + throw new TypeError( "Thenable self-resolution" ); + } + + // Support: Promises/A+ sections 2.3.3.1, 3.5 + // https://promisesaplus.com/#point-54 + // https://promisesaplus.com/#point-75 + // Retrieve `then` only once + then = returned && + + // Support: Promises/A+ section 2.3.4 + // https://promisesaplus.com/#point-64 + // Only check objects and functions for thenability + ( typeof returned === "object" || + typeof returned === "function" ) && + returned.then; + + // Handle a returned thenable + if ( isFunction( then ) ) { + + // Special processors (notify) just wait for resolution + if ( special ) { + then.call( + returned, + resolve( maxDepth, deferred, Identity, special ), + resolve( maxDepth, deferred, Thrower, special ) + ); + + // Normal processors (resolve) also hook into progress + } else { + + // ...and disregard older resolution values + maxDepth++; + + then.call( + returned, + resolve( maxDepth, deferred, Identity, special ), + resolve( maxDepth, deferred, Thrower, special ), + resolve( maxDepth, deferred, Identity, + deferred.notifyWith ) + ); + } + + // Handle all other returned values + } else { + + // Only substitute handlers pass on context + // and multiple values (non-spec behavior) + if ( handler !== Identity ) { + that = undefined; + args = [ returned ]; + } + + // Process the value(s) + // Default process is resolve + ( special || deferred.resolveWith )( that, args ); + } + }, + + // Only normal processors (resolve) catch and reject exceptions + process = special ? + mightThrow : + function() { + try { + mightThrow(); + } catch ( e ) { + + if ( jQuery.Deferred.exceptionHook ) { + jQuery.Deferred.exceptionHook( e, + process.stackTrace ); + } + + // Support: Promises/A+ section 2.3.3.3.4.1 + // https://promisesaplus.com/#point-61 + // Ignore post-resolution exceptions + if ( depth + 1 >= maxDepth ) { + + // Only substitute handlers pass on context + // and multiple values (non-spec behavior) + if ( handler !== Thrower ) { + that = undefined; + args = [ e ]; + } + + deferred.rejectWith( that, args ); + } + } + }; + + // Support: Promises/A+ section 2.3.3.3.1 + // https://promisesaplus.com/#point-57 + // Re-resolve promises immediately to dodge false rejection from + // subsequent errors + if ( depth ) { + process(); + } else { + + // Call an optional hook to record the stack, in case of exception + // since it's otherwise lost when execution goes async + if ( jQuery.Deferred.getStackHook ) { + process.stackTrace = jQuery.Deferred.getStackHook(); + } + window.setTimeout( process ); + } + }; + } + + return jQuery.Deferred( function( newDefer ) { + + // progress_handlers.add( ... ) + tuples[ 0 ][ 3 ].add( + resolve( + 0, + newDefer, + isFunction( onProgress ) ? + onProgress : + Identity, + newDefer.notifyWith + ) + ); + + // fulfilled_handlers.add( ... ) + tuples[ 1 ][ 3 ].add( + resolve( + 0, + newDefer, + isFunction( onFulfilled ) ? + onFulfilled : + Identity + ) + ); + + // rejected_handlers.add( ... ) + tuples[ 2 ][ 3 ].add( + resolve( + 0, + newDefer, + isFunction( onRejected ) ? + onRejected : + Thrower + ) + ); + } ).promise(); + }, + + // Get a promise for this deferred + // If obj is provided, the promise aspect is added to the object + promise: function( obj ) { + return obj != null ? jQuery.extend( obj, promise ) : promise; + } + }, + deferred = {}; + + // Add list-specific methods + jQuery.each( tuples, function( i, tuple ) { + var list = tuple[ 2 ], + stateString = tuple[ 5 ]; + + // promise.progress = list.add + // promise.done = list.add + // promise.fail = list.add + promise[ tuple[ 1 ] ] = list.add; + + // Handle state + if ( stateString ) { + list.add( + function() { + + // state = "resolved" (i.e., fulfilled) + // state = "rejected" + state = stateString; + }, + + // rejected_callbacks.disable + // fulfilled_callbacks.disable + tuples[ 3 - i ][ 2 ].disable, + + // rejected_handlers.disable + // fulfilled_handlers.disable + tuples[ 3 - i ][ 3 ].disable, + + // progress_callbacks.lock + tuples[ 0 ][ 2 ].lock, + + // progress_handlers.lock + tuples[ 0 ][ 3 ].lock + ); + } + + // progress_handlers.fire + // fulfilled_handlers.fire + // rejected_handlers.fire + list.add( tuple[ 3 ].fire ); + + // deferred.notify = function() { deferred.notifyWith(...) } + // deferred.resolve = function() { deferred.resolveWith(...) } + // deferred.reject = function() { deferred.rejectWith(...) } + deferred[ tuple[ 0 ] ] = function() { + deferred[ tuple[ 0 ] + "With" ]( this === deferred ? undefined : this, arguments ); + return this; + }; + + // deferred.notifyWith = list.fireWith + // deferred.resolveWith = list.fireWith + // deferred.rejectWith = list.fireWith + deferred[ tuple[ 0 ] + "With" ] = list.fireWith; + } ); + + // Make the deferred a promise + promise.promise( deferred ); + + // Call given func if any + if ( func ) { + func.call( deferred, deferred ); + } + + // All done! + return deferred; + }, + + // Deferred helper + when: function( singleValue ) { + var + + // count of uncompleted subordinates + remaining = arguments.length, + + // count of unprocessed arguments + i = remaining, + + // subordinate fulfillment data + resolveContexts = Array( i ), + resolveValues = slice.call( arguments ), + + // the master Deferred + master = jQuery.Deferred(), + + // subordinate callback factory + updateFunc = function( i ) { + return function( value ) { + resolveContexts[ i ] = this; + resolveValues[ i ] = arguments.length > 1 ? slice.call( arguments ) : value; + if ( !( --remaining ) ) { + master.resolveWith( resolveContexts, resolveValues ); + } + }; + }; + + // Single- and empty arguments are adopted like Promise.resolve + if ( remaining <= 1 ) { + adoptValue( singleValue, master.done( updateFunc( i ) ).resolve, master.reject, + !remaining ); + + // Use .then() to unwrap secondary thenables (cf. gh-3000) + if ( master.state() === "pending" || + isFunction( resolveValues[ i ] && resolveValues[ i ].then ) ) { + + return master.then(); + } + } + + // Multiple arguments are aggregated like Promise.all array elements + while ( i-- ) { + adoptValue( resolveValues[ i ], updateFunc( i ), master.reject ); + } + + return master.promise(); + } +} ); + + +// These usually indicate a programmer mistake during development, +// warn about them ASAP rather than swallowing them by default. +var rerrorNames = /^(Eval|Internal|Range|Reference|Syntax|Type|URI)Error$/; + +jQuery.Deferred.exceptionHook = function( error, stack ) { + + // Support: IE 8 - 9 only + // Console exists when dev tools are open, which can happen at any time + if ( window.console && window.console.warn && error && rerrorNames.test( error.name ) ) { + window.console.warn( "jQuery.Deferred exception: " + error.message, error.stack, stack ); + } +}; + + + + +jQuery.readyException = function( error ) { + window.setTimeout( function() { + throw error; + } ); +}; + + + + +// The deferred used on DOM ready +var readyList = jQuery.Deferred(); + +jQuery.fn.ready = function( fn ) { + + readyList + .then( fn ) + + // Wrap jQuery.readyException in a function so that the lookup + // happens at the time of error handling instead of callback + // registration. + .catch( function( error ) { + jQuery.readyException( error ); + } ); + + return this; +}; + +jQuery.extend( { + + // Is the DOM ready to be used? Set to true once it occurs. + isReady: false, + + // A counter to track how many items to wait for before + // the ready event fires. See #6781 + readyWait: 1, + + // Handle when the DOM is ready + ready: function( wait ) { + + // Abort if there are pending holds or we're already ready + if ( wait === true ? --jQuery.readyWait : jQuery.isReady ) { + return; + } + + // Remember that the DOM is ready + jQuery.isReady = true; + + // If a normal DOM Ready event fired, decrement, and wait if need be + if ( wait !== true && --jQuery.readyWait > 0 ) { + return; + } + + // If there are functions bound, to execute + readyList.resolveWith( document, [ jQuery ] ); + } +} ); + +jQuery.ready.then = readyList.then; + +// The ready event handler and self cleanup method +function completed() { + document.removeEventListener( "DOMContentLoaded", completed ); + window.removeEventListener( "load", completed ); + jQuery.ready(); +} + +// Catch cases where $(document).ready() is called +// after the browser event has already occurred. +// Support: IE <=9 - 10 only +// Older IE sometimes signals "interactive" too soon +if ( document.readyState === "complete" || + ( document.readyState !== "loading" && !document.documentElement.doScroll ) ) { + + // Handle it asynchronously to allow scripts the opportunity to delay ready + window.setTimeout( jQuery.ready ); + +} else { + + // Use the handy event callback + document.addEventListener( "DOMContentLoaded", completed ); + + // A fallback to window.onload, that will always work + window.addEventListener( "load", completed ); +} + + + + +// Multifunctional method to get and set values of a collection +// The value/s can optionally be executed if it's a function +var access = function( elems, fn, key, value, chainable, emptyGet, raw ) { + var i = 0, + len = elems.length, + bulk = key == null; + + // Sets many values + if ( toType( key ) === "object" ) { + chainable = true; + for ( i in key ) { + access( elems, fn, i, key[ i ], true, emptyGet, raw ); + } + + // Sets one value + } else if ( value !== undefined ) { + chainable = true; + + if ( !isFunction( value ) ) { + raw = true; + } + + if ( bulk ) { + + // Bulk operations run against the entire set + if ( raw ) { + fn.call( elems, value ); + fn = null; + + // ...except when executing function values + } else { + bulk = fn; + fn = function( elem, _key, value ) { + return bulk.call( jQuery( elem ), value ); + }; + } + } + + if ( fn ) { + for ( ; i < len; i++ ) { + fn( + elems[ i ], key, raw ? + value : + value.call( elems[ i ], i, fn( elems[ i ], key ) ) + ); + } + } + } + + if ( chainable ) { + return elems; + } + + // Gets + if ( bulk ) { + return fn.call( elems ); + } + + return len ? fn( elems[ 0 ], key ) : emptyGet; +}; + + +// Matches dashed string for camelizing +var rmsPrefix = /^-ms-/, + rdashAlpha = /-([a-z])/g; + +// Used by camelCase as callback to replace() +function fcamelCase( _all, letter ) { + return letter.toUpperCase(); +} + +// Convert dashed to camelCase; used by the css and data modules +// Support: IE <=9 - 11, Edge 12 - 15 +// Microsoft forgot to hump their vendor prefix (#9572) +function camelCase( string ) { + return string.replace( rmsPrefix, "ms-" ).replace( rdashAlpha, fcamelCase ); +} +var acceptData = function( owner ) { + + // Accepts only: + // - Node + // - Node.ELEMENT_NODE + // - Node.DOCUMENT_NODE + // - Object + // - Any + return owner.nodeType === 1 || owner.nodeType === 9 || !( +owner.nodeType ); +}; + + + + +function Data() { + this.expando = jQuery.expando + Data.uid++; +} + +Data.uid = 1; + +Data.prototype = { + + cache: function( owner ) { + + // Check if the owner object already has a cache + var value = owner[ this.expando ]; + + // If not, create one + if ( !value ) { + value = {}; + + // We can accept data for non-element nodes in modern browsers, + // but we should not, see #8335. + // Always return an empty object. + if ( acceptData( owner ) ) { + + // If it is a node unlikely to be stringify-ed or looped over + // use plain assignment + if ( owner.nodeType ) { + owner[ this.expando ] = value; + + // Otherwise secure it in a non-enumerable property + // configurable must be true to allow the property to be + // deleted when data is removed + } else { + Object.defineProperty( owner, this.expando, { + value: value, + configurable: true + } ); + } + } + } + + return value; + }, + set: function( owner, data, value ) { + var prop, + cache = this.cache( owner ); + + // Handle: [ owner, key, value ] args + // Always use camelCase key (gh-2257) + if ( typeof data === "string" ) { + cache[ camelCase( data ) ] = value; + + // Handle: [ owner, { properties } ] args + } else { + + // Copy the properties one-by-one to the cache object + for ( prop in data ) { + cache[ camelCase( prop ) ] = data[ prop ]; + } + } + return cache; + }, + get: function( owner, key ) { + return key === undefined ? + this.cache( owner ) : + + // Always use camelCase key (gh-2257) + owner[ this.expando ] && owner[ this.expando ][ camelCase( key ) ]; + }, + access: function( owner, key, value ) { + + // In cases where either: + // + // 1. No key was specified + // 2. A string key was specified, but no value provided + // + // Take the "read" path and allow the get method to determine + // which value to return, respectively either: + // + // 1. The entire cache object + // 2. The data stored at the key + // + if ( key === undefined || + ( ( key && typeof key === "string" ) && value === undefined ) ) { + + return this.get( owner, key ); + } + + // When the key is not a string, or both a key and value + // are specified, set or extend (existing objects) with either: + // + // 1. An object of properties + // 2. A key and value + // + this.set( owner, key, value ); + + // Since the "set" path can have two possible entry points + // return the expected data based on which path was taken[*] + return value !== undefined ? value : key; + }, + remove: function( owner, key ) { + var i, + cache = owner[ this.expando ]; + + if ( cache === undefined ) { + return; + } + + if ( key !== undefined ) { + + // Support array or space separated string of keys + if ( Array.isArray( key ) ) { + + // If key is an array of keys... + // We always set camelCase keys, so remove that. + key = key.map( camelCase ); + } else { + key = camelCase( key ); + + // If a key with the spaces exists, use it. + // Otherwise, create an array by matching non-whitespace + key = key in cache ? + [ key ] : + ( key.match( rnothtmlwhite ) || [] ); + } + + i = key.length; + + while ( i-- ) { + delete cache[ key[ i ] ]; + } + } + + // Remove the expando if there's no more data + if ( key === undefined || jQuery.isEmptyObject( cache ) ) { + + // Support: Chrome <=35 - 45 + // Webkit & Blink performance suffers when deleting properties + // from DOM nodes, so set to undefined instead + // https://bugs.chromium.org/p/chromium/issues/detail?id=378607 (bug restricted) + if ( owner.nodeType ) { + owner[ this.expando ] = undefined; + } else { + delete owner[ this.expando ]; + } + } + }, + hasData: function( owner ) { + var cache = owner[ this.expando ]; + return cache !== undefined && !jQuery.isEmptyObject( cache ); + } +}; +var dataPriv = new Data(); + +var dataUser = new Data(); + + + +// Implementation Summary +// +// 1. Enforce API surface and semantic compatibility with 1.9.x branch +// 2. Improve the module's maintainability by reducing the storage +// paths to a single mechanism. +// 3. Use the same single mechanism to support "private" and "user" data. +// 4. _Never_ expose "private" data to user code (TODO: Drop _data, _removeData) +// 5. Avoid exposing implementation details on user objects (eg. expando properties) +// 6. Provide a clear path for implementation upgrade to WeakMap in 2014 + +var rbrace = /^(?:\{[\w\W]*\}|\[[\w\W]*\])$/, + rmultiDash = /[A-Z]/g; + +function getData( data ) { + if ( data === "true" ) { + return true; + } + + if ( data === "false" ) { + return false; + } + + if ( data === "null" ) { + return null; + } + + // Only convert to a number if it doesn't change the string + if ( data === +data + "" ) { + return +data; + } + + if ( rbrace.test( data ) ) { + return JSON.parse( data ); + } + + return data; +} + +function dataAttr( elem, key, data ) { + var name; + + // If nothing was found internally, try to fetch any + // data from the HTML5 data-* attribute + if ( data === undefined && elem.nodeType === 1 ) { + name = "data-" + key.replace( rmultiDash, "-$&" ).toLowerCase(); + data = elem.getAttribute( name ); + + if ( typeof data === "string" ) { + try { + data = getData( data ); + } catch ( e ) {} + + // Make sure we set the data so it isn't changed later + dataUser.set( elem, key, data ); + } else { + data = undefined; + } + } + return data; +} + +jQuery.extend( { + hasData: function( elem ) { + return dataUser.hasData( elem ) || dataPriv.hasData( elem ); + }, + + data: function( elem, name, data ) { + return dataUser.access( elem, name, data ); + }, + + removeData: function( elem, name ) { + dataUser.remove( elem, name ); + }, + + // TODO: Now that all calls to _data and _removeData have been replaced + // with direct calls to dataPriv methods, these can be deprecated. + _data: function( elem, name, data ) { + return dataPriv.access( elem, name, data ); + }, + + _removeData: function( elem, name ) { + dataPriv.remove( elem, name ); + } +} ); + +jQuery.fn.extend( { + data: function( key, value ) { + var i, name, data, + elem = this[ 0 ], + attrs = elem && elem.attributes; + + // Gets all values + if ( key === undefined ) { + if ( this.length ) { + data = dataUser.get( elem ); + + if ( elem.nodeType === 1 && !dataPriv.get( elem, "hasDataAttrs" ) ) { + i = attrs.length; + while ( i-- ) { + + // Support: IE 11 only + // The attrs elements can be null (#14894) + if ( attrs[ i ] ) { + name = attrs[ i ].name; + if ( name.indexOf( "data-" ) === 0 ) { + name = camelCase( name.slice( 5 ) ); + dataAttr( elem, name, data[ name ] ); + } + } + } + dataPriv.set( elem, "hasDataAttrs", true ); + } + } + + return data; + } + + // Sets multiple values + if ( typeof key === "object" ) { + return this.each( function() { + dataUser.set( this, key ); + } ); + } + + return access( this, function( value ) { + var data; + + // The calling jQuery object (element matches) is not empty + // (and therefore has an element appears at this[ 0 ]) and the + // `value` parameter was not undefined. An empty jQuery object + // will result in `undefined` for elem = this[ 0 ] which will + // throw an exception if an attempt to read a data cache is made. + if ( elem && value === undefined ) { + + // Attempt to get data from the cache + // The key will always be camelCased in Data + data = dataUser.get( elem, key ); + if ( data !== undefined ) { + return data; + } + + // Attempt to "discover" the data in + // HTML5 custom data-* attrs + data = dataAttr( elem, key ); + if ( data !== undefined ) { + return data; + } + + // We tried really hard, but the data doesn't exist. + return; + } + + // Set the data... + this.each( function() { + + // We always store the camelCased key + dataUser.set( this, key, value ); + } ); + }, null, value, arguments.length > 1, null, true ); + }, + + removeData: function( key ) { + return this.each( function() { + dataUser.remove( this, key ); + } ); + } +} ); + + +jQuery.extend( { + queue: function( elem, type, data ) { + var queue; + + if ( elem ) { + type = ( type || "fx" ) + "queue"; + queue = dataPriv.get( elem, type ); + + // Speed up dequeue by getting out quickly if this is just a lookup + if ( data ) { + if ( !queue || Array.isArray( data ) ) { + queue = dataPriv.access( elem, type, jQuery.makeArray( data ) ); + } else { + queue.push( data ); + } + } + return queue || []; + } + }, + + dequeue: function( elem, type ) { + type = type || "fx"; + + var queue = jQuery.queue( elem, type ), + startLength = queue.length, + fn = queue.shift(), + hooks = jQuery._queueHooks( elem, type ), + next = function() { + jQuery.dequeue( elem, type ); + }; + + // If the fx queue is dequeued, always remove the progress sentinel + if ( fn === "inprogress" ) { + fn = queue.shift(); + startLength--; + } + + if ( fn ) { + + // Add a progress sentinel to prevent the fx queue from being + // automatically dequeued + if ( type === "fx" ) { + queue.unshift( "inprogress" ); + } + + // Clear up the last queue stop function + delete hooks.stop; + fn.call( elem, next, hooks ); + } + + if ( !startLength && hooks ) { + hooks.empty.fire(); + } + }, + + // Not public - generate a queueHooks object, or return the current one + _queueHooks: function( elem, type ) { + var key = type + "queueHooks"; + return dataPriv.get( elem, key ) || dataPriv.access( elem, key, { + empty: jQuery.Callbacks( "once memory" ).add( function() { + dataPriv.remove( elem, [ type + "queue", key ] ); + } ) + } ); + } +} ); + +jQuery.fn.extend( { + queue: function( type, data ) { + var setter = 2; + + if ( typeof type !== "string" ) { + data = type; + type = "fx"; + setter--; + } + + if ( arguments.length < setter ) { + return jQuery.queue( this[ 0 ], type ); + } + + return data === undefined ? + this : + this.each( function() { + var queue = jQuery.queue( this, type, data ); + + // Ensure a hooks for this queue + jQuery._queueHooks( this, type ); + + if ( type === "fx" && queue[ 0 ] !== "inprogress" ) { + jQuery.dequeue( this, type ); + } + } ); + }, + dequeue: function( type ) { + return this.each( function() { + jQuery.dequeue( this, type ); + } ); + }, + clearQueue: function( type ) { + return this.queue( type || "fx", [] ); + }, + + // Get a promise resolved when queues of a certain type + // are emptied (fx is the type by default) + promise: function( type, obj ) { + var tmp, + count = 1, + defer = jQuery.Deferred(), + elements = this, + i = this.length, + resolve = function() { + if ( !( --count ) ) { + defer.resolveWith( elements, [ elements ] ); + } + }; + + if ( typeof type !== "string" ) { + obj = type; + type = undefined; + } + type = type || "fx"; + + while ( i-- ) { + tmp = dataPriv.get( elements[ i ], type + "queueHooks" ); + if ( tmp && tmp.empty ) { + count++; + tmp.empty.add( resolve ); + } + } + resolve(); + return defer.promise( obj ); + } +} ); +var pnum = ( /[+-]?(?:\d*\.|)\d+(?:[eE][+-]?\d+|)/ ).source; + +var rcssNum = new RegExp( "^(?:([+-])=|)(" + pnum + ")([a-z%]*)$", "i" ); + + +var cssExpand = [ "Top", "Right", "Bottom", "Left" ]; + +var documentElement = document.documentElement; + + + + var isAttached = function( elem ) { + return jQuery.contains( elem.ownerDocument, elem ); + }, + composed = { composed: true }; + + // Support: IE 9 - 11+, Edge 12 - 18+, iOS 10.0 - 10.2 only + // Check attachment across shadow DOM boundaries when possible (gh-3504) + // Support: iOS 10.0-10.2 only + // Early iOS 10 versions support `attachShadow` but not `getRootNode`, + // leading to errors. We need to check for `getRootNode`. + if ( documentElement.getRootNode ) { + isAttached = function( elem ) { + return jQuery.contains( elem.ownerDocument, elem ) || + elem.getRootNode( composed ) === elem.ownerDocument; + }; + } +var isHiddenWithinTree = function( elem, el ) { + + // isHiddenWithinTree might be called from jQuery#filter function; + // in that case, element will be second argument + elem = el || elem; + + // Inline style trumps all + return elem.style.display === "none" || + elem.style.display === "" && + + // Otherwise, check computed style + // Support: Firefox <=43 - 45 + // Disconnected elements can have computed display: none, so first confirm that elem is + // in the document. + isAttached( elem ) && + + jQuery.css( elem, "display" ) === "none"; + }; + + + +function adjustCSS( elem, prop, valueParts, tween ) { + var adjusted, scale, + maxIterations = 20, + currentValue = tween ? + function() { + return tween.cur(); + } : + function() { + return jQuery.css( elem, prop, "" ); + }, + initial = currentValue(), + unit = valueParts && valueParts[ 3 ] || ( jQuery.cssNumber[ prop ] ? "" : "px" ), + + // Starting value computation is required for potential unit mismatches + initialInUnit = elem.nodeType && + ( jQuery.cssNumber[ prop ] || unit !== "px" && +initial ) && + rcssNum.exec( jQuery.css( elem, prop ) ); + + if ( initialInUnit && initialInUnit[ 3 ] !== unit ) { + + // Support: Firefox <=54 + // Halve the iteration target value to prevent interference from CSS upper bounds (gh-2144) + initial = initial / 2; + + // Trust units reported by jQuery.css + unit = unit || initialInUnit[ 3 ]; + + // Iteratively approximate from a nonzero starting point + initialInUnit = +initial || 1; + + while ( maxIterations-- ) { + + // Evaluate and update our best guess (doubling guesses that zero out). + // Finish if the scale equals or crosses 1 (making the old*new product non-positive). + jQuery.style( elem, prop, initialInUnit + unit ); + if ( ( 1 - scale ) * ( 1 - ( scale = currentValue() / initial || 0.5 ) ) <= 0 ) { + maxIterations = 0; + } + initialInUnit = initialInUnit / scale; + + } + + initialInUnit = initialInUnit * 2; + jQuery.style( elem, prop, initialInUnit + unit ); + + // Make sure we update the tween properties later on + valueParts = valueParts || []; + } + + if ( valueParts ) { + initialInUnit = +initialInUnit || +initial || 0; + + // Apply relative offset (+=/-=) if specified + adjusted = valueParts[ 1 ] ? + initialInUnit + ( valueParts[ 1 ] + 1 ) * valueParts[ 2 ] : + +valueParts[ 2 ]; + if ( tween ) { + tween.unit = unit; + tween.start = initialInUnit; + tween.end = adjusted; + } + } + return adjusted; +} + + +var defaultDisplayMap = {}; + +function getDefaultDisplay( elem ) { + var temp, + doc = elem.ownerDocument, + nodeName = elem.nodeName, + display = defaultDisplayMap[ nodeName ]; + + if ( display ) { + return display; + } + + temp = doc.body.appendChild( doc.createElement( nodeName ) ); + display = jQuery.css( temp, "display" ); + + temp.parentNode.removeChild( temp ); + + if ( display === "none" ) { + display = "block"; + } + defaultDisplayMap[ nodeName ] = display; + + return display; +} + +function showHide( elements, show ) { + var display, elem, + values = [], + index = 0, + length = elements.length; + + // Determine new display value for elements that need to change + for ( ; index < length; index++ ) { + elem = elements[ index ]; + if ( !elem.style ) { + continue; + } + + display = elem.style.display; + if ( show ) { + + // Since we force visibility upon cascade-hidden elements, an immediate (and slow) + // check is required in this first loop unless we have a nonempty display value (either + // inline or about-to-be-restored) + if ( display === "none" ) { + values[ index ] = dataPriv.get( elem, "display" ) || null; + if ( !values[ index ] ) { + elem.style.display = ""; + } + } + if ( elem.style.display === "" && isHiddenWithinTree( elem ) ) { + values[ index ] = getDefaultDisplay( elem ); + } + } else { + if ( display !== "none" ) { + values[ index ] = "none"; + + // Remember what we're overwriting + dataPriv.set( elem, "display", display ); + } + } + } + + // Set the display of the elements in a second loop to avoid constant reflow + for ( index = 0; index < length; index++ ) { + if ( values[ index ] != null ) { + elements[ index ].style.display = values[ index ]; + } + } + + return elements; +} + +jQuery.fn.extend( { + show: function() { + return showHide( this, true ); + }, + hide: function() { + return showHide( this ); + }, + toggle: function( state ) { + if ( typeof state === "boolean" ) { + return state ? this.show() : this.hide(); + } + + return this.each( function() { + if ( isHiddenWithinTree( this ) ) { + jQuery( this ).show(); + } else { + jQuery( this ).hide(); + } + } ); + } +} ); +var rcheckableType = ( /^(?:checkbox|radio)$/i ); + +var rtagName = ( /<([a-z][^\/\0>\x20\t\r\n\f]*)/i ); + +var rscriptType = ( /^$|^module$|\/(?:java|ecma)script/i ); + + + +( function() { + var fragment = document.createDocumentFragment(), + div = fragment.appendChild( document.createElement( "div" ) ), + input = document.createElement( "input" ); + + // Support: Android 4.0 - 4.3 only + // Check state lost if the name is set (#11217) + // Support: Windows Web Apps (WWA) + // `name` and `type` must use .setAttribute for WWA (#14901) + input.setAttribute( "type", "radio" ); + input.setAttribute( "checked", "checked" ); + input.setAttribute( "name", "t" ); + + div.appendChild( input ); + + // Support: Android <=4.1 only + // Older WebKit doesn't clone checked state correctly in fragments + support.checkClone = div.cloneNode( true ).cloneNode( true ).lastChild.checked; + + // Support: IE <=11 only + // Make sure textarea (and checkbox) defaultValue is properly cloned + div.innerHTML = ""; + support.noCloneChecked = !!div.cloneNode( true ).lastChild.defaultValue; + + // Support: IE <=9 only + // IE <=9 replaces "; + support.option = !!div.lastChild; +} )(); + + +// We have to close these tags to support XHTML (#13200) +var wrapMap = { + + // XHTML parsers do not magically insert elements in the + // same way that tag soup parsers do. So we cannot shorten + // this by omitting or other required elements. + thead: [ 1, "", "
" ], + col: [ 2, "", "
" ], + tr: [ 2, "", "
" ], + td: [ 3, "", "
" ], + + _default: [ 0, "", "" ] +}; + +wrapMap.tbody = wrapMap.tfoot = wrapMap.colgroup = wrapMap.caption = wrapMap.thead; +wrapMap.th = wrapMap.td; + +// Support: IE <=9 only +if ( !support.option ) { + wrapMap.optgroup = wrapMap.option = [ 1, "" ]; +} + + +function getAll( context, tag ) { + + // Support: IE <=9 - 11 only + // Use typeof to avoid zero-argument method invocation on host objects (#15151) + var ret; + + if ( typeof context.getElementsByTagName !== "undefined" ) { + ret = context.getElementsByTagName( tag || "*" ); + + } else if ( typeof context.querySelectorAll !== "undefined" ) { + ret = context.querySelectorAll( tag || "*" ); + + } else { + ret = []; + } + + if ( tag === undefined || tag && nodeName( context, tag ) ) { + return jQuery.merge( [ context ], ret ); + } + + return ret; +} + + +// Mark scripts as having already been evaluated +function setGlobalEval( elems, refElements ) { + var i = 0, + l = elems.length; + + for ( ; i < l; i++ ) { + dataPriv.set( + elems[ i ], + "globalEval", + !refElements || dataPriv.get( refElements[ i ], "globalEval" ) + ); + } +} + + +var rhtml = /<|&#?\w+;/; + +function buildFragment( elems, context, scripts, selection, ignored ) { + var elem, tmp, tag, wrap, attached, j, + fragment = context.createDocumentFragment(), + nodes = [], + i = 0, + l = elems.length; + + for ( ; i < l; i++ ) { + elem = elems[ i ]; + + if ( elem || elem === 0 ) { + + // Add nodes directly + if ( toType( elem ) === "object" ) { + + // Support: Android <=4.0 only, PhantomJS 1 only + // push.apply(_, arraylike) throws on ancient WebKit + jQuery.merge( nodes, elem.nodeType ? [ elem ] : elem ); + + // Convert non-html into a text node + } else if ( !rhtml.test( elem ) ) { + nodes.push( context.createTextNode( elem ) ); + + // Convert html into DOM nodes + } else { + tmp = tmp || fragment.appendChild( context.createElement( "div" ) ); + + // Deserialize a standard representation + tag = ( rtagName.exec( elem ) || [ "", "" ] )[ 1 ].toLowerCase(); + wrap = wrapMap[ tag ] || wrapMap._default; + tmp.innerHTML = wrap[ 1 ] + jQuery.htmlPrefilter( elem ) + wrap[ 2 ]; + + // Descend through wrappers to the right content + j = wrap[ 0 ]; + while ( j-- ) { + tmp = tmp.lastChild; + } + + // Support: Android <=4.0 only, PhantomJS 1 only + // push.apply(_, arraylike) throws on ancient WebKit + jQuery.merge( nodes, tmp.childNodes ); + + // Remember the top-level container + tmp = fragment.firstChild; + + // Ensure the created nodes are orphaned (#12392) + tmp.textContent = ""; + } + } + } + + // Remove wrapper from fragment + fragment.textContent = ""; + + i = 0; + while ( ( elem = nodes[ i++ ] ) ) { + + // Skip elements already in the context collection (trac-4087) + if ( selection && jQuery.inArray( elem, selection ) > -1 ) { + if ( ignored ) { + ignored.push( elem ); + } + continue; + } + + attached = isAttached( elem ); + + // Append to fragment + tmp = getAll( fragment.appendChild( elem ), "script" ); + + // Preserve script evaluation history + if ( attached ) { + setGlobalEval( tmp ); + } + + // Capture executables + if ( scripts ) { + j = 0; + while ( ( elem = tmp[ j++ ] ) ) { + if ( rscriptType.test( elem.type || "" ) ) { + scripts.push( elem ); + } + } + } + } + + return fragment; +} + + +var + rkeyEvent = /^key/, + rmouseEvent = /^(?:mouse|pointer|contextmenu|drag|drop)|click/, + rtypenamespace = /^([^.]*)(?:\.(.+)|)/; + +function returnTrue() { + return true; +} + +function returnFalse() { + return false; +} + +// Support: IE <=9 - 11+ +// focus() and blur() are asynchronous, except when they are no-op. +// So expect focus to be synchronous when the element is already active, +// and blur to be synchronous when the element is not already active. +// (focus and blur are always synchronous in other supported browsers, +// this just defines when we can count on it). +function expectSync( elem, type ) { + return ( elem === safeActiveElement() ) === ( type === "focus" ); +} + +// Support: IE <=9 only +// Accessing document.activeElement can throw unexpectedly +// https://bugs.jquery.com/ticket/13393 +function safeActiveElement() { + try { + return document.activeElement; + } catch ( err ) { } +} + +function on( elem, types, selector, data, fn, one ) { + var origFn, type; + + // Types can be a map of types/handlers + if ( typeof types === "object" ) { + + // ( types-Object, selector, data ) + if ( typeof selector !== "string" ) { + + // ( types-Object, data ) + data = data || selector; + selector = undefined; + } + for ( type in types ) { + on( elem, type, selector, data, types[ type ], one ); + } + return elem; + } + + if ( data == null && fn == null ) { + + // ( types, fn ) + fn = selector; + data = selector = undefined; + } else if ( fn == null ) { + if ( typeof selector === "string" ) { + + // ( types, selector, fn ) + fn = data; + data = undefined; + } else { + + // ( types, data, fn ) + fn = data; + data = selector; + selector = undefined; + } + } + if ( fn === false ) { + fn = returnFalse; + } else if ( !fn ) { + return elem; + } + + if ( one === 1 ) { + origFn = fn; + fn = function( event ) { + + // Can use an empty set, since event contains the info + jQuery().off( event ); + return origFn.apply( this, arguments ); + }; + + // Use same guid so caller can remove using origFn + fn.guid = origFn.guid || ( origFn.guid = jQuery.guid++ ); + } + return elem.each( function() { + jQuery.event.add( this, types, fn, data, selector ); + } ); +} + +/* + * Helper functions for managing events -- not part of the public interface. + * Props to Dean Edwards' addEvent library for many of the ideas. + */ +jQuery.event = { + + global: {}, + + add: function( elem, types, handler, data, selector ) { + + var handleObjIn, eventHandle, tmp, + events, t, handleObj, + special, handlers, type, namespaces, origType, + elemData = dataPriv.get( elem ); + + // Only attach events to objects that accept data + if ( !acceptData( elem ) ) { + return; + } + + // Caller can pass in an object of custom data in lieu of the handler + if ( handler.handler ) { + handleObjIn = handler; + handler = handleObjIn.handler; + selector = handleObjIn.selector; + } + + // Ensure that invalid selectors throw exceptions at attach time + // Evaluate against documentElement in case elem is a non-element node (e.g., document) + if ( selector ) { + jQuery.find.matchesSelector( documentElement, selector ); + } + + // Make sure that the handler has a unique ID, used to find/remove it later + if ( !handler.guid ) { + handler.guid = jQuery.guid++; + } + + // Init the element's event structure and main handler, if this is the first + if ( !( events = elemData.events ) ) { + events = elemData.events = Object.create( null ); + } + if ( !( eventHandle = elemData.handle ) ) { + eventHandle = elemData.handle = function( e ) { + + // Discard the second event of a jQuery.event.trigger() and + // when an event is called after a page has unloaded + return typeof jQuery !== "undefined" && jQuery.event.triggered !== e.type ? + jQuery.event.dispatch.apply( elem, arguments ) : undefined; + }; + } + + // Handle multiple events separated by a space + types = ( types || "" ).match( rnothtmlwhite ) || [ "" ]; + t = types.length; + while ( t-- ) { + tmp = rtypenamespace.exec( types[ t ] ) || []; + type = origType = tmp[ 1 ]; + namespaces = ( tmp[ 2 ] || "" ).split( "." ).sort(); + + // There *must* be a type, no attaching namespace-only handlers + if ( !type ) { + continue; + } + + // If event changes its type, use the special event handlers for the changed type + special = jQuery.event.special[ type ] || {}; + + // If selector defined, determine special event api type, otherwise given type + type = ( selector ? special.delegateType : special.bindType ) || type; + + // Update special based on newly reset type + special = jQuery.event.special[ type ] || {}; + + // handleObj is passed to all event handlers + handleObj = jQuery.extend( { + type: type, + origType: origType, + data: data, + handler: handler, + guid: handler.guid, + selector: selector, + needsContext: selector && jQuery.expr.match.needsContext.test( selector ), + namespace: namespaces.join( "." ) + }, handleObjIn ); + + // Init the event handler queue if we're the first + if ( !( handlers = events[ type ] ) ) { + handlers = events[ type ] = []; + handlers.delegateCount = 0; + + // Only use addEventListener if the special events handler returns false + if ( !special.setup || + special.setup.call( elem, data, namespaces, eventHandle ) === false ) { + + if ( elem.addEventListener ) { + elem.addEventListener( type, eventHandle ); + } + } + } + + if ( special.add ) { + special.add.call( elem, handleObj ); + + if ( !handleObj.handler.guid ) { + handleObj.handler.guid = handler.guid; + } + } + + // Add to the element's handler list, delegates in front + if ( selector ) { + handlers.splice( handlers.delegateCount++, 0, handleObj ); + } else { + handlers.push( handleObj ); + } + + // Keep track of which events have ever been used, for event optimization + jQuery.event.global[ type ] = true; + } + + }, + + // Detach an event or set of events from an element + remove: function( elem, types, handler, selector, mappedTypes ) { + + var j, origCount, tmp, + events, t, handleObj, + special, handlers, type, namespaces, origType, + elemData = dataPriv.hasData( elem ) && dataPriv.get( elem ); + + if ( !elemData || !( events = elemData.events ) ) { + return; + } + + // Once for each type.namespace in types; type may be omitted + types = ( types || "" ).match( rnothtmlwhite ) || [ "" ]; + t = types.length; + while ( t-- ) { + tmp = rtypenamespace.exec( types[ t ] ) || []; + type = origType = tmp[ 1 ]; + namespaces = ( tmp[ 2 ] || "" ).split( "." ).sort(); + + // Unbind all events (on this namespace, if provided) for the element + if ( !type ) { + for ( type in events ) { + jQuery.event.remove( elem, type + types[ t ], handler, selector, true ); + } + continue; + } + + special = jQuery.event.special[ type ] || {}; + type = ( selector ? special.delegateType : special.bindType ) || type; + handlers = events[ type ] || []; + tmp = tmp[ 2 ] && + new RegExp( "(^|\\.)" + namespaces.join( "\\.(?:.*\\.|)" ) + "(\\.|$)" ); + + // Remove matching events + origCount = j = handlers.length; + while ( j-- ) { + handleObj = handlers[ j ]; + + if ( ( mappedTypes || origType === handleObj.origType ) && + ( !handler || handler.guid === handleObj.guid ) && + ( !tmp || tmp.test( handleObj.namespace ) ) && + ( !selector || selector === handleObj.selector || + selector === "**" && handleObj.selector ) ) { + handlers.splice( j, 1 ); + + if ( handleObj.selector ) { + handlers.delegateCount--; + } + if ( special.remove ) { + special.remove.call( elem, handleObj ); + } + } + } + + // Remove generic event handler if we removed something and no more handlers exist + // (avoids potential for endless recursion during removal of special event handlers) + if ( origCount && !handlers.length ) { + if ( !special.teardown || + special.teardown.call( elem, namespaces, elemData.handle ) === false ) { + + jQuery.removeEvent( elem, type, elemData.handle ); + } + + delete events[ type ]; + } + } + + // Remove data and the expando if it's no longer used + if ( jQuery.isEmptyObject( events ) ) { + dataPriv.remove( elem, "handle events" ); + } + }, + + dispatch: function( nativeEvent ) { + + var i, j, ret, matched, handleObj, handlerQueue, + args = new Array( arguments.length ), + + // Make a writable jQuery.Event from the native event object + event = jQuery.event.fix( nativeEvent ), + + handlers = ( + dataPriv.get( this, "events" ) || Object.create( null ) + )[ event.type ] || [], + special = jQuery.event.special[ event.type ] || {}; + + // Use the fix-ed jQuery.Event rather than the (read-only) native event + args[ 0 ] = event; + + for ( i = 1; i < arguments.length; i++ ) { + args[ i ] = arguments[ i ]; + } + + event.delegateTarget = this; + + // Call the preDispatch hook for the mapped type, and let it bail if desired + if ( special.preDispatch && special.preDispatch.call( this, event ) === false ) { + return; + } + + // Determine handlers + handlerQueue = jQuery.event.handlers.call( this, event, handlers ); + + // Run delegates first; they may want to stop propagation beneath us + i = 0; + while ( ( matched = handlerQueue[ i++ ] ) && !event.isPropagationStopped() ) { + event.currentTarget = matched.elem; + + j = 0; + while ( ( handleObj = matched.handlers[ j++ ] ) && + !event.isImmediatePropagationStopped() ) { + + // If the event is namespaced, then each handler is only invoked if it is + // specially universal or its namespaces are a superset of the event's. + if ( !event.rnamespace || handleObj.namespace === false || + event.rnamespace.test( handleObj.namespace ) ) { + + event.handleObj = handleObj; + event.data = handleObj.data; + + ret = ( ( jQuery.event.special[ handleObj.origType ] || {} ).handle || + handleObj.handler ).apply( matched.elem, args ); + + if ( ret !== undefined ) { + if ( ( event.result = ret ) === false ) { + event.preventDefault(); + event.stopPropagation(); + } + } + } + } + } + + // Call the postDispatch hook for the mapped type + if ( special.postDispatch ) { + special.postDispatch.call( this, event ); + } + + return event.result; + }, + + handlers: function( event, handlers ) { + var i, handleObj, sel, matchedHandlers, matchedSelectors, + handlerQueue = [], + delegateCount = handlers.delegateCount, + cur = event.target; + + // Find delegate handlers + if ( delegateCount && + + // Support: IE <=9 + // Black-hole SVG instance trees (trac-13180) + cur.nodeType && + + // Support: Firefox <=42 + // Suppress spec-violating clicks indicating a non-primary pointer button (trac-3861) + // https://www.w3.org/TR/DOM-Level-3-Events/#event-type-click + // Support: IE 11 only + // ...but not arrow key "clicks" of radio inputs, which can have `button` -1 (gh-2343) + !( event.type === "click" && event.button >= 1 ) ) { + + for ( ; cur !== this; cur = cur.parentNode || this ) { + + // Don't check non-elements (#13208) + // Don't process clicks on disabled elements (#6911, #8165, #11382, #11764) + if ( cur.nodeType === 1 && !( event.type === "click" && cur.disabled === true ) ) { + matchedHandlers = []; + matchedSelectors = {}; + for ( i = 0; i < delegateCount; i++ ) { + handleObj = handlers[ i ]; + + // Don't conflict with Object.prototype properties (#13203) + sel = handleObj.selector + " "; + + if ( matchedSelectors[ sel ] === undefined ) { + matchedSelectors[ sel ] = handleObj.needsContext ? + jQuery( sel, this ).index( cur ) > -1 : + jQuery.find( sel, this, null, [ cur ] ).length; + } + if ( matchedSelectors[ sel ] ) { + matchedHandlers.push( handleObj ); + } + } + if ( matchedHandlers.length ) { + handlerQueue.push( { elem: cur, handlers: matchedHandlers } ); + } + } + } + } + + // Add the remaining (directly-bound) handlers + cur = this; + if ( delegateCount < handlers.length ) { + handlerQueue.push( { elem: cur, handlers: handlers.slice( delegateCount ) } ); + } + + return handlerQueue; + }, + + addProp: function( name, hook ) { + Object.defineProperty( jQuery.Event.prototype, name, { + enumerable: true, + configurable: true, + + get: isFunction( hook ) ? + function() { + if ( this.originalEvent ) { + return hook( this.originalEvent ); + } + } : + function() { + if ( this.originalEvent ) { + return this.originalEvent[ name ]; + } + }, + + set: function( value ) { + Object.defineProperty( this, name, { + enumerable: true, + configurable: true, + writable: true, + value: value + } ); + } + } ); + }, + + fix: function( originalEvent ) { + return originalEvent[ jQuery.expando ] ? + originalEvent : + new jQuery.Event( originalEvent ); + }, + + special: { + load: { + + // Prevent triggered image.load events from bubbling to window.load + noBubble: true + }, + click: { + + // Utilize native event to ensure correct state for checkable inputs + setup: function( data ) { + + // For mutual compressibility with _default, replace `this` access with a local var. + // `|| data` is dead code meant only to preserve the variable through minification. + var el = this || data; + + // Claim the first handler + if ( rcheckableType.test( el.type ) && + el.click && nodeName( el, "input" ) ) { + + // dataPriv.set( el, "click", ... ) + leverageNative( el, "click", returnTrue ); + } + + // Return false to allow normal processing in the caller + return false; + }, + trigger: function( data ) { + + // For mutual compressibility with _default, replace `this` access with a local var. + // `|| data` is dead code meant only to preserve the variable through minification. + var el = this || data; + + // Force setup before triggering a click + if ( rcheckableType.test( el.type ) && + el.click && nodeName( el, "input" ) ) { + + leverageNative( el, "click" ); + } + + // Return non-false to allow normal event-path propagation + return true; + }, + + // For cross-browser consistency, suppress native .click() on links + // Also prevent it if we're currently inside a leveraged native-event stack + _default: function( event ) { + var target = event.target; + return rcheckableType.test( target.type ) && + target.click && nodeName( target, "input" ) && + dataPriv.get( target, "click" ) || + nodeName( target, "a" ); + } + }, + + beforeunload: { + postDispatch: function( event ) { + + // Support: Firefox 20+ + // Firefox doesn't alert if the returnValue field is not set. + if ( event.result !== undefined && event.originalEvent ) { + event.originalEvent.returnValue = event.result; + } + } + } + } +}; + +// Ensure the presence of an event listener that handles manually-triggered +// synthetic events by interrupting progress until reinvoked in response to +// *native* events that it fires directly, ensuring that state changes have +// already occurred before other listeners are invoked. +function leverageNative( el, type, expectSync ) { + + // Missing expectSync indicates a trigger call, which must force setup through jQuery.event.add + if ( !expectSync ) { + if ( dataPriv.get( el, type ) === undefined ) { + jQuery.event.add( el, type, returnTrue ); + } + return; + } + + // Register the controller as a special universal handler for all event namespaces + dataPriv.set( el, type, false ); + jQuery.event.add( el, type, { + namespace: false, + handler: function( event ) { + var notAsync, result, + saved = dataPriv.get( this, type ); + + if ( ( event.isTrigger & 1 ) && this[ type ] ) { + + // Interrupt processing of the outer synthetic .trigger()ed event + // Saved data should be false in such cases, but might be a leftover capture object + // from an async native handler (gh-4350) + if ( !saved.length ) { + + // Store arguments for use when handling the inner native event + // There will always be at least one argument (an event object), so this array + // will not be confused with a leftover capture object. + saved = slice.call( arguments ); + dataPriv.set( this, type, saved ); + + // Trigger the native event and capture its result + // Support: IE <=9 - 11+ + // focus() and blur() are asynchronous + notAsync = expectSync( this, type ); + this[ type ](); + result = dataPriv.get( this, type ); + if ( saved !== result || notAsync ) { + dataPriv.set( this, type, false ); + } else { + result = {}; + } + if ( saved !== result ) { + + // Cancel the outer synthetic event + event.stopImmediatePropagation(); + event.preventDefault(); + return result.value; + } + + // If this is an inner synthetic event for an event with a bubbling surrogate + // (focus or blur), assume that the surrogate already propagated from triggering the + // native event and prevent that from happening again here. + // This technically gets the ordering wrong w.r.t. to `.trigger()` (in which the + // bubbling surrogate propagates *after* the non-bubbling base), but that seems + // less bad than duplication. + } else if ( ( jQuery.event.special[ type ] || {} ).delegateType ) { + event.stopPropagation(); + } + + // If this is a native event triggered above, everything is now in order + // Fire an inner synthetic event with the original arguments + } else if ( saved.length ) { + + // ...and capture the result + dataPriv.set( this, type, { + value: jQuery.event.trigger( + + // Support: IE <=9 - 11+ + // Extend with the prototype to reset the above stopImmediatePropagation() + jQuery.extend( saved[ 0 ], jQuery.Event.prototype ), + saved.slice( 1 ), + this + ) + } ); + + // Abort handling of the native event + event.stopImmediatePropagation(); + } + } + } ); +} + +jQuery.removeEvent = function( elem, type, handle ) { + + // This "if" is needed for plain objects + if ( elem.removeEventListener ) { + elem.removeEventListener( type, handle ); + } +}; + +jQuery.Event = function( src, props ) { + + // Allow instantiation without the 'new' keyword + if ( !( this instanceof jQuery.Event ) ) { + return new jQuery.Event( src, props ); + } + + // Event object + if ( src && src.type ) { + this.originalEvent = src; + this.type = src.type; + + // Events bubbling up the document may have been marked as prevented + // by a handler lower down the tree; reflect the correct value. + this.isDefaultPrevented = src.defaultPrevented || + src.defaultPrevented === undefined && + + // Support: Android <=2.3 only + src.returnValue === false ? + returnTrue : + returnFalse; + + // Create target properties + // Support: Safari <=6 - 7 only + // Target should not be a text node (#504, #13143) + this.target = ( src.target && src.target.nodeType === 3 ) ? + src.target.parentNode : + src.target; + + this.currentTarget = src.currentTarget; + this.relatedTarget = src.relatedTarget; + + // Event type + } else { + this.type = src; + } + + // Put explicitly provided properties onto the event object + if ( props ) { + jQuery.extend( this, props ); + } + + // Create a timestamp if incoming event doesn't have one + this.timeStamp = src && src.timeStamp || Date.now(); + + // Mark it as fixed + this[ jQuery.expando ] = true; +}; + +// jQuery.Event is based on DOM3 Events as specified by the ECMAScript Language Binding +// https://www.w3.org/TR/2003/WD-DOM-Level-3-Events-20030331/ecma-script-binding.html +jQuery.Event.prototype = { + constructor: jQuery.Event, + isDefaultPrevented: returnFalse, + isPropagationStopped: returnFalse, + isImmediatePropagationStopped: returnFalse, + isSimulated: false, + + preventDefault: function() { + var e = this.originalEvent; + + this.isDefaultPrevented = returnTrue; + + if ( e && !this.isSimulated ) { + e.preventDefault(); + } + }, + stopPropagation: function() { + var e = this.originalEvent; + + this.isPropagationStopped = returnTrue; + + if ( e && !this.isSimulated ) { + e.stopPropagation(); + } + }, + stopImmediatePropagation: function() { + var e = this.originalEvent; + + this.isImmediatePropagationStopped = returnTrue; + + if ( e && !this.isSimulated ) { + e.stopImmediatePropagation(); + } + + this.stopPropagation(); + } +}; + +// Includes all common event props including KeyEvent and MouseEvent specific props +jQuery.each( { + altKey: true, + bubbles: true, + cancelable: true, + changedTouches: true, + ctrlKey: true, + detail: true, + eventPhase: true, + metaKey: true, + pageX: true, + pageY: true, + shiftKey: true, + view: true, + "char": true, + code: true, + charCode: true, + key: true, + keyCode: true, + button: true, + buttons: true, + clientX: true, + clientY: true, + offsetX: true, + offsetY: true, + pointerId: true, + pointerType: true, + screenX: true, + screenY: true, + targetTouches: true, + toElement: true, + touches: true, + + which: function( event ) { + var button = event.button; + + // Add which for key events + if ( event.which == null && rkeyEvent.test( event.type ) ) { + return event.charCode != null ? event.charCode : event.keyCode; + } + + // Add which for click: 1 === left; 2 === middle; 3 === right + if ( !event.which && button !== undefined && rmouseEvent.test( event.type ) ) { + if ( button & 1 ) { + return 1; + } + + if ( button & 2 ) { + return 3; + } + + if ( button & 4 ) { + return 2; + } + + return 0; + } + + return event.which; + } +}, jQuery.event.addProp ); + +jQuery.each( { focus: "focusin", blur: "focusout" }, function( type, delegateType ) { + jQuery.event.special[ type ] = { + + // Utilize native event if possible so blur/focus sequence is correct + setup: function() { + + // Claim the first handler + // dataPriv.set( this, "focus", ... ) + // dataPriv.set( this, "blur", ... ) + leverageNative( this, type, expectSync ); + + // Return false to allow normal processing in the caller + return false; + }, + trigger: function() { + + // Force setup before trigger + leverageNative( this, type ); + + // Return non-false to allow normal event-path propagation + return true; + }, + + delegateType: delegateType + }; +} ); + +// Create mouseenter/leave events using mouseover/out and event-time checks +// so that event delegation works in jQuery. +// Do the same for pointerenter/pointerleave and pointerover/pointerout +// +// Support: Safari 7 only +// Safari sends mouseenter too often; see: +// https://bugs.chromium.org/p/chromium/issues/detail?id=470258 +// for the description of the bug (it existed in older Chrome versions as well). +jQuery.each( { + mouseenter: "mouseover", + mouseleave: "mouseout", + pointerenter: "pointerover", + pointerleave: "pointerout" +}, function( orig, fix ) { + jQuery.event.special[ orig ] = { + delegateType: fix, + bindType: fix, + + handle: function( event ) { + var ret, + target = this, + related = event.relatedTarget, + handleObj = event.handleObj; + + // For mouseenter/leave call the handler if related is outside the target. + // NB: No relatedTarget if the mouse left/entered the browser window + if ( !related || ( related !== target && !jQuery.contains( target, related ) ) ) { + event.type = handleObj.origType; + ret = handleObj.handler.apply( this, arguments ); + event.type = fix; + } + return ret; + } + }; +} ); + +jQuery.fn.extend( { + + on: function( types, selector, data, fn ) { + return on( this, types, selector, data, fn ); + }, + one: function( types, selector, data, fn ) { + return on( this, types, selector, data, fn, 1 ); + }, + off: function( types, selector, fn ) { + var handleObj, type; + if ( types && types.preventDefault && types.handleObj ) { + + // ( event ) dispatched jQuery.Event + handleObj = types.handleObj; + jQuery( types.delegateTarget ).off( + handleObj.namespace ? + handleObj.origType + "." + handleObj.namespace : + handleObj.origType, + handleObj.selector, + handleObj.handler + ); + return this; + } + if ( typeof types === "object" ) { + + // ( types-object [, selector] ) + for ( type in types ) { + this.off( type, selector, types[ type ] ); + } + return this; + } + if ( selector === false || typeof selector === "function" ) { + + // ( types [, fn] ) + fn = selector; + selector = undefined; + } + if ( fn === false ) { + fn = returnFalse; + } + return this.each( function() { + jQuery.event.remove( this, types, fn, selector ); + } ); + } +} ); + + +var + + // Support: IE <=10 - 11, Edge 12 - 13 only + // In IE/Edge using regex groups here causes severe slowdowns. + // See https://connect.microsoft.com/IE/feedback/details/1736512/ + rnoInnerhtml = /\s*$/g; + +// Prefer a tbody over its parent table for containing new rows +function manipulationTarget( elem, content ) { + if ( nodeName( elem, "table" ) && + nodeName( content.nodeType !== 11 ? content : content.firstChild, "tr" ) ) { + + return jQuery( elem ).children( "tbody" )[ 0 ] || elem; + } + + return elem; +} + +// Replace/restore the type attribute of script elements for safe DOM manipulation +function disableScript( elem ) { + elem.type = ( elem.getAttribute( "type" ) !== null ) + "/" + elem.type; + return elem; +} +function restoreScript( elem ) { + if ( ( elem.type || "" ).slice( 0, 5 ) === "true/" ) { + elem.type = elem.type.slice( 5 ); + } else { + elem.removeAttribute( "type" ); + } + + return elem; +} + +function cloneCopyEvent( src, dest ) { + var i, l, type, pdataOld, udataOld, udataCur, events; + + if ( dest.nodeType !== 1 ) { + return; + } + + // 1. Copy private data: events, handlers, etc. + if ( dataPriv.hasData( src ) ) { + pdataOld = dataPriv.get( src ); + events = pdataOld.events; + + if ( events ) { + dataPriv.remove( dest, "handle events" ); + + for ( type in events ) { + for ( i = 0, l = events[ type ].length; i < l; i++ ) { + jQuery.event.add( dest, type, events[ type ][ i ] ); + } + } + } + } + + // 2. Copy user data + if ( dataUser.hasData( src ) ) { + udataOld = dataUser.access( src ); + udataCur = jQuery.extend( {}, udataOld ); + + dataUser.set( dest, udataCur ); + } +} + +// Fix IE bugs, see support tests +function fixInput( src, dest ) { + var nodeName = dest.nodeName.toLowerCase(); + + // Fails to persist the checked state of a cloned checkbox or radio button. + if ( nodeName === "input" && rcheckableType.test( src.type ) ) { + dest.checked = src.checked; + + // Fails to return the selected option to the default selected state when cloning options + } else if ( nodeName === "input" || nodeName === "textarea" ) { + dest.defaultValue = src.defaultValue; + } +} + +function domManip( collection, args, callback, ignored ) { + + // Flatten any nested arrays + args = flat( args ); + + var fragment, first, scripts, hasScripts, node, doc, + i = 0, + l = collection.length, + iNoClone = l - 1, + value = args[ 0 ], + valueIsFunction = isFunction( value ); + + // We can't cloneNode fragments that contain checked, in WebKit + if ( valueIsFunction || + ( l > 1 && typeof value === "string" && + !support.checkClone && rchecked.test( value ) ) ) { + return collection.each( function( index ) { + var self = collection.eq( index ); + if ( valueIsFunction ) { + args[ 0 ] = value.call( this, index, self.html() ); + } + domManip( self, args, callback, ignored ); + } ); + } + + if ( l ) { + fragment = buildFragment( args, collection[ 0 ].ownerDocument, false, collection, ignored ); + first = fragment.firstChild; + + if ( fragment.childNodes.length === 1 ) { + fragment = first; + } + + // Require either new content or an interest in ignored elements to invoke the callback + if ( first || ignored ) { + scripts = jQuery.map( getAll( fragment, "script" ), disableScript ); + hasScripts = scripts.length; + + // Use the original fragment for the last item + // instead of the first because it can end up + // being emptied incorrectly in certain situations (#8070). + for ( ; i < l; i++ ) { + node = fragment; + + if ( i !== iNoClone ) { + node = jQuery.clone( node, true, true ); + + // Keep references to cloned scripts for later restoration + if ( hasScripts ) { + + // Support: Android <=4.0 only, PhantomJS 1 only + // push.apply(_, arraylike) throws on ancient WebKit + jQuery.merge( scripts, getAll( node, "script" ) ); + } + } + + callback.call( collection[ i ], node, i ); + } + + if ( hasScripts ) { + doc = scripts[ scripts.length - 1 ].ownerDocument; + + // Reenable scripts + jQuery.map( scripts, restoreScript ); + + // Evaluate executable scripts on first document insertion + for ( i = 0; i < hasScripts; i++ ) { + node = scripts[ i ]; + if ( rscriptType.test( node.type || "" ) && + !dataPriv.access( node, "globalEval" ) && + jQuery.contains( doc, node ) ) { + + if ( node.src && ( node.type || "" ).toLowerCase() !== "module" ) { + + // Optional AJAX dependency, but won't run scripts if not present + if ( jQuery._evalUrl && !node.noModule ) { + jQuery._evalUrl( node.src, { + nonce: node.nonce || node.getAttribute( "nonce" ) + }, doc ); + } + } else { + DOMEval( node.textContent.replace( rcleanScript, "" ), node, doc ); + } + } + } + } + } + } + + return collection; +} + +function remove( elem, selector, keepData ) { + var node, + nodes = selector ? jQuery.filter( selector, elem ) : elem, + i = 0; + + for ( ; ( node = nodes[ i ] ) != null; i++ ) { + if ( !keepData && node.nodeType === 1 ) { + jQuery.cleanData( getAll( node ) ); + } + + if ( node.parentNode ) { + if ( keepData && isAttached( node ) ) { + setGlobalEval( getAll( node, "script" ) ); + } + node.parentNode.removeChild( node ); + } + } + + return elem; +} + +jQuery.extend( { + htmlPrefilter: function( html ) { + return html; + }, + + clone: function( elem, dataAndEvents, deepDataAndEvents ) { + var i, l, srcElements, destElements, + clone = elem.cloneNode( true ), + inPage = isAttached( elem ); + + // Fix IE cloning issues + if ( !support.noCloneChecked && ( elem.nodeType === 1 || elem.nodeType === 11 ) && + !jQuery.isXMLDoc( elem ) ) { + + // We eschew Sizzle here for performance reasons: https://jsperf.com/getall-vs-sizzle/2 + destElements = getAll( clone ); + srcElements = getAll( elem ); + + for ( i = 0, l = srcElements.length; i < l; i++ ) { + fixInput( srcElements[ i ], destElements[ i ] ); + } + } + + // Copy the events from the original to the clone + if ( dataAndEvents ) { + if ( deepDataAndEvents ) { + srcElements = srcElements || getAll( elem ); + destElements = destElements || getAll( clone ); + + for ( i = 0, l = srcElements.length; i < l; i++ ) { + cloneCopyEvent( srcElements[ i ], destElements[ i ] ); + } + } else { + cloneCopyEvent( elem, clone ); + } + } + + // Preserve script evaluation history + destElements = getAll( clone, "script" ); + if ( destElements.length > 0 ) { + setGlobalEval( destElements, !inPage && getAll( elem, "script" ) ); + } + + // Return the cloned set + return clone; + }, + + cleanData: function( elems ) { + var data, elem, type, + special = jQuery.event.special, + i = 0; + + for ( ; ( elem = elems[ i ] ) !== undefined; i++ ) { + if ( acceptData( elem ) ) { + if ( ( data = elem[ dataPriv.expando ] ) ) { + if ( data.events ) { + for ( type in data.events ) { + if ( special[ type ] ) { + jQuery.event.remove( elem, type ); + + // This is a shortcut to avoid jQuery.event.remove's overhead + } else { + jQuery.removeEvent( elem, type, data.handle ); + } + } + } + + // Support: Chrome <=35 - 45+ + // Assign undefined instead of using delete, see Data#remove + elem[ dataPriv.expando ] = undefined; + } + if ( elem[ dataUser.expando ] ) { + + // Support: Chrome <=35 - 45+ + // Assign undefined instead of using delete, see Data#remove + elem[ dataUser.expando ] = undefined; + } + } + } + } +} ); + +jQuery.fn.extend( { + detach: function( selector ) { + return remove( this, selector, true ); + }, + + remove: function( selector ) { + return remove( this, selector ); + }, + + text: function( value ) { + return access( this, function( value ) { + return value === undefined ? + jQuery.text( this ) : + this.empty().each( function() { + if ( this.nodeType === 1 || this.nodeType === 11 || this.nodeType === 9 ) { + this.textContent = value; + } + } ); + }, null, value, arguments.length ); + }, + + append: function() { + return domManip( this, arguments, function( elem ) { + if ( this.nodeType === 1 || this.nodeType === 11 || this.nodeType === 9 ) { + var target = manipulationTarget( this, elem ); + target.appendChild( elem ); + } + } ); + }, + + prepend: function() { + return domManip( this, arguments, function( elem ) { + if ( this.nodeType === 1 || this.nodeType === 11 || this.nodeType === 9 ) { + var target = manipulationTarget( this, elem ); + target.insertBefore( elem, target.firstChild ); + } + } ); + }, + + before: function() { + return domManip( this, arguments, function( elem ) { + if ( this.parentNode ) { + this.parentNode.insertBefore( elem, this ); + } + } ); + }, + + after: function() { + return domManip( this, arguments, function( elem ) { + if ( this.parentNode ) { + this.parentNode.insertBefore( elem, this.nextSibling ); + } + } ); + }, + + empty: function() { + var elem, + i = 0; + + for ( ; ( elem = this[ i ] ) != null; i++ ) { + if ( elem.nodeType === 1 ) { + + // Prevent memory leaks + jQuery.cleanData( getAll( elem, false ) ); + + // Remove any remaining nodes + elem.textContent = ""; + } + } + + return this; + }, + + clone: function( dataAndEvents, deepDataAndEvents ) { + dataAndEvents = dataAndEvents == null ? false : dataAndEvents; + deepDataAndEvents = deepDataAndEvents == null ? dataAndEvents : deepDataAndEvents; + + return this.map( function() { + return jQuery.clone( this, dataAndEvents, deepDataAndEvents ); + } ); + }, + + html: function( value ) { + return access( this, function( value ) { + var elem = this[ 0 ] || {}, + i = 0, + l = this.length; + + if ( value === undefined && elem.nodeType === 1 ) { + return elem.innerHTML; + } + + // See if we can take a shortcut and just use innerHTML + if ( typeof value === "string" && !rnoInnerhtml.test( value ) && + !wrapMap[ ( rtagName.exec( value ) || [ "", "" ] )[ 1 ].toLowerCase() ] ) { + + value = jQuery.htmlPrefilter( value ); + + try { + for ( ; i < l; i++ ) { + elem = this[ i ] || {}; + + // Remove element nodes and prevent memory leaks + if ( elem.nodeType === 1 ) { + jQuery.cleanData( getAll( elem, false ) ); + elem.innerHTML = value; + } + } + + elem = 0; + + // If using innerHTML throws an exception, use the fallback method + } catch ( e ) {} + } + + if ( elem ) { + this.empty().append( value ); + } + }, null, value, arguments.length ); + }, + + replaceWith: function() { + var ignored = []; + + // Make the changes, replacing each non-ignored context element with the new content + return domManip( this, arguments, function( elem ) { + var parent = this.parentNode; + + if ( jQuery.inArray( this, ignored ) < 0 ) { + jQuery.cleanData( getAll( this ) ); + if ( parent ) { + parent.replaceChild( elem, this ); + } + } + + // Force callback invocation + }, ignored ); + } +} ); + +jQuery.each( { + appendTo: "append", + prependTo: "prepend", + insertBefore: "before", + insertAfter: "after", + replaceAll: "replaceWith" +}, function( name, original ) { + jQuery.fn[ name ] = function( selector ) { + var elems, + ret = [], + insert = jQuery( selector ), + last = insert.length - 1, + i = 0; + + for ( ; i <= last; i++ ) { + elems = i === last ? this : this.clone( true ); + jQuery( insert[ i ] )[ original ]( elems ); + + // Support: Android <=4.0 only, PhantomJS 1 only + // .get() because push.apply(_, arraylike) throws on ancient WebKit + push.apply( ret, elems.get() ); + } + + return this.pushStack( ret ); + }; +} ); +var rnumnonpx = new RegExp( "^(" + pnum + ")(?!px)[a-z%]+$", "i" ); + +var getStyles = function( elem ) { + + // Support: IE <=11 only, Firefox <=30 (#15098, #14150) + // IE throws on elements created in popups + // FF meanwhile throws on frame elements through "defaultView.getComputedStyle" + var view = elem.ownerDocument.defaultView; + + if ( !view || !view.opener ) { + view = window; + } + + return view.getComputedStyle( elem ); + }; + +var swap = function( elem, options, callback ) { + var ret, name, + old = {}; + + // Remember the old values, and insert the new ones + for ( name in options ) { + old[ name ] = elem.style[ name ]; + elem.style[ name ] = options[ name ]; + } + + ret = callback.call( elem ); + + // Revert the old values + for ( name in options ) { + elem.style[ name ] = old[ name ]; + } + + return ret; +}; + + +var rboxStyle = new RegExp( cssExpand.join( "|" ), "i" ); + + + +( function() { + + // Executing both pixelPosition & boxSizingReliable tests require only one layout + // so they're executed at the same time to save the second computation. + function computeStyleTests() { + + // This is a singleton, we need to execute it only once + if ( !div ) { + return; + } + + container.style.cssText = "position:absolute;left:-11111px;width:60px;" + + "margin-top:1px;padding:0;border:0"; + div.style.cssText = + "position:relative;display:block;box-sizing:border-box;overflow:scroll;" + + "margin:auto;border:1px;padding:1px;" + + "width:60%;top:1%"; + documentElement.appendChild( container ).appendChild( div ); + + var divStyle = window.getComputedStyle( div ); + pixelPositionVal = divStyle.top !== "1%"; + + // Support: Android 4.0 - 4.3 only, Firefox <=3 - 44 + reliableMarginLeftVal = roundPixelMeasures( divStyle.marginLeft ) === 12; + + // Support: Android 4.0 - 4.3 only, Safari <=9.1 - 10.1, iOS <=7.0 - 9.3 + // Some styles come back with percentage values, even though they shouldn't + div.style.right = "60%"; + pixelBoxStylesVal = roundPixelMeasures( divStyle.right ) === 36; + + // Support: IE 9 - 11 only + // Detect misreporting of content dimensions for box-sizing:border-box elements + boxSizingReliableVal = roundPixelMeasures( divStyle.width ) === 36; + + // Support: IE 9 only + // Detect overflow:scroll screwiness (gh-3699) + // Support: Chrome <=64 + // Don't get tricked when zoom affects offsetWidth (gh-4029) + div.style.position = "absolute"; + scrollboxSizeVal = roundPixelMeasures( div.offsetWidth / 3 ) === 12; + + documentElement.removeChild( container ); + + // Nullify the div so it wouldn't be stored in the memory and + // it will also be a sign that checks already performed + div = null; + } + + function roundPixelMeasures( measure ) { + return Math.round( parseFloat( measure ) ); + } + + var pixelPositionVal, boxSizingReliableVal, scrollboxSizeVal, pixelBoxStylesVal, + reliableTrDimensionsVal, reliableMarginLeftVal, + container = document.createElement( "div" ), + div = document.createElement( "div" ); + + // Finish early in limited (non-browser) environments + if ( !div.style ) { + return; + } + + // Support: IE <=9 - 11 only + // Style of cloned element affects source element cloned (#8908) + div.style.backgroundClip = "content-box"; + div.cloneNode( true ).style.backgroundClip = ""; + support.clearCloneStyle = div.style.backgroundClip === "content-box"; + + jQuery.extend( support, { + boxSizingReliable: function() { + computeStyleTests(); + return boxSizingReliableVal; + }, + pixelBoxStyles: function() { + computeStyleTests(); + return pixelBoxStylesVal; + }, + pixelPosition: function() { + computeStyleTests(); + return pixelPositionVal; + }, + reliableMarginLeft: function() { + computeStyleTests(); + return reliableMarginLeftVal; + }, + scrollboxSize: function() { + computeStyleTests(); + return scrollboxSizeVal; + }, + + // Support: IE 9 - 11+, Edge 15 - 18+ + // IE/Edge misreport `getComputedStyle` of table rows with width/height + // set in CSS while `offset*` properties report correct values. + // Behavior in IE 9 is more subtle than in newer versions & it passes + // some versions of this test; make sure not to make it pass there! + reliableTrDimensions: function() { + var table, tr, trChild, trStyle; + if ( reliableTrDimensionsVal == null ) { + table = document.createElement( "table" ); + tr = document.createElement( "tr" ); + trChild = document.createElement( "div" ); + + table.style.cssText = "position:absolute;left:-11111px"; + tr.style.height = "1px"; + trChild.style.height = "9px"; + + documentElement + .appendChild( table ) + .appendChild( tr ) + .appendChild( trChild ); + + trStyle = window.getComputedStyle( tr ); + reliableTrDimensionsVal = parseInt( trStyle.height ) > 3; + + documentElement.removeChild( table ); + } + return reliableTrDimensionsVal; + } + } ); +} )(); + + +function curCSS( elem, name, computed ) { + var width, minWidth, maxWidth, ret, + + // Support: Firefox 51+ + // Retrieving style before computed somehow + // fixes an issue with getting wrong values + // on detached elements + style = elem.style; + + computed = computed || getStyles( elem ); + + // getPropertyValue is needed for: + // .css('filter') (IE 9 only, #12537) + // .css('--customProperty) (#3144) + if ( computed ) { + ret = computed.getPropertyValue( name ) || computed[ name ]; + + if ( ret === "" && !isAttached( elem ) ) { + ret = jQuery.style( elem, name ); + } + + // A tribute to the "awesome hack by Dean Edwards" + // Android Browser returns percentage for some values, + // but width seems to be reliably pixels. + // This is against the CSSOM draft spec: + // https://drafts.csswg.org/cssom/#resolved-values + if ( !support.pixelBoxStyles() && rnumnonpx.test( ret ) && rboxStyle.test( name ) ) { + + // Remember the original values + width = style.width; + minWidth = style.minWidth; + maxWidth = style.maxWidth; + + // Put in the new values to get a computed value out + style.minWidth = style.maxWidth = style.width = ret; + ret = computed.width; + + // Revert the changed values + style.width = width; + style.minWidth = minWidth; + style.maxWidth = maxWidth; + } + } + + return ret !== undefined ? + + // Support: IE <=9 - 11 only + // IE returns zIndex value as an integer. + ret + "" : + ret; +} + + +function addGetHookIf( conditionFn, hookFn ) { + + // Define the hook, we'll check on the first run if it's really needed. + return { + get: function() { + if ( conditionFn() ) { + + // Hook not needed (or it's not possible to use it due + // to missing dependency), remove it. + delete this.get; + return; + } + + // Hook needed; redefine it so that the support test is not executed again. + return ( this.get = hookFn ).apply( this, arguments ); + } + }; +} + + +var cssPrefixes = [ "Webkit", "Moz", "ms" ], + emptyStyle = document.createElement( "div" ).style, + vendorProps = {}; + +// Return a vendor-prefixed property or undefined +function vendorPropName( name ) { + + // Check for vendor prefixed names + var capName = name[ 0 ].toUpperCase() + name.slice( 1 ), + i = cssPrefixes.length; + + while ( i-- ) { + name = cssPrefixes[ i ] + capName; + if ( name in emptyStyle ) { + return name; + } + } +} + +// Return a potentially-mapped jQuery.cssProps or vendor prefixed property +function finalPropName( name ) { + var final = jQuery.cssProps[ name ] || vendorProps[ name ]; + + if ( final ) { + return final; + } + if ( name in emptyStyle ) { + return name; + } + return vendorProps[ name ] = vendorPropName( name ) || name; +} + + +var + + // Swappable if display is none or starts with table + // except "table", "table-cell", or "table-caption" + // See here for display values: https://developer.mozilla.org/en-US/docs/CSS/display + rdisplayswap = /^(none|table(?!-c[ea]).+)/, + rcustomProp = /^--/, + cssShow = { position: "absolute", visibility: "hidden", display: "block" }, + cssNormalTransform = { + letterSpacing: "0", + fontWeight: "400" + }; + +function setPositiveNumber( _elem, value, subtract ) { + + // Any relative (+/-) values have already been + // normalized at this point + var matches = rcssNum.exec( value ); + return matches ? + + // Guard against undefined "subtract", e.g., when used as in cssHooks + Math.max( 0, matches[ 2 ] - ( subtract || 0 ) ) + ( matches[ 3 ] || "px" ) : + value; +} + +function boxModelAdjustment( elem, dimension, box, isBorderBox, styles, computedVal ) { + var i = dimension === "width" ? 1 : 0, + extra = 0, + delta = 0; + + // Adjustment may not be necessary + if ( box === ( isBorderBox ? "border" : "content" ) ) { + return 0; + } + + for ( ; i < 4; i += 2 ) { + + // Both box models exclude margin + if ( box === "margin" ) { + delta += jQuery.css( elem, box + cssExpand[ i ], true, styles ); + } + + // If we get here with a content-box, we're seeking "padding" or "border" or "margin" + if ( !isBorderBox ) { + + // Add padding + delta += jQuery.css( elem, "padding" + cssExpand[ i ], true, styles ); + + // For "border" or "margin", add border + if ( box !== "padding" ) { + delta += jQuery.css( elem, "border" + cssExpand[ i ] + "Width", true, styles ); + + // But still keep track of it otherwise + } else { + extra += jQuery.css( elem, "border" + cssExpand[ i ] + "Width", true, styles ); + } + + // If we get here with a border-box (content + padding + border), we're seeking "content" or + // "padding" or "margin" + } else { + + // For "content", subtract padding + if ( box === "content" ) { + delta -= jQuery.css( elem, "padding" + cssExpand[ i ], true, styles ); + } + + // For "content" or "padding", subtract border + if ( box !== "margin" ) { + delta -= jQuery.css( elem, "border" + cssExpand[ i ] + "Width", true, styles ); + } + } + } + + // Account for positive content-box scroll gutter when requested by providing computedVal + if ( !isBorderBox && computedVal >= 0 ) { + + // offsetWidth/offsetHeight is a rounded sum of content, padding, scroll gutter, and border + // Assuming integer scroll gutter, subtract the rest and round down + delta += Math.max( 0, Math.ceil( + elem[ "offset" + dimension[ 0 ].toUpperCase() + dimension.slice( 1 ) ] - + computedVal - + delta - + extra - + 0.5 + + // If offsetWidth/offsetHeight is unknown, then we can't determine content-box scroll gutter + // Use an explicit zero to avoid NaN (gh-3964) + ) ) || 0; + } + + return delta; +} + +function getWidthOrHeight( elem, dimension, extra ) { + + // Start with computed style + var styles = getStyles( elem ), + + // To avoid forcing a reflow, only fetch boxSizing if we need it (gh-4322). + // Fake content-box until we know it's needed to know the true value. + boxSizingNeeded = !support.boxSizingReliable() || extra, + isBorderBox = boxSizingNeeded && + jQuery.css( elem, "boxSizing", false, styles ) === "border-box", + valueIsBorderBox = isBorderBox, + + val = curCSS( elem, dimension, styles ), + offsetProp = "offset" + dimension[ 0 ].toUpperCase() + dimension.slice( 1 ); + + // Support: Firefox <=54 + // Return a confounding non-pixel value or feign ignorance, as appropriate. + if ( rnumnonpx.test( val ) ) { + if ( !extra ) { + return val; + } + val = "auto"; + } + + + // Support: IE 9 - 11 only + // Use offsetWidth/offsetHeight for when box sizing is unreliable. + // In those cases, the computed value can be trusted to be border-box. + if ( ( !support.boxSizingReliable() && isBorderBox || + + // Support: IE 10 - 11+, Edge 15 - 18+ + // IE/Edge misreport `getComputedStyle` of table rows with width/height + // set in CSS while `offset*` properties report correct values. + // Interestingly, in some cases IE 9 doesn't suffer from this issue. + !support.reliableTrDimensions() && nodeName( elem, "tr" ) || + + // Fall back to offsetWidth/offsetHeight when value is "auto" + // This happens for inline elements with no explicit setting (gh-3571) + val === "auto" || + + // Support: Android <=4.1 - 4.3 only + // Also use offsetWidth/offsetHeight for misreported inline dimensions (gh-3602) + !parseFloat( val ) && jQuery.css( elem, "display", false, styles ) === "inline" ) && + + // Make sure the element is visible & connected + elem.getClientRects().length ) { + + isBorderBox = jQuery.css( elem, "boxSizing", false, styles ) === "border-box"; + + // Where available, offsetWidth/offsetHeight approximate border box dimensions. + // Where not available (e.g., SVG), assume unreliable box-sizing and interpret the + // retrieved value as a content box dimension. + valueIsBorderBox = offsetProp in elem; + if ( valueIsBorderBox ) { + val = elem[ offsetProp ]; + } + } + + // Normalize "" and auto + val = parseFloat( val ) || 0; + + // Adjust for the element's box model + return ( val + + boxModelAdjustment( + elem, + dimension, + extra || ( isBorderBox ? "border" : "content" ), + valueIsBorderBox, + styles, + + // Provide the current computed size to request scroll gutter calculation (gh-3589) + val + ) + ) + "px"; +} + +jQuery.extend( { + + // Add in style property hooks for overriding the default + // behavior of getting and setting a style property + cssHooks: { + opacity: { + get: function( elem, computed ) { + if ( computed ) { + + // We should always get a number back from opacity + var ret = curCSS( elem, "opacity" ); + return ret === "" ? "1" : ret; + } + } + } + }, + + // Don't automatically add "px" to these possibly-unitless properties + cssNumber: { + "animationIterationCount": true, + "columnCount": true, + "fillOpacity": true, + "flexGrow": true, + "flexShrink": true, + "fontWeight": true, + "gridArea": true, + "gridColumn": true, + "gridColumnEnd": true, + "gridColumnStart": true, + "gridRow": true, + "gridRowEnd": true, + "gridRowStart": true, + "lineHeight": true, + "opacity": true, + "order": true, + "orphans": true, + "widows": true, + "zIndex": true, + "zoom": true + }, + + // Add in properties whose names you wish to fix before + // setting or getting the value + cssProps: {}, + + // Get and set the style property on a DOM Node + style: function( elem, name, value, extra ) { + + // Don't set styles on text and comment nodes + if ( !elem || elem.nodeType === 3 || elem.nodeType === 8 || !elem.style ) { + return; + } + + // Make sure that we're working with the right name + var ret, type, hooks, + origName = camelCase( name ), + isCustomProp = rcustomProp.test( name ), + style = elem.style; + + // Make sure that we're working with the right name. We don't + // want to query the value if it is a CSS custom property + // since they are user-defined. + if ( !isCustomProp ) { + name = finalPropName( origName ); + } + + // Gets hook for the prefixed version, then unprefixed version + hooks = jQuery.cssHooks[ name ] || jQuery.cssHooks[ origName ]; + + // Check if we're setting a value + if ( value !== undefined ) { + type = typeof value; + + // Convert "+=" or "-=" to relative numbers (#7345) + if ( type === "string" && ( ret = rcssNum.exec( value ) ) && ret[ 1 ] ) { + value = adjustCSS( elem, name, ret ); + + // Fixes bug #9237 + type = "number"; + } + + // Make sure that null and NaN values aren't set (#7116) + if ( value == null || value !== value ) { + return; + } + + // If a number was passed in, add the unit (except for certain CSS properties) + // The isCustomProp check can be removed in jQuery 4.0 when we only auto-append + // "px" to a few hardcoded values. + if ( type === "number" && !isCustomProp ) { + value += ret && ret[ 3 ] || ( jQuery.cssNumber[ origName ] ? "" : "px" ); + } + + // background-* props affect original clone's values + if ( !support.clearCloneStyle && value === "" && name.indexOf( "background" ) === 0 ) { + style[ name ] = "inherit"; + } + + // If a hook was provided, use that value, otherwise just set the specified value + if ( !hooks || !( "set" in hooks ) || + ( value = hooks.set( elem, value, extra ) ) !== undefined ) { + + if ( isCustomProp ) { + style.setProperty( name, value ); + } else { + style[ name ] = value; + } + } + + } else { + + // If a hook was provided get the non-computed value from there + if ( hooks && "get" in hooks && + ( ret = hooks.get( elem, false, extra ) ) !== undefined ) { + + return ret; + } + + // Otherwise just get the value from the style object + return style[ name ]; + } + }, + + css: function( elem, name, extra, styles ) { + var val, num, hooks, + origName = camelCase( name ), + isCustomProp = rcustomProp.test( name ); + + // Make sure that we're working with the right name. We don't + // want to modify the value if it is a CSS custom property + // since they are user-defined. + if ( !isCustomProp ) { + name = finalPropName( origName ); + } + + // Try prefixed name followed by the unprefixed name + hooks = jQuery.cssHooks[ name ] || jQuery.cssHooks[ origName ]; + + // If a hook was provided get the computed value from there + if ( hooks && "get" in hooks ) { + val = hooks.get( elem, true, extra ); + } + + // Otherwise, if a way to get the computed value exists, use that + if ( val === undefined ) { + val = curCSS( elem, name, styles ); + } + + // Convert "normal" to computed value + if ( val === "normal" && name in cssNormalTransform ) { + val = cssNormalTransform[ name ]; + } + + // Make numeric if forced or a qualifier was provided and val looks numeric + if ( extra === "" || extra ) { + num = parseFloat( val ); + return extra === true || isFinite( num ) ? num || 0 : val; + } + + return val; + } +} ); + +jQuery.each( [ "height", "width" ], function( _i, dimension ) { + jQuery.cssHooks[ dimension ] = { + get: function( elem, computed, extra ) { + if ( computed ) { + + // Certain elements can have dimension info if we invisibly show them + // but it must have a current display style that would benefit + return rdisplayswap.test( jQuery.css( elem, "display" ) ) && + + // Support: Safari 8+ + // Table columns in Safari have non-zero offsetWidth & zero + // getBoundingClientRect().width unless display is changed. + // Support: IE <=11 only + // Running getBoundingClientRect on a disconnected node + // in IE throws an error. + ( !elem.getClientRects().length || !elem.getBoundingClientRect().width ) ? + swap( elem, cssShow, function() { + return getWidthOrHeight( elem, dimension, extra ); + } ) : + getWidthOrHeight( elem, dimension, extra ); + } + }, + + set: function( elem, value, extra ) { + var matches, + styles = getStyles( elem ), + + // Only read styles.position if the test has a chance to fail + // to avoid forcing a reflow. + scrollboxSizeBuggy = !support.scrollboxSize() && + styles.position === "absolute", + + // To avoid forcing a reflow, only fetch boxSizing if we need it (gh-3991) + boxSizingNeeded = scrollboxSizeBuggy || extra, + isBorderBox = boxSizingNeeded && + jQuery.css( elem, "boxSizing", false, styles ) === "border-box", + subtract = extra ? + boxModelAdjustment( + elem, + dimension, + extra, + isBorderBox, + styles + ) : + 0; + + // Account for unreliable border-box dimensions by comparing offset* to computed and + // faking a content-box to get border and padding (gh-3699) + if ( isBorderBox && scrollboxSizeBuggy ) { + subtract -= Math.ceil( + elem[ "offset" + dimension[ 0 ].toUpperCase() + dimension.slice( 1 ) ] - + parseFloat( styles[ dimension ] ) - + boxModelAdjustment( elem, dimension, "border", false, styles ) - + 0.5 + ); + } + + // Convert to pixels if value adjustment is needed + if ( subtract && ( matches = rcssNum.exec( value ) ) && + ( matches[ 3 ] || "px" ) !== "px" ) { + + elem.style[ dimension ] = value; + value = jQuery.css( elem, dimension ); + } + + return setPositiveNumber( elem, value, subtract ); + } + }; +} ); + +jQuery.cssHooks.marginLeft = addGetHookIf( support.reliableMarginLeft, + function( elem, computed ) { + if ( computed ) { + return ( parseFloat( curCSS( elem, "marginLeft" ) ) || + elem.getBoundingClientRect().left - + swap( elem, { marginLeft: 0 }, function() { + return elem.getBoundingClientRect().left; + } ) + ) + "px"; + } + } +); + +// These hooks are used by animate to expand properties +jQuery.each( { + margin: "", + padding: "", + border: "Width" +}, function( prefix, suffix ) { + jQuery.cssHooks[ prefix + suffix ] = { + expand: function( value ) { + var i = 0, + expanded = {}, + + // Assumes a single number if not a string + parts = typeof value === "string" ? value.split( " " ) : [ value ]; + + for ( ; i < 4; i++ ) { + expanded[ prefix + cssExpand[ i ] + suffix ] = + parts[ i ] || parts[ i - 2 ] || parts[ 0 ]; + } + + return expanded; + } + }; + + if ( prefix !== "margin" ) { + jQuery.cssHooks[ prefix + suffix ].set = setPositiveNumber; + } +} ); + +jQuery.fn.extend( { + css: function( name, value ) { + return access( this, function( elem, name, value ) { + var styles, len, + map = {}, + i = 0; + + if ( Array.isArray( name ) ) { + styles = getStyles( elem ); + len = name.length; + + for ( ; i < len; i++ ) { + map[ name[ i ] ] = jQuery.css( elem, name[ i ], false, styles ); + } + + return map; + } + + return value !== undefined ? + jQuery.style( elem, name, value ) : + jQuery.css( elem, name ); + }, name, value, arguments.length > 1 ); + } +} ); + + +function Tween( elem, options, prop, end, easing ) { + return new Tween.prototype.init( elem, options, prop, end, easing ); +} +jQuery.Tween = Tween; + +Tween.prototype = { + constructor: Tween, + init: function( elem, options, prop, end, easing, unit ) { + this.elem = elem; + this.prop = prop; + this.easing = easing || jQuery.easing._default; + this.options = options; + this.start = this.now = this.cur(); + this.end = end; + this.unit = unit || ( jQuery.cssNumber[ prop ] ? "" : "px" ); + }, + cur: function() { + var hooks = Tween.propHooks[ this.prop ]; + + return hooks && hooks.get ? + hooks.get( this ) : + Tween.propHooks._default.get( this ); + }, + run: function( percent ) { + var eased, + hooks = Tween.propHooks[ this.prop ]; + + if ( this.options.duration ) { + this.pos = eased = jQuery.easing[ this.easing ]( + percent, this.options.duration * percent, 0, 1, this.options.duration + ); + } else { + this.pos = eased = percent; + } + this.now = ( this.end - this.start ) * eased + this.start; + + if ( this.options.step ) { + this.options.step.call( this.elem, this.now, this ); + } + + if ( hooks && hooks.set ) { + hooks.set( this ); + } else { + Tween.propHooks._default.set( this ); + } + return this; + } +}; + +Tween.prototype.init.prototype = Tween.prototype; + +Tween.propHooks = { + _default: { + get: function( tween ) { + var result; + + // Use a property on the element directly when it is not a DOM element, + // or when there is no matching style property that exists. + if ( tween.elem.nodeType !== 1 || + tween.elem[ tween.prop ] != null && tween.elem.style[ tween.prop ] == null ) { + return tween.elem[ tween.prop ]; + } + + // Passing an empty string as a 3rd parameter to .css will automatically + // attempt a parseFloat and fallback to a string if the parse fails. + // Simple values such as "10px" are parsed to Float; + // complex values such as "rotate(1rad)" are returned as-is. + result = jQuery.css( tween.elem, tween.prop, "" ); + + // Empty strings, null, undefined and "auto" are converted to 0. + return !result || result === "auto" ? 0 : result; + }, + set: function( tween ) { + + // Use step hook for back compat. + // Use cssHook if its there. + // Use .style if available and use plain properties where available. + if ( jQuery.fx.step[ tween.prop ] ) { + jQuery.fx.step[ tween.prop ]( tween ); + } else if ( tween.elem.nodeType === 1 && ( + jQuery.cssHooks[ tween.prop ] || + tween.elem.style[ finalPropName( tween.prop ) ] != null ) ) { + jQuery.style( tween.elem, tween.prop, tween.now + tween.unit ); + } else { + tween.elem[ tween.prop ] = tween.now; + } + } + } +}; + +// Support: IE <=9 only +// Panic based approach to setting things on disconnected nodes +Tween.propHooks.scrollTop = Tween.propHooks.scrollLeft = { + set: function( tween ) { + if ( tween.elem.nodeType && tween.elem.parentNode ) { + tween.elem[ tween.prop ] = tween.now; + } + } +}; + +jQuery.easing = { + linear: function( p ) { + return p; + }, + swing: function( p ) { + return 0.5 - Math.cos( p * Math.PI ) / 2; + }, + _default: "swing" +}; + +jQuery.fx = Tween.prototype.init; + +// Back compat <1.8 extension point +jQuery.fx.step = {}; + + + + +var + fxNow, inProgress, + rfxtypes = /^(?:toggle|show|hide)$/, + rrun = /queueHooks$/; + +function schedule() { + if ( inProgress ) { + if ( document.hidden === false && window.requestAnimationFrame ) { + window.requestAnimationFrame( schedule ); + } else { + window.setTimeout( schedule, jQuery.fx.interval ); + } + + jQuery.fx.tick(); + } +} + +// Animations created synchronously will run synchronously +function createFxNow() { + window.setTimeout( function() { + fxNow = undefined; + } ); + return ( fxNow = Date.now() ); +} + +// Generate parameters to create a standard animation +function genFx( type, includeWidth ) { + var which, + i = 0, + attrs = { height: type }; + + // If we include width, step value is 1 to do all cssExpand values, + // otherwise step value is 2 to skip over Left and Right + includeWidth = includeWidth ? 1 : 0; + for ( ; i < 4; i += 2 - includeWidth ) { + which = cssExpand[ i ]; + attrs[ "margin" + which ] = attrs[ "padding" + which ] = type; + } + + if ( includeWidth ) { + attrs.opacity = attrs.width = type; + } + + return attrs; +} + +function createTween( value, prop, animation ) { + var tween, + collection = ( Animation.tweeners[ prop ] || [] ).concat( Animation.tweeners[ "*" ] ), + index = 0, + length = collection.length; + for ( ; index < length; index++ ) { + if ( ( tween = collection[ index ].call( animation, prop, value ) ) ) { + + // We're done with this property + return tween; + } + } +} + +function defaultPrefilter( elem, props, opts ) { + var prop, value, toggle, hooks, oldfire, propTween, restoreDisplay, display, + isBox = "width" in props || "height" in props, + anim = this, + orig = {}, + style = elem.style, + hidden = elem.nodeType && isHiddenWithinTree( elem ), + dataShow = dataPriv.get( elem, "fxshow" ); + + // Queue-skipping animations hijack the fx hooks + if ( !opts.queue ) { + hooks = jQuery._queueHooks( elem, "fx" ); + if ( hooks.unqueued == null ) { + hooks.unqueued = 0; + oldfire = hooks.empty.fire; + hooks.empty.fire = function() { + if ( !hooks.unqueued ) { + oldfire(); + } + }; + } + hooks.unqueued++; + + anim.always( function() { + + // Ensure the complete handler is called before this completes + anim.always( function() { + hooks.unqueued--; + if ( !jQuery.queue( elem, "fx" ).length ) { + hooks.empty.fire(); + } + } ); + } ); + } + + // Detect show/hide animations + for ( prop in props ) { + value = props[ prop ]; + if ( rfxtypes.test( value ) ) { + delete props[ prop ]; + toggle = toggle || value === "toggle"; + if ( value === ( hidden ? "hide" : "show" ) ) { + + // Pretend to be hidden if this is a "show" and + // there is still data from a stopped show/hide + if ( value === "show" && dataShow && dataShow[ prop ] !== undefined ) { + hidden = true; + + // Ignore all other no-op show/hide data + } else { + continue; + } + } + orig[ prop ] = dataShow && dataShow[ prop ] || jQuery.style( elem, prop ); + } + } + + // Bail out if this is a no-op like .hide().hide() + propTween = !jQuery.isEmptyObject( props ); + if ( !propTween && jQuery.isEmptyObject( orig ) ) { + return; + } + + // Restrict "overflow" and "display" styles during box animations + if ( isBox && elem.nodeType === 1 ) { + + // Support: IE <=9 - 11, Edge 12 - 15 + // Record all 3 overflow attributes because IE does not infer the shorthand + // from identically-valued overflowX and overflowY and Edge just mirrors + // the overflowX value there. + opts.overflow = [ style.overflow, style.overflowX, style.overflowY ]; + + // Identify a display type, preferring old show/hide data over the CSS cascade + restoreDisplay = dataShow && dataShow.display; + if ( restoreDisplay == null ) { + restoreDisplay = dataPriv.get( elem, "display" ); + } + display = jQuery.css( elem, "display" ); + if ( display === "none" ) { + if ( restoreDisplay ) { + display = restoreDisplay; + } else { + + // Get nonempty value(s) by temporarily forcing visibility + showHide( [ elem ], true ); + restoreDisplay = elem.style.display || restoreDisplay; + display = jQuery.css( elem, "display" ); + showHide( [ elem ] ); + } + } + + // Animate inline elements as inline-block + if ( display === "inline" || display === "inline-block" && restoreDisplay != null ) { + if ( jQuery.css( elem, "float" ) === "none" ) { + + // Restore the original display value at the end of pure show/hide animations + if ( !propTween ) { + anim.done( function() { + style.display = restoreDisplay; + } ); + if ( restoreDisplay == null ) { + display = style.display; + restoreDisplay = display === "none" ? "" : display; + } + } + style.display = "inline-block"; + } + } + } + + if ( opts.overflow ) { + style.overflow = "hidden"; + anim.always( function() { + style.overflow = opts.overflow[ 0 ]; + style.overflowX = opts.overflow[ 1 ]; + style.overflowY = opts.overflow[ 2 ]; + } ); + } + + // Implement show/hide animations + propTween = false; + for ( prop in orig ) { + + // General show/hide setup for this element animation + if ( !propTween ) { + if ( dataShow ) { + if ( "hidden" in dataShow ) { + hidden = dataShow.hidden; + } + } else { + dataShow = dataPriv.access( elem, "fxshow", { display: restoreDisplay } ); + } + + // Store hidden/visible for toggle so `.stop().toggle()` "reverses" + if ( toggle ) { + dataShow.hidden = !hidden; + } + + // Show elements before animating them + if ( hidden ) { + showHide( [ elem ], true ); + } + + /* eslint-disable no-loop-func */ + + anim.done( function() { + + /* eslint-enable no-loop-func */ + + // The final step of a "hide" animation is actually hiding the element + if ( !hidden ) { + showHide( [ elem ] ); + } + dataPriv.remove( elem, "fxshow" ); + for ( prop in orig ) { + jQuery.style( elem, prop, orig[ prop ] ); + } + } ); + } + + // Per-property setup + propTween = createTween( hidden ? dataShow[ prop ] : 0, prop, anim ); + if ( !( prop in dataShow ) ) { + dataShow[ prop ] = propTween.start; + if ( hidden ) { + propTween.end = propTween.start; + propTween.start = 0; + } + } + } +} + +function propFilter( props, specialEasing ) { + var index, name, easing, value, hooks; + + // camelCase, specialEasing and expand cssHook pass + for ( index in props ) { + name = camelCase( index ); + easing = specialEasing[ name ]; + value = props[ index ]; + if ( Array.isArray( value ) ) { + easing = value[ 1 ]; + value = props[ index ] = value[ 0 ]; + } + + if ( index !== name ) { + props[ name ] = value; + delete props[ index ]; + } + + hooks = jQuery.cssHooks[ name ]; + if ( hooks && "expand" in hooks ) { + value = hooks.expand( value ); + delete props[ name ]; + + // Not quite $.extend, this won't overwrite existing keys. + // Reusing 'index' because we have the correct "name" + for ( index in value ) { + if ( !( index in props ) ) { + props[ index ] = value[ index ]; + specialEasing[ index ] = easing; + } + } + } else { + specialEasing[ name ] = easing; + } + } +} + +function Animation( elem, properties, options ) { + var result, + stopped, + index = 0, + length = Animation.prefilters.length, + deferred = jQuery.Deferred().always( function() { + + // Don't match elem in the :animated selector + delete tick.elem; + } ), + tick = function() { + if ( stopped ) { + return false; + } + var currentTime = fxNow || createFxNow(), + remaining = Math.max( 0, animation.startTime + animation.duration - currentTime ), + + // Support: Android 2.3 only + // Archaic crash bug won't allow us to use `1 - ( 0.5 || 0 )` (#12497) + temp = remaining / animation.duration || 0, + percent = 1 - temp, + index = 0, + length = animation.tweens.length; + + for ( ; index < length; index++ ) { + animation.tweens[ index ].run( percent ); + } + + deferred.notifyWith( elem, [ animation, percent, remaining ] ); + + // If there's more to do, yield + if ( percent < 1 && length ) { + return remaining; + } + + // If this was an empty animation, synthesize a final progress notification + if ( !length ) { + deferred.notifyWith( elem, [ animation, 1, 0 ] ); + } + + // Resolve the animation and report its conclusion + deferred.resolveWith( elem, [ animation ] ); + return false; + }, + animation = deferred.promise( { + elem: elem, + props: jQuery.extend( {}, properties ), + opts: jQuery.extend( true, { + specialEasing: {}, + easing: jQuery.easing._default + }, options ), + originalProperties: properties, + originalOptions: options, + startTime: fxNow || createFxNow(), + duration: options.duration, + tweens: [], + createTween: function( prop, end ) { + var tween = jQuery.Tween( elem, animation.opts, prop, end, + animation.opts.specialEasing[ prop ] || animation.opts.easing ); + animation.tweens.push( tween ); + return tween; + }, + stop: function( gotoEnd ) { + var index = 0, + + // If we are going to the end, we want to run all the tweens + // otherwise we skip this part + length = gotoEnd ? animation.tweens.length : 0; + if ( stopped ) { + return this; + } + stopped = true; + for ( ; index < length; index++ ) { + animation.tweens[ index ].run( 1 ); + } + + // Resolve when we played the last frame; otherwise, reject + if ( gotoEnd ) { + deferred.notifyWith( elem, [ animation, 1, 0 ] ); + deferred.resolveWith( elem, [ animation, gotoEnd ] ); + } else { + deferred.rejectWith( elem, [ animation, gotoEnd ] ); + } + return this; + } + } ), + props = animation.props; + + propFilter( props, animation.opts.specialEasing ); + + for ( ; index < length; index++ ) { + result = Animation.prefilters[ index ].call( animation, elem, props, animation.opts ); + if ( result ) { + if ( isFunction( result.stop ) ) { + jQuery._queueHooks( animation.elem, animation.opts.queue ).stop = + result.stop.bind( result ); + } + return result; + } + } + + jQuery.map( props, createTween, animation ); + + if ( isFunction( animation.opts.start ) ) { + animation.opts.start.call( elem, animation ); + } + + // Attach callbacks from options + animation + .progress( animation.opts.progress ) + .done( animation.opts.done, animation.opts.complete ) + .fail( animation.opts.fail ) + .always( animation.opts.always ); + + jQuery.fx.timer( + jQuery.extend( tick, { + elem: elem, + anim: animation, + queue: animation.opts.queue + } ) + ); + + return animation; +} + +jQuery.Animation = jQuery.extend( Animation, { + + tweeners: { + "*": [ function( prop, value ) { + var tween = this.createTween( prop, value ); + adjustCSS( tween.elem, prop, rcssNum.exec( value ), tween ); + return tween; + } ] + }, + + tweener: function( props, callback ) { + if ( isFunction( props ) ) { + callback = props; + props = [ "*" ]; + } else { + props = props.match( rnothtmlwhite ); + } + + var prop, + index = 0, + length = props.length; + + for ( ; index < length; index++ ) { + prop = props[ index ]; + Animation.tweeners[ prop ] = Animation.tweeners[ prop ] || []; + Animation.tweeners[ prop ].unshift( callback ); + } + }, + + prefilters: [ defaultPrefilter ], + + prefilter: function( callback, prepend ) { + if ( prepend ) { + Animation.prefilters.unshift( callback ); + } else { + Animation.prefilters.push( callback ); + } + } +} ); + +jQuery.speed = function( speed, easing, fn ) { + var opt = speed && typeof speed === "object" ? jQuery.extend( {}, speed ) : { + complete: fn || !fn && easing || + isFunction( speed ) && speed, + duration: speed, + easing: fn && easing || easing && !isFunction( easing ) && easing + }; + + // Go to the end state if fx are off + if ( jQuery.fx.off ) { + opt.duration = 0; + + } else { + if ( typeof opt.duration !== "number" ) { + if ( opt.duration in jQuery.fx.speeds ) { + opt.duration = jQuery.fx.speeds[ opt.duration ]; + + } else { + opt.duration = jQuery.fx.speeds._default; + } + } + } + + // Normalize opt.queue - true/undefined/null -> "fx" + if ( opt.queue == null || opt.queue === true ) { + opt.queue = "fx"; + } + + // Queueing + opt.old = opt.complete; + + opt.complete = function() { + if ( isFunction( opt.old ) ) { + opt.old.call( this ); + } + + if ( opt.queue ) { + jQuery.dequeue( this, opt.queue ); + } + }; + + return opt; +}; + +jQuery.fn.extend( { + fadeTo: function( speed, to, easing, callback ) { + + // Show any hidden elements after setting opacity to 0 + return this.filter( isHiddenWithinTree ).css( "opacity", 0 ).show() + + // Animate to the value specified + .end().animate( { opacity: to }, speed, easing, callback ); + }, + animate: function( prop, speed, easing, callback ) { + var empty = jQuery.isEmptyObject( prop ), + optall = jQuery.speed( speed, easing, callback ), + doAnimation = function() { + + // Operate on a copy of prop so per-property easing won't be lost + var anim = Animation( this, jQuery.extend( {}, prop ), optall ); + + // Empty animations, or finishing resolves immediately + if ( empty || dataPriv.get( this, "finish" ) ) { + anim.stop( true ); + } + }; + doAnimation.finish = doAnimation; + + return empty || optall.queue === false ? + this.each( doAnimation ) : + this.queue( optall.queue, doAnimation ); + }, + stop: function( type, clearQueue, gotoEnd ) { + var stopQueue = function( hooks ) { + var stop = hooks.stop; + delete hooks.stop; + stop( gotoEnd ); + }; + + if ( typeof type !== "string" ) { + gotoEnd = clearQueue; + clearQueue = type; + type = undefined; + } + if ( clearQueue ) { + this.queue( type || "fx", [] ); + } + + return this.each( function() { + var dequeue = true, + index = type != null && type + "queueHooks", + timers = jQuery.timers, + data = dataPriv.get( this ); + + if ( index ) { + if ( data[ index ] && data[ index ].stop ) { + stopQueue( data[ index ] ); + } + } else { + for ( index in data ) { + if ( data[ index ] && data[ index ].stop && rrun.test( index ) ) { + stopQueue( data[ index ] ); + } + } + } + + for ( index = timers.length; index--; ) { + if ( timers[ index ].elem === this && + ( type == null || timers[ index ].queue === type ) ) { + + timers[ index ].anim.stop( gotoEnd ); + dequeue = false; + timers.splice( index, 1 ); + } + } + + // Start the next in the queue if the last step wasn't forced. + // Timers currently will call their complete callbacks, which + // will dequeue but only if they were gotoEnd. + if ( dequeue || !gotoEnd ) { + jQuery.dequeue( this, type ); + } + } ); + }, + finish: function( type ) { + if ( type !== false ) { + type = type || "fx"; + } + return this.each( function() { + var index, + data = dataPriv.get( this ), + queue = data[ type + "queue" ], + hooks = data[ type + "queueHooks" ], + timers = jQuery.timers, + length = queue ? queue.length : 0; + + // Enable finishing flag on private data + data.finish = true; + + // Empty the queue first + jQuery.queue( this, type, [] ); + + if ( hooks && hooks.stop ) { + hooks.stop.call( this, true ); + } + + // Look for any active animations, and finish them + for ( index = timers.length; index--; ) { + if ( timers[ index ].elem === this && timers[ index ].queue === type ) { + timers[ index ].anim.stop( true ); + timers.splice( index, 1 ); + } + } + + // Look for any animations in the old queue and finish them + for ( index = 0; index < length; index++ ) { + if ( queue[ index ] && queue[ index ].finish ) { + queue[ index ].finish.call( this ); + } + } + + // Turn off finishing flag + delete data.finish; + } ); + } +} ); + +jQuery.each( [ "toggle", "show", "hide" ], function( _i, name ) { + var cssFn = jQuery.fn[ name ]; + jQuery.fn[ name ] = function( speed, easing, callback ) { + return speed == null || typeof speed === "boolean" ? + cssFn.apply( this, arguments ) : + this.animate( genFx( name, true ), speed, easing, callback ); + }; +} ); + +// Generate shortcuts for custom animations +jQuery.each( { + slideDown: genFx( "show" ), + slideUp: genFx( "hide" ), + slideToggle: genFx( "toggle" ), + fadeIn: { opacity: "show" }, + fadeOut: { opacity: "hide" }, + fadeToggle: { opacity: "toggle" } +}, function( name, props ) { + jQuery.fn[ name ] = function( speed, easing, callback ) { + return this.animate( props, speed, easing, callback ); + }; +} ); + +jQuery.timers = []; +jQuery.fx.tick = function() { + var timer, + i = 0, + timers = jQuery.timers; + + fxNow = Date.now(); + + for ( ; i < timers.length; i++ ) { + timer = timers[ i ]; + + // Run the timer and safely remove it when done (allowing for external removal) + if ( !timer() && timers[ i ] === timer ) { + timers.splice( i--, 1 ); + } + } + + if ( !timers.length ) { + jQuery.fx.stop(); + } + fxNow = undefined; +}; + +jQuery.fx.timer = function( timer ) { + jQuery.timers.push( timer ); + jQuery.fx.start(); +}; + +jQuery.fx.interval = 13; +jQuery.fx.start = function() { + if ( inProgress ) { + return; + } + + inProgress = true; + schedule(); +}; + +jQuery.fx.stop = function() { + inProgress = null; +}; + +jQuery.fx.speeds = { + slow: 600, + fast: 200, + + // Default speed + _default: 400 +}; + + +// Based off of the plugin by Clint Helfers, with permission. +// https://web.archive.org/web/20100324014747/http://blindsignals.com/index.php/2009/07/jquery-delay/ +jQuery.fn.delay = function( time, type ) { + time = jQuery.fx ? jQuery.fx.speeds[ time ] || time : time; + type = type || "fx"; + + return this.queue( type, function( next, hooks ) { + var timeout = window.setTimeout( next, time ); + hooks.stop = function() { + window.clearTimeout( timeout ); + }; + } ); +}; + + +( function() { + var input = document.createElement( "input" ), + select = document.createElement( "select" ), + opt = select.appendChild( document.createElement( "option" ) ); + + input.type = "checkbox"; + + // Support: Android <=4.3 only + // Default value for a checkbox should be "on" + support.checkOn = input.value !== ""; + + // Support: IE <=11 only + // Must access selectedIndex to make default options select + support.optSelected = opt.selected; + + // Support: IE <=11 only + // An input loses its value after becoming a radio + input = document.createElement( "input" ); + input.value = "t"; + input.type = "radio"; + support.radioValue = input.value === "t"; +} )(); + + +var boolHook, + attrHandle = jQuery.expr.attrHandle; + +jQuery.fn.extend( { + attr: function( name, value ) { + return access( this, jQuery.attr, name, value, arguments.length > 1 ); + }, + + removeAttr: function( name ) { + return this.each( function() { + jQuery.removeAttr( this, name ); + } ); + } +} ); + +jQuery.extend( { + attr: function( elem, name, value ) { + var ret, hooks, + nType = elem.nodeType; + + // Don't get/set attributes on text, comment and attribute nodes + if ( nType === 3 || nType === 8 || nType === 2 ) { + return; + } + + // Fallback to prop when attributes are not supported + if ( typeof elem.getAttribute === "undefined" ) { + return jQuery.prop( elem, name, value ); + } + + // Attribute hooks are determined by the lowercase version + // Grab necessary hook if one is defined + if ( nType !== 1 || !jQuery.isXMLDoc( elem ) ) { + hooks = jQuery.attrHooks[ name.toLowerCase() ] || + ( jQuery.expr.match.bool.test( name ) ? boolHook : undefined ); + } + + if ( value !== undefined ) { + if ( value === null ) { + jQuery.removeAttr( elem, name ); + return; + } + + if ( hooks && "set" in hooks && + ( ret = hooks.set( elem, value, name ) ) !== undefined ) { + return ret; + } + + elem.setAttribute( name, value + "" ); + return value; + } + + if ( hooks && "get" in hooks && ( ret = hooks.get( elem, name ) ) !== null ) { + return ret; + } + + ret = jQuery.find.attr( elem, name ); + + // Non-existent attributes return null, we normalize to undefined + return ret == null ? undefined : ret; + }, + + attrHooks: { + type: { + set: function( elem, value ) { + if ( !support.radioValue && value === "radio" && + nodeName( elem, "input" ) ) { + var val = elem.value; + elem.setAttribute( "type", value ); + if ( val ) { + elem.value = val; + } + return value; + } + } + } + }, + + removeAttr: function( elem, value ) { + var name, + i = 0, + + // Attribute names can contain non-HTML whitespace characters + // https://html.spec.whatwg.org/multipage/syntax.html#attributes-2 + attrNames = value && value.match( rnothtmlwhite ); + + if ( attrNames && elem.nodeType === 1 ) { + while ( ( name = attrNames[ i++ ] ) ) { + elem.removeAttribute( name ); + } + } + } +} ); + +// Hooks for boolean attributes +boolHook = { + set: function( elem, value, name ) { + if ( value === false ) { + + // Remove boolean attributes when set to false + jQuery.removeAttr( elem, name ); + } else { + elem.setAttribute( name, name ); + } + return name; + } +}; + +jQuery.each( jQuery.expr.match.bool.source.match( /\w+/g ), function( _i, name ) { + var getter = attrHandle[ name ] || jQuery.find.attr; + + attrHandle[ name ] = function( elem, name, isXML ) { + var ret, handle, + lowercaseName = name.toLowerCase(); + + if ( !isXML ) { + + // Avoid an infinite loop by temporarily removing this function from the getter + handle = attrHandle[ lowercaseName ]; + attrHandle[ lowercaseName ] = ret; + ret = getter( elem, name, isXML ) != null ? + lowercaseName : + null; + attrHandle[ lowercaseName ] = handle; + } + return ret; + }; +} ); + + + + +var rfocusable = /^(?:input|select|textarea|button)$/i, + rclickable = /^(?:a|area)$/i; + +jQuery.fn.extend( { + prop: function( name, value ) { + return access( this, jQuery.prop, name, value, arguments.length > 1 ); + }, + + removeProp: function( name ) { + return this.each( function() { + delete this[ jQuery.propFix[ name ] || name ]; + } ); + } +} ); + +jQuery.extend( { + prop: function( elem, name, value ) { + var ret, hooks, + nType = elem.nodeType; + + // Don't get/set properties on text, comment and attribute nodes + if ( nType === 3 || nType === 8 || nType === 2 ) { + return; + } + + if ( nType !== 1 || !jQuery.isXMLDoc( elem ) ) { + + // Fix name and attach hooks + name = jQuery.propFix[ name ] || name; + hooks = jQuery.propHooks[ name ]; + } + + if ( value !== undefined ) { + if ( hooks && "set" in hooks && + ( ret = hooks.set( elem, value, name ) ) !== undefined ) { + return ret; + } + + return ( elem[ name ] = value ); + } + + if ( hooks && "get" in hooks && ( ret = hooks.get( elem, name ) ) !== null ) { + return ret; + } + + return elem[ name ]; + }, + + propHooks: { + tabIndex: { + get: function( elem ) { + + // Support: IE <=9 - 11 only + // elem.tabIndex doesn't always return the + // correct value when it hasn't been explicitly set + // https://web.archive.org/web/20141116233347/http://fluidproject.org/blog/2008/01/09/getting-setting-and-removing-tabindex-values-with-javascript/ + // Use proper attribute retrieval(#12072) + var tabindex = jQuery.find.attr( elem, "tabindex" ); + + if ( tabindex ) { + return parseInt( tabindex, 10 ); + } + + if ( + rfocusable.test( elem.nodeName ) || + rclickable.test( elem.nodeName ) && + elem.href + ) { + return 0; + } + + return -1; + } + } + }, + + propFix: { + "for": "htmlFor", + "class": "className" + } +} ); + +// Support: IE <=11 only +// Accessing the selectedIndex property +// forces the browser to respect setting selected +// on the option +// The getter ensures a default option is selected +// when in an optgroup +// eslint rule "no-unused-expressions" is disabled for this code +// since it considers such accessions noop +if ( !support.optSelected ) { + jQuery.propHooks.selected = { + get: function( elem ) { + + /* eslint no-unused-expressions: "off" */ + + var parent = elem.parentNode; + if ( parent && parent.parentNode ) { + parent.parentNode.selectedIndex; + } + return null; + }, + set: function( elem ) { + + /* eslint no-unused-expressions: "off" */ + + var parent = elem.parentNode; + if ( parent ) { + parent.selectedIndex; + + if ( parent.parentNode ) { + parent.parentNode.selectedIndex; + } + } + } + }; +} + +jQuery.each( [ + "tabIndex", + "readOnly", + "maxLength", + "cellSpacing", + "cellPadding", + "rowSpan", + "colSpan", + "useMap", + "frameBorder", + "contentEditable" +], function() { + jQuery.propFix[ this.toLowerCase() ] = this; +} ); + + + + + // Strip and collapse whitespace according to HTML spec + // https://infra.spec.whatwg.org/#strip-and-collapse-ascii-whitespace + function stripAndCollapse( value ) { + var tokens = value.match( rnothtmlwhite ) || []; + return tokens.join( " " ); + } + + +function getClass( elem ) { + return elem.getAttribute && elem.getAttribute( "class" ) || ""; +} + +function classesToArray( value ) { + if ( Array.isArray( value ) ) { + return value; + } + if ( typeof value === "string" ) { + return value.match( rnothtmlwhite ) || []; + } + return []; +} + +jQuery.fn.extend( { + addClass: function( value ) { + var classes, elem, cur, curValue, clazz, j, finalValue, + i = 0; + + if ( isFunction( value ) ) { + return this.each( function( j ) { + jQuery( this ).addClass( value.call( this, j, getClass( this ) ) ); + } ); + } + + classes = classesToArray( value ); + + if ( classes.length ) { + while ( ( elem = this[ i++ ] ) ) { + curValue = getClass( elem ); + cur = elem.nodeType === 1 && ( " " + stripAndCollapse( curValue ) + " " ); + + if ( cur ) { + j = 0; + while ( ( clazz = classes[ j++ ] ) ) { + if ( cur.indexOf( " " + clazz + " " ) < 0 ) { + cur += clazz + " "; + } + } + + // Only assign if different to avoid unneeded rendering. + finalValue = stripAndCollapse( cur ); + if ( curValue !== finalValue ) { + elem.setAttribute( "class", finalValue ); + } + } + } + } + + return this; + }, + + removeClass: function( value ) { + var classes, elem, cur, curValue, clazz, j, finalValue, + i = 0; + + if ( isFunction( value ) ) { + return this.each( function( j ) { + jQuery( this ).removeClass( value.call( this, j, getClass( this ) ) ); + } ); + } + + if ( !arguments.length ) { + return this.attr( "class", "" ); + } + + classes = classesToArray( value ); + + if ( classes.length ) { + while ( ( elem = this[ i++ ] ) ) { + curValue = getClass( elem ); + + // This expression is here for better compressibility (see addClass) + cur = elem.nodeType === 1 && ( " " + stripAndCollapse( curValue ) + " " ); + + if ( cur ) { + j = 0; + while ( ( clazz = classes[ j++ ] ) ) { + + // Remove *all* instances + while ( cur.indexOf( " " + clazz + " " ) > -1 ) { + cur = cur.replace( " " + clazz + " ", " " ); + } + } + + // Only assign if different to avoid unneeded rendering. + finalValue = stripAndCollapse( cur ); + if ( curValue !== finalValue ) { + elem.setAttribute( "class", finalValue ); + } + } + } + } + + return this; + }, + + toggleClass: function( value, stateVal ) { + var type = typeof value, + isValidValue = type === "string" || Array.isArray( value ); + + if ( typeof stateVal === "boolean" && isValidValue ) { + return stateVal ? this.addClass( value ) : this.removeClass( value ); + } + + if ( isFunction( value ) ) { + return this.each( function( i ) { + jQuery( this ).toggleClass( + value.call( this, i, getClass( this ), stateVal ), + stateVal + ); + } ); + } + + return this.each( function() { + var className, i, self, classNames; + + if ( isValidValue ) { + + // Toggle individual class names + i = 0; + self = jQuery( this ); + classNames = classesToArray( value ); + + while ( ( className = classNames[ i++ ] ) ) { + + // Check each className given, space separated list + if ( self.hasClass( className ) ) { + self.removeClass( className ); + } else { + self.addClass( className ); + } + } + + // Toggle whole class name + } else if ( value === undefined || type === "boolean" ) { + className = getClass( this ); + if ( className ) { + + // Store className if set + dataPriv.set( this, "__className__", className ); + } + + // If the element has a class name or if we're passed `false`, + // then remove the whole classname (if there was one, the above saved it). + // Otherwise bring back whatever was previously saved (if anything), + // falling back to the empty string if nothing was stored. + if ( this.setAttribute ) { + this.setAttribute( "class", + className || value === false ? + "" : + dataPriv.get( this, "__className__" ) || "" + ); + } + } + } ); + }, + + hasClass: function( selector ) { + var className, elem, + i = 0; + + className = " " + selector + " "; + while ( ( elem = this[ i++ ] ) ) { + if ( elem.nodeType === 1 && + ( " " + stripAndCollapse( getClass( elem ) ) + " " ).indexOf( className ) > -1 ) { + return true; + } + } + + return false; + } +} ); + + + + +var rreturn = /\r/g; + +jQuery.fn.extend( { + val: function( value ) { + var hooks, ret, valueIsFunction, + elem = this[ 0 ]; + + if ( !arguments.length ) { + if ( elem ) { + hooks = jQuery.valHooks[ elem.type ] || + jQuery.valHooks[ elem.nodeName.toLowerCase() ]; + + if ( hooks && + "get" in hooks && + ( ret = hooks.get( elem, "value" ) ) !== undefined + ) { + return ret; + } + + ret = elem.value; + + // Handle most common string cases + if ( typeof ret === "string" ) { + return ret.replace( rreturn, "" ); + } + + // Handle cases where value is null/undef or number + return ret == null ? "" : ret; + } + + return; + } + + valueIsFunction = isFunction( value ); + + return this.each( function( i ) { + var val; + + if ( this.nodeType !== 1 ) { + return; + } + + if ( valueIsFunction ) { + val = value.call( this, i, jQuery( this ).val() ); + } else { + val = value; + } + + // Treat null/undefined as ""; convert numbers to string + if ( val == null ) { + val = ""; + + } else if ( typeof val === "number" ) { + val += ""; + + } else if ( Array.isArray( val ) ) { + val = jQuery.map( val, function( value ) { + return value == null ? "" : value + ""; + } ); + } + + hooks = jQuery.valHooks[ this.type ] || jQuery.valHooks[ this.nodeName.toLowerCase() ]; + + // If set returns undefined, fall back to normal setting + if ( !hooks || !( "set" in hooks ) || hooks.set( this, val, "value" ) === undefined ) { + this.value = val; + } + } ); + } +} ); + +jQuery.extend( { + valHooks: { + option: { + get: function( elem ) { + + var val = jQuery.find.attr( elem, "value" ); + return val != null ? + val : + + // Support: IE <=10 - 11 only + // option.text throws exceptions (#14686, #14858) + // Strip and collapse whitespace + // https://html.spec.whatwg.org/#strip-and-collapse-whitespace + stripAndCollapse( jQuery.text( elem ) ); + } + }, + select: { + get: function( elem ) { + var value, option, i, + options = elem.options, + index = elem.selectedIndex, + one = elem.type === "select-one", + values = one ? null : [], + max = one ? index + 1 : options.length; + + if ( index < 0 ) { + i = max; + + } else { + i = one ? index : 0; + } + + // Loop through all the selected options + for ( ; i < max; i++ ) { + option = options[ i ]; + + // Support: IE <=9 only + // IE8-9 doesn't update selected after form reset (#2551) + if ( ( option.selected || i === index ) && + + // Don't return options that are disabled or in a disabled optgroup + !option.disabled && + ( !option.parentNode.disabled || + !nodeName( option.parentNode, "optgroup" ) ) ) { + + // Get the specific value for the option + value = jQuery( option ).val(); + + // We don't need an array for one selects + if ( one ) { + return value; + } + + // Multi-Selects return an array + values.push( value ); + } + } + + return values; + }, + + set: function( elem, value ) { + var optionSet, option, + options = elem.options, + values = jQuery.makeArray( value ), + i = options.length; + + while ( i-- ) { + option = options[ i ]; + + /* eslint-disable no-cond-assign */ + + if ( option.selected = + jQuery.inArray( jQuery.valHooks.option.get( option ), values ) > -1 + ) { + optionSet = true; + } + + /* eslint-enable no-cond-assign */ + } + + // Force browsers to behave consistently when non-matching value is set + if ( !optionSet ) { + elem.selectedIndex = -1; + } + return values; + } + } + } +} ); + +// Radios and checkboxes getter/setter +jQuery.each( [ "radio", "checkbox" ], function() { + jQuery.valHooks[ this ] = { + set: function( elem, value ) { + if ( Array.isArray( value ) ) { + return ( elem.checked = jQuery.inArray( jQuery( elem ).val(), value ) > -1 ); + } + } + }; + if ( !support.checkOn ) { + jQuery.valHooks[ this ].get = function( elem ) { + return elem.getAttribute( "value" ) === null ? "on" : elem.value; + }; + } +} ); + + + + +// Return jQuery for attributes-only inclusion + + +support.focusin = "onfocusin" in window; + + +var rfocusMorph = /^(?:focusinfocus|focusoutblur)$/, + stopPropagationCallback = function( e ) { + e.stopPropagation(); + }; + +jQuery.extend( jQuery.event, { + + trigger: function( event, data, elem, onlyHandlers ) { + + var i, cur, tmp, bubbleType, ontype, handle, special, lastElement, + eventPath = [ elem || document ], + type = hasOwn.call( event, "type" ) ? event.type : event, + namespaces = hasOwn.call( event, "namespace" ) ? event.namespace.split( "." ) : []; + + cur = lastElement = tmp = elem = elem || document; + + // Don't do events on text and comment nodes + if ( elem.nodeType === 3 || elem.nodeType === 8 ) { + return; + } + + // focus/blur morphs to focusin/out; ensure we're not firing them right now + if ( rfocusMorph.test( type + jQuery.event.triggered ) ) { + return; + } + + if ( type.indexOf( "." ) > -1 ) { + + // Namespaced trigger; create a regexp to match event type in handle() + namespaces = type.split( "." ); + type = namespaces.shift(); + namespaces.sort(); + } + ontype = type.indexOf( ":" ) < 0 && "on" + type; + + // Caller can pass in a jQuery.Event object, Object, or just an event type string + event = event[ jQuery.expando ] ? + event : + new jQuery.Event( type, typeof event === "object" && event ); + + // Trigger bitmask: & 1 for native handlers; & 2 for jQuery (always true) + event.isTrigger = onlyHandlers ? 2 : 3; + event.namespace = namespaces.join( "." ); + event.rnamespace = event.namespace ? + new RegExp( "(^|\\.)" + namespaces.join( "\\.(?:.*\\.|)" ) + "(\\.|$)" ) : + null; + + // Clean up the event in case it is being reused + event.result = undefined; + if ( !event.target ) { + event.target = elem; + } + + // Clone any incoming data and prepend the event, creating the handler arg list + data = data == null ? + [ event ] : + jQuery.makeArray( data, [ event ] ); + + // Allow special events to draw outside the lines + special = jQuery.event.special[ type ] || {}; + if ( !onlyHandlers && special.trigger && special.trigger.apply( elem, data ) === false ) { + return; + } + + // Determine event propagation path in advance, per W3C events spec (#9951) + // Bubble up to document, then to window; watch for a global ownerDocument var (#9724) + if ( !onlyHandlers && !special.noBubble && !isWindow( elem ) ) { + + bubbleType = special.delegateType || type; + if ( !rfocusMorph.test( bubbleType + type ) ) { + cur = cur.parentNode; + } + for ( ; cur; cur = cur.parentNode ) { + eventPath.push( cur ); + tmp = cur; + } + + // Only add window if we got to document (e.g., not plain obj or detached DOM) + if ( tmp === ( elem.ownerDocument || document ) ) { + eventPath.push( tmp.defaultView || tmp.parentWindow || window ); + } + } + + // Fire handlers on the event path + i = 0; + while ( ( cur = eventPath[ i++ ] ) && !event.isPropagationStopped() ) { + lastElement = cur; + event.type = i > 1 ? + bubbleType : + special.bindType || type; + + // jQuery handler + handle = ( + dataPriv.get( cur, "events" ) || Object.create( null ) + )[ event.type ] && + dataPriv.get( cur, "handle" ); + if ( handle ) { + handle.apply( cur, data ); + } + + // Native handler + handle = ontype && cur[ ontype ]; + if ( handle && handle.apply && acceptData( cur ) ) { + event.result = handle.apply( cur, data ); + if ( event.result === false ) { + event.preventDefault(); + } + } + } + event.type = type; + + // If nobody prevented the default action, do it now + if ( !onlyHandlers && !event.isDefaultPrevented() ) { + + if ( ( !special._default || + special._default.apply( eventPath.pop(), data ) === false ) && + acceptData( elem ) ) { + + // Call a native DOM method on the target with the same name as the event. + // Don't do default actions on window, that's where global variables be (#6170) + if ( ontype && isFunction( elem[ type ] ) && !isWindow( elem ) ) { + + // Don't re-trigger an onFOO event when we call its FOO() method + tmp = elem[ ontype ]; + + if ( tmp ) { + elem[ ontype ] = null; + } + + // Prevent re-triggering of the same event, since we already bubbled it above + jQuery.event.triggered = type; + + if ( event.isPropagationStopped() ) { + lastElement.addEventListener( type, stopPropagationCallback ); + } + + elem[ type ](); + + if ( event.isPropagationStopped() ) { + lastElement.removeEventListener( type, stopPropagationCallback ); + } + + jQuery.event.triggered = undefined; + + if ( tmp ) { + elem[ ontype ] = tmp; + } + } + } + } + + return event.result; + }, + + // Piggyback on a donor event to simulate a different one + // Used only for `focus(in | out)` events + simulate: function( type, elem, event ) { + var e = jQuery.extend( + new jQuery.Event(), + event, + { + type: type, + isSimulated: true + } + ); + + jQuery.event.trigger( e, null, elem ); + } + +} ); + +jQuery.fn.extend( { + + trigger: function( type, data ) { + return this.each( function() { + jQuery.event.trigger( type, data, this ); + } ); + }, + triggerHandler: function( type, data ) { + var elem = this[ 0 ]; + if ( elem ) { + return jQuery.event.trigger( type, data, elem, true ); + } + } +} ); + + +// Support: Firefox <=44 +// Firefox doesn't have focus(in | out) events +// Related ticket - https://bugzilla.mozilla.org/show_bug.cgi?id=687787 +// +// Support: Chrome <=48 - 49, Safari <=9.0 - 9.1 +// focus(in | out) events fire after focus & blur events, +// which is spec violation - http://www.w3.org/TR/DOM-Level-3-Events/#events-focusevent-event-order +// Related ticket - https://bugs.chromium.org/p/chromium/issues/detail?id=449857 +if ( !support.focusin ) { + jQuery.each( { focus: "focusin", blur: "focusout" }, function( orig, fix ) { + + // Attach a single capturing handler on the document while someone wants focusin/focusout + var handler = function( event ) { + jQuery.event.simulate( fix, event.target, jQuery.event.fix( event ) ); + }; + + jQuery.event.special[ fix ] = { + setup: function() { + + // Handle: regular nodes (via `this.ownerDocument`), window + // (via `this.document`) & document (via `this`). + var doc = this.ownerDocument || this.document || this, + attaches = dataPriv.access( doc, fix ); + + if ( !attaches ) { + doc.addEventListener( orig, handler, true ); + } + dataPriv.access( doc, fix, ( attaches || 0 ) + 1 ); + }, + teardown: function() { + var doc = this.ownerDocument || this.document || this, + attaches = dataPriv.access( doc, fix ) - 1; + + if ( !attaches ) { + doc.removeEventListener( orig, handler, true ); + dataPriv.remove( doc, fix ); + + } else { + dataPriv.access( doc, fix, attaches ); + } + } + }; + } ); +} +var location = window.location; + +var nonce = { guid: Date.now() }; + +var rquery = ( /\?/ ); + + + +// Cross-browser xml parsing +jQuery.parseXML = function( data ) { + var xml; + if ( !data || typeof data !== "string" ) { + return null; + } + + // Support: IE 9 - 11 only + // IE throws on parseFromString with invalid input. + try { + xml = ( new window.DOMParser() ).parseFromString( data, "text/xml" ); + } catch ( e ) { + xml = undefined; + } + + if ( !xml || xml.getElementsByTagName( "parsererror" ).length ) { + jQuery.error( "Invalid XML: " + data ); + } + return xml; +}; + + +var + rbracket = /\[\]$/, + rCRLF = /\r?\n/g, + rsubmitterTypes = /^(?:submit|button|image|reset|file)$/i, + rsubmittable = /^(?:input|select|textarea|keygen)/i; + +function buildParams( prefix, obj, traditional, add ) { + var name; + + if ( Array.isArray( obj ) ) { + + // Serialize array item. + jQuery.each( obj, function( i, v ) { + if ( traditional || rbracket.test( prefix ) ) { + + // Treat each array item as a scalar. + add( prefix, v ); + + } else { + + // Item is non-scalar (array or object), encode its numeric index. + buildParams( + prefix + "[" + ( typeof v === "object" && v != null ? i : "" ) + "]", + v, + traditional, + add + ); + } + } ); + + } else if ( !traditional && toType( obj ) === "object" ) { + + // Serialize object item. + for ( name in obj ) { + buildParams( prefix + "[" + name + "]", obj[ name ], traditional, add ); + } + + } else { + + // Serialize scalar item. + add( prefix, obj ); + } +} + +// Serialize an array of form elements or a set of +// key/values into a query string +jQuery.param = function( a, traditional ) { + var prefix, + s = [], + add = function( key, valueOrFunction ) { + + // If value is a function, invoke it and use its return value + var value = isFunction( valueOrFunction ) ? + valueOrFunction() : + valueOrFunction; + + s[ s.length ] = encodeURIComponent( key ) + "=" + + encodeURIComponent( value == null ? "" : value ); + }; + + if ( a == null ) { + return ""; + } + + // If an array was passed in, assume that it is an array of form elements. + if ( Array.isArray( a ) || ( a.jquery && !jQuery.isPlainObject( a ) ) ) { + + // Serialize the form elements + jQuery.each( a, function() { + add( this.name, this.value ); + } ); + + } else { + + // If traditional, encode the "old" way (the way 1.3.2 or older + // did it), otherwise encode params recursively. + for ( prefix in a ) { + buildParams( prefix, a[ prefix ], traditional, add ); + } + } + + // Return the resulting serialization + return s.join( "&" ); +}; + +jQuery.fn.extend( { + serialize: function() { + return jQuery.param( this.serializeArray() ); + }, + serializeArray: function() { + return this.map( function() { + + // Can add propHook for "elements" to filter or add form elements + var elements = jQuery.prop( this, "elements" ); + return elements ? jQuery.makeArray( elements ) : this; + } ) + .filter( function() { + var type = this.type; + + // Use .is( ":disabled" ) so that fieldset[disabled] works + return this.name && !jQuery( this ).is( ":disabled" ) && + rsubmittable.test( this.nodeName ) && !rsubmitterTypes.test( type ) && + ( this.checked || !rcheckableType.test( type ) ); + } ) + .map( function( _i, elem ) { + var val = jQuery( this ).val(); + + if ( val == null ) { + return null; + } + + if ( Array.isArray( val ) ) { + return jQuery.map( val, function( val ) { + return { name: elem.name, value: val.replace( rCRLF, "\r\n" ) }; + } ); + } + + return { name: elem.name, value: val.replace( rCRLF, "\r\n" ) }; + } ).get(); + } +} ); + + +var + r20 = /%20/g, + rhash = /#.*$/, + rantiCache = /([?&])_=[^&]*/, + rheaders = /^(.*?):[ \t]*([^\r\n]*)$/mg, + + // #7653, #8125, #8152: local protocol detection + rlocalProtocol = /^(?:about|app|app-storage|.+-extension|file|res|widget):$/, + rnoContent = /^(?:GET|HEAD)$/, + rprotocol = /^\/\//, + + /* Prefilters + * 1) They are useful to introduce custom dataTypes (see ajax/jsonp.js for an example) + * 2) These are called: + * - BEFORE asking for a transport + * - AFTER param serialization (s.data is a string if s.processData is true) + * 3) key is the dataType + * 4) the catchall symbol "*" can be used + * 5) execution will start with transport dataType and THEN continue down to "*" if needed + */ + prefilters = {}, + + /* Transports bindings + * 1) key is the dataType + * 2) the catchall symbol "*" can be used + * 3) selection will start with transport dataType and THEN go to "*" if needed + */ + transports = {}, + + // Avoid comment-prolog char sequence (#10098); must appease lint and evade compression + allTypes = "*/".concat( "*" ), + + // Anchor tag for parsing the document origin + originAnchor = document.createElement( "a" ); + originAnchor.href = location.href; + +// Base "constructor" for jQuery.ajaxPrefilter and jQuery.ajaxTransport +function addToPrefiltersOrTransports( structure ) { + + // dataTypeExpression is optional and defaults to "*" + return function( dataTypeExpression, func ) { + + if ( typeof dataTypeExpression !== "string" ) { + func = dataTypeExpression; + dataTypeExpression = "*"; + } + + var dataType, + i = 0, + dataTypes = dataTypeExpression.toLowerCase().match( rnothtmlwhite ) || []; + + if ( isFunction( func ) ) { + + // For each dataType in the dataTypeExpression + while ( ( dataType = dataTypes[ i++ ] ) ) { + + // Prepend if requested + if ( dataType[ 0 ] === "+" ) { + dataType = dataType.slice( 1 ) || "*"; + ( structure[ dataType ] = structure[ dataType ] || [] ).unshift( func ); + + // Otherwise append + } else { + ( structure[ dataType ] = structure[ dataType ] || [] ).push( func ); + } + } + } + }; +} + +// Base inspection function for prefilters and transports +function inspectPrefiltersOrTransports( structure, options, originalOptions, jqXHR ) { + + var inspected = {}, + seekingTransport = ( structure === transports ); + + function inspect( dataType ) { + var selected; + inspected[ dataType ] = true; + jQuery.each( structure[ dataType ] || [], function( _, prefilterOrFactory ) { + var dataTypeOrTransport = prefilterOrFactory( options, originalOptions, jqXHR ); + if ( typeof dataTypeOrTransport === "string" && + !seekingTransport && !inspected[ dataTypeOrTransport ] ) { + + options.dataTypes.unshift( dataTypeOrTransport ); + inspect( dataTypeOrTransport ); + return false; + } else if ( seekingTransport ) { + return !( selected = dataTypeOrTransport ); + } + } ); + return selected; + } + + return inspect( options.dataTypes[ 0 ] ) || !inspected[ "*" ] && inspect( "*" ); +} + +// A special extend for ajax options +// that takes "flat" options (not to be deep extended) +// Fixes #9887 +function ajaxExtend( target, src ) { + var key, deep, + flatOptions = jQuery.ajaxSettings.flatOptions || {}; + + for ( key in src ) { + if ( src[ key ] !== undefined ) { + ( flatOptions[ key ] ? target : ( deep || ( deep = {} ) ) )[ key ] = src[ key ]; + } + } + if ( deep ) { + jQuery.extend( true, target, deep ); + } + + return target; +} + +/* Handles responses to an ajax request: + * - finds the right dataType (mediates between content-type and expected dataType) + * - returns the corresponding response + */ +function ajaxHandleResponses( s, jqXHR, responses ) { + + var ct, type, finalDataType, firstDataType, + contents = s.contents, + dataTypes = s.dataTypes; + + // Remove auto dataType and get content-type in the process + while ( dataTypes[ 0 ] === "*" ) { + dataTypes.shift(); + if ( ct === undefined ) { + ct = s.mimeType || jqXHR.getResponseHeader( "Content-Type" ); + } + } + + // Check if we're dealing with a known content-type + if ( ct ) { + for ( type in contents ) { + if ( contents[ type ] && contents[ type ].test( ct ) ) { + dataTypes.unshift( type ); + break; + } + } + } + + // Check to see if we have a response for the expected dataType + if ( dataTypes[ 0 ] in responses ) { + finalDataType = dataTypes[ 0 ]; + } else { + + // Try convertible dataTypes + for ( type in responses ) { + if ( !dataTypes[ 0 ] || s.converters[ type + " " + dataTypes[ 0 ] ] ) { + finalDataType = type; + break; + } + if ( !firstDataType ) { + firstDataType = type; + } + } + + // Or just use first one + finalDataType = finalDataType || firstDataType; + } + + // If we found a dataType + // We add the dataType to the list if needed + // and return the corresponding response + if ( finalDataType ) { + if ( finalDataType !== dataTypes[ 0 ] ) { + dataTypes.unshift( finalDataType ); + } + return responses[ finalDataType ]; + } +} + +/* Chain conversions given the request and the original response + * Also sets the responseXXX fields on the jqXHR instance + */ +function ajaxConvert( s, response, jqXHR, isSuccess ) { + var conv2, current, conv, tmp, prev, + converters = {}, + + // Work with a copy of dataTypes in case we need to modify it for conversion + dataTypes = s.dataTypes.slice(); + + // Create converters map with lowercased keys + if ( dataTypes[ 1 ] ) { + for ( conv in s.converters ) { + converters[ conv.toLowerCase() ] = s.converters[ conv ]; + } + } + + current = dataTypes.shift(); + + // Convert to each sequential dataType + while ( current ) { + + if ( s.responseFields[ current ] ) { + jqXHR[ s.responseFields[ current ] ] = response; + } + + // Apply the dataFilter if provided + if ( !prev && isSuccess && s.dataFilter ) { + response = s.dataFilter( response, s.dataType ); + } + + prev = current; + current = dataTypes.shift(); + + if ( current ) { + + // There's only work to do if current dataType is non-auto + if ( current === "*" ) { + + current = prev; + + // Convert response if prev dataType is non-auto and differs from current + } else if ( prev !== "*" && prev !== current ) { + + // Seek a direct converter + conv = converters[ prev + " " + current ] || converters[ "* " + current ]; + + // If none found, seek a pair + if ( !conv ) { + for ( conv2 in converters ) { + + // If conv2 outputs current + tmp = conv2.split( " " ); + if ( tmp[ 1 ] === current ) { + + // If prev can be converted to accepted input + conv = converters[ prev + " " + tmp[ 0 ] ] || + converters[ "* " + tmp[ 0 ] ]; + if ( conv ) { + + // Condense equivalence converters + if ( conv === true ) { + conv = converters[ conv2 ]; + + // Otherwise, insert the intermediate dataType + } else if ( converters[ conv2 ] !== true ) { + current = tmp[ 0 ]; + dataTypes.unshift( tmp[ 1 ] ); + } + break; + } + } + } + } + + // Apply converter (if not an equivalence) + if ( conv !== true ) { + + // Unless errors are allowed to bubble, catch and return them + if ( conv && s.throws ) { + response = conv( response ); + } else { + try { + response = conv( response ); + } catch ( e ) { + return { + state: "parsererror", + error: conv ? e : "No conversion from " + prev + " to " + current + }; + } + } + } + } + } + } + + return { state: "success", data: response }; +} + +jQuery.extend( { + + // Counter for holding the number of active queries + active: 0, + + // Last-Modified header cache for next request + lastModified: {}, + etag: {}, + + ajaxSettings: { + url: location.href, + type: "GET", + isLocal: rlocalProtocol.test( location.protocol ), + global: true, + processData: true, + async: true, + contentType: "application/x-www-form-urlencoded; charset=UTF-8", + + /* + timeout: 0, + data: null, + dataType: null, + username: null, + password: null, + cache: null, + throws: false, + traditional: false, + headers: {}, + */ + + accepts: { + "*": allTypes, + text: "text/plain", + html: "text/html", + xml: "application/xml, text/xml", + json: "application/json, text/javascript" + }, + + contents: { + xml: /\bxml\b/, + html: /\bhtml/, + json: /\bjson\b/ + }, + + responseFields: { + xml: "responseXML", + text: "responseText", + json: "responseJSON" + }, + + // Data converters + // Keys separate source (or catchall "*") and destination types with a single space + converters: { + + // Convert anything to text + "* text": String, + + // Text to html (true = no transformation) + "text html": true, + + // Evaluate text as a json expression + "text json": JSON.parse, + + // Parse text as xml + "text xml": jQuery.parseXML + }, + + // For options that shouldn't be deep extended: + // you can add your own custom options here if + // and when you create one that shouldn't be + // deep extended (see ajaxExtend) + flatOptions: { + url: true, + context: true + } + }, + + // Creates a full fledged settings object into target + // with both ajaxSettings and settings fields. + // If target is omitted, writes into ajaxSettings. + ajaxSetup: function( target, settings ) { + return settings ? + + // Building a settings object + ajaxExtend( ajaxExtend( target, jQuery.ajaxSettings ), settings ) : + + // Extending ajaxSettings + ajaxExtend( jQuery.ajaxSettings, target ); + }, + + ajaxPrefilter: addToPrefiltersOrTransports( prefilters ), + ajaxTransport: addToPrefiltersOrTransports( transports ), + + // Main method + ajax: function( url, options ) { + + // If url is an object, simulate pre-1.5 signature + if ( typeof url === "object" ) { + options = url; + url = undefined; + } + + // Force options to be an object + options = options || {}; + + var transport, + + // URL without anti-cache param + cacheURL, + + // Response headers + responseHeadersString, + responseHeaders, + + // timeout handle + timeoutTimer, + + // Url cleanup var + urlAnchor, + + // Request state (becomes false upon send and true upon completion) + completed, + + // To know if global events are to be dispatched + fireGlobals, + + // Loop variable + i, + + // uncached part of the url + uncached, + + // Create the final options object + s = jQuery.ajaxSetup( {}, options ), + + // Callbacks context + callbackContext = s.context || s, + + // Context for global events is callbackContext if it is a DOM node or jQuery collection + globalEventContext = s.context && + ( callbackContext.nodeType || callbackContext.jquery ) ? + jQuery( callbackContext ) : + jQuery.event, + + // Deferreds + deferred = jQuery.Deferred(), + completeDeferred = jQuery.Callbacks( "once memory" ), + + // Status-dependent callbacks + statusCode = s.statusCode || {}, + + // Headers (they are sent all at once) + requestHeaders = {}, + requestHeadersNames = {}, + + // Default abort message + strAbort = "canceled", + + // Fake xhr + jqXHR = { + readyState: 0, + + // Builds headers hashtable if needed + getResponseHeader: function( key ) { + var match; + if ( completed ) { + if ( !responseHeaders ) { + responseHeaders = {}; + while ( ( match = rheaders.exec( responseHeadersString ) ) ) { + responseHeaders[ match[ 1 ].toLowerCase() + " " ] = + ( responseHeaders[ match[ 1 ].toLowerCase() + " " ] || [] ) + .concat( match[ 2 ] ); + } + } + match = responseHeaders[ key.toLowerCase() + " " ]; + } + return match == null ? null : match.join( ", " ); + }, + + // Raw string + getAllResponseHeaders: function() { + return completed ? responseHeadersString : null; + }, + + // Caches the header + setRequestHeader: function( name, value ) { + if ( completed == null ) { + name = requestHeadersNames[ name.toLowerCase() ] = + requestHeadersNames[ name.toLowerCase() ] || name; + requestHeaders[ name ] = value; + } + return this; + }, + + // Overrides response content-type header + overrideMimeType: function( type ) { + if ( completed == null ) { + s.mimeType = type; + } + return this; + }, + + // Status-dependent callbacks + statusCode: function( map ) { + var code; + if ( map ) { + if ( completed ) { + + // Execute the appropriate callbacks + jqXHR.always( map[ jqXHR.status ] ); + } else { + + // Lazy-add the new callbacks in a way that preserves old ones + for ( code in map ) { + statusCode[ code ] = [ statusCode[ code ], map[ code ] ]; + } + } + } + return this; + }, + + // Cancel the request + abort: function( statusText ) { + var finalText = statusText || strAbort; + if ( transport ) { + transport.abort( finalText ); + } + done( 0, finalText ); + return this; + } + }; + + // Attach deferreds + deferred.promise( jqXHR ); + + // Add protocol if not provided (prefilters might expect it) + // Handle falsy url in the settings object (#10093: consistency with old signature) + // We also use the url parameter if available + s.url = ( ( url || s.url || location.href ) + "" ) + .replace( rprotocol, location.protocol + "//" ); + + // Alias method option to type as per ticket #12004 + s.type = options.method || options.type || s.method || s.type; + + // Extract dataTypes list + s.dataTypes = ( s.dataType || "*" ).toLowerCase().match( rnothtmlwhite ) || [ "" ]; + + // A cross-domain request is in order when the origin doesn't match the current origin. + if ( s.crossDomain == null ) { + urlAnchor = document.createElement( "a" ); + + // Support: IE <=8 - 11, Edge 12 - 15 + // IE throws exception on accessing the href property if url is malformed, + // e.g. http://example.com:80x/ + try { + urlAnchor.href = s.url; + + // Support: IE <=8 - 11 only + // Anchor's host property isn't correctly set when s.url is relative + urlAnchor.href = urlAnchor.href; + s.crossDomain = originAnchor.protocol + "//" + originAnchor.host !== + urlAnchor.protocol + "//" + urlAnchor.host; + } catch ( e ) { + + // If there is an error parsing the URL, assume it is crossDomain, + // it can be rejected by the transport if it is invalid + s.crossDomain = true; + } + } + + // Convert data if not already a string + if ( s.data && s.processData && typeof s.data !== "string" ) { + s.data = jQuery.param( s.data, s.traditional ); + } + + // Apply prefilters + inspectPrefiltersOrTransports( prefilters, s, options, jqXHR ); + + // If request was aborted inside a prefilter, stop there + if ( completed ) { + return jqXHR; + } + + // We can fire global events as of now if asked to + // Don't fire events if jQuery.event is undefined in an AMD-usage scenario (#15118) + fireGlobals = jQuery.event && s.global; + + // Watch for a new set of requests + if ( fireGlobals && jQuery.active++ === 0 ) { + jQuery.event.trigger( "ajaxStart" ); + } + + // Uppercase the type + s.type = s.type.toUpperCase(); + + // Determine if request has content + s.hasContent = !rnoContent.test( s.type ); + + // Save the URL in case we're toying with the If-Modified-Since + // and/or If-None-Match header later on + // Remove hash to simplify url manipulation + cacheURL = s.url.replace( rhash, "" ); + + // More options handling for requests with no content + if ( !s.hasContent ) { + + // Remember the hash so we can put it back + uncached = s.url.slice( cacheURL.length ); + + // If data is available and should be processed, append data to url + if ( s.data && ( s.processData || typeof s.data === "string" ) ) { + cacheURL += ( rquery.test( cacheURL ) ? "&" : "?" ) + s.data; + + // #9682: remove data so that it's not used in an eventual retry + delete s.data; + } + + // Add or update anti-cache param if needed + if ( s.cache === false ) { + cacheURL = cacheURL.replace( rantiCache, "$1" ); + uncached = ( rquery.test( cacheURL ) ? "&" : "?" ) + "_=" + ( nonce.guid++ ) + + uncached; + } + + // Put hash and anti-cache on the URL that will be requested (gh-1732) + s.url = cacheURL + uncached; + + // Change '%20' to '+' if this is encoded form body content (gh-2658) + } else if ( s.data && s.processData && + ( s.contentType || "" ).indexOf( "application/x-www-form-urlencoded" ) === 0 ) { + s.data = s.data.replace( r20, "+" ); + } + + // Set the If-Modified-Since and/or If-None-Match header, if in ifModified mode. + if ( s.ifModified ) { + if ( jQuery.lastModified[ cacheURL ] ) { + jqXHR.setRequestHeader( "If-Modified-Since", jQuery.lastModified[ cacheURL ] ); + } + if ( jQuery.etag[ cacheURL ] ) { + jqXHR.setRequestHeader( "If-None-Match", jQuery.etag[ cacheURL ] ); + } + } + + // Set the correct header, if data is being sent + if ( s.data && s.hasContent && s.contentType !== false || options.contentType ) { + jqXHR.setRequestHeader( "Content-Type", s.contentType ); + } + + // Set the Accepts header for the server, depending on the dataType + jqXHR.setRequestHeader( + "Accept", + s.dataTypes[ 0 ] && s.accepts[ s.dataTypes[ 0 ] ] ? + s.accepts[ s.dataTypes[ 0 ] ] + + ( s.dataTypes[ 0 ] !== "*" ? ", " + allTypes + "; q=0.01" : "" ) : + s.accepts[ "*" ] + ); + + // Check for headers option + for ( i in s.headers ) { + jqXHR.setRequestHeader( i, s.headers[ i ] ); + } + + // Allow custom headers/mimetypes and early abort + if ( s.beforeSend && + ( s.beforeSend.call( callbackContext, jqXHR, s ) === false || completed ) ) { + + // Abort if not done already and return + return jqXHR.abort(); + } + + // Aborting is no longer a cancellation + strAbort = "abort"; + + // Install callbacks on deferreds + completeDeferred.add( s.complete ); + jqXHR.done( s.success ); + jqXHR.fail( s.error ); + + // Get transport + transport = inspectPrefiltersOrTransports( transports, s, options, jqXHR ); + + // If no transport, we auto-abort + if ( !transport ) { + done( -1, "No Transport" ); + } else { + jqXHR.readyState = 1; + + // Send global event + if ( fireGlobals ) { + globalEventContext.trigger( "ajaxSend", [ jqXHR, s ] ); + } + + // If request was aborted inside ajaxSend, stop there + if ( completed ) { + return jqXHR; + } + + // Timeout + if ( s.async && s.timeout > 0 ) { + timeoutTimer = window.setTimeout( function() { + jqXHR.abort( "timeout" ); + }, s.timeout ); + } + + try { + completed = false; + transport.send( requestHeaders, done ); + } catch ( e ) { + + // Rethrow post-completion exceptions + if ( completed ) { + throw e; + } + + // Propagate others as results + done( -1, e ); + } + } + + // Callback for when everything is done + function done( status, nativeStatusText, responses, headers ) { + var isSuccess, success, error, response, modified, + statusText = nativeStatusText; + + // Ignore repeat invocations + if ( completed ) { + return; + } + + completed = true; + + // Clear timeout if it exists + if ( timeoutTimer ) { + window.clearTimeout( timeoutTimer ); + } + + // Dereference transport for early garbage collection + // (no matter how long the jqXHR object will be used) + transport = undefined; + + // Cache response headers + responseHeadersString = headers || ""; + + // Set readyState + jqXHR.readyState = status > 0 ? 4 : 0; + + // Determine if successful + isSuccess = status >= 200 && status < 300 || status === 304; + + // Get response data + if ( responses ) { + response = ajaxHandleResponses( s, jqXHR, responses ); + } + + // Use a noop converter for missing script + if ( !isSuccess && jQuery.inArray( "script", s.dataTypes ) > -1 ) { + s.converters[ "text script" ] = function() {}; + } + + // Convert no matter what (that way responseXXX fields are always set) + response = ajaxConvert( s, response, jqXHR, isSuccess ); + + // If successful, handle type chaining + if ( isSuccess ) { + + // Set the If-Modified-Since and/or If-None-Match header, if in ifModified mode. + if ( s.ifModified ) { + modified = jqXHR.getResponseHeader( "Last-Modified" ); + if ( modified ) { + jQuery.lastModified[ cacheURL ] = modified; + } + modified = jqXHR.getResponseHeader( "etag" ); + if ( modified ) { + jQuery.etag[ cacheURL ] = modified; + } + } + + // if no content + if ( status === 204 || s.type === "HEAD" ) { + statusText = "nocontent"; + + // if not modified + } else if ( status === 304 ) { + statusText = "notmodified"; + + // If we have data, let's convert it + } else { + statusText = response.state; + success = response.data; + error = response.error; + isSuccess = !error; + } + } else { + + // Extract error from statusText and normalize for non-aborts + error = statusText; + if ( status || !statusText ) { + statusText = "error"; + if ( status < 0 ) { + status = 0; + } + } + } + + // Set data for the fake xhr object + jqXHR.status = status; + jqXHR.statusText = ( nativeStatusText || statusText ) + ""; + + // Success/Error + if ( isSuccess ) { + deferred.resolveWith( callbackContext, [ success, statusText, jqXHR ] ); + } else { + deferred.rejectWith( callbackContext, [ jqXHR, statusText, error ] ); + } + + // Status-dependent callbacks + jqXHR.statusCode( statusCode ); + statusCode = undefined; + + if ( fireGlobals ) { + globalEventContext.trigger( isSuccess ? "ajaxSuccess" : "ajaxError", + [ jqXHR, s, isSuccess ? success : error ] ); + } + + // Complete + completeDeferred.fireWith( callbackContext, [ jqXHR, statusText ] ); + + if ( fireGlobals ) { + globalEventContext.trigger( "ajaxComplete", [ jqXHR, s ] ); + + // Handle the global AJAX counter + if ( !( --jQuery.active ) ) { + jQuery.event.trigger( "ajaxStop" ); + } + } + } + + return jqXHR; + }, + + getJSON: function( url, data, callback ) { + return jQuery.get( url, data, callback, "json" ); + }, + + getScript: function( url, callback ) { + return jQuery.get( url, undefined, callback, "script" ); + } +} ); + +jQuery.each( [ "get", "post" ], function( _i, method ) { + jQuery[ method ] = function( url, data, callback, type ) { + + // Shift arguments if data argument was omitted + if ( isFunction( data ) ) { + type = type || callback; + callback = data; + data = undefined; + } + + // The url can be an options object (which then must have .url) + return jQuery.ajax( jQuery.extend( { + url: url, + type: method, + dataType: type, + data: data, + success: callback + }, jQuery.isPlainObject( url ) && url ) ); + }; +} ); + +jQuery.ajaxPrefilter( function( s ) { + var i; + for ( i in s.headers ) { + if ( i.toLowerCase() === "content-type" ) { + s.contentType = s.headers[ i ] || ""; + } + } +} ); + + +jQuery._evalUrl = function( url, options, doc ) { + return jQuery.ajax( { + url: url, + + // Make this explicit, since user can override this through ajaxSetup (#11264) + type: "GET", + dataType: "script", + cache: true, + async: false, + global: false, + + // Only evaluate the response if it is successful (gh-4126) + // dataFilter is not invoked for failure responses, so using it instead + // of the default converter is kludgy but it works. + converters: { + "text script": function() {} + }, + dataFilter: function( response ) { + jQuery.globalEval( response, options, doc ); + } + } ); +}; + + +jQuery.fn.extend( { + wrapAll: function( html ) { + var wrap; + + if ( this[ 0 ] ) { + if ( isFunction( html ) ) { + html = html.call( this[ 0 ] ); + } + + // The elements to wrap the target around + wrap = jQuery( html, this[ 0 ].ownerDocument ).eq( 0 ).clone( true ); + + if ( this[ 0 ].parentNode ) { + wrap.insertBefore( this[ 0 ] ); + } + + wrap.map( function() { + var elem = this; + + while ( elem.firstElementChild ) { + elem = elem.firstElementChild; + } + + return elem; + } ).append( this ); + } + + return this; + }, + + wrapInner: function( html ) { + if ( isFunction( html ) ) { + return this.each( function( i ) { + jQuery( this ).wrapInner( html.call( this, i ) ); + } ); + } + + return this.each( function() { + var self = jQuery( this ), + contents = self.contents(); + + if ( contents.length ) { + contents.wrapAll( html ); + + } else { + self.append( html ); + } + } ); + }, + + wrap: function( html ) { + var htmlIsFunction = isFunction( html ); + + return this.each( function( i ) { + jQuery( this ).wrapAll( htmlIsFunction ? html.call( this, i ) : html ); + } ); + }, + + unwrap: function( selector ) { + this.parent( selector ).not( "body" ).each( function() { + jQuery( this ).replaceWith( this.childNodes ); + } ); + return this; + } +} ); + + +jQuery.expr.pseudos.hidden = function( elem ) { + return !jQuery.expr.pseudos.visible( elem ); +}; +jQuery.expr.pseudos.visible = function( elem ) { + return !!( elem.offsetWidth || elem.offsetHeight || elem.getClientRects().length ); +}; + + + + +jQuery.ajaxSettings.xhr = function() { + try { + return new window.XMLHttpRequest(); + } catch ( e ) {} +}; + +var xhrSuccessStatus = { + + // File protocol always yields status code 0, assume 200 + 0: 200, + + // Support: IE <=9 only + // #1450: sometimes IE returns 1223 when it should be 204 + 1223: 204 + }, + xhrSupported = jQuery.ajaxSettings.xhr(); + +support.cors = !!xhrSupported && ( "withCredentials" in xhrSupported ); +support.ajax = xhrSupported = !!xhrSupported; + +jQuery.ajaxTransport( function( options ) { + var callback, errorCallback; + + // Cross domain only allowed if supported through XMLHttpRequest + if ( support.cors || xhrSupported && !options.crossDomain ) { + return { + send: function( headers, complete ) { + var i, + xhr = options.xhr(); + + xhr.open( + options.type, + options.url, + options.async, + options.username, + options.password + ); + + // Apply custom fields if provided + if ( options.xhrFields ) { + for ( i in options.xhrFields ) { + xhr[ i ] = options.xhrFields[ i ]; + } + } + + // Override mime type if needed + if ( options.mimeType && xhr.overrideMimeType ) { + xhr.overrideMimeType( options.mimeType ); + } + + // X-Requested-With header + // For cross-domain requests, seeing as conditions for a preflight are + // akin to a jigsaw puzzle, we simply never set it to be sure. + // (it can always be set on a per-request basis or even using ajaxSetup) + // For same-domain requests, won't change header if already provided. + if ( !options.crossDomain && !headers[ "X-Requested-With" ] ) { + headers[ "X-Requested-With" ] = "XMLHttpRequest"; + } + + // Set headers + for ( i in headers ) { + xhr.setRequestHeader( i, headers[ i ] ); + } + + // Callback + callback = function( type ) { + return function() { + if ( callback ) { + callback = errorCallback = xhr.onload = + xhr.onerror = xhr.onabort = xhr.ontimeout = + xhr.onreadystatechange = null; + + if ( type === "abort" ) { + xhr.abort(); + } else if ( type === "error" ) { + + // Support: IE <=9 only + // On a manual native abort, IE9 throws + // errors on any property access that is not readyState + if ( typeof xhr.status !== "number" ) { + complete( 0, "error" ); + } else { + complete( + + // File: protocol always yields status 0; see #8605, #14207 + xhr.status, + xhr.statusText + ); + } + } else { + complete( + xhrSuccessStatus[ xhr.status ] || xhr.status, + xhr.statusText, + + // Support: IE <=9 only + // IE9 has no XHR2 but throws on binary (trac-11426) + // For XHR2 non-text, let the caller handle it (gh-2498) + ( xhr.responseType || "text" ) !== "text" || + typeof xhr.responseText !== "string" ? + { binary: xhr.response } : + { text: xhr.responseText }, + xhr.getAllResponseHeaders() + ); + } + } + }; + }; + + // Listen to events + xhr.onload = callback(); + errorCallback = xhr.onerror = xhr.ontimeout = callback( "error" ); + + // Support: IE 9 only + // Use onreadystatechange to replace onabort + // to handle uncaught aborts + if ( xhr.onabort !== undefined ) { + xhr.onabort = errorCallback; + } else { + xhr.onreadystatechange = function() { + + // Check readyState before timeout as it changes + if ( xhr.readyState === 4 ) { + + // Allow onerror to be called first, + // but that will not handle a native abort + // Also, save errorCallback to a variable + // as xhr.onerror cannot be accessed + window.setTimeout( function() { + if ( callback ) { + errorCallback(); + } + } ); + } + }; + } + + // Create the abort callback + callback = callback( "abort" ); + + try { + + // Do send the request (this may raise an exception) + xhr.send( options.hasContent && options.data || null ); + } catch ( e ) { + + // #14683: Only rethrow if this hasn't been notified as an error yet + if ( callback ) { + throw e; + } + } + }, + + abort: function() { + if ( callback ) { + callback(); + } + } + }; + } +} ); + + + + +// Prevent auto-execution of scripts when no explicit dataType was provided (See gh-2432) +jQuery.ajaxPrefilter( function( s ) { + if ( s.crossDomain ) { + s.contents.script = false; + } +} ); + +// Install script dataType +jQuery.ajaxSetup( { + accepts: { + script: "text/javascript, application/javascript, " + + "application/ecmascript, application/x-ecmascript" + }, + contents: { + script: /\b(?:java|ecma)script\b/ + }, + converters: { + "text script": function( text ) { + jQuery.globalEval( text ); + return text; + } + } +} ); + +// Handle cache's special case and crossDomain +jQuery.ajaxPrefilter( "script", function( s ) { + if ( s.cache === undefined ) { + s.cache = false; + } + if ( s.crossDomain ) { + s.type = "GET"; + } +} ); + +// Bind script tag hack transport +jQuery.ajaxTransport( "script", function( s ) { + + // This transport only deals with cross domain or forced-by-attrs requests + if ( s.crossDomain || s.scriptAttrs ) { + var script, callback; + return { + send: function( _, complete ) { + script = jQuery( "\r\n"; + +// inject VBScript +document.write(IEBinaryToArray_ByteStr_Script); + +global.JSZipUtils._getBinaryFromXHR = function (xhr) { + var binary = xhr.responseBody; + var byteMapping = {}; + for ( var i = 0; i < 256; i++ ) { + for ( var j = 0; j < 256; j++ ) { + byteMapping[ String.fromCharCode( i + (j << 8) ) ] = + String.fromCharCode(i) + String.fromCharCode(j); + } + } + var rawBytes = IEBinaryToArray_ByteStr(binary); + var lastChr = IEBinaryToArray_ByteStr_Last(binary); + return rawBytes.replace(/[\s\S]/g, function( match ) { + return byteMapping[match]; + }) + lastChr; +}; + +// enforcing Stuk's coding style +// vim: set shiftwidth=4 softtabstop=4: + +},{}]},{},[1]) +; diff --git a/website/static/javadoc/0.8-SNAPSHOT/jquery/jszip-utils/dist/jszip-utils-ie.min.js b/website/static/javadoc/0.8-SNAPSHOT/jquery/jszip-utils/dist/jszip-utils-ie.min.js new file mode 100644 index 0000000000..93d8bc8ef2 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/jquery/jszip-utils/dist/jszip-utils-ie.min.js @@ -0,0 +1,10 @@ +/*! + +JSZipUtils - A collection of cross-browser utilities to go along with JSZip. + + +(c) 2014 Stuart Knightley, David Duponchel +Dual licenced under the MIT license or GPLv3. See https://raw.github.com/Stuk/jszip-utils/master/LICENSE.markdown. + +*/ +!function a(b,c,d){function e(g,h){if(!c[g]){if(!b[g]){var i="function"==typeof require&&require;if(!h&&i)return i(g,!0);if(f)return f(g,!0);throw new Error("Cannot find module '"+g+"'")}var j=c[g]={exports:{}};b[g][0].call(j.exports,function(a){var c=b[g][1][a];return e(c?c:a)},j,j.exports,a,b,c,d)}return c[g].exports}for(var f="function"==typeof require&&require,g=0;g\r\n";document.write(b),a.JSZipUtils._getBinaryFromXHR=function(a){for(var b=a.responseBody,c={},d=0;256>d;d++)for(var e=0;256>e;e++)c[String.fromCharCode(d+(e<<8))]=String.fromCharCode(d)+String.fromCharCode(e);var f=IEBinaryToArray_ByteStr(b),g=IEBinaryToArray_ByteStr_Last(b);return f.replace(/[\s\S]/g,function(a){return c[a]})+g}},{}]},{},[1]); diff --git a/website/static/javadoc/0.8-SNAPSHOT/jquery/jszip-utils/dist/jszip-utils.js b/website/static/javadoc/0.8-SNAPSHOT/jquery/jszip-utils/dist/jszip-utils.js new file mode 100644 index 0000000000..775895ec92 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/jquery/jszip-utils/dist/jszip-utils.js @@ -0,0 +1,118 @@ +/*! + +JSZipUtils - A collection of cross-browser utilities to go along with JSZip. + + +(c) 2014 Stuart Knightley, David Duponchel +Dual licenced under the MIT license or GPLv3. See https://raw.github.com/Stuk/jszip-utils/master/LICENSE.markdown. + +*/ +!function(e){"object"==typeof exports?module.exports=e():"function"==typeof define&&define.amd?define(e):"undefined"!=typeof window?window.JSZipUtils=e():"undefined"!=typeof global?global.JSZipUtils=e():"undefined"!=typeof self&&(self.JSZipUtils=e())}(function(){var define,module,exports;return (function e(t,n,r){function s(o,u){if(!n[o]){if(!t[o]){var a=typeof require=="function"&&require;if(!u&&a)return a(o,!0);if(i)return i(o,!0);throw new Error("Cannot find module '"+o+"'")}var f=n[o]={exports:{}};t[o][0].call(f.exports,function(e){var n=t[o][1][e];return s(n?n:e)},f,f.exports,e,t,n,r)}return n[o].exports}var i=typeof require=="function"&&require;for(var o=0;o + +(c) 2014 Stuart Knightley, David Duponchel +Dual licenced under the MIT license or GPLv3. See https://raw.github.com/Stuk/jszip-utils/master/LICENSE.markdown. + +*/ +!function(a){"object"==typeof exports?module.exports=a():"function"==typeof define&&define.amd?define(a):"undefined"!=typeof window?window.JSZipUtils=a():"undefined"!=typeof global?global.JSZipUtils=a():"undefined"!=typeof self&&(self.JSZipUtils=a())}(function(){return function a(b,c,d){function e(g,h){if(!c[g]){if(!b[g]){var i="function"==typeof require&&require;if(!h&&i)return i(g,!0);if(f)return f(g,!0);throw new Error("Cannot find module '"+g+"'")}var j=c[g]={exports:{}};b[g][0].call(j.exports,function(a){var c=b[g][1][a];return e(c?c:a)},j,j.exports,a,b,c,d)}return c[g].exports}for(var f="function"==typeof require&&require,g=0;g + +(c) 2009-2016 Stuart Knightley +Dual licenced under the MIT license or GPLv3. See https://raw.github.com/Stuk/jszip/master/LICENSE.markdown. + +JSZip uses the library pako released under the MIT license : +https://github.com/nodeca/pako/blob/master/LICENSE +*/ + +(function(f){if(typeof exports==="object"&&typeof module!=="undefined"){module.exports=f()}else if(typeof define==="function"&&define.amd){define([],f)}else{var g;if(typeof window!=="undefined"){g=window}else if(typeof global!=="undefined"){g=global}else if(typeof self!=="undefined"){g=self}else{g=this}g.JSZip = f()}})(function(){var define,module,exports;return (function e(t,n,r){function s(o,u){if(!n[o]){if(!t[o]){var a=typeof require=="function"&&require;if(!u&&a)return a(o,!0);if(i)return i(o,!0);var f=new Error("Cannot find module '"+o+"'");throw f.code="MODULE_NOT_FOUND",f}var l=n[o]={exports:{}};t[o][0].call(l.exports,function(e){var n=t[o][1][e];return s(n?n:e)},l,l.exports,e,t,n,r)}return n[o].exports}var i=typeof require=="function"&&require;for(var o=0;o> 2; + enc2 = ((chr1 & 3) << 4) | (chr2 >> 4); + enc3 = remainingBytes > 1 ? (((chr2 & 15) << 2) | (chr3 >> 6)) : 64; + enc4 = remainingBytes > 2 ? (chr3 & 63) : 64; + + output.push(_keyStr.charAt(enc1) + _keyStr.charAt(enc2) + _keyStr.charAt(enc3) + _keyStr.charAt(enc4)); + + } + + return output.join(""); +}; + +// public method for decoding +exports.decode = function(input) { + var chr1, chr2, chr3; + var enc1, enc2, enc3, enc4; + var i = 0, resultIndex = 0; + + var dataUrlPrefix = "data:"; + + if (input.substr(0, dataUrlPrefix.length) === dataUrlPrefix) { + // This is a common error: people give a data url + // (data:image/png;base64,iVBOR...) with a {base64: true} and + // wonders why things don't work. + // We can detect that the string input looks like a data url but we + // *can't* be sure it is one: removing everything up to the comma would + // be too dangerous. + throw new Error("Invalid base64 input, it looks like a data url."); + } + + input = input.replace(/[^A-Za-z0-9\+\/\=]/g, ""); + + var totalLength = input.length * 3 / 4; + if(input.charAt(input.length - 1) === _keyStr.charAt(64)) { + totalLength--; + } + if(input.charAt(input.length - 2) === _keyStr.charAt(64)) { + totalLength--; + } + if (totalLength % 1 !== 0) { + // totalLength is not an integer, the length does not match a valid + // base64 content. That can happen if: + // - the input is not a base64 content + // - the input is *almost* a base64 content, with a extra chars at the + // beginning or at the end + // - the input uses a base64 variant (base64url for example) + throw new Error("Invalid base64 input, bad content length."); + } + var output; + if (support.uint8array) { + output = new Uint8Array(totalLength|0); + } else { + output = new Array(totalLength|0); + } + + while (i < input.length) { + + enc1 = _keyStr.indexOf(input.charAt(i++)); + enc2 = _keyStr.indexOf(input.charAt(i++)); + enc3 = _keyStr.indexOf(input.charAt(i++)); + enc4 = _keyStr.indexOf(input.charAt(i++)); + + chr1 = (enc1 << 2) | (enc2 >> 4); + chr2 = ((enc2 & 15) << 4) | (enc3 >> 2); + chr3 = ((enc3 & 3) << 6) | enc4; + + output[resultIndex++] = chr1; + + if (enc3 !== 64) { + output[resultIndex++] = chr2; + } + if (enc4 !== 64) { + output[resultIndex++] = chr3; + } + + } + + return output; +}; + +},{"./support":30,"./utils":32}],2:[function(require,module,exports){ +'use strict'; + +var external = require("./external"); +var DataWorker = require('./stream/DataWorker'); +var Crc32Probe = require('./stream/Crc32Probe'); +var DataLengthProbe = require('./stream/DataLengthProbe'); + +/** + * Represent a compressed object, with everything needed to decompress it. + * @constructor + * @param {number} compressedSize the size of the data compressed. + * @param {number} uncompressedSize the size of the data after decompression. + * @param {number} crc32 the crc32 of the decompressed file. + * @param {object} compression the type of compression, see lib/compressions.js. + * @param {String|ArrayBuffer|Uint8Array|Buffer} data the compressed data. + */ +function CompressedObject(compressedSize, uncompressedSize, crc32, compression, data) { + this.compressedSize = compressedSize; + this.uncompressedSize = uncompressedSize; + this.crc32 = crc32; + this.compression = compression; + this.compressedContent = data; +} + +CompressedObject.prototype = { + /** + * Create a worker to get the uncompressed content. + * @return {GenericWorker} the worker. + */ + getContentWorker: function () { + var worker = new DataWorker(external.Promise.resolve(this.compressedContent)) + .pipe(this.compression.uncompressWorker()) + .pipe(new DataLengthProbe("data_length")); + + var that = this; + worker.on("end", function () { + if (this.streamInfo['data_length'] !== that.uncompressedSize) { + throw new Error("Bug : uncompressed data size mismatch"); + } + }); + return worker; + }, + /** + * Create a worker to get the compressed content. + * @return {GenericWorker} the worker. + */ + getCompressedWorker: function () { + return new DataWorker(external.Promise.resolve(this.compressedContent)) + .withStreamInfo("compressedSize", this.compressedSize) + .withStreamInfo("uncompressedSize", this.uncompressedSize) + .withStreamInfo("crc32", this.crc32) + .withStreamInfo("compression", this.compression) + ; + } +}; + +/** + * Chain the given worker with other workers to compress the content with the + * given compression. + * @param {GenericWorker} uncompressedWorker the worker to pipe. + * @param {Object} compression the compression object. + * @param {Object} compressionOptions the options to use when compressing. + * @return {GenericWorker} the new worker compressing the content. + */ +CompressedObject.createWorkerFrom = function (uncompressedWorker, compression, compressionOptions) { + return uncompressedWorker + .pipe(new Crc32Probe()) + .pipe(new DataLengthProbe("uncompressedSize")) + .pipe(compression.compressWorker(compressionOptions)) + .pipe(new DataLengthProbe("compressedSize")) + .withStreamInfo("compression", compression); +}; + +module.exports = CompressedObject; + +},{"./external":6,"./stream/Crc32Probe":25,"./stream/DataLengthProbe":26,"./stream/DataWorker":27}],3:[function(require,module,exports){ +'use strict'; + +var GenericWorker = require("./stream/GenericWorker"); + +exports.STORE = { + magic: "\x00\x00", + compressWorker : function (compressionOptions) { + return new GenericWorker("STORE compression"); + }, + uncompressWorker : function () { + return new GenericWorker("STORE decompression"); + } +}; +exports.DEFLATE = require('./flate'); + +},{"./flate":7,"./stream/GenericWorker":28}],4:[function(require,module,exports){ +'use strict'; + +var utils = require('./utils'); + +/** + * The following functions come from pako, from pako/lib/zlib/crc32.js + * released under the MIT license, see pako https://github.com/nodeca/pako/ + */ + +// Use ordinary array, since untyped makes no boost here +function makeTable() { + var c, table = []; + + for(var n =0; n < 256; n++){ + c = n; + for(var k =0; k < 8; k++){ + c = ((c&1) ? (0xEDB88320 ^ (c >>> 1)) : (c >>> 1)); + } + table[n] = c; + } + + return table; +} + +// Create table on load. Just 255 signed longs. Not a problem. +var crcTable = makeTable(); + + +function crc32(crc, buf, len, pos) { + var t = crcTable, end = pos + len; + + crc = crc ^ (-1); + + for (var i = pos; i < end; i++ ) { + crc = (crc >>> 8) ^ t[(crc ^ buf[i]) & 0xFF]; + } + + return (crc ^ (-1)); // >>> 0; +} + +// That's all for the pako functions. + +/** + * Compute the crc32 of a string. + * This is almost the same as the function crc32, but for strings. Using the + * same function for the two use cases leads to horrible performances. + * @param {Number} crc the starting value of the crc. + * @param {String} str the string to use. + * @param {Number} len the length of the string. + * @param {Number} pos the starting position for the crc32 computation. + * @return {Number} the computed crc32. + */ +function crc32str(crc, str, len, pos) { + var t = crcTable, end = pos + len; + + crc = crc ^ (-1); + + for (var i = pos; i < end; i++ ) { + crc = (crc >>> 8) ^ t[(crc ^ str.charCodeAt(i)) & 0xFF]; + } + + return (crc ^ (-1)); // >>> 0; +} + +module.exports = function crc32wrapper(input, crc) { + if (typeof input === "undefined" || !input.length) { + return 0; + } + + var isArray = utils.getTypeOf(input) !== "string"; + + if(isArray) { + return crc32(crc|0, input, input.length, 0); + } else { + return crc32str(crc|0, input, input.length, 0); + } +}; + +},{"./utils":32}],5:[function(require,module,exports){ +'use strict'; +exports.base64 = false; +exports.binary = false; +exports.dir = false; +exports.createFolders = true; +exports.date = null; +exports.compression = null; +exports.compressionOptions = null; +exports.comment = null; +exports.unixPermissions = null; +exports.dosPermissions = null; + +},{}],6:[function(require,module,exports){ +/* global Promise */ +'use strict'; + +// load the global object first: +// - it should be better integrated in the system (unhandledRejection in node) +// - the environment may have a custom Promise implementation (see zone.js) +var ES6Promise = null; +if (typeof Promise !== "undefined") { + ES6Promise = Promise; +} else { + ES6Promise = require("lie"); +} + +/** + * Let the user use/change some implementations. + */ +module.exports = { + Promise: ES6Promise +}; + +},{"lie":37}],7:[function(require,module,exports){ +'use strict'; +var USE_TYPEDARRAY = (typeof Uint8Array !== 'undefined') && (typeof Uint16Array !== 'undefined') && (typeof Uint32Array !== 'undefined'); + +var pako = require("pako"); +var utils = require("./utils"); +var GenericWorker = require("./stream/GenericWorker"); + +var ARRAY_TYPE = USE_TYPEDARRAY ? "uint8array" : "array"; + +exports.magic = "\x08\x00"; + +/** + * Create a worker that uses pako to inflate/deflate. + * @constructor + * @param {String} action the name of the pako function to call : either "Deflate" or "Inflate". + * @param {Object} options the options to use when (de)compressing. + */ +function FlateWorker(action, options) { + GenericWorker.call(this, "FlateWorker/" + action); + + this._pako = null; + this._pakoAction = action; + this._pakoOptions = options; + // the `meta` object from the last chunk received + // this allow this worker to pass around metadata + this.meta = {}; +} + +utils.inherits(FlateWorker, GenericWorker); + +/** + * @see GenericWorker.processChunk + */ +FlateWorker.prototype.processChunk = function (chunk) { + this.meta = chunk.meta; + if (this._pako === null) { + this._createPako(); + } + this._pako.push(utils.transformTo(ARRAY_TYPE, chunk.data), false); +}; + +/** + * @see GenericWorker.flush + */ +FlateWorker.prototype.flush = function () { + GenericWorker.prototype.flush.call(this); + if (this._pako === null) { + this._createPako(); + } + this._pako.push([], true); +}; +/** + * @see GenericWorker.cleanUp + */ +FlateWorker.prototype.cleanUp = function () { + GenericWorker.prototype.cleanUp.call(this); + this._pako = null; +}; + +/** + * Create the _pako object. + * TODO: lazy-loading this object isn't the best solution but it's the + * quickest. The best solution is to lazy-load the worker list. See also the + * issue #446. + */ +FlateWorker.prototype._createPako = function () { + this._pako = new pako[this._pakoAction]({ + raw: true, + level: this._pakoOptions.level || -1 // default compression + }); + var self = this; + this._pako.onData = function(data) { + self.push({ + data : data, + meta : self.meta + }); + }; +}; + +exports.compressWorker = function (compressionOptions) { + return new FlateWorker("Deflate", compressionOptions); +}; +exports.uncompressWorker = function () { + return new FlateWorker("Inflate", {}); +}; + +},{"./stream/GenericWorker":28,"./utils":32,"pako":38}],8:[function(require,module,exports){ +'use strict'; + +var utils = require('../utils'); +var GenericWorker = require('../stream/GenericWorker'); +var utf8 = require('../utf8'); +var crc32 = require('../crc32'); +var signature = require('../signature'); + +/** + * Transform an integer into a string in hexadecimal. + * @private + * @param {number} dec the number to convert. + * @param {number} bytes the number of bytes to generate. + * @returns {string} the result. + */ +var decToHex = function(dec, bytes) { + var hex = "", i; + for (i = 0; i < bytes; i++) { + hex += String.fromCharCode(dec & 0xff); + dec = dec >>> 8; + } + return hex; +}; + +/** + * Generate the UNIX part of the external file attributes. + * @param {Object} unixPermissions the unix permissions or null. + * @param {Boolean} isDir true if the entry is a directory, false otherwise. + * @return {Number} a 32 bit integer. + * + * adapted from http://unix.stackexchange.com/questions/14705/the-zip-formats-external-file-attribute : + * + * TTTTsstrwxrwxrwx0000000000ADVSHR + * ^^^^____________________________ file type, see zipinfo.c (UNX_*) + * ^^^_________________________ setuid, setgid, sticky + * ^^^^^^^^^________________ permissions + * ^^^^^^^^^^______ not used ? + * ^^^^^^ DOS attribute bits : Archive, Directory, Volume label, System file, Hidden, Read only + */ +var generateUnixExternalFileAttr = function (unixPermissions, isDir) { + + var result = unixPermissions; + if (!unixPermissions) { + // I can't use octal values in strict mode, hence the hexa. + // 040775 => 0x41fd + // 0100664 => 0x81b4 + result = isDir ? 0x41fd : 0x81b4; + } + return (result & 0xFFFF) << 16; +}; + +/** + * Generate the DOS part of the external file attributes. + * @param {Object} dosPermissions the dos permissions or null. + * @param {Boolean} isDir true if the entry is a directory, false otherwise. + * @return {Number} a 32 bit integer. + * + * Bit 0 Read-Only + * Bit 1 Hidden + * Bit 2 System + * Bit 3 Volume Label + * Bit 4 Directory + * Bit 5 Archive + */ +var generateDosExternalFileAttr = function (dosPermissions, isDir) { + + // the dir flag is already set for compatibility + return (dosPermissions || 0) & 0x3F; +}; + +/** + * Generate the various parts used in the construction of the final zip file. + * @param {Object} streamInfo the hash with information about the compressed file. + * @param {Boolean} streamedContent is the content streamed ? + * @param {Boolean} streamingEnded is the stream finished ? + * @param {number} offset the current offset from the start of the zip file. + * @param {String} platform let's pretend we are this platform (change platform dependents fields) + * @param {Function} encodeFileName the function to encode the file name / comment. + * @return {Object} the zip parts. + */ +var generateZipParts = function(streamInfo, streamedContent, streamingEnded, offset, platform, encodeFileName) { + var file = streamInfo['file'], + compression = streamInfo['compression'], + useCustomEncoding = encodeFileName !== utf8.utf8encode, + encodedFileName = utils.transformTo("string", encodeFileName(file.name)), + utfEncodedFileName = utils.transformTo("string", utf8.utf8encode(file.name)), + comment = file.comment, + encodedComment = utils.transformTo("string", encodeFileName(comment)), + utfEncodedComment = utils.transformTo("string", utf8.utf8encode(comment)), + useUTF8ForFileName = utfEncodedFileName.length !== file.name.length, + useUTF8ForComment = utfEncodedComment.length !== comment.length, + dosTime, + dosDate, + extraFields = "", + unicodePathExtraField = "", + unicodeCommentExtraField = "", + dir = file.dir, + date = file.date; + + + var dataInfo = { + crc32 : 0, + compressedSize : 0, + uncompressedSize : 0 + }; + + // if the content is streamed, the sizes/crc32 are only available AFTER + // the end of the stream. + if (!streamedContent || streamingEnded) { + dataInfo.crc32 = streamInfo['crc32']; + dataInfo.compressedSize = streamInfo['compressedSize']; + dataInfo.uncompressedSize = streamInfo['uncompressedSize']; + } + + var bitflag = 0; + if (streamedContent) { + // Bit 3: the sizes/crc32 are set to zero in the local header. + // The correct values are put in the data descriptor immediately + // following the compressed data. + bitflag |= 0x0008; + } + if (!useCustomEncoding && (useUTF8ForFileName || useUTF8ForComment)) { + // Bit 11: Language encoding flag (EFS). + bitflag |= 0x0800; + } + + + var extFileAttr = 0; + var versionMadeBy = 0; + if (dir) { + // dos or unix, we set the dos dir flag + extFileAttr |= 0x00010; + } + if(platform === "UNIX") { + versionMadeBy = 0x031E; // UNIX, version 3.0 + extFileAttr |= generateUnixExternalFileAttr(file.unixPermissions, dir); + } else { // DOS or other, fallback to DOS + versionMadeBy = 0x0014; // DOS, version 2.0 + extFileAttr |= generateDosExternalFileAttr(file.dosPermissions, dir); + } + + // date + // @see http://www.delorie.com/djgpp/doc/rbinter/it/52/13.html + // @see http://www.delorie.com/djgpp/doc/rbinter/it/65/16.html + // @see http://www.delorie.com/djgpp/doc/rbinter/it/66/16.html + + dosTime = date.getUTCHours(); + dosTime = dosTime << 6; + dosTime = dosTime | date.getUTCMinutes(); + dosTime = dosTime << 5; + dosTime = dosTime | date.getUTCSeconds() / 2; + + dosDate = date.getUTCFullYear() - 1980; + dosDate = dosDate << 4; + dosDate = dosDate | (date.getUTCMonth() + 1); + dosDate = dosDate << 5; + dosDate = dosDate | date.getUTCDate(); + + if (useUTF8ForFileName) { + // set the unicode path extra field. unzip needs at least one extra + // field to correctly handle unicode path, so using the path is as good + // as any other information. This could improve the situation with + // other archive managers too. + // This field is usually used without the utf8 flag, with a non + // unicode path in the header (winrar, winzip). This helps (a bit) + // with the messy Windows' default compressed folders feature but + // breaks on p7zip which doesn't seek the unicode path extra field. + // So for now, UTF-8 everywhere ! + unicodePathExtraField = + // Version + decToHex(1, 1) + + // NameCRC32 + decToHex(crc32(encodedFileName), 4) + + // UnicodeName + utfEncodedFileName; + + extraFields += + // Info-ZIP Unicode Path Extra Field + "\x75\x70" + + // size + decToHex(unicodePathExtraField.length, 2) + + // content + unicodePathExtraField; + } + + if(useUTF8ForComment) { + + unicodeCommentExtraField = + // Version + decToHex(1, 1) + + // CommentCRC32 + decToHex(crc32(encodedComment), 4) + + // UnicodeName + utfEncodedComment; + + extraFields += + // Info-ZIP Unicode Path Extra Field + "\x75\x63" + + // size + decToHex(unicodeCommentExtraField.length, 2) + + // content + unicodeCommentExtraField; + } + + var header = ""; + + // version needed to extract + header += "\x0A\x00"; + // general purpose bit flag + header += decToHex(bitflag, 2); + // compression method + header += compression.magic; + // last mod file time + header += decToHex(dosTime, 2); + // last mod file date + header += decToHex(dosDate, 2); + // crc-32 + header += decToHex(dataInfo.crc32, 4); + // compressed size + header += decToHex(dataInfo.compressedSize, 4); + // uncompressed size + header += decToHex(dataInfo.uncompressedSize, 4); + // file name length + header += decToHex(encodedFileName.length, 2); + // extra field length + header += decToHex(extraFields.length, 2); + + + var fileRecord = signature.LOCAL_FILE_HEADER + header + encodedFileName + extraFields; + + var dirRecord = signature.CENTRAL_FILE_HEADER + + // version made by (00: DOS) + decToHex(versionMadeBy, 2) + + // file header (common to file and central directory) + header + + // file comment length + decToHex(encodedComment.length, 2) + + // disk number start + "\x00\x00" + + // internal file attributes TODO + "\x00\x00" + + // external file attributes + decToHex(extFileAttr, 4) + + // relative offset of local header + decToHex(offset, 4) + + // file name + encodedFileName + + // extra field + extraFields + + // file comment + encodedComment; + + return { + fileRecord: fileRecord, + dirRecord: dirRecord + }; +}; + +/** + * Generate the EOCD record. + * @param {Number} entriesCount the number of entries in the zip file. + * @param {Number} centralDirLength the length (in bytes) of the central dir. + * @param {Number} localDirLength the length (in bytes) of the local dir. + * @param {String} comment the zip file comment as a binary string. + * @param {Function} encodeFileName the function to encode the comment. + * @return {String} the EOCD record. + */ +var generateCentralDirectoryEnd = function (entriesCount, centralDirLength, localDirLength, comment, encodeFileName) { + var dirEnd = ""; + var encodedComment = utils.transformTo("string", encodeFileName(comment)); + + // end of central dir signature + dirEnd = signature.CENTRAL_DIRECTORY_END + + // number of this disk + "\x00\x00" + + // number of the disk with the start of the central directory + "\x00\x00" + + // total number of entries in the central directory on this disk + decToHex(entriesCount, 2) + + // total number of entries in the central directory + decToHex(entriesCount, 2) + + // size of the central directory 4 bytes + decToHex(centralDirLength, 4) + + // offset of start of central directory with respect to the starting disk number + decToHex(localDirLength, 4) + + // .ZIP file comment length + decToHex(encodedComment.length, 2) + + // .ZIP file comment + encodedComment; + + return dirEnd; +}; + +/** + * Generate data descriptors for a file entry. + * @param {Object} streamInfo the hash generated by a worker, containing information + * on the file entry. + * @return {String} the data descriptors. + */ +var generateDataDescriptors = function (streamInfo) { + var descriptor = ""; + descriptor = signature.DATA_DESCRIPTOR + + // crc-32 4 bytes + decToHex(streamInfo['crc32'], 4) + + // compressed size 4 bytes + decToHex(streamInfo['compressedSize'], 4) + + // uncompressed size 4 bytes + decToHex(streamInfo['uncompressedSize'], 4); + + return descriptor; +}; + + +/** + * A worker to concatenate other workers to create a zip file. + * @param {Boolean} streamFiles `true` to stream the content of the files, + * `false` to accumulate it. + * @param {String} comment the comment to use. + * @param {String} platform the platform to use, "UNIX" or "DOS". + * @param {Function} encodeFileName the function to encode file names and comments. + */ +function ZipFileWorker(streamFiles, comment, platform, encodeFileName) { + GenericWorker.call(this, "ZipFileWorker"); + // The number of bytes written so far. This doesn't count accumulated chunks. + this.bytesWritten = 0; + // The comment of the zip file + this.zipComment = comment; + // The platform "generating" the zip file. + this.zipPlatform = platform; + // the function to encode file names and comments. + this.encodeFileName = encodeFileName; + // Should we stream the content of the files ? + this.streamFiles = streamFiles; + // If `streamFiles` is false, we will need to accumulate the content of the + // files to calculate sizes / crc32 (and write them *before* the content). + // This boolean indicates if we are accumulating chunks (it will change a lot + // during the lifetime of this worker). + this.accumulate = false; + // The buffer receiving chunks when accumulating content. + this.contentBuffer = []; + // The list of generated directory records. + this.dirRecords = []; + // The offset (in bytes) from the beginning of the zip file for the current source. + this.currentSourceOffset = 0; + // The total number of entries in this zip file. + this.entriesCount = 0; + // the name of the file currently being added, null when handling the end of the zip file. + // Used for the emitted metadata. + this.currentFile = null; + + + + this._sources = []; +} +utils.inherits(ZipFileWorker, GenericWorker); + +/** + * @see GenericWorker.push + */ +ZipFileWorker.prototype.push = function (chunk) { + + var currentFilePercent = chunk.meta.percent || 0; + var entriesCount = this.entriesCount; + var remainingFiles = this._sources.length; + + if(this.accumulate) { + this.contentBuffer.push(chunk); + } else { + this.bytesWritten += chunk.data.length; + + GenericWorker.prototype.push.call(this, { + data : chunk.data, + meta : { + currentFile : this.currentFile, + percent : entriesCount ? (currentFilePercent + 100 * (entriesCount - remainingFiles - 1)) / entriesCount : 100 + } + }); + } +}; + +/** + * The worker started a new source (an other worker). + * @param {Object} streamInfo the streamInfo object from the new source. + */ +ZipFileWorker.prototype.openedSource = function (streamInfo) { + this.currentSourceOffset = this.bytesWritten; + this.currentFile = streamInfo['file'].name; + + var streamedContent = this.streamFiles && !streamInfo['file'].dir; + + // don't stream folders (because they don't have any content) + if(streamedContent) { + var record = generateZipParts(streamInfo, streamedContent, false, this.currentSourceOffset, this.zipPlatform, this.encodeFileName); + this.push({ + data : record.fileRecord, + meta : {percent:0} + }); + } else { + // we need to wait for the whole file before pushing anything + this.accumulate = true; + } +}; + +/** + * The worker finished a source (an other worker). + * @param {Object} streamInfo the streamInfo object from the finished source. + */ +ZipFileWorker.prototype.closedSource = function (streamInfo) { + this.accumulate = false; + var streamedContent = this.streamFiles && !streamInfo['file'].dir; + var record = generateZipParts(streamInfo, streamedContent, true, this.currentSourceOffset, this.zipPlatform, this.encodeFileName); + + this.dirRecords.push(record.dirRecord); + if(streamedContent) { + // after the streamed file, we put data descriptors + this.push({ + data : generateDataDescriptors(streamInfo), + meta : {percent:100} + }); + } else { + // the content wasn't streamed, we need to push everything now + // first the file record, then the content + this.push({ + data : record.fileRecord, + meta : {percent:0} + }); + while(this.contentBuffer.length) { + this.push(this.contentBuffer.shift()); + } + } + this.currentFile = null; +}; + +/** + * @see GenericWorker.flush + */ +ZipFileWorker.prototype.flush = function () { + + var localDirLength = this.bytesWritten; + for(var i = 0; i < this.dirRecords.length; i++) { + this.push({ + data : this.dirRecords[i], + meta : {percent:100} + }); + } + var centralDirLength = this.bytesWritten - localDirLength; + + var dirEnd = generateCentralDirectoryEnd(this.dirRecords.length, centralDirLength, localDirLength, this.zipComment, this.encodeFileName); + + this.push({ + data : dirEnd, + meta : {percent:100} + }); +}; + +/** + * Prepare the next source to be read. + */ +ZipFileWorker.prototype.prepareNextSource = function () { + this.previous = this._sources.shift(); + this.openedSource(this.previous.streamInfo); + if (this.isPaused) { + this.previous.pause(); + } else { + this.previous.resume(); + } +}; + +/** + * @see GenericWorker.registerPrevious + */ +ZipFileWorker.prototype.registerPrevious = function (previous) { + this._sources.push(previous); + var self = this; + + previous.on('data', function (chunk) { + self.processChunk(chunk); + }); + previous.on('end', function () { + self.closedSource(self.previous.streamInfo); + if(self._sources.length) { + self.prepareNextSource(); + } else { + self.end(); + } + }); + previous.on('error', function (e) { + self.error(e); + }); + return this; +}; + +/** + * @see GenericWorker.resume + */ +ZipFileWorker.prototype.resume = function () { + if(!GenericWorker.prototype.resume.call(this)) { + return false; + } + + if (!this.previous && this._sources.length) { + this.prepareNextSource(); + return true; + } + if (!this.previous && !this._sources.length && !this.generatedError) { + this.end(); + return true; + } +}; + +/** + * @see GenericWorker.error + */ +ZipFileWorker.prototype.error = function (e) { + var sources = this._sources; + if(!GenericWorker.prototype.error.call(this, e)) { + return false; + } + for(var i = 0; i < sources.length; i++) { + try { + sources[i].error(e); + } catch(e) { + // the `error` exploded, nothing to do + } + } + return true; +}; + +/** + * @see GenericWorker.lock + */ +ZipFileWorker.prototype.lock = function () { + GenericWorker.prototype.lock.call(this); + var sources = this._sources; + for(var i = 0; i < sources.length; i++) { + sources[i].lock(); + } +}; + +module.exports = ZipFileWorker; + +},{"../crc32":4,"../signature":23,"../stream/GenericWorker":28,"../utf8":31,"../utils":32}],9:[function(require,module,exports){ +'use strict'; + +var compressions = require('../compressions'); +var ZipFileWorker = require('./ZipFileWorker'); + +/** + * Find the compression to use. + * @param {String} fileCompression the compression defined at the file level, if any. + * @param {String} zipCompression the compression defined at the load() level. + * @return {Object} the compression object to use. + */ +var getCompression = function (fileCompression, zipCompression) { + + var compressionName = fileCompression || zipCompression; + var compression = compressions[compressionName]; + if (!compression) { + throw new Error(compressionName + " is not a valid compression method !"); + } + return compression; +}; + +/** + * Create a worker to generate a zip file. + * @param {JSZip} zip the JSZip instance at the right root level. + * @param {Object} options to generate the zip file. + * @param {String} comment the comment to use. + */ +exports.generateWorker = function (zip, options, comment) { + + var zipFileWorker = new ZipFileWorker(options.streamFiles, comment, options.platform, options.encodeFileName); + var entriesCount = 0; + try { + + zip.forEach(function (relativePath, file) { + entriesCount++; + var compression = getCompression(file.options.compression, options.compression); + var compressionOptions = file.options.compressionOptions || options.compressionOptions || {}; + var dir = file.dir, date = file.date; + + file._compressWorker(compression, compressionOptions) + .withStreamInfo("file", { + name : relativePath, + dir : dir, + date : date, + comment : file.comment || "", + unixPermissions : file.unixPermissions, + dosPermissions : file.dosPermissions + }) + .pipe(zipFileWorker); + }); + zipFileWorker.entriesCount = entriesCount; + } catch (e) { + zipFileWorker.error(e); + } + + return zipFileWorker; +}; + +},{"../compressions":3,"./ZipFileWorker":8}],10:[function(require,module,exports){ +'use strict'; + +/** + * Representation a of zip file in js + * @constructor + */ +function JSZip() { + // if this constructor is used without `new`, it adds `new` before itself: + if(!(this instanceof JSZip)) { + return new JSZip(); + } + + if(arguments.length) { + throw new Error("The constructor with parameters has been removed in JSZip 3.0, please check the upgrade guide."); + } + + // object containing the files : + // { + // "folder/" : {...}, + // "folder/data.txt" : {...} + // } + // NOTE: we use a null prototype because we do not + // want filenames like "toString" coming from a zip file + // to overwrite methods and attributes in a normal Object. + this.files = Object.create(null); + + this.comment = null; + + // Where we are in the hierarchy + this.root = ""; + this.clone = function() { + var newObj = new JSZip(); + for (var i in this) { + if (typeof this[i] !== "function") { + newObj[i] = this[i]; + } + } + return newObj; + }; +} +JSZip.prototype = require('./object'); +JSZip.prototype.loadAsync = require('./load'); +JSZip.support = require('./support'); +JSZip.defaults = require('./defaults'); + +// TODO find a better way to handle this version, +// a require('package.json').version doesn't work with webpack, see #327 +JSZip.version = "3.7.1"; + +JSZip.loadAsync = function (content, options) { + return new JSZip().loadAsync(content, options); +}; + +JSZip.external = require("./external"); +module.exports = JSZip; + +},{"./defaults":5,"./external":6,"./load":11,"./object":15,"./support":30}],11:[function(require,module,exports){ +'use strict'; +var utils = require('./utils'); +var external = require("./external"); +var utf8 = require('./utf8'); +var ZipEntries = require('./zipEntries'); +var Crc32Probe = require('./stream/Crc32Probe'); +var nodejsUtils = require("./nodejsUtils"); + +/** + * Check the CRC32 of an entry. + * @param {ZipEntry} zipEntry the zip entry to check. + * @return {Promise} the result. + */ +function checkEntryCRC32(zipEntry) { + return new external.Promise(function (resolve, reject) { + var worker = zipEntry.decompressed.getContentWorker().pipe(new Crc32Probe()); + worker.on("error", function (e) { + reject(e); + }) + .on("end", function () { + if (worker.streamInfo.crc32 !== zipEntry.decompressed.crc32) { + reject(new Error("Corrupted zip : CRC32 mismatch")); + } else { + resolve(); + } + }) + .resume(); + }); +} + +module.exports = function (data, options) { + var zip = this; + options = utils.extend(options || {}, { + base64: false, + checkCRC32: false, + optimizedBinaryString: false, + createFolders: false, + decodeFileName: utf8.utf8decode + }); + + if (nodejsUtils.isNode && nodejsUtils.isStream(data)) { + return external.Promise.reject(new Error("JSZip can't accept a stream when loading a zip file.")); + } + + return utils.prepareContent("the loaded zip file", data, true, options.optimizedBinaryString, options.base64) + .then(function (data) { + var zipEntries = new ZipEntries(options); + zipEntries.load(data); + return zipEntries; + }).then(function checkCRC32(zipEntries) { + var promises = [external.Promise.resolve(zipEntries)]; + var files = zipEntries.files; + if (options.checkCRC32) { + for (var i = 0; i < files.length; i++) { + promises.push(checkEntryCRC32(files[i])); + } + } + return external.Promise.all(promises); + }).then(function addFiles(results) { + var zipEntries = results.shift(); + var files = zipEntries.files; + for (var i = 0; i < files.length; i++) { + var input = files[i]; + zip.file(input.fileNameStr, input.decompressed, { + binary: true, + optimizedBinaryString: true, + date: input.date, + dir: input.dir, + comment: input.fileCommentStr.length ? input.fileCommentStr : null, + unixPermissions: input.unixPermissions, + dosPermissions: input.dosPermissions, + createFolders: options.createFolders + }); + } + if (zipEntries.zipComment.length) { + zip.comment = zipEntries.zipComment; + } + + return zip; + }); +}; + +},{"./external":6,"./nodejsUtils":14,"./stream/Crc32Probe":25,"./utf8":31,"./utils":32,"./zipEntries":33}],12:[function(require,module,exports){ +"use strict"; + +var utils = require('../utils'); +var GenericWorker = require('../stream/GenericWorker'); + +/** + * A worker that use a nodejs stream as source. + * @constructor + * @param {String} filename the name of the file entry for this stream. + * @param {Readable} stream the nodejs stream. + */ +function NodejsStreamInputAdapter(filename, stream) { + GenericWorker.call(this, "Nodejs stream input adapter for " + filename); + this._upstreamEnded = false; + this._bindStream(stream); +} + +utils.inherits(NodejsStreamInputAdapter, GenericWorker); + +/** + * Prepare the stream and bind the callbacks on it. + * Do this ASAP on node 0.10 ! A lazy binding doesn't always work. + * @param {Stream} stream the nodejs stream to use. + */ +NodejsStreamInputAdapter.prototype._bindStream = function (stream) { + var self = this; + this._stream = stream; + stream.pause(); + stream + .on("data", function (chunk) { + self.push({ + data: chunk, + meta : { + percent : 0 + } + }); + }) + .on("error", function (e) { + if(self.isPaused) { + this.generatedError = e; + } else { + self.error(e); + } + }) + .on("end", function () { + if(self.isPaused) { + self._upstreamEnded = true; + } else { + self.end(); + } + }); +}; +NodejsStreamInputAdapter.prototype.pause = function () { + if(!GenericWorker.prototype.pause.call(this)) { + return false; + } + this._stream.pause(); + return true; +}; +NodejsStreamInputAdapter.prototype.resume = function () { + if(!GenericWorker.prototype.resume.call(this)) { + return false; + } + + if(this._upstreamEnded) { + this.end(); + } else { + this._stream.resume(); + } + + return true; +}; + +module.exports = NodejsStreamInputAdapter; + +},{"../stream/GenericWorker":28,"../utils":32}],13:[function(require,module,exports){ +'use strict'; + +var Readable = require('readable-stream').Readable; + +var utils = require('../utils'); +utils.inherits(NodejsStreamOutputAdapter, Readable); + +/** +* A nodejs stream using a worker as source. +* @see the SourceWrapper in http://nodejs.org/api/stream.html +* @constructor +* @param {StreamHelper} helper the helper wrapping the worker +* @param {Object} options the nodejs stream options +* @param {Function} updateCb the update callback. +*/ +function NodejsStreamOutputAdapter(helper, options, updateCb) { + Readable.call(this, options); + this._helper = helper; + + var self = this; + helper.on("data", function (data, meta) { + if (!self.push(data)) { + self._helper.pause(); + } + if(updateCb) { + updateCb(meta); + } + }) + .on("error", function(e) { + self.emit('error', e); + }) + .on("end", function () { + self.push(null); + }); +} + + +NodejsStreamOutputAdapter.prototype._read = function() { + this._helper.resume(); +}; + +module.exports = NodejsStreamOutputAdapter; + +},{"../utils":32,"readable-stream":16}],14:[function(require,module,exports){ +'use strict'; + +module.exports = { + /** + * True if this is running in Nodejs, will be undefined in a browser. + * In a browser, browserify won't include this file and the whole module + * will be resolved an empty object. + */ + isNode : typeof Buffer !== "undefined", + /** + * Create a new nodejs Buffer from an existing content. + * @param {Object} data the data to pass to the constructor. + * @param {String} encoding the encoding to use. + * @return {Buffer} a new Buffer. + */ + newBufferFrom: function(data, encoding) { + if (Buffer.from && Buffer.from !== Uint8Array.from) { + return Buffer.from(data, encoding); + } else { + if (typeof data === "number") { + // Safeguard for old Node.js versions. On newer versions, + // Buffer.from(number) / Buffer(number, encoding) already throw. + throw new Error("The \"data\" argument must not be a number"); + } + return new Buffer(data, encoding); + } + }, + /** + * Create a new nodejs Buffer with the specified size. + * @param {Integer} size the size of the buffer. + * @return {Buffer} a new Buffer. + */ + allocBuffer: function (size) { + if (Buffer.alloc) { + return Buffer.alloc(size); + } else { + var buf = new Buffer(size); + buf.fill(0); + return buf; + } + }, + /** + * Find out if an object is a Buffer. + * @param {Object} b the object to test. + * @return {Boolean} true if the object is a Buffer, false otherwise. + */ + isBuffer : function(b){ + return Buffer.isBuffer(b); + }, + + isStream : function (obj) { + return obj && + typeof obj.on === "function" && + typeof obj.pause === "function" && + typeof obj.resume === "function"; + } +}; + +},{}],15:[function(require,module,exports){ +'use strict'; +var utf8 = require('./utf8'); +var utils = require('./utils'); +var GenericWorker = require('./stream/GenericWorker'); +var StreamHelper = require('./stream/StreamHelper'); +var defaults = require('./defaults'); +var CompressedObject = require('./compressedObject'); +var ZipObject = require('./zipObject'); +var generate = require("./generate"); +var nodejsUtils = require("./nodejsUtils"); +var NodejsStreamInputAdapter = require("./nodejs/NodejsStreamInputAdapter"); + + +/** + * Add a file in the current folder. + * @private + * @param {string} name the name of the file + * @param {String|ArrayBuffer|Uint8Array|Buffer} data the data of the file + * @param {Object} originalOptions the options of the file + * @return {Object} the new file. + */ +var fileAdd = function(name, data, originalOptions) { + // be sure sub folders exist + var dataType = utils.getTypeOf(data), + parent; + + + /* + * Correct options. + */ + + var o = utils.extend(originalOptions || {}, defaults); + o.date = o.date || new Date(); + if (o.compression !== null) { + o.compression = o.compression.toUpperCase(); + } + + if (typeof o.unixPermissions === "string") { + o.unixPermissions = parseInt(o.unixPermissions, 8); + } + + // UNX_IFDIR 0040000 see zipinfo.c + if (o.unixPermissions && (o.unixPermissions & 0x4000)) { + o.dir = true; + } + // Bit 4 Directory + if (o.dosPermissions && (o.dosPermissions & 0x0010)) { + o.dir = true; + } + + if (o.dir) { + name = forceTrailingSlash(name); + } + if (o.createFolders && (parent = parentFolder(name))) { + folderAdd.call(this, parent, true); + } + + var isUnicodeString = dataType === "string" && o.binary === false && o.base64 === false; + if (!originalOptions || typeof originalOptions.binary === "undefined") { + o.binary = !isUnicodeString; + } + + + var isCompressedEmpty = (data instanceof CompressedObject) && data.uncompressedSize === 0; + + if (isCompressedEmpty || o.dir || !data || data.length === 0) { + o.base64 = false; + o.binary = true; + data = ""; + o.compression = "STORE"; + dataType = "string"; + } + + /* + * Convert content to fit. + */ + + var zipObjectContent = null; + if (data instanceof CompressedObject || data instanceof GenericWorker) { + zipObjectContent = data; + } else if (nodejsUtils.isNode && nodejsUtils.isStream(data)) { + zipObjectContent = new NodejsStreamInputAdapter(name, data); + } else { + zipObjectContent = utils.prepareContent(name, data, o.binary, o.optimizedBinaryString, o.base64); + } + + var object = new ZipObject(name, zipObjectContent, o); + this.files[name] = object; + /* + TODO: we can't throw an exception because we have async promises + (we can have a promise of a Date() for example) but returning a + promise is useless because file(name, data) returns the JSZip + object for chaining. Should we break that to allow the user + to catch the error ? + + return external.Promise.resolve(zipObjectContent) + .then(function () { + return object; + }); + */ +}; + +/** + * Find the parent folder of the path. + * @private + * @param {string} path the path to use + * @return {string} the parent folder, or "" + */ +var parentFolder = function (path) { + if (path.slice(-1) === '/') { + path = path.substring(0, path.length - 1); + } + var lastSlash = path.lastIndexOf('/'); + return (lastSlash > 0) ? path.substring(0, lastSlash) : ""; +}; + +/** + * Returns the path with a slash at the end. + * @private + * @param {String} path the path to check. + * @return {String} the path with a trailing slash. + */ +var forceTrailingSlash = function(path) { + // Check the name ends with a / + if (path.slice(-1) !== "/") { + path += "/"; // IE doesn't like substr(-1) + } + return path; +}; + +/** + * Add a (sub) folder in the current folder. + * @private + * @param {string} name the folder's name + * @param {boolean=} [createFolders] If true, automatically create sub + * folders. Defaults to false. + * @return {Object} the new folder. + */ +var folderAdd = function(name, createFolders) { + createFolders = (typeof createFolders !== 'undefined') ? createFolders : defaults.createFolders; + + name = forceTrailingSlash(name); + + // Does this folder already exist? + if (!this.files[name]) { + fileAdd.call(this, name, null, { + dir: true, + createFolders: createFolders + }); + } + return this.files[name]; +}; + +/** +* Cross-window, cross-Node-context regular expression detection +* @param {Object} object Anything +* @return {Boolean} true if the object is a regular expression, +* false otherwise +*/ +function isRegExp(object) { + return Object.prototype.toString.call(object) === "[object RegExp]"; +} + +// return the actual prototype of JSZip +var out = { + /** + * @see loadAsync + */ + load: function() { + throw new Error("This method has been removed in JSZip 3.0, please check the upgrade guide."); + }, + + + /** + * Call a callback function for each entry at this folder level. + * @param {Function} cb the callback function: + * function (relativePath, file) {...} + * It takes 2 arguments : the relative path and the file. + */ + forEach: function(cb) { + var filename, relativePath, file; + /* jshint ignore:start */ + // ignore warning about unwanted properties because this.files is a null prototype object + for (filename in this.files) { + file = this.files[filename]; + relativePath = filename.slice(this.root.length, filename.length); + if (relativePath && filename.slice(0, this.root.length) === this.root) { // the file is in the current root + cb(relativePath, file); // TODO reverse the parameters ? need to be clean AND consistent with the filter search fn... + } + } + /* jshint ignore:end */ + }, + + /** + * Filter nested files/folders with the specified function. + * @param {Function} search the predicate to use : + * function (relativePath, file) {...} + * It takes 2 arguments : the relative path and the file. + * @return {Array} An array of matching elements. + */ + filter: function(search) { + var result = []; + this.forEach(function (relativePath, entry) { + if (search(relativePath, entry)) { // the file matches the function + result.push(entry); + } + + }); + return result; + }, + + /** + * Add a file to the zip file, or search a file. + * @param {string|RegExp} name The name of the file to add (if data is defined), + * the name of the file to find (if no data) or a regex to match files. + * @param {String|ArrayBuffer|Uint8Array|Buffer} data The file data, either raw or base64 encoded + * @param {Object} o File options + * @return {JSZip|Object|Array} this JSZip object (when adding a file), + * a file (when searching by string) or an array of files (when searching by regex). + */ + file: function(name, data, o) { + if (arguments.length === 1) { + if (isRegExp(name)) { + var regexp = name; + return this.filter(function(relativePath, file) { + return !file.dir && regexp.test(relativePath); + }); + } + else { // text + var obj = this.files[this.root + name]; + if (obj && !obj.dir) { + return obj; + } else { + return null; + } + } + } + else { // more than one argument : we have data ! + name = this.root + name; + fileAdd.call(this, name, data, o); + } + return this; + }, + + /** + * Add a directory to the zip file, or search. + * @param {String|RegExp} arg The name of the directory to add, or a regex to search folders. + * @return {JSZip} an object with the new directory as the root, or an array containing matching folders. + */ + folder: function(arg) { + if (!arg) { + return this; + } + + if (isRegExp(arg)) { + return this.filter(function(relativePath, file) { + return file.dir && arg.test(relativePath); + }); + } + + // else, name is a new folder + var name = this.root + arg; + var newFolder = folderAdd.call(this, name); + + // Allow chaining by returning a new object with this folder as the root + var ret = this.clone(); + ret.root = newFolder.name; + return ret; + }, + + /** + * Delete a file, or a directory and all sub-files, from the zip + * @param {string} name the name of the file to delete + * @return {JSZip} this JSZip object + */ + remove: function(name) { + name = this.root + name; + var file = this.files[name]; + if (!file) { + // Look for any folders + if (name.slice(-1) !== "/") { + name += "/"; + } + file = this.files[name]; + } + + if (file && !file.dir) { + // file + delete this.files[name]; + } else { + // maybe a folder, delete recursively + var kids = this.filter(function(relativePath, file) { + return file.name.slice(0, name.length) === name; + }); + for (var i = 0; i < kids.length; i++) { + delete this.files[kids[i].name]; + } + } + + return this; + }, + + /** + * Generate the complete zip file + * @param {Object} options the options to generate the zip file : + * - compression, "STORE" by default. + * - type, "base64" by default. Values are : string, base64, uint8array, arraybuffer, blob. + * @return {String|Uint8Array|ArrayBuffer|Buffer|Blob} the zip file + */ + generate: function(options) { + throw new Error("This method has been removed in JSZip 3.0, please check the upgrade guide."); + }, + + /** + * Generate the complete zip file as an internal stream. + * @param {Object} options the options to generate the zip file : + * - compression, "STORE" by default. + * - type, "base64" by default. Values are : string, base64, uint8array, arraybuffer, blob. + * @return {StreamHelper} the streamed zip file. + */ + generateInternalStream: function(options) { + var worker, opts = {}; + try { + opts = utils.extend(options || {}, { + streamFiles: false, + compression: "STORE", + compressionOptions : null, + type: "", + platform: "DOS", + comment: null, + mimeType: 'application/zip', + encodeFileName: utf8.utf8encode + }); + + opts.type = opts.type.toLowerCase(); + opts.compression = opts.compression.toUpperCase(); + + // "binarystring" is preferred but the internals use "string". + if(opts.type === "binarystring") { + opts.type = "string"; + } + + if (!opts.type) { + throw new Error("No output type specified."); + } + + utils.checkSupport(opts.type); + + // accept nodejs `process.platform` + if( + opts.platform === 'darwin' || + opts.platform === 'freebsd' || + opts.platform === 'linux' || + opts.platform === 'sunos' + ) { + opts.platform = "UNIX"; + } + if (opts.platform === 'win32') { + opts.platform = "DOS"; + } + + var comment = opts.comment || this.comment || ""; + worker = generate.generateWorker(this, opts, comment); + } catch (e) { + worker = new GenericWorker("error"); + worker.error(e); + } + return new StreamHelper(worker, opts.type || "string", opts.mimeType); + }, + /** + * Generate the complete zip file asynchronously. + * @see generateInternalStream + */ + generateAsync: function(options, onUpdate) { + return this.generateInternalStream(options).accumulate(onUpdate); + }, + /** + * Generate the complete zip file asynchronously. + * @see generateInternalStream + */ + generateNodeStream: function(options, onUpdate) { + options = options || {}; + if (!options.type) { + options.type = "nodebuffer"; + } + return this.generateInternalStream(options).toNodejsStream(onUpdate); + } +}; +module.exports = out; + +},{"./compressedObject":2,"./defaults":5,"./generate":9,"./nodejs/NodejsStreamInputAdapter":12,"./nodejsUtils":14,"./stream/GenericWorker":28,"./stream/StreamHelper":29,"./utf8":31,"./utils":32,"./zipObject":35}],16:[function(require,module,exports){ +/* + * This file is used by module bundlers (browserify/webpack/etc) when + * including a stream implementation. We use "readable-stream" to get a + * consistent behavior between nodejs versions but bundlers often have a shim + * for "stream". Using this shim greatly improve the compatibility and greatly + * reduce the final size of the bundle (only one stream implementation, not + * two). + */ +module.exports = require("stream"); + +},{"stream":undefined}],17:[function(require,module,exports){ +'use strict'; +var DataReader = require('./DataReader'); +var utils = require('../utils'); + +function ArrayReader(data) { + DataReader.call(this, data); + for(var i = 0; i < this.data.length; i++) { + data[i] = data[i] & 0xFF; + } +} +utils.inherits(ArrayReader, DataReader); +/** + * @see DataReader.byteAt + */ +ArrayReader.prototype.byteAt = function(i) { + return this.data[this.zero + i]; +}; +/** + * @see DataReader.lastIndexOfSignature + */ +ArrayReader.prototype.lastIndexOfSignature = function(sig) { + var sig0 = sig.charCodeAt(0), + sig1 = sig.charCodeAt(1), + sig2 = sig.charCodeAt(2), + sig3 = sig.charCodeAt(3); + for (var i = this.length - 4; i >= 0; --i) { + if (this.data[i] === sig0 && this.data[i + 1] === sig1 && this.data[i + 2] === sig2 && this.data[i + 3] === sig3) { + return i - this.zero; + } + } + + return -1; +}; +/** + * @see DataReader.readAndCheckSignature + */ +ArrayReader.prototype.readAndCheckSignature = function (sig) { + var sig0 = sig.charCodeAt(0), + sig1 = sig.charCodeAt(1), + sig2 = sig.charCodeAt(2), + sig3 = sig.charCodeAt(3), + data = this.readData(4); + return sig0 === data[0] && sig1 === data[1] && sig2 === data[2] && sig3 === data[3]; +}; +/** + * @see DataReader.readData + */ +ArrayReader.prototype.readData = function(size) { + this.checkOffset(size); + if(size === 0) { + return []; + } + var result = this.data.slice(this.zero + this.index, this.zero + this.index + size); + this.index += size; + return result; +}; +module.exports = ArrayReader; + +},{"../utils":32,"./DataReader":18}],18:[function(require,module,exports){ +'use strict'; +var utils = require('../utils'); + +function DataReader(data) { + this.data = data; // type : see implementation + this.length = data.length; + this.index = 0; + this.zero = 0; +} +DataReader.prototype = { + /** + * Check that the offset will not go too far. + * @param {string} offset the additional offset to check. + * @throws {Error} an Error if the offset is out of bounds. + */ + checkOffset: function(offset) { + this.checkIndex(this.index + offset); + }, + /** + * Check that the specified index will not be too far. + * @param {string} newIndex the index to check. + * @throws {Error} an Error if the index is out of bounds. + */ + checkIndex: function(newIndex) { + if (this.length < this.zero + newIndex || newIndex < 0) { + throw new Error("End of data reached (data length = " + this.length + ", asked index = " + (newIndex) + "). Corrupted zip ?"); + } + }, + /** + * Change the index. + * @param {number} newIndex The new index. + * @throws {Error} if the new index is out of the data. + */ + setIndex: function(newIndex) { + this.checkIndex(newIndex); + this.index = newIndex; + }, + /** + * Skip the next n bytes. + * @param {number} n the number of bytes to skip. + * @throws {Error} if the new index is out of the data. + */ + skip: function(n) { + this.setIndex(this.index + n); + }, + /** + * Get the byte at the specified index. + * @param {number} i the index to use. + * @return {number} a byte. + */ + byteAt: function(i) { + // see implementations + }, + /** + * Get the next number with a given byte size. + * @param {number} size the number of bytes to read. + * @return {number} the corresponding number. + */ + readInt: function(size) { + var result = 0, + i; + this.checkOffset(size); + for (i = this.index + size - 1; i >= this.index; i--) { + result = (result << 8) + this.byteAt(i); + } + this.index += size; + return result; + }, + /** + * Get the next string with a given byte size. + * @param {number} size the number of bytes to read. + * @return {string} the corresponding string. + */ + readString: function(size) { + return utils.transformTo("string", this.readData(size)); + }, + /** + * Get raw data without conversion, bytes. + * @param {number} size the number of bytes to read. + * @return {Object} the raw data, implementation specific. + */ + readData: function(size) { + // see implementations + }, + /** + * Find the last occurrence of a zip signature (4 bytes). + * @param {string} sig the signature to find. + * @return {number} the index of the last occurrence, -1 if not found. + */ + lastIndexOfSignature: function(sig) { + // see implementations + }, + /** + * Read the signature (4 bytes) at the current position and compare it with sig. + * @param {string} sig the expected signature + * @return {boolean} true if the signature matches, false otherwise. + */ + readAndCheckSignature: function(sig) { + // see implementations + }, + /** + * Get the next date. + * @return {Date} the date. + */ + readDate: function() { + var dostime = this.readInt(4); + return new Date(Date.UTC( + ((dostime >> 25) & 0x7f) + 1980, // year + ((dostime >> 21) & 0x0f) - 1, // month + (dostime >> 16) & 0x1f, // day + (dostime >> 11) & 0x1f, // hour + (dostime >> 5) & 0x3f, // minute + (dostime & 0x1f) << 1)); // second + } +}; +module.exports = DataReader; + +},{"../utils":32}],19:[function(require,module,exports){ +'use strict'; +var Uint8ArrayReader = require('./Uint8ArrayReader'); +var utils = require('../utils'); + +function NodeBufferReader(data) { + Uint8ArrayReader.call(this, data); +} +utils.inherits(NodeBufferReader, Uint8ArrayReader); + +/** + * @see DataReader.readData + */ +NodeBufferReader.prototype.readData = function(size) { + this.checkOffset(size); + var result = this.data.slice(this.zero + this.index, this.zero + this.index + size); + this.index += size; + return result; +}; +module.exports = NodeBufferReader; + +},{"../utils":32,"./Uint8ArrayReader":21}],20:[function(require,module,exports){ +'use strict'; +var DataReader = require('./DataReader'); +var utils = require('../utils'); + +function StringReader(data) { + DataReader.call(this, data); +} +utils.inherits(StringReader, DataReader); +/** + * @see DataReader.byteAt + */ +StringReader.prototype.byteAt = function(i) { + return this.data.charCodeAt(this.zero + i); +}; +/** + * @see DataReader.lastIndexOfSignature + */ +StringReader.prototype.lastIndexOfSignature = function(sig) { + return this.data.lastIndexOf(sig) - this.zero; +}; +/** + * @see DataReader.readAndCheckSignature + */ +StringReader.prototype.readAndCheckSignature = function (sig) { + var data = this.readData(4); + return sig === data; +}; +/** + * @see DataReader.readData + */ +StringReader.prototype.readData = function(size) { + this.checkOffset(size); + // this will work because the constructor applied the "& 0xff" mask. + var result = this.data.slice(this.zero + this.index, this.zero + this.index + size); + this.index += size; + return result; +}; +module.exports = StringReader; + +},{"../utils":32,"./DataReader":18}],21:[function(require,module,exports){ +'use strict'; +var ArrayReader = require('./ArrayReader'); +var utils = require('../utils'); + +function Uint8ArrayReader(data) { + ArrayReader.call(this, data); +} +utils.inherits(Uint8ArrayReader, ArrayReader); +/** + * @see DataReader.readData + */ +Uint8ArrayReader.prototype.readData = function(size) { + this.checkOffset(size); + if(size === 0) { + // in IE10, when using subarray(idx, idx), we get the array [0x00] instead of []. + return new Uint8Array(0); + } + var result = this.data.subarray(this.zero + this.index, this.zero + this.index + size); + this.index += size; + return result; +}; +module.exports = Uint8ArrayReader; + +},{"../utils":32,"./ArrayReader":17}],22:[function(require,module,exports){ +'use strict'; + +var utils = require('../utils'); +var support = require('../support'); +var ArrayReader = require('./ArrayReader'); +var StringReader = require('./StringReader'); +var NodeBufferReader = require('./NodeBufferReader'); +var Uint8ArrayReader = require('./Uint8ArrayReader'); + +/** + * Create a reader adapted to the data. + * @param {String|ArrayBuffer|Uint8Array|Buffer} data the data to read. + * @return {DataReader} the data reader. + */ +module.exports = function (data) { + var type = utils.getTypeOf(data); + utils.checkSupport(type); + if (type === "string" && !support.uint8array) { + return new StringReader(data); + } + if (type === "nodebuffer") { + return new NodeBufferReader(data); + } + if (support.uint8array) { + return new Uint8ArrayReader(utils.transformTo("uint8array", data)); + } + return new ArrayReader(utils.transformTo("array", data)); +}; + +},{"../support":30,"../utils":32,"./ArrayReader":17,"./NodeBufferReader":19,"./StringReader":20,"./Uint8ArrayReader":21}],23:[function(require,module,exports){ +'use strict'; +exports.LOCAL_FILE_HEADER = "PK\x03\x04"; +exports.CENTRAL_FILE_HEADER = "PK\x01\x02"; +exports.CENTRAL_DIRECTORY_END = "PK\x05\x06"; +exports.ZIP64_CENTRAL_DIRECTORY_LOCATOR = "PK\x06\x07"; +exports.ZIP64_CENTRAL_DIRECTORY_END = "PK\x06\x06"; +exports.DATA_DESCRIPTOR = "PK\x07\x08"; + +},{}],24:[function(require,module,exports){ +'use strict'; + +var GenericWorker = require('./GenericWorker'); +var utils = require('../utils'); + +/** + * A worker which convert chunks to a specified type. + * @constructor + * @param {String} destType the destination type. + */ +function ConvertWorker(destType) { + GenericWorker.call(this, "ConvertWorker to " + destType); + this.destType = destType; +} +utils.inherits(ConvertWorker, GenericWorker); + +/** + * @see GenericWorker.processChunk + */ +ConvertWorker.prototype.processChunk = function (chunk) { + this.push({ + data : utils.transformTo(this.destType, chunk.data), + meta : chunk.meta + }); +}; +module.exports = ConvertWorker; + +},{"../utils":32,"./GenericWorker":28}],25:[function(require,module,exports){ +'use strict'; + +var GenericWorker = require('./GenericWorker'); +var crc32 = require('../crc32'); +var utils = require('../utils'); + +/** + * A worker which calculate the crc32 of the data flowing through. + * @constructor + */ +function Crc32Probe() { + GenericWorker.call(this, "Crc32Probe"); + this.withStreamInfo("crc32", 0); +} +utils.inherits(Crc32Probe, GenericWorker); + +/** + * @see GenericWorker.processChunk + */ +Crc32Probe.prototype.processChunk = function (chunk) { + this.streamInfo.crc32 = crc32(chunk.data, this.streamInfo.crc32 || 0); + this.push(chunk); +}; +module.exports = Crc32Probe; + +},{"../crc32":4,"../utils":32,"./GenericWorker":28}],26:[function(require,module,exports){ +'use strict'; + +var utils = require('../utils'); +var GenericWorker = require('./GenericWorker'); + +/** + * A worker which calculate the total length of the data flowing through. + * @constructor + * @param {String} propName the name used to expose the length + */ +function DataLengthProbe(propName) { + GenericWorker.call(this, "DataLengthProbe for " + propName); + this.propName = propName; + this.withStreamInfo(propName, 0); +} +utils.inherits(DataLengthProbe, GenericWorker); + +/** + * @see GenericWorker.processChunk + */ +DataLengthProbe.prototype.processChunk = function (chunk) { + if(chunk) { + var length = this.streamInfo[this.propName] || 0; + this.streamInfo[this.propName] = length + chunk.data.length; + } + GenericWorker.prototype.processChunk.call(this, chunk); +}; +module.exports = DataLengthProbe; + + +},{"../utils":32,"./GenericWorker":28}],27:[function(require,module,exports){ +'use strict'; + +var utils = require('../utils'); +var GenericWorker = require('./GenericWorker'); + +// the size of the generated chunks +// TODO expose this as a public variable +var DEFAULT_BLOCK_SIZE = 16 * 1024; + +/** + * A worker that reads a content and emits chunks. + * @constructor + * @param {Promise} dataP the promise of the data to split + */ +function DataWorker(dataP) { + GenericWorker.call(this, "DataWorker"); + var self = this; + this.dataIsReady = false; + this.index = 0; + this.max = 0; + this.data = null; + this.type = ""; + + this._tickScheduled = false; + + dataP.then(function (data) { + self.dataIsReady = true; + self.data = data; + self.max = data && data.length || 0; + self.type = utils.getTypeOf(data); + if(!self.isPaused) { + self._tickAndRepeat(); + } + }, function (e) { + self.error(e); + }); +} + +utils.inherits(DataWorker, GenericWorker); + +/** + * @see GenericWorker.cleanUp + */ +DataWorker.prototype.cleanUp = function () { + GenericWorker.prototype.cleanUp.call(this); + this.data = null; +}; + +/** + * @see GenericWorker.resume + */ +DataWorker.prototype.resume = function () { + if(!GenericWorker.prototype.resume.call(this)) { + return false; + } + + if (!this._tickScheduled && this.dataIsReady) { + this._tickScheduled = true; + utils.delay(this._tickAndRepeat, [], this); + } + return true; +}; + +/** + * Trigger a tick a schedule an other call to this function. + */ +DataWorker.prototype._tickAndRepeat = function() { + this._tickScheduled = false; + if(this.isPaused || this.isFinished) { + return; + } + this._tick(); + if(!this.isFinished) { + utils.delay(this._tickAndRepeat, [], this); + this._tickScheduled = true; + } +}; + +/** + * Read and push a chunk. + */ +DataWorker.prototype._tick = function() { + + if(this.isPaused || this.isFinished) { + return false; + } + + var size = DEFAULT_BLOCK_SIZE; + var data = null, nextIndex = Math.min(this.max, this.index + size); + if (this.index >= this.max) { + // EOF + return this.end(); + } else { + switch(this.type) { + case "string": + data = this.data.substring(this.index, nextIndex); + break; + case "uint8array": + data = this.data.subarray(this.index, nextIndex); + break; + case "array": + case "nodebuffer": + data = this.data.slice(this.index, nextIndex); + break; + } + this.index = nextIndex; + return this.push({ + data : data, + meta : { + percent : this.max ? this.index / this.max * 100 : 0 + } + }); + } +}; + +module.exports = DataWorker; + +},{"../utils":32,"./GenericWorker":28}],28:[function(require,module,exports){ +'use strict'; + +/** + * A worker that does nothing but passing chunks to the next one. This is like + * a nodejs stream but with some differences. On the good side : + * - it works on IE 6-9 without any issue / polyfill + * - it weights less than the full dependencies bundled with browserify + * - it forwards errors (no need to declare an error handler EVERYWHERE) + * + * A chunk is an object with 2 attributes : `meta` and `data`. The former is an + * object containing anything (`percent` for example), see each worker for more + * details. The latter is the real data (String, Uint8Array, etc). + * + * @constructor + * @param {String} name the name of the stream (mainly used for debugging purposes) + */ +function GenericWorker(name) { + // the name of the worker + this.name = name || "default"; + // an object containing metadata about the workers chain + this.streamInfo = {}; + // an error which happened when the worker was paused + this.generatedError = null; + // an object containing metadata to be merged by this worker into the general metadata + this.extraStreamInfo = {}; + // true if the stream is paused (and should not do anything), false otherwise + this.isPaused = true; + // true if the stream is finished (and should not do anything), false otherwise + this.isFinished = false; + // true if the stream is locked to prevent further structure updates (pipe), false otherwise + this.isLocked = false; + // the event listeners + this._listeners = { + 'data':[], + 'end':[], + 'error':[] + }; + // the previous worker, if any + this.previous = null; +} + +GenericWorker.prototype = { + /** + * Push a chunk to the next workers. + * @param {Object} chunk the chunk to push + */ + push : function (chunk) { + this.emit("data", chunk); + }, + /** + * End the stream. + * @return {Boolean} true if this call ended the worker, false otherwise. + */ + end : function () { + if (this.isFinished) { + return false; + } + + this.flush(); + try { + this.emit("end"); + this.cleanUp(); + this.isFinished = true; + } catch (e) { + this.emit("error", e); + } + return true; + }, + /** + * End the stream with an error. + * @param {Error} e the error which caused the premature end. + * @return {Boolean} true if this call ended the worker with an error, false otherwise. + */ + error : function (e) { + if (this.isFinished) { + return false; + } + + if(this.isPaused) { + this.generatedError = e; + } else { + this.isFinished = true; + + this.emit("error", e); + + // in the workers chain exploded in the middle of the chain, + // the error event will go downward but we also need to notify + // workers upward that there has been an error. + if(this.previous) { + this.previous.error(e); + } + + this.cleanUp(); + } + return true; + }, + /** + * Add a callback on an event. + * @param {String} name the name of the event (data, end, error) + * @param {Function} listener the function to call when the event is triggered + * @return {GenericWorker} the current object for chainability + */ + on : function (name, listener) { + this._listeners[name].push(listener); + return this; + }, + /** + * Clean any references when a worker is ending. + */ + cleanUp : function () { + this.streamInfo = this.generatedError = this.extraStreamInfo = null; + this._listeners = []; + }, + /** + * Trigger an event. This will call registered callback with the provided arg. + * @param {String} name the name of the event (data, end, error) + * @param {Object} arg the argument to call the callback with. + */ + emit : function (name, arg) { + if (this._listeners[name]) { + for(var i = 0; i < this._listeners[name].length; i++) { + this._listeners[name][i].call(this, arg); + } + } + }, + /** + * Chain a worker with an other. + * @param {Worker} next the worker receiving events from the current one. + * @return {worker} the next worker for chainability + */ + pipe : function (next) { + return next.registerPrevious(this); + }, + /** + * Same as `pipe` in the other direction. + * Using an API with `pipe(next)` is very easy. + * Implementing the API with the point of view of the next one registering + * a source is easier, see the ZipFileWorker. + * @param {Worker} previous the previous worker, sending events to this one + * @return {Worker} the current worker for chainability + */ + registerPrevious : function (previous) { + if (this.isLocked) { + throw new Error("The stream '" + this + "' has already been used."); + } + + // sharing the streamInfo... + this.streamInfo = previous.streamInfo; + // ... and adding our own bits + this.mergeStreamInfo(); + this.previous = previous; + var self = this; + previous.on('data', function (chunk) { + self.processChunk(chunk); + }); + previous.on('end', function () { + self.end(); + }); + previous.on('error', function (e) { + self.error(e); + }); + return this; + }, + /** + * Pause the stream so it doesn't send events anymore. + * @return {Boolean} true if this call paused the worker, false otherwise. + */ + pause : function () { + if(this.isPaused || this.isFinished) { + return false; + } + this.isPaused = true; + + if(this.previous) { + this.previous.pause(); + } + return true; + }, + /** + * Resume a paused stream. + * @return {Boolean} true if this call resumed the worker, false otherwise. + */ + resume : function () { + if(!this.isPaused || this.isFinished) { + return false; + } + this.isPaused = false; + + // if true, the worker tried to resume but failed + var withError = false; + if(this.generatedError) { + this.error(this.generatedError); + withError = true; + } + if(this.previous) { + this.previous.resume(); + } + + return !withError; + }, + /** + * Flush any remaining bytes as the stream is ending. + */ + flush : function () {}, + /** + * Process a chunk. This is usually the method overridden. + * @param {Object} chunk the chunk to process. + */ + processChunk : function(chunk) { + this.push(chunk); + }, + /** + * Add a key/value to be added in the workers chain streamInfo once activated. + * @param {String} key the key to use + * @param {Object} value the associated value + * @return {Worker} the current worker for chainability + */ + withStreamInfo : function (key, value) { + this.extraStreamInfo[key] = value; + this.mergeStreamInfo(); + return this; + }, + /** + * Merge this worker's streamInfo into the chain's streamInfo. + */ + mergeStreamInfo : function () { + for(var key in this.extraStreamInfo) { + if (!this.extraStreamInfo.hasOwnProperty(key)) { + continue; + } + this.streamInfo[key] = this.extraStreamInfo[key]; + } + }, + + /** + * Lock the stream to prevent further updates on the workers chain. + * After calling this method, all calls to pipe will fail. + */ + lock: function () { + if (this.isLocked) { + throw new Error("The stream '" + this + "' has already been used."); + } + this.isLocked = true; + if (this.previous) { + this.previous.lock(); + } + }, + + /** + * + * Pretty print the workers chain. + */ + toString : function () { + var me = "Worker " + this.name; + if (this.previous) { + return this.previous + " -> " + me; + } else { + return me; + } + } +}; + +module.exports = GenericWorker; + +},{}],29:[function(require,module,exports){ +'use strict'; + +var utils = require('../utils'); +var ConvertWorker = require('./ConvertWorker'); +var GenericWorker = require('./GenericWorker'); +var base64 = require('../base64'); +var support = require("../support"); +var external = require("../external"); + +var NodejsStreamOutputAdapter = null; +if (support.nodestream) { + try { + NodejsStreamOutputAdapter = require('../nodejs/NodejsStreamOutputAdapter'); + } catch(e) {} +} + +/** + * Apply the final transformation of the data. If the user wants a Blob for + * example, it's easier to work with an U8intArray and finally do the + * ArrayBuffer/Blob conversion. + * @param {String} type the name of the final type + * @param {String|Uint8Array|Buffer} content the content to transform + * @param {String} mimeType the mime type of the content, if applicable. + * @return {String|Uint8Array|ArrayBuffer|Buffer|Blob} the content in the right format. + */ +function transformZipOutput(type, content, mimeType) { + switch(type) { + case "blob" : + return utils.newBlob(utils.transformTo("arraybuffer", content), mimeType); + case "base64" : + return base64.encode(content); + default : + return utils.transformTo(type, content); + } +} + +/** + * Concatenate an array of data of the given type. + * @param {String} type the type of the data in the given array. + * @param {Array} dataArray the array containing the data chunks to concatenate + * @return {String|Uint8Array|Buffer} the concatenated data + * @throws Error if the asked type is unsupported + */ +function concat (type, dataArray) { + var i, index = 0, res = null, totalLength = 0; + for(i = 0; i < dataArray.length; i++) { + totalLength += dataArray[i].length; + } + switch(type) { + case "string": + return dataArray.join(""); + case "array": + return Array.prototype.concat.apply([], dataArray); + case "uint8array": + res = new Uint8Array(totalLength); + for(i = 0; i < dataArray.length; i++) { + res.set(dataArray[i], index); + index += dataArray[i].length; + } + return res; + case "nodebuffer": + return Buffer.concat(dataArray); + default: + throw new Error("concat : unsupported type '" + type + "'"); + } +} + +/** + * Listen a StreamHelper, accumulate its content and concatenate it into a + * complete block. + * @param {StreamHelper} helper the helper to use. + * @param {Function} updateCallback a callback called on each update. Called + * with one arg : + * - the metadata linked to the update received. + * @return Promise the promise for the accumulation. + */ +function accumulate(helper, updateCallback) { + return new external.Promise(function (resolve, reject){ + var dataArray = []; + var chunkType = helper._internalType, + resultType = helper._outputType, + mimeType = helper._mimeType; + helper + .on('data', function (data, meta) { + dataArray.push(data); + if(updateCallback) { + updateCallback(meta); + } + }) + .on('error', function(err) { + dataArray = []; + reject(err); + }) + .on('end', function (){ + try { + var result = transformZipOutput(resultType, concat(chunkType, dataArray), mimeType); + resolve(result); + } catch (e) { + reject(e); + } + dataArray = []; + }) + .resume(); + }); +} + +/** + * An helper to easily use workers outside of JSZip. + * @constructor + * @param {Worker} worker the worker to wrap + * @param {String} outputType the type of data expected by the use + * @param {String} mimeType the mime type of the content, if applicable. + */ +function StreamHelper(worker, outputType, mimeType) { + var internalType = outputType; + switch(outputType) { + case "blob": + case "arraybuffer": + internalType = "uint8array"; + break; + case "base64": + internalType = "string"; + break; + } + + try { + // the type used internally + this._internalType = internalType; + // the type used to output results + this._outputType = outputType; + // the mime type + this._mimeType = mimeType; + utils.checkSupport(internalType); + this._worker = worker.pipe(new ConvertWorker(internalType)); + // the last workers can be rewired without issues but we need to + // prevent any updates on previous workers. + worker.lock(); + } catch(e) { + this._worker = new GenericWorker("error"); + this._worker.error(e); + } +} + +StreamHelper.prototype = { + /** + * Listen a StreamHelper, accumulate its content and concatenate it into a + * complete block. + * @param {Function} updateCb the update callback. + * @return Promise the promise for the accumulation. + */ + accumulate : function (updateCb) { + return accumulate(this, updateCb); + }, + /** + * Add a listener on an event triggered on a stream. + * @param {String} evt the name of the event + * @param {Function} fn the listener + * @return {StreamHelper} the current helper. + */ + on : function (evt, fn) { + var self = this; + + if(evt === "data") { + this._worker.on(evt, function (chunk) { + fn.call(self, chunk.data, chunk.meta); + }); + } else { + this._worker.on(evt, function () { + utils.delay(fn, arguments, self); + }); + } + return this; + }, + /** + * Resume the flow of chunks. + * @return {StreamHelper} the current helper. + */ + resume : function () { + utils.delay(this._worker.resume, [], this._worker); + return this; + }, + /** + * Pause the flow of chunks. + * @return {StreamHelper} the current helper. + */ + pause : function () { + this._worker.pause(); + return this; + }, + /** + * Return a nodejs stream for this helper. + * @param {Function} updateCb the update callback. + * @return {NodejsStreamOutputAdapter} the nodejs stream. + */ + toNodejsStream : function (updateCb) { + utils.checkSupport("nodestream"); + if (this._outputType !== "nodebuffer") { + // an object stream containing blob/arraybuffer/uint8array/string + // is strange and I don't know if it would be useful. + // I you find this comment and have a good usecase, please open a + // bug report ! + throw new Error(this._outputType + " is not supported by this method"); + } + + return new NodejsStreamOutputAdapter(this, { + objectMode : this._outputType !== "nodebuffer" + }, updateCb); + } +}; + + +module.exports = StreamHelper; + +},{"../base64":1,"../external":6,"../nodejs/NodejsStreamOutputAdapter":13,"../support":30,"../utils":32,"./ConvertWorker":24,"./GenericWorker":28}],30:[function(require,module,exports){ +'use strict'; + +exports.base64 = true; +exports.array = true; +exports.string = true; +exports.arraybuffer = typeof ArrayBuffer !== "undefined" && typeof Uint8Array !== "undefined"; +exports.nodebuffer = typeof Buffer !== "undefined"; +// contains true if JSZip can read/generate Uint8Array, false otherwise. +exports.uint8array = typeof Uint8Array !== "undefined"; + +if (typeof ArrayBuffer === "undefined") { + exports.blob = false; +} +else { + var buffer = new ArrayBuffer(0); + try { + exports.blob = new Blob([buffer], { + type: "application/zip" + }).size === 0; + } + catch (e) { + try { + var Builder = self.BlobBuilder || self.WebKitBlobBuilder || self.MozBlobBuilder || self.MSBlobBuilder; + var builder = new Builder(); + builder.append(buffer); + exports.blob = builder.getBlob('application/zip').size === 0; + } + catch (e) { + exports.blob = false; + } + } +} + +try { + exports.nodestream = !!require('readable-stream').Readable; +} catch(e) { + exports.nodestream = false; +} + +},{"readable-stream":16}],31:[function(require,module,exports){ +'use strict'; + +var utils = require('./utils'); +var support = require('./support'); +var nodejsUtils = require('./nodejsUtils'); +var GenericWorker = require('./stream/GenericWorker'); + +/** + * The following functions come from pako, from pako/lib/utils/strings + * released under the MIT license, see pako https://github.com/nodeca/pako/ + */ + +// Table with utf8 lengths (calculated by first byte of sequence) +// Note, that 5 & 6-byte values and some 4-byte values can not be represented in JS, +// because max possible codepoint is 0x10ffff +var _utf8len = new Array(256); +for (var i=0; i<256; i++) { + _utf8len[i] = (i >= 252 ? 6 : i >= 248 ? 5 : i >= 240 ? 4 : i >= 224 ? 3 : i >= 192 ? 2 : 1); +} +_utf8len[254]=_utf8len[254]=1; // Invalid sequence start + +// convert string to array (typed, when possible) +var string2buf = function (str) { + var buf, c, c2, m_pos, i, str_len = str.length, buf_len = 0; + + // count binary size + for (m_pos = 0; m_pos < str_len; m_pos++) { + c = str.charCodeAt(m_pos); + if ((c & 0xfc00) === 0xd800 && (m_pos+1 < str_len)) { + c2 = str.charCodeAt(m_pos+1); + if ((c2 & 0xfc00) === 0xdc00) { + c = 0x10000 + ((c - 0xd800) << 10) + (c2 - 0xdc00); + m_pos++; + } + } + buf_len += c < 0x80 ? 1 : c < 0x800 ? 2 : c < 0x10000 ? 3 : 4; + } + + // allocate buffer + if (support.uint8array) { + buf = new Uint8Array(buf_len); + } else { + buf = new Array(buf_len); + } + + // convert + for (i=0, m_pos = 0; i < buf_len; m_pos++) { + c = str.charCodeAt(m_pos); + if ((c & 0xfc00) === 0xd800 && (m_pos+1 < str_len)) { + c2 = str.charCodeAt(m_pos+1); + if ((c2 & 0xfc00) === 0xdc00) { + c = 0x10000 + ((c - 0xd800) << 10) + (c2 - 0xdc00); + m_pos++; + } + } + if (c < 0x80) { + /* one byte */ + buf[i++] = c; + } else if (c < 0x800) { + /* two bytes */ + buf[i++] = 0xC0 | (c >>> 6); + buf[i++] = 0x80 | (c & 0x3f); + } else if (c < 0x10000) { + /* three bytes */ + buf[i++] = 0xE0 | (c >>> 12); + buf[i++] = 0x80 | (c >>> 6 & 0x3f); + buf[i++] = 0x80 | (c & 0x3f); + } else { + /* four bytes */ + buf[i++] = 0xf0 | (c >>> 18); + buf[i++] = 0x80 | (c >>> 12 & 0x3f); + buf[i++] = 0x80 | (c >>> 6 & 0x3f); + buf[i++] = 0x80 | (c & 0x3f); + } + } + + return buf; +}; + +// Calculate max possible position in utf8 buffer, +// that will not break sequence. If that's not possible +// - (very small limits) return max size as is. +// +// buf[] - utf8 bytes array +// max - length limit (mandatory); +var utf8border = function(buf, max) { + var pos; + + max = max || buf.length; + if (max > buf.length) { max = buf.length; } + + // go back from last position, until start of sequence found + pos = max-1; + while (pos >= 0 && (buf[pos] & 0xC0) === 0x80) { pos--; } + + // Fuckup - very small and broken sequence, + // return max, because we should return something anyway. + if (pos < 0) { return max; } + + // If we came to start of buffer - that means vuffer is too small, + // return max too. + if (pos === 0) { return max; } + + return (pos + _utf8len[buf[pos]] > max) ? pos : max; +}; + +// convert array to string +var buf2string = function (buf) { + var str, i, out, c, c_len; + var len = buf.length; + + // Reserve max possible length (2 words per char) + // NB: by unknown reasons, Array is significantly faster for + // String.fromCharCode.apply than Uint16Array. + var utf16buf = new Array(len*2); + + for (out=0, i=0; i 4) { utf16buf[out++] = 0xfffd; i += c_len-1; continue; } + + // apply mask on first byte + c &= c_len === 2 ? 0x1f : c_len === 3 ? 0x0f : 0x07; + // join the rest + while (c_len > 1 && i < len) { + c = (c << 6) | (buf[i++] & 0x3f); + c_len--; + } + + // terminated by end of string? + if (c_len > 1) { utf16buf[out++] = 0xfffd; continue; } + + if (c < 0x10000) { + utf16buf[out++] = c; + } else { + c -= 0x10000; + utf16buf[out++] = 0xd800 | ((c >> 10) & 0x3ff); + utf16buf[out++] = 0xdc00 | (c & 0x3ff); + } + } + + // shrinkBuf(utf16buf, out) + if (utf16buf.length !== out) { + if(utf16buf.subarray) { + utf16buf = utf16buf.subarray(0, out); + } else { + utf16buf.length = out; + } + } + + // return String.fromCharCode.apply(null, utf16buf); + return utils.applyFromCharCode(utf16buf); +}; + + +// That's all for the pako functions. + + +/** + * Transform a javascript string into an array (typed if possible) of bytes, + * UTF-8 encoded. + * @param {String} str the string to encode + * @return {Array|Uint8Array|Buffer} the UTF-8 encoded string. + */ +exports.utf8encode = function utf8encode(str) { + if (support.nodebuffer) { + return nodejsUtils.newBufferFrom(str, "utf-8"); + } + + return string2buf(str); +}; + + +/** + * Transform a bytes array (or a representation) representing an UTF-8 encoded + * string into a javascript string. + * @param {Array|Uint8Array|Buffer} buf the data de decode + * @return {String} the decoded string. + */ +exports.utf8decode = function utf8decode(buf) { + if (support.nodebuffer) { + return utils.transformTo("nodebuffer", buf).toString("utf-8"); + } + + buf = utils.transformTo(support.uint8array ? "uint8array" : "array", buf); + + return buf2string(buf); +}; + +/** + * A worker to decode utf8 encoded binary chunks into string chunks. + * @constructor + */ +function Utf8DecodeWorker() { + GenericWorker.call(this, "utf-8 decode"); + // the last bytes if a chunk didn't end with a complete codepoint. + this.leftOver = null; +} +utils.inherits(Utf8DecodeWorker, GenericWorker); + +/** + * @see GenericWorker.processChunk + */ +Utf8DecodeWorker.prototype.processChunk = function (chunk) { + + var data = utils.transformTo(support.uint8array ? "uint8array" : "array", chunk.data); + + // 1st step, re-use what's left of the previous chunk + if (this.leftOver && this.leftOver.length) { + if(support.uint8array) { + var previousData = data; + data = new Uint8Array(previousData.length + this.leftOver.length); + data.set(this.leftOver, 0); + data.set(previousData, this.leftOver.length); + } else { + data = this.leftOver.concat(data); + } + this.leftOver = null; + } + + var nextBoundary = utf8border(data); + var usableData = data; + if (nextBoundary !== data.length) { + if (support.uint8array) { + usableData = data.subarray(0, nextBoundary); + this.leftOver = data.subarray(nextBoundary, data.length); + } else { + usableData = data.slice(0, nextBoundary); + this.leftOver = data.slice(nextBoundary, data.length); + } + } + + this.push({ + data : exports.utf8decode(usableData), + meta : chunk.meta + }); +}; + +/** + * @see GenericWorker.flush + */ +Utf8DecodeWorker.prototype.flush = function () { + if(this.leftOver && this.leftOver.length) { + this.push({ + data : exports.utf8decode(this.leftOver), + meta : {} + }); + this.leftOver = null; + } +}; +exports.Utf8DecodeWorker = Utf8DecodeWorker; + +/** + * A worker to endcode string chunks into utf8 encoded binary chunks. + * @constructor + */ +function Utf8EncodeWorker() { + GenericWorker.call(this, "utf-8 encode"); +} +utils.inherits(Utf8EncodeWorker, GenericWorker); + +/** + * @see GenericWorker.processChunk + */ +Utf8EncodeWorker.prototype.processChunk = function (chunk) { + this.push({ + data : exports.utf8encode(chunk.data), + meta : chunk.meta + }); +}; +exports.Utf8EncodeWorker = Utf8EncodeWorker; + +},{"./nodejsUtils":14,"./stream/GenericWorker":28,"./support":30,"./utils":32}],32:[function(require,module,exports){ +'use strict'; + +var support = require('./support'); +var base64 = require('./base64'); +var nodejsUtils = require('./nodejsUtils'); +var setImmediate = require('set-immediate-shim'); +var external = require("./external"); + + +/** + * Convert a string that pass as a "binary string": it should represent a byte + * array but may have > 255 char codes. Be sure to take only the first byte + * and returns the byte array. + * @param {String} str the string to transform. + * @return {Array|Uint8Array} the string in a binary format. + */ +function string2binary(str) { + var result = null; + if (support.uint8array) { + result = new Uint8Array(str.length); + } else { + result = new Array(str.length); + } + return stringToArrayLike(str, result); +} + +/** + * Create a new blob with the given content and the given type. + * @param {String|ArrayBuffer} part the content to put in the blob. DO NOT use + * an Uint8Array because the stock browser of android 4 won't accept it (it + * will be silently converted to a string, "[object Uint8Array]"). + * + * Use only ONE part to build the blob to avoid a memory leak in IE11 / Edge: + * when a large amount of Array is used to create the Blob, the amount of + * memory consumed is nearly 100 times the original data amount. + * + * @param {String} type the mime type of the blob. + * @return {Blob} the created blob. + */ +exports.newBlob = function(part, type) { + exports.checkSupport("blob"); + + try { + // Blob constructor + return new Blob([part], { + type: type + }); + } + catch (e) { + + try { + // deprecated, browser only, old way + var Builder = self.BlobBuilder || self.WebKitBlobBuilder || self.MozBlobBuilder || self.MSBlobBuilder; + var builder = new Builder(); + builder.append(part); + return builder.getBlob(type); + } + catch (e) { + + // well, fuck ?! + throw new Error("Bug : can't construct the Blob."); + } + } + + +}; +/** + * The identity function. + * @param {Object} input the input. + * @return {Object} the same input. + */ +function identity(input) { + return input; +} + +/** + * Fill in an array with a string. + * @param {String} str the string to use. + * @param {Array|ArrayBuffer|Uint8Array|Buffer} array the array to fill in (will be mutated). + * @return {Array|ArrayBuffer|Uint8Array|Buffer} the updated array. + */ +function stringToArrayLike(str, array) { + for (var i = 0; i < str.length; ++i) { + array[i] = str.charCodeAt(i) & 0xFF; + } + return array; +} + +/** + * An helper for the function arrayLikeToString. + * This contains static information and functions that + * can be optimized by the browser JIT compiler. + */ +var arrayToStringHelper = { + /** + * Transform an array of int into a string, chunk by chunk. + * See the performances notes on arrayLikeToString. + * @param {Array|ArrayBuffer|Uint8Array|Buffer} array the array to transform. + * @param {String} type the type of the array. + * @param {Integer} chunk the chunk size. + * @return {String} the resulting string. + * @throws Error if the chunk is too big for the stack. + */ + stringifyByChunk: function(array, type, chunk) { + var result = [], k = 0, len = array.length; + // shortcut + if (len <= chunk) { + return String.fromCharCode.apply(null, array); + } + while (k < len) { + if (type === "array" || type === "nodebuffer") { + result.push(String.fromCharCode.apply(null, array.slice(k, Math.min(k + chunk, len)))); + } + else { + result.push(String.fromCharCode.apply(null, array.subarray(k, Math.min(k + chunk, len)))); + } + k += chunk; + } + return result.join(""); + }, + /** + * Call String.fromCharCode on every item in the array. + * This is the naive implementation, which generate A LOT of intermediate string. + * This should be used when everything else fail. + * @param {Array|ArrayBuffer|Uint8Array|Buffer} array the array to transform. + * @return {String} the result. + */ + stringifyByChar: function(array){ + var resultStr = ""; + for(var i = 0; i < array.length; i++) { + resultStr += String.fromCharCode(array[i]); + } + return resultStr; + }, + applyCanBeUsed : { + /** + * true if the browser accepts to use String.fromCharCode on Uint8Array + */ + uint8array : (function () { + try { + return support.uint8array && String.fromCharCode.apply(null, new Uint8Array(1)).length === 1; + } catch (e) { + return false; + } + })(), + /** + * true if the browser accepts to use String.fromCharCode on nodejs Buffer. + */ + nodebuffer : (function () { + try { + return support.nodebuffer && String.fromCharCode.apply(null, nodejsUtils.allocBuffer(1)).length === 1; + } catch (e) { + return false; + } + })() + } +}; + +/** + * Transform an array-like object to a string. + * @param {Array|ArrayBuffer|Uint8Array|Buffer} array the array to transform. + * @return {String} the result. + */ +function arrayLikeToString(array) { + // Performances notes : + // -------------------- + // String.fromCharCode.apply(null, array) is the fastest, see + // see http://jsperf.com/converting-a-uint8array-to-a-string/2 + // but the stack is limited (and we can get huge arrays !). + // + // result += String.fromCharCode(array[i]); generate too many strings ! + // + // This code is inspired by http://jsperf.com/arraybuffer-to-string-apply-performance/2 + // TODO : we now have workers that split the work. Do we still need that ? + var chunk = 65536, + type = exports.getTypeOf(array), + canUseApply = true; + if (type === "uint8array") { + canUseApply = arrayToStringHelper.applyCanBeUsed.uint8array; + } else if (type === "nodebuffer") { + canUseApply = arrayToStringHelper.applyCanBeUsed.nodebuffer; + } + + if (canUseApply) { + while (chunk > 1) { + try { + return arrayToStringHelper.stringifyByChunk(array, type, chunk); + } catch (e) { + chunk = Math.floor(chunk / 2); + } + } + } + + // no apply or chunk error : slow and painful algorithm + // default browser on android 4.* + return arrayToStringHelper.stringifyByChar(array); +} + +exports.applyFromCharCode = arrayLikeToString; + + +/** + * Copy the data from an array-like to an other array-like. + * @param {Array|ArrayBuffer|Uint8Array|Buffer} arrayFrom the origin array. + * @param {Array|ArrayBuffer|Uint8Array|Buffer} arrayTo the destination array which will be mutated. + * @return {Array|ArrayBuffer|Uint8Array|Buffer} the updated destination array. + */ +function arrayLikeToArrayLike(arrayFrom, arrayTo) { + for (var i = 0; i < arrayFrom.length; i++) { + arrayTo[i] = arrayFrom[i]; + } + return arrayTo; +} + +// a matrix containing functions to transform everything into everything. +var transform = {}; + +// string to ? +transform["string"] = { + "string": identity, + "array": function(input) { + return stringToArrayLike(input, new Array(input.length)); + }, + "arraybuffer": function(input) { + return transform["string"]["uint8array"](input).buffer; + }, + "uint8array": function(input) { + return stringToArrayLike(input, new Uint8Array(input.length)); + }, + "nodebuffer": function(input) { + return stringToArrayLike(input, nodejsUtils.allocBuffer(input.length)); + } +}; + +// array to ? +transform["array"] = { + "string": arrayLikeToString, + "array": identity, + "arraybuffer": function(input) { + return (new Uint8Array(input)).buffer; + }, + "uint8array": function(input) { + return new Uint8Array(input); + }, + "nodebuffer": function(input) { + return nodejsUtils.newBufferFrom(input); + } +}; + +// arraybuffer to ? +transform["arraybuffer"] = { + "string": function(input) { + return arrayLikeToString(new Uint8Array(input)); + }, + "array": function(input) { + return arrayLikeToArrayLike(new Uint8Array(input), new Array(input.byteLength)); + }, + "arraybuffer": identity, + "uint8array": function(input) { + return new Uint8Array(input); + }, + "nodebuffer": function(input) { + return nodejsUtils.newBufferFrom(new Uint8Array(input)); + } +}; + +// uint8array to ? +transform["uint8array"] = { + "string": arrayLikeToString, + "array": function(input) { + return arrayLikeToArrayLike(input, new Array(input.length)); + }, + "arraybuffer": function(input) { + return input.buffer; + }, + "uint8array": identity, + "nodebuffer": function(input) { + return nodejsUtils.newBufferFrom(input); + } +}; + +// nodebuffer to ? +transform["nodebuffer"] = { + "string": arrayLikeToString, + "array": function(input) { + return arrayLikeToArrayLike(input, new Array(input.length)); + }, + "arraybuffer": function(input) { + return transform["nodebuffer"]["uint8array"](input).buffer; + }, + "uint8array": function(input) { + return arrayLikeToArrayLike(input, new Uint8Array(input.length)); + }, + "nodebuffer": identity +}; + +/** + * Transform an input into any type. + * The supported output type are : string, array, uint8array, arraybuffer, nodebuffer. + * If no output type is specified, the unmodified input will be returned. + * @param {String} outputType the output type. + * @param {String|Array|ArrayBuffer|Uint8Array|Buffer} input the input to convert. + * @throws {Error} an Error if the browser doesn't support the requested output type. + */ +exports.transformTo = function(outputType, input) { + if (!input) { + // undefined, null, etc + // an empty string won't harm. + input = ""; + } + if (!outputType) { + return input; + } + exports.checkSupport(outputType); + var inputType = exports.getTypeOf(input); + var result = transform[inputType][outputType](input); + return result; +}; + +/** + * Return the type of the input. + * The type will be in a format valid for JSZip.utils.transformTo : string, array, uint8array, arraybuffer. + * @param {Object} input the input to identify. + * @return {String} the (lowercase) type of the input. + */ +exports.getTypeOf = function(input) { + if (typeof input === "string") { + return "string"; + } + if (Object.prototype.toString.call(input) === "[object Array]") { + return "array"; + } + if (support.nodebuffer && nodejsUtils.isBuffer(input)) { + return "nodebuffer"; + } + if (support.uint8array && input instanceof Uint8Array) { + return "uint8array"; + } + if (support.arraybuffer && input instanceof ArrayBuffer) { + return "arraybuffer"; + } +}; + +/** + * Throw an exception if the type is not supported. + * @param {String} type the type to check. + * @throws {Error} an Error if the browser doesn't support the requested type. + */ +exports.checkSupport = function(type) { + var supported = support[type.toLowerCase()]; + if (!supported) { + throw new Error(type + " is not supported by this platform"); + } +}; + +exports.MAX_VALUE_16BITS = 65535; +exports.MAX_VALUE_32BITS = -1; // well, "\xFF\xFF\xFF\xFF\xFF\xFF\xFF\xFF" is parsed as -1 + +/** + * Prettify a string read as binary. + * @param {string} str the string to prettify. + * @return {string} a pretty string. + */ +exports.pretty = function(str) { + var res = '', + code, i; + for (i = 0; i < (str || "").length; i++) { + code = str.charCodeAt(i); + res += '\\x' + (code < 16 ? "0" : "") + code.toString(16).toUpperCase(); + } + return res; +}; + +/** + * Defer the call of a function. + * @param {Function} callback the function to call asynchronously. + * @param {Array} args the arguments to give to the callback. + */ +exports.delay = function(callback, args, self) { + setImmediate(function () { + callback.apply(self || null, args || []); + }); +}; + +/** + * Extends a prototype with an other, without calling a constructor with + * side effects. Inspired by nodejs' `utils.inherits` + * @param {Function} ctor the constructor to augment + * @param {Function} superCtor the parent constructor to use + */ +exports.inherits = function (ctor, superCtor) { + var Obj = function() {}; + Obj.prototype = superCtor.prototype; + ctor.prototype = new Obj(); +}; + +/** + * Merge the objects passed as parameters into a new one. + * @private + * @param {...Object} var_args All objects to merge. + * @return {Object} a new object with the data of the others. + */ +exports.extend = function() { + var result = {}, i, attr; + for (i = 0; i < arguments.length; i++) { // arguments is not enumerable in some browsers + for (attr in arguments[i]) { + if (arguments[i].hasOwnProperty(attr) && typeof result[attr] === "undefined") { + result[attr] = arguments[i][attr]; + } + } + } + return result; +}; + +/** + * Transform arbitrary content into a Promise. + * @param {String} name a name for the content being processed. + * @param {Object} inputData the content to process. + * @param {Boolean} isBinary true if the content is not an unicode string + * @param {Boolean} isOptimizedBinaryString true if the string content only has one byte per character. + * @param {Boolean} isBase64 true if the string content is encoded with base64. + * @return {Promise} a promise in a format usable by JSZip. + */ +exports.prepareContent = function(name, inputData, isBinary, isOptimizedBinaryString, isBase64) { + + // if inputData is already a promise, this flatten it. + var promise = external.Promise.resolve(inputData).then(function(data) { + + + var isBlob = support.blob && (data instanceof Blob || ['[object File]', '[object Blob]'].indexOf(Object.prototype.toString.call(data)) !== -1); + + if (isBlob && typeof FileReader !== "undefined") { + return new external.Promise(function (resolve, reject) { + var reader = new FileReader(); + + reader.onload = function(e) { + resolve(e.target.result); + }; + reader.onerror = function(e) { + reject(e.target.error); + }; + reader.readAsArrayBuffer(data); + }); + } else { + return data; + } + }); + + return promise.then(function(data) { + var dataType = exports.getTypeOf(data); + + if (!dataType) { + return external.Promise.reject( + new Error("Can't read the data of '" + name + "'. Is it " + + "in a supported JavaScript type (String, Blob, ArrayBuffer, etc) ?") + ); + } + // special case : it's way easier to work with Uint8Array than with ArrayBuffer + if (dataType === "arraybuffer") { + data = exports.transformTo("uint8array", data); + } else if (dataType === "string") { + if (isBase64) { + data = base64.decode(data); + } + else if (isBinary) { + // optimizedBinaryString === true means that the file has already been filtered with a 0xFF mask + if (isOptimizedBinaryString !== true) { + // this is a string, not in a base64 format. + // Be sure that this is a correct "binary string" + data = string2binary(data); + } + } + } + return data; + }); +}; + +},{"./base64":1,"./external":6,"./nodejsUtils":14,"./support":30,"set-immediate-shim":54}],33:[function(require,module,exports){ +'use strict'; +var readerFor = require('./reader/readerFor'); +var utils = require('./utils'); +var sig = require('./signature'); +var ZipEntry = require('./zipEntry'); +var utf8 = require('./utf8'); +var support = require('./support'); +// class ZipEntries {{{ +/** + * All the entries in the zip file. + * @constructor + * @param {Object} loadOptions Options for loading the stream. + */ +function ZipEntries(loadOptions) { + this.files = []; + this.loadOptions = loadOptions; +} +ZipEntries.prototype = { + /** + * Check that the reader is on the specified signature. + * @param {string} expectedSignature the expected signature. + * @throws {Error} if it is an other signature. + */ + checkSignature: function(expectedSignature) { + if (!this.reader.readAndCheckSignature(expectedSignature)) { + this.reader.index -= 4; + var signature = this.reader.readString(4); + throw new Error("Corrupted zip or bug: unexpected signature " + "(" + utils.pretty(signature) + ", expected " + utils.pretty(expectedSignature) + ")"); + } + }, + /** + * Check if the given signature is at the given index. + * @param {number} askedIndex the index to check. + * @param {string} expectedSignature the signature to expect. + * @return {boolean} true if the signature is here, false otherwise. + */ + isSignature: function(askedIndex, expectedSignature) { + var currentIndex = this.reader.index; + this.reader.setIndex(askedIndex); + var signature = this.reader.readString(4); + var result = signature === expectedSignature; + this.reader.setIndex(currentIndex); + return result; + }, + /** + * Read the end of the central directory. + */ + readBlockEndOfCentral: function() { + this.diskNumber = this.reader.readInt(2); + this.diskWithCentralDirStart = this.reader.readInt(2); + this.centralDirRecordsOnThisDisk = this.reader.readInt(2); + this.centralDirRecords = this.reader.readInt(2); + this.centralDirSize = this.reader.readInt(4); + this.centralDirOffset = this.reader.readInt(4); + + this.zipCommentLength = this.reader.readInt(2); + // warning : the encoding depends of the system locale + // On a linux machine with LANG=en_US.utf8, this field is utf8 encoded. + // On a windows machine, this field is encoded with the localized windows code page. + var zipComment = this.reader.readData(this.zipCommentLength); + var decodeParamType = support.uint8array ? "uint8array" : "array"; + // To get consistent behavior with the generation part, we will assume that + // this is utf8 encoded unless specified otherwise. + var decodeContent = utils.transformTo(decodeParamType, zipComment); + this.zipComment = this.loadOptions.decodeFileName(decodeContent); + }, + /** + * Read the end of the Zip 64 central directory. + * Not merged with the method readEndOfCentral : + * The end of central can coexist with its Zip64 brother, + * I don't want to read the wrong number of bytes ! + */ + readBlockZip64EndOfCentral: function() { + this.zip64EndOfCentralSize = this.reader.readInt(8); + this.reader.skip(4); + // this.versionMadeBy = this.reader.readString(2); + // this.versionNeeded = this.reader.readInt(2); + this.diskNumber = this.reader.readInt(4); + this.diskWithCentralDirStart = this.reader.readInt(4); + this.centralDirRecordsOnThisDisk = this.reader.readInt(8); + this.centralDirRecords = this.reader.readInt(8); + this.centralDirSize = this.reader.readInt(8); + this.centralDirOffset = this.reader.readInt(8); + + this.zip64ExtensibleData = {}; + var extraDataSize = this.zip64EndOfCentralSize - 44, + index = 0, + extraFieldId, + extraFieldLength, + extraFieldValue; + while (index < extraDataSize) { + extraFieldId = this.reader.readInt(2); + extraFieldLength = this.reader.readInt(4); + extraFieldValue = this.reader.readData(extraFieldLength); + this.zip64ExtensibleData[extraFieldId] = { + id: extraFieldId, + length: extraFieldLength, + value: extraFieldValue + }; + } + }, + /** + * Read the end of the Zip 64 central directory locator. + */ + readBlockZip64EndOfCentralLocator: function() { + this.diskWithZip64CentralDirStart = this.reader.readInt(4); + this.relativeOffsetEndOfZip64CentralDir = this.reader.readInt(8); + this.disksCount = this.reader.readInt(4); + if (this.disksCount > 1) { + throw new Error("Multi-volumes zip are not supported"); + } + }, + /** + * Read the local files, based on the offset read in the central part. + */ + readLocalFiles: function() { + var i, file; + for (i = 0; i < this.files.length; i++) { + file = this.files[i]; + this.reader.setIndex(file.localHeaderOffset); + this.checkSignature(sig.LOCAL_FILE_HEADER); + file.readLocalPart(this.reader); + file.handleUTF8(); + file.processAttributes(); + } + }, + /** + * Read the central directory. + */ + readCentralDir: function() { + var file; + + this.reader.setIndex(this.centralDirOffset); + while (this.reader.readAndCheckSignature(sig.CENTRAL_FILE_HEADER)) { + file = new ZipEntry({ + zip64: this.zip64 + }, this.loadOptions); + file.readCentralPart(this.reader); + this.files.push(file); + } + + if (this.centralDirRecords !== this.files.length) { + if (this.centralDirRecords !== 0 && this.files.length === 0) { + // We expected some records but couldn't find ANY. + // This is really suspicious, as if something went wrong. + throw new Error("Corrupted zip or bug: expected " + this.centralDirRecords + " records in central dir, got " + this.files.length); + } else { + // We found some records but not all. + // Something is wrong but we got something for the user: no error here. + // console.warn("expected", this.centralDirRecords, "records in central dir, got", this.files.length); + } + } + }, + /** + * Read the end of central directory. + */ + readEndOfCentral: function() { + var offset = this.reader.lastIndexOfSignature(sig.CENTRAL_DIRECTORY_END); + if (offset < 0) { + // Check if the content is a truncated zip or complete garbage. + // A "LOCAL_FILE_HEADER" is not required at the beginning (auto + // extractible zip for example) but it can give a good hint. + // If an ajax request was used without responseType, we will also + // get unreadable data. + var isGarbage = !this.isSignature(0, sig.LOCAL_FILE_HEADER); + + if (isGarbage) { + throw new Error("Can't find end of central directory : is this a zip file ? " + + "If it is, see https://stuk.github.io/jszip/documentation/howto/read_zip.html"); + } else { + throw new Error("Corrupted zip: can't find end of central directory"); + } + + } + this.reader.setIndex(offset); + var endOfCentralDirOffset = offset; + this.checkSignature(sig.CENTRAL_DIRECTORY_END); + this.readBlockEndOfCentral(); + + + /* extract from the zip spec : + 4) If one of the fields in the end of central directory + record is too small to hold required data, the field + should be set to -1 (0xFFFF or 0xFFFFFFFF) and the + ZIP64 format record should be created. + 5) The end of central directory record and the + Zip64 end of central directory locator record must + reside on the same disk when splitting or spanning + an archive. + */ + if (this.diskNumber === utils.MAX_VALUE_16BITS || this.diskWithCentralDirStart === utils.MAX_VALUE_16BITS || this.centralDirRecordsOnThisDisk === utils.MAX_VALUE_16BITS || this.centralDirRecords === utils.MAX_VALUE_16BITS || this.centralDirSize === utils.MAX_VALUE_32BITS || this.centralDirOffset === utils.MAX_VALUE_32BITS) { + this.zip64 = true; + + /* + Warning : the zip64 extension is supported, but ONLY if the 64bits integer read from + the zip file can fit into a 32bits integer. This cannot be solved : JavaScript represents + all numbers as 64-bit double precision IEEE 754 floating point numbers. + So, we have 53bits for integers and bitwise operations treat everything as 32bits. + see https://developer.mozilla.org/en-US/docs/JavaScript/Reference/Operators/Bitwise_Operators + and http://www.ecma-international.org/publications/files/ECMA-ST/ECMA-262.pdf section 8.5 + */ + + // should look for a zip64 EOCD locator + offset = this.reader.lastIndexOfSignature(sig.ZIP64_CENTRAL_DIRECTORY_LOCATOR); + if (offset < 0) { + throw new Error("Corrupted zip: can't find the ZIP64 end of central directory locator"); + } + this.reader.setIndex(offset); + this.checkSignature(sig.ZIP64_CENTRAL_DIRECTORY_LOCATOR); + this.readBlockZip64EndOfCentralLocator(); + + // now the zip64 EOCD record + if (!this.isSignature(this.relativeOffsetEndOfZip64CentralDir, sig.ZIP64_CENTRAL_DIRECTORY_END)) { + // console.warn("ZIP64 end of central directory not where expected."); + this.relativeOffsetEndOfZip64CentralDir = this.reader.lastIndexOfSignature(sig.ZIP64_CENTRAL_DIRECTORY_END); + if (this.relativeOffsetEndOfZip64CentralDir < 0) { + throw new Error("Corrupted zip: can't find the ZIP64 end of central directory"); + } + } + this.reader.setIndex(this.relativeOffsetEndOfZip64CentralDir); + this.checkSignature(sig.ZIP64_CENTRAL_DIRECTORY_END); + this.readBlockZip64EndOfCentral(); + } + + var expectedEndOfCentralDirOffset = this.centralDirOffset + this.centralDirSize; + if (this.zip64) { + expectedEndOfCentralDirOffset += 20; // end of central dir 64 locator + expectedEndOfCentralDirOffset += 12 /* should not include the leading 12 bytes */ + this.zip64EndOfCentralSize; + } + + var extraBytes = endOfCentralDirOffset - expectedEndOfCentralDirOffset; + + if (extraBytes > 0) { + // console.warn(extraBytes, "extra bytes at beginning or within zipfile"); + if (this.isSignature(endOfCentralDirOffset, sig.CENTRAL_FILE_HEADER)) { + // The offsets seem wrong, but we have something at the specified offset. + // So… we keep it. + } else { + // the offset is wrong, update the "zero" of the reader + // this happens if data has been prepended (crx files for example) + this.reader.zero = extraBytes; + } + } else if (extraBytes < 0) { + throw new Error("Corrupted zip: missing " + Math.abs(extraBytes) + " bytes."); + } + }, + prepareReader: function(data) { + this.reader = readerFor(data); + }, + /** + * Read a zip file and create ZipEntries. + * @param {String|ArrayBuffer|Uint8Array|Buffer} data the binary string representing a zip file. + */ + load: function(data) { + this.prepareReader(data); + this.readEndOfCentral(); + this.readCentralDir(); + this.readLocalFiles(); + } +}; +// }}} end of ZipEntries +module.exports = ZipEntries; + +},{"./reader/readerFor":22,"./signature":23,"./support":30,"./utf8":31,"./utils":32,"./zipEntry":34}],34:[function(require,module,exports){ +'use strict'; +var readerFor = require('./reader/readerFor'); +var utils = require('./utils'); +var CompressedObject = require('./compressedObject'); +var crc32fn = require('./crc32'); +var utf8 = require('./utf8'); +var compressions = require('./compressions'); +var support = require('./support'); + +var MADE_BY_DOS = 0x00; +var MADE_BY_UNIX = 0x03; + +/** + * Find a compression registered in JSZip. + * @param {string} compressionMethod the method magic to find. + * @return {Object|null} the JSZip compression object, null if none found. + */ +var findCompression = function(compressionMethod) { + for (var method in compressions) { + if (!compressions.hasOwnProperty(method)) { + continue; + } + if (compressions[method].magic === compressionMethod) { + return compressions[method]; + } + } + return null; +}; + +// class ZipEntry {{{ +/** + * An entry in the zip file. + * @constructor + * @param {Object} options Options of the current file. + * @param {Object} loadOptions Options for loading the stream. + */ +function ZipEntry(options, loadOptions) { + this.options = options; + this.loadOptions = loadOptions; +} +ZipEntry.prototype = { + /** + * say if the file is encrypted. + * @return {boolean} true if the file is encrypted, false otherwise. + */ + isEncrypted: function() { + // bit 1 is set + return (this.bitFlag & 0x0001) === 0x0001; + }, + /** + * say if the file has utf-8 filename/comment. + * @return {boolean} true if the filename/comment is in utf-8, false otherwise. + */ + useUTF8: function() { + // bit 11 is set + return (this.bitFlag & 0x0800) === 0x0800; + }, + /** + * Read the local part of a zip file and add the info in this object. + * @param {DataReader} reader the reader to use. + */ + readLocalPart: function(reader) { + var compression, localExtraFieldsLength; + + // we already know everything from the central dir ! + // If the central dir data are false, we are doomed. + // On the bright side, the local part is scary : zip64, data descriptors, both, etc. + // The less data we get here, the more reliable this should be. + // Let's skip the whole header and dash to the data ! + reader.skip(22); + // in some zip created on windows, the filename stored in the central dir contains \ instead of /. + // Strangely, the filename here is OK. + // I would love to treat these zip files as corrupted (see http://www.info-zip.org/FAQ.html#backslashes + // or APPNOTE#4.4.17.1, "All slashes MUST be forward slashes '/'") but there are a lot of bad zip generators... + // Search "unzip mismatching "local" filename continuing with "central" filename version" on + // the internet. + // + // I think I see the logic here : the central directory is used to display + // content and the local directory is used to extract the files. Mixing / and \ + // may be used to display \ to windows users and use / when extracting the files. + // Unfortunately, this lead also to some issues : http://seclists.org/fulldisclosure/2009/Sep/394 + this.fileNameLength = reader.readInt(2); + localExtraFieldsLength = reader.readInt(2); // can't be sure this will be the same as the central dir + // the fileName is stored as binary data, the handleUTF8 method will take care of the encoding. + this.fileName = reader.readData(this.fileNameLength); + reader.skip(localExtraFieldsLength); + + if (this.compressedSize === -1 || this.uncompressedSize === -1) { + throw new Error("Bug or corrupted zip : didn't get enough information from the central directory " + "(compressedSize === -1 || uncompressedSize === -1)"); + } + + compression = findCompression(this.compressionMethod); + if (compression === null) { // no compression found + throw new Error("Corrupted zip : compression " + utils.pretty(this.compressionMethod) + " unknown (inner file : " + utils.transformTo("string", this.fileName) + ")"); + } + this.decompressed = new CompressedObject(this.compressedSize, this.uncompressedSize, this.crc32, compression, reader.readData(this.compressedSize)); + }, + + /** + * Read the central part of a zip file and add the info in this object. + * @param {DataReader} reader the reader to use. + */ + readCentralPart: function(reader) { + this.versionMadeBy = reader.readInt(2); + reader.skip(2); + // this.versionNeeded = reader.readInt(2); + this.bitFlag = reader.readInt(2); + this.compressionMethod = reader.readString(2); + this.date = reader.readDate(); + this.crc32 = reader.readInt(4); + this.compressedSize = reader.readInt(4); + this.uncompressedSize = reader.readInt(4); + var fileNameLength = reader.readInt(2); + this.extraFieldsLength = reader.readInt(2); + this.fileCommentLength = reader.readInt(2); + this.diskNumberStart = reader.readInt(2); + this.internalFileAttributes = reader.readInt(2); + this.externalFileAttributes = reader.readInt(4); + this.localHeaderOffset = reader.readInt(4); + + if (this.isEncrypted()) { + throw new Error("Encrypted zip are not supported"); + } + + // will be read in the local part, see the comments there + reader.skip(fileNameLength); + this.readExtraFields(reader); + this.parseZIP64ExtraField(reader); + this.fileComment = reader.readData(this.fileCommentLength); + }, + + /** + * Parse the external file attributes and get the unix/dos permissions. + */ + processAttributes: function () { + this.unixPermissions = null; + this.dosPermissions = null; + var madeBy = this.versionMadeBy >> 8; + + // Check if we have the DOS directory flag set. + // We look for it in the DOS and UNIX permissions + // but some unknown platform could set it as a compatibility flag. + this.dir = this.externalFileAttributes & 0x0010 ? true : false; + + if(madeBy === MADE_BY_DOS) { + // first 6 bits (0 to 5) + this.dosPermissions = this.externalFileAttributes & 0x3F; + } + + if(madeBy === MADE_BY_UNIX) { + this.unixPermissions = (this.externalFileAttributes >> 16) & 0xFFFF; + // the octal permissions are in (this.unixPermissions & 0x01FF).toString(8); + } + + // fail safe : if the name ends with a / it probably means a folder + if (!this.dir && this.fileNameStr.slice(-1) === '/') { + this.dir = true; + } + }, + + /** + * Parse the ZIP64 extra field and merge the info in the current ZipEntry. + * @param {DataReader} reader the reader to use. + */ + parseZIP64ExtraField: function(reader) { + + if (!this.extraFields[0x0001]) { + return; + } + + // should be something, preparing the extra reader + var extraReader = readerFor(this.extraFields[0x0001].value); + + // I really hope that these 64bits integer can fit in 32 bits integer, because js + // won't let us have more. + if (this.uncompressedSize === utils.MAX_VALUE_32BITS) { + this.uncompressedSize = extraReader.readInt(8); + } + if (this.compressedSize === utils.MAX_VALUE_32BITS) { + this.compressedSize = extraReader.readInt(8); + } + if (this.localHeaderOffset === utils.MAX_VALUE_32BITS) { + this.localHeaderOffset = extraReader.readInt(8); + } + if (this.diskNumberStart === utils.MAX_VALUE_32BITS) { + this.diskNumberStart = extraReader.readInt(4); + } + }, + /** + * Read the central part of a zip file and add the info in this object. + * @param {DataReader} reader the reader to use. + */ + readExtraFields: function(reader) { + var end = reader.index + this.extraFieldsLength, + extraFieldId, + extraFieldLength, + extraFieldValue; + + if (!this.extraFields) { + this.extraFields = {}; + } + + while (reader.index + 4 < end) { + extraFieldId = reader.readInt(2); + extraFieldLength = reader.readInt(2); + extraFieldValue = reader.readData(extraFieldLength); + + this.extraFields[extraFieldId] = { + id: extraFieldId, + length: extraFieldLength, + value: extraFieldValue + }; + } + + reader.setIndex(end); + }, + /** + * Apply an UTF8 transformation if needed. + */ + handleUTF8: function() { + var decodeParamType = support.uint8array ? "uint8array" : "array"; + if (this.useUTF8()) { + this.fileNameStr = utf8.utf8decode(this.fileName); + this.fileCommentStr = utf8.utf8decode(this.fileComment); + } else { + var upath = this.findExtraFieldUnicodePath(); + if (upath !== null) { + this.fileNameStr = upath; + } else { + // ASCII text or unsupported code page + var fileNameByteArray = utils.transformTo(decodeParamType, this.fileName); + this.fileNameStr = this.loadOptions.decodeFileName(fileNameByteArray); + } + + var ucomment = this.findExtraFieldUnicodeComment(); + if (ucomment !== null) { + this.fileCommentStr = ucomment; + } else { + // ASCII text or unsupported code page + var commentByteArray = utils.transformTo(decodeParamType, this.fileComment); + this.fileCommentStr = this.loadOptions.decodeFileName(commentByteArray); + } + } + }, + + /** + * Find the unicode path declared in the extra field, if any. + * @return {String} the unicode path, null otherwise. + */ + findExtraFieldUnicodePath: function() { + var upathField = this.extraFields[0x7075]; + if (upathField) { + var extraReader = readerFor(upathField.value); + + // wrong version + if (extraReader.readInt(1) !== 1) { + return null; + } + + // the crc of the filename changed, this field is out of date. + if (crc32fn(this.fileName) !== extraReader.readInt(4)) { + return null; + } + + return utf8.utf8decode(extraReader.readData(upathField.length - 5)); + } + return null; + }, + + /** + * Find the unicode comment declared in the extra field, if any. + * @return {String} the unicode comment, null otherwise. + */ + findExtraFieldUnicodeComment: function() { + var ucommentField = this.extraFields[0x6375]; + if (ucommentField) { + var extraReader = readerFor(ucommentField.value); + + // wrong version + if (extraReader.readInt(1) !== 1) { + return null; + } + + // the crc of the comment changed, this field is out of date. + if (crc32fn(this.fileComment) !== extraReader.readInt(4)) { + return null; + } + + return utf8.utf8decode(extraReader.readData(ucommentField.length - 5)); + } + return null; + } +}; +module.exports = ZipEntry; + +},{"./compressedObject":2,"./compressions":3,"./crc32":4,"./reader/readerFor":22,"./support":30,"./utf8":31,"./utils":32}],35:[function(require,module,exports){ +'use strict'; + +var StreamHelper = require('./stream/StreamHelper'); +var DataWorker = require('./stream/DataWorker'); +var utf8 = require('./utf8'); +var CompressedObject = require('./compressedObject'); +var GenericWorker = require('./stream/GenericWorker'); + +/** + * A simple object representing a file in the zip file. + * @constructor + * @param {string} name the name of the file + * @param {String|ArrayBuffer|Uint8Array|Buffer} data the data + * @param {Object} options the options of the file + */ +var ZipObject = function(name, data, options) { + this.name = name; + this.dir = options.dir; + this.date = options.date; + this.comment = options.comment; + this.unixPermissions = options.unixPermissions; + this.dosPermissions = options.dosPermissions; + + this._data = data; + this._dataBinary = options.binary; + // keep only the compression + this.options = { + compression : options.compression, + compressionOptions : options.compressionOptions + }; +}; + +ZipObject.prototype = { + /** + * Create an internal stream for the content of this object. + * @param {String} type the type of each chunk. + * @return StreamHelper the stream. + */ + internalStream: function (type) { + var result = null, outputType = "string"; + try { + if (!type) { + throw new Error("No output type specified."); + } + outputType = type.toLowerCase(); + var askUnicodeString = outputType === "string" || outputType === "text"; + if (outputType === "binarystring" || outputType === "text") { + outputType = "string"; + } + result = this._decompressWorker(); + + var isUnicodeString = !this._dataBinary; + + if (isUnicodeString && !askUnicodeString) { + result = result.pipe(new utf8.Utf8EncodeWorker()); + } + if (!isUnicodeString && askUnicodeString) { + result = result.pipe(new utf8.Utf8DecodeWorker()); + } + } catch (e) { + result = new GenericWorker("error"); + result.error(e); + } + + return new StreamHelper(result, outputType, ""); + }, + + /** + * Prepare the content in the asked type. + * @param {String} type the type of the result. + * @param {Function} onUpdate a function to call on each internal update. + * @return Promise the promise of the result. + */ + async: function (type, onUpdate) { + return this.internalStream(type).accumulate(onUpdate); + }, + + /** + * Prepare the content as a nodejs stream. + * @param {String} type the type of each chunk. + * @param {Function} onUpdate a function to call on each internal update. + * @return Stream the stream. + */ + nodeStream: function (type, onUpdate) { + return this.internalStream(type || "nodebuffer").toNodejsStream(onUpdate); + }, + + /** + * Return a worker for the compressed content. + * @private + * @param {Object} compression the compression object to use. + * @param {Object} compressionOptions the options to use when compressing. + * @return Worker the worker. + */ + _compressWorker: function (compression, compressionOptions) { + if ( + this._data instanceof CompressedObject && + this._data.compression.magic === compression.magic + ) { + return this._data.getCompressedWorker(); + } else { + var result = this._decompressWorker(); + if(!this._dataBinary) { + result = result.pipe(new utf8.Utf8EncodeWorker()); + } + return CompressedObject.createWorkerFrom(result, compression, compressionOptions); + } + }, + /** + * Return a worker for the decompressed content. + * @private + * @return Worker the worker. + */ + _decompressWorker : function () { + if (this._data instanceof CompressedObject) { + return this._data.getContentWorker(); + } else if (this._data instanceof GenericWorker) { + return this._data; + } else { + return new DataWorker(this._data); + } + } +}; + +var removedMethods = ["asText", "asBinary", "asNodeBuffer", "asUint8Array", "asArrayBuffer"]; +var removedFn = function () { + throw new Error("This method has been removed in JSZip 3.0, please check the upgrade guide."); +}; + +for(var i = 0; i < removedMethods.length; i++) { + ZipObject.prototype[removedMethods[i]] = removedFn; +} +module.exports = ZipObject; + +},{"./compressedObject":2,"./stream/DataWorker":27,"./stream/GenericWorker":28,"./stream/StreamHelper":29,"./utf8":31}],36:[function(require,module,exports){ +(function (global){ +'use strict'; +var Mutation = global.MutationObserver || global.WebKitMutationObserver; + +var scheduleDrain; + +{ + if (Mutation) { + var called = 0; + var observer = new Mutation(nextTick); + var element = global.document.createTextNode(''); + observer.observe(element, { + characterData: true + }); + scheduleDrain = function () { + element.data = (called = ++called % 2); + }; + } else if (!global.setImmediate && typeof global.MessageChannel !== 'undefined') { + var channel = new global.MessageChannel(); + channel.port1.onmessage = nextTick; + scheduleDrain = function () { + channel.port2.postMessage(0); + }; + } else if ('document' in global && 'onreadystatechange' in global.document.createElement('script')) { + scheduleDrain = function () { + + // Create a + + + + + + + + + +
+ +
+ +
+
+ +

Interface Connection

+
+
+
+
    +
  • +
    +
    All Superinterfaces:
    +
    AutoCloseable
    +
    +
    +
    All Known Implementing Classes:
    +
    FlussConnection
    +
    +
    +
    @PublicEvolving
    +@ThreadSafe
    +public interface Connection
    +extends AutoCloseable
    +
    A cluster connection encapsulating lower level individual connections to actual Fluss servers. + Connections are instantiated through the ConnectionFactory class. The lifecycle of the + connection is managed by the caller, who has to close() the connection to release the + resources. The connection object contains logic to find the Coordinator, locate table buckets out + on the cluster, keeps a cache of locations and then knows how to re-calibrate after they move. + The individual connections to servers, meta cache, etc. are all shared by the Table and + Admin instances obtained from this connection. + +

    Connection creation is a heavy-weight operation. Connection implementations are thread-safe, + so that the client can create a connection once, and share it with different threads. + Table and Admin instances, on the other hand, are light-weight and are not thread-safe. + Typically, a single connection per client application is instantiated and every thread will + obtain its own Table instance. Caching or pooling of Table and Admin is + not recommended.

    +
    +
    Since:
    +
    0.1
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Method Summary

      + + + + + + + + + + + + + + + + + + + + + + + + + + + +
      All Methods Instance Methods Abstract Methods 
      Modifier and TypeMethodDescription
      voidclose() +
      Close the connection and release all resources.
      +
      AdmingetAdmin() +
      Retrieve a new Admin client to administer a Fluss cluster.
      +
      org.apache.fluss.config.ConfigurationgetConfiguration() +
      Retrieve the configuration used to create this connection.
      +
      TablegetTable​(org.apache.fluss.metadata.TablePath tablePath) +
      Retrieve a new Table client to operate data in table.
      +
      +
    • +
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getConfiguration

        +
        org.apache.fluss.config.Configuration getConfiguration()
        +
        Retrieve the configuration used to create this connection.
        +
      • +
      + + + +
        +
      • +

        getAdmin

        +
        Admin getAdmin()
        +
        Retrieve a new Admin client to administer a Fluss cluster.
        +
      • +
      + + + +
        +
      • +

        getTable

        +
        Table getTable​(org.apache.fluss.metadata.TablePath tablePath)
        +
        Retrieve a new Table client to operate data in table.
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/ConnectionFactory.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/ConnectionFactory.html new file mode 100644 index 0000000000..5848e8c490 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/ConnectionFactory.html @@ -0,0 +1,314 @@ + + + + + +ConnectionFactory (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class ConnectionFactory

+
+
+ +
+
    +
  • +
    +
    @PublicEvolving
    +public class ConnectionFactory
    +extends Object
    +
    A non-instantiable class that manages creation of Connections. Managing the lifecycle of + the Connections to the cluster is the responsibility of the caller. From a Connection, Admin implementations are retrieved with Connection.getAdmin().
    +
    +
    Since:
    +
    0.1
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        createConnection

        +
        public static Connection createConnection​(org.apache.fluss.config.Configuration conf)
        +
        Creates a new Connection to the Fluss cluster. The given configuration at least needs + to contain "bootstrap.servers" to discover the Fluss cluster. Here is a simple example: + +
        
        + Configuration conf = new Configuration();
        + conf.setString("bootstrap.servers", "localhost:9092");
        + Connection connection = ConnectionFactory.createConnection(conf);
        + Admin admin = connection.getAdmin();
        + try {
        +    // Use the admin as needed, for a single operation and a single thread
        +  } finally {
        +    admin.close();
        +    connection.close();
        +  }
        + 
        +
      • +
      + + + +
        +
      • +

        createConnection

        +
        public static Connection createConnection​(org.apache.fluss.config.Configuration conf,
        +                                          org.apache.fluss.metrics.registry.MetricRegistry metricRegistry)
        +
        Create a new Connection to the Fluss cluster with registering metrics to the given + metricRegistry. It's mainly used for client to register metrics to external metrics + system. + +

        See more comments in method createConnection(Configuration)

        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/FlussConnection.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/FlussConnection.html new file mode 100644 index 0000000000..62fa30c299 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/FlussConnection.html @@ -0,0 +1,435 @@ + + + + + +FlussConnection (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlussConnection

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    AutoCloseable, Connection
    +
    +
    +
    public final class FlussConnection
    +extends Object
    +implements Connection
    +
    A connection to Fluss cluster, and holds the client session resources.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getConfiguration

        +
        public org.apache.fluss.config.Configuration getConfiguration()
        +
        Description copied from interface: Connection
        +
        Retrieve the configuration used to create this connection.
        +
        +
        Specified by:
        +
        getConfiguration in interface Connection
        +
        +
      • +
      + + + +
        +
      • +

        getAdmin

        +
        public Admin getAdmin()
        +
        Description copied from interface: Connection
        +
        Retrieve a new Admin client to administer a Fluss cluster.
        +
        +
        Specified by:
        +
        getAdmin in interface Connection
        +
        +
      • +
      + + + +
        +
      • +

        getTable

        +
        public Table getTable​(org.apache.fluss.metadata.TablePath tablePath)
        +
        Description copied from interface: Connection
        +
        Retrieve a new Table client to operate data in table.
        +
        +
        Specified by:
        +
        getTable in interface Connection
        +
        +
      • +
      + + + +
        +
      • +

        getRpcClient

        +
        public org.apache.fluss.rpc.RpcClient getRpcClient()
        +
      • +
      + + + + + + + +
        +
      • +

        getClientMetricGroup

        +
        public org.apache.fluss.rpc.metrics.ClientMetricGroup getClientMetricGroup()
        +
      • +
      + + + +
        +
      • +

        getOrCreateWriterClient

        +
        public WriterClient getOrCreateWriterClient()
        +
      • +
      + + + +
        +
      • +

        getOrCreateLookupClient

        +
        public LookupClient getOrCreateLookupClient()
        +
      • +
      + + + +
        +
      • +

        getOrCreateRemoteFileDownloader

        +
        public RemoteFileDownloader getOrCreateRemoteFileDownloader()
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/Admin.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/Admin.html new file mode 100644 index 0000000000..8c9115ccd0 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/Admin.html @@ -0,0 +1,1040 @@ + + + + + +Admin (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface Admin

+
+
+
+
    +
  • +
    +
    All Superinterfaces:
    +
    AutoCloseable
    +
    +
    +
    All Known Implementing Classes:
    +
    FlussAdmin
    +
    +
    +
    @PublicEvolving
    +public interface Admin
    +extends AutoCloseable
    +
    The administrative client for Fluss, which supports managing and inspecting tables, servers, + configurations and ACLs.
    +
    +
    Since:
    +
    0.1
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getServerNodes

        +
        CompletableFuture<List<org.apache.fluss.cluster.ServerNode>> getServerNodes()
        +
        Get the current server node information. asynchronously.
        +
      • +
      + + + +
        +
      • +

        getTableSchema

        +
        CompletableFuture<org.apache.fluss.metadata.SchemaInfo> getTableSchema​(org.apache.fluss.metadata.TablePath tablePath)
        +
        Get the latest table schema of the given table asynchronously. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • TableNotExistException if the table does not exist. +
        +
        +
        Parameters:
        +
        tablePath - the table path of the table.
        +
        +
      • +
      + + + +
        +
      • +

        getTableSchema

        +
        CompletableFuture<org.apache.fluss.metadata.SchemaInfo> getTableSchema​(org.apache.fluss.metadata.TablePath tablePath,
        +                                                                       int schemaId)
        +
        Get the specific table schema of the given table by schema id asynchronously. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • TableNotExistException if the table does not exist. +
        • SchemaNotExistException if the schema does not exist. +
        +
        +
        Parameters:
        +
        tablePath - the table path of the table.
        +
        +
      • +
      + + + +
        +
      • +

        createDatabase

        +
        CompletableFuture<Void> createDatabase​(String databaseName,
        +                                       org.apache.fluss.metadata.DatabaseDescriptor databaseDescriptor,
        +                                       boolean ignoreIfExists)
        +
        Create a new database asynchronously. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • DatabaseAlreadyExistException if the database already exists and + ignoreIfExists is false. +
        +
        +
        Parameters:
        +
        databaseName - The name of the database to create.
        +
        databaseDescriptor - The descriptor of the database to create.
        +
        ignoreIfExists - Flag to specify behavior when a database with the given name already + exists: if set to false, throw a DatabaseAlreadyExistException, if set to true, do + nothing.
        +
        Throws:
        +
        org.apache.fluss.exception.InvalidDatabaseException - if the database name is invalid, e.g., contains illegal + characters, or exceeds the maximum length.
        +
        +
      • +
      + + + +
        +
      • +

        getDatabaseInfo

        +
        CompletableFuture<org.apache.fluss.metadata.DatabaseInfo> getDatabaseInfo​(String databaseName)
        +
        Get the database with the given database name asynchronously. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • DatabaseNotExistException if the database does not exist. +
        +
        +
        Parameters:
        +
        databaseName - The database name of the database.
        +
        +
      • +
      + + + +
        +
      • +

        dropDatabase

        +
        CompletableFuture<Void> dropDatabase​(String databaseName,
        +                                     boolean ignoreIfNotExists,
        +                                     boolean cascade)
        +
        Drop the database with the given name asynchronously. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • DatabaseNotExistException if the database does not exist and + ignoreIfNotExists is false. +
        • DatabaseNotEmptyException if the database is not empty and cascade is + false. +
        +
        +
        Parameters:
        +
        databaseName - The name of the database to delete.
        +
        ignoreIfNotExists - Flag to specify behavior when a database with the given name does + not exist: if set to false, throw a DatabaseNotExistException, if set to true, do + nothing.
        +
        cascade - Flag to specify whether to delete all tables in the database.
        +
        +
      • +
      + + + +
        +
      • +

        databaseExists

        +
        CompletableFuture<Boolean> databaseExists​(String databaseName)
        +
        Get whether database exists asynchronously.
        +
        +
        Parameters:
        +
        databaseName - The name of the database to check.
        +
        +
      • +
      + + + + + + + +
        +
      • +

        createTable

        +
        CompletableFuture<Void> createTable​(org.apache.fluss.metadata.TablePath tablePath,
        +                                    org.apache.fluss.metadata.TableDescriptor tableDescriptor,
        +                                    boolean ignoreIfExists)
        +                             throws org.apache.fluss.exception.InvalidTableException,
        +                                    org.apache.fluss.exception.InvalidDatabaseException
        +
        Create a new table asynchronously. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • DatabaseNotExistException if the database in the table path does not exist. +
        • TableAlreadyExistException if the table already exists and + ignoreIfExists is false. +
        • InvalidReplicationFactorException if the table's replication factor is larger + than the number of available tablet servers. +
        +
        +
        Parameters:
        +
        tablePath - The tablePath of the table.
        +
        tableDescriptor - The table to create.
        +
        Throws:
        +
        org.apache.fluss.exception.InvalidTableException - if the table name is invalid, e.g., contains illegal + characters, or exceeds the maximum length.
        +
        org.apache.fluss.exception.InvalidDatabaseException - if the database name is invalid, e.g., contains illegal + characters, or exceeds the maximum length.
        +
        +
      • +
      + + + +
        +
      • +

        getTableInfo

        +
        CompletableFuture<org.apache.fluss.metadata.TableInfo> getTableInfo​(org.apache.fluss.metadata.TablePath tablePath)
        +
        Get the table with the given table path asynchronously. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • TableNotExistException if the table does not exist. +
        +
        +
        Parameters:
        +
        tablePath - The table path of the table.
        +
        +
      • +
      + + + +
        +
      • +

        dropTable

        +
        CompletableFuture<Void> dropTable​(org.apache.fluss.metadata.TablePath tablePath,
        +                                  boolean ignoreIfNotExists)
        +
        Drop the table with the given table path asynchronously. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • TableNotExistException if the table does not exist and + ignoreIfNotExists is false. +
        +
        +
        Parameters:
        +
        tablePath - The table path of the table.
        +
        ignoreIfNotExists - Flag to specify behavior when a table with the given name does not + exist: if set to false, throw a TableNotExistException, if set to true, do nothing.
        +
        +
      • +
      + + + +
        +
      • +

        tableExists

        +
        CompletableFuture<Boolean> tableExists​(org.apache.fluss.metadata.TablePath tablePath)
        +
        Get whether table exists asynchronously.
        +
        +
        Parameters:
        +
        tablePath - The table path of the table.
        +
        +
      • +
      + + + +
        +
      • +

        listTables

        +
        CompletableFuture<List<String>> listTables​(String databaseName)
        +
        List all tables in the given database in fluss cluster asynchronously. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • DatabaseNotExistException if the database does not exist. +
        +
        +
        Parameters:
        +
        databaseName - The name of the database.
        +
        +
      • +
      + + + +
        +
      • +

        listPartitionInfos

        +
        CompletableFuture<List<org.apache.fluss.metadata.PartitionInfo>> listPartitionInfos​(org.apache.fluss.metadata.TablePath tablePath)
        +
        List all partitions in the given table in fluss cluster asynchronously. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • TableNotExistException if the table does not exist. +
        • TableNotPartitionedException if the table is not partitioned. +
        +
        +
        Parameters:
        +
        tablePath - The path of the table.
        +
        +
      • +
      + + + +
        +
      • +

        listPartitionInfos

        +
        CompletableFuture<List<org.apache.fluss.metadata.PartitionInfo>> listPartitionInfos​(org.apache.fluss.metadata.TablePath tablePath,
        +                                                                                    org.apache.fluss.metadata.PartitionSpec partialPartitionSpec)
        +
        List all partitions in fluss cluster that are under the given table and the given partial + PartitionSpec asynchronously. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • TableNotExistException if the table does not exist. +
        • TableNotPartitionedException if the table is not partitioned. +
        • InvalidPartitionException if the input partition spec is invalid. +
        +
        +
        Parameters:
        +
        tablePath - The path of the table.
        +
        partialPartitionSpec - Part of table partition spec
        +
        +
      • +
      + + + +
        +
      • +

        createPartition

        +
        CompletableFuture<Void> createPartition​(org.apache.fluss.metadata.TablePath tablePath,
        +                                        org.apache.fluss.metadata.PartitionSpec partitionSpec,
        +                                        boolean ignoreIfExists)
        +
        Create a new partition for a partitioned table. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • TableNotExistException if the table does not exist. +
        • TableNotPartitionedException if the table is not partitioned. +
        • PartitionAlreadyExistsException if the partition already exists and + ignoreIfExists is false. +
        • InvalidPartitionException if the input partition spec is invalid. +
        • TooManyPartitionsException if the number of partitions is larger than the + maximum number of partitions of one table, see ConfigOptions.MAX_PARTITION_NUM. +
        • TooManyBucketsException if the number of buckets is larger than the maximum + number of buckets of one table, see ConfigOptions.MAX_BUCKET_NUM. +
        +
        +
        Parameters:
        +
        tablePath - The table path of the table.
        +
        partitionSpec - The partition spec to add.
        +
        ignoreIfExists - Flag to specify behavior when a partition with the given name already + exists: if set to false, throw a PartitionAlreadyExistsException, if set to true, do + nothing.
        +
        +
      • +
      + + + +
        +
      • +

        dropPartition

        +
        CompletableFuture<Void> dropPartition​(org.apache.fluss.metadata.TablePath tablePath,
        +                                      org.apache.fluss.metadata.PartitionSpec partitionSpec,
        +                                      boolean ignoreIfNotExists)
        +
        Drop a partition from a partitioned table. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • TableNotExistException if the table does not exist. +
        • TableNotPartitionedException if the table is not partitioned. +
        • PartitionNotExistException if the partition not exists and + ignoreIfExists is false. +
        • InvalidPartitionException if the input partition spec is invalid. +
        +
        +
        Parameters:
        +
        tablePath - The table path of the table.
        +
        partitionSpec - The partition spec to drop.
        +
        ignoreIfNotExists - Flag to specify behavior when a partition with the given name does + not exist: if set to false, throw a PartitionNotExistException, if set to true, do + nothing.
        +
        +
      • +
      + + + +
        +
      • +

        getLatestKvSnapshots

        +
        CompletableFuture<KvSnapshots> getLatestKvSnapshots​(org.apache.fluss.metadata.TablePath tablePath)
        +
        Get the latest kv snapshots of the given table asynchronously. A kv snapshot is a snapshot of + a bucket of a primary key table at a certain point in time. Therefore, there are at-most + N snapshots for a primary key table, N is the number of buckets. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • TableNotExistException if the table does not exist. +
        • NonPrimaryKeyTableException if the table is not a primary key table. +
        • PartitionNotExistException if the table is partitioned, use getLatestKvSnapshots(TablePath, String) instead to get the latest kv snapshot of a + partition of a partitioned table. +
        • +
        +
        +
        Parameters:
        +
        tablePath - the table path of the table.
        +
        +
      • +
      + + + +
        +
      • +

        getLatestKvSnapshots

        +
        CompletableFuture<KvSnapshots> getLatestKvSnapshots​(org.apache.fluss.metadata.TablePath tablePath,
        +                                                    String partitionName)
        +
        Get the latest kv snapshots of the given table partition asynchronously. A kv snapshot is a + snapshot of a bucket of a primary key table at a certain point in time. Therefore, there are + at-most N snapshots for a partition of a primary key table, N is the number + of buckets. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • TableNotExistException if the table does not exist. +
        • NonPrimaryKeyTableException if the table is not a primary key table. +
        • PartitionNotExistException if the partition does not exist +
        • TableNotPartitionedException if the table is not partitioned, use getLatestKvSnapshots(TablePath) instead to get the latest kv snapshots for a + non-partitioned table. +
        +
        +
        Parameters:
        +
        tablePath - the table path of the table.
        +
        partitionName - the partition name, see ResolvedPartitionSpec.getPartitionName() + for the format of the partition name.
        +
        +
      • +
      + + + +
        +
      • +

        getKvSnapshotMetadata

        +
        CompletableFuture<KvSnapshotMetadata> getKvSnapshotMetadata​(org.apache.fluss.metadata.TableBucket bucket,
        +                                                            long snapshotId)
        +
        Get the kv snapshot metadata of the given kv snapshot asynchronously. The kv snapshot + metadata including the snapshot files for the kv tablet and the log offset for the changelog + at the snapshot time. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • KvSnapshotNotExistException if the snapshot does not exist. +
        +
        +
        Parameters:
        +
        bucket - the table bucket of the kv snapshot.
        +
        snapshotId - the snapshot id.
        +
        +
      • +
      + + + +
        +
      • +

        getLatestLakeSnapshot

        +
        CompletableFuture<LakeSnapshot> getLatestLakeSnapshot​(org.apache.fluss.metadata.TablePath tablePath)
        +
        Get table lake snapshot info of the given table asynchronously. + +

        It'll get the latest snapshot for all the buckets of the table. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • TableNotExistException if the table does not exist. +
        • LakeTableSnapshotNotExistException if no any lake snapshot exist. +
        +
        +
        Parameters:
        +
        tablePath - the table path of the table.
        +
        +
      • +
      + + + +
        +
      • +

        listOffsets

        +
        ListOffsetsResult listOffsets​(org.apache.fluss.metadata.TablePath tablePath,
        +                              Collection<Integer> buckets,
        +                              OffsetSpec offsetSpec)
        +
        List offset for the specified buckets. This operation enables to find the beginning offset, + end offset as well as the offset matching a timestamp in buckets.
        +
        +
        Parameters:
        +
        tablePath - the table path of the table.
        +
        buckets - the buckets to fetch offset.
        +
        offsetSpec - the offset spec to fetch.
        +
        +
      • +
      + + + +
        +
      • +

        listOffsets

        +
        ListOffsetsResult listOffsets​(org.apache.fluss.metadata.TablePath tablePath,
        +                              String partitionName,
        +                              Collection<Integer> buckets,
        +                              OffsetSpec offsetSpec)
        +
        List offset for the specified buckets. This operation enables to find the beginning offset, + end offset as well as the offset matching a timestamp in buckets.
        +
        +
        Parameters:
        +
        tablePath - the table path of the table.
        +
        partitionName - the partition name of the partition,see ResolvedPartitionSpec.getPartitionName() * for the format of the partition name.
        +
        buckets - the buckets to fetch offset.
        +
        offsetSpec - the offset spec to fetch.
        +
        +
      • +
      + + + +
        +
      • +

        listAcls

        +
        CompletableFuture<Collection<org.apache.fluss.security.acl.AclBinding>> listAcls​(org.apache.fluss.security.acl.AclBindingFilter aclBindingFilter)
        +
        Retrieves ACL entries filtered by principal for the specified resource. + +

        1. Validates the user has 'describe' permission on the resource. 2. Returns entries + matching the principal if permitted; throws an exception otherwise.

        +
        +
        Returns:
        +
        A CompletableFuture containing the filtered ACL entries.
        +
        +
      • +
      + + + +
        +
      • +

        createAcls

        +
        CreateAclsResult createAcls​(Collection<org.apache.fluss.security.acl.AclBinding> aclBindings)
        +
        Creates multiple ACL entries in a single atomic operation. + +

        1. Validates the user has 'alter' permission on the resource. 2. Creates the ACL entries + if valid and permitted. + +

        Each entry in aclBindings must have a valid principal, operation and permission.

        +
        +
        Parameters:
        +
        aclBindings - List of ACL entries to create.
        +
        Returns:
        +
        A CompletableFuture indicating completion of the operation.
        +
        +
      • +
      + + + +
        +
      • +

        dropAcls

        +
        DropAclsResult dropAcls​(Collection<org.apache.fluss.security.acl.AclBindingFilter> filters)
        +
        Removes multiple ACL entries in a single atomic operation. + +

        1. Validates the user has 'alter' permission on the resource. 2. Removes entries only if + they exactly match the provided entries (principal, operation, permission). 3. Does not + remove entries if any of the ACL entries do not exist.

        +
        +
        Parameters:
        +
        filters - List of ACL entries to remove.
        +
        Returns:
        +
        A CompletableFuture indicating completion of the operation.
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/CreateAclsResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/CreateAclsResult.html new file mode 100644 index 0000000000..ed5c10e6c1 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/CreateAclsResult.html @@ -0,0 +1,385 @@ + + + + + +CreateAclsResult (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class CreateAclsResult

+
+
+ +
+
    +
  • +
    +
    public class CreateAclsResult
    +extends Object
    +
    Represents the result of a batch ACL operation, managing asynchronous completion of individual + ACL operations. + +

    This class tracks the execution status of multiple ACL operations (e.g., create/drop) by + associating each AclBinding with its corresponding CompletableFuture. It + processes RPC responses to complete or fail individual futures based on server-side results.

    +
    +
    Since:
    +
    0.6
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        CreateAclsResult

        +
        public CreateAclsResult​(Collection<org.apache.fluss.security.acl.AclBinding> aclBindings)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + + + + + +
        +
      • +

        getFutures

        +
        public Map<org.apache.fluss.security.acl.AclBinding,​CompletableFuture<Void>> getFutures()
        +
        Gets the map of ACL bindings to their associated futures.
        +
        +
        Returns:
        +
        The map of ACL bindings to futures.
        +
        +
      • +
      + + + +
        +
      • +

        complete

        +
        public void complete​(List<org.apache.fluss.rpc.messages.PbCreateAclRespInfo> pbAclRespInfos)
        +
        Completes individual futures based on RPC response information. + +

        For each PbCreateAclRespInfo in the collection, Completes the future with success + or failure based on the response's error code.

        +
        +
        Parameters:
        +
        pbAclRespInfos - Collection of protobuf response messages containing ACL operation + results.
        +
        +
      • +
      + + + +
        +
      • +

        completeExceptionally

        +
        public void completeExceptionally​(Throwable t)
        +
        Marks all futures as exceptionally completed with the provided throwable. + +

        This method propagates a common exception (e.g., network error) to all tracked futures.

        +
        +
        Parameters:
        +
        t - The throwable to propagate to all futures
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.FilterResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.FilterResult.html new file mode 100644 index 0000000000..515fe92f51 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.FilterResult.html @@ -0,0 +1,292 @@ + + + + + +DropAclsResult.FilterResult (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class DropAclsResult.FilterResult

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.admin.DropAclsResult.FilterResult
    • +
    +
  • +
+
+
    +
  • +
    +
    Enclosing class:
    +
    DropAclsResult
    +
    +
    +
    public static class DropAclsResult.FilterResult
    +extends Object
    +
    A class containing either the deleted ACL binding or an exception if the delete failed.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        binding

        +
        public org.apache.fluss.security.acl.AclBinding binding()
        +
        Return the deleted ACL binding or null if there was an error.
        +
      • +
      + + + +
        +
      • +

        exception

        +
        @Nullable
        +public org.apache.fluss.exception.ApiException exception()
        +
        Return an exception if the ACL delete was not successful or null if it was.
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.FilterResults.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.FilterResults.html new file mode 100644 index 0000000000..cea8bc22d2 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.FilterResults.html @@ -0,0 +1,274 @@ + + + + + +DropAclsResult.FilterResults (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class DropAclsResult.FilterResults

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.admin.DropAclsResult.FilterResults
    • +
    +
  • +
+
+
    +
  • +
    +
    Enclosing class:
    +
    DropAclsResult
    +
    +
    +
    public static class DropAclsResult.FilterResults
    +extends Object
    +
    A class containing the results of the delete ACLs operation.
    +
  • +
+
+
+ +
+
+ +
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.html new file mode 100644 index 0000000000..56535cd81a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.html @@ -0,0 +1,352 @@ + + + + + +DropAclsResult (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class DropAclsResult

+
+
+ +
+
    +
  • +
    +
    public class DropAclsResult
    +extends Object
    +
    Represents the result of a drop ACLs operation.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        all

        +
        public CompletableFuture<Collection<org.apache.fluss.security.acl.AclBinding>> all()
        +
        Return a future which succeeds only if all the ACLs deletions succeed, and which contains all + the deleted ACLs. Note that it if the filters don't match any ACLs, this is not considered an + error.
        +
      • +
      + + + +
        +
      • +

        completeExceptionally

        +
        public void completeExceptionally​(Throwable t)
        +
      • +
      + + + +
        +
      • +

        complete

        +
        public void complete​(List<org.apache.fluss.rpc.messages.PbDropAclsFilterResult> results)
        +
      • +
      + + + +
        +
      • +

        values

        +
        public Map<org.apache.fluss.security.acl.AclBindingFilter,​CompletableFuture<DropAclsResult.FilterResults>> values()
        +
        Return a map from acl filters to futures which can be used to check the status of the + deletions by each filter.
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/FlussAdmin.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/FlussAdmin.html new file mode 100644 index 0000000000..14af9dc59b --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/FlussAdmin.html @@ -0,0 +1,1173 @@ + + + + + +FlussAdmin (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlussAdmin

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    AutoCloseable, Admin
    +
    +
    +
    public class FlussAdmin
    +extends Object
    +implements Admin
    +
    The default implementation of Admin. + +

    This class is thread-safe. The API of this class is evolving, see Admin for details.

    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlussAdmin

        +
        public FlussAdmin​(org.apache.fluss.rpc.RpcClient client,
        +                  MetadataUpdater metadataUpdater)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getServerNodes

        +
        public CompletableFuture<List<org.apache.fluss.cluster.ServerNode>> getServerNodes()
        +
        Description copied from interface: Admin
        +
        Get the current server node information. asynchronously.
        +
        +
        Specified by:
        +
        getServerNodes in interface Admin
        +
        +
      • +
      + + + +
        +
      • +

        getTableSchema

        +
        public CompletableFuture<org.apache.fluss.metadata.SchemaInfo> getTableSchema​(org.apache.fluss.metadata.TablePath tablePath)
        +
        Description copied from interface: Admin
        +
        Get the latest table schema of the given table asynchronously. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • TableNotExistException if the table does not exist. +
        +
        +
        Specified by:
        +
        getTableSchema in interface Admin
        +
        Parameters:
        +
        tablePath - the table path of the table.
        +
        +
      • +
      + + + +
        +
      • +

        getTableSchema

        +
        public CompletableFuture<org.apache.fluss.metadata.SchemaInfo> getTableSchema​(org.apache.fluss.metadata.TablePath tablePath,
        +                                                                              int schemaId)
        +
        Description copied from interface: Admin
        +
        Get the specific table schema of the given table by schema id asynchronously. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • TableNotExistException if the table does not exist. +
        • SchemaNotExistException if the schema does not exist. +
        +
        +
        Specified by:
        +
        getTableSchema in interface Admin
        +
        Parameters:
        +
        tablePath - the table path of the table.
        +
        +
      • +
      + + + +
        +
      • +

        createDatabase

        +
        public CompletableFuture<Void> createDatabase​(String databaseName,
        +                                              org.apache.fluss.metadata.DatabaseDescriptor databaseDescriptor,
        +                                              boolean ignoreIfExists)
        +
        Description copied from interface: Admin
        +
        Create a new database asynchronously. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • DatabaseAlreadyExistException if the database already exists and + ignoreIfExists is false. +
        +
        +
        Specified by:
        +
        createDatabase in interface Admin
        +
        Parameters:
        +
        databaseName - The name of the database to create.
        +
        databaseDescriptor - The descriptor of the database to create.
        +
        ignoreIfExists - Flag to specify behavior when a database with the given name already + exists: if set to false, throw a DatabaseAlreadyExistException, if set to true, do + nothing.
        +
        +
      • +
      + + + +
        +
      • +

        getDatabaseInfo

        +
        public CompletableFuture<org.apache.fluss.metadata.DatabaseInfo> getDatabaseInfo​(String databaseName)
        +
        Description copied from interface: Admin
        +
        Get the database with the given database name asynchronously. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • DatabaseNotExistException if the database does not exist. +
        +
        +
        Specified by:
        +
        getDatabaseInfo in interface Admin
        +
        Parameters:
        +
        databaseName - The database name of the database.
        +
        +
      • +
      + + + +
        +
      • +

        dropDatabase

        +
        public CompletableFuture<Void> dropDatabase​(String databaseName,
        +                                            boolean ignoreIfNotExists,
        +                                            boolean cascade)
        +
        Description copied from interface: Admin
        +
        Drop the database with the given name asynchronously. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • DatabaseNotExistException if the database does not exist and + ignoreIfNotExists is false. +
        • DatabaseNotEmptyException if the database is not empty and cascade is + false. +
        +
        +
        Specified by:
        +
        dropDatabase in interface Admin
        +
        Parameters:
        +
        databaseName - The name of the database to delete.
        +
        ignoreIfNotExists - Flag to specify behavior when a database with the given name does + not exist: if set to false, throw a DatabaseNotExistException, if set to true, do + nothing.
        +
        cascade - Flag to specify whether to delete all tables in the database.
        +
        +
      • +
      + + + +
        +
      • +

        databaseExists

        +
        public CompletableFuture<Boolean> databaseExists​(String databaseName)
        +
        Description copied from interface: Admin
        +
        Get whether database exists asynchronously.
        +
        +
        Specified by:
        +
        databaseExists in interface Admin
        +
        Parameters:
        +
        databaseName - The name of the database to check.
        +
        +
      • +
      + + + + + + + +
        +
      • +

        createTable

        +
        public CompletableFuture<Void> createTable​(org.apache.fluss.metadata.TablePath tablePath,
        +                                           org.apache.fluss.metadata.TableDescriptor tableDescriptor,
        +                                           boolean ignoreIfExists)
        +
        Description copied from interface: Admin
        +
        Create a new table asynchronously. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • DatabaseNotExistException if the database in the table path does not exist. +
        • TableAlreadyExistException if the table already exists and + ignoreIfExists is false. +
        • InvalidReplicationFactorException if the table's replication factor is larger + than the number of available tablet servers. +
        +
        +
        Specified by:
        +
        createTable in interface Admin
        +
        Parameters:
        +
        tablePath - The tablePath of the table.
        +
        tableDescriptor - The table to create.
        +
        +
      • +
      + + + +
        +
      • +

        getTableInfo

        +
        public CompletableFuture<org.apache.fluss.metadata.TableInfo> getTableInfo​(org.apache.fluss.metadata.TablePath tablePath)
        +
        Description copied from interface: Admin
        +
        Get the table with the given table path asynchronously. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • TableNotExistException if the table does not exist. +
        +
        +
        Specified by:
        +
        getTableInfo in interface Admin
        +
        Parameters:
        +
        tablePath - The table path of the table.
        +
        +
      • +
      + + + +
        +
      • +

        dropTable

        +
        public CompletableFuture<Void> dropTable​(org.apache.fluss.metadata.TablePath tablePath,
        +                                         boolean ignoreIfNotExists)
        +
        Description copied from interface: Admin
        +
        Drop the table with the given table path asynchronously. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • TableNotExistException if the table does not exist and + ignoreIfNotExists is false. +
        +
        +
        Specified by:
        +
        dropTable in interface Admin
        +
        Parameters:
        +
        tablePath - The table path of the table.
        +
        ignoreIfNotExists - Flag to specify behavior when a table with the given name does not + exist: if set to false, throw a TableNotExistException, if set to true, do nothing.
        +
        +
      • +
      + + + +
        +
      • +

        tableExists

        +
        public CompletableFuture<Boolean> tableExists​(org.apache.fluss.metadata.TablePath tablePath)
        +
        Description copied from interface: Admin
        +
        Get whether table exists asynchronously.
        +
        +
        Specified by:
        +
        tableExists in interface Admin
        +
        Parameters:
        +
        tablePath - The table path of the table.
        +
        +
      • +
      + + + +
        +
      • +

        listTables

        +
        public CompletableFuture<List<String>> listTables​(String databaseName)
        +
        Description copied from interface: Admin
        +
        List all tables in the given database in fluss cluster asynchronously. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • DatabaseNotExistException if the database does not exist. +
        +
        +
        Specified by:
        +
        listTables in interface Admin
        +
        Parameters:
        +
        databaseName - The name of the database.
        +
        +
      • +
      + + + +
        +
      • +

        listPartitionInfos

        +
        public CompletableFuture<List<org.apache.fluss.metadata.PartitionInfo>> listPartitionInfos​(org.apache.fluss.metadata.TablePath tablePath)
        +
        Description copied from interface: Admin
        +
        List all partitions in the given table in fluss cluster asynchronously. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • TableNotExistException if the table does not exist. +
        • TableNotPartitionedException if the table is not partitioned. +
        +
        +
        Specified by:
        +
        listPartitionInfos in interface Admin
        +
        Parameters:
        +
        tablePath - The path of the table.
        +
        +
      • +
      + + + +
        +
      • +

        listPartitionInfos

        +
        public CompletableFuture<List<org.apache.fluss.metadata.PartitionInfo>> listPartitionInfos​(org.apache.fluss.metadata.TablePath tablePath,
        +                                                                                           org.apache.fluss.metadata.PartitionSpec partitionSpec)
        +
        Description copied from interface: Admin
        +
        List all partitions in fluss cluster that are under the given table and the given partial + PartitionSpec asynchronously. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • TableNotExistException if the table does not exist. +
        • TableNotPartitionedException if the table is not partitioned. +
        • InvalidPartitionException if the input partition spec is invalid. +
        +
        +
        Specified by:
        +
        listPartitionInfos in interface Admin
        +
        Parameters:
        +
        tablePath - The path of the table.
        +
        partitionSpec - Part of table partition spec
        +
        +
      • +
      + + + +
        +
      • +

        createPartition

        +
        public CompletableFuture<Void> createPartition​(org.apache.fluss.metadata.TablePath tablePath,
        +                                               org.apache.fluss.metadata.PartitionSpec partitionSpec,
        +                                               boolean ignoreIfExists)
        +
        Description copied from interface: Admin
        +
        Create a new partition for a partitioned table. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • TableNotExistException if the table does not exist. +
        • TableNotPartitionedException if the table is not partitioned. +
        • PartitionAlreadyExistsException if the partition already exists and + ignoreIfExists is false. +
        • InvalidPartitionException if the input partition spec is invalid. +
        • TooManyPartitionsException if the number of partitions is larger than the + maximum number of partitions of one table, see ConfigOptions.MAX_PARTITION_NUM. +
        • TooManyBucketsException if the number of buckets is larger than the maximum + number of buckets of one table, see ConfigOptions.MAX_BUCKET_NUM. +
        +
        +
        Specified by:
        +
        createPartition in interface Admin
        +
        Parameters:
        +
        tablePath - The table path of the table.
        +
        partitionSpec - The partition spec to add.
        +
        ignoreIfExists - Flag to specify behavior when a partition with the given name already + exists: if set to false, throw a PartitionAlreadyExistsException, if set to true, do + nothing.
        +
        +
      • +
      + + + +
        +
      • +

        dropPartition

        +
        public CompletableFuture<Void> dropPartition​(org.apache.fluss.metadata.TablePath tablePath,
        +                                             org.apache.fluss.metadata.PartitionSpec partitionSpec,
        +                                             boolean ignoreIfNotExists)
        +
        Description copied from interface: Admin
        +
        Drop a partition from a partitioned table. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • TableNotExistException if the table does not exist. +
        • TableNotPartitionedException if the table is not partitioned. +
        • PartitionNotExistException if the partition not exists and + ignoreIfExists is false. +
        • InvalidPartitionException if the input partition spec is invalid. +
        +
        +
        Specified by:
        +
        dropPartition in interface Admin
        +
        Parameters:
        +
        tablePath - The table path of the table.
        +
        partitionSpec - The partition spec to drop.
        +
        ignoreIfNotExists - Flag to specify behavior when a partition with the given name does + not exist: if set to false, throw a PartitionNotExistException, if set to true, do + nothing.
        +
        +
      • +
      + + + +
        +
      • +

        getLatestKvSnapshots

        +
        public CompletableFuture<KvSnapshots> getLatestKvSnapshots​(org.apache.fluss.metadata.TablePath tablePath)
        +
        Description copied from interface: Admin
        +
        Get the latest kv snapshots of the given table asynchronously. A kv snapshot is a snapshot of + a bucket of a primary key table at a certain point in time. Therefore, there are at-most + N snapshots for a primary key table, N is the number of buckets. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • TableNotExistException if the table does not exist. +
        • NonPrimaryKeyTableException if the table is not a primary key table. +
        • PartitionNotExistException if the table is partitioned, use Admin.getLatestKvSnapshots(TablePath, String) instead to get the latest kv snapshot of a + partition of a partitioned table. +
        • +
        +
        +
        Specified by:
        +
        getLatestKvSnapshots in interface Admin
        +
        Parameters:
        +
        tablePath - the table path of the table.
        +
        +
      • +
      + + + +
        +
      • +

        getLatestKvSnapshots

        +
        public CompletableFuture<KvSnapshots> getLatestKvSnapshots​(org.apache.fluss.metadata.TablePath tablePath,
        +                                                           String partitionName)
        +
        Description copied from interface: Admin
        +
        Get the latest kv snapshots of the given table partition asynchronously. A kv snapshot is a + snapshot of a bucket of a primary key table at a certain point in time. Therefore, there are + at-most N snapshots for a partition of a primary key table, N is the number + of buckets. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • TableNotExistException if the table does not exist. +
        • NonPrimaryKeyTableException if the table is not a primary key table. +
        • PartitionNotExistException if the partition does not exist +
        • TableNotPartitionedException if the table is not partitioned, use Admin.getLatestKvSnapshots(TablePath) instead to get the latest kv snapshots for a + non-partitioned table. +
        +
        +
        Specified by:
        +
        getLatestKvSnapshots in interface Admin
        +
        Parameters:
        +
        tablePath - the table path of the table.
        +
        partitionName - the partition name, see ResolvedPartitionSpec.getPartitionName() + for the format of the partition name.
        +
        +
      • +
      + + + +
        +
      • +

        getKvSnapshotMetadata

        +
        public CompletableFuture<KvSnapshotMetadata> getKvSnapshotMetadata​(org.apache.fluss.metadata.TableBucket bucket,
        +                                                                   long snapshotId)
        +
        Description copied from interface: Admin
        +
        Get the kv snapshot metadata of the given kv snapshot asynchronously. The kv snapshot + metadata including the snapshot files for the kv tablet and the log offset for the changelog + at the snapshot time. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • KvSnapshotNotExistException if the snapshot does not exist. +
        +
        +
        Specified by:
        +
        getKvSnapshotMetadata in interface Admin
        +
        Parameters:
        +
        bucket - the table bucket of the kv snapshot.
        +
        snapshotId - the snapshot id.
        +
        +
      • +
      + + + +
        +
      • +

        getLatestLakeSnapshot

        +
        public CompletableFuture<LakeSnapshot> getLatestLakeSnapshot​(org.apache.fluss.metadata.TablePath tablePath)
        +
        Description copied from interface: Admin
        +
        Get table lake snapshot info of the given table asynchronously. + +

        It'll get the latest snapshot for all the buckets of the table. + +

        The following exceptions can be anticipated when calling get() on returned future. + +

          +
        • TableNotExistException if the table does not exist. +
        • LakeTableSnapshotNotExistException if no any lake snapshot exist. +
        +
        +
        Specified by:
        +
        getLatestLakeSnapshot in interface Admin
        +
        Parameters:
        +
        tablePath - the table path of the table.
        +
        +
      • +
      + + + +
        +
      • +

        listOffsets

        +
        public ListOffsetsResult listOffsets​(org.apache.fluss.metadata.TablePath tablePath,
        +                                     Collection<Integer> buckets,
        +                                     OffsetSpec offsetSpec)
        +
        Description copied from interface: Admin
        +
        List offset for the specified buckets. This operation enables to find the beginning offset, + end offset as well as the offset matching a timestamp in buckets.
        +
        +
        Specified by:
        +
        listOffsets in interface Admin
        +
        Parameters:
        +
        tablePath - the table path of the table.
        +
        buckets - the buckets to fetch offset.
        +
        offsetSpec - the offset spec to fetch.
        +
        +
      • +
      + + + +
        +
      • +

        listOffsets

        +
        public ListOffsetsResult listOffsets​(org.apache.fluss.metadata.TablePath tablePath,
        +                                     String partitionName,
        +                                     Collection<Integer> buckets,
        +                                     OffsetSpec offsetSpec)
        +
        Description copied from interface: Admin
        +
        List offset for the specified buckets. This operation enables to find the beginning offset, + end offset as well as the offset matching a timestamp in buckets.
        +
        +
        Specified by:
        +
        listOffsets in interface Admin
        +
        Parameters:
        +
        tablePath - the table path of the table.
        +
        partitionName - the partition name of the partition,see ResolvedPartitionSpec.getPartitionName() * for the format of the partition name.
        +
        buckets - the buckets to fetch offset.
        +
        offsetSpec - the offset spec to fetch.
        +
        +
      • +
      + + + +
        +
      • +

        listAcls

        +
        public CompletableFuture<Collection<org.apache.fluss.security.acl.AclBinding>> listAcls​(org.apache.fluss.security.acl.AclBindingFilter aclBindingFilter)
        +
        Description copied from interface: Admin
        +
        Retrieves ACL entries filtered by principal for the specified resource. + +

        1. Validates the user has 'describe' permission on the resource. 2. Returns entries + matching the principal if permitted; throws an exception otherwise.

        +
        +
        Specified by:
        +
        listAcls in interface Admin
        +
        Returns:
        +
        A CompletableFuture containing the filtered ACL entries.
        +
        +
      • +
      + + + +
        +
      • +

        createAcls

        +
        public CreateAclsResult createAcls​(Collection<org.apache.fluss.security.acl.AclBinding> aclBindings)
        +
        Description copied from interface: Admin
        +
        Creates multiple ACL entries in a single atomic operation. + +

        1. Validates the user has 'alter' permission on the resource. 2. Creates the ACL entries + if valid and permitted. + +

        Each entry in aclBindings must have a valid principal, operation and permission.

        +
        +
        Specified by:
        +
        createAcls in interface Admin
        +
        Parameters:
        +
        aclBindings - List of ACL entries to create.
        +
        Returns:
        +
        A CompletableFuture indicating completion of the operation.
        +
        +
      • +
      + + + +
        +
      • +

        dropAcls

        +
        public DropAclsResult dropAcls​(Collection<org.apache.fluss.security.acl.AclBindingFilter> filters)
        +
        Description copied from interface: Admin
        +
        Removes multiple ACL entries in a single atomic operation. + +

        1. Validates the user has 'alter' permission on the resource. 2. Removes entries only if + they exactly match the provided entries (principal, operation, permission). 3. Does not + remove entries if any of the ACL entries do not exist.

        +
        +
        Specified by:
        +
        dropAcls in interface Admin
        +
        Parameters:
        +
        filters - List of ACL entries to remove.
        +
        Returns:
        +
        A CompletableFuture indicating completion of the operation.
        +
        +
      • +
      + + + +
        +
      • +

        close

        +
        public void close()
        +
        +
        Specified by:
        +
        close in interface AutoCloseable
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/ListOffsetsResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/ListOffsetsResult.html new file mode 100644 index 0000000000..d1cd4527fd --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/ListOffsetsResult.html @@ -0,0 +1,326 @@ + + + + + +ListOffsetsResult (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class ListOffsetsResult

+
+
+ +
+
    +
  • +
    +
    @PublicEvolving
    +public class ListOffsetsResult
    +extends Object
    +
    Result of list offsets request.
    +
    +
    Since:
    +
    0.2
    +
    +
  • +
+
+
+ +
+
+ +
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.EarliestSpec.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.EarliestSpec.html new file mode 100644 index 0000000000..4de7b8e4f5 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.EarliestSpec.html @@ -0,0 +1,312 @@ + + + + + +OffsetSpec.EarliestSpec (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class OffsetSpec.EarliestSpec

+
+
+ +
+
    +
  • +
    +
    Enclosing class:
    +
    OffsetSpec
    +
    +
    +
    public static class OffsetSpec.EarliestSpec
    +extends OffsetSpec
    +
    Earliest offset spec.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        EarliestSpec

        +
        public EarliestSpec()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.LatestSpec.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.LatestSpec.html new file mode 100644 index 0000000000..2c28048181 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.LatestSpec.html @@ -0,0 +1,312 @@ + + + + + +OffsetSpec.LatestSpec (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class OffsetSpec.LatestSpec

+
+
+ +
+
    +
  • +
    +
    Enclosing class:
    +
    OffsetSpec
    +
    +
    +
    public static class OffsetSpec.LatestSpec
    +extends OffsetSpec
    +
    latest offset spec.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LatestSpec

        +
        public LatestSpec()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.TimestampSpec.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.TimestampSpec.html new file mode 100644 index 0000000000..ce8f828fe7 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.TimestampSpec.html @@ -0,0 +1,350 @@ + + + + + +OffsetSpec.TimestampSpec (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class OffsetSpec.TimestampSpec

+
+
+ +
+
    +
  • +
    +
    Enclosing class:
    +
    OffsetSpec
    +
    +
    +
    public static class OffsetSpec.TimestampSpec
    +extends OffsetSpec
    +
    timestamp offset spec.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        TimestampSpec

        +
        public TimestampSpec​(long timestamp)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getTimestamp

        +
        public long getTimestamp()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.html new file mode 100644 index 0000000000..209c725dad --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.html @@ -0,0 +1,399 @@ + + + + + +OffsetSpec (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class OffsetSpec

+
+
+ +
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + +
        +
      • +

        LIST_EARLIEST_OFFSET

        +
        public static final int LIST_EARLIEST_OFFSET
        +
        +
        See Also:
        +
        Constant Field Values
        +
        +
      • +
      + + + +
        +
      • +

        LIST_LATEST_OFFSET

        +
        public static final int LIST_LATEST_OFFSET
        +
        +
        See Also:
        +
        Constant Field Values
        +
        +
      • +
      + + + +
        +
      • +

        LIST_OFFSET_FROM_TIMESTAMP

        +
        public static final int LIST_OFFSET_FROM_TIMESTAMP
        +
        +
        See Also:
        +
        Constant Field Values
        +
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        OffsetSpec

        +
        public OffsetSpec()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/Admin.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/Admin.html new file mode 100644 index 0000000000..b28444501a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/Admin.html @@ -0,0 +1,413 @@ + + + + + +Uses of Interface org.apache.fluss.client.admin.Admin (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.client.admin.Admin

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/CreateAclsResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/CreateAclsResult.html new file mode 100644 index 0000000000..86a49a89a5 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/CreateAclsResult.html @@ -0,0 +1,199 @@ + + + + + +Uses of Class org.apache.fluss.client.admin.CreateAclsResult (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.admin.CreateAclsResult

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.FilterResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.FilterResult.html new file mode 100644 index 0000000000..fa3bfcc6d6 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.FilterResult.html @@ -0,0 +1,194 @@ + + + + + +Uses of Class org.apache.fluss.client.admin.DropAclsResult.FilterResult (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.admin.DropAclsResult.FilterResult

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.FilterResults.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.FilterResults.html new file mode 100644 index 0000000000..209d0e8a17 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.FilterResults.html @@ -0,0 +1,195 @@ + + + + + +Uses of Class org.apache.fluss.client.admin.DropAclsResult.FilterResults (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.admin.DropAclsResult.FilterResults

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.html new file mode 100644 index 0000000000..69f1a8dd38 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.html @@ -0,0 +1,199 @@ + + + + + +Uses of Class org.apache.fluss.client.admin.DropAclsResult (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.admin.DropAclsResult

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/FlussAdmin.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/FlussAdmin.html new file mode 100644 index 0000000000..9cf453a2cf --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/FlussAdmin.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.admin.FlussAdmin (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.admin.FlussAdmin

+
+
No usage of org.apache.fluss.client.admin.FlussAdmin
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/ListOffsetsResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/ListOffsetsResult.html new file mode 100644 index 0000000000..a91a047c1d --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/ListOffsetsResult.html @@ -0,0 +1,221 @@ + + + + + +Uses of Class org.apache.fluss.client.admin.ListOffsetsResult (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.admin.ListOffsetsResult

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.EarliestSpec.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.EarliestSpec.html new file mode 100644 index 0000000000..aaa11f8ee2 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.EarliestSpec.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.admin.OffsetSpec.EarliestSpec (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.admin.OffsetSpec.EarliestSpec

+
+
No usage of org.apache.fluss.client.admin.OffsetSpec.EarliestSpec
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.LatestSpec.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.LatestSpec.html new file mode 100644 index 0000000000..dc1bb528dd --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.LatestSpec.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.admin.OffsetSpec.LatestSpec (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.admin.OffsetSpec.LatestSpec

+
+
No usage of org.apache.fluss.client.admin.OffsetSpec.LatestSpec
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.TimestampSpec.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.TimestampSpec.html new file mode 100644 index 0000000000..73c30490d0 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.TimestampSpec.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.admin.OffsetSpec.TimestampSpec (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.admin.OffsetSpec.TimestampSpec

+
+
No usage of org.apache.fluss.client.admin.OffsetSpec.TimestampSpec
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.html new file mode 100644 index 0000000000..158176b94c --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.html @@ -0,0 +1,281 @@ + + + + + +Uses of Class org.apache.fluss.client.admin.OffsetSpec (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.admin.OffsetSpec

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-summary.html new file mode 100644 index 0000000000..1e2f9fbf88 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-summary.html @@ -0,0 +1,241 @@ + + + + + +org.apache.fluss.client.admin (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.client.admin

+
+
+
    +
  • + + + + + + + + + + + + +
    Interface Summary 
    InterfaceDescription
    Admin +
    The administrative client for Fluss, which supports managing and inspecting tables, servers, + configurations and ACLs.
    +
    +
  • +
  • + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Class Summary 
    ClassDescription
    CreateAclsResult +
    Represents the result of a batch ACL operation, managing asynchronous completion of individual + ACL operations.
    +
    DropAclsResult +
    Represents the result of a drop ACLs operation.
    +
    DropAclsResult.FilterResult +
    A class containing either the deleted ACL binding or an exception if the delete failed.
    +
    DropAclsResult.FilterResults +
    A class containing the results of the delete ACLs operation.
    +
    FlussAdmin +
    The default implementation of Admin.
    +
    ListOffsetsResult +
    Result of list offsets request.
    +
    OffsetSpec +
    Offset spec.
    +
    OffsetSpec.EarliestSpec +
    Earliest offset spec.
    +
    OffsetSpec.LatestSpec +
    latest offset spec.
    +
    OffsetSpec.TimestampSpec +
    timestamp offset spec.
    +
    +
  • +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-tree.html new file mode 100644 index 0000000000..25cf96694a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-tree.html @@ -0,0 +1,185 @@ + + + + + +org.apache.fluss.client.admin Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.client.admin

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+

Interface Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-use.html new file mode 100644 index 0000000000..039ff4fe58 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-use.html @@ -0,0 +1,414 @@ + + + + + +Uses of Package org.apache.fluss.client.admin (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.client.admin

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/Connection.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/Connection.html new file mode 100644 index 0000000000..9885f40391 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/Connection.html @@ -0,0 +1,246 @@ + + + + + +Uses of Interface org.apache.fluss.client.Connection (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.client.Connection

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/ConnectionFactory.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/ConnectionFactory.html new file mode 100644 index 0000000000..be7483606e --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/ConnectionFactory.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.ConnectionFactory (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.ConnectionFactory

+
+
No usage of org.apache.fluss.client.ConnectionFactory
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/FlussConnection.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/FlussConnection.html new file mode 100644 index 0000000000..103483b660 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/FlussConnection.html @@ -0,0 +1,217 @@ + + + + + +Uses of Class org.apache.fluss.client.FlussConnection (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.FlussConnection

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/AbstractLookupBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/AbstractLookupBatch.html new file mode 100644 index 0000000000..1b5c6204ad --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/AbstractLookupBatch.html @@ -0,0 +1,412 @@ + + + + + +AbstractLookupBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class AbstractLookupBatch<T>

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.lookup.AbstractLookupBatch<T>
    • +
    +
  • +
+
+
    +
  • +
    +
    @Internal
    +public abstract class AbstractLookupBatch<T>
    +extends Object
    +
    An abstract lookup batch.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    + +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        AbstractLookupBatch

        +
        public AbstractLookupBatch​(org.apache.fluss.metadata.TableBucket tableBucket)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        complete

        +
        public abstract void complete​(List<T> values)
        +
        Complete the lookup operations using given values .
        +
      • +
      + + + + + + + + + + + +
        +
      • +

        tableBucket

        +
        public org.apache.fluss.metadata.TableBucket tableBucket()
        +
      • +
      + + + +
        +
      • +

        completeExceptionally

        +
        public void completeExceptionally​(Exception exception)
        +
        Complete the get operations with given exception.
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/AbstractLookupQuery.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/AbstractLookupQuery.html new file mode 100644 index 0000000000..cdb368599e --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/AbstractLookupQuery.html @@ -0,0 +1,356 @@ + + + + + +AbstractLookupQuery (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class AbstractLookupQuery<T>

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.lookup.AbstractLookupQuery<T>
    • +
    +
  • +
+
+
    +
  • +
    +
    Direct Known Subclasses:
    +
    LookupQuery, PrefixLookupQuery
    +
    +
    +
    @Internal
    +public abstract class AbstractLookupQuery<T>
    +extends Object
    +
    Abstract Class to represent a lookup operation.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        AbstractLookupQuery

        +
        public AbstractLookupQuery​(org.apache.fluss.metadata.TableBucket tableBucket,
        +                           byte[] key)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        key

        +
        public byte[] key()
        +
      • +
      + + + +
        +
      • +

        tableBucket

        +
        public org.apache.fluss.metadata.TableBucket tableBucket()
        +
      • +
      + + + +
        +
      • +

        lookupType

        +
        public abstract LookupType lookupType()
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/Lookup.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/Lookup.html new file mode 100644 index 0000000000..d8a9c3bf51 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/Lookup.html @@ -0,0 +1,354 @@ + + + + + +Lookup (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface Lookup

+
+
+
+
    +
  • +
    +
    All Known Implementing Classes:
    +
    TableLookup
    +
    +
    +
    @PublicEvolving
    +public interface Lookup
    +
    Used to configure and create a Lookuper to lookup rows of a primary key table. The built + Lookuper can be a primary key lookuper that lookups by the primary key, or a prefix key lookup + that lookups by the prefix key of the primary key. + +

    Lookup objects are immutable and can be shared between threads. Refinement methods, + like lookupBy(java.util.List<java.lang.String>), create new Lookup instances. + +

    Example1: Create a Primary Key Lookuper. Given a table with primary key column [k STRING]. + +

    
    + Lookuper lookuper = table.newLookup().createLookuper();
    + CompletableFuture<LookupResult> resultFuture = lookuper.lookup(GenericRow.of("key1"));
    + resultFuture.get().getRows().forEach(row -> {
    +    System.out.println(row);
    + });
    + 
    + +

    Example2: Create a Prefix Key Lookuper. Given a table with primary key column [a INT, b + STRING, c BIGINT] and bucket key [a, b]. + +

    
    + Lookuper lookuper = table.newLookup().lookupBy("a", "b").createLookuper();
    + CompletableFuture<LookupResult> resultFuture = lookuper.lookup(GenericRow.of(1, "b1"));
    + resultFuture.get().getRows().forEach(row -> {
    +   System.out.println(row);
    + });
    + 
    +
    +
    Since:
    +
    0.6
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        lookupBy

        +
        Lookup lookupBy​(List<String> lookupColumnNames)
        +
        Returns a new Lookup instance with the given lookup columns. By default, the lookup columns + are the primary key columns of the table. The specified columns must be a prefix subset of + the physical primary key columns (i.e., the primary key columns exclude partition columns). + +

        Note: Currently, if you want to lookup with specified lookup columns (i.e., prefix key + lookup), the table you created must both define the primary key and the bucket key, in + addition, the lookup columns needs to be equals with bucket key, and to be a part of the + primary key and must be a prefix of the primary key. For example, if a table has fields + [a,b,c,d], and the primary key is set to [a, b, c], with the bucket key set to [a, b], then + the schema of the lookup columns would also be [a, b]. This pattern can create a Prefix Key + Lookuper lookup rows by prefix of the primary key. + +

        TODO: currently, the interface only support bucket key as the prefix key to lookup. + Generalize the prefix key lookup to support any prefix keys. + +

        We also support prefix key lookup for partitioned tables. The schema of the lookup columns + should contain partition fields and bucket key. In addition, the schema of the lookup columns + excluded partition fields should be a prefix of primary key excluded partition fields.

        +
        +
        Parameters:
        +
        lookupColumnNames - the specified columns to lookup by
        +
        +
      • +
      + + + +
        +
      • +

        lookupBy

        +
        default Lookup lookupBy​(String... lookupColumnNames)
        +
        +
        Parameters:
        +
        lookupColumnNames - the specified columns to lookup by
        +
        See Also:
        +
        for more details.
        +
        +
      • +
      + + + +
        +
      • +

        createLookuper

        +
        Lookuper createLookuper()
        +
        Creates a Lookuper instance to lookup rows of a primary key table by the specified + lookup columns. By default, the lookup columns are the primary key columns, but can be + changed with (lookupBy(List)) method.
        +
        +
        Returns:
        +
        the lookuper
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupBatch.html new file mode 100644 index 0000000000..06b041d1ee --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupBatch.html @@ -0,0 +1,370 @@ + + + + + +LookupBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LookupBatch

+
+
+ +
+
    +
  • +
    +
    @Internal
    +public class LookupBatch
    +extends Object
    +
    A batch that contains the lookup operations that send to same tablet bucket together.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LookupBatch

        +
        public LookupBatch​(org.apache.fluss.metadata.TableBucket tableBucket)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        addLookup

        +
        public void addLookup​(LookupQuery lookup)
        +
      • +
      + + + + + + + +
        +
      • +

        tableBucket

        +
        public org.apache.fluss.metadata.TableBucket tableBucket()
        +
      • +
      + + + +
        +
      • +

        complete

        +
        public void complete​(List<byte[]> values)
        +
        Complete the lookup operations using given values .
        +
      • +
      + + + +
        +
      • +

        completeExceptionally

        +
        public void completeExceptionally​(Exception exception)
        +
        Complete the lookup operations with given exception.
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupClient.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupClient.html new file mode 100644 index 0000000000..ff1eef4b16 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupClient.html @@ -0,0 +1,397 @@ + + + + + +LookupClient (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LookupClient

+
+
+ +
+
    +
  • +
    +
    @ThreadSafe
    +@Internal
    +public class LookupClient
    +extends Object
    +
    A client that lookups value of keys from server. + +

    The lookup client contains of a queue of pending lookup operations and background I/O threads + that is responsible for turning these lookup operations into network requests and transmitting + them to the cluster. + +

    The lookup(TableBucket, byte[]) method is asynchronous, when called, it adds the + lookup operation to a queue of pending lookup operations and immediately returns. This allows the + lookup operations to batch together individual lookup operations for efficiency.

    +
  • +
+
+
+ +
+
+
    +
  • + +
    + +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LookupClient

        +
        public LookupClient​(org.apache.fluss.config.Configuration conf,
        +                    MetadataUpdater metadataUpdater)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        lookup

        +
        public CompletableFuture<byte[]> lookup​(org.apache.fluss.metadata.TableBucket tableBucket,
        +                                        byte[] keyBytes)
        +
      • +
      + + + +
        +
      • +

        prefixLookup

        +
        public CompletableFuture<List<byte[]>> prefixLookup​(org.apache.fluss.metadata.TableBucket tableBucket,
        +                                                    byte[] keyBytes)
        +
      • +
      + + + +
        +
      • +

        close

        +
        public void close​(Duration timeout)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupQuery.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupQuery.html new file mode 100644 index 0000000000..4fc6064837 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupQuery.html @@ -0,0 +1,303 @@ + + + + + +LookupQuery (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LookupQuery

+
+
+ +
+
    +
  • +
    +
    @Internal
    +public class LookupQuery
    +extends AbstractLookupQuery<byte[]>
    +
    Class to represent a Lookup operation, it contains the table bucket that the key should lookup + from, the bytes of the key, and a future for the lookup operation.
    +
  • +
+
+
+ +
+
+ +
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupResult.html new file mode 100644 index 0000000000..489cc5438f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupResult.html @@ -0,0 +1,395 @@ + + + + + +LookupResult (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LookupResult

+
+
+ +
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LookupResult

        +
        public LookupResult​(@Nullable
        +                    org.apache.fluss.row.InternalRow row)
        +
      • +
      + + + +
        +
      • +

        LookupResult

        +
        public LookupResult​(List<org.apache.fluss.row.InternalRow> rowList)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getRowList

        +
        public List<org.apache.fluss.row.InternalRow> getRowList()
        +
      • +
      + + + +
        +
      • +

        getSingletonRow

        +
        @Nullable
        +public org.apache.fluss.row.InternalRow getSingletonRow()
        +
      • +
      + + + +
        +
      • +

        equals

        +
        public boolean equals​(Object o)
        +
        +
        Overrides:
        +
        equals in class Object
        +
        +
      • +
      + + + +
        +
      • +

        hashCode

        +
        public int hashCode()
        +
        +
        Overrides:
        +
        hashCode in class Object
        +
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupType.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupType.html new file mode 100644 index 0000000000..618eee9430 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupType.html @@ -0,0 +1,380 @@ + + + + + +LookupType (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Enum LookupType

+
+
+ +
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Enum Constant Detail

      + + + +
        +
      • +

        LOOKUP

        +
        public static final LookupType LOOKUP
        +
      • +
      + + + +
        +
      • +

        PREFIX_LOOKUP

        +
        public static final LookupType PREFIX_LOOKUP
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        values

        +
        public static LookupType[] values()
        +
        Returns an array containing the constants of this enum type, in +the order they are declared. This method may be used to iterate +over the constants as follows: +
        +for (LookupType c : LookupType.values())
        +    System.out.println(c);
        +
        +
        +
        Returns:
        +
        an array containing the constants of this enum type, in the order they are declared
        +
        +
      • +
      + + + +
        +
      • +

        valueOf

        +
        public static LookupType valueOf​(String name)
        +
        Returns the enum constant of this type with the specified name. +The string must match exactly an identifier used to declare an +enum constant in this type. (Extraneous whitespace characters are +not permitted.)
        +
        +
        Parameters:
        +
        name - the name of the enum constant to be returned.
        +
        Returns:
        +
        the enum constant with the specified name
        +
        Throws:
        +
        IllegalArgumentException - if this enum type has no constant with the specified name
        +
        NullPointerException - if the argument is null
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/Lookuper.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/Lookuper.html new file mode 100644 index 0000000000..e1055c8b17 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/Lookuper.html @@ -0,0 +1,269 @@ + + + + + +Lookuper (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface Lookuper

+
+
+
+
    +
  • +
    +
    @PublicEvolving
    +public interface Lookuper
    +
    The lookup-er is used to lookup row of a primary key table by primary key or prefix key.
    +
    +
    Since:
    +
    0.6
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        lookup

        +
        CompletableFuture<LookupResult> lookup​(org.apache.fluss.row.InternalRow lookupKey)
        +
        Lookups certain row from the given lookup key. + +

        The lookup key must be a primary key if the lookuper is a Primary Key Lookuper (created by + table.newLookup().createLookuper()), or be the prefix key if the lookuper is a Prefix + Key Lookuper (created by table.newLookup().lookupBy(prefixKeys).createLookuper()).

        +
        +
        Parameters:
        +
        lookupKey - the lookup key.
        +
        Returns:
        +
        the result of lookup.
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookup.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookup.html new file mode 100644 index 0000000000..1446848f55 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookup.html @@ -0,0 +1,312 @@ + + + + + +PrefixLookup (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class PrefixLookup

+
+
+ +
+
    +
  • +
    +
    @PublicEvolving
    +public class PrefixLookup
    +extends Object
    +
    Used to describe the operation to prefix lookup by PrefixLookuper to a primary key table.
    +
    +
    Since:
    +
    0.6
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        PrefixLookup

        +
        public PrefixLookup​(List<String> lookupColumnNames)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getLookupColumnNames

        +
        public List<String> getLookupColumnNames()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookupBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookupBatch.html new file mode 100644 index 0000000000..6bcda800e8 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookupBatch.html @@ -0,0 +1,368 @@ + + + + + +PrefixLookupBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class PrefixLookupBatch

+
+
+ +
+
    +
  • +
    +
    @Internal
    +public class PrefixLookupBatch
    +extends Object
    +
    A batch that contains the prefix lookup operations that send to same destination and some table + together.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        PrefixLookupBatch

        +
        public PrefixLookupBatch​(org.apache.fluss.metadata.TableBucket tableBucket)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + + + + + + + + + +
        +
      • +

        tableBucket

        +
        public org.apache.fluss.metadata.TableBucket tableBucket()
        +
      • +
      + + + +
        +
      • +

        complete

        +
        public void complete​(List<List<byte[]>> values)
        +
      • +
      + + + +
        +
      • +

        completeExceptionally

        +
        public void completeExceptionally​(Exception exception)
        +
        Complete the get operations with given exception.
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookupQuery.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookupQuery.html new file mode 100644 index 0000000000..67db91fbca --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookupQuery.html @@ -0,0 +1,303 @@ + + + + + +PrefixLookupQuery (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class PrefixLookupQuery

+
+
+ +
+
    +
  • +
    +
    @Internal
    +public class PrefixLookupQuery
    +extends AbstractLookupQuery<List<byte[]>>
    +
    Class to represent a prefix lookup operation, it contains the table id, bucketNums and related + CompletableFuture.
    +
  • +
+
+
+ +
+
+ +
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/TableLookup.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/TableLookup.html new file mode 100644 index 0000000000..8b0c7a3a23 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/TableLookup.html @@ -0,0 +1,377 @@ + + + + + +TableLookup (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TableLookup

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    Lookup
    +
    +
    +
    public class TableLookup
    +extends Object
    +implements Lookup
    +
    API for configuring and creating Lookuper.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        TableLookup

        +
        public TableLookup​(org.apache.fluss.metadata.TableInfo tableInfo,
        +                   MetadataUpdater metadataUpdater,
        +                   LookupClient lookupClient)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        lookupBy

        +
        public Lookup lookupBy​(List<String> lookupColumnNames)
        +
        Description copied from interface: Lookup
        +
        Returns a new Lookup instance with the given lookup columns. By default, the lookup columns + are the primary key columns of the table. The specified columns must be a prefix subset of + the physical primary key columns (i.e., the primary key columns exclude partition columns). + +

        Note: Currently, if you want to lookup with specified lookup columns (i.e., prefix key + lookup), the table you created must both define the primary key and the bucket key, in + addition, the lookup columns needs to be equals with bucket key, and to be a part of the + primary key and must be a prefix of the primary key. For example, if a table has fields + [a,b,c,d], and the primary key is set to [a, b, c], with the bucket key set to [a, b], then + the schema of the lookup columns would also be [a, b]. This pattern can create a Prefix Key + Lookuper lookup rows by prefix of the primary key. + +

        TODO: currently, the interface only support bucket key as the prefix key to lookup. + Generalize the prefix key lookup to support any prefix keys. + +

        We also support prefix key lookup for partitioned tables. The schema of the lookup columns + should contain partition fields and bucket key. In addition, the schema of the lookup columns + excluded partition fields should be a prefix of primary key excluded partition fields.

        +
        +
        Specified by:
        +
        lookupBy in interface Lookup
        +
        Parameters:
        +
        lookupColumnNames - the specified columns to lookup by
        +
        +
      • +
      + + + +
        +
      • +

        createLookuper

        +
        public Lookuper createLookuper()
        +
        Description copied from interface: Lookup
        +
        Creates a Lookuper instance to lookup rows of a primary key table by the specified + lookup columns. By default, the lookup columns are the primary key columns, but can be + changed with (Lookup.lookupBy(List)) method.
        +
        +
        Specified by:
        +
        createLookuper in interface Lookup
        +
        Returns:
        +
        the lookuper
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/AbstractLookupBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/AbstractLookupBatch.html new file mode 100644 index 0000000000..30dbc32a46 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/AbstractLookupBatch.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.lookup.AbstractLookupBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.lookup.AbstractLookupBatch

+
+
No usage of org.apache.fluss.client.lookup.AbstractLookupBatch
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/AbstractLookupQuery.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/AbstractLookupQuery.html new file mode 100644 index 0000000000..400c0c9cd2 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/AbstractLookupQuery.html @@ -0,0 +1,248 @@ + + + + + +Uses of Class org.apache.fluss.client.lookup.AbstractLookupQuery (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.lookup.AbstractLookupQuery

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/Lookup.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/Lookup.html new file mode 100644 index 0000000000..dfb691b6a3 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/Lookup.html @@ -0,0 +1,255 @@ + + + + + +Uses of Interface org.apache.fluss.client.lookup.Lookup (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.client.lookup.Lookup

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupBatch.html new file mode 100644 index 0000000000..948d2ce7a1 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupBatch.html @@ -0,0 +1,193 @@ + + + + + +Uses of Class org.apache.fluss.client.lookup.LookupBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.lookup.LookupBatch

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupClient.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupClient.html new file mode 100644 index 0000000000..4b508f55ba --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupClient.html @@ -0,0 +1,218 @@ + + + + + +Uses of Class org.apache.fluss.client.lookup.LookupClient (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.lookup.LookupClient

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupQuery.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupQuery.html new file mode 100644 index 0000000000..20b01d6de4 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupQuery.html @@ -0,0 +1,207 @@ + + + + + +Uses of Class org.apache.fluss.client.lookup.LookupQuery (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.lookup.LookupQuery

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupResult.html new file mode 100644 index 0000000000..dd71ef8ac5 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupResult.html @@ -0,0 +1,194 @@ + + + + + +Uses of Class org.apache.fluss.client.lookup.LookupResult (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.lookup.LookupResult

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupType.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupType.html new file mode 100644 index 0000000000..75301e71b9 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupType.html @@ -0,0 +1,246 @@ + + + + + +Uses of Class org.apache.fluss.client.lookup.LookupType (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.lookup.LookupType

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/Lookuper.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/Lookuper.html new file mode 100644 index 0000000000..a70953e7f2 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/Lookuper.html @@ -0,0 +1,200 @@ + + + + + +Uses of Interface org.apache.fluss.client.lookup.Lookuper (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.client.lookup.Lookuper

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookup.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookup.html new file mode 100644 index 0000000000..e57351cf06 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookup.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.lookup.PrefixLookup (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.lookup.PrefixLookup

+
+
No usage of org.apache.fluss.client.lookup.PrefixLookup
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookupBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookupBatch.html new file mode 100644 index 0000000000..dee71159a8 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookupBatch.html @@ -0,0 +1,193 @@ + + + + + +Uses of Class org.apache.fluss.client.lookup.PrefixLookupBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.lookup.PrefixLookupBatch

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookupQuery.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookupQuery.html new file mode 100644 index 0000000000..821ae6a97b --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookupQuery.html @@ -0,0 +1,207 @@ + + + + + +Uses of Class org.apache.fluss.client.lookup.PrefixLookupQuery (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.lookup.PrefixLookupQuery

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/TableLookup.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/TableLookup.html new file mode 100644 index 0000000000..64aa9eb0ec --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/TableLookup.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.lookup.TableLookup (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.lookup.TableLookup

+
+
No usage of org.apache.fluss.client.lookup.TableLookup
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-summary.html new file mode 100644 index 0000000000..3f3e4d96dd --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-summary.html @@ -0,0 +1,265 @@ + + + + + +org.apache.fluss.client.lookup (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.client.lookup

+
+
+
    +
  • + + + + + + + + + + + + + + + + +
    Interface Summary 
    InterfaceDescription
    Lookup +
    Used to configure and create a Lookuper to lookup rows of a primary key table.
    +
    Lookuper +
    The lookup-er is used to lookup row of a primary key table by primary key or prefix key.
    +
    +
  • +
  • + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Class Summary 
    ClassDescription
    AbstractLookupBatch<T> +
    An abstract lookup batch.
    +
    AbstractLookupQuery<T> +
    Abstract Class to represent a lookup operation.
    +
    LookupBatch +
    A batch that contains the lookup operations that send to same tablet bucket together.
    +
    LookupClient +
    A client that lookups value of keys from server.
    +
    LookupQuery +
    Class to represent a Lookup operation, it contains the table bucket that the key should lookup + from, the bytes of the key, and a future for the lookup operation.
    +
    LookupResult + +
    PrefixLookup +
    Used to describe the operation to prefix lookup by PrefixLookuper to a primary key table.
    +
    PrefixLookupBatch +
    A batch that contains the prefix lookup operations that send to same destination and some table + together.
    +
    PrefixLookupQuery +
    Class to represent a prefix lookup operation, it contains the table id, bucketNums and related + CompletableFuture.
    +
    TableLookup +
    API for configuring and creating Lookuper.
    +
    +
  • +
  • + + + + + + + + + + + + +
    Enum Summary 
    EnumDescription
    LookupType +
    Enum to represent the type of lookup operation.
    +
    +
  • +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-tree.html new file mode 100644 index 0000000000..6ebb77a7af --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-tree.html @@ -0,0 +1,196 @@ + + + + + +org.apache.fluss.client.lookup Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.client.lookup

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+

Interface Hierarchy

+
    +
  • org.apache.fluss.client.lookup.Lookup
  • +
  • org.apache.fluss.client.lookup.Lookuper
  • +
+
+
+

Enum Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-use.html new file mode 100644 index 0000000000..e8b7161486 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-use.html @@ -0,0 +1,328 @@ + + + + + +Uses of Package org.apache.fluss.client.lookup (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.client.lookup

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/KvSnapshotMetadata.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/KvSnapshotMetadata.html new file mode 100644 index 0000000000..efc9ff38d7 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/KvSnapshotMetadata.html @@ -0,0 +1,354 @@ + + + + + +KvSnapshotMetadata (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class KvSnapshotMetadata

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.metadata.KvSnapshotMetadata
    • +
    +
  • +
+
+
    +
  • +
    +
    @PublicEvolving
    +public class KvSnapshotMetadata
    +extends Object
    +
    A class to represent the snapshot info of a kv tablet (bucket). It contains: + +
      +
    • The snapshot files of the bucket. +
    • The log offset corresponding to the snapshot. +
    + +

    To read the full data of a bucket, it requires to read the snapshot and the log from the + corresponding log offset.

    +
    +
    Since:
    +
    0.6
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        KvSnapshotMetadata

        +
        public KvSnapshotMetadata​(List<org.apache.fluss.fs.FsPathAndFileName> snapshotFiles,
        +                          long logOffset)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getSnapshotFiles

        +
        public List<org.apache.fluss.fs.FsPathAndFileName> getSnapshotFiles()
        +
      • +
      + + + +
        +
      • +

        getLogOffset

        +
        public long getLogOffset()
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/KvSnapshots.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/KvSnapshots.html new file mode 100644 index 0000000000..96a8da7bcc --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/KvSnapshots.html @@ -0,0 +1,385 @@ + + + + + +KvSnapshots (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class KvSnapshots

+
+
+ +
+
    +
  • +
    +
    @PublicEvolving
    +public class KvSnapshots
    +extends Object
    +
    A class representing the kv snapshots of a table or a partition. It contains multiple snapshots + for each kv tablet (bucket).
    +
    +
    Since:
    +
    0.6
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        KvSnapshots

        +
        public KvSnapshots​(long tableId,
        +                   @Nullable
        +                   Long partitionId,
        +                   Map<Integer,​Long> snapshotIds,
        +                   Map<Integer,​Long> logOffsets)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getTableId

        +
        public long getTableId()
        +
      • +
      + + + +
        +
      • +

        getPartitionId

        +
        @Nullable
        +public Long getPartitionId()
        +
      • +
      + + + +
        +
      • +

        getBucketIds

        +
        public Set<Integer> getBucketIds()
        +
      • +
      + + + +
        +
      • +

        getSnapshotId

        +
        public OptionalLong getSnapshotId​(int bucketId)
        +
        Get the latest snapshot id for this kv tablet (bucket), or empty if there are no snapshots.
        +
      • +
      + + + +
        +
      • +

        getLogOffset

        +
        public OptionalLong getLogOffset​(int bucketId)
        +
        Get the log offset to read after the snapshot for this kv tablet (bucket), or empty if there + are no snapshots.
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/LakeSnapshot.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/LakeSnapshot.html new file mode 100644 index 0000000000..7779ded118 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/LakeSnapshot.html @@ -0,0 +1,363 @@ + + + + + +LakeSnapshot (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LakeSnapshot

+
+
+ +
+
    +
  • +
    +
    @PublicEvolving
    +public class LakeSnapshot
    +extends Object
    +
    A class representing the lake snapshot information of a table. It contains: +
  • The snapshot id and the log offset for each bucket.
+
+
Since:
+
0.3
+
+ + +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LakeSnapshot

        +
        public LakeSnapshot​(long snapshotId,
        +                    Map<org.apache.fluss.metadata.TableBucket,​Long> tableBucketsOffset,
        +                    Map<Long,​String> partitionNameById)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getSnapshotId

        +
        public long getSnapshotId()
        +
      • +
      + + + +
        +
      • +

        getTableBucketsOffset

        +
        public Map<org.apache.fluss.metadata.TableBucket,​Long> getTableBucketsOffset()
        +
      • +
      + + + +
        +
      • +

        getPartitionNameById

        +
        public Map<Long,​String> getPartitionNameById()
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+ +
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/MetadataUpdater.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/MetadataUpdater.html new file mode 100644 index 0000000000..d62e985e38 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/MetadataUpdater.html @@ -0,0 +1,721 @@ + + + + + +MetadataUpdater (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class MetadataUpdater

+
+
+ +
+
    +
  • +
    +
    public class MetadataUpdater
    +extends Object
    +
    The updater to initialize and update client metadata.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + +
        +
      • +

        cluster

        +
        protected volatile org.apache.fluss.cluster.Cluster cluster
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        MetadataUpdater

        +
        public MetadataUpdater​(org.apache.fluss.config.Configuration configuration,
        +                       org.apache.fluss.rpc.RpcClient rpcClient)
        +
      • +
      + + + +
        +
      • +

        MetadataUpdater

        +
        @VisibleForTesting
        +public MetadataUpdater​(org.apache.fluss.rpc.RpcClient rpcClient,
        +                       org.apache.fluss.cluster.Cluster cluster)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getCluster

        +
        public org.apache.fluss.cluster.Cluster getCluster()
        +
      • +
      + + + +
        +
      • +

        getCoordinatorServer

        +
        @Nullable
        +public org.apache.fluss.cluster.ServerNode getCoordinatorServer()
        +
      • +
      + + + +
        +
      • +

        getTableId

        +
        public long getTableId​(org.apache.fluss.metadata.TablePath tablePath)
        +
      • +
      + + + +
        +
      • +

        getPartitionId

        +
        public Optional<Long> getPartitionId​(org.apache.fluss.metadata.PhysicalTablePath physicalTablePath)
        +
      • +
      + + + +
        +
      • +

        getPartitionIdOrElseThrow

        +
        public Long getPartitionIdOrElseThrow​(org.apache.fluss.metadata.PhysicalTablePath physicalTablePath)
        +
      • +
      + + + +
        +
      • +

        getTableInfoOrElseThrow

        +
        public org.apache.fluss.metadata.TableInfo getTableInfoOrElseThrow​(org.apache.fluss.metadata.TablePath tablePath)
        +
      • +
      + + + +
        +
      • +

        getBucketLocation

        +
        public Optional<org.apache.fluss.cluster.BucketLocation> getBucketLocation​(org.apache.fluss.metadata.TableBucket tableBucket)
        +
      • +
      + + + +
        +
      • +

        getTableInfoOrElseThrow

        +
        public org.apache.fluss.metadata.TableInfo getTableInfoOrElseThrow​(long tableId)
        +
      • +
      + + + +
        +
      • +

        leaderFor

        +
        public int leaderFor​(org.apache.fluss.metadata.TableBucket tableBucket)
        +
      • +
      + + + +
        +
      • +

        getRandomTabletServer

        +
        @Nullable
        +public org.apache.fluss.cluster.ServerNode getRandomTabletServer()
        +
      • +
      + + + +
        +
      • +

        newCoordinatorServerClient

        +
        public org.apache.fluss.rpc.gateway.CoordinatorGateway newCoordinatorServerClient()
        +
      • +
      + + + +
        +
      • +

        newRandomTabletServerClient

        +
        public org.apache.fluss.rpc.gateway.TabletServerGateway newRandomTabletServerClient()
        +
      • +
      + + + +
        +
      • +

        newTabletServerClientForNode

        +
        @Nullable
        +public org.apache.fluss.rpc.gateway.TabletServerGateway newTabletServerClientForNode​(int serverId)
        +
      • +
      + + + +
        +
      • +

        checkAndUpdateTableMetadata

        +
        public void checkAndUpdateTableMetadata​(Set<org.apache.fluss.metadata.TablePath> tablePaths)
        +
      • +
      + + + +
        +
      • +

        checkAndUpdatePartitionMetadata

        +
        public boolean checkAndUpdatePartitionMetadata​(org.apache.fluss.metadata.PhysicalTablePath physicalTablePath)
        +                                        throws org.apache.fluss.exception.PartitionNotExistException
        +
        Check the partition exists in metadata cache, if not, try to update the metadata cache, if + not exist yet, throw exception. + +

        and update partition metadata .

        +
        +
        Throws:
        +
        org.apache.fluss.exception.PartitionNotExistException
        +
        +
      • +
      + + + +
        +
      • +

        checkAndUpdateMetadata

        +
        public void checkAndUpdateMetadata​(org.apache.fluss.metadata.TablePath tablePath,
        +                                   org.apache.fluss.metadata.TableBucket tableBucket)
        +
        Check the table/partition info for the given table bucket exist in metadata cache, if not, + try to update the metadata cache.
        +
      • +
      + + + +
        +
      • +

        checkAndUpdatePartitionMetadata

        +
        public void checkAndUpdatePartitionMetadata​(org.apache.fluss.metadata.TablePath tablePath,
        +                                            Collection<Long> partitionIds)
        +
        Check the partitions info for the given partition ids exist in metadata cache, if not, try to + update the metadata cache. + +

        Note: it'll assume the partition ids belong to the given tablePath

        +
      • +
      + + + +
        +
      • +

        updateTableOrPartitionMetadata

        +
        public void updateTableOrPartitionMetadata​(org.apache.fluss.metadata.TablePath tablePath,
        +                                           @Nullable
        +                                           Long partitionId)
        +
      • +
      + + + +
        +
      • +

        updatePhysicalTableMetadata

        +
        public void updatePhysicalTableMetadata​(Set<org.apache.fluss.metadata.PhysicalTablePath> physicalTablePaths)
        +
        Update the table or partition metadata info.
        +
      • +
      + + + +
        +
      • +

        updateMetadata

        +
        @VisibleForTesting
        +protected void updateMetadata​(@Nullable
        +                              Set<org.apache.fluss.metadata.TablePath> tablePaths,
        +                              @Nullable
        +                              Collection<org.apache.fluss.metadata.PhysicalTablePath> tablePartitionNames,
        +                              @Nullable
        +                              Collection<Long> tablePartitionIds)
        +                       throws org.apache.fluss.exception.PartitionNotExistException
        +
        +
        Throws:
        +
        org.apache.fluss.exception.PartitionNotExistException
        +
        +
      • +
      + + + +
        +
      • +

        invalidPhysicalTableBucketMeta

        +
        public void invalidPhysicalTableBucketMeta​(Set<org.apache.fluss.metadata.PhysicalTablePath> physicalTablesToInvalid)
        +
        Invalid the bucket metadata for the given physical table paths.
        +
      • +
      + + + +
        +
      • +

        getPhysicalTablePathByIds

        +
        public Set<org.apache.fluss.metadata.PhysicalTablePath> getPhysicalTablePathByIds​(@Nullable
        +                                                                                  Collection<Long> tableId,
        +                                                                                  @Nullable
        +                                                                                  Collection<org.apache.fluss.metadata.TablePartition> tablePartitions)
        +
        Get the table physical paths by table ids and partition ids.
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/KvSnapshotMetadata.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/KvSnapshotMetadata.html new file mode 100644 index 0000000000..3371e6022f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/KvSnapshotMetadata.html @@ -0,0 +1,227 @@ + + + + + +Uses of Class org.apache.fluss.client.metadata.KvSnapshotMetadata (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.metadata.KvSnapshotMetadata

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/KvSnapshots.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/KvSnapshots.html new file mode 100644 index 0000000000..79d0606216 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/KvSnapshots.html @@ -0,0 +1,239 @@ + + + + + +Uses of Class org.apache.fluss.client.metadata.KvSnapshots (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.metadata.KvSnapshots

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/LakeSnapshot.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/LakeSnapshot.html new file mode 100644 index 0000000000..7cbf96d092 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/LakeSnapshot.html @@ -0,0 +1,225 @@ + + + + + +Uses of Class org.apache.fluss.client.metadata.LakeSnapshot (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.metadata.LakeSnapshot

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/MetadataUpdater.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/MetadataUpdater.html new file mode 100644 index 0000000000..521c06b7dd --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/MetadataUpdater.html @@ -0,0 +1,398 @@ + + + + + +Uses of Class org.apache.fluss.client.metadata.MetadataUpdater (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.metadata.MetadataUpdater

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-summary.html new file mode 100644 index 0000000000..2912039719 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-summary.html @@ -0,0 +1,186 @@ + + + + + +org.apache.fluss.client.metadata (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.client.metadata

+
+
+
    +
  • + + + + + + + + + + + + + + + + + + + + + + + + +
    Class Summary 
    ClassDescription
    KvSnapshotMetadata +
    A class to represent the snapshot info of a kv tablet (bucket).
    +
    KvSnapshots +
    A class representing the kv snapshots of a table or a partition.
    +
    LakeSnapshot +
    A class representing the lake snapshot information of a table.
    +
    MetadataUpdater +
    The updater to initialize and update client metadata.
    +
    +
  • +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-tree.html new file mode 100644 index 0000000000..e3f50d5f39 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-tree.html @@ -0,0 +1,166 @@ + + + + + +org.apache.fluss.client.metadata Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.client.metadata

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-use.html new file mode 100644 index 0000000000..691feed838 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-use.html @@ -0,0 +1,359 @@ + + + + + +Uses of Package org.apache.fluss.client.metadata (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.client.metadata

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/ScannerMetricGroup.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/ScannerMetricGroup.html new file mode 100644 index 0000000000..56df1dee93 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/ScannerMetricGroup.html @@ -0,0 +1,494 @@ + + + + + +ScannerMetricGroup (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class ScannerMetricGroup

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.metrics.groups.AbstractMetricGroup
    • +
    • +
        +
      • org.apache.fluss.client.metrics.ScannerMetricGroup
      • +
      +
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.fluss.metrics.groups.MetricGroup
    +
    +
    +
    @Internal
    +public class ScannerMetricGroup
    +extends org.apache.fluss.metrics.groups.AbstractMetricGroup
    +
    The metric group for scanner, including LogScanner and BatchScanner.
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Nested Class Summary

      +
        +
      • + + +

        Nested classes/interfaces inherited from class org.apache.fluss.metrics.groups.AbstractMetricGroup

        +org.apache.fluss.metrics.groups.AbstractMetricGroup.ChildType
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Field Summary

      +
        +
      • + + +

        Fields inherited from class org.apache.fluss.metrics.groups.AbstractMetricGroup

        +LOG, parent, registry, variables
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Summary

      + + + + + + + + + + +
      Constructors 
      ConstructorDescription
      ScannerMetricGroup​(org.apache.fluss.rpc.metrics.ClientMetricGroup parent, + org.apache.fluss.metadata.TablePath tablePath) 
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        ScannerMetricGroup

        +
        public ScannerMetricGroup​(org.apache.fluss.rpc.metrics.ClientMetricGroup parent,
        +                          org.apache.fluss.metadata.TablePath tablePath)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        fetchRequestCount

        +
        public org.apache.fluss.metrics.Counter fetchRequestCount()
        +
      • +
      + + + +
        +
      • +

        bytesPerRequest

        +
        public org.apache.fluss.metrics.Histogram bytesPerRequest()
        +
      • +
      + + + +
        +
      • +

        remoteFetchBytes

        +
        public org.apache.fluss.metrics.Counter remoteFetchBytes()
        +
      • +
      + + + +
        +
      • +

        remoteFetchRequestCount

        +
        public org.apache.fluss.metrics.Counter remoteFetchRequestCount()
        +
      • +
      + + + +
        +
      • +

        remoteFetchErrorCount

        +
        public org.apache.fluss.metrics.Counter remoteFetchErrorCount()
        +
      • +
      + + + +
        +
      • +

        recordPollStart

        +
        public void recordPollStart​(long pollStartMs)
        +
      • +
      + + + +
        +
      • +

        recordPollEnd

        +
        public void recordPollEnd​(long pollEndMs)
        +
      • +
      + + + +
        +
      • +

        updateFetchLatency

        +
        public void updateFetchLatency​(long latencyInMs)
        +
      • +
      + + + +
        +
      • +

        getGroupName

        +
        protected String getGroupName​(org.apache.fluss.metrics.CharacterFilter filter)
        +
        +
        Specified by:
        +
        getGroupName in class org.apache.fluss.metrics.groups.AbstractMetricGroup
        +
        +
      • +
      + + + +
        +
      • +

        putVariables

        +
        protected final void putVariables​(Map<String,​String> variables)
        +
        +
        Overrides:
        +
        putVariables in class org.apache.fluss.metrics.groups.AbstractMetricGroup
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/WriterMetricGroup.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/WriterMetricGroup.html new file mode 100644 index 0000000000..c852040807 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/WriterMetricGroup.html @@ -0,0 +1,460 @@ + + + + + +WriterMetricGroup (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class WriterMetricGroup

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.metrics.groups.AbstractMetricGroup
    • +
    • +
        +
      • org.apache.fluss.client.metrics.WriterMetricGroup
      • +
      +
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.fluss.metrics.groups.MetricGroup
    +
    +
    +
    @Internal
    +public class WriterMetricGroup
    +extends org.apache.fluss.metrics.groups.AbstractMetricGroup
    +
    Metrics for WriterClient.
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Nested Class Summary

      +
        +
      • + + +

        Nested classes/interfaces inherited from class org.apache.fluss.metrics.groups.AbstractMetricGroup

        +org.apache.fluss.metrics.groups.AbstractMetricGroup.ChildType
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Field Summary

      +
        +
      • + + +

        Fields inherited from class org.apache.fluss.metrics.groups.AbstractMetricGroup

        +LOG, parent, registry, variables
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Summary

      + + + + + + + + + + +
      Constructors 
      ConstructorDescription
      WriterMetricGroup​(org.apache.fluss.rpc.metrics.ClientMetricGroup parent) 
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        WriterMetricGroup

        +
        public WriterMetricGroup​(org.apache.fluss.rpc.metrics.ClientMetricGroup parent)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        setBatchQueueTimeMs

        +
        public void setBatchQueueTimeMs​(long batchQueueTimeMs)
        +
      • +
      + + + +
        +
      • +

        setSendLatencyInMs

        +
        public void setSendLatencyInMs​(long sendLatencyInMs)
        +
      • +
      + + + +
        +
      • +

        recordsRetryTotal

        +
        public org.apache.fluss.metrics.Counter recordsRetryTotal()
        +
      • +
      + + + +
        +
      • +

        recordsSendTotal

        +
        public org.apache.fluss.metrics.Counter recordsSendTotal()
        +
      • +
      + + + +
        +
      • +

        bytesPerBatch

        +
        public org.apache.fluss.metrics.Histogram bytesPerBatch()
        +
      • +
      + + + +
        +
      • +

        bytesSendTotal

        +
        public org.apache.fluss.metrics.Counter bytesSendTotal()
        +
      • +
      + + + +
        +
      • +

        recordPerBatch

        +
        public org.apache.fluss.metrics.Histogram recordPerBatch()
        +
      • +
      + + + +
        +
      • +

        getGroupName

        +
        protected String getGroupName​(org.apache.fluss.metrics.CharacterFilter filter)
        +
        +
        Specified by:
        +
        getGroupName in class org.apache.fluss.metrics.groups.AbstractMetricGroup
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/class-use/ScannerMetricGroup.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/class-use/ScannerMetricGroup.html new file mode 100644 index 0000000000..b55b238b64 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/class-use/ScannerMetricGroup.html @@ -0,0 +1,203 @@ + + + + + +Uses of Class org.apache.fluss.client.metrics.ScannerMetricGroup (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.metrics.ScannerMetricGroup

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/class-use/WriterMetricGroup.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/class-use/WriterMetricGroup.html new file mode 100644 index 0000000000..d393a72f6a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/class-use/WriterMetricGroup.html @@ -0,0 +1,197 @@ + + + + + +Uses of Class org.apache.fluss.client.metrics.WriterMetricGroup (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.metrics.WriterMetricGroup

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-summary.html new file mode 100644 index 0000000000..bf07f802d3 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-summary.html @@ -0,0 +1,174 @@ + + + + + +org.apache.fluss.client.metrics (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.client.metrics

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-tree.html new file mode 100644 index 0000000000..ccebcb4fb1 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-tree.html @@ -0,0 +1,168 @@ + + + + + +org.apache.fluss.client.metrics Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.client.metrics

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+
    +
  • java.lang.Object +
      +
    • org.apache.fluss.metrics.groups.AbstractMetricGroup (implements org.apache.fluss.metrics.groups.MetricGroup) + +
    • +
    +
  • +
+
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-use.html new file mode 100644 index 0000000000..ee24d19b1a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-use.html @@ -0,0 +1,208 @@ + + + + + +Uses of Package org.apache.fluss.client.metrics (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.client.metrics

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-summary.html new file mode 100644 index 0000000000..444abc2cbf --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-summary.html @@ -0,0 +1,191 @@ + + + + + +org.apache.fluss.client (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.client

+
+
+
    +
  • + + + + + + + + + + + + +
    Interface Summary 
    InterfaceDescription
    Connection +
    A cluster connection encapsulating lower level individual connections to actual Fluss servers.
    +
    +
  • +
  • + + + + + + + + + + + + + + + + +
    Class Summary 
    ClassDescription
    ConnectionFactory +
    A non-instantiable class that manages creation of Connections.
    +
    FlussConnection +
    A connection to Fluss cluster, and holds the client session resources.
    +
    +
  • +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-tree.html new file mode 100644 index 0000000000..d44d4c30ce --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-tree.html @@ -0,0 +1,174 @@ + + + + + +org.apache.fluss.client Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.client

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+

Interface Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-use.html new file mode 100644 index 0000000000..df0ee0dd15 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-use.html @@ -0,0 +1,254 @@ + + + + + +Uses of Package org.apache.fluss.client (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.client

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/FlussTable.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/FlussTable.html new file mode 100644 index 0000000000..50c06a5414 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/FlussTable.html @@ -0,0 +1,451 @@ + + + + + +FlussTable (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlussTable

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    AutoCloseable, Table
    +
    +
    +
    @PublicEvolving
    +public class FlussTable
    +extends Object
    +implements Table
    +
    The base impl of Table.
    +
    +
    Since:
    +
    0.1
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlussTable

        +
        public FlussTable​(FlussConnection conn,
        +                  org.apache.fluss.metadata.TablePath tablePath,
        +                  org.apache.fluss.metadata.TableInfo tableInfo)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getTableInfo

        +
        public org.apache.fluss.metadata.TableInfo getTableInfo()
        +
        Description copied from interface: Table
        +
        Get the TableInfo for this table. + +

        Note: the table info of this Table is set during the creation of this Table and will not be updated after that, even if the table info of the table has been + changed. Therefore, if there are any changes to the table info, it may be necessary to + reconstruct the Table.

        +
        +
        Specified by:
        +
        getTableInfo in interface Table
        +
        +
      • +
      + + + +
        +
      • +

        newScan

        +
        public Scan newScan()
        +
        Description copied from interface: Table
        +
        Creates a new Scan for this table to configure and create a scanner to scan data for + this table. The scanner can be a log scanner to continuously read streaming log data or a + batch scanner to read batch data.
        +
        +
        Specified by:
        +
        newScan in interface Table
        +
        +
      • +
      + + + +
        +
      • +

        newLookup

        +
        public Lookup newLookup()
        +
        Description copied from interface: Table
        +
        Creates a new Lookup for this table to configure and create a Lookuper to + lookup data for this table by primary key or a prefix of primary key.
        +
        +
        Specified by:
        +
        newLookup in interface Table
        +
        +
      • +
      + + + +
        +
      • +

        newAppend

        +
        public Append newAppend()
        +
        Description copied from interface: Table
        +
        Creates a new Append to build a AppendWriter to append data to this table + (requires to be a Log Table).
        +
        +
        Specified by:
        +
        newAppend in interface Table
        +
        +
      • +
      + + + +
        +
      • +

        newUpsert

        +
        public Upsert newUpsert()
        +
        Description copied from interface: Table
        +
        Creates a new Upsert to build a UpsertWriter to upsert and delete data to + this table (requires to be a Primary Key Table).
        +
        +
        Specified by:
        +
        newUpsert in interface Table
        +
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/Table.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/Table.html new file mode 100644 index 0000000000..03552fa610 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/Table.html @@ -0,0 +1,358 @@ + + + + + +Table (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface Table

+
+
+
+
    +
  • +
    +
    All Superinterfaces:
    +
    AutoCloseable
    +
    +
    +
    All Known Implementing Classes:
    +
    FlussTable
    +
    +
    +
    @PublicEvolving
    +public interface Table
    +extends AutoCloseable
    +
    Used to communicate with a single Fluss table. Obtain an instance from a Connection. + +

    Table can be used to get, put, delete or scan data from a fluss table.

    +
    +
    Since:
    +
    0.1
    +
    +
  • +
+
+
+
    +
  • + +
    + +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getTableInfo

        +
        org.apache.fluss.metadata.TableInfo getTableInfo()
        +
        Get the TableInfo for this table. + +

        Note: the table info of this Table is set during the creation of this Table and will not be updated after that, even if the table info of the table has been + changed. Therefore, if there are any changes to the table info, it may be necessary to + reconstruct the Table.

        +
      • +
      + + + +
        +
      • +

        newScan

        +
        Scan newScan()
        +
        Creates a new Scan for this table to configure and create a scanner to scan data for + this table. The scanner can be a log scanner to continuously read streaming log data or a + batch scanner to read batch data.
        +
      • +
      + + + +
        +
      • +

        newLookup

        +
        Lookup newLookup()
        +
        Creates a new Lookup for this table to configure and create a Lookuper to + lookup data for this table by primary key or a prefix of primary key.
        +
      • +
      + + + +
        +
      • +

        newAppend

        +
        Append newAppend()
        +
        Creates a new Append to build a AppendWriter to append data to this table + (requires to be a Log Table).
        +
      • +
      + + + +
        +
      • +

        newUpsert

        +
        Upsert newUpsert()
        +
        Creates a new Upsert to build a UpsertWriter to upsert and delete data to + this table (requires to be a Primary Key Table).
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/class-use/FlussTable.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/class-use/FlussTable.html new file mode 100644 index 0000000000..3a9ebf5a3e --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/class-use/FlussTable.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.table.FlussTable (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.FlussTable

+
+
No usage of org.apache.fluss.client.table.FlussTable
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/class-use/Table.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/class-use/Table.html new file mode 100644 index 0000000000..51bae671c7 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/class-use/Table.html @@ -0,0 +1,306 @@ + + + + + +Uses of Interface org.apache.fluss.client.table.Table (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.client.table.Table

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/PartitionGetter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/PartitionGetter.html new file mode 100644 index 0000000000..5c0474d514 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/PartitionGetter.html @@ -0,0 +1,309 @@ + + + + + +PartitionGetter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class PartitionGetter

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.table.getter.PartitionGetter
    • +
    +
  • +
+
+
    +
  • +
    +
    public class PartitionGetter
    +extends Object
    +
    A getter to get partition name from a row.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        PartitionGetter

        +
        public PartitionGetter​(org.apache.fluss.types.RowType rowType,
        +                       List<String> partitionKeys)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getPartition

        +
        public String getPartition​(org.apache.fluss.row.InternalRow row)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/class-use/PartitionGetter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/class-use/PartitionGetter.html new file mode 100644 index 0000000000..237574e2a2 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/class-use/PartitionGetter.html @@ -0,0 +1,197 @@ + + + + + +Uses of Class org.apache.fluss.client.table.getter.PartitionGetter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.getter.PartitionGetter

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-summary.html new file mode 100644 index 0000000000..bddc3fd654 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-summary.html @@ -0,0 +1,168 @@ + + + + + +org.apache.fluss.client.table.getter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.client.table.getter

+
+
+
    +
  • + + + + + + + + + + + + +
    Class Summary 
    ClassDescription
    PartitionGetter +
    A getter to get partition name from a row.
    +
    +
  • +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-tree.html new file mode 100644 index 0000000000..27b4de7e79 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-tree.html @@ -0,0 +1,163 @@ + + + + + +org.apache.fluss.client.table.getter Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.client.table.getter

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-use.html new file mode 100644 index 0000000000..c86e5968cb --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-use.html @@ -0,0 +1,185 @@ + + + + + +Uses of Package org.apache.fluss.client.table.getter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.client.table.getter

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-summary.html new file mode 100644 index 0000000000..4c6b88ad7f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-summary.html @@ -0,0 +1,185 @@ + + + + + +org.apache.fluss.client.table (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.client.table

+
+
+
    +
  • + + + + + + + + + + + + +
    Interface Summary 
    InterfaceDescription
    Table +
    Used to communicate with a single Fluss table.
    +
    +
  • +
  • + + + + + + + + + + + + +
    Class Summary 
    ClassDescription
    FlussTable +
    The base impl of Table.
    +
    +
  • +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-tree.html new file mode 100644 index 0000000000..0c3b04ded2 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-tree.html @@ -0,0 +1,173 @@ + + + + + +org.apache.fluss.client.table Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.client.table

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+
    +
  • java.lang.Object +
      +
    • org.apache.fluss.client.table.FlussTable (implements org.apache.fluss.client.table.Table)
    • +
    +
  • +
+
+
+

Interface Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-use.html new file mode 100644 index 0000000000..f8fd961b49 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-use.html @@ -0,0 +1,277 @@ + + + + + +Uses of Package org.apache.fluss.client.table (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.client.table

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/RemoteFileDownloader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/RemoteFileDownloader.html new file mode 100644 index 0000000000..97dd7905de --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/RemoteFileDownloader.html @@ -0,0 +1,435 @@ + + + + + +RemoteFileDownloader (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class RemoteFileDownloader

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.table.scanner.RemoteFileDownloader
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    Closeable, AutoCloseable
    +
    +
    +
    public class RemoteFileDownloader
    +extends Object
    +implements Closeable
    +
    The downloader that has a IO thread pool to download the remote files (like kv snapshots files, + log segment files).
    +
  • +
+
+
+ +
+
+
    +
  • + +
    + +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        RemoteFileDownloader

        +
        public RemoteFileDownloader​(int threadNum)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        downloadFileAsync

        +
        public CompletableFuture<Long> downloadFileAsync​(org.apache.fluss.fs.FsPathAndFileName fsPathAndFileName,
        +                                                 Path targetDirectory)
        +
        Downloads the file from the given remote file path to the target directory asynchronously, + returns a Future object of the number of downloaded bytes. The Future will fail if the + download fails after retrying for RETRY_COUNT times.
        +
      • +
      + + + +
        +
      • +

        downloadFile

        +
        protected long downloadFile​(Path targetFilePath,
        +                            org.apache.fluss.fs.FsPath remoteFilePath)
        +                     throws IOException
        +
        Copies the file from a remote file path to the given target file path, returns the number of + downloaded bytes.
        +
        +
        Throws:
        +
        IOException
        +
        +
      • +
      + + + + + + + +
        +
      • +

        transferAllToDirectory

        +
        public void transferAllToDirectory​(List<org.apache.fluss.fs.FsPathAndFileName> fsPathAndFileNames,
        +                                   Path targetDirectory,
        +                                   org.apache.fluss.utils.CloseableRegistry closeableRegistry)
        +                            throws IOException
        +
        +
        Throws:
        +
        IOException
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/Scan.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/Scan.html new file mode 100644 index 0000000000..1f1ab5c3b2 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/Scan.html @@ -0,0 +1,374 @@ + + + + + +Scan (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface Scan

+
+
+
+
    +
  • +
    +
    All Known Implementing Classes:
    +
    TableScan
    +
    +
    +
    @PublicEvolving
    +public interface Scan
    +
    Used to configure and create a scanner to scan data for a table. + +

    Scan objects are immutable and can be shared between threads. Refinement methods, like + project(int[]) and limit(int), create new Scan instances.

    +
    +
    Since:
    +
    0.6
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Method Summary

      + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
      All Methods Instance Methods Abstract Methods 
      Modifier and TypeMethodDescription
      BatchScannercreateBatchScanner​(org.apache.fluss.metadata.TableBucket tableBucket) +
      Creates a BatchScanner to read current data in the given table bucket for this scan.
      +
      BatchScannercreateBatchScanner​(org.apache.fluss.metadata.TableBucket tableBucket, + long snapshotId) +
      Creates a BatchScanner to read given snapshot data in the given table bucket for this + scan.
      +
      LogScannercreateLogScanner() +
      Creates a LogScanner to continuously read log data for this scan.
      +
      Scanlimit​(int rowNumber) +
      Returns a new scan from this that will read the given limited row number.
      +
      Scanproject​(int[] projectedColumns) +
      Returns a new scan from this that will read the given data columns.
      +
      Scanproject​(List<String> projectedColumnNames) +
      Returns a new scan from this that will read the given data columns.
      +
      +
    • +
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        project

        +
        Scan project​(@Nullable
        +             int[] projectedColumns)
        +
        Returns a new scan from this that will read the given data columns.
        +
        +
        Parameters:
        +
        projectedColumns - the selected column indexes
        +
        +
      • +
      + + + +
        +
      • +

        project

        +
        Scan project​(List<String> projectedColumnNames)
        +
        Returns a new scan from this that will read the given data columns.
        +
        +
        Parameters:
        +
        projectedColumnNames - the selected column names
        +
        +
      • +
      + + + +
        +
      • +

        limit

        +
        Scan limit​(int rowNumber)
        +
        Returns a new scan from this that will read the given limited row number.
        +
        +
        Parameters:
        +
        rowNumber - the limited row number to read
        +
        +
      • +
      + + + +
        +
      • +

        createLogScanner

        +
        LogScanner createLogScanner()
        +
        Creates a LogScanner to continuously read log data for this scan. + +

        Note: this API doesn't support pre-configured with limit(int).

        +
      • +
      + + + +
        +
      • +

        createBatchScanner

        +
        BatchScanner createBatchScanner​(org.apache.fluss.metadata.TableBucket tableBucket)
        +
        Creates a BatchScanner to read current data in the given table bucket for this scan. + +

        Note: this API doesn't support pre-configured with project(int[]).

        +
      • +
      + + + +
        +
      • +

        createBatchScanner

        +
        BatchScanner createBatchScanner​(org.apache.fluss.metadata.TableBucket tableBucket,
        +                                long snapshotId)
        +
        Creates a BatchScanner to read given snapshot data in the given table bucket for this + scan. + +

        Note: this API doesn't support pre-configured with project(int[]) and limit(int) and only support for Primary Key Tables.

        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/ScanRecord.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/ScanRecord.html new file mode 100644 index 0000000000..8c7501644f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/ScanRecord.html @@ -0,0 +1,447 @@ + + + + + +ScanRecord (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class ScanRecord

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.fluss.record.LogRecord
    +
    +
    +
    @Internal
    +public class ScanRecord
    +extends Object
    +implements org.apache.fluss.record.LogRecord
    +
    one scan record.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        ScanRecord

        +
        public ScanRecord​(org.apache.fluss.row.InternalRow row)
        +
      • +
      + + + +
        +
      • +

        ScanRecord

        +
        public ScanRecord​(long offset,
        +                  long timestamp,
        +                  org.apache.fluss.record.ChangeType changeType,
        +                  org.apache.fluss.row.InternalRow row)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        logOffset

        +
        public long logOffset()
        +
        The position of this record in the corresponding fluss table bucket.
        +
        +
        Specified by:
        +
        logOffset in interface org.apache.fluss.record.LogRecord
        +
        +
      • +
      + + + +
        +
      • +

        timestamp

        +
        public long timestamp()
        +
        +
        Specified by:
        +
        timestamp in interface org.apache.fluss.record.LogRecord
        +
        +
      • +
      + + + +
        +
      • +

        getChangeType

        +
        public org.apache.fluss.record.ChangeType getChangeType()
        +
        +
        Specified by:
        +
        getChangeType in interface org.apache.fluss.record.LogRecord
        +
        +
      • +
      + + + +
        +
      • +

        getRow

        +
        public org.apache.fluss.row.InternalRow getRow()
        +
        +
        Specified by:
        +
        getRow in interface org.apache.fluss.record.LogRecord
        +
        +
      • +
      + + + +
        +
      • +

        equals

        +
        public boolean equals​(Object o)
        +
        +
        Overrides:
        +
        equals in class Object
        +
        +
      • +
      + + + +
        +
      • +

        hashCode

        +
        public int hashCode()
        +
        +
        Overrides:
        +
        hashCode in class Object
        +
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/TableScan.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/TableScan.html new file mode 100644 index 0000000000..057f3828ff --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/TableScan.html @@ -0,0 +1,449 @@ + + + + + +TableScan (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TableScan

+
+
+ +
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        TableScan

        +
        public TableScan​(FlussConnection conn,
        +                 org.apache.fluss.metadata.TableInfo tableInfo)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        project

        +
        public Scan project​(@Nullable
        +                    int[] projectedColumns)
        +
        Description copied from interface: Scan
        +
        Returns a new scan from this that will read the given data columns.
        +
        +
        Specified by:
        +
        project in interface Scan
        +
        Parameters:
        +
        projectedColumns - the selected column indexes
        +
        +
      • +
      + + + +
        +
      • +

        project

        +
        public Scan project​(List<String> projectedColumnNames)
        +
        Description copied from interface: Scan
        +
        Returns a new scan from this that will read the given data columns.
        +
        +
        Specified by:
        +
        project in interface Scan
        +
        Parameters:
        +
        projectedColumnNames - the selected column names
        +
        +
      • +
      + + + +
        +
      • +

        limit

        +
        public Scan limit​(int rowNumber)
        +
        Description copied from interface: Scan
        +
        Returns a new scan from this that will read the given limited row number.
        +
        +
        Specified by:
        +
        limit in interface Scan
        +
        Parameters:
        +
        rowNumber - the limited row number to read
        +
        +
      • +
      + + + +
        +
      • +

        createLogScanner

        +
        public LogScanner createLogScanner()
        +
        Description copied from interface: Scan
        +
        Creates a LogScanner to continuously read log data for this scan. + +

        Note: this API doesn't support pre-configured with Scan.limit(int).

        +
        +
        Specified by:
        +
        createLogScanner in interface Scan
        +
        +
      • +
      + + + +
        +
      • +

        createBatchScanner

        +
        public BatchScanner createBatchScanner​(org.apache.fluss.metadata.TableBucket tableBucket)
        +
        Description copied from interface: Scan
        +
        Creates a BatchScanner to read current data in the given table bucket for this scan. + +

        Note: this API doesn't support pre-configured with Scan.project(int[]).

        +
        +
        Specified by:
        +
        createBatchScanner in interface Scan
        +
        +
      • +
      + + + +
        +
      • +

        createBatchScanner

        +
        public BatchScanner createBatchScanner​(org.apache.fluss.metadata.TableBucket tableBucket,
        +                                       long snapshotId)
        +
        Description copied from interface: Scan
        +
        Creates a BatchScanner to read given snapshot data in the given table bucket for this + scan. + +

        Note: this API doesn't support pre-configured with Scan.project(int[]) and Scan.limit(int) and only support for Primary Key Tables.

        +
        +
        Specified by:
        +
        createBatchScanner in interface Scan
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/BatchScanUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/BatchScanUtils.html new file mode 100644 index 0000000000..8d98ac7d97 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/BatchScanUtils.html @@ -0,0 +1,367 @@ + + + + + +BatchScanUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class BatchScanUtils

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.table.scanner.batch.BatchScanUtils
    • +
    +
  • +
+
+
    +
  • +
    +
    public class BatchScanUtils
    +extends Object
    +
    Utility class for batch scan.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        BatchScanUtils

        +
        public BatchScanUtils()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        collectRows

        +
        public static List<org.apache.fluss.row.InternalRow> collectRows​(BatchScanner scanner)
        +
        Collect all rows from the scanner.
        +
      • +
      + + + +
        +
      • +

        collectAllRows

        +
        public static List<org.apache.fluss.row.InternalRow> collectAllRows​(List<BatchScanner> scanners)
        +
        Collect all rows from the scanners.
        +
      • +
      + + + +
        +
      • +

        collectLimitedRows

        +
        public static List<org.apache.fluss.row.InternalRow> collectLimitedRows​(List<BatchScanner> scanners,
        +                                                                        int limit)
        +
        Collect rows from the scanners until the number of rows reaches the limit.
        +
      • +
      + + + +
        +
      • +

        toList

        +
        public static <T> List<T> toList​(org.apache.fluss.utils.CloseableIterator<T> iterator)
        +
        Collect all rows from the iterator and close it.
        +
        +
        Returns:
        +
        return a list of rows
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/BatchScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/BatchScanner.html new file mode 100644 index 0000000000..d97399b796 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/BatchScanner.html @@ -0,0 +1,305 @@ + + + + + +BatchScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface BatchScanner

+
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Summary

      + + + + + + + + + + + + + + + + + +
      All Methods Instance Methods Abstract Methods 
      Modifier and TypeMethodDescription
      voidclose() +
      Closes the scanner and should release all resources.
      +
      org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow>pollBatch​(Duration timeout) +
      Poll one batch records.
      +
      +
    • +
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        pollBatch

        +
        @Nullable
        +org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow> pollBatch​(Duration timeout)
        +                                                                              throws IOException
        +
        Poll one batch records. The method should return null when reaching the end of the input.
        +
        +
        Parameters:
        +
        timeout - The maximum time to block (must not be greater than Long.MAX_VALUE + milliseconds)
        +
        Throws:
        +
        IOException
        +
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/KvSnapshotBatchScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/KvSnapshotBatchScanner.html new file mode 100644 index 0000000000..bf91387213 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/KvSnapshotBatchScanner.html @@ -0,0 +1,423 @@ + + + + + +KvSnapshotBatchScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class KvSnapshotBatchScanner

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.table.scanner.batch.KvSnapshotBatchScanner
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    Closeable, AutoCloseable, BatchScanner
    +
    +
    +
    @Internal
    +public class KvSnapshotBatchScanner
    +extends Object
    +implements BatchScanner
    +
    A scanner to scan the snapshot data of a kv bucket. + +

    When the scanner is created, It will download the given snapshot files from the remote storage + and create a reader to read the files into records asynchronously. + +

    In the BatchScanner.pollBatch(Duration) method: + +

      +
    • if the reader is not ready in given time, return an empty iterator +
    • If the reader is ready, always return the reader if there remains any data in the reader. + Otherwise, return null +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Field Summary

      + + + + + + + + + + + + +
      Fields 
      Modifier and TypeFieldDescription
      static org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow>NO_DATA_AVAILABLE 
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Summary

      + + + + + + + + + + +
      Constructors 
      ConstructorDescription
      KvSnapshotBatchScanner​(org.apache.fluss.types.RowType tableRowType, + org.apache.fluss.metadata.TableBucket tableBucket, + List<org.apache.fluss.fs.FsPathAndFileName> fsPathAndFileNames, + int[] projectedFields, + String scannerTmpDir, + org.apache.fluss.metadata.KvFormat kvFormat, + RemoteFileDownloader remoteFileDownloader) 
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + +
        +
      • +

        NO_DATA_AVAILABLE

        +
        public static final org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow> NO_DATA_AVAILABLE
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        KvSnapshotBatchScanner

        +
        public KvSnapshotBatchScanner​(org.apache.fluss.types.RowType tableRowType,
        +                              org.apache.fluss.metadata.TableBucket tableBucket,
        +                              List<org.apache.fluss.fs.FsPathAndFileName> fsPathAndFileNames,
        +                              @Nullable
        +                              int[] projectedFields,
        +                              String scannerTmpDir,
        +                              org.apache.fluss.metadata.KvFormat kvFormat,
        +                              RemoteFileDownloader remoteFileDownloader)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        pollBatch

        +
        @Nullable
        +public org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow> pollBatch​(Duration timeout)
        +                                                                                     throws IOException
        +
        Fetch data from snapshot. + +

        If the snapshot file reader is not ready in given maximum block time, return an empty + iterator. If the reader is ready, always return the reader if there remains any record in the + reader, otherwise, return null.

        +
        +
        Specified by:
        +
        pollBatch in interface BatchScanner
        +
        Parameters:
        +
        timeout - The maximum time to block (must not be greater than Long.MAX_VALUE + milliseconds)
        +
        Throws:
        +
        IOException
        +
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/LimitBatchScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/LimitBatchScanner.html new file mode 100644 index 0000000000..f4c95f0785 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/LimitBatchScanner.html @@ -0,0 +1,365 @@ + + + + + +LimitBatchScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LimitBatchScanner

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.table.scanner.batch.LimitBatchScanner
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LimitBatchScanner

        +
        public LimitBatchScanner​(org.apache.fluss.metadata.TableInfo tableInfo,
        +                         org.apache.fluss.metadata.TableBucket tableBucket,
        +                         MetadataUpdater metadataUpdater,
        +                         @Nullable
        +                         int[] projectedFields,
        +                         int limit)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        pollBatch

        +
        @Nullable
        +public org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow> pollBatch​(Duration timeout)
        +                                                                                     throws IOException
        +
        Description copied from interface: BatchScanner
        +
        Poll one batch records. The method should return null when reaching the end of the input.
        +
        +
        Specified by:
        +
        pollBatch in interface BatchScanner
        +
        Parameters:
        +
        timeout - The maximum time to block (must not be greater than Long.MAX_VALUE + milliseconds)
        +
        Throws:
        +
        IOException
        +
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/BatchScanUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/BatchScanUtils.html new file mode 100644 index 0000000000..75777dd2d8 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/BatchScanUtils.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.table.scanner.batch.BatchScanUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.scanner.batch.BatchScanUtils

+
+
No usage of org.apache.fluss.client.table.scanner.batch.BatchScanUtils
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/BatchScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/BatchScanner.html new file mode 100644 index 0000000000..6be498059a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/BatchScanner.html @@ -0,0 +1,358 @@ + + + + + +Uses of Interface org.apache.fluss.client.table.scanner.batch.BatchScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.client.table.scanner.batch.BatchScanner

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/KvSnapshotBatchScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/KvSnapshotBatchScanner.html new file mode 100644 index 0000000000..75269c070d --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/KvSnapshotBatchScanner.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.table.scanner.batch.KvSnapshotBatchScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.scanner.batch.KvSnapshotBatchScanner

+
+
No usage of org.apache.fluss.client.table.scanner.batch.KvSnapshotBatchScanner
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/LimitBatchScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/LimitBatchScanner.html new file mode 100644 index 0000000000..7499305f36 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/LimitBatchScanner.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.table.scanner.batch.LimitBatchScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.scanner.batch.LimitBatchScanner

+
+
No usage of org.apache.fluss.client.table.scanner.batch.LimitBatchScanner
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-summary.html new file mode 100644 index 0000000000..579e32d9b7 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-summary.html @@ -0,0 +1,197 @@ + + + + + +org.apache.fluss.client.table.scanner.batch (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.client.table.scanner.batch

+
+
+
    +
  • + + + + + + + + + + + + +
    Interface Summary 
    InterfaceDescription
    BatchScanner +
    The scanner that reads records form a table in a batch fashion.
    +
    +
  • +
  • + + + + + + + + + + + + + + + + + + + + +
    Class Summary 
    ClassDescription
    BatchScanUtils +
    Utility class for batch scan.
    +
    KvSnapshotBatchScanner +
    A scanner to scan the snapshot data of a kv bucket.
    +
    LimitBatchScanner +
    A BatchScanner implementation that scans a limited number of records from a table.
    +
    +
  • +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-tree.html new file mode 100644 index 0000000000..c3eaf1e8a8 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-tree.html @@ -0,0 +1,179 @@ + + + + + +org.apache.fluss.client.table.scanner.batch Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.client.table.scanner.batch

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+

Interface Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-use.html new file mode 100644 index 0000000000..0277e6da45 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-use.html @@ -0,0 +1,254 @@ + + + + + +Uses of Package org.apache.fluss.client.table.scanner.batch (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.client.table.scanner.batch

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/RemoteFileDownloader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/RemoteFileDownloader.html new file mode 100644 index 0000000000..cafc462740 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/RemoteFileDownloader.html @@ -0,0 +1,268 @@ + + + + + +Uses of Class org.apache.fluss.client.table.scanner.RemoteFileDownloader (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.scanner.RemoteFileDownloader

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/Scan.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/Scan.html new file mode 100644 index 0000000000..3a07484684 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/Scan.html @@ -0,0 +1,274 @@ + + + + + +Uses of Interface org.apache.fluss.client.table.scanner.Scan (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.client.table.scanner.Scan

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/ScanRecord.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/ScanRecord.html new file mode 100644 index 0000000000..b99c0ef1f1 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/ScanRecord.html @@ -0,0 +1,351 @@ + + + + + +Uses of Class org.apache.fluss.client.table.scanner.ScanRecord (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.scanner.ScanRecord

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/TableScan.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/TableScan.html new file mode 100644 index 0000000000..c6ab351270 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/TableScan.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.table.scanner.TableScan (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.scanner.TableScan

+
+
No usage of org.apache.fluss.client.table.scanner.TableScan
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetchBuffer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetchBuffer.html new file mode 100644 index 0000000000..6bf6232f39 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetchBuffer.html @@ -0,0 +1,340 @@ + + + + + +LogFetchBuffer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LogFetchBuffer

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.table.scanner.log.LogFetchBuffer
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    AutoCloseable
    +
    +
    +
    @ThreadSafe
    +@Internal
    +public class LogFetchBuffer
    +extends Object
    +implements AutoCloseable
    +
    LogFetchBuffer buffers up the results from the tablet server + responses as they are received. It's essentially a wrapper around a Queue of + CompletedFetch. There is at most one LogFetchBuffer per bucket in the queue. + +

    Note: this class is thread-safe with the intention that the data will + be created by a background thread and consumed by the application thread.

    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LogFetchBuffer

        +
        public LogFetchBuffer()
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetchCollector.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetchCollector.html new file mode 100644 index 0000000000..7b6d0349e4 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetchCollector.html @@ -0,0 +1,329 @@ + + + + + +LogFetchCollector (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LogFetchCollector

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.table.scanner.log.LogFetchCollector
    • +
    +
  • +
+
+
    +
  • +
    +
    @ThreadSafe
    +@Internal
    +public class LogFetchCollector
    +extends Object
    +
    LogFetchCollector operates at the LogRecordBatch level, as that is what is stored + in the LogFetchBuffer. Each LogRecord in the LogRecordBatch is converted + to a ScanRecord and added to the returned LogFetcher.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LogFetchCollector

        +
        public LogFetchCollector​(org.apache.fluss.metadata.TablePath tablePath,
        +                         LogScannerStatus logScannerStatus,
        +                         org.apache.fluss.config.Configuration conf,
        +                         MetadataUpdater metadataUpdater)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        collectFetch

        +
        public Map<org.apache.fluss.metadata.TableBucket,​List<ScanRecord>> collectFetch​(LogFetchBuffer logFetchBuffer)
        +
        Return the fetched log records, empty the record buffer and update the consumed position. + +

        NOTE: returning empty records guarantees the consumed position are NOT updated.

        +
        +
        Returns:
        +
        The fetched records per partition
        +
        Throws:
        +
        org.apache.fluss.exception.LogOffsetOutOfRangeException - If there is OffsetOutOfRange error in fetchResponse and + the defaultResetPolicy is NONE
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetcher.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetcher.html new file mode 100644 index 0000000000..b52d9827ac --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetcher.html @@ -0,0 +1,423 @@ + + + + + +LogFetcher (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LogFetcher

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.table.scanner.log.LogFetcher
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LogFetcher

        +
        public LogFetcher​(org.apache.fluss.metadata.TableInfo tableInfo,
        +                  @Nullable
        +                  org.apache.fluss.utils.Projection projection,
        +                  LogScannerStatus logScannerStatus,
        +                  org.apache.fluss.config.Configuration conf,
        +                  MetadataUpdater metadataUpdater,
        +                  ScannerMetricGroup scannerMetricGroup,
        +                  RemoteFileDownloader remoteFileDownloader)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        hasAvailableFetches

        +
        public boolean hasAvailableFetches()
        +
        Return whether we have any completed fetches that are fetch-able. This method is thread-safe.
        +
        +
        Returns:
        +
        true if there are completed fetches that can be returned, false otherwise
        +
        +
      • +
      + + + +
        +
      • +

        collectFetch

        +
        public Map<org.apache.fluss.metadata.TableBucket,​List<ScanRecord>> collectFetch()
        +
      • +
      + + + +
        +
      • +

        sendFetches

        +
        public void sendFetches()
        +
        Set up a fetch request for any node that we have assigned buckets for which doesn't already + have an in-flight fetch or pending fetch data.
        +
      • +
      + + + +
        +
      • +

        awaitNotEmpty

        +
        public boolean awaitNotEmpty​(long deadlineNanos)
        +
        +
        Parameters:
        +
        deadlineNanos - the deadline time to wait until
        +
        Returns:
        +
        false if the waiting time detectably elapsed before return from the method, else true
        +
        +
      • +
      + + + +
        +
      • +

        wakeup

        +
        public void wakeup()
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScan.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScan.html new file mode 100644 index 0000000000..5e6e5c3a6c --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScan.html @@ -0,0 +1,334 @@ + + + + + +LogScan (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LogScan

+
+
+ +
+
    +
  • +
    +
    @PublicEvolving
    +public class LogScan
    +extends Object
    +
    Used to describe the operation to scan log data by LogScanner to a table.
    +
    +
    Since:
    +
    0.1
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LogScan

        +
        public LogScan()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        withProjectedFields

        +
        public LogScan withProjectedFields​(int[] projectedFields)
        +
        Returns a new instance of LogScan description with column projection.
        +
        +
        Parameters:
        +
        projectedFields - the projection fields
        +
        +
      • +
      + + + +
        +
      • +

        getProjectedFields

        +
        @Nullable
        +public int[] getProjectedFields()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScanner.html new file mode 100644 index 0000000000..89c4845c60 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScanner.html @@ -0,0 +1,503 @@ + + + + + +LogScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface LogScanner

+
+
+
+
    +
  • +
    +
    All Superinterfaces:
    +
    AutoCloseable
    +
    +
    +
    All Known Implementing Classes:
    +
    LogScannerImpl
    +
    +
    +
    @PublicEvolving
    +public interface LogScanner
    +extends AutoCloseable
    +
    The scanner is used to scan log data of specify table from Fluss.
    +
    +
    Since:
    +
    0.1
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Field Summary

      + + + + + + + + + + + + +
      Fields 
      Modifier and TypeFieldDescription
      static longEARLIEST_OFFSET +
      The earliest offset to fetch from.
      +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Summary

      + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
      All Methods Instance Methods Abstract Methods Default Methods 
      Modifier and TypeMethodDescription
      ScanRecordspoll​(Duration timeout) +
      Poll log data from tablet server.
      +
      voidsubscribe​(int bucket, + long offset) +
      Subscribe to the given table bucket in given offset dynamically.
      +
      voidsubscribe​(long partitionId, + int bucket, + long offset) +
      Subscribe to the given partitioned table bucket in given offset dynamically.
      +
      default voidsubscribeFromBeginning​(int bucket) +
      Subscribe to the given table buckets from beginning dynamically.
      +
      default voidsubscribeFromBeginning​(long partitionId, + int bucket) +
      Subscribe to the given partitioned table bucket from beginning dynamically.
      +
      voidunsubscribe​(long partitionId, + int bucket) +
      Unsubscribe from the given bucket of given partition dynamically.
      +
      voidwakeup() +
      Wake up the log scanner in case the fetcher thread in log scanner is blocking in poll(Duration timeout).
      +
      + +
    • +
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + +
        +
      • +

        EARLIEST_OFFSET

        +
        static final long EARLIEST_OFFSET
        +
        The earliest offset to fetch from. Fluss uses "-2" to indicate fetching from log start + offset.
        +
        +
        See Also:
        +
        Constant Field Values
        +
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        poll

        +
        ScanRecords poll​(Duration timeout)
        +
        Poll log data from tablet server. + +

        On each poll, scanner will try to use the last scanned offset as the starting offset and + fetch sequentially. The latest offset will be cached in memory, and if there is no offset + cached in memory, scanner will start either from the earliest offset or from the latest + offset for each bucket according to the policy.

        +
        +
        Parameters:
        +
        timeout - the timeout to poll.
        +
        Returns:
        +
        the result of poll.
        +
        Throws:
        +
        IllegalStateException - if the scanner is not subscribed to any buckets to + read from.
        +
        +
      • +
      + + + +
        +
      • +

        subscribe

        +
        void subscribe​(int bucket,
        +               long offset)
        +
        Subscribe to the given table bucket in given offset dynamically. If the table bucket is + already subscribed, the offset will be updated. + +

        Please use subscribe(long, int, long) to subscribe a partitioned table.

        +
        +
        Parameters:
        +
        bucket - the table bucket to subscribe.
        +
        offset - the offset to start from.
        +
        Throws:
        +
        IllegalStateException - if the table is a partitioned table.
        +
        +
      • +
      + + + +
        +
      • +

        subscribeFromBeginning

        +
        default void subscribeFromBeginning​(int bucket)
        +
        Subscribe to the given table buckets from beginning dynamically. If the table bucket is + already subscribed, the start offset will be updated. + +

        It equals to call subscribe(int, long) with offset EARLIEST_OFFSET for + the subscribed bucket. + +

        Please use subscribeFromBeginning(long, int) to subscribe a partitioned table.

        +
        +
        Parameters:
        +
        bucket - the table bucket to subscribe.
        +
        Throws:
        +
        IllegalStateException - if the table is a partitioned table.
        +
        +
      • +
      + + + +
        +
      • +

        subscribe

        +
        void subscribe​(long partitionId,
        +               int bucket,
        +               long offset)
        +
        Subscribe to the given partitioned table bucket in given offset dynamically. If the table + bucket is already subscribed, the offset will be updated. + +

        Please use subscribe(int, long) to subscribe a non-partitioned table.

        +
        +
        Parameters:
        +
        partitionId - the partition id of the table partition to subscribe.
        +
        bucket - the table bucket to subscribe.
        +
        offset - the offset to start from.
        +
        Throws:
        +
        IllegalStateException - if the table is a non-partitioned table.
        +
        +
      • +
      + + + +
        +
      • +

        unsubscribe

        +
        void unsubscribe​(long partitionId,
        +                 int bucket)
        +
        Unsubscribe from the given bucket of given partition dynamically.
        +
        +
        Parameters:
        +
        partitionId - the partition id of the table partition to unsubscribe.
        +
        bucket - the table bucket to unsubscribe.
        +
        Throws:
        +
        IllegalStateException - if the table is a non-partitioned table.
        +
        +
      • +
      + + + +
        +
      • +

        subscribeFromBeginning

        +
        default void subscribeFromBeginning​(long partitionId,
        +                                    int bucket)
        +
        Subscribe to the given partitioned table bucket from beginning dynamically. If the table + bucket is already subscribed, the start offset will be updated. + +

        It equals to call subscribe(long, int, long) with offset EARLIEST_OFFSET + for the subscribed bucket. + +

        Please use subscribeFromBeginning(int) to subscribe a non-partitioned table.

        +
        +
        Parameters:
        +
        bucket - the table bucket to subscribe.
        +
        Throws:
        +
        IllegalStateException - if the table is a non-partitioned table.
        +
        +
      • +
      + + + +
        +
      • +

        wakeup

        +
        void wakeup()
        +
        Wake up the log scanner in case the fetcher thread in log scanner is blocking in poll(Duration timeout).
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScannerImpl.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScannerImpl.html new file mode 100644 index 0000000000..f63ef7841d --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScannerImpl.html @@ -0,0 +1,502 @@ + + + + + +LogScannerImpl (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LogScannerImpl

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.table.scanner.log.LogScannerImpl
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LogScannerImpl

        +
        public LogScannerImpl​(org.apache.fluss.config.Configuration conf,
        +                      org.apache.fluss.metadata.TableInfo tableInfo,
        +                      MetadataUpdater metadataUpdater,
        +                      org.apache.fluss.rpc.metrics.ClientMetricGroup clientMetricGroup,
        +                      RemoteFileDownloader remoteFileDownloader,
        +                      @Nullable
        +                      int[] projectedFields)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        poll

        +
        public ScanRecords poll​(Duration timeout)
        +
        Description copied from interface: LogScanner
        +
        Poll log data from tablet server. + +

        On each poll, scanner will try to use the last scanned offset as the starting offset and + fetch sequentially. The latest offset will be cached in memory, and if there is no offset + cached in memory, scanner will start either from the earliest offset or from the latest + offset for each bucket according to the policy.

        +
        +
        Specified by:
        +
        poll in interface LogScanner
        +
        Parameters:
        +
        timeout - the timeout to poll.
        +
        Returns:
        +
        the result of poll.
        +
        +
      • +
      + + + +
        +
      • +

        subscribe

        +
        public void subscribe​(int bucket,
        +                      long offset)
        +
        Description copied from interface: LogScanner
        +
        Subscribe to the given table bucket in given offset dynamically. If the table bucket is + already subscribed, the offset will be updated. + +

        Please use LogScanner.subscribe(long, int, long) to subscribe a partitioned table.

        +
        +
        Specified by:
        +
        subscribe in interface LogScanner
        +
        Parameters:
        +
        bucket - the table bucket to subscribe.
        +
        offset - the offset to start from.
        +
        +
      • +
      + + + +
        +
      • +

        subscribe

        +
        public void subscribe​(long partitionId,
        +                      int bucket,
        +                      long offset)
        +
        Description copied from interface: LogScanner
        +
        Subscribe to the given partitioned table bucket in given offset dynamically. If the table + bucket is already subscribed, the offset will be updated. + +

        Please use LogScanner.subscribe(int, long) to subscribe a non-partitioned table.

        +
        +
        Specified by:
        +
        subscribe in interface LogScanner
        +
        Parameters:
        +
        partitionId - the partition id of the table partition to subscribe.
        +
        bucket - the table bucket to subscribe.
        +
        offset - the offset to start from.
        +
        +
      • +
      + + + +
        +
      • +

        unsubscribe

        +
        public void unsubscribe​(long partitionId,
        +                        int bucket)
        +
        Description copied from interface: LogScanner
        +
        Unsubscribe from the given bucket of given partition dynamically.
        +
        +
        Specified by:
        +
        unsubscribe in interface LogScanner
        +
        Parameters:
        +
        partitionId - the partition id of the table partition to unsubscribe.
        +
        bucket - the table bucket to unsubscribe.
        +
        +
      • +
      + + + + + + + +
        +
      • +

        close

        +
        public void close()
        +
        +
        Specified by:
        +
        close in interface AutoCloseable
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScannerStatus.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScannerStatus.html new file mode 100644 index 0000000000..7e944852be --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScannerStatus.html @@ -0,0 +1,271 @@ + + + + + +LogScannerStatus (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LogScannerStatus

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.table.scanner.log.LogScannerStatus
    • +
    +
  • +
+
+
    +
  • +
    +
    @ThreadSafe
    +@Internal
    +public class LogScannerStatus
    +extends Object
    +
    The status of a LogScanner.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LogScannerStatus

        +
        public LogScannerStatus()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/RemoteLogDownloadFuture.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/RemoteLogDownloadFuture.html new file mode 100644 index 0000000000..887caf6a81 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/RemoteLogDownloadFuture.html @@ -0,0 +1,351 @@ + + + + + +RemoteLogDownloadFuture (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class RemoteLogDownloadFuture

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.table.scanner.log.RemoteLogDownloadFuture
    • +
    +
  • +
+
+
    +
  • +
    +
    public class RemoteLogDownloadFuture
    +extends Object
    +
    Represents the future of a remote log download request.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    + +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        isDone

        +
        public boolean isDone()
        +
      • +
      + + + +
        +
      • +

        getFileLogRecords

        +
        public org.apache.fluss.record.FileLogRecords getFileLogRecords​(int startPosition)
        +
      • +
      + + + +
        +
      • +

        getRecycleCallback

        +
        public Runnable getRecycleCallback()
        +
      • +
      + + + +
        +
      • +

        onComplete

        +
        public void onComplete​(Runnable callback)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/RemoteLogDownloader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/RemoteLogDownloader.html new file mode 100644 index 0000000000..3ef589fba8 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/RemoteLogDownloader.html @@ -0,0 +1,378 @@ + + + + + +RemoteLogDownloader (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class RemoteLogDownloader

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.table.scanner.log.RemoteLogDownloader
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    Closeable, AutoCloseable
    +
    +
    +
    @ThreadSafe
    +@Internal
    +public class RemoteLogDownloader
    +extends Object
    +implements Closeable
    +
    Downloader to read remote log files to local disk.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        RemoteLogDownloader

        +
        public RemoteLogDownloader​(org.apache.fluss.metadata.TablePath tablePath,
        +                           org.apache.fluss.config.Configuration conf,
        +                           RemoteFileDownloader remoteFileDownloader,
        +                           ScannerMetricGroup scannerMetricGroup)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        start

        +
        public void start()
        +
      • +
      + + + +
        +
      • +

        requestRemoteLog

        +
        public RemoteLogDownloadFuture requestRemoteLog​(org.apache.fluss.fs.FsPath logTabletDir,
        +                                                org.apache.fluss.remote.RemoteLogSegment segment)
        +
        Request to fetch remote log segment to local. This method is non-blocking.
        +
      • +
      + + + + + + + +
        +
      • +

        getFsPathAndFileName

        +
        protected static org.apache.fluss.fs.FsPathAndFileName getFsPathAndFileName​(org.apache.fluss.fs.FsPath remoteLogTabletDir,
        +                                                                            org.apache.fluss.remote.RemoteLogSegment segment)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/ScanRecords.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/ScanRecords.html new file mode 100644 index 0000000000..ea78b81fd8 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/ScanRecords.html @@ -0,0 +1,445 @@ + + + + + +ScanRecords (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class ScanRecords

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.table.scanner.log.ScanRecords
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    + +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        ScanRecords

        +
        public ScanRecords​(Map<org.apache.fluss.metadata.TableBucket,​List<ScanRecord>> records)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        records

        +
        public List<ScanRecord> records​(org.apache.fluss.metadata.TableBucket scanBucket)
        +
        Get just the records for the given bucketId.
        +
        +
        Parameters:
        +
        scanBucket - The bucket to get records for
        +
        +
      • +
      + + + +
        +
      • +

        buckets

        +
        public Set<org.apache.fluss.metadata.TableBucket> buckets()
        +
        Get the bucket ids which have records contained in this record set.
        +
        +
        Returns:
        +
        the set of partitions with data in this record set (maybe empty if no data was + returned)
        +
        +
      • +
      + + + +
        +
      • +

        count

        +
        public int count()
        +
        The number of records for all buckets.
        +
      • +
      + + + +
        +
      • +

        isEmpty

        +
        public boolean isEmpty()
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetchBuffer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetchBuffer.html new file mode 100644 index 0000000000..f112afd140 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetchBuffer.html @@ -0,0 +1,194 @@ + + + + + +Uses of Class org.apache.fluss.client.table.scanner.log.LogFetchBuffer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.scanner.log.LogFetchBuffer

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetchCollector.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetchCollector.html new file mode 100644 index 0000000000..1ff1177f3d --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetchCollector.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.table.scanner.log.LogFetchCollector (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.scanner.log.LogFetchCollector

+
+
No usage of org.apache.fluss.client.table.scanner.log.LogFetchCollector
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetcher.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetcher.html new file mode 100644 index 0000000000..386a2a300a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetcher.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.table.scanner.log.LogFetcher (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.scanner.log.LogFetcher

+
+
No usage of org.apache.fluss.client.table.scanner.log.LogFetcher
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScan.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScan.html new file mode 100644 index 0000000000..05a1f661fe --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScan.html @@ -0,0 +1,194 @@ + + + + + +Uses of Class org.apache.fluss.client.table.scanner.log.LogScan (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.scanner.log.LogScan

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScanner.html new file mode 100644 index 0000000000..687fc1c7a3 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScanner.html @@ -0,0 +1,227 @@ + + + + + +Uses of Interface org.apache.fluss.client.table.scanner.log.LogScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.client.table.scanner.log.LogScanner

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScannerImpl.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScannerImpl.html new file mode 100644 index 0000000000..4d28490e78 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScannerImpl.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.table.scanner.log.LogScannerImpl (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.scanner.log.LogScannerImpl

+
+
No usage of org.apache.fluss.client.table.scanner.log.LogScannerImpl
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScannerStatus.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScannerStatus.html new file mode 100644 index 0000000000..7bc84832b0 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScannerStatus.html @@ -0,0 +1,203 @@ + + + + + +Uses of Class org.apache.fluss.client.table.scanner.log.LogScannerStatus (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.scanner.log.LogScannerStatus

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/RemoteLogDownloadFuture.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/RemoteLogDownloadFuture.html new file mode 100644 index 0000000000..0fb9637dbc --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/RemoteLogDownloadFuture.html @@ -0,0 +1,195 @@ + + + + + +Uses of Class org.apache.fluss.client.table.scanner.log.RemoteLogDownloadFuture (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.scanner.log.RemoteLogDownloadFuture

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/RemoteLogDownloader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/RemoteLogDownloader.html new file mode 100644 index 0000000000..7edfade352 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/RemoteLogDownloader.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.table.scanner.log.RemoteLogDownloader (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.scanner.log.RemoteLogDownloader

+
+
No usage of org.apache.fluss.client.table.scanner.log.RemoteLogDownloader
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/ScanRecords.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/ScanRecords.html new file mode 100644 index 0000000000..1df5840ed6 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/ScanRecords.html @@ -0,0 +1,214 @@ + + + + + +Uses of Class org.apache.fluss.client.table.scanner.log.ScanRecords (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.scanner.log.ScanRecords

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-summary.html new file mode 100644 index 0000000000..ae4e7c6a9b --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-summary.html @@ -0,0 +1,235 @@ + + + + + +org.apache.fluss.client.table.scanner.log (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.client.table.scanner.log

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-tree.html new file mode 100644 index 0000000000..ca0798d3f8 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-tree.html @@ -0,0 +1,181 @@ + + + + + +org.apache.fluss.client.table.scanner.log Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.client.table.scanner.log

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+

Interface Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-use.html new file mode 100644 index 0000000000..b00018f3f1 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-use.html @@ -0,0 +1,239 @@ + + + + + +Uses of Package org.apache.fluss.client.table.scanner.log (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.client.table.scanner.log

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-summary.html new file mode 100644 index 0000000000..14fee667ce --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-summary.html @@ -0,0 +1,198 @@ + + + + + +org.apache.fluss.client.table.scanner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.client.table.scanner

+
+
+
    +
  • + + + + + + + + + + + + +
    Interface Summary 
    InterfaceDescription
    Scan +
    Used to configure and create a scanner to scan data for a table.
    +
    +
  • +
  • + + + + + + + + + + + + + + + + + + + + +
    Class Summary 
    ClassDescription
    RemoteFileDownloader +
    The downloader that has a IO thread pool to download the remote files (like kv snapshots files, + log segment files).
    +
    ScanRecord +
    one scan record.
    +
    TableScan +
    API for configuring and creating LogScanner and BatchScanner.
    +
    +
  • +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-tree.html new file mode 100644 index 0000000000..950aedd9f5 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-tree.html @@ -0,0 +1,171 @@ + + + + + +org.apache.fluss.client.table.scanner Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.client.table.scanner

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+
    +
  • java.lang.Object +
      +
    • org.apache.fluss.client.table.scanner.RemoteFileDownloader (implements java.io.Closeable)
    • +
    • org.apache.fluss.client.table.scanner.ScanRecord (implements org.apache.fluss.record.LogRecord)
    • +
    • org.apache.fluss.client.table.scanner.TableScan (implements org.apache.fluss.client.table.scanner.Scan)
    • +
    +
  • +
+
+
+

Interface Hierarchy

+
    +
  • org.apache.fluss.client.table.scanner.Scan
  • +
+
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-use.html new file mode 100644 index 0000000000..208e6182c9 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-use.html @@ -0,0 +1,332 @@ + + + + + +Uses of Package org.apache.fluss.client.table.scanner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.client.table.scanner

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AbstractTableWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AbstractTableWriter.html new file mode 100644 index 0000000000..ca23cb24b7 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AbstractTableWriter.html @@ -0,0 +1,443 @@ + + + + + +AbstractTableWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class AbstractTableWriter

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.table.writer.AbstractTableWriter
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + +
        +
      • +

        tablePath

        +
        protected final org.apache.fluss.metadata.TablePath tablePath
        +
      • +
      + + + +
        +
      • +

        writerClient

        +
        protected final WriterClient writerClient
        +
      • +
      + + + +
        +
      • +

        fieldCount

        +
        protected final int fieldCount
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        AbstractTableWriter

        +
        protected AbstractTableWriter​(org.apache.fluss.metadata.TablePath tablePath,
        +                              org.apache.fluss.metadata.TableInfo tableInfo,
        +                              WriterClient writerClient)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        flush

        +
        public void flush()
        +
        Flush data written that have not yet been sent to the server, forcing the client to send the + requests to server and blocks on the completion of the requests associated with these + records. A request is considered completed when it is successfully acknowledged according to + the ConfigOptions.CLIENT_WRITER_ACKS configuration you have specified or else it + results in an error.
        +
        +
        Specified by:
        +
        flush in interface TableWriter
        +
        +
      • +
      + + + + + + + +
        +
      • +

        getPhysicalPath

        +
        protected org.apache.fluss.metadata.PhysicalTablePath getPhysicalPath​(org.apache.fluss.row.InternalRow row)
        +
      • +
      + + + +
        +
      • +

        checkFieldCount

        +
        protected void checkFieldCount​(org.apache.fluss.row.InternalRow row)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/Append.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/Append.html new file mode 100644 index 0000000000..06f795ca1e --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/Append.html @@ -0,0 +1,265 @@ + + + + + +Append (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface Append

+
+
+
+
    +
  • +
    +
    All Known Implementing Classes:
    +
    TableAppend
    +
    +
    +
    @PublicEvolving
    +public interface Append
    +
    Used to configure and create a AppendWriter to write data to a Log Table. + +

    Append objects are immutable and can be shared between threads.

    +
    +
    Since:
    +
    0.6
    +
    +
  • +
+
+
+ +
+
+ +
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AppendResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AppendResult.html new file mode 100644 index 0000000000..79f99a359e --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AppendResult.html @@ -0,0 +1,274 @@ + + + + + +AppendResult (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class AppendResult

+
+
+ +
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        AppendResult

        +
        public AppendResult()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AppendWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AppendWriter.html new file mode 100644 index 0000000000..269dcbba63 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AppendWriter.html @@ -0,0 +1,277 @@ + + + + + +AppendWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface AppendWriter

+
+
+
+
    +
  • +
    +
    All Superinterfaces:
    +
    TableWriter
    +
    +
    +
    @PublicEvolving
    +public interface AppendWriter
    +extends TableWriter
    +
    The writer to write data to the log table.
    +
    +
    Since:
    +
    0.2
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        append

        +
        CompletableFuture<AppendResult> append​(org.apache.fluss.row.InternalRow row)
        +
        Append row into a Log Table.
        +
        +
        Parameters:
        +
        row - the row to append.
        +
        Returns:
        +
        A CompletableFuture that always returns append result when complete normally.
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/DeleteResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/DeleteResult.html new file mode 100644 index 0000000000..5279c4e11d --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/DeleteResult.html @@ -0,0 +1,274 @@ + + + + + +DeleteResult (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class DeleteResult

+
+
+ +
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        DeleteResult

        +
        public DeleteResult()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableAppend.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableAppend.html new file mode 100644 index 0000000000..0eb32d06da --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableAppend.html @@ -0,0 +1,324 @@ + + + + + +TableAppend (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TableAppend

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    Append
    +
    +
    +
    public class TableAppend
    +extends Object
    +implements Append
    +
    API for configuring and creating AppendWriter.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        TableAppend

        +
        public TableAppend​(org.apache.fluss.metadata.TablePath tablePath,
        +                   org.apache.fluss.metadata.TableInfo tableInfo,
        +                   WriterClient writerClient)
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableUpsert.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableUpsert.html new file mode 100644 index 0000000000..5e424bdbd4 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableUpsert.html @@ -0,0 +1,384 @@ + + + + + +TableUpsert (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TableUpsert

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    Upsert
    +
    +
    +
    public class TableUpsert
    +extends Object
    +implements Upsert
    +
    API for configuring and creating UpsertWriter.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        TableUpsert

        +
        public TableUpsert​(org.apache.fluss.metadata.TablePath tablePath,
        +                   org.apache.fluss.metadata.TableInfo tableInfo,
        +                   WriterClient writerClient)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        partialUpdate

        +
        public Upsert partialUpdate​(@Nullable
        +                            int[] targetColumns)
        +
        Description copied from interface: Upsert
        +
        Apply partial update columns and returns a new Upsert instance. + +

        For UpsertWriter.upsert(InternalRow) operation, only the specified columns will be + updated and other columns will remain unchanged if the row exists or set to null if the row + doesn't exist. + +

        For UpsertWriter.delete(InternalRow) operation, the entire row will not be + removed, but only the specified columns except primary key will be set to null. The entire + row will be removed when all columns except primary key are null after a UpsertWriter.delete(InternalRow) operation. + +

        Note: The specified columns must be a contains all columns of primary key, and all columns + except primary key should be nullable.

        +
        +
        Specified by:
        +
        partialUpdate in interface Upsert
        +
        Parameters:
        +
        targetColumns - the column indexes to partial update
        +
        +
      • +
      + + + +
        +
      • +

        partialUpdate

        +
        public Upsert partialUpdate​(String... targetColumnNames)
        +
        +
        Specified by:
        +
        partialUpdate in interface Upsert
        +
        Parameters:
        +
        targetColumnNames - the column names to partial update
        +
        See Also:
        +
        for more details.
        +
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableWriter.html new file mode 100644 index 0000000000..6bd9d9f062 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableWriter.html @@ -0,0 +1,275 @@ + + + + + +TableWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface TableWriter

+
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Summary

      + + + + + + + + + + + + +
      All Methods Instance Methods Abstract Methods 
      Modifier and TypeMethodDescription
      voidflush() +
      Flush data written that have not yet been sent to the server, forcing the client to send the + requests to server and blocks on the completion of the requests associated with these + records.
      +
      +
    • +
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        flush

        +
        void flush()
        +
        Flush data written that have not yet been sent to the server, forcing the client to send the + requests to server and blocks on the completion of the requests associated with these + records. A request is considered completed when it is successfully acknowledged according to + the ConfigOptions.CLIENT_WRITER_ACKS configuration you have specified or else it + results in an error.
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/Upsert.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/Upsert.html new file mode 100644 index 0000000000..4b46328362 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/Upsert.html @@ -0,0 +1,322 @@ + + + + + +Upsert (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface Upsert

+
+
+
+
    +
  • +
    +
    All Known Implementing Classes:
    +
    TableUpsert
    +
    +
    +
    @PublicEvolving
    +public interface Upsert
    +
    Used to configure and create UpsertWriter to upsert and delete data to a Primary Key + Table. + +

    Upsert objects are immutable and can be shared between threads. Refinement methods, + like partialUpdate(int[]), create new Upsert instances.

    +
    +
    Since:
    +
    0.6
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        partialUpdate

        +
        Upsert partialUpdate​(@Nullable
        +                     int[] targetColumns)
        +
        Apply partial update columns and returns a new Upsert instance. + +

        For UpsertWriter.upsert(InternalRow) operation, only the specified columns will be + updated and other columns will remain unchanged if the row exists or set to null if the row + doesn't exist. + +

        For UpsertWriter.delete(InternalRow) operation, the entire row will not be + removed, but only the specified columns except primary key will be set to null. The entire + row will be removed when all columns except primary key are null after a UpsertWriter.delete(InternalRow) operation. + +

        Note: The specified columns must be a contains all columns of primary key, and all columns + except primary key should be nullable.

        +
        +
        Parameters:
        +
        targetColumns - the column indexes to partial update
        +
        +
      • +
      + + + +
        +
      • +

        partialUpdate

        +
        Upsert partialUpdate​(String... targetColumnNames)
        +
        +
        Parameters:
        +
        targetColumnNames - the column names to partial update
        +
        See Also:
        +
        for more details.
        +
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertResult.html new file mode 100644 index 0000000000..4666920ba2 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertResult.html @@ -0,0 +1,274 @@ + + + + + +UpsertResult (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class UpsertResult

+
+
+ +
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        UpsertResult

        +
        public UpsertResult()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertWrite.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertWrite.html new file mode 100644 index 0000000000..26f95ac8f9 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertWrite.html @@ -0,0 +1,352 @@ + + + + + +UpsertWrite (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class UpsertWrite

+
+
+ +
+
    +
  • +
    +
    @PublicEvolving
    +public class UpsertWrite
    +extends Object
    +
    Used to describe the operation to write data by UpsertWriter to a table.
    +
    +
    Since:
    +
    0.2
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        UpsertWrite

        +
        public UpsertWrite()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        withPartialUpdate

        +
        public UpsertWrite withPartialUpdate​(int[] targetColumns)
        +
        Returns a new instance of UpsertWrite description with partial update the specified columns. + +

        For UpsertWriter.upsert(InternalRow) operation, only the specified columns will be + updated and other columns will remain unchanged if the row exists or set to null if the row + doesn't exist. + +

        For UpsertWriter.delete(InternalRow) operation, the entire row will not be + removed, but only the specified columns except primary key will be set to null. The entire + row will be removed when all columns except primary key are null after a UpsertWriter.delete(InternalRow) operation. + +

        Note: The specified columns must be a contains all columns of primary key, and all columns + except primary key should be nullable.

        +
        +
        Parameters:
        +
        targetColumns - the columns to partial update,
        +
        +
      • +
      + + + +
        +
      • +

        getPartialUpdateColumns

        +
        @Nullable
        +public int[] getPartialUpdateColumns()
        +
        Returns the columns to partial update. Returns null if update all columns.
        +
        +
        Returns:
        +
        the columns to partial update.
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertWriter.html new file mode 100644 index 0000000000..572b271623 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertWriter.html @@ -0,0 +1,302 @@ + + + + + +UpsertWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface UpsertWriter

+
+
+
+
    +
  • +
    +
    All Superinterfaces:
    +
    TableWriter
    +
    +
    +
    @PublicEvolving
    +public interface UpsertWriter
    +extends TableWriter
    +
    The writer to write data to the primary key table.
    +
    +
    Since:
    +
    0.2
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        upsert

        +
        CompletableFuture<UpsertResult> upsert​(org.apache.fluss.row.InternalRow row)
        +
        Inserts row into Fluss table if they do not already exist, or updates them if they do exist.
        +
        +
        Parameters:
        +
        row - the row to upsert.
        +
        Returns:
        +
        A CompletableFuture that always returns upsert result when complete normally.
        +
        +
      • +
      + + + +
        +
      • +

        delete

        +
        CompletableFuture<DeleteResult> delete​(org.apache.fluss.row.InternalRow row)
        +
        Delete certain row by the input row in Fluss table, the input row must contain the primary + key.
        +
        +
        Parameters:
        +
        row - the row to delete.
        +
        Returns:
        +
        A CompletableFuture that always delete result when complete normally.
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AbstractTableWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AbstractTableWriter.html new file mode 100644 index 0000000000..73ea9aa792 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AbstractTableWriter.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.table.writer.AbstractTableWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.writer.AbstractTableWriter

+
+
No usage of org.apache.fluss.client.table.writer.AbstractTableWriter
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/Append.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/Append.html new file mode 100644 index 0000000000..085bff6a69 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/Append.html @@ -0,0 +1,228 @@ + + + + + +Uses of Interface org.apache.fluss.client.table.writer.Append (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.client.table.writer.Append

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AppendResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AppendResult.html new file mode 100644 index 0000000000..723de1052b --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AppendResult.html @@ -0,0 +1,194 @@ + + + + + +Uses of Class org.apache.fluss.client.table.writer.AppendResult (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.writer.AppendResult

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AppendWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AppendWriter.html new file mode 100644 index 0000000000..06c8fbae39 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AppendWriter.html @@ -0,0 +1,199 @@ + + + + + +Uses of Interface org.apache.fluss.client.table.writer.AppendWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.client.table.writer.AppendWriter

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/DeleteResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/DeleteResult.html new file mode 100644 index 0000000000..b4caa93de4 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/DeleteResult.html @@ -0,0 +1,195 @@ + + + + + +Uses of Class org.apache.fluss.client.table.writer.DeleteResult (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.writer.DeleteResult

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableAppend.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableAppend.html new file mode 100644 index 0000000000..444502f386 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableAppend.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.table.writer.TableAppend (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.writer.TableAppend

+
+
No usage of org.apache.fluss.client.table.writer.TableAppend
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableUpsert.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableUpsert.html new file mode 100644 index 0000000000..105586ccef --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableUpsert.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.table.writer.TableUpsert (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.writer.TableUpsert

+
+
No usage of org.apache.fluss.client.table.writer.TableUpsert
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableWriter.html new file mode 100644 index 0000000000..b359c9b2dc --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableWriter.html @@ -0,0 +1,218 @@ + + + + + +Uses of Interface org.apache.fluss.client.table.writer.TableWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.client.table.writer.TableWriter

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/Upsert.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/Upsert.html new file mode 100644 index 0000000000..bb4715456e --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/Upsert.html @@ -0,0 +1,260 @@ + + + + + +Uses of Interface org.apache.fluss.client.table.writer.Upsert (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.client.table.writer.Upsert

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertResult.html new file mode 100644 index 0000000000..0a4ccc2776 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertResult.html @@ -0,0 +1,194 @@ + + + + + +Uses of Class org.apache.fluss.client.table.writer.UpsertResult (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.writer.UpsertResult

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertWrite.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertWrite.html new file mode 100644 index 0000000000..77198995ee --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertWrite.html @@ -0,0 +1,194 @@ + + + + + +Uses of Class org.apache.fluss.client.table.writer.UpsertWrite (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.table.writer.UpsertWrite

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertWriter.html new file mode 100644 index 0000000000..dd17066ba6 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertWriter.html @@ -0,0 +1,200 @@ + + + + + +Uses of Interface org.apache.fluss.client.table.writer.UpsertWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.client.table.writer.UpsertWriter

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-summary.html new file mode 100644 index 0000000000..ffa217be8d --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-summary.html @@ -0,0 +1,246 @@ + + + + + +org.apache.fluss.client.table.writer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.client.table.writer

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-tree.html new file mode 100644 index 0000000000..e1bb348e82 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-tree.html @@ -0,0 +1,182 @@ + + + + + +org.apache.fluss.client.table.writer Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.client.table.writer

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+

Interface Hierarchy

+
    +
  • org.apache.fluss.client.table.writer.Append
  • +
  • org.apache.fluss.client.table.writer.TableWriter + +
  • +
  • org.apache.fluss.client.table.writer.Upsert
  • +
+
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-use.html new file mode 100644 index 0000000000..8cc4618d20 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-use.html @@ -0,0 +1,264 @@ + + + + + +Uses of Package org.apache.fluss.client.table.writer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.client.table.writer

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/DefaultSecurityTokenManager.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/DefaultSecurityTokenManager.html new file mode 100644 index 0000000000..d7b3a2302f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/DefaultSecurityTokenManager.html @@ -0,0 +1,360 @@ + + + + + +DefaultSecurityTokenManager (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class DefaultSecurityTokenManager

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.token.DefaultSecurityTokenManager
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        DefaultSecurityTokenManager

        +
        public DefaultSecurityTokenManager​(org.apache.fluss.config.Configuration configuration,
        +                                   SecurityTokenProvider securityTokenProvider)
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/DefaultSecurityTokenProvider.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/DefaultSecurityTokenProvider.html new file mode 100644 index 0000000000..d8d96b7f18 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/DefaultSecurityTokenProvider.html @@ -0,0 +1,325 @@ + + + + + +DefaultSecurityTokenProvider (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class DefaultSecurityTokenProvider

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.token.DefaultSecurityTokenProvider
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        DefaultSecurityTokenProvider

        +
        public DefaultSecurityTokenProvider​(org.apache.fluss.rpc.gateway.AdminReadOnlyGateway adminReadOnlyGateway)
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/SecurityTokenManager.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/SecurityTokenManager.html new file mode 100644 index 0000000000..cf24d051cf --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/SecurityTokenManager.html @@ -0,0 +1,280 @@ + + + + + +SecurityTokenManager (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface SecurityTokenManager

+
+
+
+
    +
  • +
    +
    All Known Implementing Classes:
    +
    DefaultSecurityTokenManager
    +
    +
    +
    public interface SecurityTokenManager
    +
    API for security token manager.
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Method Summary

      + + + + + + + + + + + + + + + + + +
      All Methods Instance Methods Abstract Methods 
      Modifier and TypeMethodDescription
      voidstart() +
      Creates a re-occurring task which obtains new tokens.
      +
      voidstop() +
      Stops re-occurring token obtain task.
      +
      +
    • +
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        start

        +
        void start()
        +    throws Exception
        +
        Creates a re-occurring task which obtains new tokens.
        +
        +
        Throws:
        +
        Exception
        +
        +
      • +
      + + + +
        +
      • +

        stop

        +
        void stop()
        +
        Stops re-occurring token obtain task.
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/SecurityTokenProvider.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/SecurityTokenProvider.html new file mode 100644 index 0000000000..949327705f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/SecurityTokenProvider.html @@ -0,0 +1,265 @@ + + + + + +SecurityTokenProvider (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface SecurityTokenProvider

+
+
+
+
    +
  • +
    +
    All Known Implementing Classes:
    +
    DefaultSecurityTokenProvider
    +
    +
    +
    public interface SecurityTokenProvider
    +
    Security token provider API.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        obtainSecurityToken

        +
        org.apache.fluss.fs.token.ObtainedSecurityToken obtainSecurityToken()
        +                                                             throws Exception
        +
        Obtain security token.
        +
        +
        Returns:
        +
        the obtained security token.
        +
        Throws:
        +
        Exception
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/DefaultSecurityTokenManager.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/DefaultSecurityTokenManager.html new file mode 100644 index 0000000000..5fee791735 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/DefaultSecurityTokenManager.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.token.DefaultSecurityTokenManager (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.token.DefaultSecurityTokenManager

+
+
No usage of org.apache.fluss.client.token.DefaultSecurityTokenManager
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/DefaultSecurityTokenProvider.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/DefaultSecurityTokenProvider.html new file mode 100644 index 0000000000..616cb1a45a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/DefaultSecurityTokenProvider.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.token.DefaultSecurityTokenProvider (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.token.DefaultSecurityTokenProvider

+
+
No usage of org.apache.fluss.client.token.DefaultSecurityTokenProvider
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/SecurityTokenManager.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/SecurityTokenManager.html new file mode 100644 index 0000000000..43e5a197aa --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/SecurityTokenManager.html @@ -0,0 +1,194 @@ + + + + + +Uses of Interface org.apache.fluss.client.token.SecurityTokenManager (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.client.token.SecurityTokenManager

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/SecurityTokenProvider.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/SecurityTokenProvider.html new file mode 100644 index 0000000000..f44ca74b92 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/SecurityTokenProvider.html @@ -0,0 +1,208 @@ + + + + + +Uses of Interface org.apache.fluss.client.token.SecurityTokenProvider (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.client.token.SecurityTokenProvider

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-summary.html new file mode 100644 index 0000000000..1595497445 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-summary.html @@ -0,0 +1,197 @@ + + + + + +org.apache.fluss.client.token (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.client.token

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-tree.html new file mode 100644 index 0000000000..44327a8226 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-tree.html @@ -0,0 +1,171 @@ + + + + + +org.apache.fluss.client.token Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.client.token

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+

Interface Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-use.html new file mode 100644 index 0000000000..a5c0a11515 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-use.html @@ -0,0 +1,191 @@ + + + + + +Uses of Package org.apache.fluss.client.token (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.client.token

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/ClientRpcMessageUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/ClientRpcMessageUtils.html new file mode 100644 index 0000000000..3198c1c8ad --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/ClientRpcMessageUtils.html @@ -0,0 +1,556 @@ + + + + + +ClientRpcMessageUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class ClientRpcMessageUtils

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.utils.ClientRpcMessageUtils
    • +
    +
  • +
+
+
    +
  • +
    +
    public class ClientRpcMessageUtils
    +extends Object
    +
    Utils for making rpc request/response from inner object or convert inner class to rpc + request/response for client.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        ClientRpcMessageUtils

        +
        public ClientRpcMessageUtils()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        makeProduceLogRequest

        +
        public static org.apache.fluss.rpc.messages.ProduceLogRequest makeProduceLogRequest​(long tableId,
        +                                                                                    int acks,
        +                                                                                    int maxRequestTimeoutMs,
        +                                                                                    List<ReadyWriteBatch> readyBatches)
        +
      • +
      + + + +
        +
      • +

        makePutKvRequest

        +
        public static org.apache.fluss.rpc.messages.PutKvRequest makePutKvRequest​(long tableId,
        +                                                                          int acks,
        +                                                                          int maxRequestTimeoutMs,
        +                                                                          List<ReadyWriteBatch> readyWriteBatches)
        +
      • +
      + + + +
        +
      • +

        makeLookupRequest

        +
        public static org.apache.fluss.rpc.messages.LookupRequest makeLookupRequest​(long tableId,
        +                                                                            Collection<LookupBatch> lookupBatches)
        +
      • +
      + + + +
        +
      • +

        makePrefixLookupRequest

        +
        public static org.apache.fluss.rpc.messages.PrefixLookupRequest makePrefixLookupRequest​(long tableId,
        +                                                                                        Collection<PrefixLookupBatch> lookupBatches)
        +
      • +
      + + + +
        +
      • +

        toKvSnapshots

        +
        public static KvSnapshots toKvSnapshots​(org.apache.fluss.rpc.messages.GetLatestKvSnapshotsResponse response)
        +
      • +
      + + + +
        +
      • +

        toKvSnapshotMetadata

        +
        public static KvSnapshotMetadata toKvSnapshotMetadata​(org.apache.fluss.rpc.messages.GetKvSnapshotMetadataResponse response)
        +
      • +
      + + + +
        +
      • +

        toLakeTableSnapshotInfo

        +
        public static LakeSnapshot toLakeTableSnapshotInfo​(org.apache.fluss.rpc.messages.GetLatestLakeSnapshotResponse response)
        +
      • +
      + + + +
        +
      • +

        toFsPathAndFileName

        +
        public static List<org.apache.fluss.fs.FsPathAndFileName> toFsPathAndFileName​(List<org.apache.fluss.rpc.messages.PbRemotePathAndLocalFile> pbFileHandles)
        +
      • +
      + + + +
        +
      • +

        toSecurityToken

        +
        public static org.apache.fluss.fs.token.ObtainedSecurityToken toSecurityToken​(org.apache.fluss.rpc.messages.GetFileSystemSecurityTokenResponse response)
        +
      • +
      + + + +
        +
      • +

        makeMetadataRequest

        +
        public static org.apache.fluss.rpc.messages.MetadataRequest makeMetadataRequest​(@Nullable
        +                                                                                Set<org.apache.fluss.metadata.TablePath> tablePaths,
        +                                                                                @Nullable
        +                                                                                Collection<org.apache.fluss.metadata.PhysicalTablePath> tablePathPartitionNames,
        +                                                                                @Nullable
        +                                                                                Collection<Long> tablePathPartitionIds)
        +
      • +
      + + + +
        +
      • +

        makeListOffsetsRequest

        +
        public static org.apache.fluss.rpc.messages.ListOffsetsRequest makeListOffsetsRequest​(long tableId,
        +                                                                                      @Nullable
        +                                                                                      Long partitionId,
        +                                                                                      List<Integer> bucketIdList,
        +                                                                                      OffsetSpec offsetSpec)
        +
      • +
      + + + +
        +
      • +

        makeCreatePartitionRequest

        +
        public static org.apache.fluss.rpc.messages.CreatePartitionRequest makeCreatePartitionRequest​(org.apache.fluss.metadata.TablePath tablePath,
        +                                                                                              org.apache.fluss.metadata.PartitionSpec partitionSpec,
        +                                                                                              boolean ignoreIfNotExists)
        +
      • +
      + + + +
        +
      • +

        makeDropPartitionRequest

        +
        public static org.apache.fluss.rpc.messages.DropPartitionRequest makeDropPartitionRequest​(org.apache.fluss.metadata.TablePath tablePath,
        +                                                                                          org.apache.fluss.metadata.PartitionSpec partitionSpec,
        +                                                                                          boolean ignoreIfNotExists)
        +
      • +
      + + + +
        +
      • +

        toPartitionInfos

        +
        public static List<org.apache.fluss.metadata.PartitionInfo> toPartitionInfos​(org.apache.fluss.rpc.messages.ListPartitionInfosResponse response)
        +
      • +
      + + + +
        +
      • +

        toKeyValueMap

        +
        public static Map<String,​String> toKeyValueMap​(List<org.apache.fluss.rpc.messages.PbKeyValue> pbKeyValues)
        +
      • +
      + + + +
        +
      • +

        makePbPartitionSpec

        +
        public static org.apache.fluss.rpc.messages.PbPartitionSpec makePbPartitionSpec​(org.apache.fluss.metadata.PartitionSpec partitionSpec)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/ClientUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/ClientUtils.html new file mode 100644 index 0000000000..ed680cdd9e --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/ClientUtils.html @@ -0,0 +1,343 @@ + + + + + +ClientUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class ClientUtils

+
+
+ +
+
    +
  • +
    +
    public final class ClientUtils
    +extends Object
    +
    Utils for Fluss Client.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + + + + + +
        +
      • +

        getHost

        +
        public static String getHost​(String address)
        +
        Extracts the hostname from a "host:port" address string.
        +
        +
        Parameters:
        +
        address - address string to parse
        +
        Returns:
        +
        hostname or null if the given address is incorrect
        +
        +
      • +
      + + + +
        +
      • +

        getPort

        +
        public static Integer getPort​(String address)
        +
        Extracts the port number from a "host:port" address string.
        +
        +
        Parameters:
        +
        address - address string to parse
        +
        Returns:
        +
        port number or null if the given address is incorrect
        +
        +
      • +
      + + + +
        +
      • +

        getPartitionId

        +
        public static Long getPartitionId​(org.apache.fluss.row.InternalRow row,
        +                                  PartitionGetter partitionGetter,
        +                                  org.apache.fluss.metadata.TablePath tablePath,
        +                                  MetadataUpdater metadataUpdater)
        +                           throws org.apache.fluss.exception.PartitionNotExistException
        +
        Return the id of the partition the row belongs to. It'll try to update the metadata if the + partition doesn't exist. If the partition doesn't exist yet after update metadata, it'll + throw PartitionNotExistException.
        +
        +
        Throws:
        +
        org.apache.fluss.exception.PartitionNotExistException
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/MetadataUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/MetadataUtils.html new file mode 100644 index 0000000000..66feafba5f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/MetadataUtils.html @@ -0,0 +1,420 @@ + + + + + +MetadataUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class MetadataUtils

+
+
+ +
+
    +
  • +
    +
    public class MetadataUtils
    +extends Object
    +
    Utils for metadata for client.
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Summary

      + + + + + + + + + + +
      Constructors 
      ConstructorDescription
      MetadataUtils() 
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Summary

      + + + + + + + + + + + + + + + + + + + + + + + + + + + +
      All Methods Static Methods Concrete Methods 
      Modifier and TypeMethodDescription
      static org.apache.fluss.cluster.ServerNodegetOneAvailableTabletServerNode​(org.apache.fluss.cluster.Cluster cluster) 
      static org.apache.fluss.cluster.ClustersendMetadataRequestAndRebuildCluster​(org.apache.fluss.cluster.Cluster cluster, + org.apache.fluss.rpc.RpcClient client, + Set<org.apache.fluss.metadata.TablePath> tablePaths, + Collection<org.apache.fluss.metadata.PhysicalTablePath> tablePartitionNames, + Collection<Long> tablePartitionIds) +
      Partial update cluster, means we will rebuild the cluster by sending metadata request to + request the input tables/partitions in physicalTablePaths, after that add those + tables/partitions into cluster.
      +
      static org.apache.fluss.cluster.ClustersendMetadataRequestAndRebuildCluster​(org.apache.fluss.rpc.gateway.AdminReadOnlyGateway gateway, + boolean partialUpdate, + org.apache.fluss.cluster.Cluster originCluster, + Set<org.apache.fluss.metadata.TablePath> tablePaths, + Collection<org.apache.fluss.metadata.PhysicalTablePath> tablePartitions, + Collection<Long> tablePartitionIds) +
      maybe partial update cluster.
      +
      static org.apache.fluss.cluster.ClustersendMetadataRequestAndRebuildCluster​(org.apache.fluss.rpc.gateway.AdminReadOnlyGateway gateway, + Set<org.apache.fluss.metadata.TablePath> tablePaths) +
      full update cluster, means we will rebuild the cluster by clearing all cached table in + cluster, and then send metadata request to request the input tables in tablePaths, after that + add those table into cluster.
      +
      + +
    • +
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        MetadataUtils

        +
        public MetadataUtils()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        sendMetadataRequestAndRebuildCluster

        +
        public static org.apache.fluss.cluster.Cluster sendMetadataRequestAndRebuildCluster​(org.apache.fluss.rpc.gateway.AdminReadOnlyGateway gateway,
        +                                                                                    Set<org.apache.fluss.metadata.TablePath> tablePaths)
        +                                                                             throws ExecutionException,
        +                                                                                    InterruptedException,
        +                                                                                    TimeoutException
        +
        full update cluster, means we will rebuild the cluster by clearing all cached table in + cluster, and then send metadata request to request the input tables in tablePaths, after that + add those table into cluster.
        +
        +
        Throws:
        +
        ExecutionException
        +
        InterruptedException
        +
        TimeoutException
        +
        +
      • +
      + + + +
        +
      • +

        sendMetadataRequestAndRebuildCluster

        +
        public static org.apache.fluss.cluster.Cluster sendMetadataRequestAndRebuildCluster​(org.apache.fluss.cluster.Cluster cluster,
        +                                                                                    org.apache.fluss.rpc.RpcClient client,
        +                                                                                    @Nullable
        +                                                                                    Set<org.apache.fluss.metadata.TablePath> tablePaths,
        +                                                                                    @Nullable
        +                                                                                    Collection<org.apache.fluss.metadata.PhysicalTablePath> tablePartitionNames,
        +                                                                                    @Nullable
        +                                                                                    Collection<Long> tablePartitionIds)
        +                                                                             throws ExecutionException,
        +                                                                                    InterruptedException,
        +                                                                                    TimeoutException
        +
        Partial update cluster, means we will rebuild the cluster by sending metadata request to + request the input tables/partitions in physicalTablePaths, after that add those + tables/partitions into cluster. The origin tables/partitions in cluster will not be cleared, + but will be updated.
        +
        +
        Throws:
        +
        ExecutionException
        +
        InterruptedException
        +
        TimeoutException
        +
        +
      • +
      + + + + + + + +
        +
      • +

        getOneAvailableTabletServerNode

        +
        public static org.apache.fluss.cluster.ServerNode getOneAvailableTabletServerNode​(org.apache.fluss.cluster.Cluster cluster)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/ClientRpcMessageUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/ClientRpcMessageUtils.html new file mode 100644 index 0000000000..bb1d01f261 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/ClientRpcMessageUtils.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.utils.ClientRpcMessageUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.utils.ClientRpcMessageUtils

+
+
No usage of org.apache.fluss.client.utils.ClientRpcMessageUtils
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/ClientUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/ClientUtils.html new file mode 100644 index 0000000000..f13ca48e77 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/ClientUtils.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.utils.ClientUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.utils.ClientUtils

+
+
No usage of org.apache.fluss.client.utils.ClientUtils
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/MetadataUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/MetadataUtils.html new file mode 100644 index 0000000000..4351d0c34b --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/MetadataUtils.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.utils.MetadataUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.utils.MetadataUtils

+
+
No usage of org.apache.fluss.client.utils.MetadataUtils
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-summary.html new file mode 100644 index 0000000000..263757778c --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-summary.html @@ -0,0 +1,181 @@ + + + + + +org.apache.fluss.client.utils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.client.utils

+
+
+
    +
  • + + + + + + + + + + + + + + + + + + + + +
    Class Summary 
    ClassDescription
    ClientRpcMessageUtils +
    Utils for making rpc request/response from inner object or convert inner class to rpc + request/response for client.
    +
    ClientUtils +
    Utils for Fluss Client.
    +
    MetadataUtils +
    Utils for metadata for client.
    +
    +
  • +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-tree.html new file mode 100644 index 0000000000..3065645f24 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-tree.html @@ -0,0 +1,165 @@ + + + + + +org.apache.fluss.client.utils Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.client.utils

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-use.html new file mode 100644 index 0000000000..e62e4f11c1 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-use.html @@ -0,0 +1,148 @@ + + + + + +Uses of Package org.apache.fluss.client.utils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.client.utils

+
+
No usage of org.apache.fluss.client.utils
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/ArrowLogWriteBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/ArrowLogWriteBatch.html new file mode 100644 index 0000000000..b977cc6057 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/ArrowLogWriteBatch.html @@ -0,0 +1,597 @@ + + + + + +ArrowLogWriteBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class ArrowLogWriteBatch

+
+
+ +
+
    +
  • +
    +
    @NotThreadSafe
    +@Internal
    +public class ArrowLogWriteBatch
    +extends WriteBatch
    +
    A batch of log records managed in ARROW format that is or will be sent to server by ProduceLogRequest. + +

    This class is not thread safe and external synchronization must be used when modifying it.

    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        ArrowLogWriteBatch

        +
        public ArrowLogWriteBatch​(int bucketId,
        +                          org.apache.fluss.metadata.PhysicalTablePath physicalTablePath,
        +                          int schemaId,
        +                          org.apache.fluss.row.arrow.ArrowWriter arrowWriter,
        +                          org.apache.fluss.memory.AbstractPagedOutputView outputView,
        +                          long createdMs)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        tryAppend

        +
        public boolean tryAppend​(WriteRecord writeRecord,
        +                         WriteCallback callback)
        +                  throws Exception
        +
        Description copied from class: WriteBatch
        +
        try to append one write record to the record batch.
        +
        +
        Specified by:
        +
        tryAppend in class WriteBatch
        +
        Parameters:
        +
        writeRecord - the record to write
        +
        callback - the callback to send back to writer
        +
        Returns:
        +
        true if append success, false if the batch is full.
        +
        Throws:
        +
        Exception
        +
        +
      • +
      + + + +
        +
      • +

        build

        +
        public org.apache.fluss.record.bytesview.BytesView build()
        +
        Description copied from class: WriteBatch
        +
        Gets the memory segment bytes view of the batch. This includes the latest updated WriteBatch.setWriterState(long, int) in the bytes view.
        +
        +
        Specified by:
        +
        build in class WriteBatch
        +
        +
      • +
      + + + + + + + +
        +
      • +

        isClosed

        +
        public boolean isClosed()
        +
        Description copied from class: WriteBatch
        +
        check if the batch is closed.
        +
        +
        Specified by:
        +
        isClosed in class WriteBatch
        +
        Returns:
        +
        true if closed, false otherwise
        +
        +
      • +
      + + + +
        +
      • +

        estimatedSizeInBytes

        +
        public int estimatedSizeInBytes()
        +
        Description copied from class: WriteBatch
        +
        Get an estimate of the number of bytes written to the underlying buffer. The returned value + is exactly correct if the record set is not compressed or if the batch has been WriteBatch.build().
        +
        +
        Specified by:
        +
        estimatedSizeInBytes in class WriteBatch
        +
        +
      • +
      + + + +
        +
      • +

        pooledMemorySegments

        +
        public List<org.apache.fluss.memory.MemorySegment> pooledMemorySegments()
        +
        Description copied from class: WriteBatch
        +
        get pooled memory segments to de-allocate. After produceLog/PutKv acks, the WriteBatch need to de-allocate the allocated pooled MemorySegments back to MemorySegmentPool for reusing.
        +
        +
        Specified by:
        +
        pooledMemorySegments in class WriteBatch
        +
        Returns:
        +
        the pooled memory segment this batch allocated
        +
        +
      • +
      + + + +
        +
      • +

        setWriterState

        +
        public void setWriterState​(long writerId,
        +                           int batchSequence)
        +
        +
        Specified by:
        +
        setWriterState in class WriteBatch
        +
        +
      • +
      + + + +
        +
      • +

        resetWriterState

        +
        public void resetWriterState​(long writerId,
        +                             int batchSequence)
        +
        +
        Overrides:
        +
        resetWriterState in class WriteBatch
        +
        +
      • +
      + + + +
        +
      • +

        writerId

        +
        public long writerId()
        +
        +
        Specified by:
        +
        writerId in class WriteBatch
        +
        +
      • +
      + + + + + + + + +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/DynamicPartitionCreator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/DynamicPartitionCreator.html new file mode 100644 index 0000000000..6527436850 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/DynamicPartitionCreator.html @@ -0,0 +1,314 @@ + + + + + +DynamicPartitionCreator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class DynamicPartitionCreator

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.write.DynamicPartitionCreator
    • +
    +
  • +
+
+
    +
  • +
    +
    @ThreadSafe
    +public class DynamicPartitionCreator
    +extends Object
    +
    A creator to create partition when dynamic partition create enable for table.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        DynamicPartitionCreator

        +
        public DynamicPartitionCreator​(MetadataUpdater metadataUpdater,
        +                               Admin admin,
        +                               boolean dynamicPartitionEnabled,
        +                               Consumer<Throwable> fatalErrorHandler)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        checkAndCreatePartitionAsync

        +
        public void checkAndCreatePartitionAsync​(org.apache.fluss.metadata.PhysicalTablePath physicalTablePath)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/DynamicWriteBatchSizeEstimator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/DynamicWriteBatchSizeEstimator.html new file mode 100644 index 0000000000..aa6048863b --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/DynamicWriteBatchSizeEstimator.html @@ -0,0 +1,329 @@ + + + + + +DynamicWriteBatchSizeEstimator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class DynamicWriteBatchSizeEstimator

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.write.DynamicWriteBatchSizeEstimator
    • +
    +
  • +
+
+
    +
  • +
    +
    @Internal
    +@ThreadSafe
    +public class DynamicWriteBatchSizeEstimator
    +extends Object
    +
    An estimator to estimate the buffer usage of a writeBatch.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        DynamicWriteBatchSizeEstimator

        +
        public DynamicWriteBatchSizeEstimator​(boolean dynamicBatchSizeEnabled,
        +                                      int maxBatchSize,
        +                                      int pageSize)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        updateEstimation

        +
        public void updateEstimation​(org.apache.fluss.metadata.PhysicalTablePath physicalTablePath,
        +                             int observedBatchSize)
        +
      • +
      + + + +
        +
      • +

        getEstimatedBatchSize

        +
        public int getEstimatedBatchSize​(org.apache.fluss.metadata.PhysicalTablePath physicalTablePath)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/HashBucketAssigner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/HashBucketAssigner.html new file mode 100644 index 0000000000..bf7a4e19fd --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/HashBucketAssigner.html @@ -0,0 +1,370 @@ + + + + + +HashBucketAssigner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class HashBucketAssigner

+
+
+ +
+
    +
  • +
    +
    @Internal
    +public class HashBucketAssigner
    +extends Object
    +
    Hash bucket assigner.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        HashBucketAssigner

        +
        public HashBucketAssigner​(int numBuckets)
        +
      • +
      + + + +
        +
      • +

        HashBucketAssigner

        +
        public HashBucketAssigner​(int numBuckets,
        +                          org.apache.fluss.bucketing.BucketingFunction function)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        assignBucket

        +
        public int assignBucket​(byte[] bucketKeys)
        +
      • +
      + + + +
        +
      • +

        assignBucket

        +
        public int assignBucket​(@Nullable
        +                        byte[] bucketKey,
        +                        org.apache.fluss.cluster.Cluster cluster)
        +
      • +
      + + + +
        +
      • +

        abortIfBatchFull

        +
        public boolean abortIfBatchFull()
        +
      • +
      + + + +
        +
      • +

        onNewBatch

        +
        public void onNewBatch​(org.apache.fluss.cluster.Cluster cluster,
        +                       int prevBucketId)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceBucketEntry.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceBucketEntry.html new file mode 100644 index 0000000000..9b9641e3dd --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceBucketEntry.html @@ -0,0 +1,224 @@ + + + + + +IdempotenceBucketEntry (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class IdempotenceBucketEntry

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.write.IdempotenceBucketEntry
    • +
    +
  • +
+
+
    +
  • +
    +
    @Internal
    +public class IdempotenceBucketEntry
    +extends Object
    +
    Entry to store the idempotence information of each table-bucket.
    +
  • +
+
+
+ +
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceBucketMap.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceBucketMap.html new file mode 100644 index 0000000000..e95b2cbf6b --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceBucketMap.html @@ -0,0 +1,270 @@ + + + + + +IdempotenceBucketMap (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class IdempotenceBucketMap

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.write.IdempotenceBucketMap
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        IdempotenceBucketMap

        +
        public IdempotenceBucketMap()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceManager.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceManager.html new file mode 100644 index 0000000000..c6fdd26670 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceManager.html @@ -0,0 +1,276 @@ + + + + + +IdempotenceManager (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class IdempotenceManager

+
+
+ +
+
    +
  • +
    +
    @Internal
    +@ThreadSafe
    +public class IdempotenceManager
    +extends Object
    +
    A class which manages the idempotence in writer keeps the state necessary to ensure idempotent of + writer.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        IdempotenceManager

        +
        public IdempotenceManager​(boolean idempotenceEnabled,
        +                          int maxInflightRequestsPerBucket,
        +                          org.apache.fluss.rpc.gateway.TabletServerGateway tabletServerGateway)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IndexedLogWriteBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IndexedLogWriteBatch.html new file mode 100644 index 0000000000..7a44311887 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IndexedLogWriteBatch.html @@ -0,0 +1,597 @@ + + + + + +IndexedLogWriteBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class IndexedLogWriteBatch

+
+
+ +
+
    +
  • +
    +
    @NotThreadSafe
    +@Internal
    +public final class IndexedLogWriteBatch
    +extends WriteBatch
    +
    A batch of log records managed in INDEXED format that is or will be sent to server by ProduceLogRequest. + +

    This class is not thread safe and external synchronization must be used when modifying it.

    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        IndexedLogWriteBatch

        +
        public IndexedLogWriteBatch​(int bucketId,
        +                            org.apache.fluss.metadata.PhysicalTablePath physicalTablePath,
        +                            int schemaId,
        +                            int writeLimit,
        +                            org.apache.fluss.memory.AbstractPagedOutputView outputView,
        +                            long createdMs)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        tryAppend

        +
        public boolean tryAppend​(WriteRecord writeRecord,
        +                         WriteCallback callback)
        +                  throws Exception
        +
        Description copied from class: WriteBatch
        +
        try to append one write record to the record batch.
        +
        +
        Specified by:
        +
        tryAppend in class WriteBatch
        +
        Parameters:
        +
        writeRecord - the record to write
        +
        callback - the callback to send back to writer
        +
        Returns:
        +
        true if append success, false if the batch is full.
        +
        Throws:
        +
        Exception
        +
        +
      • +
      + + + +
        +
      • +

        build

        +
        public org.apache.fluss.record.bytesview.BytesView build()
        +
        Description copied from class: WriteBatch
        +
        Gets the memory segment bytes view of the batch. This includes the latest updated WriteBatch.setWriterState(long, int) in the bytes view.
        +
        +
        Specified by:
        +
        build in class WriteBatch
        +
        +
      • +
      + + + +
        +
      • +

        isClosed

        +
        public boolean isClosed()
        +
        Description copied from class: WriteBatch
        +
        check if the batch is closed.
        +
        +
        Specified by:
        +
        isClosed in class WriteBatch
        +
        Returns:
        +
        true if closed, false otherwise
        +
        +
      • +
      + + + + + + + +
        +
      • +

        pooledMemorySegments

        +
        public List<org.apache.fluss.memory.MemorySegment> pooledMemorySegments()
        +
        Description copied from class: WriteBatch
        +
        get pooled memory segments to de-allocate. After produceLog/PutKv acks, the WriteBatch need to de-allocate the allocated pooled MemorySegments back to MemorySegmentPool for reusing.
        +
        +
        Specified by:
        +
        pooledMemorySegments in class WriteBatch
        +
        Returns:
        +
        the pooled memory segment this batch allocated
        +
        +
      • +
      + + + +
        +
      • +

        setWriterState

        +
        public void setWriterState​(long writerId,
        +                           int batchSequence)
        +
        +
        Specified by:
        +
        setWriterState in class WriteBatch
        +
        +
      • +
      + + + +
        +
      • +

        writerId

        +
        public long writerId()
        +
        +
        Specified by:
        +
        writerId in class WriteBatch
        +
        +
      • +
      + + + + + + + + + + + +
        +
      • +

        resetWriterState

        +
        public void resetWriterState​(long writerId,
        +                             int batchSequence)
        +
        +
        Overrides:
        +
        resetWriterState in class WriteBatch
        +
        +
      • +
      + + + +
        +
      • +

        estimatedSizeInBytes

        +
        public int estimatedSizeInBytes()
        +
        Description copied from class: WriteBatch
        +
        Get an estimate of the number of bytes written to the underlying buffer. The returned value + is exactly correct if the record set is not compressed or if the batch has been WriteBatch.build().
        +
        +
        Specified by:
        +
        estimatedSizeInBytes in class WriteBatch
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/KvWriteBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/KvWriteBatch.html new file mode 100644 index 0000000000..ce2a15f4ec --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/KvWriteBatch.html @@ -0,0 +1,617 @@ + + + + + +KvWriteBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class KvWriteBatch

+
+
+ +
+
    +
  • +
    +
    @NotThreadSafe
    +@Internal
    +public class KvWriteBatch
    +extends WriteBatch
    +
    A batch of kv records that is or will be sent to server by PutKvRequest. + +

    This class is not thread safe and external synchronization must be used when modifying it.

    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        KvWriteBatch

        +
        public KvWriteBatch​(int bucketId,
        +                    org.apache.fluss.metadata.PhysicalTablePath physicalTablePath,
        +                    int schemaId,
        +                    org.apache.fluss.metadata.KvFormat kvFormat,
        +                    int writeLimit,
        +                    org.apache.fluss.memory.AbstractPagedOutputView outputView,
        +                    @Nullable
        +                    int[] targetColumns,
        +                    long createdMs)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        tryAppend

        +
        public boolean tryAppend​(WriteRecord writeRecord,
        +                         WriteCallback callback)
        +                  throws Exception
        +
        Description copied from class: WriteBatch
        +
        try to append one write record to the record batch.
        +
        +
        Specified by:
        +
        tryAppend in class WriteBatch
        +
        Parameters:
        +
        writeRecord - the record to write
        +
        callback - the callback to send back to writer
        +
        Returns:
        +
        true if append success, false if the batch is full.
        +
        Throws:
        +
        Exception
        +
        +
      • +
      + + + +
        +
      • +

        getTargetColumns

        +
        @Nullable
        +public int[] getTargetColumns()
        +
      • +
      + + + +
        +
      • +

        build

        +
        public org.apache.fluss.record.bytesview.BytesView build()
        +
        Description copied from class: WriteBatch
        +
        Gets the memory segment bytes view of the batch. This includes the latest updated WriteBatch.setWriterState(long, int) in the bytes view.
        +
        +
        Specified by:
        +
        build in class WriteBatch
        +
        +
      • +
      + + + + + + + +
        +
      • +

        isClosed

        +
        public boolean isClosed()
        +
        Description copied from class: WriteBatch
        +
        check if the batch is closed.
        +
        +
        Specified by:
        +
        isClosed in class WriteBatch
        +
        Returns:
        +
        true if closed, false otherwise
        +
        +
      • +
      + + + +
        +
      • +

        estimatedSizeInBytes

        +
        public int estimatedSizeInBytes()
        +
        Description copied from class: WriteBatch
        +
        Get an estimate of the number of bytes written to the underlying buffer. The returned value + is exactly correct if the record set is not compressed or if the batch has been WriteBatch.build().
        +
        +
        Specified by:
        +
        estimatedSizeInBytes in class WriteBatch
        +
        +
      • +
      + + + +
        +
      • +

        pooledMemorySegments

        +
        public List<org.apache.fluss.memory.MemorySegment> pooledMemorySegments()
        +
        Description copied from class: WriteBatch
        +
        get pooled memory segments to de-allocate. After produceLog/PutKv acks, the WriteBatch need to de-allocate the allocated pooled MemorySegments back to MemorySegmentPool for reusing.
        +
        +
        Specified by:
        +
        pooledMemorySegments in class WriteBatch
        +
        Returns:
        +
        the pooled memory segment this batch allocated
        +
        +
      • +
      + + + +
        +
      • +

        setWriterState

        +
        public void setWriterState​(long writerId,
        +                           int batchSequence)
        +
        +
        Specified by:
        +
        setWriterState in class WriteBatch
        +
        +
      • +
      + + + +
        +
      • +

        writerId

        +
        public long writerId()
        +
        +
        Specified by:
        +
        writerId in class WriteBatch
        +
        +
      • +
      + + + + + + + + + + + +
        +
      • +

        resetWriterState

        +
        public void resetWriterState​(long writerId,
        +                             int batchSequence)
        +
        +
        Overrides:
        +
        resetWriterState in class WriteBatch
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/ReadyWriteBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/ReadyWriteBatch.html new file mode 100644 index 0000000000..a2ea9f7a02 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/ReadyWriteBatch.html @@ -0,0 +1,325 @@ + + + + + +ReadyWriteBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class ReadyWriteBatch

+
+
+ +
+
    +
  • +
    +
    public class ReadyWriteBatch
    +extends Object
    +
    The WriteBatch already ready in sender. The difference with WriteBatch is that + the partitionId of the tableBucket in ReadyWriteBatch has already been determined in the dynamic + partition create scenario
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        ReadyWriteBatch

        +
        public ReadyWriteBatch​(org.apache.fluss.metadata.TableBucket tableBucket,
        +                       WriteBatch writeBatch)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        tableBucket

        +
        public org.apache.fluss.metadata.TableBucket tableBucket()
        +
      • +
      + + + +
        +
      • +

        writeBatch

        +
        public WriteBatch writeBatch()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.ReadyCheckResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.ReadyCheckResult.html new file mode 100644 index 0000000000..6d51bc30f6 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.ReadyCheckResult.html @@ -0,0 +1,347 @@ + + + + + +RecordAccumulator.ReadyCheckResult (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class RecordAccumulator.ReadyCheckResult

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.write.RecordAccumulator.ReadyCheckResult
    • +
    +
  • +
+
+
    +
  • +
    +
    Enclosing class:
    +
    RecordAccumulator
    +
    +
    +
    public static final class RecordAccumulator.ReadyCheckResult
    +extends Object
    +
    The set of nodes that have at leader one complete record batch in the accumulator.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + +
        +
      • +

        readyNodes

        +
        public final Set<Integer> readyNodes
        +
      • +
      + + + +
        +
      • +

        nextReadyCheckDelayMs

        +
        public final long nextReadyCheckDelayMs
        +
      • +
      + + + +
        +
      • +

        unknownLeaderTables

        +
        public final Set<org.apache.fluss.metadata.PhysicalTablePath> unknownLeaderTables
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        ReadyCheckResult

        +
        public ReadyCheckResult​(Set<Integer> readyNodes,
        +                        long nextReadyCheckDelayMs,
        +                        Set<org.apache.fluss.metadata.PhysicalTablePath> unknownLeaderTables)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.RecordAppendResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.RecordAppendResult.html new file mode 100644 index 0000000000..d8351ff7e1 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.RecordAppendResult.html @@ -0,0 +1,350 @@ + + + + + +RecordAccumulator.RecordAppendResult (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class RecordAccumulator.RecordAppendResult

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.write.RecordAccumulator.RecordAppendResult
    • +
    +
  • +
+
+
    +
  • +
    +
    Enclosing class:
    +
    RecordAccumulator
    +
    +
    +
    public static final class RecordAccumulator.RecordAppendResult
    +extends Object
    +
    Metadata about a record just appended to the record accumulator.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + +
        +
      • +

        batchIsFull

        +
        public final boolean batchIsFull
        +
      • +
      + + + +
        +
      • +

        newBatchCreated

        +
        public final boolean newBatchCreated
        +
      • +
      + + + +
        +
      • +

        abortRecordForNewBatch

        +
        public final boolean abortRecordForNewBatch
        +
        Whether this record was abort because the new batch created in record accumulator.
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        RecordAppendResult

        +
        public RecordAppendResult​(boolean batchIsFull,
        +                          boolean newBatchCreated,
        +                          boolean abortRecordForNewBatch)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.html new file mode 100644 index 0000000000..2c51fab987 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.html @@ -0,0 +1,541 @@ + + + + + +RecordAccumulator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class RecordAccumulator

+
+
+ +
+
    +
  • +
    +
    @Internal
    +public final class RecordAccumulator
    +extends Object
    +
    This class act as a queue that accumulates records into WriteBatch instances to be sent + to tablet servers.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        append

        +
        public RecordAccumulator.RecordAppendResult append​(WriteRecord writeRecord,
        +                                                   WriteCallback callback,
        +                                                   org.apache.fluss.cluster.Cluster cluster,
        +                                                   int bucketId,
        +                                                   boolean abortIfBatchFull)
        +                                            throws Exception
        +
        Add a record to the accumulator, return to append result. + +

        The append result will contain the future metadata, and flag for whether the appended + batch is full or a new batch is created.

        +
        +
        Throws:
        +
        Exception
        +
        +
      • +
      + + + +
        +
      • +

        ready

        +
        public RecordAccumulator.ReadyCheckResult ready​(org.apache.fluss.cluster.Cluster cluster)
        +
        Get a list of nodes whose buckets are ready to be sent. + +

        Also return the flag for whether there are any unknown leaders for the accumulated bucket + batches. + +

        A destination node is ready to send data if: + +

        +     1.There is at least one bucket that is not backing off its send.
        +     2.The record set is full
        +     3.The record set has sat in the accumulator for at least lingerMs milliseconds
        +     4.The accumulator is out of memory and threads are blocking waiting for data (in
        +     this case all buckets are immediately considered ready).
        +     5.The accumulator has been closed
        + 
        +
      • +
      + + + +
        +
      • +

        drain

        +
        public Map<Integer,​List<ReadyWriteBatch>> drain​(org.apache.fluss.cluster.Cluster cluster,
        +                                                      Set<Integer> nodes,
        +                                                      int maxSize)
        +                                               throws Exception
        +
        Drain all the data for the given nodes and collate them into a list of batches that will fit + within the specified size on a per-node basis. This method attempts to avoid choosing the + same table-node over and over.
        +
        +
        Parameters:
        +
        cluster - The current cluster metadata
        +
        nodes - The list of node to drain
        +
        maxSize - The maximum number of bytes to drain
        +
        Returns:
        +
        A list of ReadyWriteBatch for each node specified with total size less than + the requested maxSize.
        +
        Throws:
        +
        Exception
        +
        +
      • +
      + + + +
        +
      • +

        reEnqueue

        +
        public void reEnqueue​(ReadyWriteBatch readyWriteBatch)
        +
      • +
      + + + +
        +
      • +

        abortBatches

        +
        public void abortBatches​(Exception reason)
        +
        Abort all incomplete batches (whether they have been sent or not).
        +
      • +
      + + + +
        +
      • +

        hasUnDrained

        +
        public boolean hasUnDrained()
        +
        Check whether there are any batches which haven't been drained.
        +
      • +
      + + + +
        +
      • +

        hasIncomplete

        +
        public boolean hasIncomplete()
        +
        Check whether there are any pending batches (whether sent or unsent).
        +
      • +
      + + + +
        +
      • +

        beginFlush

        +
        public void beginFlush()
        +
        Initiate the flushing of data from the accumulator...this makes all requests immediately + ready.
        +
      • +
      + + + +
        +
      • +

        awaitFlushCompletion

        +
        public void awaitFlushCompletion()
        +                          throws InterruptedException
        +
        Mark all buckets as ready to send and block until to send is complete.
        +
        +
        Throws:
        +
        InterruptedException
        +
        +
      • +
      + + + +
        +
      • +

        deallocate

        +
        public void deallocate​(WriteBatch batch)
        +
        Deallocate the record batch.
        +
      • +
      + + + +
        +
      • +

        getPhysicalTablePathsInBatches

        +
        public Set<org.apache.fluss.metadata.PhysicalTablePath> getPhysicalTablePathsInBatches()
        +
      • +
      + + + +
        +
      • +

        close

        +
        public void close()
        +
        Close this accumulator and force all the record buffers to be drained.
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RoundRobinBucketAssigner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RoundRobinBucketAssigner.html new file mode 100644 index 0000000000..19cb757a81 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RoundRobinBucketAssigner.html @@ -0,0 +1,355 @@ + + + + + +RoundRobinBucketAssigner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class RoundRobinBucketAssigner

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.write.RoundRobinBucketAssigner
    • +
    +
  • +
+
+
    +
  • +
    +
    @Internal
    +public class RoundRobinBucketAssigner
    +extends Object
    +
    The bucket assigner use round-robin strategy.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        RoundRobinBucketAssigner

        +
        public RoundRobinBucketAssigner​(org.apache.fluss.metadata.PhysicalTablePath physicalTablePath)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        assignBucket

        +
        public int assignBucket​(org.apache.fluss.cluster.Cluster cluster)
        +
      • +
      + + + +
        +
      • +

        abortIfBatchFull

        +
        public boolean abortIfBatchFull()
        +
      • +
      + + + +
        +
      • +

        onNewBatch

        +
        public void onNewBatch​(org.apache.fluss.cluster.Cluster cluster,
        +                       int prevBucketId)
        +
      • +
      + + + +
        +
      • +

        assignBucket

        +
        public int assignBucket​(@Nullable
        +                        byte[] bucketKey,
        +                        org.apache.fluss.cluster.Cluster cluster)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/Sender.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/Sender.html new file mode 100644 index 0000000000..fb5e8ffc4e --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/Sender.html @@ -0,0 +1,402 @@ + + + + + +Sender (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class Sender

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    Runnable
    +
    +
    +
    public class Sender
    +extends Object
    +implements Runnable
    +
    This background thread handles the sending of produce requests to the tablet server. This thread + makes metadata requests to renew its view of the cluster and then sends produce requests to the + appropriate nodes.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    + +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        run

        +
        public void run()
        +
        +
        Specified by:
        +
        run in interface Runnable
        +
        +
      • +
      + + + +
        +
      • +

        runOnce

        +
        public void runOnce()
        +             throws Exception
        +
        Run a single iteration of sending.
        +
        +
        Throws:
        +
        Exception
        +
        +
      • +
      + + + +
        +
      • +

        isRunning

        +
        public boolean isRunning()
        +
      • +
      + + + +
        +
      • +

        forceClose

        +
        public void forceClose()
        +
        Closes the sender without sending out any pending messages.
        +
      • +
      + + + +
        +
      • +

        initiateClose

        +
        public void initiateClose()
        +
        Start closing the sender (won't actually complete until all data is sent out).
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/StickyBucketAssigner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/StickyBucketAssigner.html new file mode 100644 index 0000000000..fa79f19320 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/StickyBucketAssigner.html @@ -0,0 +1,356 @@ + + + + + +StickyBucketAssigner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class StickyBucketAssigner

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.write.StickyBucketAssigner
    • +
    +
  • +
+
+
    +
  • +
    +
    @Internal
    +public class StickyBucketAssigner
    +extends Object
    +
    The bucket assigner with sticky strategy. The assigned bucket id maybe changed only if one new + batch created in record accumulator. Otherwise, we will always return the same bucket id.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        StickyBucketAssigner

        +
        public StickyBucketAssigner​(org.apache.fluss.metadata.PhysicalTablePath physicalTablePath)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        assignBucket

        +
        public int assignBucket​(org.apache.fluss.cluster.Cluster cluster)
        +
      • +
      + + + +
        +
      • +

        abortIfBatchFull

        +
        public boolean abortIfBatchFull()
        +
      • +
      + + + +
        +
      • +

        onNewBatch

        +
        public void onNewBatch​(org.apache.fluss.cluster.Cluster cluster,
        +                       int prevBucketId)
        +
      • +
      + + + +
        +
      • +

        assignBucket

        +
        public int assignBucket​(@Nullable
        +                        byte[] bucketKey,
        +                        org.apache.fluss.cluster.Cluster cluster)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteBatch.RequestFuture.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteBatch.RequestFuture.html new file mode 100644 index 0000000000..bfb3e02837 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteBatch.RequestFuture.html @@ -0,0 +1,336 @@ + + + + + +WriteBatch.RequestFuture (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class WriteBatch.RequestFuture

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.client.write.WriteBatch.RequestFuture
    • +
    +
  • +
+
+
    +
  • +
    +
    Enclosing class:
    +
    WriteBatch
    +
    +
    +
    public static class WriteBatch.RequestFuture
    +extends Object
    +
    The future for this batch.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        RequestFuture

        +
        public RequestFuture()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        done

        +
        public void done()
        +
        Mark this request as complete and unblock any threads waiting on its completion.
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteBatch.html new file mode 100644 index 0000000000..74610acf96 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteBatch.html @@ -0,0 +1,763 @@ + + + + + +WriteBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class WriteBatch

+
+
+ +
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + + + + + +
        +
      • +

        reopened

        +
        protected boolean reopened
        +
      • +
      + + + +
        +
      • +

        recordCount

        +
        protected int recordCount
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        WriteBatch

        +
        public WriteBatch​(int bucketId,
        +                  org.apache.fluss.metadata.PhysicalTablePath physicalTablePath,
        +                  long createdMs)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        tryAppend

        +
        public abstract boolean tryAppend​(WriteRecord writeRecord,
        +                                  WriteCallback callback)
        +                           throws Exception
        +
        try to append one write record to the record batch.
        +
        +
        Parameters:
        +
        writeRecord - the record to write
        +
        callback - the callback to send back to writer
        +
        Returns:
        +
        true if append success, false if the batch is full.
        +
        Throws:
        +
        Exception
        +
        +
      • +
      + + + +
        +
      • +

        build

        +
        public abstract org.apache.fluss.record.bytesview.BytesView build()
        +
        Gets the memory segment bytes view of the batch. This includes the latest updated setWriterState(long, int) in the bytes view.
        +
      • +
      + + + +
        +
      • +

        close

        +
        public abstract void close()
        +                    throws Exception
        +
        close the batch to not append new records.
        +
        +
        Throws:
        +
        Exception
        +
        +
      • +
      + + + +
        +
      • +

        isClosed

        +
        public abstract boolean isClosed()
        +
        check if the batch is closed.
        +
        +
        Returns:
        +
        true if closed, false otherwise
        +
        +
      • +
      + + + +
        +
      • +

        estimatedSizeInBytes

        +
        public abstract int estimatedSizeInBytes()
        +
        Get an estimate of the number of bytes written to the underlying buffer. The returned value + is exactly correct if the record set is not compressed or if the batch has been build().
        +
      • +
      + + + +
        +
      • +

        pooledMemorySegments

        +
        public abstract List<org.apache.fluss.memory.MemorySegment> pooledMemorySegments()
        +
        get pooled memory segments to de-allocate. After produceLog/PutKv acks, the WriteBatch need to de-allocate the allocated pooled MemorySegments back to MemorySegmentPool for reusing.
        +
        +
        Returns:
        +
        the pooled memory segment this batch allocated
        +
        +
      • +
      + + + +
        +
      • +

        setWriterState

        +
        public abstract void setWriterState​(long writerId,
        +                                    int batchSequence)
        +
      • +
      + + + +
        +
      • +

        writerId

        +
        public abstract long writerId()
        +
      • +
      + + + +
        +
      • +

        batchSequence

        +
        public abstract int batchSequence()
        +
      • +
      + + + +
        +
      • +

        abortRecordAppends

        +
        public abstract void abortRecordAppends()
        +
      • +
      + + + +
        +
      • +

        hasBatchSequence

        +
        public boolean hasBatchSequence()
        +
      • +
      + + + +
        +
      • +

        resetWriterState

        +
        public void resetWriterState​(long writerId,
        +                             int batchSequence)
        +
      • +
      + + + +
        +
      • +

        abort

        +
        public void abort​(Exception exception)
        +
        Abort the batch and complete the future and callbacks.
        +
      • +
      + + + +
        +
      • +

        sequenceHasBeenReset

        +
        public boolean sequenceHasBeenReset()
        +
      • +
      + + + +
        +
      • +

        bucketId

        +
        public int bucketId()
        +
      • +
      + + + +
        +
      • +

        physicalTablePath

        +
        public org.apache.fluss.metadata.PhysicalTablePath physicalTablePath()
        +
      • +
      + + + + + + + +
        +
      • +

        waitedTimeMs

        +
        public long waitedTimeMs​(long nowMs)
        +
      • +
      + + + +
        +
      • +

        getRecordCount

        +
        public int getRecordCount()
        +
      • +
      + + + +
        +
      • +

        getQueueTimeMs

        +
        public long getQueueTimeMs()
        +
      • +
      + + + +
        +
      • +

        complete

        +
        public boolean complete()
        +
        Complete the batch successfully.
        +
      • +
      + + + +
        +
      • +

        completeExceptionally

        +
        public boolean completeExceptionally​(Exception exception)
        +
        Complete the batch exceptionally. The provided exception will be used for each record future + contained in the batch.
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteCallback.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteCallback.html new file mode 100644 index 0000000000..8ec2431bac --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteCallback.html @@ -0,0 +1,255 @@ + + + + + +WriteCallback (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface WriteCallback

+
+
+
+
    +
  • +
    +
    @Internal
    +public interface WriteCallback
    +
    A callback interface that the user can implement to allow code to execute when the write request + is complete. This callback will generally execute in the background I/O thread, so it should be + fast.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        onCompletion

        +
        void onCompletion​(@Nullable
        +                  Exception exception)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteFormat.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteFormat.html new file mode 100644 index 0000000000..8438b335c9 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteFormat.html @@ -0,0 +1,393 @@ + + + + + +WriteFormat (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Enum WriteFormat

+
+
+ +
+ +
+
+ +
+
+
    +
  • + +
    + +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        values

        +
        public static WriteFormat[] values()
        +
        Returns an array containing the constants of this enum type, in +the order they are declared. This method may be used to iterate +over the constants as follows: +
        +for (WriteFormat c : WriteFormat.values())
        +    System.out.println(c);
        +
        +
        +
        Returns:
        +
        an array containing the constants of this enum type, in the order they are declared
        +
        +
      • +
      + + + +
        +
      • +

        valueOf

        +
        public static WriteFormat valueOf​(String name)
        +
        Returns the enum constant of this type with the specified name. +The string must match exactly an identifier used to declare an +enum constant in this type. (Extraneous whitespace characters are +not permitted.)
        +
        +
        Parameters:
        +
        name - the name of the enum constant to be returned.
        +
        Returns:
        +
        the enum constant with the specified name
        +
        Throws:
        +
        IllegalArgumentException - if this enum type has no constant with the specified name
        +
        NullPointerException - if the argument is null
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteRecord.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteRecord.html new file mode 100644 index 0000000000..a897784627 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteRecord.html @@ -0,0 +1,461 @@ + + + + + +WriteRecord (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class WriteRecord

+
+
+ +
+
    +
  • +
    +
    @Internal
    +public final class WriteRecord
    +extends Object
    +
    A record to write to a table. It can represent an upsert operation, a delete operation, or an + append operation.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        forUpsert

        +
        public static WriteRecord forUpsert​(org.apache.fluss.metadata.PhysicalTablePath tablePath,
        +                                    org.apache.fluss.row.BinaryRow row,
        +                                    byte[] key,
        +                                    byte[] bucketKey,
        +                                    @Nullable
        +                                    int[] targetColumns)
        +
        Create a write record for upsert operation and partial-upsert operation.
        +
      • +
      + + + +
        +
      • +

        forDelete

        +
        public static WriteRecord forDelete​(org.apache.fluss.metadata.PhysicalTablePath tablePath,
        +                                    byte[] key,
        +                                    byte[] bucketKey,
        +                                    @Nullable
        +                                    int[] targetColumns)
        +
        Create a write record for delete operation and partial-delete update.
        +
      • +
      + + + +
        +
      • +

        forIndexedAppend

        +
        public static WriteRecord forIndexedAppend​(org.apache.fluss.metadata.PhysicalTablePath tablePath,
        +                                           org.apache.fluss.row.indexed.IndexedRow row,
        +                                           @Nullable
        +                                           byte[] bucketKey)
        +
        Create a write record for append operation for indexed format.
        +
      • +
      + + + +
        +
      • +

        forArrowAppend

        +
        public static WriteRecord forArrowAppend​(org.apache.fluss.metadata.PhysicalTablePath tablePath,
        +                                         org.apache.fluss.row.InternalRow row,
        +                                         @Nullable
        +                                         byte[] bucketKey)
        +
        Creates a write record for append operation for Arrow format.
        +
      • +
      + + + +
        +
      • +

        getPhysicalTablePath

        +
        public org.apache.fluss.metadata.PhysicalTablePath getPhysicalTablePath()
        +
      • +
      + + + +
        +
      • +

        getKey

        +
        @Nullable
        +public byte[] getKey()
        +
      • +
      + + + +
        +
      • +

        getBucketKey

        +
        @Nullable
        +public byte[] getBucketKey()
        +
      • +
      + + + +
        +
      • +

        getRow

        +
        @Nullable
        +public org.apache.fluss.row.InternalRow getRow()
        +
      • +
      + + + +
        +
      • +

        getTargetColumns

        +
        @Nullable
        +public int[] getTargetColumns()
        +
      • +
      + + + +
        +
      • +

        getWriteFormat

        +
        public WriteFormat getWriteFormat()
        +
      • +
      + + + +
        +
      • +

        getEstimatedSizeInBytes

        +
        public int getEstimatedSizeInBytes()
        +
        Get the estimated size in bytes of the record with batch header.
        +
        +
        Returns:
        +
        the estimated size in bytes of the record with batch header
        +
        Throws:
        +
        IllegalStateException - if the estimated size in bytes is not supported for the write + format
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriterClient.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriterClient.html new file mode 100644 index 0000000000..faee5f1fa1 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriterClient.html @@ -0,0 +1,419 @@ + + + + + +WriterClient (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class WriterClient

+
+
+ +
+
    +
  • +
    +
    @ThreadSafe
    +@Internal
    +public class WriterClient
    +extends Object
    +
    A client that write records to server. + +

    The writer consists of a pool of buffer space that holds records that haven't yet been + transmitted to the tablet server as well as a background I/O thread that is responsible for + turning these records into requests and transmitting them to the cluster. Failure to close the + WriterClient after use will leak these resources. + +

    The send method is asynchronous. When called, it adds the log record to a buffer of pending + record sends and immediately returns. This allows the wrote record to batch together individual + records for efficiency.

    +
  • +
+
+
+
    +
  • + +
    + +
    + +
    +
      +
    • + + +

      Constructor Summary

      + + + + + + + + + + +
      Constructors 
      ConstructorDescription
      WriterClient​(org.apache.fluss.config.Configuration conf, + MetadataUpdater metadataUpdater, + org.apache.fluss.rpc.metrics.ClientMetricGroup clientMetricGroup, + Admin admin) 
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
    +
  • + +
    + +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        WriterClient

        +
        public WriterClient​(org.apache.fluss.config.Configuration conf,
        +                    MetadataUpdater metadataUpdater,
        +                    org.apache.fluss.rpc.metrics.ClientMetricGroup clientMetricGroup,
        +                    Admin admin)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        send

        +
        public void send​(WriteRecord record,
        +                 WriteCallback callback)
        +
        Asynchronously send a record to a table and invoke the provided callback when to send has + been acknowledged.
        +
      • +
      + + + +
        +
      • +

        flush

        +
        public void flush()
        +
        Invoking this method makes all buffered records immediately available to send (even if + linger.ms is greater than 0) and blocks on the completion of the requests associated + with these records. The post-condition of flush() is that any previously sent + record will have completed (e.g. Future.isDone() == true). A request is + considered completed when it is successfully acknowledged according to the acks + configuration you have specified or else it results in an error. + +

        Other threads can continue sending records while one thread is blocked waiting for a flush + call to complete, however no guarantee is made about the completion of records sent after the + flush call begins.

        +
      • +
      + + + +
        +
      • +

        close

        +
        public void close​(Duration timeout)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/ArrowLogWriteBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/ArrowLogWriteBatch.html new file mode 100644 index 0000000000..6402d55796 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/ArrowLogWriteBatch.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.write.ArrowLogWriteBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.write.ArrowLogWriteBatch

+
+
No usage of org.apache.fluss.client.write.ArrowLogWriteBatch
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/DynamicPartitionCreator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/DynamicPartitionCreator.html new file mode 100644 index 0000000000..45e3d22c05 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/DynamicPartitionCreator.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.write.DynamicPartitionCreator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.write.DynamicPartitionCreator

+
+
No usage of org.apache.fluss.client.write.DynamicPartitionCreator
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/DynamicWriteBatchSizeEstimator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/DynamicWriteBatchSizeEstimator.html new file mode 100644 index 0000000000..170647c8b5 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/DynamicWriteBatchSizeEstimator.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.write.DynamicWriteBatchSizeEstimator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.write.DynamicWriteBatchSizeEstimator

+
+
No usage of org.apache.fluss.client.write.DynamicWriteBatchSizeEstimator
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/HashBucketAssigner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/HashBucketAssigner.html new file mode 100644 index 0000000000..c54e8bc87a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/HashBucketAssigner.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.write.HashBucketAssigner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.write.HashBucketAssigner

+
+
No usage of org.apache.fluss.client.write.HashBucketAssigner
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceBucketEntry.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceBucketEntry.html new file mode 100644 index 0000000000..4f01b83fda --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceBucketEntry.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.write.IdempotenceBucketEntry (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.write.IdempotenceBucketEntry

+
+
No usage of org.apache.fluss.client.write.IdempotenceBucketEntry
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceBucketMap.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceBucketMap.html new file mode 100644 index 0000000000..1638841664 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceBucketMap.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.write.IdempotenceBucketMap (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.write.IdempotenceBucketMap

+
+
No usage of org.apache.fluss.client.write.IdempotenceBucketMap
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceManager.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceManager.html new file mode 100644 index 0000000000..c54387dcc6 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceManager.html @@ -0,0 +1,197 @@ + + + + + +Uses of Class org.apache.fluss.client.write.IdempotenceManager (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.write.IdempotenceManager

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IndexedLogWriteBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IndexedLogWriteBatch.html new file mode 100644 index 0000000000..d7d1d13ae5 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IndexedLogWriteBatch.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.write.IndexedLogWriteBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.write.IndexedLogWriteBatch

+
+
No usage of org.apache.fluss.client.write.IndexedLogWriteBatch
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/KvWriteBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/KvWriteBatch.html new file mode 100644 index 0000000000..1ff1e9eeec --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/KvWriteBatch.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.write.KvWriteBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.write.KvWriteBatch

+
+
No usage of org.apache.fluss.client.write.KvWriteBatch
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/ReadyWriteBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/ReadyWriteBatch.html new file mode 100644 index 0000000000..b6bb435e15 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/ReadyWriteBatch.html @@ -0,0 +1,249 @@ + + + + + +Uses of Class org.apache.fluss.client.write.ReadyWriteBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.write.ReadyWriteBatch

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.ReadyCheckResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.ReadyCheckResult.html new file mode 100644 index 0000000000..46d4a8fb32 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.ReadyCheckResult.html @@ -0,0 +1,194 @@ + + + + + +Uses of Class org.apache.fluss.client.write.RecordAccumulator.ReadyCheckResult (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.write.RecordAccumulator.ReadyCheckResult

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.RecordAppendResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.RecordAppendResult.html new file mode 100644 index 0000000000..d65d4073a0 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.RecordAppendResult.html @@ -0,0 +1,198 @@ + + + + + +Uses of Class org.apache.fluss.client.write.RecordAccumulator.RecordAppendResult (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.write.RecordAccumulator.RecordAppendResult

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.html new file mode 100644 index 0000000000..37e8d34a73 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.html @@ -0,0 +1,197 @@ + + + + + +Uses of Class org.apache.fluss.client.write.RecordAccumulator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.write.RecordAccumulator

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RoundRobinBucketAssigner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RoundRobinBucketAssigner.html new file mode 100644 index 0000000000..d0e8c090f4 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RoundRobinBucketAssigner.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.write.RoundRobinBucketAssigner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.write.RoundRobinBucketAssigner

+
+
No usage of org.apache.fluss.client.write.RoundRobinBucketAssigner
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/Sender.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/Sender.html new file mode 100644 index 0000000000..f04b48f1dd --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/Sender.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.write.Sender (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.write.Sender

+
+
No usage of org.apache.fluss.client.write.Sender
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/StickyBucketAssigner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/StickyBucketAssigner.html new file mode 100644 index 0000000000..27f4aa71fa --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/StickyBucketAssigner.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.client.write.StickyBucketAssigner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.write.StickyBucketAssigner

+
+
No usage of org.apache.fluss.client.write.StickyBucketAssigner
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteBatch.RequestFuture.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteBatch.RequestFuture.html new file mode 100644 index 0000000000..ccb00ed1d7 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteBatch.RequestFuture.html @@ -0,0 +1,192 @@ + + + + + +Uses of Class org.apache.fluss.client.write.WriteBatch.RequestFuture (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.write.WriteBatch.RequestFuture

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteBatch.html new file mode 100644 index 0000000000..804445429f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteBatch.html @@ -0,0 +1,254 @@ + + + + + +Uses of Class org.apache.fluss.client.write.WriteBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.write.WriteBatch

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteCallback.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteCallback.html new file mode 100644 index 0000000000..d2f4d2195c --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteCallback.html @@ -0,0 +1,248 @@ + + + + + +Uses of Interface org.apache.fluss.client.write.WriteCallback (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.client.write.WriteCallback

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteFormat.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteFormat.html new file mode 100644 index 0000000000..bb433ec798 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteFormat.html @@ -0,0 +1,207 @@ + + + + + +Uses of Class org.apache.fluss.client.write.WriteFormat (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.write.WriteFormat

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteRecord.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteRecord.html new file mode 100644 index 0000000000..ede1e65e35 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteRecord.html @@ -0,0 +1,308 @@ + + + + + +Uses of Class org.apache.fluss.client.write.WriteRecord (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.write.WriteRecord

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriterClient.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriterClient.html new file mode 100644 index 0000000000..cbf1d1a245 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriterClient.html @@ -0,0 +1,245 @@ + + + + + +Uses of Class org.apache.fluss.client.write.WriterClient (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.client.write.WriterClient

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-summary.html new file mode 100644 index 0000000000..5ed026a837 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-summary.html @@ -0,0 +1,319 @@ + + + + + +org.apache.fluss.client.write (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.client.write

+
+
+
    +
  • + + + + + + + + + + + + +
    Interface Summary 
    InterfaceDescription
    WriteCallback +
    A callback interface that the user can implement to allow code to execute when the write request + is complete.
    +
    +
  • +
  • + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Class Summary 
    ClassDescription
    ArrowLogWriteBatch +
    A batch of log records managed in ARROW format that is or will be sent to server by ProduceLogRequest.
    +
    DynamicPartitionCreator +
    A creator to create partition when dynamic partition create enable for table.
    +
    DynamicWriteBatchSizeEstimator +
    An estimator to estimate the buffer usage of a writeBatch.
    +
    HashBucketAssigner +
    Hash bucket assigner.
    +
    IdempotenceBucketEntry +
    Entry to store the idempotence information of each table-bucket.
    +
    IdempotenceBucketMap +
    Map to manage IdempotenceBucketEntry of all table-bucket.
    +
    IdempotenceManager +
    A class which manages the idempotence in writer keeps the state necessary to ensure idempotent of + writer.
    +
    IndexedLogWriteBatch +
    A batch of log records managed in INDEXED format that is or will be sent to server by ProduceLogRequest.
    +
    KvWriteBatch +
    A batch of kv records that is or will be sent to server by PutKvRequest.
    +
    ReadyWriteBatch +
    The WriteBatch already ready in sender.
    +
    RecordAccumulator +
    This class act as a queue that accumulates records into WriteBatch instances to be sent + to tablet servers.
    +
    RecordAccumulator.ReadyCheckResult +
    The set of nodes that have at leader one complete record batch in the accumulator.
    +
    RecordAccumulator.RecordAppendResult +
    Metadata about a record just appended to the record accumulator.
    +
    RoundRobinBucketAssigner +
    The bucket assigner use round-robin strategy.
    +
    Sender +
    This background thread handles the sending of produce requests to the tablet server.
    +
    StickyBucketAssigner +
    The bucket assigner with sticky strategy.
    +
    WriteBatch +
    The abstract write batch contains write callback object to wait write request feedback.
    +
    WriteBatch.RequestFuture +
    The future for this batch.
    +
    WriterClient +
    A client that write records to server.
    +
    WriteRecord +
    A record to write to a table.
    +
    +
  • +
  • + + + + + + + + + + + + +
    Enum Summary 
    EnumDescription
    WriteFormat +
    The format of the write record.
    +
    +
  • +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-tree.html new file mode 100644 index 0000000000..cbb0ac96a7 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-tree.html @@ -0,0 +1,205 @@ + + + + + +org.apache.fluss.client.write Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.client.write

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+

Interface Hierarchy

+ +
+
+

Enum Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-use.html new file mode 100644 index 0000000000..da9cf93515 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-use.html @@ -0,0 +1,317 @@ + + + + + +Uses of Package org.apache.fluss.client.write (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.client.write

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/FlinkConnectorOptions.ScanStartupMode.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/FlinkConnectorOptions.ScanStartupMode.html new file mode 100644 index 0000000000..59f8fce695 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/FlinkConnectorOptions.ScanStartupMode.html @@ -0,0 +1,446 @@ + + + + + +FlinkConnectorOptions.ScanStartupMode (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Enum FlinkConnectorOptions.ScanStartupMode

+
+
+ +
+ +
+
+ +
+
+
    +
  • + +
    + +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        values

        +
        public static FlinkConnectorOptions.ScanStartupMode[] values()
        +
        Returns an array containing the constants of this enum type, in +the order they are declared. This method may be used to iterate +over the constants as follows: +
        +for (FlinkConnectorOptions.ScanStartupMode c : FlinkConnectorOptions.ScanStartupMode.values())
        +    System.out.println(c);
        +
        +
        +
        Returns:
        +
        an array containing the constants of this enum type, in the order they are declared
        +
        +
      • +
      + + + +
        +
      • +

        valueOf

        +
        public static FlinkConnectorOptions.ScanStartupMode valueOf​(String name)
        +
        Returns the enum constant of this type with the specified name. +The string must match exactly an identifier used to declare an +enum constant in this type. (Extraneous whitespace characters are +not permitted.)
        +
        +
        Parameters:
        +
        name - the name of the enum constant to be returned.
        +
        Returns:
        +
        the enum constant with the specified name
        +
        Throws:
        +
        IllegalArgumentException - if this enum type has no constant with the specified name
        +
        NullPointerException - if the argument is null
        +
        +
      • +
      + + + + + + + +
        +
      • +

        getDescription

        +
        public org.apache.flink.configuration.description.InlineElement getDescription()
        +
        +
        Specified by:
        +
        getDescription in interface org.apache.flink.configuration.DescribedEnum
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/FlinkConnectorOptions.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/FlinkConnectorOptions.html new file mode 100644 index 0000000000..898bc12894 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/FlinkConnectorOptions.html @@ -0,0 +1,476 @@ + + + + + +FlinkConnectorOptions (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkConnectorOptions

+
+
+ +
+
    +
  • +
    +
    public class FlinkConnectorOptions
    +extends Object
    +
    Options for flink connector.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + +
        +
      • +

        BUCKET_NUMBER

        +
        public static final org.apache.flink.configuration.ConfigOption<Integer> BUCKET_NUMBER
        +
      • +
      + + + +
        +
      • +

        BUCKET_KEY

        +
        public static final org.apache.flink.configuration.ConfigOption<String> BUCKET_KEY
        +
      • +
      + + + +
        +
      • +

        BOOTSTRAP_SERVERS

        +
        public static final org.apache.flink.configuration.ConfigOption<String> BOOTSTRAP_SERVERS
        +
      • +
      + + + +
        +
      • +

        LOOKUP_ASYNC

        +
        public static final org.apache.flink.configuration.ConfigOption<Boolean> LOOKUP_ASYNC
        +
      • +
      + + + + + + + +
        +
      • +

        SCAN_STARTUP_TIMESTAMP

        +
        public static final org.apache.flink.configuration.ConfigOption<String> SCAN_STARTUP_TIMESTAMP
        +
      • +
      + + + +
        +
      • +

        SCAN_PARTITION_DISCOVERY_INTERVAL

        +
        public static final org.apache.flink.configuration.ConfigOption<Duration> SCAN_PARTITION_DISCOVERY_INTERVAL
        +
      • +
      + + + +
        +
      • +

        SINK_IGNORE_DELETE

        +
        public static final org.apache.flink.configuration.ConfigOption<Boolean> SINK_IGNORE_DELETE
        +
      • +
      + + + +
        +
      • +

        SINK_BUCKET_SHUFFLE

        +
        public static final org.apache.flink.configuration.ConfigOption<Boolean> SINK_BUCKET_SHUFFLE
        +
      • +
      + + + +
        +
      • +

        TABLE_OPTIONS

        +
        public static final List<org.apache.flink.configuration.ConfigOption<?>> TABLE_OPTIONS
        +
      • +
      + + + +
        +
      • +

        CLIENT_OPTIONS

        +
        public static final List<org.apache.flink.configuration.ConfigOption<?>> CLIENT_OPTIONS
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlinkConnectorOptions

        +
        public FlinkConnectorOptions()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/CatalogTableAdapter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/CatalogTableAdapter.html new file mode 100644 index 0000000000..f906e27d30 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/CatalogTableAdapter.html @@ -0,0 +1,315 @@ + + + + + +CatalogTableAdapter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class CatalogTableAdapter

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.adapter.CatalogTableAdapter
    • +
    +
  • +
+
+
    +
  • +
    +
    public class CatalogTableAdapter
    +extends Object
    +
    A adapter for CatalogTable constructor. TODO: remove this class when no longer support + flink 1.18 and 1.19.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        CatalogTableAdapter

        +
        public CatalogTableAdapter()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        toCatalogTable

        +
        public static org.apache.flink.table.catalog.CatalogTable toCatalogTable​(org.apache.flink.table.api.Schema schema,
        +                                                                         @Nullable
        +                                                                         String comment,
        +                                                                         List<String> partitionKeys,
        +                                                                         Map<String,​String> options)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/SingleThreadFetcherManagerAdapter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/SingleThreadFetcherManagerAdapter.html new file mode 100644 index 0000000000..cfcca073db --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/SingleThreadFetcherManagerAdapter.html @@ -0,0 +1,320 @@ + + + + + +SingleThreadFetcherManagerAdapter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class SingleThreadFetcherManagerAdapter<E,​SplitT extends org.apache.flink.api.connector.source.SourceSplit>

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager<E,​SplitT>
    • +
    • +
        +
      • org.apache.flink.connector.base.source.reader.fetcher.SingleThreadFetcherManager<E,​SplitT>
      • +
      • +
          +
        • org.apache.fluss.flink.adapter.SingleThreadFetcherManagerAdapter<E,​SplitT>
        • +
        +
      • +
      +
    • +
    +
  • +
+
+
    +
  • +
    +
    Direct Known Subclasses:
    +
    FlinkSourceFetcherManager
    +
    +
    +
    public class SingleThreadFetcherManagerAdapter<E,​SplitT extends org.apache.flink.api.connector.source.SourceSplit>
    +extends org.apache.flink.connector.base.source.reader.fetcher.SingleThreadFetcherManager<E,​SplitT>
    +
    Adapter for SingleThreadFetcherManager.TODO: remove it until not supported in flink 1.18.
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Field Summary

      +
        +
      • + + +

        Fields inherited from class org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager

        +fetchers
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Summary

      + + + + + + + + + + +
      Constructors 
      ConstructorDescription
      SingleThreadFetcherManagerAdapter​(org.apache.flink.connector.base.source.reader.synchronization.FutureCompletingBlockingQueue<org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<E>> elementsQueue, + Supplier<org.apache.flink.connector.base.source.reader.splitreader.SplitReader<E,​SplitT>> splitReaderSupplier, + org.apache.flink.configuration.Configuration configuration, + Consumer<Collection<String>> splitFinishedHook) 
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Summary

      +
        +
      • + + +

        Methods inherited from class org.apache.flink.connector.base.source.reader.fetcher.SingleThreadFetcherManager

        +addSplits, getRunningFetcher, removeSplits
      • +
      +
        +
      • + + +

        Methods inherited from class org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager

        +checkErrors, close, createSplitFetcher, getNumAliveFetchers, getQueue, maybeShutdownFinishedFetchers, pauseOrResumeSplits, startFetcher
      • +
      + +
    • +
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        SingleThreadFetcherManagerAdapter

        +
        public SingleThreadFetcherManagerAdapter​(org.apache.flink.connector.base.source.reader.synchronization.FutureCompletingBlockingQueue<org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<E>> elementsQueue,
        +                                         Supplier<org.apache.flink.connector.base.source.reader.splitreader.SplitReader<E,​SplitT>> splitReaderSupplier,
        +                                         org.apache.flink.configuration.Configuration configuration,
        +                                         Consumer<Collection<String>> splitFinishedHook)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/SingleThreadMultiplexSourceReaderBaseAdapter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/SingleThreadMultiplexSourceReaderBaseAdapter.html new file mode 100644 index 0000000000..57b0023240 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/SingleThreadMultiplexSourceReaderBaseAdapter.html @@ -0,0 +1,334 @@ + + + + + +SingleThreadMultiplexSourceReaderBaseAdapter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class SingleThreadMultiplexSourceReaderBaseAdapter<E,​T,​SplitT extends org.apache.flink.api.connector.source.SourceSplit,​SplitStateT>

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.flink.connector.base.source.reader.SourceReaderBase<E,​T,​SplitT,​SplitStateT>
    • +
    • +
        +
      • org.apache.flink.connector.base.source.reader.SingleThreadMultiplexSourceReaderBase<E,​T,​SplitT,​SplitStateT>
      • +
      • +
          +
        • org.apache.fluss.flink.adapter.SingleThreadMultiplexSourceReaderBaseAdapter<E,​T,​SplitT,​SplitStateT>
        • +
        +
      • +
      +
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    AutoCloseable, org.apache.flink.api.common.state.CheckpointListener, org.apache.flink.api.connector.source.SourceReader<T,​SplitT>
    +
    +
    +
    Direct Known Subclasses:
    +
    FlinkSourceReader
    +
    +
    +
    public abstract class SingleThreadMultiplexSourceReaderBaseAdapter<E,​T,​SplitT extends org.apache.flink.api.connector.source.SourceSplit,​SplitStateT>
    +extends org.apache.flink.connector.base.source.reader.SingleThreadMultiplexSourceReaderBase<E,​T,​SplitT,​SplitStateT>
    +
    Adapter for SingleThreadMultiplexSourceReaderBase.TODO: remove it until not supported in + flink 1.18.
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Field Summary

      +
        +
      • + + +

        Fields inherited from class org.apache.flink.connector.base.source.reader.SourceReaderBase

        +config, context, eofRecordEvaluator, options, recordEmitter, splitFetcherManager
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Summary

      + + + + + + + + + + +
      Constructors 
      ConstructorDescription
      SingleThreadMultiplexSourceReaderBaseAdapter​(org.apache.flink.connector.base.source.reader.synchronization.FutureCompletingBlockingQueue<org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<E>> elementsQueue, + org.apache.flink.connector.base.source.reader.fetcher.SingleThreadFetcherManager<E,​SplitT> splitFetcherManager, + org.apache.flink.connector.base.source.reader.RecordEmitter<E,​T,​SplitStateT> recordEmitter, + org.apache.flink.configuration.Configuration config, + org.apache.flink.api.connector.source.SourceReaderContext context) 
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Summary

      +
        +
      • + + +

        Methods inherited from class org.apache.flink.connector.base.source.reader.SourceReaderBase

        +addSplits, close, getNumberOfCurrentlyAssignedSplits, handleSourceEvents, initializedState, isAvailable, notifyNoMoreSplits, onSplitFinished, pauseOrResumeSplits, pollNext, snapshotState, start, toSplitType
      • +
      + +
        +
      • + + +

        Methods inherited from interface org.apache.flink.api.common.state.CheckpointListener

        +notifyCheckpointAborted
      • +
      +
        +
      • + + +

        Methods inherited from interface org.apache.flink.api.connector.source.SourceReader

        +notifyCheckpointComplete
      • +
      +
    • +
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        SingleThreadMultiplexSourceReaderBaseAdapter

        +
        public SingleThreadMultiplexSourceReaderBaseAdapter​(org.apache.flink.connector.base.source.reader.synchronization.FutureCompletingBlockingQueue<org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<E>> elementsQueue,
        +                                                    org.apache.flink.connector.base.source.reader.fetcher.SingleThreadFetcherManager<E,​SplitT> splitFetcherManager,
        +                                                    org.apache.flink.connector.base.source.reader.RecordEmitter<E,​T,​SplitStateT> recordEmitter,
        +                                                    org.apache.flink.configuration.Configuration config,
        +                                                    org.apache.flink.api.connector.source.SourceReaderContext context)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/CatalogTableAdapter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/CatalogTableAdapter.html new file mode 100644 index 0000000000..208b884445 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/CatalogTableAdapter.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.adapter.CatalogTableAdapter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.adapter.CatalogTableAdapter

+
+
No usage of org.apache.fluss.flink.adapter.CatalogTableAdapter
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/SingleThreadFetcherManagerAdapter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/SingleThreadFetcherManagerAdapter.html new file mode 100644 index 0000000000..b3236fc615 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/SingleThreadFetcherManagerAdapter.html @@ -0,0 +1,194 @@ + + + + + +Uses of Class org.apache.fluss.flink.adapter.SingleThreadFetcherManagerAdapter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.adapter.SingleThreadFetcherManagerAdapter

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/SingleThreadMultiplexSourceReaderBaseAdapter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/SingleThreadMultiplexSourceReaderBaseAdapter.html new file mode 100644 index 0000000000..ff25e2207c --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/SingleThreadMultiplexSourceReaderBaseAdapter.html @@ -0,0 +1,194 @@ + + + + + +Uses of Class org.apache.fluss.flink.adapter.SingleThreadMultiplexSourceReaderBaseAdapter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.adapter.SingleThreadMultiplexSourceReaderBaseAdapter

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-summary.html new file mode 100644 index 0000000000..d3e6c47b31 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-summary.html @@ -0,0 +1,181 @@ + + + + + +org.apache.fluss.flink.adapter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.adapter

+
+
+
    +
  • + + + + + + + + + + + + + + + + + + + + +
    Class Summary 
    ClassDescription
    CatalogTableAdapter +
    A adapter for CatalogTable constructor.
    +
    SingleThreadFetcherManagerAdapter<E,​SplitT extends org.apache.flink.api.connector.source.SourceSplit> +
    Adapter for SingleThreadFetcherManager.TODO: remove it until not supported in flink 1.18.
    +
    SingleThreadMultiplexSourceReaderBaseAdapter<E,​T,​SplitT extends org.apache.flink.api.connector.source.SourceSplit,​SplitStateT> +
    Adapter for SingleThreadMultiplexSourceReaderBase.TODO: remove it until not supported in + flink 1.18.
    +
    +
  • +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-tree.html new file mode 100644 index 0000000000..b4633f7b55 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-tree.html @@ -0,0 +1,181 @@ + + + + + +org.apache.fluss.flink.adapter Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.adapter

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+
    +
  • java.lang.Object +
      +
    • org.apache.fluss.flink.adapter.CatalogTableAdapter
    • +
    • org.apache.flink.connector.base.source.reader.SourceReaderBase<E,​T,​SplitT,​SplitStateT> (implements org.apache.flink.api.connector.source.SourceReader<T,​SplitT>) +
        +
      • org.apache.flink.connector.base.source.reader.SingleThreadMultiplexSourceReaderBase<E,​T,​SplitT,​SplitStateT> + +
      • +
      +
    • +
    • org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager<E,​SplitT> +
        +
      • org.apache.flink.connector.base.source.reader.fetcher.SingleThreadFetcherManager<E,​SplitT> + +
      • +
      +
    • +
    +
  • +
+
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-use.html new file mode 100644 index 0000000000..4513815309 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-use.html @@ -0,0 +1,209 @@ + + + + + +Uses of Package org.apache.fluss.flink.adapter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.adapter

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalog.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalog.html new file mode 100644 index 0000000000..2306ae203e --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalog.html @@ -0,0 +1,1463 @@ + + + + + +FlinkCatalog (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkCatalog

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.flink.table.catalog.AbstractCatalog
    • +
    • +
        +
      • org.apache.fluss.flink.catalog.FlinkCatalog
      • +
      +
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.table.catalog.Catalog
    +
    +
    +
    public class FlinkCatalog
    +extends org.apache.flink.table.catalog.AbstractCatalog
    +
    A Flink Catalog for fluss. + +

    Currently, this class must extend the internal Flink class AbstractCatalog because an + incompatibility bug ( FLINK-38030) in flink 2.0.0. + +

    TODO: Once this issue is resolved in a future version of Flink (likely 2.1+), refactor this + class to implement the public interface Catalog instead of + extending the internal class AbstractCatalog.

    +
  • +
+
+
+
    +
  • + +
    + +
    + +
    + +
    + +
    +
      +
    • + + +

      Method Summary

      + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
      All Methods Instance Methods Concrete Methods 
      Modifier and TypeMethodDescription
      voidalterDatabase​(String databaseName, + org.apache.flink.table.catalog.CatalogDatabase catalogDatabase, + boolean b) 
      voidalterFunction​(org.apache.flink.table.catalog.ObjectPath objectPath, + org.apache.flink.table.catalog.CatalogFunction catalogFunction, + boolean b) 
      voidalterPartition​(org.apache.flink.table.catalog.ObjectPath objectPath, + org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec, + org.apache.flink.table.catalog.CatalogPartition catalogPartition, + boolean b) 
      voidalterPartitionColumnStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath, + org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec, + org.apache.flink.table.catalog.stats.CatalogColumnStatistics catalogColumnStatistics, + boolean b) 
      voidalterPartitionStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath, + org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec, + org.apache.flink.table.catalog.stats.CatalogTableStatistics catalogTableStatistics, + boolean b) 
      voidalterTable​(org.apache.flink.table.catalog.ObjectPath objectPath, + org.apache.flink.table.catalog.CatalogBaseTable catalogBaseTable, + boolean b) 
      voidalterTableColumnStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath, + org.apache.flink.table.catalog.stats.CatalogColumnStatistics catalogColumnStatistics, + boolean b) 
      voidalterTableStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath, + org.apache.flink.table.catalog.stats.CatalogTableStatistics catalogTableStatistics, + boolean b) 
      voidclose() 
      voidcreateDatabase​(String databaseName, + org.apache.flink.table.catalog.CatalogDatabase database, + boolean ignoreIfExists) 
      voidcreateFunction​(org.apache.flink.table.catalog.ObjectPath objectPath, + org.apache.flink.table.catalog.CatalogFunction catalogFunction, + boolean b) 
      voidcreatePartition​(org.apache.flink.table.catalog.ObjectPath objectPath, + org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec, + org.apache.flink.table.catalog.CatalogPartition catalogPartition, + boolean b) 
      voidcreateTable​(org.apache.flink.table.catalog.ObjectPath objectPath, + org.apache.flink.table.catalog.CatalogBaseTable table, + boolean ignoreIfExist) 
      booleandatabaseExists​(String databaseName) 
      voiddropDatabase​(String databaseName, + boolean ignoreIfNotExists, + boolean cascade) 
      voiddropFunction​(org.apache.flink.table.catalog.ObjectPath objectPath, + boolean b) 
      voiddropPartition​(org.apache.flink.table.catalog.ObjectPath objectPath, + org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec, + boolean b) 
      voiddropTable​(org.apache.flink.table.catalog.ObjectPath objectPath, + boolean ignoreIfNotExists) 
      booleanfunctionExists​(org.apache.flink.table.catalog.ObjectPath objectPath) 
      org.apache.flink.table.catalog.CatalogDatabasegetDatabase​(String databaseName) 
      Optional<org.apache.flink.table.factories.Factory>getFactory() 
      org.apache.flink.table.catalog.CatalogFunctiongetFunction​(org.apache.flink.table.catalog.ObjectPath functionPath) 
      protected org.apache.flink.table.catalog.CatalogBaseTablegetLakeTable​(String databaseName, + String tableName, + org.apache.fluss.config.Configuration properties) 
      org.apache.flink.table.catalog.CatalogPartitiongetPartition​(org.apache.flink.table.catalog.ObjectPath objectPath, + org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec) 
      org.apache.flink.table.catalog.stats.CatalogColumnStatisticsgetPartitionColumnStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath, + org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec) 
      org.apache.flink.table.catalog.stats.CatalogTableStatisticsgetPartitionStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath, + org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec) 
      org.apache.flink.table.procedures.ProceduregetProcedure​(org.apache.flink.table.catalog.ObjectPath procedurePath) 
      Map<String,​String>getSecurityConfigs() 
      org.apache.flink.table.catalog.CatalogBaseTablegetTable​(org.apache.flink.table.catalog.ObjectPath objectPath) 
      org.apache.flink.table.catalog.stats.CatalogColumnStatisticsgetTableColumnStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath) 
      org.apache.flink.table.catalog.stats.CatalogTableStatisticsgetTableStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath) 
      List<String>listDatabases() 
      List<String>listFunctions​(String s) 
      List<org.apache.flink.table.catalog.CatalogPartitionSpec>listPartitions​(org.apache.flink.table.catalog.ObjectPath objectPath) 
      List<org.apache.flink.table.catalog.CatalogPartitionSpec>listPartitions​(org.apache.flink.table.catalog.ObjectPath objectPath, + org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec) 
      List<org.apache.flink.table.catalog.CatalogPartitionSpec>listPartitionsByFilter​(org.apache.flink.table.catalog.ObjectPath objectPath, + List<org.apache.flink.table.expressions.Expression> list) 
      List<String>listProcedures​(String dbName) 
      List<String>listTables​(String databaseName) 
      List<String>listViews​(String s) 
      voidopen() 
      booleanpartitionExists​(org.apache.flink.table.catalog.ObjectPath objectPath, + org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec) 
      voidrenameTable​(org.apache.flink.table.catalog.ObjectPath objectPath, + String s, + boolean b) 
      booleantableExists​(org.apache.flink.table.catalog.ObjectPath objectPath) 
      protected org.apache.fluss.metadata.TablePathtoTablePath​(org.apache.flink.table.catalog.ObjectPath objectPath) 
      +
        +
      • + + +

        Methods inherited from class org.apache.flink.table.catalog.AbstractCatalog

        +getDefaultDatabase, getName
      • +
      + +
        +
      • + + +

        Methods inherited from interface org.apache.flink.table.catalog.Catalog

        +alterTable, bulkGetPartitionColumnStatistics, bulkGetPartitionStatistics, dropDatabase, getFunctionDefinitionFactory, getTable, getTableFactory, supportsManagedTable
      • +
      +
    • +
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + + + + + +
        +
      • +

        classLoader

        +
        protected final ClassLoader classLoader
        +
      • +
      + + + +
        +
      • +

        catalogName

        +
        protected final String catalogName
        +
      • +
      + + + +
        +
      • +

        defaultDatabase

        +
        protected final String defaultDatabase
        +
      • +
      + + + +
        +
      • +

        bootstrapServers

        +
        protected final String bootstrapServers
        +
      • +
      + + + +
        +
      • +

        connection

        +
        protected Connection connection
        +
      • +
      + + + +
        +
      • +

        admin

        +
        protected Admin admin
        +
      • +
      +
    • +
    +
    + +
    + +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getFactory

        +
        public Optional<org.apache.flink.table.factories.Factory> getFactory()
        +
      • +
      + + + +
        +
      • +

        open

        +
        public void open()
        +          throws org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        close

        +
        public void close()
        +           throws org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        listDatabases

        +
        public List<String> listDatabases()
        +                           throws org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        getDatabase

        +
        public org.apache.flink.table.catalog.CatalogDatabase getDatabase​(String databaseName)
        +                                                           throws org.apache.flink.table.catalog.exceptions.DatabaseNotExistException,
        +                                                                  org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.DatabaseNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        databaseExists

        +
        public boolean databaseExists​(String databaseName)
        +                       throws org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        createDatabase

        +
        public void createDatabase​(String databaseName,
        +                           org.apache.flink.table.catalog.CatalogDatabase database,
        +                           boolean ignoreIfExists)
        +                    throws org.apache.flink.table.catalog.exceptions.DatabaseAlreadyExistException,
        +                           org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.DatabaseAlreadyExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        dropDatabase

        +
        public void dropDatabase​(String databaseName,
        +                         boolean ignoreIfNotExists,
        +                         boolean cascade)
        +                  throws org.apache.flink.table.catalog.exceptions.DatabaseNotExistException,
        +                         org.apache.flink.table.catalog.exceptions.DatabaseNotEmptyException,
        +                         org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.DatabaseNotExistException
        +
        org.apache.flink.table.catalog.exceptions.DatabaseNotEmptyException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        alterDatabase

        +
        public void alterDatabase​(String databaseName,
        +                          org.apache.flink.table.catalog.CatalogDatabase catalogDatabase,
        +                          boolean b)
        +                   throws org.apache.flink.table.catalog.exceptions.DatabaseNotExistException,
        +                          org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.DatabaseNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        listTables

        +
        public List<String> listTables​(String databaseName)
        +                        throws org.apache.flink.table.catalog.exceptions.DatabaseNotExistException,
        +                               org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.DatabaseNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        listViews

        +
        public List<String> listViews​(String s)
        +                       throws org.apache.flink.table.catalog.exceptions.DatabaseNotExistException,
        +                              org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.DatabaseNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        getTable

        +
        public org.apache.flink.table.catalog.CatalogBaseTable getTable​(org.apache.flink.table.catalog.ObjectPath objectPath)
        +                                                         throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
        +                                                                org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.TableNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        getLakeTable

        +
        protected org.apache.flink.table.catalog.CatalogBaseTable getLakeTable​(String databaseName,
        +                                                                       String tableName,
        +                                                                       org.apache.fluss.config.Configuration properties)
        +                                                                throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
        +                                                                       org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.TableNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        tableExists

        +
        public boolean tableExists​(org.apache.flink.table.catalog.ObjectPath objectPath)
        +                    throws org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        dropTable

        +
        public void dropTable​(org.apache.flink.table.catalog.ObjectPath objectPath,
        +                      boolean ignoreIfNotExists)
        +               throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
        +                      org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.TableNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        renameTable

        +
        public void renameTable​(org.apache.flink.table.catalog.ObjectPath objectPath,
        +                        String s,
        +                        boolean b)
        +                 throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
        +                        org.apache.flink.table.catalog.exceptions.TableAlreadyExistException,
        +                        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.TableNotExistException
        +
        org.apache.flink.table.catalog.exceptions.TableAlreadyExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        createTable

        +
        public void createTable​(org.apache.flink.table.catalog.ObjectPath objectPath,
        +                        org.apache.flink.table.catalog.CatalogBaseTable table,
        +                        boolean ignoreIfExist)
        +                 throws org.apache.flink.table.catalog.exceptions.TableAlreadyExistException,
        +                        org.apache.flink.table.catalog.exceptions.DatabaseNotExistException,
        +                        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.TableAlreadyExistException
        +
        org.apache.flink.table.catalog.exceptions.DatabaseNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        alterTable

        +
        public void alterTable​(org.apache.flink.table.catalog.ObjectPath objectPath,
        +                       org.apache.flink.table.catalog.CatalogBaseTable catalogBaseTable,
        +                       boolean b)
        +                throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
        +                       org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.TableNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        listPartitions

        +
        public List<org.apache.flink.table.catalog.CatalogPartitionSpec> listPartitions​(org.apache.flink.table.catalog.ObjectPath objectPath)
        +                                                                         throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
        +                                                                                org.apache.flink.table.catalog.exceptions.TableNotPartitionedException,
        +                                                                                org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.TableNotExistException
        +
        org.apache.flink.table.catalog.exceptions.TableNotPartitionedException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        listPartitions

        +
        public List<org.apache.flink.table.catalog.CatalogPartitionSpec> listPartitions​(org.apache.flink.table.catalog.ObjectPath objectPath,
        +                                                                                org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec)
        +                                                                         throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
        +                                                                                org.apache.flink.table.catalog.exceptions.TableNotPartitionedException,
        +                                                                                org.apache.flink.table.catalog.exceptions.PartitionSpecInvalidException,
        +                                                                                org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.TableNotExistException
        +
        org.apache.flink.table.catalog.exceptions.TableNotPartitionedException
        +
        org.apache.flink.table.catalog.exceptions.PartitionSpecInvalidException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        listPartitionsByFilter

        +
        public List<org.apache.flink.table.catalog.CatalogPartitionSpec> listPartitionsByFilter​(org.apache.flink.table.catalog.ObjectPath objectPath,
        +                                                                                        List<org.apache.flink.table.expressions.Expression> list)
        +                                                                                 throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
        +                                                                                        org.apache.flink.table.catalog.exceptions.TableNotPartitionedException,
        +                                                                                        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.TableNotExistException
        +
        org.apache.flink.table.catalog.exceptions.TableNotPartitionedException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        getPartition

        +
        public org.apache.flink.table.catalog.CatalogPartition getPartition​(org.apache.flink.table.catalog.ObjectPath objectPath,
        +                                                                    org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec)
        +                                                             throws org.apache.flink.table.catalog.exceptions.PartitionNotExistException,
        +                                                                    org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.PartitionNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        partitionExists

        +
        public boolean partitionExists​(org.apache.flink.table.catalog.ObjectPath objectPath,
        +                               org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec)
        +                        throws org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        createPartition

        +
        public void createPartition​(org.apache.flink.table.catalog.ObjectPath objectPath,
        +                            org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec,
        +                            org.apache.flink.table.catalog.CatalogPartition catalogPartition,
        +                            boolean b)
        +                     throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
        +                            org.apache.flink.table.catalog.exceptions.TableNotPartitionedException,
        +                            org.apache.flink.table.catalog.exceptions.PartitionSpecInvalidException,
        +                            org.apache.flink.table.catalog.exceptions.PartitionAlreadyExistsException,
        +                            org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.TableNotExistException
        +
        org.apache.flink.table.catalog.exceptions.TableNotPartitionedException
        +
        org.apache.flink.table.catalog.exceptions.PartitionSpecInvalidException
        +
        org.apache.flink.table.catalog.exceptions.PartitionAlreadyExistsException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        dropPartition

        +
        public void dropPartition​(org.apache.flink.table.catalog.ObjectPath objectPath,
        +                          org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec,
        +                          boolean b)
        +                   throws org.apache.flink.table.catalog.exceptions.PartitionNotExistException,
        +                          org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.PartitionNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        alterPartition

        +
        public void alterPartition​(org.apache.flink.table.catalog.ObjectPath objectPath,
        +                           org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec,
        +                           org.apache.flink.table.catalog.CatalogPartition catalogPartition,
        +                           boolean b)
        +                    throws org.apache.flink.table.catalog.exceptions.PartitionNotExistException,
        +                           org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.PartitionNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        listFunctions

        +
        public List<String> listFunctions​(String s)
        +                           throws org.apache.flink.table.catalog.exceptions.DatabaseNotExistException,
        +                                  org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.DatabaseNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        getFunction

        +
        public org.apache.flink.table.catalog.CatalogFunction getFunction​(org.apache.flink.table.catalog.ObjectPath functionPath)
        +                                                           throws org.apache.flink.table.catalog.exceptions.FunctionNotExistException,
        +                                                                  org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.FunctionNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        functionExists

        +
        public boolean functionExists​(org.apache.flink.table.catalog.ObjectPath objectPath)
        +                       throws org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        createFunction

        +
        public void createFunction​(org.apache.flink.table.catalog.ObjectPath objectPath,
        +                           org.apache.flink.table.catalog.CatalogFunction catalogFunction,
        +                           boolean b)
        +                    throws org.apache.flink.table.catalog.exceptions.FunctionAlreadyExistException,
        +                           org.apache.flink.table.catalog.exceptions.DatabaseNotExistException,
        +                           org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.FunctionAlreadyExistException
        +
        org.apache.flink.table.catalog.exceptions.DatabaseNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        alterFunction

        +
        public void alterFunction​(org.apache.flink.table.catalog.ObjectPath objectPath,
        +                          org.apache.flink.table.catalog.CatalogFunction catalogFunction,
        +                          boolean b)
        +                   throws org.apache.flink.table.catalog.exceptions.FunctionNotExistException,
        +                          org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.FunctionNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        dropFunction

        +
        public void dropFunction​(org.apache.flink.table.catalog.ObjectPath objectPath,
        +                         boolean b)
        +                  throws org.apache.flink.table.catalog.exceptions.FunctionNotExistException,
        +                         org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.FunctionNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        getTableStatistics

        +
        public org.apache.flink.table.catalog.stats.CatalogTableStatistics getTableStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath)
        +                                                                               throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
        +                                                                                      org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.TableNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        getTableColumnStatistics

        +
        public org.apache.flink.table.catalog.stats.CatalogColumnStatistics getTableColumnStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath)
        +                                                                                      throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
        +                                                                                             org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.TableNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        getPartitionStatistics

        +
        public org.apache.flink.table.catalog.stats.CatalogTableStatistics getPartitionStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath,
        +                                                                                          org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec)
        +                                                                                   throws org.apache.flink.table.catalog.exceptions.PartitionNotExistException,
        +                                                                                          org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.PartitionNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        getPartitionColumnStatistics

        +
        public org.apache.flink.table.catalog.stats.CatalogColumnStatistics getPartitionColumnStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath,
        +                                                                                                 org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec)
        +                                                                                          throws org.apache.flink.table.catalog.exceptions.PartitionNotExistException,
        +                                                                                                 org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.PartitionNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        alterTableStatistics

        +
        public void alterTableStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath,
        +                                 org.apache.flink.table.catalog.stats.CatalogTableStatistics catalogTableStatistics,
        +                                 boolean b)
        +                          throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
        +                                 org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.TableNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        alterTableColumnStatistics

        +
        public void alterTableColumnStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath,
        +                                       org.apache.flink.table.catalog.stats.CatalogColumnStatistics catalogColumnStatistics,
        +                                       boolean b)
        +                                throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
        +                                       org.apache.flink.table.catalog.exceptions.CatalogException,
        +                                       org.apache.flink.table.catalog.exceptions.TablePartitionedException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.TableNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        org.apache.flink.table.catalog.exceptions.TablePartitionedException
        +
        +
      • +
      + + + +
        +
      • +

        alterPartitionStatistics

        +
        public void alterPartitionStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath,
        +                                     org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec,
        +                                     org.apache.flink.table.catalog.stats.CatalogTableStatistics catalogTableStatistics,
        +                                     boolean b)
        +                              throws org.apache.flink.table.catalog.exceptions.PartitionNotExistException,
        +                                     org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.PartitionNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        alterPartitionColumnStatistics

        +
        public void alterPartitionColumnStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath,
        +                                           org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec,
        +                                           org.apache.flink.table.catalog.stats.CatalogColumnStatistics catalogColumnStatistics,
        +                                           boolean b)
        +                                    throws org.apache.flink.table.catalog.exceptions.PartitionNotExistException,
        +                                           org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.PartitionNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        toTablePath

        +
        protected org.apache.fluss.metadata.TablePath toTablePath​(org.apache.flink.table.catalog.ObjectPath objectPath)
        +
      • +
      + + + +
        +
      • +

        listProcedures

        +
        public List<String> listProcedures​(String dbName)
        +                            throws org.apache.flink.table.catalog.exceptions.DatabaseNotExistException,
        +                                   org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.DatabaseNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        getProcedure

        +
        public org.apache.flink.table.procedures.Procedure getProcedure​(org.apache.flink.table.catalog.ObjectPath procedurePath)
        +                                                         throws org.apache.flink.table.catalog.exceptions.ProcedureNotExistException,
        +                                                                org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.ProcedureNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      + + + +
        +
      • +

        getSecurityConfigs

        +
        @VisibleForTesting
        +public Map<String,​String> getSecurityConfigs()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalogFactory.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalogFactory.html new file mode 100644 index 0000000000..10fa409d9d --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalogFactory.html @@ -0,0 +1,446 @@ + + + + + +FlinkCatalogFactory (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkCatalogFactory

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.catalog.FlinkCatalogFactory
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.table.factories.CatalogFactory, org.apache.flink.table.factories.Factory, org.apache.flink.table.factories.TableFactory
    +
    +
    +
    public class FlinkCatalogFactory
    +extends Object
    +implements org.apache.flink.table.factories.CatalogFactory
    +
    Factory for FlinkCatalog.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    + +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlinkCatalogFactory

        +
        public FlinkCatalogFactory()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        factoryIdentifier

        +
        public String factoryIdentifier()
        +
        +
        Specified by:
        +
        factoryIdentifier in interface org.apache.flink.table.factories.CatalogFactory
        +
        Specified by:
        +
        factoryIdentifier in interface org.apache.flink.table.factories.Factory
        +
        +
      • +
      + + + +
        +
      • +

        requiredOptions

        +
        public Set<org.apache.flink.configuration.ConfigOption<?>> requiredOptions()
        +
        +
        Specified by:
        +
        requiredOptions in interface org.apache.flink.table.factories.CatalogFactory
        +
        Specified by:
        +
        requiredOptions in interface org.apache.flink.table.factories.Factory
        +
        +
      • +
      + + + +
        +
      • +

        optionalOptions

        +
        public Set<org.apache.flink.configuration.ConfigOption<?>> optionalOptions()
        +
        +
        Specified by:
        +
        optionalOptions in interface org.apache.flink.table.factories.CatalogFactory
        +
        Specified by:
        +
        optionalOptions in interface org.apache.flink.table.factories.Factory
        +
        +
      • +
      + + + +
        +
      • +

        createCatalog

        +
        public FlinkCatalog createCatalog​(org.apache.flink.table.factories.CatalogFactory.Context context)
        +
        +
        Specified by:
        +
        createCatalog in interface org.apache.flink.table.factories.CatalogFactory
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalogOptions.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalogOptions.html new file mode 100644 index 0000000000..43470494bc --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalogOptions.html @@ -0,0 +1,271 @@ + + + + + +FlinkCatalogOptions (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkCatalogOptions

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.catalog.FlinkCatalogOptions
    • +
    +
  • +
+
+
    +
  • +
    +
    public class FlinkCatalogOptions
    +extends Object
    +
    Options for flink catalog.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + +
        +
      • +

        DEFAULT_DATABASE

        +
        public static final org.apache.flink.configuration.ConfigOption<String> DEFAULT_DATABASE
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkTableFactory.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkTableFactory.html new file mode 100644 index 0000000000..3acc063ab3 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkTableFactory.html @@ -0,0 +1,412 @@ + + + + + +FlinkTableFactory (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkTableFactory

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.table.factories.DynamicTableFactory, org.apache.flink.table.factories.DynamicTableSinkFactory, org.apache.flink.table.factories.DynamicTableSourceFactory, org.apache.flink.table.factories.Factory
    +
    +
    +
    public class FlinkTableFactory
    +extends Object
    +implements org.apache.flink.table.factories.DynamicTableSourceFactory, org.apache.flink.table.factories.DynamicTableSinkFactory
    +
    Factory to create table source and table sink for Fluss.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlinkTableFactory

        +
        public FlinkTableFactory()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        createDynamicTableSource

        +
        public org.apache.flink.table.connector.source.DynamicTableSource createDynamicTableSource​(org.apache.flink.table.factories.DynamicTableFactory.Context context)
        +
        +
        Specified by:
        +
        createDynamicTableSource in interface org.apache.flink.table.factories.DynamicTableSourceFactory
        +
        +
      • +
      + + + +
        +
      • +

        createDynamicTableSink

        +
        public org.apache.flink.table.connector.sink.DynamicTableSink createDynamicTableSink​(org.apache.flink.table.factories.DynamicTableFactory.Context context)
        +
        +
        Specified by:
        +
        createDynamicTableSink in interface org.apache.flink.table.factories.DynamicTableSinkFactory
        +
        +
      • +
      + + + +
        +
      • +

        factoryIdentifier

        +
        public String factoryIdentifier()
        +
        +
        Specified by:
        +
        factoryIdentifier in interface org.apache.flink.table.factories.Factory
        +
        +
      • +
      + + + +
        +
      • +

        requiredOptions

        +
        public Set<org.apache.flink.configuration.ConfigOption<?>> requiredOptions()
        +
        +
        Specified by:
        +
        requiredOptions in interface org.apache.flink.table.factories.Factory
        +
        +
      • +
      + + + +
        +
      • +

        optionalOptions

        +
        public Set<org.apache.flink.configuration.ConfigOption<?>> optionalOptions()
        +
        +
        Specified by:
        +
        optionalOptions in interface org.apache.flink.table.factories.Factory
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalog.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalog.html new file mode 100644 index 0000000000..030fa1be56 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalog.html @@ -0,0 +1,192 @@ + + + + + +Uses of Class org.apache.fluss.flink.catalog.FlinkCatalog (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.catalog.FlinkCatalog

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalogFactory.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalogFactory.html new file mode 100644 index 0000000000..2fcddfb73f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalogFactory.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.catalog.FlinkCatalogFactory (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.catalog.FlinkCatalogFactory

+
+
No usage of org.apache.fluss.flink.catalog.FlinkCatalogFactory
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalogOptions.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalogOptions.html new file mode 100644 index 0000000000..5c770b2e96 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalogOptions.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.catalog.FlinkCatalogOptions (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.catalog.FlinkCatalogOptions

+
+
No usage of org.apache.fluss.flink.catalog.FlinkCatalogOptions
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkTableFactory.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkTableFactory.html new file mode 100644 index 0000000000..49ec529ce6 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkTableFactory.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.catalog.FlinkTableFactory (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.catalog.FlinkTableFactory

+
+
No usage of org.apache.fluss.flink.catalog.FlinkTableFactory
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-summary.html new file mode 100644 index 0000000000..ce294770f5 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-summary.html @@ -0,0 +1,186 @@ + + + + + +org.apache.fluss.flink.catalog (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.catalog

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-tree.html new file mode 100644 index 0000000000..06a58001b7 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-tree.html @@ -0,0 +1,170 @@ + + + + + +org.apache.fluss.flink.catalog Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.catalog

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+
    +
  • java.lang.Object +
      +
    • org.apache.flink.table.catalog.AbstractCatalog (implements org.apache.flink.table.catalog.Catalog) + +
    • +
    • org.apache.fluss.flink.catalog.FlinkCatalogFactory (implements org.apache.flink.table.factories.CatalogFactory)
    • +
    • org.apache.fluss.flink.catalog.FlinkCatalogOptions
    • +
    • org.apache.fluss.flink.catalog.FlinkTableFactory (implements org.apache.flink.table.factories.DynamicTableSinkFactory, org.apache.flink.table.factories.DynamicTableSourceFactory)
    • +
    +
  • +
+
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-use.html new file mode 100644 index 0000000000..d252449fe8 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-use.html @@ -0,0 +1,185 @@ + + + + + +Uses of Package org.apache.fluss.flink.catalog (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.catalog

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/class-use/FlinkConnectorOptions.ScanStartupMode.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/class-use/FlinkConnectorOptions.ScanStartupMode.html new file mode 100644 index 0000000000..713da7e1ec --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/class-use/FlinkConnectorOptions.ScanStartupMode.html @@ -0,0 +1,243 @@ + + + + + +Uses of Class org.apache.fluss.flink.FlinkConnectorOptions.ScanStartupMode (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.FlinkConnectorOptions.ScanStartupMode

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/class-use/FlinkConnectorOptions.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/class-use/FlinkConnectorOptions.html new file mode 100644 index 0000000000..e67feffe7a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/class-use/FlinkConnectorOptions.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.FlinkConnectorOptions (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.FlinkConnectorOptions

+
+
No usage of org.apache.fluss.flink.FlinkConnectorOptions
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeCatalog.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeCatalog.html new file mode 100644 index 0000000000..1bb2510369 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeCatalog.html @@ -0,0 +1,318 @@ + + + + + +LakeCatalog (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LakeCatalog

+
+
+ +
+
    +
  • +
    +
    public class LakeCatalog
    +extends Object
    +
    A lake catalog to delegate the operations on lake table.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    + +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getTable

        +
        public org.apache.flink.table.catalog.CatalogBaseTable getTable​(org.apache.flink.table.catalog.ObjectPath objectPath)
        +                                                         throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
        +                                                                org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
        Throws:
        +
        org.apache.flink.table.catalog.exceptions.TableNotExistException
        +
        org.apache.flink.table.catalog.exceptions.CatalogException
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeRecordRecordEmitter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeRecordRecordEmitter.html new file mode 100644 index 0000000000..c1ff70840f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeRecordRecordEmitter.html @@ -0,0 +1,311 @@ + + + + + +LakeRecordRecordEmitter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LakeRecordRecordEmitter<OUT>

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.lake.LakeRecordRecordEmitter<OUT>
    • +
    +
  • +
+
+
    +
  • +
    +
    public class LakeRecordRecordEmitter<OUT>
    +extends Object
    +
    The emitter to emit record from lake split.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LakeRecordRecordEmitter

        +
        public LakeRecordRecordEmitter​(BiConsumer<ScanRecord,​org.apache.flink.api.connector.source.SourceOutput<OUT>> sourceOutputFunc)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        emitRecord

        +
        public void emitRecord​(SourceSplitState splitState,
        +                       org.apache.flink.api.connector.source.SourceOutput<OUT> sourceOutput,
        +                       RecordAndPos recordAndPos)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitGenerator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitGenerator.html new file mode 100644 index 0000000000..ab2438b53d --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitGenerator.html @@ -0,0 +1,330 @@ + + + + + +LakeSplitGenerator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LakeSplitGenerator

+
+
+ +
+
    +
  • +
    +
    public class LakeSplitGenerator
    +extends Object
    +
    A generator for lake splits.
    +
  • +
+
+
+ +
+
+ +
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitReaderGenerator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitReaderGenerator.html new file mode 100644 index 0000000000..217d601269 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitReaderGenerator.html @@ -0,0 +1,329 @@ + + + + + +LakeSplitReaderGenerator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LakeSplitReaderGenerator

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.lake.LakeSplitReaderGenerator
    • +
    +
  • +
+
+
    +
  • +
    +
    public class LakeSplitReaderGenerator
    +extends Object
    +
    A generator to generate reader for lake split.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LakeSplitReaderGenerator

        +
        public LakeSplitReaderGenerator​(Table table,
        +                                @Nullable
        +                                int[] projectedFields,
        +                                @Nullable
        +                                org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource)
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitSerializer.html new file mode 100644 index 0000000000..39cfbbbbc1 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitSerializer.html @@ -0,0 +1,340 @@ + + + + + +LakeSplitSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LakeSplitSerializer

+
+
+ +
+
    +
  • +
    +
    public class LakeSplitSerializer
    +extends Object
    +
    A serializer for lake split.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LakeSplitSerializer

        +
        public LakeSplitSerializer​(org.apache.fluss.lake.serializer.SimpleVersionedSerializer<org.apache.fluss.lake.source.LakeSplit> sourceSplitSerializer)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + + + + + +
        +
      • +

        deserialize

        +
        public SourceSplitBase deserialize​(byte splitKind,
        +                                   org.apache.fluss.metadata.TableBucket tableBucket,
        +                                   @Nullable
        +                                   String partition,
        +                                   org.apache.flink.core.memory.DataInputDeserializer input)
        +                            throws IOException
        +
        +
        Throws:
        +
        IOException
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitStateInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitStateInitializer.html new file mode 100644 index 0000000000..2df0c2c4b2 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitStateInitializer.html @@ -0,0 +1,307 @@ + + + + + +LakeSplitStateInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LakeSplitStateInitializer

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.lake.LakeSplitStateInitializer
    • +
    +
  • +
+
+
    +
  • +
    +
    public class LakeSplitStateInitializer
    +extends Object
    +
    The state initializer for lake split.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LakeSplitStateInitializer

        +
        public LakeSplitStateInitializer()
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeTableFactory.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeTableFactory.html new file mode 100644 index 0000000000..ab55e3d90f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeTableFactory.html @@ -0,0 +1,309 @@ + + + + + +LakeTableFactory (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LakeTableFactory

+
+
+ +
+
    +
  • +
    +
    public class LakeTableFactory
    +extends Object
    +
    A factory to create DynamicTableSource for lake table.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LakeTableFactory

        +
        public LakeTableFactory()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        createDynamicTableSource

        +
        public org.apache.flink.table.connector.source.DynamicTableSource createDynamicTableSource​(org.apache.flink.table.factories.DynamicTableFactory.Context context,
        +                                                                                           String tableName)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeCatalog.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeCatalog.html new file mode 100644 index 0000000000..f852d878f8 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeCatalog.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.lake.LakeCatalog (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.lake.LakeCatalog

+
+
No usage of org.apache.fluss.flink.lake.LakeCatalog
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeRecordRecordEmitter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeRecordRecordEmitter.html new file mode 100644 index 0000000000..08283101e1 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeRecordRecordEmitter.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.lake.LakeRecordRecordEmitter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.lake.LakeRecordRecordEmitter

+
+
No usage of org.apache.fluss.flink.lake.LakeRecordRecordEmitter
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitGenerator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitGenerator.html new file mode 100644 index 0000000000..ea625b4ad0 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitGenerator.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.lake.LakeSplitGenerator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.lake.LakeSplitGenerator

+
+
No usage of org.apache.fluss.flink.lake.LakeSplitGenerator
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitReaderGenerator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitReaderGenerator.html new file mode 100644 index 0000000000..39072450d6 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitReaderGenerator.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.lake.LakeSplitReaderGenerator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.lake.LakeSplitReaderGenerator

+
+
No usage of org.apache.fluss.flink.lake.LakeSplitReaderGenerator
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitSerializer.html new file mode 100644 index 0000000000..89ca9d3711 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitSerializer.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.lake.LakeSplitSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.lake.LakeSplitSerializer

+
+
No usage of org.apache.fluss.flink.lake.LakeSplitSerializer
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitStateInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitStateInitializer.html new file mode 100644 index 0000000000..cf9dd9dd1a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitStateInitializer.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.lake.LakeSplitStateInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.lake.LakeSplitStateInitializer

+
+
No usage of org.apache.fluss.flink.lake.LakeSplitStateInitializer
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeTableFactory.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeTableFactory.html new file mode 100644 index 0000000000..a0feff2eab --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeTableFactory.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.lake.LakeTableFactory (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.lake.LakeTableFactory

+
+
No usage of org.apache.fluss.flink.lake.LakeTableFactory
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-summary.html new file mode 100644 index 0000000000..1629a25457 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-summary.html @@ -0,0 +1,204 @@ + + + + + +org.apache.fluss.flink.lake (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.lake

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-tree.html new file mode 100644 index 0000000000..433c85bd89 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-tree.html @@ -0,0 +1,169 @@ + + + + + +org.apache.fluss.flink.lake Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.lake

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-use.html new file mode 100644 index 0000000000..c48882c657 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-use.html @@ -0,0 +1,148 @@ + + + + + +Uses of Package org.apache.fluss.flink.lake (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.lake

+
+
No usage of org.apache.fluss.flink.lake
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/IndexedLakeSplitRecordIterator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/IndexedLakeSplitRecordIterator.html new file mode 100644 index 0000000000..d726d01215 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/IndexedLakeSplitRecordIterator.html @@ -0,0 +1,393 @@ + + + + + +IndexedLakeSplitRecordIterator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class IndexedLakeSplitRecordIterator

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.lake.reader.IndexedLakeSplitRecordIterator
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    Closeable, AutoCloseable, Iterator<org.apache.fluss.row.InternalRow>, org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow>
    +
    +
    +
    public class IndexedLakeSplitRecordIterator
    +extends Object
    +implements org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow>
    +
    An iterator wrapper that converts LogRecord objects to InternalRow objects while tracking the + current LakeSplit index being processed. + +

    This class serves as an adapter between the underlying LogRecord iterator and the InternalRow + interface expected by consumers. It maintains reference to the specific LakeSplit index that is + currently being iterated. + +

    Primary responsibilities: + +

      +
    • Wraps a LogRecord iterator and exposes InternalRow objects +
    • Preserves the index of the LakeSplit being processed +
    • Provides clean resource management through Closeable interface +
    • Maintains iterator semantics for sequential data access +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        IndexedLakeSplitRecordIterator

        +
        public IndexedLakeSplitRecordIterator​(org.apache.fluss.utils.CloseableIterator<org.apache.fluss.record.LogRecord> logRecordIterators,
        +                                      int currentLakeSplitIndex)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getCurrentLakeSplitIndex

        +
        public int getCurrentLakeSplitIndex()
        +
      • +
      + + + +
        +
      • +

        close

        +
        public void close()
        +
        +
        Specified by:
        +
        close in interface AutoCloseable
        +
        Specified by:
        +
        close in interface Closeable
        +
        Specified by:
        +
        close in interface org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow>
        +
        +
      • +
      + + + +
        +
      • +

        hasNext

        +
        public boolean hasNext()
        +
        +
        Specified by:
        +
        hasNext in interface Iterator<org.apache.fluss.row.InternalRow>
        +
        +
      • +
      + + + +
        +
      • +

        next

        +
        public org.apache.fluss.row.InternalRow next()
        +
        +
        Specified by:
        +
        next in interface Iterator<org.apache.fluss.row.InternalRow>
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/KeyValueRow.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/KeyValueRow.html new file mode 100644 index 0000000000..08278b13e8 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/KeyValueRow.html @@ -0,0 +1,339 @@ + + + + + +KeyValueRow (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class KeyValueRow

+
+
+ +
+
    +
  • +
    +
    public class KeyValueRow
    +extends Object
    +
    An InternalRow with the key part.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        KeyValueRow

        +
        public KeyValueRow​(int[] keyIndexes,
        +                   org.apache.fluss.row.InternalRow valueRow,
        +                   boolean isDelete)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        isDelete

        +
        public boolean isDelete()
        +
      • +
      + + + +
        +
      • +

        keyRow

        +
        public org.apache.fluss.row.InternalRow keyRow()
        +
      • +
      + + + +
        +
      • +

        valueRow

        +
        public org.apache.fluss.row.InternalRow valueRow()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/LakeSnapshotAndLogSplitScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/LakeSnapshotAndLogSplitScanner.html new file mode 100644 index 0000000000..1d0daa3306 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/LakeSnapshotAndLogSplitScanner.html @@ -0,0 +1,363 @@ + + + + + +LakeSnapshotAndLogSplitScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LakeSnapshotAndLogSplitScanner

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.lake.reader.LakeSnapshotAndLogSplitScanner
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LakeSnapshotAndLogSplitScanner

        +
        public LakeSnapshotAndLogSplitScanner​(Table table,
        +                                      org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource,
        +                                      LakeSnapshotAndFlussLogSplit lakeSnapshotAndFlussLogSplit,
        +                                      @Nullable
        +                                      int[] projectedFields)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        pollBatch

        +
        @Nullable
        +public org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow> pollBatch​(Duration timeout)
        +                                                                                     throws IOException
        +
        Description copied from interface: BatchScanner
        +
        Poll one batch records. The method should return null when reaching the end of the input.
        +
        +
        Specified by:
        +
        pollBatch in interface BatchScanner
        +
        Parameters:
        +
        timeout - The maximum time to block (must not be greater than Long.MAX_VALUE + milliseconds)
        +
        Throws:
        +
        IOException
        +
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/LakeSnapshotScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/LakeSnapshotScanner.html new file mode 100644 index 0000000000..7beb53751c --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/LakeSnapshotScanner.html @@ -0,0 +1,358 @@ + + + + + +LakeSnapshotScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LakeSnapshotScanner

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.lake.reader.LakeSnapshotScanner
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LakeSnapshotScanner

        +
        public LakeSnapshotScanner​(org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource,
        +                           LakeSnapshotSplit lakeSnapshotSplit)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        pollBatch

        +
        @Nullable
        +public org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow> pollBatch​(Duration timeout)
        +                                                                                     throws IOException
        +
        Description copied from interface: BatchScanner
        +
        Poll one batch records. The method should return null when reaching the end of the input.
        +
        +
        Specified by:
        +
        pollBatch in interface BatchScanner
        +
        Parameters:
        +
        timeout - The maximum time to block (must not be greater than Long.MAX_VALUE + milliseconds)
        +
        Throws:
        +
        IOException
        +
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/SeekableLakeSnapshotSplitScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/SeekableLakeSnapshotSplitScanner.html new file mode 100644 index 0000000000..d46ff11ce9 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/SeekableLakeSnapshotSplitScanner.html @@ -0,0 +1,373 @@ + + + + + +SeekableLakeSnapshotSplitScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class SeekableLakeSnapshotSplitScanner

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.lake.reader.SeekableLakeSnapshotSplitScanner
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    Closeable, AutoCloseable, BatchScanner
    +
    +
    +
    public class SeekableLakeSnapshotSplitScanner
    +extends Object
    +implements BatchScanner
    +
    A scanner that supports seeking to a specific LakeSplit and reading from that point. + +

    This scanner enables direct positioning to any LakeSplit in the list using its index, and then + sequentially reads data starting from that specific split. It provides fine-grained control over + where to begin the scanning process. + +

    Key capabilities: + +

      +
    • Direct seeking to any LakeSplit by index +
    • Sequential reading starting from the sought split +
    • Precise positioning within the split collection +
    • Resumable scanning from arbitrary positions +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        SeekableLakeSnapshotSplitScanner

        +
        public SeekableLakeSnapshotSplitScanner​(org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource,
        +                                        List<org.apache.fluss.lake.source.LakeSplit> lakeSplits,
        +                                        int currentLakeSplitIndex)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        pollBatch

        +
        @Nullable
        +public org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow> pollBatch​(Duration timeout)
        +                                                                                     throws IOException
        +
        Description copied from interface: BatchScanner
        +
        Poll one batch records. The method should return null when reaching the end of the input.
        +
        +
        Specified by:
        +
        pollBatch in interface BatchScanner
        +
        Parameters:
        +
        timeout - The maximum time to block (must not be greater than Long.MAX_VALUE + milliseconds)
        +
        Throws:
        +
        IOException
        +
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/IndexedLakeSplitRecordIterator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/IndexedLakeSplitRecordIterator.html new file mode 100644 index 0000000000..4516c7f2b3 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/IndexedLakeSplitRecordIterator.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.lake.reader.IndexedLakeSplitRecordIterator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.lake.reader.IndexedLakeSplitRecordIterator

+
+
No usage of org.apache.fluss.flink.lake.reader.IndexedLakeSplitRecordIterator
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/KeyValueRow.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/KeyValueRow.html new file mode 100644 index 0000000000..a9c2f25fb0 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/KeyValueRow.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.lake.reader.KeyValueRow (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.lake.reader.KeyValueRow

+
+
No usage of org.apache.fluss.flink.lake.reader.KeyValueRow
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/LakeSnapshotAndLogSplitScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/LakeSnapshotAndLogSplitScanner.html new file mode 100644 index 0000000000..34e03e956e --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/LakeSnapshotAndLogSplitScanner.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.lake.reader.LakeSnapshotAndLogSplitScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.lake.reader.LakeSnapshotAndLogSplitScanner

+
+
No usage of org.apache.fluss.flink.lake.reader.LakeSnapshotAndLogSplitScanner
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/LakeSnapshotScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/LakeSnapshotScanner.html new file mode 100644 index 0000000000..5e0c53a699 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/LakeSnapshotScanner.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.lake.reader.LakeSnapshotScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.lake.reader.LakeSnapshotScanner

+
+
No usage of org.apache.fluss.flink.lake.reader.LakeSnapshotScanner
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/SeekableLakeSnapshotSplitScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/SeekableLakeSnapshotSplitScanner.html new file mode 100644 index 0000000000..b7a6d1173f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/SeekableLakeSnapshotSplitScanner.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.lake.reader.SeekableLakeSnapshotSplitScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.lake.reader.SeekableLakeSnapshotSplitScanner

+
+
No usage of org.apache.fluss.flink.lake.reader.SeekableLakeSnapshotSplitScanner
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-summary.html new file mode 100644 index 0000000000..76c18f9c2b --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-summary.html @@ -0,0 +1,193 @@ + + + + + +org.apache.fluss.flink.lake.reader (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.lake.reader

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-tree.html new file mode 100644 index 0000000000..62e38a9240 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-tree.html @@ -0,0 +1,167 @@ + + + + + +org.apache.fluss.flink.lake.reader Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.lake.reader

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-use.html new file mode 100644 index 0000000000..cb01e963c5 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-use.html @@ -0,0 +1,148 @@ + + + + + +Uses of Package org.apache.fluss.flink.lake.reader (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.lake.reader

+
+
No usage of org.apache.fluss.flink.lake.reader
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/LakeSnapshotAndFlussLogSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/LakeSnapshotAndFlussLogSplit.html new file mode 100644 index 0000000000..645fdc89da --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/LakeSnapshotAndFlussLogSplit.html @@ -0,0 +1,624 @@ + + + + + +LakeSnapshotAndFlussLogSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LakeSnapshotAndFlussLogSplit

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.api.connector.source.SourceSplit
    +
    +
    +
    public class LakeSnapshotAndFlussLogSplit
    +extends SourceSplitBase
    +
    A split mixing Lake snapshot and Fluss log.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + +
        +
      • +

        LAKE_SNAPSHOT_FLUSS_LOG_SPLIT_KIND

        +
        public static final byte LAKE_SNAPSHOT_FLUSS_LOG_SPLIT_KIND
        +
        +
        See Also:
        +
        Constant Field Values
        +
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LakeSnapshotAndFlussLogSplit

        +
        public LakeSnapshotAndFlussLogSplit​(org.apache.fluss.metadata.TableBucket tableBucket,
        +                                    @Nullable
        +                                    String partitionName,
        +                                    @Nullable
        +                                    List<org.apache.fluss.lake.source.LakeSplit> snapshotSplits,
        +                                    long startingOffset,
        +                                    long stoppingOffset)
        +
      • +
      + + + +
        +
      • +

        LakeSnapshotAndFlussLogSplit

        +
        public LakeSnapshotAndFlussLogSplit​(org.apache.fluss.metadata.TableBucket tableBucket,
        +                                    @Nullable
        +                                    String partitionName,
        +                                    @Nullable
        +                                    List<org.apache.fluss.lake.source.LakeSplit> snapshotSplits,
        +                                    long startingOffset,
        +                                    long stoppingOffset,
        +                                    long recordsToSkip,
        +                                    int currentLakeSplitIndex,
        +                                    boolean isLakeSplitFinished)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + + + + + +
        +
      • +

        updateWithCurrentLakeSplitIndex

        +
        public LakeSnapshotAndFlussLogSplit updateWithCurrentLakeSplitIndex​(int currentLakeSplitIndex)
        +
      • +
      + + + + + + + + + + + +
        +
      • +

        getRecordsToSkip

        +
        public long getRecordsToSkip()
        +
      • +
      + + + +
        +
      • +

        getStartingOffset

        +
        public long getStartingOffset()
        +
      • +
      + + + +
        +
      • +

        isLakeSplitFinished

        +
        public boolean isLakeSplitFinished()
        +
      • +
      + + + +
        +
      • +

        getStoppingOffset

        +
        public Optional<Long> getStoppingOffset()
        +
      • +
      + + + + + + + +
        +
      • +

        isStreaming

        +
        public boolean isStreaming()
        +
      • +
      + + + + + + + +
        +
      • +

        splitId

        +
        public String splitId()
        +
      • +
      + + + +
        +
      • +

        getLakeSplits

        +
        @Nullable
        +public List<org.apache.fluss.lake.source.LakeSplit> getLakeSplits()
        +
      • +
      + + + +
        +
      • +

        getCurrentLakeSplitIndex

        +
        public int getCurrentLakeSplitIndex()
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/LakeSnapshotSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/LakeSnapshotSplit.html new file mode 100644 index 0000000000..af027ee0df --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/LakeSnapshotSplit.html @@ -0,0 +1,501 @@ + + + + + +LakeSnapshotSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LakeSnapshotSplit

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.api.connector.source.SourceSplit
    +
    +
    +
    public class LakeSnapshotSplit
    +extends SourceSplitBase
    +
    A split for reading a snapshot of lake.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + +
        +
      • +

        LAKE_SNAPSHOT_SPLIT_KIND

        +
        public static final byte LAKE_SNAPSHOT_SPLIT_KIND
        +
        +
        See Also:
        +
        Constant Field Values
        +
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LakeSnapshotSplit

        +
        public LakeSnapshotSplit​(org.apache.fluss.metadata.TableBucket tableBucket,
        +                         @Nullable
        +                         String partitionName,
        +                         org.apache.fluss.lake.source.LakeSplit lakeSplit,
        +                         int splitIndex)
        +
      • +
      + + + +
        +
      • +

        LakeSnapshotSplit

        +
        public LakeSnapshotSplit​(org.apache.fluss.metadata.TableBucket tableBucket,
        +                         @Nullable
        +                         String partitionName,
        +                         org.apache.fluss.lake.source.LakeSplit lakeSplit,
        +                         int splitIndex,
        +                         long recordsToSkip)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getLakeSplit

        +
        public org.apache.fluss.lake.source.LakeSplit getLakeSplit()
        +
      • +
      + + + +
        +
      • +

        getRecordsToSkip

        +
        public long getRecordsToSkip()
        +
      • +
      + + + +
        +
      • +

        getSplitIndex

        +
        public int getSplitIndex()
        +
      • +
      + + + +
        +
      • +

        splitId

        +
        public String splitId()
        +
      • +
      + + + + + + + + + + + + +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/class-use/LakeSnapshotAndFlussLogSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/class-use/LakeSnapshotAndFlussLogSplit.html new file mode 100644 index 0000000000..71eedad85c --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/class-use/LakeSnapshotAndFlussLogSplit.html @@ -0,0 +1,258 @@ + + + + + +Uses of Class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/class-use/LakeSnapshotSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/class-use/LakeSnapshotSplit.html new file mode 100644 index 0000000000..90ef942ece --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/class-use/LakeSnapshotSplit.html @@ -0,0 +1,215 @@ + + + + + +Uses of Class org.apache.fluss.flink.lake.split.LakeSnapshotSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.lake.split.LakeSnapshotSplit

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-summary.html new file mode 100644 index 0000000000..c3cb71f828 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-summary.html @@ -0,0 +1,174 @@ + + + + + +org.apache.fluss.flink.lake.split (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.lake.split

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-tree.html new file mode 100644 index 0000000000..8cc69f148f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-tree.html @@ -0,0 +1,168 @@ + + + + + +org.apache.fluss.flink.lake.split Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.lake.split

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-use.html new file mode 100644 index 0000000000..cbd3b7b60c --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-use.html @@ -0,0 +1,243 @@ + + + + + +Uses of Package org.apache.fluss.flink.lake.split (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.lake.split

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/LakeSnapshotAndFlussLogSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/LakeSnapshotAndFlussLogSplitState.html new file mode 100644 index 0000000000..a6559716cc --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/LakeSnapshotAndFlussLogSplitState.html @@ -0,0 +1,365 @@ + + + + + +LakeSnapshotAndFlussLogSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LakeSnapshotAndFlussLogSplitState

+
+
+ +
+ +
+
+ +
+
+
    +
  • + +
    + +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        setRecordsToSkip

        +
        public void setRecordsToSkip​(long recordsToSkip)
        +
      • +
      + + + +
        +
      • +

        setCurrentLakeSplitIndex

        +
        public void setCurrentLakeSplitIndex​(int currentLakeSplitIndex)
        +
      • +
      + + + +
        +
      • +

        setNextLogOffset

        +
        public void setNextLogOffset​(long nextOffset)
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/LakeSnapshotSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/LakeSnapshotSplitState.html new file mode 100644 index 0000000000..e8af1ac39e --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/LakeSnapshotSplitState.html @@ -0,0 +1,337 @@ + + + + + +LakeSnapshotSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LakeSnapshotSplitState

+
+
+ +
+ +
+
+ +
+
+ +
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/class-use/LakeSnapshotAndFlussLogSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/class-use/LakeSnapshotAndFlussLogSplitState.html new file mode 100644 index 0000000000..ed62ff9265 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/class-use/LakeSnapshotAndFlussLogSplitState.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.lake.state.LakeSnapshotAndFlussLogSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.lake.state.LakeSnapshotAndFlussLogSplitState

+
+
No usage of org.apache.fluss.flink.lake.state.LakeSnapshotAndFlussLogSplitState
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/class-use/LakeSnapshotSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/class-use/LakeSnapshotSplitState.html new file mode 100644 index 0000000000..cc43984171 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/class-use/LakeSnapshotSplitState.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.lake.state.LakeSnapshotSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.lake.state.LakeSnapshotSplitState

+
+
No usage of org.apache.fluss.flink.lake.state.LakeSnapshotSplitState
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-summary.html new file mode 100644 index 0000000000..f8d1c51552 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-summary.html @@ -0,0 +1,174 @@ + + + + + +org.apache.fluss.flink.lake.state (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.lake.state

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-tree.html new file mode 100644 index 0000000000..a24cfa7be7 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-tree.html @@ -0,0 +1,168 @@ + + + + + +org.apache.fluss.flink.lake.state Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.lake.state

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-use.html new file mode 100644 index 0000000000..b746e9fbd3 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-use.html @@ -0,0 +1,148 @@ + + + + + +Uses of Package org.apache.fluss.flink.lake.state (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.lake.state

+
+
No usage of org.apache.fluss.flink.lake.state
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkCounter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkCounter.html new file mode 100644 index 0000000000..9138323bba --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkCounter.html @@ -0,0 +1,395 @@ + + + + + +FlinkCounter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkCounter

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.metrics.Counter, org.apache.flink.metrics.Metric
    +
    +
    +
    public class FlinkCounter
    +extends Object
    +implements org.apache.flink.metrics.Counter
    +
    An implementation of Flink's Counter which wraps Fluss's Counter.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlinkCounter

        +
        public FlinkCounter​(org.apache.fluss.metrics.Counter wrapped)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        inc

        +
        public void inc()
        +
        +
        Specified by:
        +
        inc in interface org.apache.flink.metrics.Counter
        +
        +
      • +
      + + + +
        +
      • +

        inc

        +
        public void inc​(long n)
        +
        +
        Specified by:
        +
        inc in interface org.apache.flink.metrics.Counter
        +
        +
      • +
      + + + +
        +
      • +

        dec

        +
        public void dec()
        +
        +
        Specified by:
        +
        dec in interface org.apache.flink.metrics.Counter
        +
        +
      • +
      + + + +
        +
      • +

        dec

        +
        public void dec​(long n)
        +
        +
        Specified by:
        +
        dec in interface org.apache.flink.metrics.Counter
        +
        +
      • +
      + + + +
        +
      • +

        getCount

        +
        public long getCount()
        +
        +
        Specified by:
        +
        getCount in interface org.apache.flink.metrics.Counter
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkGauge.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkGauge.html new file mode 100644 index 0000000000..6375c809e5 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkGauge.html @@ -0,0 +1,323 @@ + + + + + +FlinkGauge (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkGauge<T>

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.metrics.Gauge<T>, org.apache.flink.metrics.Metric
    +
    +
    +
    public class FlinkGauge<T>
    +extends Object
    +implements org.apache.flink.metrics.Gauge<T>
    +
    An implementation of Flink's Gauge which wraps Fluss's Gauge.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlinkGauge

        +
        public FlinkGauge​(org.apache.fluss.metrics.Gauge<T> wrapped)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getValue

        +
        public T getValue()
        +
        +
        Specified by:
        +
        getValue in interface org.apache.flink.metrics.Gauge<T>
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkHistogram.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkHistogram.html new file mode 100644 index 0000000000..fbbf179c37 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkHistogram.html @@ -0,0 +1,359 @@ + + + + + +FlinkHistogram (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkHistogram

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.metrics.Histogram, org.apache.flink.metrics.Metric
    +
    +
    +
    public class FlinkHistogram
    +extends Object
    +implements org.apache.flink.metrics.Histogram
    +
    An implementation of Flink's Histogram which wraps Fluss's Histogram.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlinkHistogram

        +
        public FlinkHistogram​(org.apache.fluss.metrics.Histogram wrapped)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        update

        +
        public void update​(long n)
        +
        +
        Specified by:
        +
        update in interface org.apache.flink.metrics.Histogram
        +
        +
      • +
      + + + +
        +
      • +

        getCount

        +
        public long getCount()
        +
        +
        Specified by:
        +
        getCount in interface org.apache.flink.metrics.Histogram
        +
        +
      • +
      + + + +
        +
      • +

        getStatistics

        +
        public org.apache.flink.metrics.HistogramStatistics getStatistics()
        +
        +
        Specified by:
        +
        getStatistics in interface org.apache.flink.metrics.Histogram
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkMeter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkMeter.html new file mode 100644 index 0000000000..840eba1721 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkMeter.html @@ -0,0 +1,377 @@ + + + + + +FlinkMeter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkMeter

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.metrics.Meter, org.apache.flink.metrics.Metric
    +
    +
    +
    public class FlinkMeter
    +extends Object
    +implements org.apache.flink.metrics.Meter
    +
    An implementation of Flink's Meter which wraps Fluss's Meter.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlinkMeter

        +
        public FlinkMeter​(org.apache.fluss.metrics.Meter wrapped)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        markEvent

        +
        public void markEvent()
        +
        +
        Specified by:
        +
        markEvent in interface org.apache.flink.metrics.Meter
        +
        +
      • +
      + + + +
        +
      • +

        markEvent

        +
        public void markEvent​(long n)
        +
        +
        Specified by:
        +
        markEvent in interface org.apache.flink.metrics.Meter
        +
        +
      • +
      + + + +
        +
      • +

        getRate

        +
        public double getRate()
        +
        +
        Specified by:
        +
        getRate in interface org.apache.flink.metrics.Meter
        +
        +
      • +
      + + + +
        +
      • +

        getCount

        +
        public long getCount()
        +
        +
        Specified by:
        +
        getCount in interface org.apache.flink.metrics.Meter
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkMetricRegistry.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkMetricRegistry.html new file mode 100644 index 0000000000..a011ff7b87 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkMetricRegistry.html @@ -0,0 +1,477 @@ + + + + + +FlinkMetricRegistry (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkMetricRegistry

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.metrics.FlinkMetricRegistry
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    AutoCloseable, org.apache.fluss.metrics.registry.MetricRegistry, org.apache.fluss.utils.AutoCloseableAsync
    +
    +
    +
    public class FlinkMetricRegistry
    +extends Object
    +implements org.apache.fluss.metrics.registry.MetricRegistry
    +
    An implementation of MetricRegistry which registers all metrics into Flink's metric + system. It's mainly used for Fluss client to register its metrics to Flink's metric system. + +

    All metrics of the Fluss client are registered under group "fluss", which is a child group of + OperatorMetricGroup. + +

    For example, the metrics sendLatencyMs will be reported in metric: + "{some_parent_groups}.operator.fluss.client_writer.client_id.sendLatencyMs"

    +
  • +
+
+
+ +
+
+
    +
  • + +
    + +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlinkMetricRegistry

        +
        public FlinkMetricRegistry​(org.apache.flink.metrics.MetricGroup flinkOperatorMetricGroup)
        +
      • +
      + + + +
        +
      • +

        FlinkMetricRegistry

        +
        public FlinkMetricRegistry​(org.apache.flink.metrics.MetricGroup flinkOperatorMetricGroup,
        +                           Set<String> exposedMetricNames)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getNumberReporters

        +
        public int getNumberReporters()
        +
        +
        Specified by:
        +
        getNumberReporters in interface org.apache.fluss.metrics.registry.MetricRegistry
        +
        +
      • +
      + + + +
        +
      • +

        register

        +
        public void register​(org.apache.fluss.metrics.Metric metric,
        +                     String metricName,
        +                     org.apache.fluss.metrics.groups.AbstractMetricGroup group)
        +
        +
        Specified by:
        +
        register in interface org.apache.fluss.metrics.registry.MetricRegistry
        +
        +
      • +
      + + + +
        +
      • +

        getFlussMetric

        +
        public org.apache.fluss.metrics.Metric getFlussMetric​(String metricName)
        +
        Exposes the metrics of Fluss metics group for flink.
        +
      • +
      + + + +
        +
      • +

        unregister

        +
        public void unregister​(org.apache.fluss.metrics.Metric metric,
        +                       String metricName,
        +                       org.apache.fluss.metrics.groups.AbstractMetricGroup group)
        +
        +
        Specified by:
        +
        unregister in interface org.apache.fluss.metrics.registry.MetricRegistry
        +
        +
      • +
      + + + +
        +
      • +

        closeAsync

        +
        public CompletableFuture<Void> closeAsync()
        +
        +
        Specified by:
        +
        closeAsync in interface org.apache.fluss.utils.AutoCloseableAsync
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkCounter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkCounter.html new file mode 100644 index 0000000000..4da2cba43b --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkCounter.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.metrics.FlinkCounter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.metrics.FlinkCounter

+
+
No usage of org.apache.fluss.flink.metrics.FlinkCounter
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkGauge.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkGauge.html new file mode 100644 index 0000000000..18ae20f5c8 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkGauge.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.metrics.FlinkGauge (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.metrics.FlinkGauge

+
+
No usage of org.apache.fluss.flink.metrics.FlinkGauge
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkHistogram.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkHistogram.html new file mode 100644 index 0000000000..e655003d0d --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkHistogram.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.metrics.FlinkHistogram (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.metrics.FlinkHistogram

+
+
No usage of org.apache.fluss.flink.metrics.FlinkHistogram
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkMeter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkMeter.html new file mode 100644 index 0000000000..2363d21a3e --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkMeter.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.metrics.FlinkMeter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.metrics.FlinkMeter

+
+
No usage of org.apache.fluss.flink.metrics.FlinkMeter
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkMetricRegistry.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkMetricRegistry.html new file mode 100644 index 0000000000..90af203b33 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkMetricRegistry.html @@ -0,0 +1,192 @@ + + + + + +Uses of Class org.apache.fluss.flink.metrics.FlinkMetricRegistry (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.metrics.FlinkMetricRegistry

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-summary.html new file mode 100644 index 0000000000..f56c2ad488 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-summary.html @@ -0,0 +1,193 @@ + + + + + +org.apache.fluss.flink.metrics (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.metrics

+
+
+
    +
  • + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Class Summary 
    ClassDescription
    FlinkCounter +
    An implementation of Flink's Counter which wraps Fluss's Counter.
    +
    FlinkGauge<T> +
    An implementation of Flink's Gauge which wraps Fluss's Gauge.
    +
    FlinkHistogram +
    An implementation of Flink's Histogram which wraps Fluss's Histogram.
    +
    FlinkMeter +
    An implementation of Flink's Meter which wraps Fluss's Meter.
    +
    FlinkMetricRegistry +
    An implementation of MetricRegistry which registers all metrics into Flink's metric + system.
    +
    +
  • +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-tree.html new file mode 100644 index 0000000000..244a85acd2 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-tree.html @@ -0,0 +1,167 @@ + + + + + +org.apache.fluss.flink.metrics Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.metrics

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+
    +
  • java.lang.Object +
      +
    • org.apache.fluss.flink.metrics.FlinkCounter (implements org.apache.flink.metrics.Counter)
    • +
    • org.apache.fluss.flink.metrics.FlinkGauge<T> (implements org.apache.flink.metrics.Gauge<T>)
    • +
    • org.apache.fluss.flink.metrics.FlinkHistogram (implements org.apache.flink.metrics.Histogram)
    • +
    • org.apache.fluss.flink.metrics.FlinkMeter (implements org.apache.flink.metrics.Meter)
    • +
    • org.apache.fluss.flink.metrics.FlinkMetricRegistry (implements org.apache.fluss.metrics.registry.MetricRegistry)
    • +
    +
  • +
+
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-use.html new file mode 100644 index 0000000000..c6b70f1d48 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-use.html @@ -0,0 +1,186 @@ + + + + + +Uses of Package org.apache.fluss.flink.metrics (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.metrics

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-summary.html new file mode 100644 index 0000000000..acab3a57ef --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-summary.html @@ -0,0 +1,185 @@ + + + + + +org.apache.fluss.flink (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-tree.html new file mode 100644 index 0000000000..72302d1b1f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-tree.html @@ -0,0 +1,177 @@ + + + + + +org.apache.fluss.flink Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+

Enum Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-use.html new file mode 100644 index 0000000000..78c86f7bdc --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-use.html @@ -0,0 +1,208 @@ + + + + + +Uses of Package org.apache.fluss.flink (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/AbstractAclProcedure.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/AbstractAclProcedure.html new file mode 100644 index 0000000000..7a6158f929 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/AbstractAclProcedure.html @@ -0,0 +1,419 @@ + + + + + +AbstractAclProcedure (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class AbstractAclProcedure

+
+
+ +
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        AbstractAclProcedure

        +
        public AbstractAclProcedure()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        internalCall

        +
        protected String[] internalCall​(@Nullable
        +                                String resource,
        +                                @Nullable
        +                                String permission,
        +                                @Nullable
        +                                String principal,
        +                                @Nullable
        +                                String operation,
        +                                @Nullable
        +                                String host)
        +                         throws Exception
        +
        Invokes the ACL operation as a Flink table procedure. + +

        This method serves as the entry point for executing ACL operations (ADD, DROP, LIST) + through the Flink SQL procedure interface. It delegates execution to the internalCall method.

        +
        +
        Parameters:
        +
        resource - Resource on which the ACL operation applies. The format must be one of: +
          +
        • cluster - cluster level +
        • cluster.db_name - database level +
        • cluster.db_name.table_name - table level +
        +
        permission - Permission type to grant or revoke. Common values include in PermissionType.
        +
        principal - Principal (user or role) to apply the ACL to. Accepts: +
          +
        • ANY +
        • ALL +
        • PrincipalType:PrincipalName, e.g., User:alice +
        +
        operation - Operation type applied on the resource. Common values include in OperationType.
        +
        Returns:
        +
        An array of strings representing the result of the operation: +
          +
        • ["success"] for ADD/DROP operations upon success. +
        • For LIST operations, returns a list of formatted ACL entries as strings. +
        +
        Throws:
        +
        ExecutionException - if an error occurs during the execution of the ACL operation.
        +
        InterruptedException - if the current thread is interrupted while waiting for the + operation to complete.
        +
        Exception
        +
        +
      • +
      + + + +
        +
      • +

        aclOperation

        +
        protected abstract String[] aclOperation​(org.apache.fluss.security.acl.Resource resource,
        +                                         org.apache.fluss.security.acl.PermissionType permission,
        +                                         org.apache.fluss.security.acl.FlussPrincipal flussPrincipal,
        +                                         org.apache.fluss.security.acl.OperationType operationType,
        +                                         String host)
        +                                  throws Exception
        +
        +
        Throws:
        +
        Exception
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/AddAclProcedure.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/AddAclProcedure.html new file mode 100644 index 0000000000..0d1b896e61 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/AddAclProcedure.html @@ -0,0 +1,397 @@ + + + + + +AddAclProcedure (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class AddAclProcedure

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.table.procedures.Procedure
    +
    +
    +
    public class AddAclProcedure
    +extends AbstractAclProcedure
    +
    Procedure to add acl.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        AddAclProcedure

        +
        public AddAclProcedure()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + + + + + +
        +
      • +

        aclOperation

        +
        protected String[] aclOperation​(org.apache.fluss.security.acl.Resource resource,
        +                                org.apache.fluss.security.acl.PermissionType permission,
        +                                org.apache.fluss.security.acl.FlussPrincipal flussPrincipal,
        +                                org.apache.fluss.security.acl.OperationType operationType,
        +                                String host)
        +                         throws Exception
        +
        +
        Specified by:
        +
        aclOperation in class AbstractAclProcedure
        +
        Throws:
        +
        Exception
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/DropAclProcedure.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/DropAclProcedure.html new file mode 100644 index 0000000000..3aabc9c04a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/DropAclProcedure.html @@ -0,0 +1,401 @@ + + + + + +DropAclProcedure (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class DropAclProcedure

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.table.procedures.Procedure
    +
    +
    +
    public class DropAclProcedure
    +extends AbstractAclProcedure
    +
    Procedure to drop acl.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        DropAclProcedure

        +
        public DropAclProcedure()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        call

        +
        public String[] call​(org.apache.flink.table.procedure.ProcedureContext context,
        +                     @Nullable
        +                     String resource,
        +                     @Nullable
        +                     String permission,
        +                     @Nullable
        +                     String principal,
        +                     @Nullable
        +                     String operation,
        +                     @Nullable
        +                     String host)
        +              throws Exception
        +
        +
        Throws:
        +
        Exception
        +
        +
      • +
      + + + +
        +
      • +

        aclOperation

        +
        protected String[] aclOperation​(org.apache.fluss.security.acl.Resource resource,
        +                                org.apache.fluss.security.acl.PermissionType permission,
        +                                org.apache.fluss.security.acl.FlussPrincipal flussPrincipal,
        +                                org.apache.fluss.security.acl.OperationType operationType,
        +                                String host)
        +                         throws Exception
        +
        +
        Specified by:
        +
        aclOperation in class AbstractAclProcedure
        +
        Throws:
        +
        Exception
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ListAclProcedure.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ListAclProcedure.html new file mode 100644 index 0000000000..38ef0eee0a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ListAclProcedure.html @@ -0,0 +1,401 @@ + + + + + +ListAclProcedure (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class ListAclProcedure

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.table.procedures.Procedure
    +
    +
    +
    public class ListAclProcedure
    +extends AbstractAclProcedure
    +
    Procedure to list acl.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        ListAclProcedure

        +
        public ListAclProcedure()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        call

        +
        public String[] call​(org.apache.flink.table.procedure.ProcedureContext context,
        +                     @Nullable
        +                     String resource,
        +                     @Nullable
        +                     String permission,
        +                     @Nullable
        +                     String principal,
        +                     @Nullable
        +                     String operation,
        +                     @Nullable
        +                     String host)
        +              throws Exception
        +
        +
        Throws:
        +
        Exception
        +
        +
      • +
      + + + +
        +
      • +

        aclOperation

        +
        protected String[] aclOperation​(org.apache.fluss.security.acl.Resource resource,
        +                                org.apache.fluss.security.acl.PermissionType permission,
        +                                org.apache.fluss.security.acl.FlussPrincipal flussPrincipal,
        +                                org.apache.fluss.security.acl.OperationType operationType,
        +                                String host)
        +                         throws Exception
        +
        +
        Specified by:
        +
        aclOperation in class AbstractAclProcedure
        +
        Throws:
        +
        Exception
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ProcedureBase.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ProcedureBase.html new file mode 100644 index 0000000000..aaa755d9cc --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ProcedureBase.html @@ -0,0 +1,358 @@ + + + + + +ProcedureBase (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class ProcedureBase

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.table.procedures.Procedure
    +
    +
    +
    Direct Known Subclasses:
    +
    AbstractAclProcedure
    +
    +
    +
    public abstract class ProcedureBase
    +extends Object
    +implements org.apache.flink.table.procedures.Procedure
    +
    Base implementation for flink Procedure.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + +
        +
      • +

        admin

        +
        protected Admin admin
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        ProcedureBase

        +
        public ProcedureBase()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        withAdmin

        +
        public void withAdmin​(Admin admin)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ProcedureManager.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ProcedureManager.html new file mode 100644 index 0000000000..2d9fec3eaf --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ProcedureManager.html @@ -0,0 +1,323 @@ + + + + + +ProcedureManager (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class ProcedureManager

+
+
+ +
+
    +
  • +
    +
    public class ProcedureManager
    +extends Object
    +
    ProcedureUtil to load procedure.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        ProcedureManager

        +
        public ProcedureManager()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        listProcedures

        +
        public static List<String> listProcedures()
        +
      • +
      + + + +
        +
      • +

        getProcedure

        +
        public static Optional<org.apache.flink.table.procedures.Procedure> getProcedure​(Admin admin,
        +                                                                                 org.apache.flink.table.catalog.ObjectPath procedurePath)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/AbstractAclProcedure.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/AbstractAclProcedure.html new file mode 100644 index 0000000000..afab9c8d15 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/AbstractAclProcedure.html @@ -0,0 +1,208 @@ + + + + + +Uses of Class org.apache.fluss.flink.procedure.AbstractAclProcedure (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.procedure.AbstractAclProcedure

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/AddAclProcedure.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/AddAclProcedure.html new file mode 100644 index 0000000000..1231d74227 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/AddAclProcedure.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.procedure.AddAclProcedure (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.procedure.AddAclProcedure

+
+
No usage of org.apache.fluss.flink.procedure.AddAclProcedure
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/DropAclProcedure.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/DropAclProcedure.html new file mode 100644 index 0000000000..09eee4cd64 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/DropAclProcedure.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.procedure.DropAclProcedure (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.procedure.DropAclProcedure

+
+
No usage of org.apache.fluss.flink.procedure.DropAclProcedure
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ListAclProcedure.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ListAclProcedure.html new file mode 100644 index 0000000000..4aa65cc577 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ListAclProcedure.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.procedure.ListAclProcedure (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.procedure.ListAclProcedure

+
+
No usage of org.apache.fluss.flink.procedure.ListAclProcedure
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ProcedureBase.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ProcedureBase.html new file mode 100644 index 0000000000..9a10f1c89d --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ProcedureBase.html @@ -0,0 +1,215 @@ + + + + + +Uses of Class org.apache.fluss.flink.procedure.ProcedureBase (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.procedure.ProcedureBase

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ProcedureManager.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ProcedureManager.html new file mode 100644 index 0000000000..ca411c3606 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ProcedureManager.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.procedure.ProcedureManager (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.procedure.ProcedureManager

+
+
No usage of org.apache.fluss.flink.procedure.ProcedureManager
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-summary.html new file mode 100644 index 0000000000..2bec0feff4 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-summary.html @@ -0,0 +1,198 @@ + + + + + +org.apache.fluss.flink.procedure (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.procedure

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-tree.html new file mode 100644 index 0000000000..7606c3a71f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-tree.html @@ -0,0 +1,174 @@ + + + + + +org.apache.fluss.flink.procedure Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.procedure

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-use.html new file mode 100644 index 0000000000..dfecb1472b --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-use.html @@ -0,0 +1,191 @@ + + + + + +Uses of Package org.apache.fluss.flink.procedure (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.procedure

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/FlinkAsFlussRow.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/FlinkAsFlussRow.html new file mode 100644 index 0000000000..a24da9fd54 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/FlinkAsFlussRow.html @@ -0,0 +1,643 @@ + + + + + +FlinkAsFlussRow (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkAsFlussRow

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.fluss.row.InternalRow
    +
    +
    +
    public class FlinkAsFlussRow
    +extends Object
    +implements org.apache.fluss.row.InternalRow
    +
    Wraps a Flink RowData as a Fluss InternalRow.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlinkAsFlussRow

        +
        public FlinkAsFlussRow()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        replace

        +
        public FlinkAsFlussRow replace​(org.apache.flink.table.data.RowData flinkRow)
        +
      • +
      + + + +
        +
      • +

        getFieldCount

        +
        public int getFieldCount()
        +
        +
        Specified by:
        +
        getFieldCount in interface org.apache.fluss.row.InternalRow
        +
        +
      • +
      + + + +
        +
      • +

        isNullAt

        +
        public boolean isNullAt​(int pos)
        +
        +
        Specified by:
        +
        isNullAt in interface org.apache.fluss.row.InternalRow
        +
        +
      • +
      + + + +
        +
      • +

        getBoolean

        +
        public boolean getBoolean​(int pos)
        +
        +
        Specified by:
        +
        getBoolean in interface org.apache.fluss.row.InternalRow
        +
        +
      • +
      + + + +
        +
      • +

        getByte

        +
        public byte getByte​(int pos)
        +
        +
        Specified by:
        +
        getByte in interface org.apache.fluss.row.InternalRow
        +
        +
      • +
      + + + +
        +
      • +

        getShort

        +
        public short getShort​(int pos)
        +
        +
        Specified by:
        +
        getShort in interface org.apache.fluss.row.InternalRow
        +
        +
      • +
      + + + +
        +
      • +

        getInt

        +
        public int getInt​(int pos)
        +
        +
        Specified by:
        +
        getInt in interface org.apache.fluss.row.InternalRow
        +
        +
      • +
      + + + +
        +
      • +

        getLong

        +
        public long getLong​(int pos)
        +
        +
        Specified by:
        +
        getLong in interface org.apache.fluss.row.InternalRow
        +
        +
      • +
      + + + +
        +
      • +

        getFloat

        +
        public float getFloat​(int pos)
        +
        +
        Specified by:
        +
        getFloat in interface org.apache.fluss.row.InternalRow
        +
        +
      • +
      + + + +
        +
      • +

        getDouble

        +
        public double getDouble​(int pos)
        +
        +
        Specified by:
        +
        getDouble in interface org.apache.fluss.row.InternalRow
        +
        +
      • +
      + + + +
        +
      • +

        getChar

        +
        public org.apache.fluss.row.BinaryString getChar​(int pos,
        +                                                 int length)
        +
        +
        Specified by:
        +
        getChar in interface org.apache.fluss.row.InternalRow
        +
        +
      • +
      + + + +
        +
      • +

        getString

        +
        public org.apache.fluss.row.BinaryString getString​(int pos)
        +
        +
        Specified by:
        +
        getString in interface org.apache.fluss.row.InternalRow
        +
        +
      • +
      + + + +
        +
      • +

        getDecimal

        +
        public org.apache.fluss.row.Decimal getDecimal​(int pos,
        +                                               int precision,
        +                                               int scale)
        +
        +
        Specified by:
        +
        getDecimal in interface org.apache.fluss.row.InternalRow
        +
        +
      • +
      + + + +
        +
      • +

        fromFlinkDecimal

        +
        public static org.apache.fluss.row.Decimal fromFlinkDecimal​(org.apache.flink.table.data.DecimalData decimal)
        +
      • +
      + + + +
        +
      • +

        getTimestampNtz

        +
        public org.apache.fluss.row.TimestampNtz getTimestampNtz​(int pos,
        +                                                         int precision)
        +
        +
        Specified by:
        +
        getTimestampNtz in interface org.apache.fluss.row.InternalRow
        +
        +
      • +
      + + + +
        +
      • +

        getTimestampLtz

        +
        public org.apache.fluss.row.TimestampLtz getTimestampLtz​(int pos,
        +                                                         int precision)
        +
        +
        Specified by:
        +
        getTimestampLtz in interface org.apache.fluss.row.InternalRow
        +
        +
      • +
      + + + +
        +
      • +

        getBinary

        +
        public byte[] getBinary​(int pos,
        +                        int length)
        +
        +
        Specified by:
        +
        getBinary in interface org.apache.fluss.row.InternalRow
        +
        +
      • +
      + + + +
        +
      • +

        getBytes

        +
        public byte[] getBytes​(int pos)
        +
        +
        Specified by:
        +
        getBytes in interface org.apache.fluss.row.InternalRow
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/OperationType.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/OperationType.html new file mode 100644 index 0000000000..49bb4d2285 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/OperationType.html @@ -0,0 +1,432 @@ + + + + + +OperationType (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Enum OperationType

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    Serializable, Comparable<OperationType>
    +
    +
    +
    public enum OperationType
    +extends Enum<OperationType>
    +
    Enumeration of row operation types used in Fluss-Flink data processing. + +

    This enum represents the type of operation associated with a row, such as an append (insert), + upsert (update or insert), delete, or ignore. It is used to indicate how a row should be + interpreted or processed in downstream systems. + +

      +
    • APPEND - Represents an append-only (insert) operation. +
    • UPSERT - Represents an upsert operation (update or insert). +
    • DELETE - Represents a delete operation. +
    • IGNORE - Represents an operation that should be ignored. +
    +
    +
    See Also:
    +
    RowWithOp
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Enum Constant Detail

      + + + +
        +
      • +

        APPEND

        +
        public static final OperationType APPEND
        +
        Represents an append-only (insert) operation.
        +
      • +
      + + + +
        +
      • +

        UPSERT

        +
        public static final OperationType UPSERT
        +
        Represents an upsert operation (update or insert).
        +
      • +
      + + + +
        +
      • +

        DELETE

        +
        public static final OperationType DELETE
        +
        Represents a delete operation.
        +
      • +
      + + + +
        +
      • +

        IGNORE

        +
        public static final OperationType IGNORE
        +
        Represents an operation that should be ignored.
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        values

        +
        public static OperationType[] values()
        +
        Returns an array containing the constants of this enum type, in +the order they are declared. This method may be used to iterate +over the constants as follows: +
        +for (OperationType c : OperationType.values())
        +    System.out.println(c);
        +
        +
        +
        Returns:
        +
        an array containing the constants of this enum type, in the order they are declared
        +
        +
      • +
      + + + +
        +
      • +

        valueOf

        +
        public static OperationType valueOf​(String name)
        +
        Returns the enum constant of this type with the specified name. +The string must match exactly an identifier used to declare an +enum constant in this type. (Extraneous whitespace characters are +not permitted.)
        +
        +
        Parameters:
        +
        name - the name of the enum constant to be returned.
        +
        Returns:
        +
        the enum constant with the specified name
        +
        Throws:
        +
        IllegalArgumentException - if this enum type has no constant with the specified name
        +
        NullPointerException - if the argument is null
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/RowWithOp.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/RowWithOp.html new file mode 100644 index 0000000000..97e4da77b0 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/RowWithOp.html @@ -0,0 +1,406 @@ + + + + + +RowWithOp (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class RowWithOp

+
+
+ +
+
    +
  • +
    +
    public class RowWithOp
    +extends Object
    +
    A wrapper class that associates an InternalRow with an OperationType for use in + Fluss-Flink data processing. + +

    This class is used to represent a row of data along with its corresponding operation type, + such as APPEND, UPSERT, or DELETE, as defined by OperationType.

    +
    +
    See Also:
    +
    InternalRow, +OperationType
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        RowWithOp

        +
        public RowWithOp​(org.apache.fluss.row.InternalRow row,
        +                 @Nullable
        +                 OperationType opType)
        +
        Constructs a RowWithOp with the specified internal row and operation type.
        +
        +
        Parameters:
        +
        row - the internal row data (must not be null)
        +
        opType - the operation type (must not be null)
        +
        Throws:
        +
        NullPointerException - if row or opType is null
        +
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getRow

        +
        public org.apache.fluss.row.InternalRow getRow()
        +
        Returns the internal row data.
        +
        +
        Returns:
        +
        the internal row
        +
        +
      • +
      + + + +
        +
      • +

        getOperationType

        +
        public OperationType getOperationType()
        +
        Returns the operation type associated with this row.
        +
        +
        Returns:
        +
        the operation type
        +
        +
      • +
      + + + +
        +
      • +

        equals

        +
        public boolean equals​(Object o)
        +
        Indicates whether some other object is "equal to" this one. Two RowWithOp objects are + considered equal if their internal rows and operation types are equal.
        +
        +
        Overrides:
        +
        equals in class Object
        +
        Parameters:
        +
        o - the reference object with which to compare
        +
        Returns:
        +
        true if this object is the same as the obj argument; false otherwise
        +
        +
      • +
      + + + +
        +
      • +

        hashCode

        +
        public int hashCode()
        +
        Returns a hash code value for the object.
        +
        +
        Overrides:
        +
        hashCode in class Object
        +
        Returns:
        +
        a hash code value for this object
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/FlinkAsFlussRow.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/FlinkAsFlussRow.html new file mode 100644 index 0000000000..9720e91a3f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/FlinkAsFlussRow.html @@ -0,0 +1,192 @@ + + + + + +Uses of Class org.apache.fluss.flink.row.FlinkAsFlussRow (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.row.FlinkAsFlussRow

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/OperationType.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/OperationType.html new file mode 100644 index 0000000000..689f0fb241 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/OperationType.html @@ -0,0 +1,225 @@ + + + + + +Uses of Class org.apache.fluss.flink.row.OperationType (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.row.OperationType

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/RowWithOp.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/RowWithOp.html new file mode 100644 index 0000000000..f1f9d30a53 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/RowWithOp.html @@ -0,0 +1,201 @@ + + + + + +Uses of Class org.apache.fluss.flink.row.RowWithOp (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.row.RowWithOp

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-summary.html new file mode 100644 index 0000000000..8c2ec0c37b --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-summary.html @@ -0,0 +1,192 @@ + + + + + +org.apache.fluss.flink.row (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.row

+
+
+
    +
  • + + + + + + + + + + + + + + + + +
    Class Summary 
    ClassDescription
    FlinkAsFlussRow +
    Wraps a Flink RowData as a Fluss InternalRow.
    +
    RowWithOp +
    A wrapper class that associates an InternalRow with an OperationType for use in + Fluss-Flink data processing.
    +
    +
  • +
  • + + + + + + + + + + + + +
    Enum Summary 
    EnumDescription
    OperationType +
    Enumeration of row operation types used in Fluss-Flink data processing.
    +
    +
  • +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-tree.html new file mode 100644 index 0000000000..d4423f8e93 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-tree.html @@ -0,0 +1,178 @@ + + + + + +org.apache.fluss.flink.row Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.row

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+
    +
  • java.lang.Object +
      +
    • org.apache.fluss.flink.row.FlinkAsFlussRow (implements org.apache.fluss.row.InternalRow)
    • +
    • org.apache.fluss.flink.row.RowWithOp
    • +
    +
  • +
+
+
+

Enum Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-use.html new file mode 100644 index 0000000000..5117a575e5 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-use.html @@ -0,0 +1,215 @@ + + + + + +Uses of Package org.apache.fluss.flink.row (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.row

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/ChannelComputer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/ChannelComputer.html new file mode 100644 index 0000000000..9dd9070b73 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/ChannelComputer.html @@ -0,0 +1,315 @@ + + + + + +ChannelComputer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface ChannelComputer<T>

+
+
+
+
    +
  • +
    +
    Type Parameters:
    +
    T - type of record
    +
    +
    +
    All Superinterfaces:
    +
    Serializable
    +
    +
    +
    All Known Implementing Classes:
    +
    FlinkRowDataChannelComputer
    +
    +
    +
    public interface ChannelComputer<T>
    +extends Serializable
    +
    A utility class to compute which downstream channel a given record should be sent to before flink + sink.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        setup

        +
        void setup​(int numChannels)
        +
      • +
      + + + + + +
        +
      • +

        channel

        +
        int channel​(T record)
        +
      • +
      + + + +
        +
      • +

        select

        +
        static int select​(String partitionName,
        +                  int bucket,
        +                  int numChannels)
        +
      • +
      + + + +
        +
      • +

        select

        +
        static int select​(int bucket,
        +                  int numChannels)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkRowDataChannelComputer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkRowDataChannelComputer.html new file mode 100644 index 0000000000..71bf412e89 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkRowDataChannelComputer.html @@ -0,0 +1,369 @@ + + + + + +FlinkRowDataChannelComputer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkRowDataChannelComputer<InputT>

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.sink.FlinkRowDataChannelComputer<InputT>
    • +
    +
  • +
+
+ +
+
+ +
+
+ +
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkStreamPartitioner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkStreamPartitioner.html new file mode 100644 index 0000000000..10bddbc51b --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkStreamPartitioner.html @@ -0,0 +1,454 @@ + + + + + +FlinkStreamPartitioner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkStreamPartitioner<T>

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.flink.streaming.runtime.partitioner.StreamPartitioner<T>
    • +
    • +
        +
      • org.apache.fluss.flink.sink.FlinkStreamPartitioner<T>
      • +
      +
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    Serializable, org.apache.flink.runtime.io.network.api.writer.ChannelSelector<org.apache.flink.runtime.plugable.SerializationDelegate<org.apache.flink.streaming.runtime.streamrecord.StreamRecord<T>>>
    +
    +
    +
    public class FlinkStreamPartitioner<T>
    +extends org.apache.flink.streaming.runtime.partitioner.StreamPartitioner<T>
    +
    A StreamPartitioner which wraps a ChannelComputer.
    +
    +
    See Also:
    +
    Serialized Form
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Field Summary

      +
        +
      • + + +

        Fields inherited from class org.apache.flink.streaming.runtime.partitioner.StreamPartitioner

        +numberOfChannels
      • +
      +
    • +
    +
    + +
    + +
    + +
    +
      +
    • + + +

      Method Summary

      + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
      All Methods Static Methods Instance Methods Concrete Methods 
      Modifier and TypeMethodDescription
      org.apache.flink.streaming.runtime.partitioner.StreamPartitioner<T>copy() 
      org.apache.flink.runtime.io.network.api.writer.SubtaskStateMappergetDownstreamSubtaskStateMapper() 
      booleanisPointwise() 
      static <T> org.apache.flink.streaming.api.datastream.DataStream<T>partition​(org.apache.flink.streaming.api.datastream.DataStream<T> input, + ChannelComputer<T> channelComputer, + Integer parallelism) 
      intselectChannel​(org.apache.flink.runtime.plugable.SerializationDelegate<org.apache.flink.streaming.runtime.streamrecord.StreamRecord<T>> record) 
      voidsetup​(int numberOfChannels) 
      StringtoString() 
      +
        +
      • + + +

        Methods inherited from class org.apache.flink.streaming.runtime.partitioner.StreamPartitioner

        +disableUnalignedCheckpoints, equals, getUpstreamSubtaskStateMapper, hashCode, isBroadcast, isSupportsUnalignedCheckpoint
      • +
      + +
    • +
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlinkStreamPartitioner

        +
        public FlinkStreamPartitioner​(ChannelComputer<T> channelComputer)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        setup

        +
        public void setup​(int numberOfChannels)
        +
        +
        Specified by:
        +
        setup in interface org.apache.flink.runtime.io.network.api.writer.ChannelSelector<T>
        +
        Overrides:
        +
        setup in class org.apache.flink.streaming.runtime.partitioner.StreamPartitioner<T>
        +
        +
      • +
      + + + +
        +
      • +

        copy

        +
        public org.apache.flink.streaming.runtime.partitioner.StreamPartitioner<T> copy()
        +
        +
        Specified by:
        +
        copy in class org.apache.flink.streaming.runtime.partitioner.StreamPartitioner<T>
        +
        +
      • +
      + + + +
        +
      • +

        getDownstreamSubtaskStateMapper

        +
        public org.apache.flink.runtime.io.network.api.writer.SubtaskStateMapper getDownstreamSubtaskStateMapper()
        +
        +
        Specified by:
        +
        getDownstreamSubtaskStateMapper in class org.apache.flink.streaming.runtime.partitioner.StreamPartitioner<T>
        +
        +
      • +
      + + + +
        +
      • +

        isPointwise

        +
        public boolean isPointwise()
        +
        +
        Specified by:
        +
        isPointwise in class org.apache.flink.streaming.runtime.partitioner.StreamPartitioner<T>
        +
        +
      • +
      + + + + + + + +
        +
      • +

        selectChannel

        +
        public int selectChannel​(org.apache.flink.runtime.plugable.SerializationDelegate<org.apache.flink.streaming.runtime.streamrecord.StreamRecord<T>> record)
        +
      • +
      + + + +
        +
      • +

        partition

        +
        public static <T> org.apache.flink.streaming.api.datastream.DataStream<T> partition​(org.apache.flink.streaming.api.datastream.DataStream<T> input,
        +                                                                                    ChannelComputer<T> channelComputer,
        +                                                                                    Integer parallelism)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkTableSink.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkTableSink.html new file mode 100644 index 0000000000..e4aec6207b --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkTableSink.html @@ -0,0 +1,541 @@ + + + + + +FlinkTableSink (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkTableSink

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.table.connector.sink.abilities.SupportsDeletePushDown, org.apache.flink.table.connector.sink.abilities.SupportsPartitioning, org.apache.flink.table.connector.sink.abilities.SupportsRowLevelDelete, org.apache.flink.table.connector.sink.abilities.SupportsRowLevelUpdate, org.apache.flink.table.connector.sink.DynamicTableSink
    +
    +
    +
    public class FlinkTableSink
    +extends Object
    +implements org.apache.flink.table.connector.sink.DynamicTableSink, org.apache.flink.table.connector.sink.abilities.SupportsPartitioning, org.apache.flink.table.connector.sink.abilities.SupportsDeletePushDown, org.apache.flink.table.connector.sink.abilities.SupportsRowLevelDelete, org.apache.flink.table.connector.sink.abilities.SupportsRowLevelUpdate
    +
    A Flink DynamicTableSink.
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Nested Class Summary

      +
        +
      • + + +

        Nested classes/interfaces inherited from interface org.apache.flink.table.connector.sink.DynamicTableSink

        +org.apache.flink.table.connector.sink.DynamicTableSink.Context, org.apache.flink.table.connector.sink.DynamicTableSink.DataStructureConverter, org.apache.flink.table.connector.sink.DynamicTableSink.SinkRuntimeProvider
      • +
      +
        +
      • + + +

        Nested classes/interfaces inherited from interface org.apache.flink.table.connector.sink.abilities.SupportsRowLevelDelete

        +org.apache.flink.table.connector.sink.abilities.SupportsRowLevelDelete.RowLevelDeleteInfo, org.apache.flink.table.connector.sink.abilities.SupportsRowLevelDelete.RowLevelDeleteMode
      • +
      +
        +
      • + + +

        Nested classes/interfaces inherited from interface org.apache.flink.table.connector.sink.abilities.SupportsRowLevelUpdate

        +org.apache.flink.table.connector.sink.abilities.SupportsRowLevelUpdate.RowLevelUpdateInfo, org.apache.flink.table.connector.sink.abilities.SupportsRowLevelUpdate.RowLevelUpdateMode
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Summary

      + + + + + + + + + + +
      Constructors 
      ConstructorDescription
      FlinkTableSink​(org.apache.fluss.metadata.TablePath tablePath, + org.apache.fluss.config.Configuration flussConfig, + org.apache.flink.table.types.logical.RowType tableRowType, + int[] primaryKeyIndexes, + List<String> partitionKeys, + boolean streaming, + org.apache.fluss.metadata.MergeEngineType mergeEngineType, + org.apache.fluss.metadata.DataLakeFormat lakeFormat, + boolean ignoreDelete, + int numBucket, + List<String> bucketKeys, + boolean shuffleByBucketId) 
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlinkTableSink

        +
        public FlinkTableSink​(org.apache.fluss.metadata.TablePath tablePath,
        +                      org.apache.fluss.config.Configuration flussConfig,
        +                      org.apache.flink.table.types.logical.RowType tableRowType,
        +                      int[] primaryKeyIndexes,
        +                      List<String> partitionKeys,
        +                      boolean streaming,
        +                      @Nullable
        +                      org.apache.fluss.metadata.MergeEngineType mergeEngineType,
        +                      @Nullable
        +                      org.apache.fluss.metadata.DataLakeFormat lakeFormat,
        +                      boolean ignoreDelete,
        +                      int numBucket,
        +                      List<String> bucketKeys,
        +                      boolean shuffleByBucketId)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getChangelogMode

        +
        public org.apache.flink.table.connector.ChangelogMode getChangelogMode​(org.apache.flink.table.connector.ChangelogMode requestedMode)
        +
        +
        Specified by:
        +
        getChangelogMode in interface org.apache.flink.table.connector.sink.DynamicTableSink
        +
        +
      • +
      + + + +
        +
      • +

        getSinkRuntimeProvider

        +
        public org.apache.flink.table.connector.sink.DynamicTableSink.SinkRuntimeProvider getSinkRuntimeProvider​(org.apache.flink.table.connector.sink.DynamicTableSink.Context context)
        +
        +
        Specified by:
        +
        getSinkRuntimeProvider in interface org.apache.flink.table.connector.sink.DynamicTableSink
        +
        +
      • +
      + + + +
        +
      • +

        copy

        +
        public org.apache.flink.table.connector.sink.DynamicTableSink copy()
        +
        +
        Specified by:
        +
        copy in interface org.apache.flink.table.connector.sink.DynamicTableSink
        +
        +
      • +
      + + + +
        +
      • +

        asSummaryString

        +
        public String asSummaryString()
        +
        +
        Specified by:
        +
        asSummaryString in interface org.apache.flink.table.connector.sink.DynamicTableSink
        +
        +
      • +
      + + + +
        +
      • +

        applyStaticPartition

        +
        public void applyStaticPartition​(Map<String,​String> partition)
        +
        +
        Specified by:
        +
        applyStaticPartition in interface org.apache.flink.table.connector.sink.abilities.SupportsPartitioning
        +
        +
      • +
      + + + +
        +
      • +

        applyDeleteFilters

        +
        public boolean applyDeleteFilters​(List<org.apache.flink.table.expressions.ResolvedExpression> filters)
        +
        +
        Specified by:
        +
        applyDeleteFilters in interface org.apache.flink.table.connector.sink.abilities.SupportsDeletePushDown
        +
        +
      • +
      + + + +
        +
      • +

        executeDeletion

        +
        public Optional<Long> executeDeletion()
        +
        +
        Specified by:
        +
        executeDeletion in interface org.apache.flink.table.connector.sink.abilities.SupportsDeletePushDown
        +
        +
      • +
      + + + +
        +
      • +

        applyRowLevelDelete

        +
        public org.apache.flink.table.connector.sink.abilities.SupportsRowLevelDelete.RowLevelDeleteInfo applyRowLevelDelete​(@Nullable
        +                                                                                                                     org.apache.flink.table.connector.RowLevelModificationScanContext rowLevelModificationScanContext)
        +
        +
        Specified by:
        +
        applyRowLevelDelete in interface org.apache.flink.table.connector.sink.abilities.SupportsRowLevelDelete
        +
        +
      • +
      + + + +
        +
      • +

        applyRowLevelUpdate

        +
        public org.apache.flink.table.connector.sink.abilities.SupportsRowLevelUpdate.RowLevelUpdateInfo applyRowLevelUpdate​(List<org.apache.flink.table.catalog.Column> updatedColumns,
        +                                                                                                                     @Nullable
        +                                                                                                                     org.apache.flink.table.connector.RowLevelModificationScanContext rowLevelModificationScanContext)
        +
        +
        Specified by:
        +
        applyRowLevelUpdate in interface org.apache.flink.table.connector.sink.abilities.SupportsRowLevelUpdate
        +
        +
      • +
      + + + +
        +
      • +

        getBucketKeys

        +
        @VisibleForTesting
        +public List<String> getBucketKeys()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlussSink.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlussSink.html new file mode 100644 index 0000000000..fc87dbd901 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlussSink.html @@ -0,0 +1,376 @@ + + + + + +FlussSink (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlussSink<InputT>

+
+
+ +
+
    +
  • +
    +
    Type Parameters:
    +
    InputT - the type of input elements accepted by the sink
    +
    +
    +
    All Implemented Interfaces:
    +
    Serializable, org.apache.flink.api.connector.sink2.Sink<InputT>, org.apache.flink.streaming.api.connector.sink2.SupportsPreWriteTopology<InputT>
    +
    +
    +
    @PublicEvolving
    +public class FlussSink<InputT>
    +extends Object
    +
    FlussSink is a specialized Flink sink for writing data to Fluss. + +

    This class extends FlinkSink and provides a builder for constructing Fluss sink + instances with custom configurations. It is intended to be used as the main entry point for + integrating Fluss as a sink in Flink data pipelines.

    +
    +
    Since:
    +
    0.7
    +
    See Also:
    +
    Serialized Form
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Nested Class Summary

      +
        +
      • + + +

        Nested classes/interfaces inherited from interface org.apache.flink.api.connector.sink2.Sink

        +org.apache.flink.api.connector.sink2.Sink.InitContext, org.apache.flink.api.connector.sink2.Sink.InitContextWrapper
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        builder

        +
        public static <T> FlussSinkBuilder<T> builder()
        +
        Creates a new FlussSinkBuilder instance for building a FlussSink.
        +
        +
        Type Parameters:
        +
        T - the type of input elements
        +
        Returns:
        +
        a new FlussSinkBuilder instance
        +
        +
      • +
      + + + +
        +
      • +

        createWriter

        +
        @Deprecated
        +public org.apache.flink.api.connector.sink2.SinkWriter<InputT> createWriter​(org.apache.flink.api.connector.sink2.Sink.InitContext context)
        +                                                                     throws IOException
        +
        Deprecated.
        +
        +
        Specified by:
        +
        createWriter in interface org.apache.flink.api.connector.sink2.Sink<InputT>
        +
        Throws:
        +
        IOException
        +
        +
      • +
      + + + +
        +
      • +

        createWriter

        +
        public org.apache.flink.api.connector.sink2.SinkWriter<InputT> createWriter​(org.apache.flink.api.connector.sink2.WriterInitContext context)
        +                                                                     throws IOException
        +
        +
        Specified by:
        +
        createWriter in interface org.apache.flink.api.connector.sink2.Sink<InputT>
        +
        Throws:
        +
        IOException
        +
        +
      • +
      + + + +
        +
      • +

        addPreWriteTopology

        +
        public org.apache.flink.streaming.api.datastream.DataStream<InputT> addPreWriteTopology​(org.apache.flink.streaming.api.datastream.DataStream<InputT> input)
        +
        +
        Specified by:
        +
        addPreWriteTopology in interface org.apache.flink.streaming.api.connector.sink2.SupportsPreWriteTopology<InputT>
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlussSinkBuilder.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlussSinkBuilder.html new file mode 100644 index 0000000000..d282dbba89 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlussSinkBuilder.html @@ -0,0 +1,455 @@ + + + + + +FlussSinkBuilder (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlussSinkBuilder<InputT>

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.sink.FlussSinkBuilder<InputT>
    • +
    +
  • +
+
+
    +
  • +
    +
    Type Parameters:
    +
    InputT - > The input type of records to be written to Fluss
    +
    +
    +
    @PublicEvolving
    +public class FlussSinkBuilder<InputT>
    +extends Object
    +
    Builder for creating and configuring Fluss sink connectors for Apache Flink. + +

    The builder supports automatic schema inference from POJO classes using reflection and + provides options for customizing data conversion logic through custom converters. + +

    Example usage: + +

    
    + FlinkSink<Order> sink = new FlussSinkBuilder<Order>()
    +          .setBootstrapServers(bootstrapServers)
    +          .setTable(tableName)
    +          .setDatabase(databaseName)
    +          .setRowType(orderRowType)
    +          .setSerializationSchema(new OrderSerializationSchema())
    +          .build())
    + 
    +
    +
    Since:
    +
    0.7
    +
    +
  • +
+
+
+ +
+
+ +
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/ChannelComputer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/ChannelComputer.html new file mode 100644 index 0000000000..ed7448d8d5 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/ChannelComputer.html @@ -0,0 +1,224 @@ + + + + + +Uses of Interface org.apache.fluss.flink.sink.ChannelComputer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.flink.sink.ChannelComputer

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkRowDataChannelComputer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkRowDataChannelComputer.html new file mode 100644 index 0000000000..b900c2955e --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkRowDataChannelComputer.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.sink.FlinkRowDataChannelComputer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.sink.FlinkRowDataChannelComputer

+
+
No usage of org.apache.fluss.flink.sink.FlinkRowDataChannelComputer
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkStreamPartitioner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkStreamPartitioner.html new file mode 100644 index 0000000000..dc77de29c7 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkStreamPartitioner.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.sink.FlinkStreamPartitioner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.sink.FlinkStreamPartitioner

+
+
No usage of org.apache.fluss.flink.sink.FlinkStreamPartitioner
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkTableSink.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkTableSink.html new file mode 100644 index 0000000000..604643d4b5 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkTableSink.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.sink.FlinkTableSink (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.sink.FlinkTableSink

+
+
No usage of org.apache.fluss.flink.sink.FlinkTableSink
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlussSink.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlussSink.html new file mode 100644 index 0000000000..7f59abc3c6 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlussSink.html @@ -0,0 +1,194 @@ + + + + + +Uses of Class org.apache.fluss.flink.sink.FlussSink (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.sink.FlussSink

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlussSinkBuilder.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlussSinkBuilder.html new file mode 100644 index 0000000000..e00ce2f85d --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlussSinkBuilder.html @@ -0,0 +1,244 @@ + + + + + +Uses of Class org.apache.fluss.flink.sink.FlussSinkBuilder (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.sink.FlussSinkBuilder

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-summary.html new file mode 100644 index 0000000000..8b207fc299 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-summary.html @@ -0,0 +1,210 @@ + + + + + +org.apache.fluss.flink.sink (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.sink

+
+
+
    +
  • + + + + + + + + + + + + +
    Interface Summary 
    InterfaceDescription
    ChannelComputer<T> +
    A utility class to compute which downstream channel a given record should be sent to before flink + sink.
    +
    +
  • +
  • + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Class Summary 
    ClassDescription
    FlinkRowDataChannelComputer<InputT> +
    ChannelComputer for flink RowData.
    +
    FlinkStreamPartitioner<T> +
    A StreamPartitioner which wraps a ChannelComputer.
    +
    FlinkTableSink +
    A Flink DynamicTableSink.
    +
    FlussSink<InputT> +
    FlussSink is a specialized Flink sink for writing data to Fluss.
    +
    FlussSinkBuilder<InputT> +
    Builder for creating and configuring Fluss sink connectors for Apache Flink.
    +
    +
  • +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-tree.html new file mode 100644 index 0000000000..2902cda3a2 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-tree.html @@ -0,0 +1,181 @@ + + + + + +org.apache.fluss.flink.sink Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.sink

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+
    +
  • java.lang.Object +
      +
    • org.apache.fluss.flink.sink.FlinkRowDataChannelComputer<InputT> (implements org.apache.fluss.flink.sink.ChannelComputer<T>)
    • +
    • org.apache.fluss.flink.sink.FlinkTableSink (implements org.apache.flink.table.connector.sink.DynamicTableSink, org.apache.flink.table.connector.sink.abilities.SupportsDeletePushDown, org.apache.flink.table.connector.sink.abilities.SupportsPartitioning, org.apache.flink.table.connector.sink.abilities.SupportsRowLevelDelete, org.apache.flink.table.connector.sink.abilities.SupportsRowLevelUpdate)
    • +
    • org.apache.fluss.flink.sink.FlussSink<InputT>
    • +
    • org.apache.fluss.flink.sink.FlussSinkBuilder<InputT>
    • +
    • org.apache.flink.streaming.runtime.partitioner.StreamPartitioner<T> (implements org.apache.flink.runtime.io.network.api.writer.ChannelSelector<T>, java.io.Serializable) + +
    • +
    +
  • +
+
+
+

Interface Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-use.html new file mode 100644 index 0000000000..a2d6f78228 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-use.html @@ -0,0 +1,198 @@ + + + + + +Uses of Package org.apache.fluss.flink.sink (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.sink

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/FlussSerializationSchema.InitializationContext.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/FlussSerializationSchema.InitializationContext.html new file mode 100644 index 0000000000..baa1d71267 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/FlussSerializationSchema.InitializationContext.html @@ -0,0 +1,272 @@ + + + + + +FlussSerializationSchema.InitializationContext (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface FlussSerializationSchema.InitializationContext

+
+
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getRowSchema

        +
        org.apache.fluss.types.RowType getRowSchema()
        +
        Returns the target row schema.
        +
        +
        Returns:
        +
        The schema of the target row.
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/FlussSerializationSchema.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/FlussSerializationSchema.html new file mode 100644 index 0000000000..3521e0ddc4 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/FlussSerializationSchema.html @@ -0,0 +1,331 @@ + + + + + +FlussSerializationSchema (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface FlussSerializationSchema<T>

+
+
+
+
    +
  • +
    +
    Type Parameters:
    +
    T - The type to be serialized.
    +
    +
    +
    All Superinterfaces:
    +
    Serializable
    +
    +
    +
    All Known Implementing Classes:
    +
    RowDataSerializationSchema
    +
    +
    +
    @PublicEvolving
    +public interface FlussSerializationSchema<T>
    +extends Serializable
    +
    A serialization schema for Fluss.
    +
  • +
+
+
+ +
+
+ +
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/RowDataSerializationSchema.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/RowDataSerializationSchema.html new file mode 100644 index 0000000000..536429d799 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/RowDataSerializationSchema.html @@ -0,0 +1,384 @@ + + + + + +RowDataSerializationSchema (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class RowDataSerializationSchema

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.sink.serializer.RowDataSerializationSchema
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        RowDataSerializationSchema

        +
        public RowDataSerializationSchema​(boolean isAppendOnly,
        +                                  boolean ignoreDelete)
        +
        Constructs a new RowSerializationSchema.
        +
        +
        Parameters:
        +
        isAppendOnly - whether the schema is append-only (only INSERTs allowed)
        +
        ignoreDelete - whether to ignore DELETE and UPDATE_BEFORE operations
        +
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/SerializerInitContextImpl.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/SerializerInitContextImpl.html new file mode 100644 index 0000000000..423042074a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/SerializerInitContextImpl.html @@ -0,0 +1,322 @@ + + + + + +SerializerInitContextImpl (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class SerializerInitContextImpl

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.sink.serializer.SerializerInitContextImpl
    • +
    +
  • +
+
+ +
+
+ +
+
+ +
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/FlussSerializationSchema.InitializationContext.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/FlussSerializationSchema.InitializationContext.html new file mode 100644 index 0000000000..bd2e20a02c --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/FlussSerializationSchema.InitializationContext.html @@ -0,0 +1,218 @@ + + + + + +Uses of Interface org.apache.fluss.flink.sink.serializer.FlussSerializationSchema.InitializationContext (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.flink.sink.serializer.FlussSerializationSchema.InitializationContext

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/FlussSerializationSchema.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/FlussSerializationSchema.html new file mode 100644 index 0000000000..8c2adb196f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/FlussSerializationSchema.html @@ -0,0 +1,294 @@ + + + + + +Uses of Interface org.apache.fluss.flink.sink.serializer.FlussSerializationSchema (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.flink.sink.serializer.FlussSerializationSchema

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/RowDataSerializationSchema.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/RowDataSerializationSchema.html new file mode 100644 index 0000000000..860f84473a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/RowDataSerializationSchema.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.sink.serializer.RowDataSerializationSchema (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.sink.serializer.RowDataSerializationSchema

+
+
No usage of org.apache.fluss.flink.sink.serializer.RowDataSerializationSchema
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/SerializerInitContextImpl.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/SerializerInitContextImpl.html new file mode 100644 index 0000000000..669248f2bc --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/SerializerInitContextImpl.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.sink.serializer.SerializerInitContextImpl (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.sink.serializer.SerializerInitContextImpl

+
+
No usage of org.apache.fluss.flink.sink.serializer.SerializerInitContextImpl
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-summary.html new file mode 100644 index 0000000000..4c86309875 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-summary.html @@ -0,0 +1,197 @@ + + + + + +org.apache.fluss.flink.sink.serializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.sink.serializer

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-tree.html new file mode 100644 index 0000000000..cc4e06974d --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-tree.html @@ -0,0 +1,175 @@ + + + + + +org.apache.fluss.flink.sink.serializer Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.sink.serializer

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+

Interface Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-use.html new file mode 100644 index 0000000000..3d62d01fa2 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-use.html @@ -0,0 +1,237 @@ + + + + + +Uses of Package org.apache.fluss.flink.sink.serializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.sink.serializer

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/AppendSinkWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/AppendSinkWriter.html new file mode 100644 index 0000000000..f731236f7d --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/AppendSinkWriter.html @@ -0,0 +1,399 @@ + + + + + +AppendSinkWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class AppendSinkWriter<InputT>

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    AutoCloseable, org.apache.flink.api.connector.sink2.SinkWriter<InputT>
    +
    +
    +
    public class AppendSinkWriter<InputT>
    +extends FlinkSinkWriter<InputT>
    +
    An append only sink writer for fluss log table.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        AppendSinkWriter

        +
        public AppendSinkWriter​(org.apache.fluss.metadata.TablePath tablePath,
        +                        org.apache.fluss.config.Configuration flussConfig,
        +                        org.apache.flink.table.types.logical.RowType tableRowType,
        +                        org.apache.flink.api.common.operators.MailboxExecutor mailboxExecutor,
        +                        FlussSerializationSchema<InputT> serializationSchema)
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/FlinkSinkWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/FlinkSinkWriter.html new file mode 100644 index 0000000000..b41947e562 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/FlinkSinkWriter.html @@ -0,0 +1,593 @@ + + + + + +FlinkSinkWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkSinkWriter<InputT>

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.sink.writer.FlinkSinkWriter<InputT>
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    AutoCloseable, org.apache.flink.api.connector.sink2.SinkWriter<InputT>
    +
    +
    +
    Direct Known Subclasses:
    +
    AppendSinkWriter, UpsertSinkWriter
    +
    +
    +
    public abstract class FlinkSinkWriter<InputT>
    +extends Object
    +implements org.apache.flink.api.connector.sink2.SinkWriter<InputT>
    +
    Base class for Flink SinkWriter implementations in Fluss.
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Nested Class Summary

      +
        +
      • + + +

        Nested classes/interfaces inherited from interface org.apache.flink.api.connector.sink2.SinkWriter

        +org.apache.flink.api.connector.sink2.SinkWriter.Context
      • +
      +
    • +
    +
    + +
    + +
    + +
    +
      +
    • + + +

      Constructor Summary

      + + + + + + + + + + + + + + +
      Constructors 
      ConstructorDescription
      FlinkSinkWriter​(org.apache.fluss.metadata.TablePath tablePath, + org.apache.fluss.config.Configuration flussConfig, + org.apache.flink.table.types.logical.RowType tableRowType, + int[] targetColumns, + org.apache.flink.api.common.operators.MailboxExecutor mailboxExecutor, + FlussSerializationSchema<InputT> serializationSchema) 
      FlinkSinkWriter​(org.apache.fluss.metadata.TablePath tablePath, + org.apache.fluss.config.Configuration flussConfig, + org.apache.flink.table.types.logical.RowType tableRowType, + org.apache.flink.api.common.operators.MailboxExecutor mailboxExecutor, + FlussSerializationSchema<InputT> serializationSchema) 
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + +
        +
      • +

        LOG

        +
        protected static final org.slf4j.Logger LOG
        +
      • +
      + + + +
        +
      • +

        tableRowType

        +
        protected final org.apache.flink.table.types.logical.RowType tableRowType
        +
      • +
      + + + +
        +
      • +

        targetColumnIndexes

        +
        @Nullable
        +protected final int[] targetColumnIndexes
        +
      • +
      + + + +
        +
      • +

        table

        +
        protected transient Table table
        +
      • +
      + + + + + + + +
        +
      • +

        metricGroup

        +
        protected transient org.apache.flink.metrics.groups.SinkWriterMetricGroup metricGroup
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlinkSinkWriter

        +
        public FlinkSinkWriter​(org.apache.fluss.metadata.TablePath tablePath,
        +                       org.apache.fluss.config.Configuration flussConfig,
        +                       org.apache.flink.table.types.logical.RowType tableRowType,
        +                       org.apache.flink.api.common.operators.MailboxExecutor mailboxExecutor,
        +                       FlussSerializationSchema<InputT> serializationSchema)
        +
      • +
      + + + +
        +
      • +

        FlinkSinkWriter

        +
        public FlinkSinkWriter​(org.apache.fluss.metadata.TablePath tablePath,
        +                       org.apache.fluss.config.Configuration flussConfig,
        +                       org.apache.flink.table.types.logical.RowType tableRowType,
        +                       @Nullable
        +                       int[] targetColumns,
        +                       org.apache.flink.api.common.operators.MailboxExecutor mailboxExecutor,
        +                       FlussSerializationSchema<InputT> serializationSchema)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        initialize

        +
        public void initialize​(org.apache.flink.metrics.groups.SinkWriterMetricGroup metricGroup)
        +
      • +
      + + + +
        +
      • +

        initMetrics

        +
        protected void initMetrics()
        +
      • +
      + + + + + + + + + + + + + + + + + +
        +
      • +

        checkAsyncException

        +
        protected void checkAsyncException()
        +                            throws IOException
        +
        This method should only be invoked in the mailbox thread since the counter is not volatile. + Logic needs to be invoked by write AND flush since we support various semantics.
        +
        +
        Throws:
        +
        IOException
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/UpsertSinkWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/UpsertSinkWriter.html new file mode 100644 index 0000000000..c543826c1e --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/UpsertSinkWriter.html @@ -0,0 +1,404 @@ + + + + + +UpsertSinkWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class UpsertSinkWriter<InputT>

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    AutoCloseable, org.apache.flink.api.connector.sink2.SinkWriter<InputT>
    +
    +
    +
    public class UpsertSinkWriter<InputT>
    +extends FlinkSinkWriter<InputT>
    +
    An upsert sink writer or fluss primary key table.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        UpsertSinkWriter

        +
        public UpsertSinkWriter​(org.apache.fluss.metadata.TablePath tablePath,
        +                        org.apache.fluss.config.Configuration flussConfig,
        +                        org.apache.flink.table.types.logical.RowType tableRowType,
        +                        @Nullable
        +                        int[] targetColumnIndexes,
        +                        org.apache.flink.api.common.operators.MailboxExecutor mailboxExecutor,
        +                        FlussSerializationSchema<InputT> flussSerializationSchema)
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/AppendSinkWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/AppendSinkWriter.html new file mode 100644 index 0000000000..aa1be3168e --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/AppendSinkWriter.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.sink.writer.AppendSinkWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.sink.writer.AppendSinkWriter

+
+
No usage of org.apache.fluss.flink.sink.writer.AppendSinkWriter
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/FlinkSinkWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/FlinkSinkWriter.html new file mode 100644 index 0000000000..bb7e8cb4d2 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/FlinkSinkWriter.html @@ -0,0 +1,201 @@ + + + + + +Uses of Class org.apache.fluss.flink.sink.writer.FlinkSinkWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.sink.writer.FlinkSinkWriter

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/UpsertSinkWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/UpsertSinkWriter.html new file mode 100644 index 0000000000..eaa562a56d --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/UpsertSinkWriter.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.sink.writer.UpsertSinkWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.sink.writer.UpsertSinkWriter

+
+
No usage of org.apache.fluss.flink.sink.writer.UpsertSinkWriter
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-summary.html new file mode 100644 index 0000000000..dd95f239c9 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-summary.html @@ -0,0 +1,180 @@ + + + + + +org.apache.fluss.flink.sink.writer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.sink.writer

+
+
+
    +
  • + + + + + + + + + + + + + + + + + + + + +
    Class Summary 
    ClassDescription
    AppendSinkWriter<InputT> +
    An append only sink writer for fluss log table.
    +
    FlinkSinkWriter<InputT> +
    Base class for Flink SinkWriter implementations in Fluss.
    +
    UpsertSinkWriter<InputT> +
    An upsert sink writer or fluss primary key table.
    +
    +
  • +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-tree.html new file mode 100644 index 0000000000..a50e0445c0 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-tree.html @@ -0,0 +1,168 @@ + + + + + +org.apache.fluss.flink.sink.writer Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.sink.writer

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+
    +
  • java.lang.Object +
      +
    • org.apache.fluss.flink.sink.writer.FlinkSinkWriter<InputT> (implements org.apache.flink.api.connector.sink2.SinkWriter<InputT>) + +
    • +
    +
  • +
+
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-use.html new file mode 100644 index 0000000000..25d1f56f15 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-use.html @@ -0,0 +1,185 @@ + + + + + +Uses of Package org.apache.fluss.flink.sink.writer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.sink.writer

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlinkSource.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlinkSource.html new file mode 100644 index 0000000000..4aef171c5e --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlinkSource.html @@ -0,0 +1,550 @@ + + + + + +FlinkSource (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkSource<OUT>

+
+
+ +
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + + + + + +
        +
      • +

        scanPartitionDiscoveryIntervalMs

        +
        protected final long scanPartitionDiscoveryIntervalMs
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlinkSource

        +
        public FlinkSource​(org.apache.fluss.config.Configuration flussConf,
        +                   org.apache.fluss.metadata.TablePath tablePath,
        +                   boolean hasPrimaryKey,
        +                   boolean isPartitioned,
        +                   org.apache.fluss.types.RowType sourceOutputType,
        +                   @Nullable
        +                   int[] projectedFields,
        +                   OffsetsInitializer offsetsInitializer,
        +                   long scanPartitionDiscoveryIntervalMs,
        +                   FlussDeserializationSchema<OUT> deserializationSchema,
        +                   boolean streaming,
        +                   List<PushdownUtils.FieldEqual> partitionFilters)
        +
      • +
      + + + +
        +
      • +

        FlinkSource

        +
        public FlinkSource​(org.apache.fluss.config.Configuration flussConf,
        +                   org.apache.fluss.metadata.TablePath tablePath,
        +                   boolean hasPrimaryKey,
        +                   boolean isPartitioned,
        +                   org.apache.fluss.types.RowType sourceOutputType,
        +                   @Nullable
        +                   int[] projectedFields,
        +                   OffsetsInitializer offsetsInitializer,
        +                   long scanPartitionDiscoveryIntervalMs,
        +                   FlussDeserializationSchema<OUT> deserializationSchema,
        +                   boolean streaming,
        +                   List<PushdownUtils.FieldEqual> partitionFilters,
        +                   org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getBoundedness

        +
        public org.apache.flink.api.connector.source.Boundedness getBoundedness()
        +
        +
        Specified by:
        +
        getBoundedness in interface org.apache.flink.api.connector.source.Source<OUT,​SourceSplitBase,​SourceEnumeratorState>
        +
        +
      • +
      + + + + + + + + + + + +
        +
      • +

        getSplitSerializer

        +
        public org.apache.flink.core.io.SimpleVersionedSerializer<SourceSplitBase> getSplitSerializer()
        +
        +
        Specified by:
        +
        getSplitSerializer in interface org.apache.flink.api.connector.source.Source<OUT,​SourceSplitBase,​SourceEnumeratorState>
        +
        +
      • +
      + + + +
        +
      • +

        getEnumeratorCheckpointSerializer

        +
        public org.apache.flink.core.io.SimpleVersionedSerializer<SourceEnumeratorState> getEnumeratorCheckpointSerializer()
        +
        +
        Specified by:
        +
        getEnumeratorCheckpointSerializer in interface org.apache.flink.api.connector.source.Source<OUT,​SourceSplitBase,​SourceEnumeratorState>
        +
        +
      • +
      + + + +
        +
      • +

        createReader

        +
        public org.apache.flink.api.connector.source.SourceReader<OUT,​SourceSplitBase> createReader​(org.apache.flink.api.connector.source.SourceReaderContext context)
        +                                                                                           throws Exception
        +
        +
        Specified by:
        +
        createReader in interface org.apache.flink.api.connector.source.SourceReaderFactory<OUT,​SourceSplitBase>
        +
        Throws:
        +
        Exception
        +
        +
      • +
      + + + +
        +
      • +

        getProducedType

        +
        public org.apache.flink.api.common.typeinfo.TypeInformation<OUT> getProducedType()
        +
        +
        Specified by:
        +
        getProducedType in interface org.apache.flink.api.java.typeutils.ResultTypeQueryable<OUT>
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlinkTableSource.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlinkTableSource.html new file mode 100644 index 0000000000..fd3bd18574 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlinkTableSource.html @@ -0,0 +1,704 @@ + + + + + +FlinkTableSource (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkTableSource

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.table.connector.source.abilities.SupportsAggregatePushDown, org.apache.flink.table.connector.source.abilities.SupportsFilterPushDown, org.apache.flink.table.connector.source.abilities.SupportsLimitPushDown, org.apache.flink.table.connector.source.abilities.SupportsProjectionPushDown, org.apache.flink.table.connector.source.abilities.SupportsRowLevelModificationScan, org.apache.flink.table.connector.source.DynamicTableSource, org.apache.flink.table.connector.source.LookupTableSource, org.apache.flink.table.connector.source.ScanTableSource
    +
    +
    +
    public class FlinkTableSource
    +extends Object
    +implements org.apache.flink.table.connector.source.ScanTableSource, org.apache.flink.table.connector.source.abilities.SupportsProjectionPushDown, org.apache.flink.table.connector.source.abilities.SupportsFilterPushDown, org.apache.flink.table.connector.source.LookupTableSource, org.apache.flink.table.connector.source.abilities.SupportsRowLevelModificationScan, org.apache.flink.table.connector.source.abilities.SupportsLimitPushDown, org.apache.flink.table.connector.source.abilities.SupportsAggregatePushDown
    +
    Flink table source to scan Fluss data.
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Nested Class Summary

      +
        +
      • + + +

        Nested classes/interfaces inherited from interface org.apache.flink.table.connector.source.DynamicTableSource

        +org.apache.flink.table.connector.source.DynamicTableSource.Context, org.apache.flink.table.connector.source.DynamicTableSource.DataStructureConverter
      • +
      +
        +
      • + + +

        Nested classes/interfaces inherited from interface org.apache.flink.table.connector.source.LookupTableSource

        +org.apache.flink.table.connector.source.LookupTableSource.LookupContext, org.apache.flink.table.connector.source.LookupTableSource.LookupRuntimeProvider
      • +
      +
        +
      • + + +

        Nested classes/interfaces inherited from interface org.apache.flink.table.connector.source.ScanTableSource

        +org.apache.flink.table.connector.source.ScanTableSource.ScanContext, org.apache.flink.table.connector.source.ScanTableSource.ScanRuntimeProvider
      • +
      +
        +
      • + + +

        Nested classes/interfaces inherited from interface org.apache.flink.table.connector.source.abilities.SupportsFilterPushDown

        +org.apache.flink.table.connector.source.abilities.SupportsFilterPushDown.Result
      • +
      +
        +
      • + + +

        Nested classes/interfaces inherited from interface org.apache.flink.table.connector.source.abilities.SupportsRowLevelModificationScan

        +org.apache.flink.table.connector.source.abilities.SupportsRowLevelModificationScan.RowLevelModificationType
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Field Summary

      + + + + + + + + + + + + + + + + + +
      Fields 
      Modifier and TypeFieldDescription
      static org.slf4j.LoggerLOG 
      protected booleanselectRowCount 
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Summary

      + + + + + + + + + + +
      Constructors 
      ConstructorDescription
      FlinkTableSource​(org.apache.fluss.metadata.TablePath tablePath, + org.apache.fluss.config.Configuration flussConfig, + org.apache.flink.table.types.logical.RowType tableOutputType, + int[] primaryKeyIndexes, + int[] bucketKeyIndexes, + int[] partitionKeyIndexes, + boolean streaming, + FlinkConnectorOptionsUtils.StartupOptions startupOptions, + int lookupMaxRetryTimes, + boolean lookupAsync, + org.apache.flink.table.connector.source.lookup.cache.LookupCache cache, + long scanPartitionDiscoveryIntervalMs, + boolean isDataLakeEnabled, + org.apache.fluss.metadata.MergeEngineType mergeEngineType, + Map<String,​String> tableOptions) 
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Summary

      + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
      All Methods Instance Methods Concrete Methods 
      Modifier and TypeMethodDescription
      booleanapplyAggregates​(List<int[]> groupingSets, + List<org.apache.flink.table.expressions.AggregateExpression> aggregateExpressions, + org.apache.flink.table.types.DataType dataType) 
      org.apache.flink.table.connector.source.abilities.SupportsFilterPushDown.ResultapplyFilters​(List<org.apache.flink.table.expressions.ResolvedExpression> filters) 
      voidapplyLimit​(long limit) 
      voidapplyProjection​(int[][] projectedFields, + org.apache.flink.table.types.DataType producedDataType) 
      org.apache.flink.table.connector.RowLevelModificationScanContextapplyRowLevelModificationScan​(org.apache.flink.table.connector.source.abilities.SupportsRowLevelModificationScan.RowLevelModificationType rowLevelModificationType, + org.apache.flink.table.connector.RowLevelModificationScanContext rowLevelModificationScanContext) 
      StringasSummaryString() 
      org.apache.flink.table.connector.source.DynamicTableSourcecopy() 
      int[]getBucketKeyIndexes() 
      org.apache.flink.table.connector.source.lookup.cache.LookupCachegetCache() 
      org.apache.flink.table.connector.ChangelogModegetChangelogMode() 
      org.apache.flink.table.connector.source.LookupTableSource.LookupRuntimeProvidergetLookupRuntimeProvider​(org.apache.flink.table.connector.source.LookupTableSource.LookupContext context) 
      int[]getPartitionKeyIndexes() 
      int[]getPrimaryKeyIndexes() 
      org.apache.flink.table.connector.source.ScanTableSource.ScanRuntimeProvidergetScanRuntimeProvider​(org.apache.flink.table.connector.source.ScanTableSource.ScanContext scanContext) 
      booleansupportsNestedProjection() 
      + +
        +
      • + + +

        Methods inherited from interface org.apache.flink.table.connector.source.abilities.SupportsProjectionPushDown

        +applyProjection
      • +
      +
    • +
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + +
        +
      • +

        LOG

        +
        public static final org.slf4j.Logger LOG
        +
      • +
      + + + +
        +
      • +

        selectRowCount

        +
        protected boolean selectRowCount
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlinkTableSource

        +
        public FlinkTableSource​(org.apache.fluss.metadata.TablePath tablePath,
        +                        org.apache.fluss.config.Configuration flussConfig,
        +                        org.apache.flink.table.types.logical.RowType tableOutputType,
        +                        int[] primaryKeyIndexes,
        +                        int[] bucketKeyIndexes,
        +                        int[] partitionKeyIndexes,
        +                        boolean streaming,
        +                        FlinkConnectorOptionsUtils.StartupOptions startupOptions,
        +                        int lookupMaxRetryTimes,
        +                        boolean lookupAsync,
        +                        @Nullable
        +                        org.apache.flink.table.connector.source.lookup.cache.LookupCache cache,
        +                        long scanPartitionDiscoveryIntervalMs,
        +                        boolean isDataLakeEnabled,
        +                        @Nullable
        +                        org.apache.fluss.metadata.MergeEngineType mergeEngineType,
        +                        Map<String,​String> tableOptions)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getChangelogMode

        +
        public org.apache.flink.table.connector.ChangelogMode getChangelogMode()
        +
        +
        Specified by:
        +
        getChangelogMode in interface org.apache.flink.table.connector.source.ScanTableSource
        +
        +
      • +
      + + + +
        +
      • +

        getScanRuntimeProvider

        +
        public org.apache.flink.table.connector.source.ScanTableSource.ScanRuntimeProvider getScanRuntimeProvider​(org.apache.flink.table.connector.source.ScanTableSource.ScanContext scanContext)
        +
        +
        Specified by:
        +
        getScanRuntimeProvider in interface org.apache.flink.table.connector.source.ScanTableSource
        +
        +
      • +
      + + + +
        +
      • +

        getLookupRuntimeProvider

        +
        public org.apache.flink.table.connector.source.LookupTableSource.LookupRuntimeProvider getLookupRuntimeProvider​(org.apache.flink.table.connector.source.LookupTableSource.LookupContext context)
        +
        +
        Specified by:
        +
        getLookupRuntimeProvider in interface org.apache.flink.table.connector.source.LookupTableSource
        +
        +
      • +
      + + + +
        +
      • +

        copy

        +
        public org.apache.flink.table.connector.source.DynamicTableSource copy()
        +
        +
        Specified by:
        +
        copy in interface org.apache.flink.table.connector.source.DynamicTableSource
        +
        +
      • +
      + + + +
        +
      • +

        asSummaryString

        +
        public String asSummaryString()
        +
        +
        Specified by:
        +
        asSummaryString in interface org.apache.flink.table.connector.source.DynamicTableSource
        +
        +
      • +
      + + + +
        +
      • +

        supportsNestedProjection

        +
        public boolean supportsNestedProjection()
        +
        +
        Specified by:
        +
        supportsNestedProjection in interface org.apache.flink.table.connector.source.abilities.SupportsProjectionPushDown
        +
        +
      • +
      + + + +
        +
      • +

        applyProjection

        +
        public void applyProjection​(int[][] projectedFields,
        +                            org.apache.flink.table.types.DataType producedDataType)
        +
        +
        Specified by:
        +
        applyProjection in interface org.apache.flink.table.connector.source.abilities.SupportsProjectionPushDown
        +
        +
      • +
      + + + +
        +
      • +

        applyFilters

        +
        public org.apache.flink.table.connector.source.abilities.SupportsFilterPushDown.Result applyFilters​(List<org.apache.flink.table.expressions.ResolvedExpression> filters)
        +
        +
        Specified by:
        +
        applyFilters in interface org.apache.flink.table.connector.source.abilities.SupportsFilterPushDown
        +
        +
      • +
      + + + +
        +
      • +

        applyRowLevelModificationScan

        +
        public org.apache.flink.table.connector.RowLevelModificationScanContext applyRowLevelModificationScan​(org.apache.flink.table.connector.source.abilities.SupportsRowLevelModificationScan.RowLevelModificationType rowLevelModificationType,
        +                                                                                                      @Nullable
        +                                                                                                      org.apache.flink.table.connector.RowLevelModificationScanContext rowLevelModificationScanContext)
        +
        +
        Specified by:
        +
        applyRowLevelModificationScan in interface org.apache.flink.table.connector.source.abilities.SupportsRowLevelModificationScan
        +
        +
      • +
      + + + +
        +
      • +

        applyLimit

        +
        public void applyLimit​(long limit)
        +
        +
        Specified by:
        +
        applyLimit in interface org.apache.flink.table.connector.source.abilities.SupportsLimitPushDown
        +
        +
      • +
      + + + +
        +
      • +

        applyAggregates

        +
        public boolean applyAggregates​(List<int[]> groupingSets,
        +                               List<org.apache.flink.table.expressions.AggregateExpression> aggregateExpressions,
        +                               org.apache.flink.table.types.DataType dataType)
        +
        +
        Specified by:
        +
        applyAggregates in interface org.apache.flink.table.connector.source.abilities.SupportsAggregatePushDown
        +
        +
      • +
      + + + +
        +
      • +

        getCache

        +
        @VisibleForTesting
        +@Nullable
        +public org.apache.flink.table.connector.source.lookup.cache.LookupCache getCache()
        +
      • +
      + + + +
        +
      • +

        getPrimaryKeyIndexes

        +
        @VisibleForTesting
        +public int[] getPrimaryKeyIndexes()
        +
      • +
      + + + +
        +
      • +

        getBucketKeyIndexes

        +
        @VisibleForTesting
        +public int[] getBucketKeyIndexes()
        +
      • +
      + + + +
        +
      • +

        getPartitionKeyIndexes

        +
        @VisibleForTesting
        +public int[] getPartitionKeyIndexes()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlussSource.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlussSource.html new file mode 100644 index 0000000000..7dcf087164 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlussSource.html @@ -0,0 +1,338 @@ + + + + + +FlussSource (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlussSource<OUT>

+
+
+ +
+
    +
  • +
    +
    Type Parameters:
    +
    OUT - The type of records produced by this source
    +
    +
    +
    All Implemented Interfaces:
    +
    Serializable, org.apache.flink.api.connector.source.Source<OUT,​SourceSplitBase,​SourceEnumeratorState>, org.apache.flink.api.connector.source.SourceReaderFactory<OUT,​SourceSplitBase>, org.apache.flink.api.java.typeutils.ResultTypeQueryable
    +
    +
    +
    public class FlussSource<OUT>
    +extends FlinkSource<OUT>
    +
    A Flink DataStream source implementation for reading data from Fluss tables. + +

    This class extends the FlinkSource base class and implements + ResultTypeQueryable to provide type information for Flink's type system. + +

    Sample usage: + +

    
    + FlussSource<Order> flussSource = FlussSource.<Order>builder()
    +     .setBootstrapServers("localhost:9092")
    +     .setDatabase("mydb")
    +     .setTable("orders")
    +     .setProjectedFields("orderId", "amount")
    +     .setStartingOffsets(OffsetsInitializer.earliest())
    +     .setScanPartitionDiscoveryIntervalMs(1000L)
    +     .setDeserializationSchema(new OrderDeserializationSchema())
    +     .build();
    +
    + DataStreamSource<Order> stream = env.fromSource(
    +     flussSource,
    +     WatermarkStrategy.noWatermarks(),
    +     "Fluss Source"
    + );
    + 
    +
    +
    See Also:
    +
    Serialized Form
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        builder

        +
        public static <T> FlussSourceBuilder<T> builder()
        +
        Get a FlussSourceBuilder to build a FlussSource.
        +
        +
        Returns:
        +
        a Fluss source builder.
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlussSourceBuilder.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlussSourceBuilder.html new file mode 100644 index 0000000000..e3e9e95db6 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlussSourceBuilder.html @@ -0,0 +1,547 @@ + + + + + +FlussSourceBuilder (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlussSourceBuilder<OUT>

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.FlussSourceBuilder<OUT>
    • +
    +
  • +
+
+
    +
  • +
    +
    Type Parameters:
    +
    OUT - The type of records produced by the source being built
    +
    +
    +
    public class FlussSourceBuilder<OUT>
    +extends Object
    +
    Builder class for creating FlussSource instances. + +

    The builder allows for step-by-step configuration of a Fluss source connector. It handles the + setup of connection parameters, table metadata retrieval, and source configuration. + +

    Sample usage: + +

    
    + FlussSource<Order> source = FlussSource.<Order>builder()
    +     .setBootstrapServers("localhost:9092")
    +     .setDatabase("mydb")
    +     .setTable("orders")
    +     .setProjectedFields("orderId", "amount")
    +     .setScanPartitionDiscoveryIntervalMs(1000L)
    +     .setStartingOffsets(OffsetsInitializer.earliest())
    +     .setDeserializationSchema(new OrderDeserializationSchema())
    +     .build();
    + 
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlussSourceBuilder

        +
        public FlussSourceBuilder()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        setBootstrapServers

        +
        public FlussSourceBuilder<OUT> setBootstrapServers​(String bootstrapServers)
        +
        Sets the bootstrap servers for the Fluss source connection. + +

        This is a required parameter.

        +
        +
        Parameters:
        +
        bootstrapServers - bootstrap server addresses
        +
        Returns:
        +
        this builder
        +
        +
      • +
      + + + +
        +
      • +

        setDatabase

        +
        public FlussSourceBuilder<OUT> setDatabase​(String database)
        +
        Sets the database name for the Fluss source. + +

        This is a required parameter.

        +
        +
        Parameters:
        +
        database - name of the database
        +
        Returns:
        +
        this builder
        +
        +
      • +
      + + + +
        +
      • +

        setTable

        +
        public FlussSourceBuilder<OUT> setTable​(String table)
        +
        Sets the table name for the Fluss source. + +

        This is a required parameter.

        +
        +
        Parameters:
        +
        table - name of the table
        +
        Returns:
        +
        this builder
        +
        +
      • +
      + + + +
        +
      • +

        setScanPartitionDiscoveryIntervalMs

        +
        public FlussSourceBuilder<OUT> setScanPartitionDiscoveryIntervalMs​(long scanPartitionDiscoveryIntervalMs)
        +
        Sets the scan partition discovery interval in milliseconds. + +

        If not specified, the default value from FlinkConnectorOptions.SCAN_PARTITION_DISCOVERY_INTERVAL is used.

        +
        +
        Parameters:
        +
        scanPartitionDiscoveryIntervalMs - interval in milliseconds
        +
        Returns:
        +
        this builder
        +
        +
      • +
      + + + +
        +
      • +

        setStartingOffsets

        +
        public FlussSourceBuilder<OUT> setStartingOffsets​(OffsetsInitializer offsetsInitializer)
        +
        Sets the starting offsets strategy for the Fluss source. + +

        If not specified, OffsetsInitializer.full() is used by default.

        +
        +
        Parameters:
        +
        offsetsInitializer - the strategy for determining starting offsets
        +
        Returns:
        +
        this builder
        +
        +
      • +
      + + + +
        +
      • +

        setDeserializationSchema

        +
        public FlussSourceBuilder<OUT> setDeserializationSchema​(FlussDeserializationSchema<OUT> deserializationSchema)
        +
        Sets the deserialization schema for converting Fluss records to output records. + +

        This is a required parameter.

        +
        +
        Parameters:
        +
        deserializationSchema - the deserialization schema to use
        +
        Returns:
        +
        this builder
        +
        +
      • +
      + + + +
        +
      • +

        setProjectedFields

        +
        public FlussSourceBuilder<OUT> setProjectedFields​(String... projectedFieldNames)
        +
        Sets the projected fields for this source using field names. + +

        Projection allows selecting a subset of fields from the table. Without projection, all + fields from the table are included.

        +
        +
        Parameters:
        +
        projectedFieldNames - names of the fields to project
        +
        Returns:
        +
        this builder
        +
        Throws:
        +
        NullPointerException - if projectedFieldNames is null
        +
        +
      • +
      + + + +
        +
      • +

        setFlussConfig

        +
        public FlussSourceBuilder<OUT> setFlussConfig​(org.apache.fluss.config.Configuration flussConf)
        +
        Sets custom Fluss configuration properties for the source connector. + +

        If not specified, an empty configuration will be created and populated with required + properties. Any configuration set through this method will be merged with table-specific + properties retrieved from the Fluss system.

        +
        +
        Parameters:
        +
        flussConf - the configuration to use
        +
        Returns:
        +
        this builder
        +
        +
      • +
      + + + +
        +
      • +

        build

        +
        public FlussSource<OUT> build()
        +
        Builds and returns a new FlussSource instance with the configured properties. + +

        This method validates all required parameters, connects to the Fluss system to retrieve + table metadata, and constructs a configured source.

        +
        +
        Returns:
        +
        a new FlussSource instance
        +
        Throws:
        +
        NullPointerException - if any required parameter is missing
        +
        IllegalArgumentException - if any parameter is invalid
        +
        RuntimeException - if connection to Fluss fails or the table cannot be found
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlinkSource.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlinkSource.html new file mode 100644 index 0000000000..f08faf14b9 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlinkSource.html @@ -0,0 +1,194 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.FlinkSource (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.FlinkSource

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlinkTableSource.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlinkTableSource.html new file mode 100644 index 0000000000..c670250364 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlinkTableSource.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.FlinkTableSource (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.FlinkTableSource

+
+
No usage of org.apache.fluss.flink.source.FlinkTableSource
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlussSource.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlussSource.html new file mode 100644 index 0000000000..d8a927b34d --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlussSource.html @@ -0,0 +1,194 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.FlussSource (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.FlussSource

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlussSourceBuilder.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlussSourceBuilder.html new file mode 100644 index 0000000000..1262dd4965 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlussSourceBuilder.html @@ -0,0 +1,250 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.FlussSourceBuilder (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.FlussSourceBuilder

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/DeserializerInitContextImpl.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/DeserializerInitContextImpl.html new file mode 100644 index 0000000000..af6e9287d7 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/DeserializerInitContextImpl.html @@ -0,0 +1,380 @@ + + + + + +DeserializerInitContextImpl (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class DeserializerInitContextImpl

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.deserializer.DeserializerInitContextImpl
    • +
    +
  • +
+
+ +
+
+ +
+
+ +
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/FlussDeserializationSchema.InitializationContext.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/FlussDeserializationSchema.InitializationContext.html new file mode 100644 index 0000000000..24038ddf98 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/FlussDeserializationSchema.InitializationContext.html @@ -0,0 +1,322 @@ + + + + + +FlussDeserializationSchema.InitializationContext (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface FlussDeserializationSchema.InitializationContext

+
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Summary

      + + + + + + + + + + + + + + + + + + + + + + +
      All Methods Instance Methods Abstract Methods 
      Modifier and TypeMethodDescription
      org.apache.flink.metrics.MetricGroupgetMetricGroup() +
      Returns the metric group for the parallel subtask of the source that runs this FlussDeserializationSchema.
      +
      org.apache.fluss.types.RowTypegetRowSchema() +
      Returns the schema of the LogRecord.getRow().
      +
      org.apache.flink.util.UserCodeClassLoadergetUserCodeClassLoader() +
      Gets the UserCodeClassLoader to load classes that are not in system's classpath, + but are part of the jar file of a user job.
      +
      +
    • +
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getMetricGroup

        +
        org.apache.flink.metrics.MetricGroup getMetricGroup()
        +
        Returns the metric group for the parallel subtask of the source that runs this FlussDeserializationSchema. + +

        Instances of this class can be used to register new metrics with Flink and to create a + nested hierarchy based on the group names. See MetricGroup for more information + for the metrics system.

        +
        +
        See Also:
        +
        MetricGroup
        +
        +
      • +
      + + + +
        +
      • +

        getUserCodeClassLoader

        +
        org.apache.flink.util.UserCodeClassLoader getUserCodeClassLoader()
        +
        Gets the UserCodeClassLoader to load classes that are not in system's classpath, + but are part of the jar file of a user job.
        +
        +
        See Also:
        +
        UserCodeClassLoader
        +
        +
      • +
      + + + +
        +
      • +

        getRowSchema

        +
        org.apache.fluss.types.RowType getRowSchema()
        +
        Returns the schema of the LogRecord.getRow().
        +
        +
        Returns:
        +
        The schema of the LogRecord.getRow().
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/FlussDeserializationSchema.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/FlussDeserializationSchema.html new file mode 100644 index 0000000000..40cd6d4637 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/FlussDeserializationSchema.html @@ -0,0 +1,357 @@ + + + + + +FlussDeserializationSchema (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface FlussDeserializationSchema<T>

+
+
+
+
    +
  • +
    +
    Type Parameters:
    +
    T - The type created by the deserialization schema.
    +
    +
    +
    All Superinterfaces:
    +
    Serializable
    +
    +
    +
    All Known Implementing Classes:
    +
    JsonStringDeserializationSchema, RowDataDeserializationSchema
    +
    +
    +
    @PublicEvolving
    +public interface FlussDeserializationSchema<T>
    +extends Serializable
    +
    Interface for deserialization schema used to deserialize LogRecord objects into specific + data types.
    +
    +
    Since:
    +
    0.7
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + + + + + +
        +
      • +

        deserialize

        +
        T deserialize​(org.apache.fluss.record.LogRecord record)
        +       throws Exception
        +
        Deserializes a LogRecord into an object of type T.
        +
        +
        Parameters:
        +
        record - The Fluss record to deserialize.
        +
        Returns:
        +
        The deserialized object.
        +
        Throws:
        +
        Exception - If the deserialization fails.
        +
        +
      • +
      + + + +
        +
      • +

        getProducedType

        +
        org.apache.flink.api.common.typeinfo.TypeInformation<T> getProducedType​(org.apache.fluss.types.RowType rowSchema)
        +
        Gets the data type (as a TypeInformation) produced by this deserializer.
        +
        +
        Parameters:
        +
        rowSchema - The schema of the LogRecord.getRow().
        +
        Returns:
        +
        The data type produced by this deserializer.
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/JsonStringDeserializationSchema.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/JsonStringDeserializationSchema.html new file mode 100644 index 0000000000..9852c079b0 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/JsonStringDeserializationSchema.html @@ -0,0 +1,435 @@ + + + + + +JsonStringDeserializationSchema (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class JsonStringDeserializationSchema

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.deserializer.JsonStringDeserializationSchema
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    Serializable, FlussDeserializationSchema<String>
    +
    +
    +
    @PublicEvolving
    +public class JsonStringDeserializationSchema
    +extends Object
    +implements FlussDeserializationSchema<String>
    +
    A deserialization schema that converts LogRecord objects to JSON strings. + +

    This implementation serializes Fluss records into JSON strings, making it useful for + debugging, logging, or when the downstream processing requires string-based JSON data. The schema + preserves important metadata such as offset, timestamp, and change type along with the actual row + data. + +

    The resulting JSON has the following structure: + +

    
    + {
    +   "offset": <record_offset>,
    +   "timestamp": <record_timestamp>,
    +   "changeType": <APPEND_ONLY|INSERT|UPDATE_BEFORE|UPDATE_AFTER|DELETE>,
    +   "row": <string_representation_of_row>
    + }
    + 
    + +

    Usage example: + +

    
    + FlussSource<String> source = FlussSource.builder()
    +     .setDeserializationSchema(new JsonStringDeserializationSchema())
    +     .build();
    + 
    +
    +
    Since:
    +
    0.7
    +
    See Also:
    +
    Serialized Form
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        JsonStringDeserializationSchema

        +
        public JsonStringDeserializationSchema()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        open

        +
        public void open​(FlussDeserializationSchema.InitializationContext context)
        +          throws Exception
        +
        Initializes the JSON serialization mechanism. + +

        This method creates a new ObjectMapper instance and configures it with: + +

          +
        • JavaTimeModule for proper serialization of date/time objects +
        • Configuration to render dates in ISO-8601 format rather than timestamps +
        +
        +
        Specified by:
        +
        open in interface FlussDeserializationSchema<String>
        +
        Parameters:
        +
        context - Contextual information for initialization (not used in this implementation)
        +
        Throws:
        +
        Exception - if initialization fails
        +
        +
      • +
      + + + +
        +
      • +

        deserialize

        +
        public String deserialize​(org.apache.fluss.record.LogRecord record)
        +                   throws Exception
        +
        Deserializes a LogRecord into a JSON String. + +

        The method extracts key information from the record (offset, timestamp, change type, and + row data) and serializes it as a JSON string.

        +
        +
        Specified by:
        +
        deserialize in interface FlussDeserializationSchema<String>
        +
        Parameters:
        +
        record - The Fluss LogRecord to deserialize
        +
        Returns:
        +
        JSON string representation of the record
        +
        Throws:
        +
        Exception - If JSON serialization fails
        +
        +
      • +
      + + + +
        +
      • +

        getProducedType

        +
        public org.apache.flink.api.common.typeinfo.TypeInformation<String> getProducedType​(org.apache.fluss.types.RowType rowSchema)
        +
        Returns the TypeInformation for the produced String type.
        +
        +
        Specified by:
        +
        getProducedType in interface FlussDeserializationSchema<String>
        +
        Parameters:
        +
        rowSchema - The schema of the LogRecord.getRow().
        +
        Returns:
        +
        TypeInformation for String class
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/RowDataDeserializationSchema.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/RowDataDeserializationSchema.html new file mode 100644 index 0000000000..233aba8656 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/RowDataDeserializationSchema.html @@ -0,0 +1,414 @@ + + + + + +RowDataDeserializationSchema (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class RowDataDeserializationSchema

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.deserializer.RowDataDeserializationSchema
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    Serializable, FlussDeserializationSchema<org.apache.flink.table.data.RowData>
    +
    +
    +
    @PublicEvolving
    +public class RowDataDeserializationSchema
    +extends Object
    +implements FlussDeserializationSchema<org.apache.flink.table.data.RowData>
    +
    A deserialization schema that converts LogRecord objects to Flink's RowData + format. + +

    This implementation takes a RowType in its constructor and uses a FlussRowToFlinkRowConverter to transform Fluss records into Flink's internal row representation. + +

    Usage example: + +

    
    + FlussSource<RowData> source = FlussSource.builder()
    +     .setDeserializationSchema(new RowDataDeserializationSchema())
    +     .build();
    + 
    +
    +
    Since:
    +
    0.7
    +
    See Also:
    +
    Serialized Form
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        RowDataDeserializationSchema

        +
        public RowDataDeserializationSchema()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + + + + + +
        +
      • +

        deserialize

        +
        public org.apache.flink.table.data.RowData deserialize​(org.apache.fluss.record.LogRecord record)
        +                                                throws Exception
        +
        Deserializes a LogRecord into a Flink RowData object.
        +
        +
        Specified by:
        +
        deserialize in interface FlussDeserializationSchema<org.apache.flink.table.data.RowData>
        +
        Parameters:
        +
        record - The Fluss LogRecord to deserialize
        +
        Returns:
        +
        The deserialized RowData
        +
        Throws:
        +
        Exception - If deserialization fails or if the record is not a valid ScanRecord
        +
        +
      • +
      + + + +
        +
      • +

        getProducedType

        +
        public org.apache.flink.api.common.typeinfo.TypeInformation<org.apache.flink.table.data.RowData> getProducedType​(org.apache.fluss.types.RowType rowSchema)
        +
        Returns the TypeInformation for the produced RowData type.
        +
        +
        Specified by:
        +
        getProducedType in interface FlussDeserializationSchema<org.apache.flink.table.data.RowData>
        +
        Parameters:
        +
        rowSchema - The schema of the LogRecord.getRow().
        +
        Returns:
        +
        The data type produced by this deserializer.
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/DeserializerInitContextImpl.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/DeserializerInitContextImpl.html new file mode 100644 index 0000000000..ded3fa0b21 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/DeserializerInitContextImpl.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.deserializer.DeserializerInitContextImpl (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.deserializer.DeserializerInitContextImpl

+
+
No usage of org.apache.fluss.flink.source.deserializer.DeserializerInitContextImpl
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/FlussDeserializationSchema.InitializationContext.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/FlussDeserializationSchema.InitializationContext.html new file mode 100644 index 0000000000..5a8866e2f6 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/FlussDeserializationSchema.InitializationContext.html @@ -0,0 +1,225 @@ + + + + + +Uses of Interface org.apache.fluss.flink.source.deserializer.FlussDeserializationSchema.InitializationContext (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.flink.source.deserializer.FlussDeserializationSchema.InitializationContext

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/FlussDeserializationSchema.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/FlussDeserializationSchema.html new file mode 100644 index 0000000000..e616a323c9 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/FlussDeserializationSchema.html @@ -0,0 +1,292 @@ + + + + + +Uses of Interface org.apache.fluss.flink.source.deserializer.FlussDeserializationSchema (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.flink.source.deserializer.FlussDeserializationSchema

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/JsonStringDeserializationSchema.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/JsonStringDeserializationSchema.html new file mode 100644 index 0000000000..4bee001ebc --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/JsonStringDeserializationSchema.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.deserializer.JsonStringDeserializationSchema (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.deserializer.JsonStringDeserializationSchema

+
+
No usage of org.apache.fluss.flink.source.deserializer.JsonStringDeserializationSchema
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/RowDataDeserializationSchema.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/RowDataDeserializationSchema.html new file mode 100644 index 0000000000..7dd401fdc8 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/RowDataDeserializationSchema.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.deserializer.RowDataDeserializationSchema (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.deserializer.RowDataDeserializationSchema

+
+
No usage of org.apache.fluss.flink.source.deserializer.RowDataDeserializationSchema
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-summary.html new file mode 100644 index 0000000000..38e10cb557 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-summary.html @@ -0,0 +1,205 @@ + + + + + +org.apache.fluss.flink.source.deserializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.source.deserializer

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-tree.html new file mode 100644 index 0000000000..178095e152 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-tree.html @@ -0,0 +1,176 @@ + + + + + +org.apache.fluss.flink.source.deserializer Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.source.deserializer

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+

Interface Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-use.html new file mode 100644 index 0000000000..6e12518ca7 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-use.html @@ -0,0 +1,240 @@ + + + + + +Uses of Package org.apache.fluss.flink.source.deserializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.source.deserializer

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/FlinkRecordEmitter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/FlinkRecordEmitter.html new file mode 100644 index 0000000000..054773ae42 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/FlinkRecordEmitter.html @@ -0,0 +1,327 @@ + + + + + +FlinkRecordEmitter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkRecordEmitter<OUT>

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.emitter.FlinkRecordEmitter<OUT>
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.connector.base.source.reader.RecordEmitter<RecordAndPos,​OUT,​SourceSplitState>
    +
    +
    +
    public class FlinkRecordEmitter<OUT>
    +extends Object
    +implements org.apache.flink.connector.base.source.reader.RecordEmitter<RecordAndPos,​OUT,​SourceSplitState>
    +
    The RecordEmitter implementation for FlinkSourceReader. + +

    During emitting records: + +

    when the record is from snapshot data, it'll update the records number to skip which helps to + skip the records has been read while restoring in reading snapshot data phase. + +

    when the record is from log data, it'll update the offset

    +
  • +
+
+
+ +
+
+
    +
  • + +
    + +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        emitRecord

        +
        public void emitRecord​(RecordAndPos recordAndPosition,
        +                       org.apache.flink.api.connector.source.SourceOutput<OUT> sourceOutput,
        +                       SourceSplitState splitState)
        +
        +
        Specified by:
        +
        emitRecord in interface org.apache.flink.connector.base.source.reader.RecordEmitter<RecordAndPos,​OUT,​SourceSplitState>
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/class-use/FlinkRecordEmitter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/class-use/FlinkRecordEmitter.html new file mode 100644 index 0000000000..5cbd53ddeb --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/class-use/FlinkRecordEmitter.html @@ -0,0 +1,198 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.emitter.FlinkRecordEmitter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.emitter.FlinkRecordEmitter

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-summary.html new file mode 100644 index 0000000000..277114fc8f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-summary.html @@ -0,0 +1,168 @@ + + + + + +org.apache.fluss.flink.source.emitter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.source.emitter

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-tree.html new file mode 100644 index 0000000000..e2cb4b1aed --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-tree.html @@ -0,0 +1,163 @@ + + + + + +org.apache.fluss.flink.source.emitter Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.source.emitter

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+
    +
  • java.lang.Object +
      +
    • org.apache.fluss.flink.source.emitter.FlinkRecordEmitter<OUT> (implements org.apache.flink.connector.base.source.reader.RecordEmitter<E,​T,​SplitStateT>)
    • +
    +
  • +
+
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-use.html new file mode 100644 index 0000000000..4b74eb88a1 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-use.html @@ -0,0 +1,185 @@ + + + + + +Uses of Package org.apache.fluss.flink.source.emitter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.source.emitter

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/FlinkSourceEnumerator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/FlinkSourceEnumerator.html new file mode 100644 index 0000000000..dc735a125e --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/FlinkSourceEnumerator.html @@ -0,0 +1,580 @@ + + + + + +FlinkSourceEnumerator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkSourceEnumerator

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    AutoCloseable, org.apache.flink.api.common.state.CheckpointListener, org.apache.flink.api.connector.source.SplitEnumerator<SourceSplitBase,​SourceEnumeratorState>
    +
    +
    +
    public class FlinkSourceEnumerator
    +extends Object
    +implements org.apache.flink.api.connector.source.SplitEnumerator<SourceSplitBase,​SourceEnumeratorState>
    +
    An implementation of SplitEnumerator for the data of Fluss. + +

    The enumerator is responsible for: + +

      +
    • Get the all splits(snapshot split + log split) for a table of Fluss to be read. +
    • Assign the splits to readers with the guarantee that the splits belong to the same bucket + will be assigned to same reader. +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Summary

      + + + + + + + + + + + + + + + + + + +
      Constructors 
      ConstructorDescription
      FlinkSourceEnumerator​(org.apache.fluss.metadata.TablePath tablePath, + org.apache.fluss.config.Configuration flussConf, + boolean hasPrimaryKey, + boolean isPartitioned, + org.apache.flink.api.connector.source.SplitEnumeratorContext<SourceSplitBase> context, + Set<org.apache.fluss.metadata.TableBucket> assignedTableBuckets, + Map<Long,​String> assignedPartitions, + List<SourceSplitBase> pendingHybridLakeFlussSplits, + OffsetsInitializer startingOffsetsInitializer, + long scanPartitionDiscoveryIntervalMs, + boolean streaming, + List<PushdownUtils.FieldEqual> partitionFilters, + org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource) 
      FlinkSourceEnumerator​(org.apache.fluss.metadata.TablePath tablePath, + org.apache.fluss.config.Configuration flussConf, + boolean hasPrimaryKey, + boolean isPartitioned, + org.apache.flink.api.connector.source.SplitEnumeratorContext<SourceSplitBase> context, + OffsetsInitializer startingOffsetsInitializer, + long scanPartitionDiscoveryIntervalMs, + boolean streaming, + List<PushdownUtils.FieldEqual> partitionFilters) 
      FlinkSourceEnumerator​(org.apache.fluss.metadata.TablePath tablePath, + org.apache.fluss.config.Configuration flussConf, + boolean hasPrimaryKey, + boolean isPartitioned, + org.apache.flink.api.connector.source.SplitEnumeratorContext<SourceSplitBase> context, + OffsetsInitializer startingOffsetsInitializer, + long scanPartitionDiscoveryIntervalMs, + boolean streaming, + List<PushdownUtils.FieldEqual> partitionFilters, + org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource) 
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlinkSourceEnumerator

        +
        public FlinkSourceEnumerator​(org.apache.fluss.metadata.TablePath tablePath,
        +                             org.apache.fluss.config.Configuration flussConf,
        +                             boolean hasPrimaryKey,
        +                             boolean isPartitioned,
        +                             org.apache.flink.api.connector.source.SplitEnumeratorContext<SourceSplitBase> context,
        +                             OffsetsInitializer startingOffsetsInitializer,
        +                             long scanPartitionDiscoveryIntervalMs,
        +                             boolean streaming,
        +                             List<PushdownUtils.FieldEqual> partitionFilters)
        +
      • +
      + + + +
        +
      • +

        FlinkSourceEnumerator

        +
        public FlinkSourceEnumerator​(org.apache.fluss.metadata.TablePath tablePath,
        +                             org.apache.fluss.config.Configuration flussConf,
        +                             boolean hasPrimaryKey,
        +                             boolean isPartitioned,
        +                             org.apache.flink.api.connector.source.SplitEnumeratorContext<SourceSplitBase> context,
        +                             OffsetsInitializer startingOffsetsInitializer,
        +                             long scanPartitionDiscoveryIntervalMs,
        +                             boolean streaming,
        +                             List<PushdownUtils.FieldEqual> partitionFilters,
        +                             @Nullable
        +                             org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource)
        +
      • +
      + + + +
        +
      • +

        FlinkSourceEnumerator

        +
        public FlinkSourceEnumerator​(org.apache.fluss.metadata.TablePath tablePath,
        +                             org.apache.fluss.config.Configuration flussConf,
        +                             boolean hasPrimaryKey,
        +                             boolean isPartitioned,
        +                             org.apache.flink.api.connector.source.SplitEnumeratorContext<SourceSplitBase> context,
        +                             Set<org.apache.fluss.metadata.TableBucket> assignedTableBuckets,
        +                             Map<Long,​String> assignedPartitions,
        +                             List<SourceSplitBase> pendingHybridLakeFlussSplits,
        +                             OffsetsInitializer startingOffsetsInitializer,
        +                             long scanPartitionDiscoveryIntervalMs,
        +                             boolean streaming,
        +                             List<PushdownUtils.FieldEqual> partitionFilters,
        +                             @Nullable
        +                             org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + + + + + +
        +
      • +

        getSplitOwner

        +
        @VisibleForTesting
        +protected int getSplitOwner​(SourceSplitBase split)
        +
        Returns the index of the target subtask that a specific split should be assigned to. + +

        The resulting distribution of splits of a single table has the following contract: + +

          +
        • 1. Splits in same bucket are assigned to same subtask +
        • 2. Uniformly distributed across subtasks +
        • 3. For partitioned table, the buckets in same partition are round-robin distributed + (strictly clockwise w.r.t. ascending subtask indices) by using the partition id as the + offset from a starting index. The starting index is the index of the subtask which + bucket 0 of the partition will be assigned to, determined using the partition id to + make sure the partitions' buckets of a table are distributed uniformly +
        +
        +
        Parameters:
        +
        split - the split to assign.
        +
        Returns:
        +
        the id of the subtask that owns the split.
        +
        +
      • +
      + + + +
        +
      • +

        handleSplitRequest

        +
        public void handleSplitRequest​(int subtaskId,
        +                               @Nullable
        +                               String requesterHostname)
        +
        +
        Specified by:
        +
        handleSplitRequest in interface org.apache.flink.api.connector.source.SplitEnumerator<SourceSplitBase,​SourceEnumeratorState>
        +
        +
      • +
      + + + +
        +
      • +

        handleSourceEvent

        +
        public void handleSourceEvent​(int subtaskId,
        +                              org.apache.flink.api.connector.source.SourceEvent sourceEvent)
        +
        +
        Specified by:
        +
        handleSourceEvent in interface org.apache.flink.api.connector.source.SplitEnumerator<SourceSplitBase,​SourceEnumeratorState>
        +
        +
      • +
      + + + + + + + +
        +
      • +

        addReader

        +
        public void addReader​(int subtaskId)
        +
        +
        Specified by:
        +
        addReader in interface org.apache.flink.api.connector.source.SplitEnumerator<SourceSplitBase,​SourceEnumeratorState>
        +
        +
      • +
      + + + + + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/class-use/FlinkSourceEnumerator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/class-use/FlinkSourceEnumerator.html new file mode 100644 index 0000000000..2df23534ee --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/class-use/FlinkSourceEnumerator.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator

+
+
No usage of org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/BucketOffsetsRetrieverImpl.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/BucketOffsetsRetrieverImpl.html new file mode 100644 index 0000000000..2b1458677a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/BucketOffsetsRetrieverImpl.html @@ -0,0 +1,365 @@ + + + + + +BucketOffsetsRetrieverImpl (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class BucketOffsetsRetrieverImpl

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.enumerator.initializer.BucketOffsetsRetrieverImpl
    • +
    +
  • +
+
+ +
+
+ +
+
+ +
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/LatestOffsetsInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/LatestOffsetsInitializer.html new file mode 100644 index 0000000000..a2736bfaeb --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/LatestOffsetsInitializer.html @@ -0,0 +1,357 @@ + + + + + +LatestOffsetsInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LatestOffsetsInitializer

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.enumerator.initializer.LatestOffsetsInitializer
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LatestOffsetsInitializer

        +
        public LatestOffsetsInitializer()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getBucketOffsets

        +
        public Map<Integer,​Long> getBucketOffsets​(@Nullable
        +                                                String partitionName,
        +                                                Collection<Integer> buckets,
        +                                                OffsetsInitializer.BucketOffsetsRetriever bucketOffsetsRetriever)
        +
        Description copied from interface: OffsetsInitializer
        +
        Get the initial offsets for the given fluss buckets. These offsets will be used as starting + offsets of the fluss buckets.
        +
        +
        Specified by:
        +
        getBucketOffsets in interface OffsetsInitializer
        +
        Parameters:
        +
        partitionName - the partition name of the buckets if they are partitioned. Otherwise, + null.
        +
        buckets - the fluss buckets to get the starting offsets.
        +
        bucketOffsetsRetriever - a helper to retrieve information of the fluss buckets.
        +
        Returns:
        +
        A mapping from fluss bucket to their offsets to start scanning from.
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/NoStoppingOffsetsInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/NoStoppingOffsetsInitializer.html new file mode 100644 index 0000000000..c894aa0b06 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/NoStoppingOffsetsInitializer.html @@ -0,0 +1,356 @@ + + + + + +NoStoppingOffsetsInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class NoStoppingOffsetsInitializer

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.enumerator.initializer.NoStoppingOffsetsInitializer
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        NoStoppingOffsetsInitializer

        +
        public NoStoppingOffsetsInitializer()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getBucketOffsets

        +
        public Map<Integer,​Long> getBucketOffsets​(@Nullable
        +                                                String partitionName,
        +                                                Collection<Integer> buckets,
        +                                                OffsetsInitializer.BucketOffsetsRetriever bucketOffsetsRetriever)
        +
        Description copied from interface: OffsetsInitializer
        +
        Get the initial offsets for the given fluss buckets. These offsets will be used as starting + offsets of the fluss buckets.
        +
        +
        Specified by:
        +
        getBucketOffsets in interface OffsetsInitializer
        +
        Parameters:
        +
        partitionName - the partition name of the buckets if they are partitioned. Otherwise, + null.
        +
        buckets - the fluss buckets to get the starting offsets.
        +
        bucketOffsetsRetriever - a helper to retrieve information of the fluss buckets.
        +
        Returns:
        +
        A mapping from fluss bucket to their offsets to start scanning from.
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/OffsetsInitializer.BucketOffsetsRetriever.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/OffsetsInitializer.BucketOffsetsRetriever.html new file mode 100644 index 0000000000..bb48bcea13 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/OffsetsInitializer.BucketOffsetsRetriever.html @@ -0,0 +1,299 @@ + + + + + +OffsetsInitializer.BucketOffsetsRetriever (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface OffsetsInitializer.BucketOffsetsRetriever

+
+
+
+
    +
  • +
    +
    All Known Implementing Classes:
    +
    BucketOffsetsRetrieverImpl
    +
    +
    +
    Enclosing interface:
    +
    OffsetsInitializer
    +
    +
    +
    public static interface OffsetsInitializer.BucketOffsetsRetriever
    +
    An interface that provides necessary information to the OffsetsInitializer to get the + initial offsets of the fluss buckets.
    +
  • +
+
+
+ +
+
+ +
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/OffsetsInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/OffsetsInitializer.html new file mode 100644 index 0000000000..d2a8709429 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/OffsetsInitializer.html @@ -0,0 +1,410 @@ + + + + + +OffsetsInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface OffsetsInitializer

+
+
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getBucketOffsets

        +
        Map<Integer,​Long> getBucketOffsets​(@Nullable
        +                                         String partitionName,
        +                                         Collection<Integer> buckets,
        +                                         OffsetsInitializer.BucketOffsetsRetriever bucketOffsetsRetriever)
        +
        Get the initial offsets for the given fluss buckets. These offsets will be used as starting + offsets of the fluss buckets.
        +
        +
        Parameters:
        +
        partitionName - the partition name of the buckets if they are partitioned. Otherwise, + null.
        +
        buckets - the fluss buckets to get the starting offsets.
        +
        bucketOffsetsRetriever - a helper to retrieve information of the fluss buckets.
        +
        Returns:
        +
        A mapping from fluss bucket to their offsets to start scanning from.
        +
        +
      • +
      + + + + + + + + + + + +
        +
      • +

        full

        +
        static OffsetsInitializer full()
        +
        Get an OffsetsInitializer which performs a full snapshot on the table upon first + startup, and continue to read log with the offset to the snapshot. + +

        If the table to read is a log table, the full snapshot means reading from the earliest log + offset which means "full" OffsetsInitializer equal to the earliest() + OffsetsInitializer. If the table to read is a primary key table, the full snapshot means + reading the latest snapshot which materializes all changes on the table.

        +
        +
        Returns:
        +
        an OffsetsInitializer which initializes the offsets to snapshot offsets.
        +
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/SnapshotOffsetsInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/SnapshotOffsetsInitializer.html new file mode 100644 index 0000000000..dae9ff6d1f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/SnapshotOffsetsInitializer.html @@ -0,0 +1,357 @@ + + + + + +SnapshotOffsetsInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class SnapshotOffsetsInitializer

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.enumerator.initializer.SnapshotOffsetsInitializer
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        SnapshotOffsetsInitializer

        +
        public SnapshotOffsetsInitializer()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getBucketOffsets

        +
        public Map<Integer,​Long> getBucketOffsets​(@Nullable
        +                                                String partitionName,
        +                                                Collection<Integer> buckets,
        +                                                OffsetsInitializer.BucketOffsetsRetriever bucketOffsetsRetriever)
        +
        For table with primary key. This method will be invoked only when the kv snapshot not exists.
        +
        +
        Specified by:
        +
        getBucketOffsets in interface OffsetsInitializer
        +
        Parameters:
        +
        partitionName - the partition name of the buckets if they are partitioned. Otherwise, + null.
        +
        buckets - the fluss buckets to get the starting offsets.
        +
        bucketOffsetsRetriever - a helper to retrieve information of the fluss buckets.
        +
        Returns:
        +
        A mapping from fluss bucket to their offsets to start scanning from.
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/TimestampOffsetsInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/TimestampOffsetsInitializer.html new file mode 100644 index 0000000000..dab4adf861 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/TimestampOffsetsInitializer.html @@ -0,0 +1,356 @@ + + + + + +TimestampOffsetsInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TimestampOffsetsInitializer

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.enumerator.initializer.TimestampOffsetsInitializer
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        TimestampOffsetsInitializer

        +
        public TimestampOffsetsInitializer​(long timestamp)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getBucketOffsets

        +
        public Map<Integer,​Long> getBucketOffsets​(@Nullable
        +                                                String partitionName,
        +                                                Collection<Integer> buckets,
        +                                                OffsetsInitializer.BucketOffsetsRetriever bucketOffsetsRetriever)
        +
        Description copied from interface: OffsetsInitializer
        +
        Get the initial offsets for the given fluss buckets. These offsets will be used as starting + offsets of the fluss buckets.
        +
        +
        Specified by:
        +
        getBucketOffsets in interface OffsetsInitializer
        +
        Parameters:
        +
        partitionName - the partition name of the buckets if they are partitioned. Otherwise, + null.
        +
        buckets - the fluss buckets to get the starting offsets.
        +
        bucketOffsetsRetriever - a helper to retrieve information of the fluss buckets.
        +
        Returns:
        +
        A mapping from fluss bucket to their offsets to start scanning from.
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/BucketOffsetsRetrieverImpl.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/BucketOffsetsRetrieverImpl.html new file mode 100644 index 0000000000..2215f82d13 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/BucketOffsetsRetrieverImpl.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.enumerator.initializer.BucketOffsetsRetrieverImpl (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.enumerator.initializer.BucketOffsetsRetrieverImpl

+
+
No usage of org.apache.fluss.flink.source.enumerator.initializer.BucketOffsetsRetrieverImpl
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/LatestOffsetsInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/LatestOffsetsInitializer.html new file mode 100644 index 0000000000..642d051545 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/LatestOffsetsInitializer.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.enumerator.initializer.LatestOffsetsInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.enumerator.initializer.LatestOffsetsInitializer

+
+
No usage of org.apache.fluss.flink.source.enumerator.initializer.LatestOffsetsInitializer
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/NoStoppingOffsetsInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/NoStoppingOffsetsInitializer.html new file mode 100644 index 0000000000..3870876bc6 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/NoStoppingOffsetsInitializer.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.enumerator.initializer.NoStoppingOffsetsInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.enumerator.initializer.NoStoppingOffsetsInitializer

+
+
No usage of org.apache.fluss.flink.source.enumerator.initializer.NoStoppingOffsetsInitializer
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/OffsetsInitializer.BucketOffsetsRetriever.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/OffsetsInitializer.BucketOffsetsRetriever.html new file mode 100644 index 0000000000..7dac3b93c6 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/OffsetsInitializer.BucketOffsetsRetriever.html @@ -0,0 +1,273 @@ + + + + + +Uses of Interface org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer.BucketOffsetsRetriever (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer.BucketOffsetsRetriever

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/OffsetsInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/OffsetsInitializer.html new file mode 100644 index 0000000000..ebd9feca7b --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/OffsetsInitializer.html @@ -0,0 +1,431 @@ + + + + + +Uses of Interface org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/SnapshotOffsetsInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/SnapshotOffsetsInitializer.html new file mode 100644 index 0000000000..364841325c --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/SnapshotOffsetsInitializer.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.enumerator.initializer.SnapshotOffsetsInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.enumerator.initializer.SnapshotOffsetsInitializer

+
+
No usage of org.apache.fluss.flink.source.enumerator.initializer.SnapshotOffsetsInitializer
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/TimestampOffsetsInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/TimestampOffsetsInitializer.html new file mode 100644 index 0000000000..1522024ec2 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/TimestampOffsetsInitializer.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.enumerator.initializer.TimestampOffsetsInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.enumerator.initializer.TimestampOffsetsInitializer

+
+
No usage of org.apache.fluss.flink.source.enumerator.initializer.TimestampOffsetsInitializer
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-summary.html new file mode 100644 index 0000000000..a7ebe193da --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-summary.html @@ -0,0 +1,217 @@ + + + + + +org.apache.fluss.flink.source.enumerator.initializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.source.enumerator.initializer

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-tree.html new file mode 100644 index 0000000000..e48355cd6f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-tree.html @@ -0,0 +1,178 @@ + + + + + +org.apache.fluss.flink.source.enumerator.initializer Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.source.enumerator.initializer

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+

Interface Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-use.html new file mode 100644 index 0000000000..529cbdaf1a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-use.html @@ -0,0 +1,268 @@ + + + + + +Uses of Package org.apache.fluss.flink.source.enumerator.initializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.source.enumerator.initializer

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-summary.html new file mode 100644 index 0000000000..6e9be7be81 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-summary.html @@ -0,0 +1,168 @@ + + + + + +org.apache.fluss.flink.source.enumerator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.source.enumerator

+
+
+
    +
  • + + + + + + + + + + + + +
    Class Summary 
    ClassDescription
    FlinkSourceEnumerator +
    An implementation of SplitEnumerator for the data of Fluss.
    +
    +
  • +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-tree.html new file mode 100644 index 0000000000..499e58e6b0 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-tree.html @@ -0,0 +1,163 @@ + + + + + +org.apache.fluss.flink.source.enumerator Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.source.enumerator

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+
    +
  • java.lang.Object +
      +
    • org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator (implements org.apache.flink.api.connector.source.SplitEnumerator<SplitT,​CheckpointT>)
    • +
    +
  • +
+
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-use.html new file mode 100644 index 0000000000..0356c36bf7 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-use.html @@ -0,0 +1,148 @@ + + + + + +Uses of Package org.apache.fluss.flink.source.enumerator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.source.enumerator

+
+
No usage of org.apache.fluss.flink.source.enumerator
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/PartitionBucketsUnsubscribedEvent.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/PartitionBucketsUnsubscribedEvent.html new file mode 100644 index 0000000000..63459e3ab6 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/PartitionBucketsUnsubscribedEvent.html @@ -0,0 +1,371 @@ + + + + + +PartitionBucketsUnsubscribedEvent (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class PartitionBucketsUnsubscribedEvent

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.event.PartitionBucketsUnsubscribedEvent
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    Serializable, org.apache.flink.api.connector.source.SourceEvent
    +
    +
    +
    public class PartitionBucketsUnsubscribedEvent
    +extends Object
    +implements org.apache.flink.api.connector.source.SourceEvent
    +
    An event send from reader to enumerator to indicate the splits of the partition buckets have been + removed by the reader. It contains the table buckets that have been removed from the reader.
    +
    +
    See Also:
    +
    Serialized Form
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        PartitionBucketsUnsubscribedEvent

        +
        public PartitionBucketsUnsubscribedEvent​(Collection<org.apache.fluss.metadata.TableBucket> removedTableBuckets)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getRemovedTableBuckets

        +
        public Collection<org.apache.fluss.metadata.TableBucket> getRemovedTableBuckets()
        +
      • +
      + + + +
        +
      • +

        equals

        +
        public boolean equals​(Object o)
        +
        +
        Overrides:
        +
        equals in class Object
        +
        +
      • +
      + + + +
        +
      • +

        hashCode

        +
        public int hashCode()
        +
        +
        Overrides:
        +
        hashCode in class Object
        +
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/PartitionsRemovedEvent.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/PartitionsRemovedEvent.html new file mode 100644 index 0000000000..6702436025 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/PartitionsRemovedEvent.html @@ -0,0 +1,337 @@ + + + + + +PartitionsRemovedEvent (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class PartitionsRemovedEvent

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.event.PartitionsRemovedEvent
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    Serializable, org.apache.flink.api.connector.source.SourceEvent
    +
    +
    +
    public class PartitionsRemovedEvent
    +extends Object
    +implements org.apache.flink.api.connector.source.SourceEvent
    +
    A source event to represent partitions is removed to send from enumerator to reader. + +

    It contains the partition bucket of the removed partitions that has been assigned to the + reader.

    +
    +
    See Also:
    +
    Serialized Form
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        PartitionsRemovedEvent

        +
        public PartitionsRemovedEvent​(Map<Long,​String> removedPartitions)
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/class-use/PartitionBucketsUnsubscribedEvent.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/class-use/PartitionBucketsUnsubscribedEvent.html new file mode 100644 index 0000000000..938210a694 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/class-use/PartitionBucketsUnsubscribedEvent.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.event.PartitionBucketsUnsubscribedEvent (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.event.PartitionBucketsUnsubscribedEvent

+
+
No usage of org.apache.fluss.flink.source.event.PartitionBucketsUnsubscribedEvent
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/class-use/PartitionsRemovedEvent.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/class-use/PartitionsRemovedEvent.html new file mode 100644 index 0000000000..decfc9e303 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/class-use/PartitionsRemovedEvent.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.event.PartitionsRemovedEvent (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.event.PartitionsRemovedEvent

+
+
No usage of org.apache.fluss.flink.source.event.PartitionsRemovedEvent
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-summary.html new file mode 100644 index 0000000000..d3901f7075 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-summary.html @@ -0,0 +1,175 @@ + + + + + +org.apache.fluss.flink.source.event (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.source.event

+
+
+
    +
  • + + + + + + + + + + + + + + + + +
    Class Summary 
    ClassDescription
    PartitionBucketsUnsubscribedEvent +
    An event send from reader to enumerator to indicate the splits of the partition buckets have been + removed by the reader.
    +
    PartitionsRemovedEvent +
    A source event to represent partitions is removed to send from enumerator to reader.
    +
    +
  • +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-tree.html new file mode 100644 index 0000000000..8dc88334db --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-tree.html @@ -0,0 +1,164 @@ + + + + + +org.apache.fluss.flink.source.event Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.source.event

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-use.html new file mode 100644 index 0000000000..950dcb582c --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-use.html @@ -0,0 +1,148 @@ + + + + + +Uses of Package org.apache.fluss.flink.source.event (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.source.event

+
+
No usage of org.apache.fluss.flink.source.event
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/FlinkAsyncLookupFunction.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/FlinkAsyncLookupFunction.html new file mode 100644 index 0000000000..66248f02ea --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/FlinkAsyncLookupFunction.html @@ -0,0 +1,417 @@ + + + + + +FlinkAsyncLookupFunction (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkAsyncLookupFunction

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.flink.table.functions.UserDefinedFunction
    • +
    • +
        +
      • org.apache.flink.table.functions.AsyncTableFunction<org.apache.flink.table.data.RowData>
      • +
      • +
          +
        • org.apache.flink.table.functions.AsyncLookupFunction
        • +
        • +
            +
          • org.apache.fluss.flink.source.lookup.FlinkAsyncLookupFunction
          • +
          +
        • +
        +
      • +
      +
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    Serializable, org.apache.flink.table.functions.FunctionDefinition
    +
    +
    +
    public class FlinkAsyncLookupFunction
    +extends org.apache.flink.table.functions.AsyncLookupFunction
    +
    A flink async lookup function for fluss.
    +
    +
    See Also:
    +
    Serialized Form
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Summary

      + + + + + + + + + + +
      Constructors 
      ConstructorDescription
      FlinkAsyncLookupFunction​(org.apache.fluss.config.Configuration flussConfig, + org.apache.fluss.metadata.TablePath tablePath, + org.apache.flink.table.types.logical.RowType flinkRowType, + int maxRetryTimes, + LookupNormalizer lookupNormalizer, + int[] projection) 
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Summary

      + + + + + + + + + + + + + + + + + + + + + + +
      All Methods Instance Methods Concrete Methods 
      Modifier and TypeMethodDescription
      CompletableFuture<Collection<org.apache.flink.table.data.RowData>>asyncLookup​(org.apache.flink.table.data.RowData keyRow) +
      The invoke entry point of lookup function.
      +
      voidclose() 
      voidopen​(org.apache.flink.table.functions.FunctionContext context) 
      +
        +
      • + + +

        Methods inherited from class org.apache.flink.table.functions.AsyncLookupFunction

        +eval
      • +
      +
        +
      • + + +

        Methods inherited from class org.apache.flink.table.functions.AsyncTableFunction

        +getKind, getTypeInference
      • +
      +
        +
      • + + +

        Methods inherited from class org.apache.flink.table.functions.UserDefinedFunction

        +functionIdentifier, toString
      • +
      + +
        +
      • + + +

        Methods inherited from interface org.apache.flink.table.functions.FunctionDefinition

        +getRequirements, isDeterministic, supportsConstantFolding
      • +
      +
    • +
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlinkAsyncLookupFunction

        +
        public FlinkAsyncLookupFunction​(org.apache.fluss.config.Configuration flussConfig,
        +                                org.apache.fluss.metadata.TablePath tablePath,
        +                                org.apache.flink.table.types.logical.RowType flinkRowType,
        +                                int maxRetryTimes,
        +                                LookupNormalizer lookupNormalizer,
        +                                @Nullable
        +                                int[] projection)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        open

        +
        public void open​(org.apache.flink.table.functions.FunctionContext context)
        +
        +
        Overrides:
        +
        open in class org.apache.flink.table.functions.UserDefinedFunction
        +
        +
      • +
      + + + +
        +
      • +

        asyncLookup

        +
        public CompletableFuture<Collection<org.apache.flink.table.data.RowData>> asyncLookup​(org.apache.flink.table.data.RowData keyRow)
        +
        The invoke entry point of lookup function.
        +
        +
        Specified by:
        +
        asyncLookup in class org.apache.flink.table.functions.AsyncLookupFunction
        +
        Parameters:
        +
        keyRow - A RowData that wraps lookup keys.
        +
        +
      • +
      + + + +
        +
      • +

        close

        +
        public void close()
        +           throws Exception
        +
        +
        Overrides:
        +
        close in class org.apache.flink.table.functions.UserDefinedFunction
        +
        Throws:
        +
        Exception
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/FlinkLookupFunction.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/FlinkLookupFunction.html new file mode 100644 index 0000000000..c51e6bc2f6 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/FlinkLookupFunction.html @@ -0,0 +1,418 @@ + + + + + +FlinkLookupFunction (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkLookupFunction

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.flink.table.functions.UserDefinedFunction
    • +
    • +
        +
      • org.apache.flink.table.functions.TableFunction<org.apache.flink.table.data.RowData>
      • +
      • +
          +
        • org.apache.flink.table.functions.LookupFunction
        • +
        • +
            +
          • org.apache.fluss.flink.source.lookup.FlinkLookupFunction
          • +
          +
        • +
        +
      • +
      +
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    Serializable, org.apache.flink.table.functions.FunctionDefinition
    +
    +
    +
    public class FlinkLookupFunction
    +extends org.apache.flink.table.functions.LookupFunction
    +
    A flink lookup function for fluss.
    +
    +
    See Also:
    +
    Serialized Form
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Summary

      + + + + + + + + + + +
      Constructors 
      ConstructorDescription
      FlinkLookupFunction​(org.apache.fluss.config.Configuration flussConfig, + org.apache.fluss.metadata.TablePath tablePath, + org.apache.flink.table.types.logical.RowType flinkRowType, + int maxRetryTimes, + LookupNormalizer lookupNormalizer, + int[] projection) 
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Summary

      + + + + + + + + + + + + + + + + + + + + + + +
      All Methods Instance Methods Concrete Methods 
      Modifier and TypeMethodDescription
      voidclose() 
      Collection<org.apache.flink.table.data.RowData>lookup​(org.apache.flink.table.data.RowData keyRow) +
      The invoke entry point of lookup function.
      +
      voidopen​(org.apache.flink.table.functions.FunctionContext context) 
      +
        +
      • + + +

        Methods inherited from class org.apache.flink.table.functions.LookupFunction

        +eval
      • +
      +
        +
      • + + +

        Methods inherited from class org.apache.flink.table.functions.TableFunction

        +collect, finish, getKind, getParameterTypes, getResultType, getTypeInference, setCollector
      • +
      +
        +
      • + + +

        Methods inherited from class org.apache.flink.table.functions.UserDefinedFunction

        +functionIdentifier, toString
      • +
      + +
        +
      • + + +

        Methods inherited from interface org.apache.flink.table.functions.FunctionDefinition

        +getRequirements, isDeterministic, supportsConstantFolding
      • +
      +
    • +
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlinkLookupFunction

        +
        public FlinkLookupFunction​(org.apache.fluss.config.Configuration flussConfig,
        +                           org.apache.fluss.metadata.TablePath tablePath,
        +                           org.apache.flink.table.types.logical.RowType flinkRowType,
        +                           int maxRetryTimes,
        +                           LookupNormalizer lookupNormalizer,
        +                           @Nullable
        +                           int[] projection)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        open

        +
        public void open​(org.apache.flink.table.functions.FunctionContext context)
        +
        +
        Overrides:
        +
        open in class org.apache.flink.table.functions.UserDefinedFunction
        +
        +
      • +
      + + + +
        +
      • +

        lookup

        +
        public Collection<org.apache.flink.table.data.RowData> lookup​(org.apache.flink.table.data.RowData keyRow)
        +
        The invoke entry point of lookup function.
        +
        +
        Specified by:
        +
        lookup in class org.apache.flink.table.functions.LookupFunction
        +
        Parameters:
        +
        keyRow - - A RowData that wraps lookup keys. Currently only support single + rowkey.
        +
        +
      • +
      + + + +
        +
      • +

        close

        +
        public void close()
        +           throws Exception
        +
        +
        Overrides:
        +
        close in class org.apache.flink.table.functions.UserDefinedFunction
        +
        Throws:
        +
        Exception
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/LookupNormalizer.RemainingFilter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/LookupNormalizer.RemainingFilter.html new file mode 100644 index 0000000000..0a1ad4e986 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/LookupNormalizer.RemainingFilter.html @@ -0,0 +1,271 @@ + + + + + +LookupNormalizer.RemainingFilter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LookupNormalizer.RemainingFilter

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.lookup.LookupNormalizer.RemainingFilter
    • +
    +
  • +
+
+
    +
  • +
    +
    Enclosing class:
    +
    LookupNormalizer
    +
    +
    +
    public static class LookupNormalizer.RemainingFilter
    +extends Object
    +
    A filter to check if the lookup result matches the remaining conditions.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        isMatch

        +
        public boolean isMatch​(org.apache.flink.table.data.RowData result)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/LookupNormalizer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/LookupNormalizer.html new file mode 100644 index 0000000000..fd9823cec3 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/LookupNormalizer.html @@ -0,0 +1,418 @@ + + + + + +LookupNormalizer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LookupNormalizer

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.lookup.LookupNormalizer
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    Serializable
    +
    +
    +
    public class LookupNormalizer
    +extends Object
    +implements Serializable
    +
    A utility class to normalize the lookup key row to match the Fluss key fields order and drop the + lookup result that doesn't match remaining conditions. + +

    For example, if we have a Fluss table with the following schema: + [id: int, name: string, age: int, score: double] with primary key (name, id). And a lookup + condition dim.id = src.id AND dim.name = src.name AND dim.age = 32. The lookup key + row will be [1001, "Alice", 32]. We need to normalize the lookup key row into + ["Alice", 1001], and construct a remaining filter for {age == 32}.

    +
    +
    See Also:
    +
    Serialized Form
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getLookupType

        +
        public LookupType getLookupType()
        +
        Returns the lookup type (primary key lookup, or prefix key lookup) requested from Flink to + Fluss.
        +
      • +
      + + + +
        +
      • +

        getLookupKeyIndexes

        +
        public int[] getLookupKeyIndexes()
        +
        Returns the indexes of the normalized lookup keys.
        +
      • +
      + + + +
        +
      • +

        normalizeLookupKey

        +
        public org.apache.flink.table.data.RowData normalizeLookupKey​(org.apache.flink.table.data.RowData lookupKey)
        +
        Normalize the lookup key row to match the request key and the key fields order.
        +
      • +
      + + + + + + + +
        +
      • +

        createPrimaryKeyLookupNormalizer

        +
        public static LookupNormalizer createPrimaryKeyLookupNormalizer​(int[] primaryKeys,
        +                                                                org.apache.flink.table.types.logical.RowType schema)
        +
        Create a LookupNormalizer for primary key lookup.
        +
      • +
      + + + +
        +
      • +

        validateAndCreateLookupNormalizer

        +
        public static LookupNormalizer validateAndCreateLookupNormalizer​(int[][] lookupKeyIndexes,
        +                                                                 int[] primaryKeys,
        +                                                                 int[] bucketKeys,
        +                                                                 int[] partitionKeys,
        +                                                                 org.apache.flink.table.types.logical.RowType schema,
        +                                                                 @Nullable
        +                                                                 int[] projectedFields)
        +
        Validate the lookup key indexes and primary keys, and create a LookupNormalizer.
        +
        +
        Parameters:
        +
        lookupKeyIndexes - the indexes of the lookup keys in the table row
        +
        primaryKeys - the indexes of the primary keys of the table
        +
        bucketKeys - the indexes of the bucket keys of the table, must be a part of primary keys
        +
        partitionKeys - the indexes of the partition keys of the table, maybe empty if the table + is not partitioned
        +
        schema - the schema of the table
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/FlinkAsyncLookupFunction.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/FlinkAsyncLookupFunction.html new file mode 100644 index 0000000000..967ef4597f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/FlinkAsyncLookupFunction.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.lookup.FlinkAsyncLookupFunction (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.lookup.FlinkAsyncLookupFunction

+
+
No usage of org.apache.fluss.flink.source.lookup.FlinkAsyncLookupFunction
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/FlinkLookupFunction.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/FlinkLookupFunction.html new file mode 100644 index 0000000000..ead4608f83 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/FlinkLookupFunction.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.lookup.FlinkLookupFunction (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.lookup.FlinkLookupFunction

+
+
No usage of org.apache.fluss.flink.source.lookup.FlinkLookupFunction
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/LookupNormalizer.RemainingFilter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/LookupNormalizer.RemainingFilter.html new file mode 100644 index 0000000000..f30416656f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/LookupNormalizer.RemainingFilter.html @@ -0,0 +1,192 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.lookup.LookupNormalizer.RemainingFilter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.lookup.LookupNormalizer.RemainingFilter

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/LookupNormalizer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/LookupNormalizer.html new file mode 100644 index 0000000000..c45a3480c2 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/LookupNormalizer.html @@ -0,0 +1,234 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.lookup.LookupNormalizer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.lookup.LookupNormalizer

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-summary.html new file mode 100644 index 0000000000..ffa8c63d7a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-summary.html @@ -0,0 +1,187 @@ + + + + + +org.apache.fluss.flink.source.lookup (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.source.lookup

+
+
+
    +
  • + + + + + + + + + + + + + + + + + + + + + + + + +
    Class Summary 
    ClassDescription
    FlinkAsyncLookupFunction +
    A flink async lookup function for fluss.
    +
    FlinkLookupFunction +
    A flink lookup function for fluss.
    +
    LookupNormalizer +
    A utility class to normalize the lookup key row to match the Fluss key fields order and drop the + lookup result that doesn't match remaining conditions.
    +
    LookupNormalizer.RemainingFilter +
    A filter to check if the lookup result matches the remaining conditions.
    +
    +
  • +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-tree.html new file mode 100644 index 0000000000..1d32c1ef8d --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-tree.html @@ -0,0 +1,186 @@ + + + + + +org.apache.fluss.flink.source.lookup Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.source.lookup

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+
    +
  • java.lang.Object +
      +
    • org.apache.fluss.flink.source.lookup.LookupNormalizer (implements java.io.Serializable)
    • +
    • org.apache.fluss.flink.source.lookup.LookupNormalizer.RemainingFilter
    • +
    • org.apache.flink.table.functions.UserDefinedFunction (implements org.apache.flink.table.functions.FunctionDefinition, java.io.Serializable) +
        +
      • org.apache.flink.table.functions.AsyncTableFunction<T> +
          +
        • org.apache.flink.table.functions.AsyncLookupFunction + +
        • +
        +
      • +
      • org.apache.flink.table.functions.TableFunction<T> +
          +
        • org.apache.flink.table.functions.LookupFunction + +
        • +
        +
      • +
      +
    • +
    +
  • +
+
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-use.html new file mode 100644 index 0000000000..0d1586ab58 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-use.html @@ -0,0 +1,192 @@ + + + + + +Uses of Package org.apache.fluss.flink.source.lookup (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.source.lookup

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/FlinkSourceReaderMetrics.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/FlinkSourceReaderMetrics.html new file mode 100644 index 0000000000..2756ea3a52 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/FlinkSourceReaderMetrics.html @@ -0,0 +1,523 @@ + + + + + +FlinkSourceReaderMetrics (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkSourceReaderMetrics

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
    • +
    +
  • +
+
+
    +
  • +
    +
    public class FlinkSourceReaderMetrics
    +extends Object
    +
    A collection class for handling metrics in FlinkSourceReader of Fluss. + +

    All metrics of the source reader are registered under group "fluss.reader", which is a child + group of OperatorMetricGroup. Metrics related to a + specific table bucket will be registered in the group: + +

    "fluss.reader.bucket.{bucket_id}" for non-partitioned bucket or + "fluss.reader.partition.{partition_id}.bucket.{bucket_id}" for partitioned bucket. + +

    For example, current consuming offset of table "my-table" and bucket 1 will be reported in + metric: "{some_parent_groups}.operator.fluss.reader.table.my-table.bucket.1.currentOffset"

    +
  • +
+
+
+ +
+
+
    +
  • + +
    + +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlinkSourceReaderMetrics

        +
        public FlinkSourceReaderMetrics​(org.apache.flink.metrics.groups.SourceReaderMetricGroup sourceReaderMetricGroup)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        reportRecordEventTime

        +
        public void reportRecordEventTime​(long lag)
        +
      • +
      + + + +
        +
      • +

        registerTableBucket

        +
        public void registerTableBucket​(org.apache.fluss.metadata.TableBucket tableBucket)
        +
      • +
      + + + +
        +
      • +

        recordCurrentOffset

        +
        public void recordCurrentOffset​(org.apache.fluss.metadata.TableBucket tb,
        +                                long offset)
        +
        Update current consuming offset of the given TableBucket.
        +
        +
        Parameters:
        +
        tb - Updating table bucket
        +
        offset - Current consuming offset
        +
        +
      • +
      + + + +
        +
      • +

        getSourceReaderMetricGroup

        +
        public org.apache.flink.metrics.groups.SourceReaderMetricGroup getSourceReaderMetricGroup()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/class-use/FlinkSourceReaderMetrics.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/class-use/FlinkSourceReaderMetrics.html new file mode 100644 index 0000000000..ca451b4e20 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/class-use/FlinkSourceReaderMetrics.html @@ -0,0 +1,242 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-summary.html new file mode 100644 index 0000000000..b46b2a0a88 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-summary.html @@ -0,0 +1,168 @@ + + + + + +org.apache.fluss.flink.source.metrics (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.source.metrics

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-tree.html new file mode 100644 index 0000000000..e61cff8b26 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-tree.html @@ -0,0 +1,163 @@ + + + + + +org.apache.fluss.flink.source.metrics Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.source.metrics

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-use.html new file mode 100644 index 0000000000..e5028d7525 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-use.html @@ -0,0 +1,185 @@ + + + + + +Uses of Package org.apache.fluss.flink.source.metrics (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.source.metrics

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-summary.html new file mode 100644 index 0000000000..b97639b357 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-summary.html @@ -0,0 +1,186 @@ + + + + + +org.apache.fluss.flink.source (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.source

+
+
+
    +
  • + + + + + + + + + + + + + + + + + + + + + + + + +
    Class Summary 
    ClassDescription
    FlinkSource<OUT> +
    Flink source for Fluss.
    +
    FlinkTableSource +
    Flink table source to scan Fluss data.
    +
    FlussSource<OUT> +
    A Flink DataStream source implementation for reading data from Fluss tables.
    +
    FlussSourceBuilder<OUT> +
    Builder class for creating FlussSource instances.
    +
    +
  • +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-tree.html new file mode 100644 index 0000000000..2bd9bc2ffc --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-tree.html @@ -0,0 +1,169 @@ + + + + + +org.apache.fluss.flink.source Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.source

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+
    +
  • java.lang.Object +
      +
    • org.apache.fluss.flink.source.FlinkSource<OUT> (implements org.apache.flink.api.java.typeutils.ResultTypeQueryable<T>, org.apache.flink.api.connector.source.Source<T,​SplitT,​EnumChkT>) + +
    • +
    • org.apache.fluss.flink.source.FlinkTableSource (implements org.apache.flink.table.connector.source.LookupTableSource, org.apache.flink.table.connector.source.ScanTableSource, org.apache.flink.table.connector.source.abilities.SupportsAggregatePushDown, org.apache.flink.table.connector.source.abilities.SupportsFilterPushDown, org.apache.flink.table.connector.source.abilities.SupportsLimitPushDown, org.apache.flink.table.connector.source.abilities.SupportsProjectionPushDown, org.apache.flink.table.connector.source.abilities.SupportsRowLevelModificationScan)
    • +
    • org.apache.fluss.flink.source.FlussSourceBuilder<OUT>
    • +
    +
  • +
+
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-use.html new file mode 100644 index 0000000000..7d941008d1 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-use.html @@ -0,0 +1,197 @@ + + + + + +Uses of Package org.apache.fluss.flink.source (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.source

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/BoundedSplitReader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/BoundedSplitReader.html new file mode 100644 index 0000000000..4b736cc8a2 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/BoundedSplitReader.html @@ -0,0 +1,350 @@ + + + + + +BoundedSplitReader (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class BoundedSplitReader

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.reader.BoundedSplitReader
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        BoundedSplitReader

        +
        public BoundedSplitReader​(BatchScanner splitScanner,
        +                          long toSkip)
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkRecordsWithSplitIds.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkRecordsWithSplitIds.html new file mode 100644 index 0000000000..c83920bbd7 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkRecordsWithSplitIds.html @@ -0,0 +1,444 @@ + + + + + +FlinkRecordsWithSplitIds (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkRecordsWithSplitIds

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<RecordAndPos>
    +
    +
    +
    public class FlinkRecordsWithSplitIds
    +extends Object
    +implements org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<RecordAndPos>
    +
    An implementation of RecordsWithSplitIds which contains records from multiple splits.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    + +
    + +
    +
      +
    • + + +

      Method Detail

      + + + + + + + +
        +
      • +

        setTableBucketStoppingOffset

        +
        public void setTableBucketStoppingOffset​(org.apache.fluss.metadata.TableBucket tableBucket,
        +                                         long stoppingOffset)
        +
      • +
      + + + +
        +
      • +

        nextSplit

        +
        @Nullable
        +public String nextSplit()
        +
        +
        Specified by:
        +
        nextSplit in interface org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<RecordAndPos>
        +
        +
      • +
      + + + +
        +
      • +

        nextRecordFromSplit

        +
        @Nullable
        +public RecordAndPos nextRecordFromSplit()
        +
        +
        Specified by:
        +
        nextRecordFromSplit in interface org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<RecordAndPos>
        +
        +
      • +
      + + + +
        +
      • +

        recycle

        +
        public void recycle()
        +
        +
        Specified by:
        +
        recycle in interface org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<RecordAndPos>
        +
        +
      • +
      + + + +
        +
      • +

        finishedSplits

        +
        public Set<String> finishedSplits()
        +
        +
        Specified by:
        +
        finishedSplits in interface org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<RecordAndPos>
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkSourceReader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkSourceReader.html new file mode 100644 index 0000000000..54db07bfc5 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkSourceReader.html @@ -0,0 +1,443 @@ + + + + + +FlinkSourceReader (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkSourceReader<OUT>

+
+
+ +
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Field Summary

      +
        +
      • + + +

        Fields inherited from class org.apache.flink.connector.base.source.reader.SourceReaderBase

        +config, context, eofRecordEvaluator, options, recordEmitter, splitFetcherManager
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Summary

      + + + + + + + + + + +
      Constructors 
      ConstructorDescription
      FlinkSourceReader​(org.apache.flink.connector.base.source.reader.synchronization.FutureCompletingBlockingQueue<org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<RecordAndPos>> elementsQueue, + org.apache.fluss.config.Configuration flussConfig, + org.apache.fluss.metadata.TablePath tablePath, + org.apache.fluss.types.RowType sourceOutputType, + org.apache.flink.api.connector.source.SourceReaderContext context, + int[] projectedFields, + FlinkSourceReaderMetrics flinkSourceReaderMetrics, + FlinkRecordEmitter<OUT> recordEmitter, + org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource) 
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlinkSourceReader

        +
        public FlinkSourceReader​(org.apache.flink.connector.base.source.reader.synchronization.FutureCompletingBlockingQueue<org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<RecordAndPos>> elementsQueue,
        +                         org.apache.fluss.config.Configuration flussConfig,
        +                         org.apache.fluss.metadata.TablePath tablePath,
        +                         org.apache.fluss.types.RowType sourceOutputType,
        +                         org.apache.flink.api.connector.source.SourceReaderContext context,
        +                         @Nullable
        +                         int[] projectedFields,
        +                         FlinkSourceReaderMetrics flinkSourceReaderMetrics,
        +                         FlinkRecordEmitter<OUT> recordEmitter,
        +                         org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource)
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkSourceSplitReader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkSourceSplitReader.html new file mode 100644 index 0000000000..09017063a3 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkSourceSplitReader.html @@ -0,0 +1,411 @@ + + + + + +FlinkSourceSplitReader (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkSourceSplitReader

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.reader.FlinkSourceSplitReader
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlinkSourceSplitReader

        +
        public FlinkSourceSplitReader​(org.apache.fluss.config.Configuration flussConf,
        +                              org.apache.fluss.metadata.TablePath tablePath,
        +                              org.apache.fluss.types.RowType sourceOutputType,
        +                              @Nullable
        +                              int[] projectedFields,
        +                              FlinkSourceReaderMetrics flinkSourceReaderMetrics,
        +                              @Nullable
        +                              org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + + + + + +
        +
      • +

        handleSplitsChanges

        +
        public void handleSplitsChanges​(org.apache.flink.connector.base.source.reader.splitreader.SplitsChange<SourceSplitBase> splitsChanges)
        +
        +
        Specified by:
        +
        handleSplitsChanges in interface org.apache.flink.connector.base.source.reader.splitreader.SplitReader<RecordAndPos,​SourceSplitBase>
        +
        +
      • +
      + + + +
        +
      • +

        removePartitions

        +
        public Set<org.apache.fluss.metadata.TableBucket> removePartitions​(Map<Long,​String> removedPartitions)
        +
      • +
      + + + +
        +
      • +

        wakeUp

        +
        public void wakeUp()
        +
        +
        Specified by:
        +
        wakeUp in interface org.apache.flink.connector.base.source.reader.splitreader.SplitReader<RecordAndPos,​SourceSplitBase>
        +
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/MutableRecordAndPos.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/MutableRecordAndPos.html new file mode 100644 index 0000000000..b2faec87e0 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/MutableRecordAndPos.html @@ -0,0 +1,359 @@ + + + + + +MutableRecordAndPos (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class MutableRecordAndPos

+
+
+ +
+
    +
  • +
    +
    public class MutableRecordAndPos
    +extends RecordAndPos
    +
    A mutable version of the RecordAndPos. + +

    This mutable object is useful in cases where only once instance of a RecordAndPos is + needed at a time.

    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        MutableRecordAndPos

        +
        public MutableRecordAndPos()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        setRecord

        +
        public void setRecord​(ScanRecord scanRecord,
        +                      long readRecordsCount)
        +
      • +
      + + + +
        +
      • +

        setRecord

        +
        public void setRecord​(ScanRecord scanRecord,
        +                      long readRecordsCount,
        +                      int currentIteratorIndex)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/RecordAndPos.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/RecordAndPos.html new file mode 100644 index 0000000000..160e6cada2 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/RecordAndPos.html @@ -0,0 +1,544 @@ + + + + + +RecordAndPos (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class RecordAndPos

+
+
+ +
+
    +
  • +
    +
    Direct Known Subclasses:
    +
    MutableRecordAndPos
    +
    +
    +
    public class RecordAndPos
    +extends Object
    +
    A record wrapping a Fluss ScanRecord, the readRecordsCount when the record is + from reading snapshot, the code currentSplitIndex where the record is from when the Flink + source split read by split reader contains multiple splits(splittable unit), like LakeSnapshotAndFlussLogSplit contains multiple LakeSplits. + +

    When the record is from reading log, readRecordsCount will always be NO_READ_RECORDS_COUNT. + +

    The readRecordsCount defines the point in the snapshot reader AFTER the record. Record + processing and updating checkpointed state happens atomically. The position points to where the + reader should resume after this record is processed. + +

    For example, the very first record in a snapshot split could have a readRecordsCount + of one. + +

    It's produced by FlinkSourceSplitReader and emitted to FlinkRecordEmitter

    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + +
        +
      • +

        NO_READ_RECORDS_COUNT

        +
        public static final long NO_READ_RECORDS_COUNT
        +
        +
        See Also:
        +
        Constant Field Values
        +
        +
      • +
      + + + +
        +
      • +

        DEFAULT_SPLIT_INDEX

        +
        protected static final int DEFAULT_SPLIT_INDEX
        +
        +
        See Also:
        +
        Constant Field Values
        +
        +
      • +
      + + + +
        +
      • +

        scanRecord

        +
        protected ScanRecord scanRecord
        +
      • +
      + + + +
        +
      • +

        readRecordsCount

        +
        protected long readRecordsCount
        +
      • +
      + + + +
        +
      • +

        currentSplitIndex

        +
        protected int currentSplitIndex
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        RecordAndPos

        +
        public RecordAndPos​(ScanRecord scanRecord)
        +
      • +
      + + + +
        +
      • +

        RecordAndPos

        +
        public RecordAndPos​(ScanRecord scanRecord,
        +                    long readRecordsCount)
        +
      • +
      + + + +
        +
      • +

        RecordAndPos

        +
        public RecordAndPos​(ScanRecord scanRecord,
        +                    long readRecordsCount,
        +                    int currentSplitIndex)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        readRecordsCount

        +
        public long readRecordsCount()
        +
      • +
      + + + +
        +
      • +

        getCurrentSplitIndex

        +
        public int getCurrentSplitIndex()
        +
      • +
      + + + + + + + +
        +
      • +

        equals

        +
        public boolean equals​(Object o)
        +
        +
        Overrides:
        +
        equals in class Object
        +
        +
      • +
      + + + +
        +
      • +

        hashCode

        +
        public int hashCode()
        +
        +
        Overrides:
        +
        hashCode in class Object
        +
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/SplitScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/SplitScanner.html new file mode 100644 index 0000000000..0193fa904f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/SplitScanner.html @@ -0,0 +1,269 @@ + + + + + +SplitScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface SplitScanner

+
+
+
+
    +
  • +
    +
    All Superinterfaces:
    +
    AutoCloseable
    +
    +
    +
    public interface SplitScanner
    +extends AutoCloseable
    +
    A scanner to scan split to records .
    +
  • +
+
+
+ +
+
+ +
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/BoundedSplitReader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/BoundedSplitReader.html new file mode 100644 index 0000000000..cb8dd8605a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/BoundedSplitReader.html @@ -0,0 +1,192 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.reader.BoundedSplitReader (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.reader.BoundedSplitReader

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkRecordsWithSplitIds.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkRecordsWithSplitIds.html new file mode 100644 index 0000000000..4c98b94075 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkRecordsWithSplitIds.html @@ -0,0 +1,192 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkSourceReader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkSourceReader.html new file mode 100644 index 0000000000..834663d646 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkSourceReader.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.reader.FlinkSourceReader (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.reader.FlinkSourceReader

+
+
No usage of org.apache.fluss.flink.source.reader.FlinkSourceReader
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkSourceSplitReader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkSourceSplitReader.html new file mode 100644 index 0000000000..a145442292 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkSourceSplitReader.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.reader.FlinkSourceSplitReader (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.reader.FlinkSourceSplitReader

+
+
No usage of org.apache.fluss.flink.source.reader.FlinkSourceSplitReader
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/MutableRecordAndPos.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/MutableRecordAndPos.html new file mode 100644 index 0000000000..e0d8dd985c --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/MutableRecordAndPos.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.reader.MutableRecordAndPos (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.reader.MutableRecordAndPos

+
+
No usage of org.apache.fluss.flink.source.reader.MutableRecordAndPos
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/RecordAndPos.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/RecordAndPos.html new file mode 100644 index 0000000000..9692165f7b --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/RecordAndPos.html @@ -0,0 +1,359 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.reader.RecordAndPos (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.reader.RecordAndPos

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/SplitScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/SplitScanner.html new file mode 100644 index 0000000000..b6fdc28a3c --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/SplitScanner.html @@ -0,0 +1,148 @@ + + + + + +Uses of Interface org.apache.fluss.flink.source.reader.SplitScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.flink.source.reader.SplitScanner

+
+
No usage of org.apache.fluss.flink.source.reader.SplitScanner
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/FlinkSourceFetcherManager.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/FlinkSourceFetcherManager.html new file mode 100644 index 0000000000..bda5e4e9a5 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/FlinkSourceFetcherManager.html @@ -0,0 +1,373 @@ + + + + + +FlinkSourceFetcherManager (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkSourceFetcherManager

+
+
+ +
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Field Summary

      +
        +
      • + + +

        Fields inherited from class org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager

        +fetchers
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Summary

      + + + + + + + + + + +
      Constructors 
      ConstructorDescription
      FlinkSourceFetcherManager​(org.apache.flink.connector.base.source.reader.synchronization.FutureCompletingBlockingQueue<org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<RecordAndPos>> elementsQueue, + Supplier<org.apache.flink.connector.base.source.reader.splitreader.SplitReader<RecordAndPos,​SourceSplitBase>> splitReaderSupplier, + Consumer<Collection<String>> splitFinishedHook) +
      Creates a new SplitFetcherManager with a single I/O threads.
      +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Summary

      + + + + + + + + + + + + +
      All Methods Instance Methods Concrete Methods 
      Modifier and TypeMethodDescription
      voidremovePartitions​(Map<Long,​String> removedPartitions, + Consumer<Set<org.apache.fluss.metadata.TableBucket>> unsubscribeTableBucketsCallback) 
      +
        +
      • + + +

        Methods inherited from class org.apache.flink.connector.base.source.reader.fetcher.SingleThreadFetcherManager

        +addSplits, getRunningFetcher, removeSplits
      • +
      +
        +
      • + + +

        Methods inherited from class org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager

        +checkErrors, close, createSplitFetcher, getNumAliveFetchers, getQueue, maybeShutdownFinishedFetchers, pauseOrResumeSplits, startFetcher
      • +
      + +
    • +
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlinkSourceFetcherManager

        +
        public FlinkSourceFetcherManager​(org.apache.flink.connector.base.source.reader.synchronization.FutureCompletingBlockingQueue<org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<RecordAndPos>> elementsQueue,
        +                                 Supplier<org.apache.flink.connector.base.source.reader.splitreader.SplitReader<RecordAndPos,​SourceSplitBase>> splitReaderSupplier,
        +                                 Consumer<Collection<String>> splitFinishedHook)
        +
        Creates a new SplitFetcherManager with a single I/O threads.
        +
        +
        Parameters:
        +
        elementsQueue - The queue that is used to hand over data from the I/O thread (the + fetchers) to the reader (which emits the records and book-keeps the state. This must be + the same queue instance that is also passed to the SourceReaderBase.
        +
        splitReaderSupplier - The factory for the split reader that connects to the source + system.
        +
        splitFinishedHook - Hook for handling finished splits in split fetchers.
        +
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        removePartitions

        +
        public void removePartitions​(Map<Long,​String> removedPartitions,
        +                             Consumer<Set<org.apache.fluss.metadata.TableBucket>> unsubscribeTableBucketsCallback)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/class-use/FlinkSourceFetcherManager.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/class-use/FlinkSourceFetcherManager.html new file mode 100644 index 0000000000..141d952b72 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/class-use/FlinkSourceFetcherManager.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.reader.fetcher.FlinkSourceFetcherManager (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.reader.fetcher.FlinkSourceFetcherManager

+
+
No usage of org.apache.fluss.flink.source.reader.fetcher.FlinkSourceFetcherManager
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-summary.html new file mode 100644 index 0000000000..f1cdd0eb29 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-summary.html @@ -0,0 +1,168 @@ + + + + + +org.apache.fluss.flink.source.reader.fetcher (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.source.reader.fetcher

+
+
+
    +
  • + + + + + + + + + + + + +
    Class Summary 
    ClassDescription
    FlinkSourceFetcherManager +
    The SplitFetcherManager for Fluss source.
    +
    +
  • +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-tree.html new file mode 100644 index 0000000000..cf0935b731 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-tree.html @@ -0,0 +1,175 @@ + + + + + +org.apache.fluss.flink.source.reader.fetcher Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.source.reader.fetcher

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+
    +
  • java.lang.Object +
      +
    • org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager<E,​SplitT> + +
    • +
    +
  • +
+
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-use.html new file mode 100644 index 0000000000..8045c94b9d --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-use.html @@ -0,0 +1,148 @@ + + + + + +Uses of Package org.apache.fluss.flink.source.reader.fetcher (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.source.reader.fetcher

+
+
No usage of org.apache.fluss.flink.source.reader.fetcher
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-summary.html new file mode 100644 index 0000000000..73fd8a869d --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-summary.html @@ -0,0 +1,217 @@ + + + + + +org.apache.fluss.flink.source.reader (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.source.reader

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-tree.html new file mode 100644 index 0000000000..9f51ef1381 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-tree.html @@ -0,0 +1,193 @@ + + + + + +org.apache.fluss.flink.source.reader Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.source.reader

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+
    +
  • java.lang.Object +
      +
    • org.apache.fluss.flink.source.reader.BoundedSplitReader (implements java.lang.AutoCloseable)
    • +
    • org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds (implements org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<E>)
    • +
    • org.apache.fluss.flink.source.reader.FlinkSourceSplitReader (implements org.apache.flink.connector.base.source.reader.splitreader.SplitReader<E,​SplitT>)
    • +
    • org.apache.fluss.flink.source.reader.RecordAndPos + +
    • +
    • org.apache.flink.connector.base.source.reader.SourceReaderBase<E,​T,​SplitT,​SplitStateT> (implements org.apache.flink.api.connector.source.SourceReader<T,​SplitT>) + +
    • +
    +
  • +
+
+
+

Interface Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-use.html new file mode 100644 index 0000000000..3ae499b518 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-use.html @@ -0,0 +1,274 @@ + + + + + +Uses of Package org.apache.fluss.flink.source.reader (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.source.reader

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/HybridSnapshotLogSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/HybridSnapshotLogSplit.html new file mode 100644 index 0000000000..90a99a24bc --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/HybridSnapshotLogSplit.html @@ -0,0 +1,476 @@ + + + + + +HybridSnapshotLogSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class HybridSnapshotLogSplit

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.api.connector.source.SourceSplit
    +
    +
    +
    public class HybridSnapshotLogSplit
    +extends SnapshotSplit
    +
    The hybrid split for first reading the snapshot files and then switch to read the cdc log from a + specified offset. + +

    Only used for primary key table which will be of snapshot phase and incremental phase of + reading.

    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        HybridSnapshotLogSplit

        +
        public HybridSnapshotLogSplit​(org.apache.fluss.metadata.TableBucket tableBucket,
        +                              @Nullable
        +                              String partitionName,
        +                              long snapshotId,
        +                              long logStartingOffset)
        +
      • +
      + + + +
        +
      • +

        HybridSnapshotLogSplit

        +
        public HybridSnapshotLogSplit​(org.apache.fluss.metadata.TableBucket tableBucket,
        +                              @Nullable
        +                              String partitionName,
        +                              long snapshotId,
        +                              long recordsToSkip,
        +                              boolean isSnapshotFinished,
        +                              long logStartingOffset)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getLogStartingOffset

        +
        public long getLogStartingOffset()
        +
      • +
      + + + +
        +
      • +

        isSnapshotFinished

        +
        public boolean isSnapshotFinished()
        +
      • +
      + + + +
        +
      • +

        splitId

        +
        public String splitId()
        +
      • +
      + + + + + + + + + + + + +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/HybridSnapshotLogSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/HybridSnapshotLogSplitState.html new file mode 100644 index 0000000000..9a8c5bf285 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/HybridSnapshotLogSplitState.html @@ -0,0 +1,368 @@ + + + + + +HybridSnapshotLogSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class HybridSnapshotLogSplitState

+
+
+ +
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        HybridSnapshotLogSplitState

        +
        public HybridSnapshotLogSplitState​(HybridSnapshotLogSplit hybridSnapshotLogSplit)
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/LogSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/LogSplit.html new file mode 100644 index 0000000000..62f212715b --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/LogSplit.html @@ -0,0 +1,501 @@ + + + + + +LogSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LogSplit

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.api.connector.source.SourceSplit
    +
    +
    +
    public class LogSplit
    +extends SourceSplitBase
    +
    The split for log. It's used to describe the log data of a table bucket.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    + +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LogSplit

        +
        public LogSplit​(org.apache.fluss.metadata.TableBucket tableBucket,
        +                @Nullable
        +                String partitionName,
        +                long startingOffset)
        +
      • +
      + + + +
        +
      • +

        LogSplit

        +
        public LogSplit​(org.apache.fluss.metadata.TableBucket tableBucket,
        +                @Nullable
        +                String partitionName,
        +                long startingOffset,
        +                long stoppingOffset)
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/LogSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/LogSplitState.html new file mode 100644 index 0000000000..443bfd3ceb --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/LogSplitState.html @@ -0,0 +1,354 @@ + + + + + +LogSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LogSplitState

+
+
+ +
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LogSplitState

        +
        public LogSplitState​(LogSplit split)
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SnapshotSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SnapshotSplit.html new file mode 100644 index 0000000000..16e1d08c70 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SnapshotSplit.html @@ -0,0 +1,496 @@ + + + + + +SnapshotSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class SnapshotSplit

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.api.connector.source.SourceSplit
    +
    +
    +
    Direct Known Subclasses:
    +
    HybridSnapshotLogSplit
    +
    +
    +
    public abstract class SnapshotSplit
    +extends SourceSplitBase
    +
    The split for snapshot. It's used to describe a snapshot of a table bucket.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + +
        +
      • +

        recordsToSkip

        +
        protected final long recordsToSkip
        +
        The records to skip when reading the snapshot.
        +
      • +
      + + + +
        +
      • +

        snapshotId

        +
        protected final long snapshotId
        +
        The snapshot id. It's used to identify the snapshot for a kv bucket.
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        SnapshotSplit

        +
        public SnapshotSplit​(org.apache.fluss.metadata.TableBucket tableBucket,
        +                     @Nullable
        +                     String partitionName,
        +                     long snapshotId,
        +                     long recordsToSkip)
        +
      • +
      + + + +
        +
      • +

        SnapshotSplit

        +
        public SnapshotSplit​(org.apache.fluss.metadata.TableBucket tableBucket,
        +                     @Nullable
        +                     String partitionName,
        +                     long snapshotId)
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitBase.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitBase.html new file mode 100644 index 0000000000..0ebe839fa9 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitBase.html @@ -0,0 +1,582 @@ + + + + + +SourceSplitBase (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class SourceSplitBase

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.split.SourceSplitBase
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + +
        +
      • +

        HYBRID_SNAPSHOT_SPLIT_FLAG

        +
        public static final byte HYBRID_SNAPSHOT_SPLIT_FLAG
        +
        +
        See Also:
        +
        Constant Field Values
        +
        +
      • +
      + + + + + + + +
        +
      • +

        tableBucket

        +
        protected final org.apache.fluss.metadata.TableBucket tableBucket
        +
      • +
      + + + +
        +
      • +

        partitionName

        +
        @Nullable
        +protected final String partitionName
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        SourceSplitBase

        +
        public SourceSplitBase​(org.apache.fluss.metadata.TableBucket tableBucket,
        +                       @Nullable
        +                       String partitionName)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        toSplitId

        +
        protected static String toSplitId​(String splitPrefix,
        +                                  org.apache.fluss.metadata.TableBucket tableBucket)
        +
      • +
      + + + +
        +
      • +

        getTableBucket

        +
        public org.apache.fluss.metadata.TableBucket getTableBucket()
        +
      • +
      + + + +
        +
      • +

        getPartitionName

        +
        @Nullable
        +public String getPartitionName()
        +
      • +
      + + + +
        +
      • +

        isLogSplit

        +
        public final boolean isLogSplit()
        +
        Checks whether this split is a log split.
        +
      • +
      + + + +
        +
      • +

        isLakeSplit

        +
        public boolean isLakeSplit()
        +
      • +
      + + + +
        +
      • +

        isHybridSnapshotLogSplit

        +
        public final boolean isHybridSnapshotLogSplit()
        +
        Checks whether this split is a hybrid snapshot and log split.
        +
      • +
      + + + + + + + +
        +
      • +

        asLogSplit

        +
        public final LogSplit asLogSplit()
        +
        Casts this split into a LogSplit.
        +
      • +
      + + + +
        +
      • +

        splitKind

        +
        protected byte splitKind()
        +
      • +
      + + + +
        +
      • +

        equals

        +
        public boolean equals​(Object o)
        +
        +
        Overrides:
        +
        equals in class Object
        +
        +
      • +
      + + + +
        +
      • +

        hashCode

        +
        public int hashCode()
        +
        +
        Overrides:
        +
        hashCode in class Object
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitSerializer.html new file mode 100644 index 0000000000..6d6f0e9736 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitSerializer.html @@ -0,0 +1,362 @@ + + + + + +SourceSplitSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class SourceSplitSerializer

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.split.SourceSplitSerializer
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.core.io.SimpleVersionedSerializer<SourceSplitBase>, org.apache.flink.core.io.Versioned
    +
    +
    +
    public class SourceSplitSerializer
    +extends Object
    +implements org.apache.flink.core.io.SimpleVersionedSerializer<SourceSplitBase>
    +
    A serializer for the SourceSplitBase.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        SourceSplitSerializer

        +
        public SourceSplitSerializer​(org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getVersion

        +
        public int getVersion()
        +
        +
        Specified by:
        +
        getVersion in interface org.apache.flink.core.io.SimpleVersionedSerializer<SourceSplitBase>
        +
        Specified by:
        +
        getVersion in interface org.apache.flink.core.io.Versioned
        +
        +
      • +
      + + + + + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitState.html new file mode 100644 index 0000000000..400493ba79 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitState.html @@ -0,0 +1,435 @@ + + + + + +SourceSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class SourceSplitState

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.split.SourceSplitState
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    + +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        SourceSplitState

        +
        public SourceSplitState​(SourceSplitBase split)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        isHybridSnapshotLogSplitState

        +
        public final boolean isHybridSnapshotLogSplitState()
        +
        Checks whether this split state is a hybrid snapshot log split state.
        +
      • +
      + + + +
        +
      • +

        isLogSplitState

        +
        public final boolean isLogSplitState()
        +
        Checks whether this split state is a log split state.
        +
      • +
      + + + + + + + + + + + + + + + +
        +
      • +

        isLakeSplit

        +
        public boolean isLakeSplit()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/HybridSnapshotLogSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/HybridSnapshotLogSplit.html new file mode 100644 index 0000000000..c080edbace --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/HybridSnapshotLogSplit.html @@ -0,0 +1,212 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.split.HybridSnapshotLogSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.split.HybridSnapshotLogSplit

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/HybridSnapshotLogSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/HybridSnapshotLogSplitState.html new file mode 100644 index 0000000000..ea807e9b17 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/HybridSnapshotLogSplitState.html @@ -0,0 +1,194 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.split.HybridSnapshotLogSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.split.HybridSnapshotLogSplitState

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/LogSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/LogSplit.html new file mode 100644 index 0000000000..351bbac144 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/LogSplit.html @@ -0,0 +1,212 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.split.LogSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.split.LogSplit

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/LogSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/LogSplitState.html new file mode 100644 index 0000000000..d8da50545d --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/LogSplitState.html @@ -0,0 +1,194 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.split.LogSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.split.LogSplitState

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SnapshotSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SnapshotSplit.html new file mode 100644 index 0000000000..af40d11ab9 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SnapshotSplit.html @@ -0,0 +1,195 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.split.SnapshotSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.split.SnapshotSplit

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitBase.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitBase.html new file mode 100644 index 0000000000..f8213f5101 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitBase.html @@ -0,0 +1,733 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.split.SourceSplitBase (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.split.SourceSplitBase

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitSerializer.html new file mode 100644 index 0000000000..89a1e988d1 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitSerializer.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.split.SourceSplitSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.split.SourceSplitSerializer

+
+
No usage of org.apache.fluss.flink.source.split.SourceSplitSerializer
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitState.html new file mode 100644 index 0000000000..d9656dc70c --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitState.html @@ -0,0 +1,364 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.split.SourceSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.split.SourceSplitState

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-summary.html new file mode 100644 index 0000000000..3a1363e0c8 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-summary.html @@ -0,0 +1,211 @@ + + + + + +org.apache.fluss.flink.source.split (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.source.split

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-tree.html new file mode 100644 index 0000000000..b1b4a7559f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-tree.html @@ -0,0 +1,179 @@ + + + + + +org.apache.fluss.flink.source.split Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.source.split

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-use.html new file mode 100644 index 0000000000..d7054e5819 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-use.html @@ -0,0 +1,447 @@ + + + + + +Uses of Package org.apache.fluss.flink.source.split (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.source.split

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/FlussSourceEnumeratorStateSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/FlussSourceEnumeratorStateSerializer.html new file mode 100644 index 0000000000..59013f3387 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/FlussSourceEnumeratorStateSerializer.html @@ -0,0 +1,362 @@ + + + + + +FlussSourceEnumeratorStateSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlussSourceEnumeratorStateSerializer

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.state.FlussSourceEnumeratorStateSerializer
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.core.io.SimpleVersionedSerializer<SourceEnumeratorState>, org.apache.flink.core.io.Versioned
    +
    +
    +
    public class FlussSourceEnumeratorStateSerializer
    +extends Object
    +implements org.apache.flink.core.io.SimpleVersionedSerializer<SourceEnumeratorState>
    +
    A serializer for SourceEnumeratorState.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlussSourceEnumeratorStateSerializer

        +
        public FlussSourceEnumeratorStateSerializer​(org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource)
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/SourceEnumeratorState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/SourceEnumeratorState.html new file mode 100644 index 0000000000..f95f9ba6cf --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/SourceEnumeratorState.html @@ -0,0 +1,395 @@ + + + + + +SourceEnumeratorState (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class SourceEnumeratorState

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.source.state.SourceEnumeratorState
    • +
    +
  • +
+
+
    +
  • +
    +
    public class SourceEnumeratorState
    +extends Object
    +
    A checkpoint of the current state of the containing the buckets that is already assigned.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        SourceEnumeratorState

        +
        public SourceEnumeratorState​(Set<org.apache.fluss.metadata.TableBucket> assignedBuckets,
        +                             Map<Long,​String> assignedPartitions,
        +                             @Nullable
        +                             List<SourceSplitBase> remainingHybridLakeFlussSplits)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getAssignedBuckets

        +
        public Set<org.apache.fluss.metadata.TableBucket> getAssignedBuckets()
        +
      • +
      + + + +
        +
      • +

        getAssignedPartitions

        +
        public Map<Long,​String> getAssignedPartitions()
        +
      • +
      + + + +
        +
      • +

        getRemainingHybridLakeFlussSplits

        +
        @Nullable
        +public List<SourceSplitBase> getRemainingHybridLakeFlussSplits()
        +
      • +
      + + + +
        +
      • +

        equals

        +
        public boolean equals​(Object o)
        +
        +
        Overrides:
        +
        equals in class Object
        +
        +
      • +
      + + + +
        +
      • +

        hashCode

        +
        public int hashCode()
        +
        +
        Overrides:
        +
        hashCode in class Object
        +
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/class-use/FlussSourceEnumeratorStateSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/class-use/FlussSourceEnumeratorStateSerializer.html new file mode 100644 index 0000000000..b69d118611 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/class-use/FlussSourceEnumeratorStateSerializer.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.state.FlussSourceEnumeratorStateSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.state.FlussSourceEnumeratorStateSerializer

+
+
No usage of org.apache.fluss.flink.source.state.FlussSourceEnumeratorStateSerializer
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/class-use/SourceEnumeratorState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/class-use/SourceEnumeratorState.html new file mode 100644 index 0000000000..81c4c57e16 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/class-use/SourceEnumeratorState.html @@ -0,0 +1,287 @@ + + + + + +Uses of Class org.apache.fluss.flink.source.state.SourceEnumeratorState (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.source.state.SourceEnumeratorState

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-summary.html new file mode 100644 index 0000000000..a479c5f931 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-summary.html @@ -0,0 +1,174 @@ + + + + + +org.apache.fluss.flink.source.state (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.source.state

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-tree.html new file mode 100644 index 0000000000..632b79ba81 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-tree.html @@ -0,0 +1,164 @@ + + + + + +org.apache.fluss.flink.source.state Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.source.state

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-use.html new file mode 100644 index 0000000000..b5005bce1b --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-use.html @@ -0,0 +1,231 @@ + + + + + +Uses of Package org.apache.fluss.flink.source.state (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.source.state

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/LakeTieringJobBuilder.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/LakeTieringJobBuilder.html new file mode 100644 index 0000000000..e5c55dbcf5 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/LakeTieringJobBuilder.html @@ -0,0 +1,292 @@ + + + + + +LakeTieringJobBuilder (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LakeTieringJobBuilder

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.tiering.LakeTieringJobBuilder
    • +
    +
  • +
+
+
    +
  • +
    +
    public class LakeTieringJobBuilder
    +extends Object
    +
    The builder to build Flink lake tiering job.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        newBuilder

        +
        public static LakeTieringJobBuilder newBuilder​(org.apache.flink.streaming.api.environment.StreamExecutionEnvironment env,
        +                                               org.apache.fluss.config.Configuration flussConfig,
        +                                               org.apache.fluss.config.Configuration dataLakeConfig,
        +                                               String dataLakeFormat)
        +
      • +
      + + + +
        +
      • +

        build

        +
        public org.apache.flink.core.execution.JobClient build()
        +                                                throws Exception
        +
        +
        Throws:
        +
        Exception
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/class-use/LakeTieringJobBuilder.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/class-use/LakeTieringJobBuilder.html new file mode 100644 index 0000000000..2ceba7bc0d --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/class-use/LakeTieringJobBuilder.html @@ -0,0 +1,195 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.LakeTieringJobBuilder (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.LakeTieringJobBuilder

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/CommittableMessage.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/CommittableMessage.html new file mode 100644 index 0000000000..be474eaf01 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/CommittableMessage.html @@ -0,0 +1,318 @@ + + + + + +CommittableMessage (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class CommittableMessage<Committable>

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.tiering.committer.CommittableMessage<Committable>
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + + + +
        +
      • +

        CommittableMessage

        +
        public CommittableMessage​(Committable committable)
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/CommittableMessageTypeInfo.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/CommittableMessageTypeInfo.html new file mode 100644 index 0000000000..3565d859ca --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/CommittableMessageTypeInfo.html @@ -0,0 +1,485 @@ + + + + + +CommittableMessageTypeInfo (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class CommittableMessageTypeInfo<Committable>

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.flink.api.common.typeinfo.TypeInformation<CommittableMessage<Committable>>
    • +
    • +
        +
      • org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo<Committable>
      • +
      +
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        of

        +
        public static <Committable> org.apache.flink.api.common.typeinfo.TypeInformation<CommittableMessage<Committable>> of​(org.apache.flink.util.function.SerializableSupplier<org.apache.fluss.lake.serializer.SimpleVersionedSerializer<Committable>> committableSerializerFactory)
        +
      • +
      + + + +
        +
      • +

        isBasicType

        +
        public boolean isBasicType()
        +
        +
        Specified by:
        +
        isBasicType in class org.apache.flink.api.common.typeinfo.TypeInformation<CommittableMessage<Committable>>
        +
        +
      • +
      + + + +
        +
      • +

        isTupleType

        +
        public boolean isTupleType()
        +
        +
        Specified by:
        +
        isTupleType in class org.apache.flink.api.common.typeinfo.TypeInformation<CommittableMessage<Committable>>
        +
        +
      • +
      + + + +
        +
      • +

        getArity

        +
        public int getArity()
        +
        +
        Specified by:
        +
        getArity in class org.apache.flink.api.common.typeinfo.TypeInformation<CommittableMessage<Committable>>
        +
        +
      • +
      + + + +
        +
      • +

        getTotalFields

        +
        public int getTotalFields()
        +
        +
        Specified by:
        +
        getTotalFields in class org.apache.flink.api.common.typeinfo.TypeInformation<CommittableMessage<Committable>>
        +
        +
      • +
      + + + + + + + +
        +
      • +

        isKeyType

        +
        public boolean isKeyType()
        +
        +
        Specified by:
        +
        isKeyType in class org.apache.flink.api.common.typeinfo.TypeInformation<CommittableMessage<Committable>>
        +
        +
      • +
      + + + +
        +
      • +

        createSerializer

        +
        public org.apache.flink.api.common.typeutils.TypeSerializer<CommittableMessage<Committable>> createSerializer​(org.apache.flink.api.common.ExecutionConfig executionConfig)
        +
        +
        Specified by:
        +
        createSerializer in class org.apache.flink.api.common.typeinfo.TypeInformation<CommittableMessage<Committable>>
        +
        +
      • +
      + + + + + + + +
        +
      • +

        equals

        +
        public boolean equals​(Object obj)
        +
        +
        Specified by:
        +
        equals in class org.apache.flink.api.common.typeinfo.TypeInformation<CommittableMessage<Committable>>
        +
        +
      • +
      + + + +
        +
      • +

        hashCode

        +
        public int hashCode()
        +
        +
        Specified by:
        +
        hashCode in class org.apache.flink.api.common.typeinfo.TypeInformation<CommittableMessage<Committable>>
        +
        +
      • +
      + + + +
        +
      • +

        canEqual

        +
        public boolean canEqual​(Object obj)
        +
        +
        Specified by:
        +
        canEqual in class org.apache.flink.api.common.typeinfo.TypeInformation<CommittableMessage<Committable>>
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/FlussTableLakeSnapshotCommitter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/FlussTableLakeSnapshotCommitter.html new file mode 100644 index 0000000000..fd5c357406 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/FlussTableLakeSnapshotCommitter.html @@ -0,0 +1,354 @@ + + + + + +FlussTableLakeSnapshotCommitter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlussTableLakeSnapshotCommitter

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.tiering.committer.FlussTableLakeSnapshotCommitter
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    AutoCloseable
    +
    +
    +
    public class FlussTableLakeSnapshotCommitter
    +extends Object
    +implements AutoCloseable
    +
    Committer to commit FlussTableLakeSnapshot of lake to Fluss.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlussTableLakeSnapshotCommitter

        +
        public FlussTableLakeSnapshotCommitter​(org.apache.fluss.config.Configuration flussConf)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        open

        +
        public void open()
        +
      • +
      + + + +
        +
      • +

        commit

        +
        public void commit​(long tableId,
        +                   org.apache.fluss.lake.committer.CommittedLakeSnapshot committedLakeSnapshot)
        +            throws IOException
        +
        +
        Throws:
        +
        IOException
        +
        +
      • +
      + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitOperator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitOperator.html new file mode 100644 index 0000000000..955d05fd87 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitOperator.html @@ -0,0 +1,499 @@ + + + + + +TieringCommitOperator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TieringCommitOperator<WriteResult,​Committable>

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.flink.streaming.api.operators.AbstractStreamOperator<CommittableMessage<Committable>>
    • +
    • +
        +
      • org.apache.fluss.flink.tiering.committer.TieringCommitOperator<WriteResult,​Committable>
      • +
      +
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    Serializable, org.apache.flink.api.common.state.CheckpointListener, org.apache.flink.streaming.api.operators.Input<TableBucketWriteResult<WriteResult>>, org.apache.flink.streaming.api.operators.KeyContext, org.apache.flink.streaming.api.operators.KeyContextHandler, org.apache.flink.streaming.api.operators.OneInputStreamOperator<TableBucketWriteResult<WriteResult>,​CommittableMessage<Committable>>, org.apache.flink.streaming.api.operators.SetupableStreamOperator<CommittableMessage<Committable>>, org.apache.flink.streaming.api.operators.StreamOperator<CommittableMessage<Committable>>, org.apache.flink.streaming.api.operators.StreamOperatorStateHandler.CheckpointedStreamOperator, org.apache.flink.streaming.api.operators.YieldingOperator<CommittableMessage<Committable>>
    +
    +
    +
    public class TieringCommitOperator<WriteResult,​Committable>
    +extends org.apache.flink.streaming.api.operators.AbstractStreamOperator<CommittableMessage<Committable>>
    +implements org.apache.flink.streaming.api.operators.OneInputStreamOperator<TableBucketWriteResult<WriteResult>,​CommittableMessage<Committable>>
    +
    A Flink operator to aggregate TieringCommitOperators by table to TieringCommitOperator which will + then be committed to lake & Fluss cluster. + +

    It will collect all TableBucketWriteResults which wraps TieringCommitOperator written by + LakeWriter in TieringSource operator. + +

    When it collects all TableBucketWriteResults of a round of tiering for a table, it + will combine all the TieringCommitOperators to TieringCommitOperator via method LakeCommitter.toCommittable(List), and then call method LakeCommitter.commit(Object, + Map) to commit to lake. + +

    Finally, it will also commit the committed lake snapshot to Fluss cluster to make Fluss aware + of the tiering progress.

    +
    +
    See Also:
    +
    Serialized Form
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Field Summary

      +
        +
      • + + +

        Fields inherited from class org.apache.flink.streaming.api.operators.AbstractStreamOperator

        +chainingStrategy, config, lastRecordAttributes1, lastRecordAttributes2, latencyStats, LOG, metrics, output, processingTimeService, stateHandler, stateKeySelector1, stateKeySelector2, timeServiceManager
      • +
      +
    • +
    +
    + +
    + +
    + +
    +
      +
    • + + +

      Method Summary

      + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
      All Methods Static Methods Instance Methods Concrete Methods 
      Modifier and TypeMethodDescription
      voidclose() 
      voidopen() 
      voidprocessElement​(org.apache.flink.streaming.runtime.streamrecord.StreamRecord<TableBucketWriteResult<WriteResult>> streamRecord) 
      voidsetup​(org.apache.flink.streaming.runtime.tasks.StreamTask<?,​?> containingTask, + org.apache.flink.streaming.api.graph.StreamConfig config, + org.apache.flink.streaming.api.operators.Output<org.apache.flink.streaming.runtime.streamrecord.StreamRecord<CommittableMessage<Committable>>> output) 
      static Map<String,​String>toBucketOffsetsProperty​(Map<org.apache.fluss.metadata.TableBucket,​Long> tableBucketOffsets, + Map<Long,​String> partitionNameById, + List<String> partitionKeys) 
      +
        +
      • + + +

        Methods inherited from class org.apache.flink.streaming.api.operators.AbstractStreamOperator

        +finish, getChainingStrategy, getContainingTask, getCurrentKey, getExecutionConfig, getInternalTimerService, getKeyedStateBackend, getKeyedStateStore, getMetricGroup, getOperatorConfig, getOperatorID, getOperatorName, getOperatorStateBackend, getOrCreateKeyedState, getPartitionedState, getPartitionedState, getProcessingTimeService, getRuntimeContext, getStateKeySelector1, getStateKeySelector2, getTimeServiceManager, getUserCodeClassloader, hasKeyContext1, hasKeyContext2, initializeState, initializeState, isUsingCustomRawKeyedState, notifyCheckpointAborted, notifyCheckpointComplete, prepareSnapshotPreBarrier, processLatencyMarker, processLatencyMarker1, processLatencyMarker2, processRecordAttributes, processRecordAttributes1, processRecordAttributes2, processWatermark, processWatermark1, processWatermark2, processWatermarkStatus, processWatermarkStatus1, processWatermarkStatus2, reportOrForwardLatencyMarker, setChainingStrategy, setCurrentKey, setKeyContextElement1, setKeyContextElement2, setMailboxExecutor, setProcessingTimeService, snapshotState, snapshotState, useSplittableTimers
      • +
      + +
        +
      • + + +

        Methods inherited from interface org.apache.flink.api.common.state.CheckpointListener

        +notifyCheckpointAborted, notifyCheckpointComplete
      • +
      +
        +
      • + + +

        Methods inherited from interface org.apache.flink.streaming.api.operators.Input

        +processLatencyMarker, processRecordAttributes, processWatermark, processWatermarkStatus
      • +
      +
        +
      • + + +

        Methods inherited from interface org.apache.flink.streaming.api.operators.KeyContext

        +getCurrentKey, setCurrentKey
      • +
      +
        +
      • + + +

        Methods inherited from interface org.apache.flink.streaming.api.operators.KeyContextHandler

        +hasKeyContext
      • +
      +
        +
      • + + +

        Methods inherited from interface org.apache.flink.streaming.api.operators.OneInputStreamOperator

        +setKeyContextElement
      • +
      +
        +
      • + + +

        Methods inherited from interface org.apache.flink.streaming.api.operators.StreamOperator

        +finish, getMetricGroup, getOperatorAttributes, getOperatorID, initializeState, prepareSnapshotPreBarrier, setKeyContextElement1, setKeyContextElement2, snapshotState
      • +
      +
    • +
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        TieringCommitOperator

        +
        public TieringCommitOperator​(org.apache.flink.streaming.api.operators.StreamOperatorParameters<CommittableMessage<Committable>> parameters,
        +                             org.apache.fluss.config.Configuration flussConf,
        +                             org.apache.fluss.lake.writer.LakeTieringFactory<WriteResult,​Committable> lakeTieringFactory)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        setup

        +
        public void setup​(org.apache.flink.streaming.runtime.tasks.StreamTask<?,​?> containingTask,
        +                  org.apache.flink.streaming.api.graph.StreamConfig config,
        +                  org.apache.flink.streaming.api.operators.Output<org.apache.flink.streaming.runtime.streamrecord.StreamRecord<CommittableMessage<Committable>>> output)
        +
        +
        Specified by:
        +
        setup in interface org.apache.flink.streaming.api.operators.SetupableStreamOperator<WriteResult>
        +
        Overrides:
        +
        setup in class org.apache.flink.streaming.api.operators.AbstractStreamOperator<CommittableMessage<Committable>>
        +
        +
      • +
      + + + +
        +
      • +

        open

        +
        public void open()
        +
        +
        Specified by:
        +
        open in interface org.apache.flink.streaming.api.operators.StreamOperator<WriteResult>
        +
        Overrides:
        +
        open in class org.apache.flink.streaming.api.operators.AbstractStreamOperator<CommittableMessage<Committable>>
        +
        +
      • +
      + + + +
        +
      • +

        processElement

        +
        public void processElement​(org.apache.flink.streaming.runtime.streamrecord.StreamRecord<TableBucketWriteResult<WriteResult>> streamRecord)
        +                    throws Exception
        +
        +
        Specified by:
        +
        processElement in interface org.apache.flink.streaming.api.operators.Input<WriteResult>
        +
        Throws:
        +
        Exception
        +
        +
      • +
      + + + + + + + +
        +
      • +

        close

        +
        public void close()
        +           throws Exception
        +
        +
        Specified by:
        +
        close in interface org.apache.flink.streaming.api.operators.StreamOperator<WriteResult>
        +
        Overrides:
        +
        close in class org.apache.flink.streaming.api.operators.AbstractStreamOperator<CommittableMessage<Committable>>
        +
        Throws:
        +
        Exception
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitOperatorFactory.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitOperatorFactory.html new file mode 100644 index 0000000000..9ae414a189 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitOperatorFactory.html @@ -0,0 +1,376 @@ + + + + + +TieringCommitOperatorFactory (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TieringCommitOperatorFactory<WriteResult,​Committable>

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.flink.streaming.api.operators.AbstractStreamOperatorFactory<CommittableMessage<Committable>>
    • +
    • +
        +
      • org.apache.fluss.flink.tiering.committer.TieringCommitOperatorFactory<WriteResult,​Committable>
      • +
      +
    • +
    +
  • +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Field Summary

      +
        +
      • + + +

        Fields inherited from class org.apache.flink.streaming.api.operators.AbstractStreamOperatorFactory

        +chainingStrategy, processingTimeService
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Summary

      + + + + + + + + + + +
      Constructors 
      ConstructorDescription
      TieringCommitOperatorFactory​(org.apache.fluss.config.Configuration flussConfig, + org.apache.fluss.lake.writer.LakeTieringFactory<WriteResult,​Committable> lakeTieringFactory) 
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        TieringCommitOperatorFactory

        +
        public TieringCommitOperatorFactory​(org.apache.fluss.config.Configuration flussConfig,
        +                                    org.apache.fluss.lake.writer.LakeTieringFactory<WriteResult,​Committable> lakeTieringFactory)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        createStreamOperator

        +
        public <T extends org.apache.flink.streaming.api.operators.StreamOperator<CommittableMessage<Committable>>> T createStreamOperator​(org.apache.flink.streaming.api.operators.StreamOperatorParameters<CommittableMessage<Committable>> parameters)
        +
        +
        Specified by:
        +
        createStreamOperator in interface org.apache.flink.streaming.api.operators.StreamOperatorFactory<WriteResult>
        +
        +
      • +
      + + + +
        +
      • +

        getStreamOperatorClass

        +
        public Class<? extends org.apache.flink.streaming.api.operators.StreamOperator> getStreamOperatorClass​(ClassLoader classLoader)
        +
        +
        Specified by:
        +
        getStreamOperatorClass in interface org.apache.flink.streaming.api.operators.StreamOperatorFactory<WriteResult>
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitterInitContext.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitterInitContext.html new file mode 100644 index 0000000000..fb7679febb --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitterInitContext.html @@ -0,0 +1,316 @@ + + + + + +TieringCommitterInitContext (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TieringCommitterInitContext

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.tiering.committer.TieringCommitterInitContext
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.fluss.lake.committer.CommitterInitContext
    +
    +
    +
    public class TieringCommitterInitContext
    +extends Object
    +implements org.apache.fluss.lake.committer.CommitterInitContext
    +
    The CommitterInitContext implementation for LakeCommitter.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        TieringCommitterInitContext

        +
        public TieringCommitterInitContext​(org.apache.fluss.metadata.TablePath tablePath)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        tablePath

        +
        public org.apache.fluss.metadata.TablePath tablePath()
        +
        +
        Specified by:
        +
        tablePath in interface org.apache.fluss.lake.committer.CommitterInitContext
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/CommittableMessage.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/CommittableMessage.html new file mode 100644 index 0000000000..76ebd86258 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/CommittableMessage.html @@ -0,0 +1,254 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.committer.CommittableMessage (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.committer.CommittableMessage

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/CommittableMessageTypeInfo.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/CommittableMessageTypeInfo.html new file mode 100644 index 0000000000..cf00c5e948 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/CommittableMessageTypeInfo.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo

+
+
No usage of org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/FlussTableLakeSnapshotCommitter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/FlussTableLakeSnapshotCommitter.html new file mode 100644 index 0000000000..26d52e1e42 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/FlussTableLakeSnapshotCommitter.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.committer.FlussTableLakeSnapshotCommitter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.committer.FlussTableLakeSnapshotCommitter

+
+
No usage of org.apache.fluss.flink.tiering.committer.FlussTableLakeSnapshotCommitter
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitOperator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitOperator.html new file mode 100644 index 0000000000..ca3e77cd2a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitOperator.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.committer.TieringCommitOperator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.committer.TieringCommitOperator

+
+
No usage of org.apache.fluss.flink.tiering.committer.TieringCommitOperator
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitOperatorFactory.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitOperatorFactory.html new file mode 100644 index 0000000000..69ea4c4910 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitOperatorFactory.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.committer.TieringCommitOperatorFactory (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.committer.TieringCommitOperatorFactory

+
+
No usage of org.apache.fluss.flink.tiering.committer.TieringCommitOperatorFactory
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitterInitContext.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitterInitContext.html new file mode 100644 index 0000000000..f0f3b237e3 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitterInitContext.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.committer.TieringCommitterInitContext (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.committer.TieringCommitterInitContext

+
+
No usage of org.apache.fluss.flink.tiering.committer.TieringCommitterInitContext
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-summary.html new file mode 100644 index 0000000000..481cf9e146 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-summary.html @@ -0,0 +1,199 @@ + + + + + +org.apache.fluss.flink.tiering.committer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.tiering.committer

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-tree.html new file mode 100644 index 0000000000..bb0c1c75c4 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-tree.html @@ -0,0 +1,180 @@ + + + + + +org.apache.fluss.flink.tiering.committer Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.tiering.committer

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+
    +
  • java.lang.Object +
      +
    • org.apache.flink.streaming.api.operators.AbstractStreamOperator<OUT> (implements org.apache.flink.streaming.api.operators.KeyContextHandler, java.io.Serializable, org.apache.flink.streaming.api.operators.SetupableStreamOperator<OUT>, org.apache.flink.streaming.api.operators.StreamOperator<OUT>, org.apache.flink.streaming.api.operators.StreamOperatorStateHandler.CheckpointedStreamOperator, org.apache.flink.streaming.api.operators.YieldingOperator<OUT>) +
        +
      • org.apache.fluss.flink.tiering.committer.TieringCommitOperator<WriteResult,​Committable> (implements org.apache.flink.streaming.api.operators.OneInputStreamOperator<IN,​OUT>)
      • +
      +
    • +
    • org.apache.flink.streaming.api.operators.AbstractStreamOperatorFactory<OUT> (implements org.apache.flink.streaming.runtime.tasks.ProcessingTimeServiceAware, org.apache.flink.streaming.api.operators.StreamOperatorFactory<OUT>) +
        +
      • org.apache.fluss.flink.tiering.committer.TieringCommitOperatorFactory<WriteResult,​Committable> (implements org.apache.flink.streaming.api.operators.OneInputStreamOperatorFactory<IN,​OUT>)
      • +
      +
    • +
    • org.apache.fluss.flink.tiering.committer.CommittableMessage<Committable> (implements java.io.Serializable)
    • +
    • org.apache.fluss.flink.tiering.committer.FlussTableLakeSnapshotCommitter (implements java.lang.AutoCloseable)
    • +
    • org.apache.fluss.flink.tiering.committer.TieringCommitterInitContext (implements org.apache.fluss.lake.committer.CommitterInitContext)
    • +
    • org.apache.flink.api.common.typeinfo.TypeInformation<T> (implements java.io.Serializable) + +
    • +
    +
  • +
+
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-use.html new file mode 100644 index 0000000000..ec2b9b3423 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-use.html @@ -0,0 +1,185 @@ + + + + + +Uses of Package org.apache.fluss.flink.tiering.committer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.tiering.committer

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/FailedTieringEvent.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/FailedTieringEvent.html new file mode 100644 index 0000000000..e563a36a04 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/FailedTieringEvent.html @@ -0,0 +1,332 @@ + + + + + +FailedTieringEvent (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FailedTieringEvent

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.tiering.event.FailedTieringEvent
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    Serializable, org.apache.flink.api.connector.source.SourceEvent
    +
    +
    +
    public class FailedTieringEvent
    +extends Object
    +implements org.apache.flink.api.connector.source.SourceEvent
    +
    SourceEvent used to represent a Fluss table is failed during tiering.
    +
    +
    See Also:
    +
    Serialized Form
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FailedTieringEvent

        +
        public FailedTieringEvent​(long tableId,
        +                          String failReason)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getTableId

        +
        public long getTableId()
        +
      • +
      + + + +
        +
      • +

        failReason

        +
        public String failReason()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/FinishedTieringEvent.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/FinishedTieringEvent.html new file mode 100644 index 0000000000..c5e2342b00 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/FinishedTieringEvent.html @@ -0,0 +1,316 @@ + + + + + +FinishedTieringEvent (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FinishedTieringEvent

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.tiering.event.FinishedTieringEvent
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    Serializable, org.apache.flink.api.connector.source.SourceEvent
    +
    +
    +
    public class FinishedTieringEvent
    +extends Object
    +implements org.apache.flink.api.connector.source.SourceEvent
    +
    SourceEvent used to represent a Fluss table has been tiered finished.
    +
    +
    See Also:
    +
    Serialized Form
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FinishedTieringEvent

        +
        public FinishedTieringEvent​(long tableId)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getTableId

        +
        public long getTableId()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/TieringRestoreEvent.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/TieringRestoreEvent.html new file mode 100644 index 0000000000..79b77386e6 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/TieringRestoreEvent.html @@ -0,0 +1,278 @@ + + + + + +TieringRestoreEvent (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TieringRestoreEvent

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.tiering.event.TieringRestoreEvent
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    Serializable, org.apache.flink.api.connector.source.SourceEvent
    +
    +
    +
    public class TieringRestoreEvent
    +extends Object
    +implements org.apache.flink.api.connector.source.SourceEvent
    +
    SourceEvent used to represent tiering is restoring.
    +
    +
    See Also:
    +
    Serialized Form
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        TieringRestoreEvent

        +
        public TieringRestoreEvent()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/FailedTieringEvent.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/FailedTieringEvent.html new file mode 100644 index 0000000000..63e3497efd --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/FailedTieringEvent.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.event.FailedTieringEvent (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.event.FailedTieringEvent

+
+
No usage of org.apache.fluss.flink.tiering.event.FailedTieringEvent
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/FinishedTieringEvent.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/FinishedTieringEvent.html new file mode 100644 index 0000000000..f735faa31a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/FinishedTieringEvent.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.event.FinishedTieringEvent (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.event.FinishedTieringEvent

+
+
No usage of org.apache.fluss.flink.tiering.event.FinishedTieringEvent
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/TieringRestoreEvent.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/TieringRestoreEvent.html new file mode 100644 index 0000000000..05989a26c2 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/TieringRestoreEvent.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.event.TieringRestoreEvent (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.event.TieringRestoreEvent

+
+
No usage of org.apache.fluss.flink.tiering.event.TieringRestoreEvent
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-summary.html new file mode 100644 index 0000000000..dfa22d3083 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-summary.html @@ -0,0 +1,180 @@ + + + + + +org.apache.fluss.flink.tiering.event (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.tiering.event

+
+
+
    +
  • + + + + + + + + + + + + + + + + + + + + +
    Class Summary 
    ClassDescription
    FailedTieringEvent +
    SourceEvent used to represent a Fluss table is failed during tiering.
    +
    FinishedTieringEvent +
    SourceEvent used to represent a Fluss table has been tiered finished.
    +
    TieringRestoreEvent +
    SourceEvent used to represent tiering is restoring.
    +
    +
  • +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-tree.html new file mode 100644 index 0000000000..9becc30f22 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-tree.html @@ -0,0 +1,165 @@ + + + + + +org.apache.fluss.flink.tiering.event Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.tiering.event

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+
    +
  • java.lang.Object +
      +
    • org.apache.fluss.flink.tiering.event.FailedTieringEvent (implements org.apache.flink.api.connector.source.SourceEvent)
    • +
    • org.apache.fluss.flink.tiering.event.FinishedTieringEvent (implements org.apache.flink.api.connector.source.SourceEvent)
    • +
    • org.apache.fluss.flink.tiering.event.TieringRestoreEvent (implements org.apache.flink.api.connector.source.SourceEvent)
    • +
    +
  • +
+
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-use.html new file mode 100644 index 0000000000..5a4d6075d0 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-use.html @@ -0,0 +1,148 @@ + + + + + +Uses of Package org.apache.fluss.flink.tiering.event (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.tiering.event

+
+
No usage of org.apache.fluss.flink.tiering.event
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-summary.html new file mode 100644 index 0000000000..04a9a2fb1f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-summary.html @@ -0,0 +1,168 @@ + + + + + +org.apache.fluss.flink.tiering (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.tiering

+
+
+
    +
  • + + + + + + + + + + + + +
    Class Summary 
    ClassDescription
    LakeTieringJobBuilder +
    The builder to build Flink lake tiering job.
    +
    +
  • +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-tree.html new file mode 100644 index 0000000000..df81712c57 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-tree.html @@ -0,0 +1,163 @@ + + + + + +org.apache.fluss.flink.tiering Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.tiering

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-use.html new file mode 100644 index 0000000000..db2b4586b0 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-use.html @@ -0,0 +1,185 @@ + + + + + +Uses of Package org.apache.fluss.flink.tiering (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.tiering

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResult.html new file mode 100644 index 0000000000..cded453a08 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResult.html @@ -0,0 +1,405 @@ + + + + + +TableBucketWriteResult (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TableBucketWriteResult<WriteResult>

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.tiering.source.TableBucketWriteResult<WriteResult>
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    Serializable
    +
    +
    +
    public class TableBucketWriteResult<WriteResult>
    +extends Object
    +implements Serializable
    +
    This class contains the TableBucketWriteResult of LakeWriter, the table path and the bucket + that the write result is for, the end log offset of tiering, the total number of write results in + one round of tiering. It'll be passed to downstream committer operator to collect all the write + results of a table and do commit.
    +
    +
    See Also:
    +
    Serialized Form
    +
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + + + +
        +
      • +

        TableBucketWriteResult

        +
        public TableBucketWriteResult​(org.apache.fluss.metadata.TablePath tablePath,
        +                              org.apache.fluss.metadata.TableBucket tableBucket,
        +                              @Nullable
        +                              String partitionName,
        +                              @Nullable
        +                              WriteResult writeResult,
        +                              long logEndOffset,
        +                              int numberOfWriteResults)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        tablePath

        +
        public org.apache.fluss.metadata.TablePath tablePath()
        +
      • +
      + + + +
        +
      • +

        tableBucket

        +
        public org.apache.fluss.metadata.TableBucket tableBucket()
        +
      • +
      + + + +
        +
      • +

        partitionName

        +
        @Nullable
        +public String partitionName()
        +
      • +
      + + + +
        +
      • +

        writeResult

        +
        @Nullable
        +public WriteResult writeResult()
        +
      • +
      + + + +
        +
      • +

        numberOfWriteResults

        +
        public int numberOfWriteResults()
        +
      • +
      + + + +
        +
      • +

        logEndOffset

        +
        public long logEndOffset()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultEmitter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultEmitter.html new file mode 100644 index 0000000000..2eb6827391 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultEmitter.html @@ -0,0 +1,320 @@ + + + + + +TableBucketWriteResultEmitter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TableBucketWriteResultEmitter<WriteResult>

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.tiering.source.TableBucketWriteResultEmitter<WriteResult>
    • +
    +
  • +
+
+ +
+
+ +
+
+ +
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultSerializer.html new file mode 100644 index 0000000000..5b3ddd4ece --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultSerializer.html @@ -0,0 +1,362 @@ + + + + + +TableBucketWriteResultSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TableBucketWriteResultSerializer<WriteResult>

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.tiering.source.TableBucketWriteResultSerializer<WriteResult>
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.core.io.SimpleVersionedSerializer<TableBucketWriteResult<WriteResult>>, org.apache.flink.core.io.Versioned
    +
    +
    +
    public class TableBucketWriteResultSerializer<WriteResult>
    +extends Object
    +implements org.apache.flink.core.io.SimpleVersionedSerializer<TableBucketWriteResult<WriteResult>>
    +
    The serializer for TableBucketWriteResult.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        TableBucketWriteResultSerializer

        +
        public TableBucketWriteResultSerializer​(org.apache.fluss.lake.serializer.SimpleVersionedSerializer<WriteResult> writeResultSerializer)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getVersion

        +
        public int getVersion()
        +
        +
        Specified by:
        +
        getVersion in interface org.apache.flink.core.io.SimpleVersionedSerializer<WriteResult>
        +
        Specified by:
        +
        getVersion in interface org.apache.flink.core.io.Versioned
        +
        +
      • +
      + + + + + + + + +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultTypeInfo.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultTypeInfo.html new file mode 100644 index 0000000000..37c42bdab6 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultTypeInfo.html @@ -0,0 +1,485 @@ + + + + + +TableBucketWriteResultTypeInfo (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TableBucketWriteResultTypeInfo<WriteResult>

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.flink.api.common.typeinfo.TypeInformation<TableBucketWriteResult<WriteResult>>
    • +
    • +
        +
      • org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo<WriteResult>
      • +
      +
    • +
    +
  • +
+
+ +
+
+ +
+
+ +
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSource.Builder.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSource.Builder.html new file mode 100644 index 0000000000..0fd59d2c86 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSource.Builder.html @@ -0,0 +1,327 @@ + + + + + +TieringSource.Builder (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TieringSource.Builder<WriteResult>

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.tiering.source.TieringSource.Builder<WriteResult>
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        Builder

        +
        public Builder​(org.apache.fluss.config.Configuration flussConf,
        +               org.apache.fluss.lake.writer.LakeTieringFactory<WriteResult,​?> lakeTieringFactory)
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSource.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSource.html new file mode 100644 index 0000000000..4b6563c3ed --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSource.html @@ -0,0 +1,514 @@ + + + + + +TieringSource (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TieringSource<WriteResult>

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.tiering.source.TieringSource<WriteResult>
    • +
    +
  • +
+
+ +
+
+ +
+
+ +
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSourceOptions.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSourceOptions.html new file mode 100644 index 0000000000..d3d667a0c4 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSourceOptions.html @@ -0,0 +1,329 @@ + + + + + +TieringSourceOptions (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TieringSourceOptions

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.tiering.source.TieringSourceOptions
    • +
    +
  • +
+
+
    +
  • +
    +
    public class TieringSourceOptions
    +extends Object
    +
    Configuration options for the TieringSource.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + + + + + +
        +
      • +

        POLL_TIERING_TABLE_INTERVAL

        +
        public static final org.apache.fluss.config.ConfigOption<Duration> POLL_TIERING_TABLE_INTERVAL
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        TieringSourceOptions

        +
        public TieringSourceOptions()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSourceReader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSourceReader.html new file mode 100644 index 0000000000..05d7dd6881 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSourceReader.html @@ -0,0 +1,446 @@ + + + + + +TieringSourceReader (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TieringSourceReader<WriteResult>

+
+
+ +
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Field Summary

      +
        +
      • + + +

        Fields inherited from class org.apache.flink.connector.base.source.reader.SourceReaderBase

        +config, context, eofRecordEvaluator, options, recordEmitter, splitFetcherManager
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Summary

      + + + + + + + + + + +
      Constructors 
      ConstructorDescription
      TieringSourceReader​(org.apache.flink.api.connector.source.SourceReaderContext context, + org.apache.fluss.config.Configuration flussConf, + org.apache.fluss.lake.writer.LakeTieringFactory<WriteResult,​?> lakeTieringFactory) 
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+ +
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSplitReader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSplitReader.html new file mode 100644 index 0000000000..b1c1b86349 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSplitReader.html @@ -0,0 +1,385 @@ + + + + + +TieringSplitReader (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TieringSplitReader<WriteResult>

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.tiering.source.TieringSplitReader<WriteResult>
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    AutoCloseable, org.apache.flink.connector.base.source.reader.splitreader.SplitReader<TableBucketWriteResult<WriteResult>,​TieringSplit>
    +
    +
    +
    public class TieringSplitReader<WriteResult>
    +extends Object
    +implements org.apache.flink.connector.base.source.reader.splitreader.SplitReader<TableBucketWriteResult<WriteResult>,​TieringSplit>
    +
    The SplitReader implementation which will read Fluss and write to lake.
    +
  • +
+
+
+ +
+
+ +
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringWriterInitContext.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringWriterInitContext.html new file mode 100644 index 0000000000..b4faab11cf --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringWriterInitContext.html @@ -0,0 +1,378 @@ + + + + + +TieringWriterInitContext (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TieringWriterInitContext

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.tiering.source.TieringWriterInitContext
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.fluss.lake.writer.WriterInitContext
    +
    +
    +
    public class TieringWriterInitContext
    +extends Object
    +implements org.apache.fluss.lake.writer.WriterInitContext
    +
    The implementation of WriterInitContext.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        TieringWriterInitContext

        +
        public TieringWriterInitContext​(org.apache.fluss.metadata.TablePath tablePath,
        +                                org.apache.fluss.metadata.TableBucket tableBucket,
        +                                @Nullable
        +                                String partition,
        +                                org.apache.fluss.metadata.TableInfo tableInfo)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        tablePath

        +
        public org.apache.fluss.metadata.TablePath tablePath()
        +
        +
        Specified by:
        +
        tablePath in interface org.apache.fluss.lake.writer.WriterInitContext
        +
        +
      • +
      + + + +
        +
      • +

        tableBucket

        +
        public org.apache.fluss.metadata.TableBucket tableBucket()
        +
        +
        Specified by:
        +
        tableBucket in interface org.apache.fluss.lake.writer.WriterInitContext
        +
        +
      • +
      + + + +
        +
      • +

        partition

        +
        @Nullable
        +public String partition()
        +
        +
        Specified by:
        +
        partition in interface org.apache.fluss.lake.writer.WriterInitContext
        +
        +
      • +
      + + + +
        +
      • +

        tableInfo

        +
        public org.apache.fluss.metadata.TableInfo tableInfo()
        +
        +
        Specified by:
        +
        tableInfo in interface org.apache.fluss.lake.writer.WriterInitContext
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResult.html new file mode 100644 index 0000000000..23010e9cb9 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResult.html @@ -0,0 +1,293 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.source.TableBucketWriteResult (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.source.TableBucketWriteResult

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultEmitter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultEmitter.html new file mode 100644 index 0000000000..518b79232f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultEmitter.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.source.TableBucketWriteResultEmitter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.source.TableBucketWriteResultEmitter

+
+
No usage of org.apache.fluss.flink.tiering.source.TableBucketWriteResultEmitter
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultSerializer.html new file mode 100644 index 0000000000..bac06c06c1 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultSerializer.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.source.TableBucketWriteResultSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.source.TableBucketWriteResultSerializer

+
+
No usage of org.apache.fluss.flink.tiering.source.TableBucketWriteResultSerializer
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultTypeInfo.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultTypeInfo.html new file mode 100644 index 0000000000..10688df6da --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultTypeInfo.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo

+
+
No usage of org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSource.Builder.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSource.Builder.html new file mode 100644 index 0000000000..4e5b587baa --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSource.Builder.html @@ -0,0 +1,192 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.source.TieringSource.Builder (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.source.TieringSource.Builder

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSource.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSource.html new file mode 100644 index 0000000000..2396b530af --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSource.html @@ -0,0 +1,192 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.source.TieringSource (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.source.TieringSource

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSourceOptions.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSourceOptions.html new file mode 100644 index 0000000000..b9f0769313 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSourceOptions.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.source.TieringSourceOptions (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.source.TieringSourceOptions

+
+
No usage of org.apache.fluss.flink.tiering.source.TieringSourceOptions
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSourceReader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSourceReader.html new file mode 100644 index 0000000000..b02e21be8c --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSourceReader.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.source.TieringSourceReader (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.source.TieringSourceReader

+
+
No usage of org.apache.fluss.flink.tiering.source.TieringSourceReader
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSplitReader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSplitReader.html new file mode 100644 index 0000000000..1443bbc623 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSplitReader.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.source.TieringSplitReader (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.source.TieringSplitReader

+
+
No usage of org.apache.fluss.flink.tiering.source.TieringSplitReader
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringWriterInitContext.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringWriterInitContext.html new file mode 100644 index 0000000000..374ebb3728 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringWriterInitContext.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.source.TieringWriterInitContext (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.source.TieringWriterInitContext

+
+
No usage of org.apache.fluss.flink.tiering.source.TieringWriterInitContext
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/TieringSourceEnumerator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/TieringSourceEnumerator.html new file mode 100644 index 0000000000..a09c14db92 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/TieringSourceEnumerator.html @@ -0,0 +1,468 @@ + + + + + +TieringSourceEnumerator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TieringSourceEnumerator

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    AutoCloseable, org.apache.flink.api.common.state.CheckpointListener, org.apache.flink.api.connector.source.SplitEnumerator<TieringSplit,​TieringSourceEnumeratorState>
    +
    +
    +
    public class TieringSourceEnumerator
    +extends Object
    +implements org.apache.flink.api.connector.source.SplitEnumerator<TieringSplit,​TieringSourceEnumeratorState>
    +
    An implementation of SplitEnumerator used to request TieringSplit from Fluss + Cluster. + +

    The enumerator is responsible for: + +

      +
    • Register the Tiering Service job that the current TieringSourceEnumerator belongs to with + the Fluss Cluster when the Flink Tiering job starts up. +
    • Request Fluss table splits from Fluss Cluster and assigns to SourceReader to tier. +
    • Un-Register the Tiering Service job that the current TieringSourceEnumerator belongs to + with the Fluss Cluster when the Flink Tiering job shutdown as much as possible. +
    +
  • +
+
+
+ +
+
+ +
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/class-use/TieringSourceEnumerator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/class-use/TieringSourceEnumerator.html new file mode 100644 index 0000000000..77b0bdc411 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/class-use/TieringSourceEnumerator.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator

+
+
No usage of org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-summary.html new file mode 100644 index 0000000000..c79bfe407c --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-summary.html @@ -0,0 +1,169 @@ + + + + + +org.apache.fluss.flink.tiering.source.enumerator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.tiering.source.enumerator

+
+
+
    +
  • + + + + + + + + + + + + +
    Class Summary 
    ClassDescription
    TieringSourceEnumerator +
    An implementation of SplitEnumerator used to request TieringSplit from Fluss + Cluster.
    +
    +
  • +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-tree.html new file mode 100644 index 0000000000..295c459f12 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-tree.html @@ -0,0 +1,163 @@ + + + + + +org.apache.fluss.flink.tiering.source.enumerator Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.tiering.source.enumerator

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+
    +
  • java.lang.Object +
      +
    • org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator (implements org.apache.flink.api.connector.source.SplitEnumerator<SplitT,​CheckpointT>)
    • +
    +
  • +
+
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-use.html new file mode 100644 index 0000000000..107d0e6ccd --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-use.html @@ -0,0 +1,148 @@ + + + + + +Uses of Package org.apache.fluss.flink.tiering.source.enumerator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.tiering.source.enumerator

+
+
No usage of org.apache.fluss.flink.tiering.source.enumerator
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-summary.html new file mode 100644 index 0000000000..67afe12196 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-summary.html @@ -0,0 +1,224 @@ + + + + + +org.apache.fluss.flink.tiering.source (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.tiering.source

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-tree.html new file mode 100644 index 0000000000..a21b107201 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-tree.html @@ -0,0 +1,184 @@ + + + + + +org.apache.fluss.flink.tiering.source Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.tiering.source

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+
    +
  • java.lang.Object +
      +
    • org.apache.flink.connector.base.source.reader.SourceReaderBase<E,​T,​SplitT,​SplitStateT> (implements org.apache.flink.api.connector.source.SourceReader<T,​SplitT>) +
        +
      • org.apache.flink.connector.base.source.reader.SingleThreadMultiplexSourceReaderBase<E,​T,​SplitT,​SplitStateT> + +
      • +
      +
    • +
    • org.apache.fluss.flink.tiering.source.TableBucketWriteResult<WriteResult> (implements java.io.Serializable)
    • +
    • org.apache.fluss.flink.tiering.source.TableBucketWriteResultEmitter<WriteResult> (implements org.apache.flink.connector.base.source.reader.RecordEmitter<E,​T,​SplitStateT>)
    • +
    • org.apache.fluss.flink.tiering.source.TableBucketWriteResultSerializer<WriteResult> (implements org.apache.flink.core.io.SimpleVersionedSerializer<E>)
    • +
    • org.apache.fluss.flink.tiering.source.TieringSource<WriteResult> (implements org.apache.flink.api.connector.source.Source<T,​SplitT,​EnumChkT>)
    • +
    • org.apache.fluss.flink.tiering.source.TieringSource.Builder<WriteResult>
    • +
    • org.apache.fluss.flink.tiering.source.TieringSourceOptions
    • +
    • org.apache.fluss.flink.tiering.source.TieringSplitReader<WriteResult> (implements org.apache.flink.connector.base.source.reader.splitreader.SplitReader<E,​SplitT>)
    • +
    • org.apache.fluss.flink.tiering.source.TieringWriterInitContext (implements org.apache.fluss.lake.writer.WriterInitContext)
    • +
    • org.apache.flink.api.common.typeinfo.TypeInformation<T> (implements java.io.Serializable) + +
    • +
    +
  • +
+
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-use.html new file mode 100644 index 0000000000..f786fcfc8b --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-use.html @@ -0,0 +1,224 @@ + + + + + +Uses of Package org.apache.fluss.flink.tiering.source (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.tiering.source

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringLogSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringLogSplit.html new file mode 100644 index 0000000000..c0d364ee74 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringLogSplit.html @@ -0,0 +1,473 @@ + + + + + +TieringLogSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TieringLogSplit

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.api.connector.source.SourceSplit
    +
    +
    +
    public class TieringLogSplit
    +extends TieringSplit
    +
    The table split for tiering service. It's used to describe the log data of a table bucket.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        TieringLogSplit

        +
        public TieringLogSplit​(org.apache.fluss.metadata.TablePath tablePath,
        +                       org.apache.fluss.metadata.TableBucket tableBucket,
        +                       @Nullable
        +                       String partitionName,
        +                       long startingOffset,
        +                       long stoppingOffset)
        +
      • +
      + + + +
        +
      • +

        TieringLogSplit

        +
        public TieringLogSplit​(org.apache.fluss.metadata.TablePath tablePath,
        +                       org.apache.fluss.metadata.TableBucket tableBucket,
        +                       @Nullable
        +                       String partitionName,
        +                       long startingOffset,
        +                       long stoppingOffset,
        +                       int numberOfSplits)
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSnapshotSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSnapshotSplit.html new file mode 100644 index 0000000000..f3a0c3ab4e --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSnapshotSplit.html @@ -0,0 +1,474 @@ + + + + + +TieringSnapshotSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TieringSnapshotSplit

+
+
+ +
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.api.connector.source.SourceSplit
    +
    +
    +
    public class TieringSnapshotSplit
    +extends TieringSplit
    +
    The table split for tiering service. It's used to describe the snapshot data of a primary key + table bucket.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        TieringSnapshotSplit

        +
        public TieringSnapshotSplit​(org.apache.fluss.metadata.TablePath tablePath,
        +                            org.apache.fluss.metadata.TableBucket tableBucket,
        +                            @Nullable
        +                            String partitionName,
        +                            long snapshotId,
        +                            long logOffsetOfSnapshot)
        +
      • +
      + + + +
        +
      • +

        TieringSnapshotSplit

        +
        public TieringSnapshotSplit​(org.apache.fluss.metadata.TablePath tablePath,
        +                            org.apache.fluss.metadata.TableBucket tableBucket,
        +                            @Nullable
        +                            String partitionName,
        +                            long snapshotId,
        +                            long logOffsetOfSnapshot,
        +                            int numberOfSplits)
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplit.html new file mode 100644 index 0000000000..7a1de92bb8 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplit.html @@ -0,0 +1,660 @@ + + + + + +TieringSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TieringSplit

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.tiering.source.split.TieringSplit
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.api.connector.source.SourceSplit
    +
    +
    +
    Direct Known Subclasses:
    +
    TieringLogSplit, TieringSnapshotSplit
    +
    +
    +
    public abstract class TieringSplit
    +extends Object
    +implements org.apache.flink.api.connector.source.SourceSplit
    +
    The base table split for tiering service.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + +
        +
      • +

        TIERING_SNAPSHOT_SPLIT_FLAG

        +
        public static final byte TIERING_SNAPSHOT_SPLIT_FLAG
        +
        +
        See Also:
        +
        Constant Field Values
        +
        +
      • +
      + + + +
        +
      • +

        TIERING_LOG_SPLIT_FLAG

        +
        public static final byte TIERING_LOG_SPLIT_FLAG
        +
        +
        See Also:
        +
        Constant Field Values
        +
        +
      • +
      + + + +
        +
      • +

        UNKNOWN_NUMBER_OF_SPLITS

        +
        protected static final int UNKNOWN_NUMBER_OF_SPLITS
        +
        +
        See Also:
        +
        Constant Field Values
        +
        +
      • +
      + + + +
        +
      • +

        tablePath

        +
        protected final org.apache.fluss.metadata.TablePath tablePath
        +
      • +
      + + + +
        +
      • +

        tableBucket

        +
        protected final org.apache.fluss.metadata.TableBucket tableBucket
        +
      • +
      + + + +
        +
      • +

        partitionName

        +
        @Nullable
        +protected final String partitionName
        +
      • +
      + + + +
        +
      • +

        numberOfSplits

        +
        protected final int numberOfSplits
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        TieringSplit

        +
        public TieringSplit​(org.apache.fluss.metadata.TablePath tablePath,
        +                    org.apache.fluss.metadata.TableBucket tableBucket,
        +                    @Nullable
        +                    String partitionName,
        +                    int numberOfSplits)
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        isTieringSnapshotSplit

        +
        public final boolean isTieringSnapshotSplit()
        +
        Checks whether this split is a primary key table split to tier.
        +
      • +
      + + + + + + + +
        +
      • +

        isTieringLogSplit

        +
        public final boolean isTieringLogSplit()
        +
        Checks whether this split is a log split to tier.
        +
      • +
      + + + + + + + +
        +
      • +

        splitKind

        +
        protected byte splitKind()
        +
      • +
      + + + +
        +
      • +

        getNumberOfSplits

        +
        public int getNumberOfSplits()
        +
      • +
      + + + +
        +
      • +

        toSplitId

        +
        protected static String toSplitId​(String splitPrefix,
        +                                  org.apache.fluss.metadata.TableBucket tableBucket)
        +
      • +
      + + + +
        +
      • +

        getTablePath

        +
        public org.apache.fluss.metadata.TablePath getTablePath()
        +
      • +
      + + + +
        +
      • +

        getTableBucket

        +
        public org.apache.fluss.metadata.TableBucket getTableBucket()
        +
      • +
      + + + +
        +
      • +

        getPartitionName

        +
        @Nullable
        +public String getPartitionName()
        +
      • +
      + + + +
        +
      • +

        copy

        +
        public abstract TieringSplit copy​(int numberOfSplits)
        +
      • +
      + + + +
        +
      • +

        equals

        +
        public boolean equals​(Object object)
        +
        +
        Overrides:
        +
        equals in class Object
        +
        +
      • +
      + + + +
        +
      • +

        hashCode

        +
        public int hashCode()
        +
        +
        Overrides:
        +
        hashCode in class Object
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplitGenerator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplitGenerator.html new file mode 100644 index 0000000000..56c432b7e9 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplitGenerator.html @@ -0,0 +1,312 @@ + + + + + +TieringSplitGenerator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TieringSplitGenerator

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.tiering.source.split.TieringSplitGenerator
    • +
    +
  • +
+
+
    +
  • +
    +
    public class TieringSplitGenerator
    +extends Object
    +
    A generator for lake splits.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        TieringSplitGenerator

        +
        public TieringSplitGenerator​(Admin flussAdmin)
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplitSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplitSerializer.html new file mode 100644 index 0000000000..5da2d5a8a7 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplitSerializer.html @@ -0,0 +1,404 @@ + + + + + +TieringSplitSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TieringSplitSerializer

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.tiering.source.split.TieringSplitSerializer
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    org.apache.flink.core.io.SimpleVersionedSerializer<TieringSplit>, org.apache.flink.core.io.Versioned
    +
    +
    +
    public class TieringSplitSerializer
    +extends Object
    +implements org.apache.flink.core.io.SimpleVersionedSerializer<TieringSplit>
    +
    A serializer for the TieringSplit.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    + +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        TieringSplitSerializer

        +
        public TieringSplitSerializer()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getVersion

        +
        public int getVersion()
        +
        +
        Specified by:
        +
        getVersion in interface org.apache.flink.core.io.SimpleVersionedSerializer<TieringSplit>
        +
        Specified by:
        +
        getVersion in interface org.apache.flink.core.io.Versioned
        +
        +
      • +
      + + + + + + + +
        +
      • +

        deserialize

        +
        public TieringSplit deserialize​(int version,
        +                                byte[] serialized)
        +                         throws IOException
        +
        +
        Specified by:
        +
        deserialize in interface org.apache.flink.core.io.SimpleVersionedSerializer<TieringSplit>
        +
        Throws:
        +
        IOException
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringLogSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringLogSplit.html new file mode 100644 index 0000000000..2d747eed87 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringLogSplit.html @@ -0,0 +1,199 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.source.split.TieringLogSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.source.split.TieringLogSplit

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSnapshotSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSnapshotSplit.html new file mode 100644 index 0000000000..0f15d3f778 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSnapshotSplit.html @@ -0,0 +1,199 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.source.split.TieringSnapshotSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.source.split.TieringSnapshotSplit

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplit.html new file mode 100644 index 0000000000..2e7ff6ba79 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplit.html @@ -0,0 +1,452 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.source.split.TieringSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.source.split.TieringSplit

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplitGenerator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplitGenerator.html new file mode 100644 index 0000000000..e761d1dfd3 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplitGenerator.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.source.split.TieringSplitGenerator (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.source.split.TieringSplitGenerator

+
+
No usage of org.apache.fluss.flink.tiering.source.split.TieringSplitGenerator
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplitSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplitSerializer.html new file mode 100644 index 0000000000..6dbd1382dd --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplitSerializer.html @@ -0,0 +1,192 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.source.split.TieringSplitSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.source.split.TieringSplitSerializer

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-summary.html new file mode 100644 index 0000000000..6aa578aedf --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-summary.html @@ -0,0 +1,192 @@ + + + + + +org.apache.fluss.flink.tiering.source.split (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.tiering.source.split

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-tree.html new file mode 100644 index 0000000000..6234a6edc8 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-tree.html @@ -0,0 +1,170 @@ + + + + + +org.apache.fluss.flink.tiering.source.split Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.tiering.source.split

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+
    +
  • java.lang.Object +
      +
    • org.apache.fluss.flink.tiering.source.split.TieringSplit (implements org.apache.flink.api.connector.source.SourceSplit) + +
    • +
    • org.apache.fluss.flink.tiering.source.split.TieringSplitGenerator
    • +
    • org.apache.fluss.flink.tiering.source.split.TieringSplitSerializer (implements org.apache.flink.core.io.SimpleVersionedSerializer<E>)
    • +
    +
  • +
+
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-use.html new file mode 100644 index 0000000000..f3414c3da6 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-use.html @@ -0,0 +1,272 @@ + + + + + +Uses of Package org.apache.fluss.flink.tiering.source.split (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.tiering.source.split

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSourceEnumeratorState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSourceEnumeratorState.html new file mode 100644 index 0000000000..1b89a480c5 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSourceEnumeratorState.html @@ -0,0 +1,329 @@ + + + + + +TieringSourceEnumeratorState (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TieringSourceEnumeratorState

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorState
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        TieringSourceEnumeratorState

        +
        public TieringSourceEnumeratorState()
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSourceEnumeratorStateSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSourceEnumeratorStateSerializer.html new file mode 100644 index 0000000000..d5251a8376 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSourceEnumeratorStateSerializer.html @@ -0,0 +1,404 @@ + + + + + +TieringSourceEnumeratorStateSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TieringSourceEnumeratorStateSerializer

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorStateSerializer
    • +
    +
  • +
+
+ +
+
+ +
+
+ +
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSplitState.html new file mode 100644 index 0000000000..84f796833b --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSplitState.html @@ -0,0 +1,353 @@ + + + + + +TieringSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class TieringSplitState

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.tiering.source.state.TieringSplitState
    • +
    +
  • +
+
+
    +
  • +
    +
    public class TieringSplitState
    +extends Object
    +
    The state of a TieringSplit. + +

    Note: The tiering service adopts a stateless design and does not store any progress + information in state during checkpoints. All splits are re-requested from the Fluss cluster in + case of failover.

    +
  • +
+
+
+ +
+
+
    +
  • + +
    + +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        TieringSplitState

        +
        public TieringSplitState​(TieringSplit tieringSplit)
        +
      • +
      +
    • +
    +
    + +
    + +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSourceEnumeratorState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSourceEnumeratorState.html new file mode 100644 index 0000000000..090c958c1a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSourceEnumeratorState.html @@ -0,0 +1,287 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorState (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorState

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSourceEnumeratorStateSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSourceEnumeratorStateSerializer.html new file mode 100644 index 0000000000..1d1957f129 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSourceEnumeratorStateSerializer.html @@ -0,0 +1,192 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorStateSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorStateSerializer

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSplitState.html new file mode 100644 index 0000000000..d388abca60 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSplitState.html @@ -0,0 +1,230 @@ + + + + + +Uses of Class org.apache.fluss.flink.tiering.source.state.TieringSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.tiering.source.state.TieringSplitState

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-summary.html new file mode 100644 index 0000000000..bdfe225f86 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-summary.html @@ -0,0 +1,180 @@ + + + + + +org.apache.fluss.flink.tiering.source.state (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.tiering.source.state

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-tree.html new file mode 100644 index 0000000000..180acf46ee --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-tree.html @@ -0,0 +1,165 @@ + + + + + +org.apache.fluss.flink.tiering.source.state Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.tiering.source.state

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-use.html new file mode 100644 index 0000000000..b9fefc7117 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-use.html @@ -0,0 +1,243 @@ + + + + + +Uses of Package org.apache.fluss.flink.tiering.source.state (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.tiering.source.state

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/CatalogExceptionUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/CatalogExceptionUtils.html new file mode 100644 index 0000000000..e7beacfa47 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/CatalogExceptionUtils.html @@ -0,0 +1,407 @@ + + + + + +CatalogExceptionUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class CatalogExceptionUtils

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.utils.CatalogExceptionUtils
    • +
    +
  • +
+
+
    +
  • +
    +
    public class CatalogExceptionUtils
    +extends Object
    +
    Utility class for catalog exceptions.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        isDatabaseNotExist

        +
        public static boolean isDatabaseNotExist​(Throwable throwable)
        +
      • +
      + + + +
        +
      • +

        isDatabaseNotEmpty

        +
        public static boolean isDatabaseNotEmpty​(Throwable throwable)
        +
      • +
      + + + +
        +
      • +

        isDatabaseAlreadyExist

        +
        public static boolean isDatabaseAlreadyExist​(Throwable throwable)
        +
      • +
      + + + +
        +
      • +

        isTableNotExist

        +
        public static boolean isTableNotExist​(Throwable throwable)
        +
      • +
      + + + +
        +
      • +

        isTableAlreadyExist

        +
        public static boolean isTableAlreadyExist​(Throwable throwable)
        +
      • +
      + + + +
        +
      • +

        isLakeTableAlreadyExist

        +
        public static boolean isLakeTableAlreadyExist​(Throwable throwable)
        +
      • +
      + + + +
        +
      • +

        isTableInvalid

        +
        public static boolean isTableInvalid​(Throwable throwable)
        +
      • +
      + + + +
        +
      • +

        isTableNotPartitioned

        +
        public static boolean isTableNotPartitioned​(Throwable throwable)
        +
      • +
      + + + +
        +
      • +

        isPartitionAlreadyExists

        +
        public static boolean isPartitionAlreadyExists​(Throwable throwable)
        +
      • +
      + + + +
        +
      • +

        isPartitionNotExist

        +
        public static boolean isPartitionNotExist​(Throwable throwable)
        +
      • +
      + + + +
        +
      • +

        isPartitionInvalid

        +
        public static boolean isPartitionInvalid​(Throwable throwable)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/CatalogPropertiesUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/CatalogPropertiesUtils.html new file mode 100644 index 0000000000..3a1c3b6ce5 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/CatalogPropertiesUtils.html @@ -0,0 +1,379 @@ + + + + + +CatalogPropertiesUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class CatalogPropertiesUtils

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.utils.CatalogPropertiesUtils
    • +
    +
  • +
+
+
    +
  • +
    +
    public class CatalogPropertiesUtils
    +extends Object
    +
    Utilities for de/serializing Catalog objects into a map of string properties. + +

    Copied from Flink.

    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + + + + + +
        +
      • +

        deserializeWatermark

        +
        public static void deserializeWatermark​(Map<String,​String> map,
        +                                        org.apache.flink.table.api.Schema.Builder builder)
        +
      • +
      + + + +
        +
      • +

        deserializeComputedColumn

        +
        public static void deserializeComputedColumn​(Map<String,​String> map,
        +                                             int columnIndex,
        +                                             org.apache.flink.table.api.Schema.Builder builder)
        +
      • +
      + + + +
        +
      • +

        serializeComputedColumns

        +
        public static void serializeComputedColumns​(Map<String,​String> map,
        +                                            List<org.apache.flink.table.catalog.Column> columns)
        +
      • +
      + + + +
        +
      • +

        nonPhysicalColumnsCount

        +
        public static int nonPhysicalColumnsCount​(Map<String,​String> tableOptions,
        +                                          List<String> physicalColumns)
        +
      • +
      + + + +
        +
      • +

        serializeComputedColumns

        +
        public static String[] serializeComputedColumns​(List<org.apache.flink.table.catalog.Column> columns)
        +
      • +
      + + + +
        +
      • +

        serializeWatermarkSpecs

        +
        public static void serializeWatermarkSpecs​(Map<String,​String> map,
        +                                           List<org.apache.flink.table.catalog.WatermarkSpec> specs)
        +
      • +
      + + + +
        +
      • +

        columnKey

        +
        public static String columnKey​(int index)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/DataLakeUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/DataLakeUtils.html new file mode 100644 index 0000000000..55a6d08dd7 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/DataLakeUtils.html @@ -0,0 +1,321 @@ + + + + + +DataLakeUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class DataLakeUtils

+
+
+ +
+
    +
  • +
    +
    public class DataLakeUtils
    +extends Object
    +
    Utility class for accessing data lake related configurations.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        DataLakeUtils

        +
        public DataLakeUtils()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        extractLakeCatalogProperties

        +
        public static Map<String,​String> extractLakeCatalogProperties​(org.apache.fluss.config.Configuration tableOptions)
        +
      • +
      + + + +
        +
      • +

        getDatalakeFormat

        +
        public static Optional<org.apache.fluss.metadata.DataLakeFormat> getDatalakeFormat​(org.apache.flink.configuration.ReadableConfig tableOptions)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConnectorOptionsUtils.StartupOptions.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConnectorOptionsUtils.StartupOptions.html new file mode 100644 index 0000000000..907e5a103f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConnectorOptionsUtils.StartupOptions.html @@ -0,0 +1,329 @@ + + + + + +FlinkConnectorOptionsUtils.StartupOptions (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkConnectorOptionsUtils.StartupOptions

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils.StartupOptions
    • +
    +
  • +
+
+
    +
  • +
    +
    Enclosing class:
    +
    FlinkConnectorOptionsUtils
    +
    +
    +
    public static class FlinkConnectorOptionsUtils.StartupOptions
    +extends Object
    +
    Fluss startup options. *
    +
  • +
+
+
+ +
+
+
    +
  • + +
    + +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        StartupOptions

        +
        public StartupOptions()
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConnectorOptionsUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConnectorOptionsUtils.html new file mode 100644 index 0000000000..013bb8b14a --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConnectorOptionsUtils.html @@ -0,0 +1,424 @@ + + + + + +FlinkConnectorOptionsUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkConnectorOptionsUtils

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlinkConnectorOptionsUtils

        +
        public FlinkConnectorOptionsUtils()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        getLocalTimeZone

        +
        public static ZoneId getLocalTimeZone​(String timeZone)
        +
      • +
      + + + +
        +
      • +

        validateTableSourceOptions

        +
        public static void validateTableSourceOptions​(org.apache.flink.configuration.ReadableConfig tableOptions)
        +
      • +
      + + + + + + + +
        +
      • +

        getBucketKeys

        +
        public static List<String> getBucketKeys​(org.apache.flink.configuration.ReadableConfig tableOptions)
        +
      • +
      + + + +
        +
      • +

        getBucketKeyIndexes

        +
        public static int[] getBucketKeyIndexes​(org.apache.flink.configuration.ReadableConfig tableOptions,
        +                                        org.apache.flink.table.types.logical.RowType schema)
        +
      • +
      + + + +
        +
      • +

        parseTimestamp

        +
        public static long parseTimestamp​(String timestampStr,
        +                                  String optionKey,
        +                                  ZoneId timeZone)
        +
        Parses timestamp String to Long. + +

        timestamp String format was given as following: + +

        +     scan.startup.timestamp = 1678883047356
        +     scan.startup.timestamp = 2023-12-09 23:09:12
        + 
        +
        +
        Returns:
        +
        timestamp as long value
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConversions.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConversions.html new file mode 100644 index 0000000000..4113968e04 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConversions.html @@ -0,0 +1,422 @@ + + + + + +FlinkConversions (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkConversions

+
+
+ +
+
    +
  • +
    +
    public class FlinkConversions
    +extends Object
    +
    Utils for conversion between Flink and Fluss.
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Method Summary

      + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
      All Methods Static Methods Concrete Methods 
      Modifier and TypeMethodDescription
      static <T> org.apache.flink.configuration.ConfigOption<T>toFlinkOption​(org.apache.fluss.config.ConfigOption<T> flussOption) +
      Convert Fluss's ConfigOption to Flink's ConfigOption.
      +
      static List<org.apache.flink.configuration.ConfigOption<?>>toFlinkOptions​(Collection<org.apache.fluss.config.ConfigOption<?>> flussOption) +
      Convert Fluss's ConfigOptions to Flink's ConfigOptions.
      +
      static org.apache.flink.types.RowKindtoFlinkRowKind​(org.apache.fluss.record.ChangeType changeType) 
      static org.apache.flink.table.types.logical.RowTypetoFlinkRowType​(org.apache.fluss.types.RowType flussRowType) +
      Convert Fluss's RowType to Flink's RowType.
      +
      static org.apache.flink.table.catalog.CatalogTabletoFlinkTable​(org.apache.fluss.metadata.TableInfo tableInfo) +
      Convert Fluss's table to Flink's table.
      +
      static org.apache.flink.table.types.DataTypetoFlinkType​(org.apache.fluss.types.DataType flussDataType) +
      Convert Fluss's type to Flink's type.
      +
      static org.apache.fluss.metadata.DatabaseDescriptortoFlussDatabase​(org.apache.flink.table.catalog.CatalogDatabase catalogDatabase) +
      Convert Flink's table to Fluss's database.
      +
      static org.apache.fluss.types.RowTypetoFlussRowType​(org.apache.flink.table.types.logical.RowType flinkRowType) +
      Convert Flink's RowType to Fluss' RowType.
      +
      static org.apache.fluss.metadata.TableDescriptortoFlussTable​(org.apache.flink.table.catalog.ResolvedCatalogTable catalogTable) +
      Convert Flink's table to Fluss's table.
      +
      static org.apache.fluss.types.DataTypetoFlussType​(org.apache.flink.table.types.DataType flinkDataType) +
      Convert Flink's physical type to Fluss' type.
      +
      + +
    • +
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        toFlinkType

        +
        @VisibleForTesting
        +public static org.apache.flink.table.types.DataType toFlinkType​(org.apache.fluss.types.DataType flussDataType)
        +
        Convert Fluss's type to Flink's type.
        +
      • +
      + + + +
        +
      • +

        toFlinkRowType

        +
        public static org.apache.flink.table.types.logical.RowType toFlinkRowType​(org.apache.fluss.types.RowType flussRowType)
        +
        Convert Fluss's RowType to Flink's RowType.
        +
      • +
      + + + +
        +
      • +

        toFlussType

        +
        @VisibleForTesting
        +public static org.apache.fluss.types.DataType toFlussType​(org.apache.flink.table.types.DataType flinkDataType)
        +
        Convert Flink's physical type to Fluss' type.
        +
      • +
      + + + +
        +
      • +

        toFlussRowType

        +
        public static org.apache.fluss.types.RowType toFlussRowType​(org.apache.flink.table.types.logical.RowType flinkRowType)
        +
        Convert Flink's RowType to Fluss' RowType.
        +
      • +
      + + + +
        +
      • +

        toFlinkTable

        +
        public static org.apache.flink.table.catalog.CatalogTable toFlinkTable​(org.apache.fluss.metadata.TableInfo tableInfo)
        +
        Convert Fluss's table to Flink's table.
        +
      • +
      + + + +
        +
      • +

        toFlussTable

        +
        public static org.apache.fluss.metadata.TableDescriptor toFlussTable​(org.apache.flink.table.catalog.ResolvedCatalogTable catalogTable)
        +
        Convert Flink's table to Fluss's table.
        +
      • +
      + + + +
        +
      • +

        toFlussDatabase

        +
        public static org.apache.fluss.metadata.DatabaseDescriptor toFlussDatabase​(org.apache.flink.table.catalog.CatalogDatabase catalogDatabase)
        +
        Convert Flink's table to Fluss's database.
        +
      • +
      + + + +
        +
      • +

        toFlinkOptions

        +
        public static List<org.apache.flink.configuration.ConfigOption<?>> toFlinkOptions​(Collection<org.apache.fluss.config.ConfigOption<?>> flussOption)
        +
        Convert Fluss's ConfigOptions to Flink's ConfigOptions.
        +
      • +
      + + + +
        +
      • +

        toFlinkOption

        +
        public static <T> org.apache.flink.configuration.ConfigOption<T> toFlinkOption​(org.apache.fluss.config.ConfigOption<T> flussOption)
        +
        Convert Fluss's ConfigOption to Flink's ConfigOption.
        +
      • +
      + + + +
        +
      • +

        toFlinkRowKind

        +
        public static org.apache.flink.types.RowKind toFlinkRowKind​(org.apache.fluss.record.ChangeType changeType)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkRowToFlussRowConverter.FlussSerializationConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkRowToFlussRowConverter.FlussSerializationConverter.html new file mode 100644 index 0000000000..b9e7393b33 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkRowToFlussRowConverter.FlussSerializationConverter.html @@ -0,0 +1,275 @@ + + + + + +FlinkRowToFlussRowConverter.FlussSerializationConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface FlinkRowToFlussRowConverter.FlussSerializationConverter

+
+
+
+
    +
  • +
    +
    All Superinterfaces:
    +
    Serializable
    +
    +
    +
    Enclosing class:
    +
    FlinkRowToFlussRowConverter
    +
    +
    +
    Functional Interface:
    +
    This is a functional interface and can therefore be used as the assignment target for a lambda expression or method reference.
    +
    +
    +
    @FunctionalInterface
    +public static interface FlinkRowToFlussRowConverter.FlussSerializationConverter
    +extends Serializable
    +
    Runtime converter to convert field in Flink's RowData to Fluss's InternalRow + type object.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        serialize

        +
        Object serialize​(Object flinkField)
        +
        Convert a Flink field object of RowData to the Fluss's internal data structure + object.
        +
        +
        Parameters:
        +
        flinkField - A single field of a RowData
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkRowToFlussRowConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkRowToFlussRowConverter.html new file mode 100644 index 0000000000..156ba82ca7 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkRowToFlussRowConverter.html @@ -0,0 +1,371 @@ + + + + + +FlinkRowToFlussRowConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkRowToFlussRowConverter

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.utils.FlinkRowToFlussRowConverter
    • +
    +
  • +
+
+
    +
  • +
    +
    All Implemented Interfaces:
    +
    AutoCloseable
    +
    +
    +
    @Deprecated
    +public class FlinkRowToFlussRowConverter
    +extends Object
    +implements AutoCloseable
    +
    Deprecated. +
    This class is deprecated and will be removed in future versions. It is recommended to + use the FlinkAsFlussRow to wraps Flink's RowData as Fluss's InternalRow.
    +
    +
    A converter to convert Flink's RowData to Fluss's InternalRow.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    + +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkUtils.html new file mode 100644 index 0000000000..a7bbc1a7d3 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkUtils.html @@ -0,0 +1,317 @@ + + + + + +FlinkUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlinkUtils

+
+
+ +
+
    +
  • +
    +
    public class FlinkUtils
    +extends Object
    +
    Utils for Flink classes.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FlinkUtils

        +
        public FlinkUtils()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        projectRowType

        +
        public static org.apache.flink.table.types.logical.RowType projectRowType​(org.apache.flink.table.types.logical.RowType rowType,
        +                                                                          int[] projection)
        +
        Returns projected RowType by given projection indexes over original RowType.
        +
        +
        Parameters:
        +
        rowType - the original row type
        +
        projection - the projection indexes
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToFlinkRowConverter.FlussDeserializationConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToFlinkRowConverter.FlussDeserializationConverter.html new file mode 100644 index 0000000000..32b142dd94 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToFlinkRowConverter.FlussDeserializationConverter.html @@ -0,0 +1,275 @@ + + + + + +FlussRowToFlinkRowConverter.FlussDeserializationConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface FlussRowToFlinkRowConverter.FlussDeserializationConverter

+
+
+
+
    +
  • +
    +
    All Superinterfaces:
    +
    Serializable
    +
    +
    +
    Enclosing class:
    +
    FlussRowToFlinkRowConverter
    +
    +
    +
    Functional Interface:
    +
    This is a functional interface and can therefore be used as the assignment target for a lambda expression or method reference.
    +
    +
    +
    @FunctionalInterface
    +public static interface FlussRowToFlinkRowConverter.FlussDeserializationConverter
    +extends Serializable
    +
    Runtime converter to convert field in Fluss's InternalRow to Flink's RowData + type object.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        deserialize

        +
        Object deserialize​(Object flussField)
        +
        Convert a Fluss field object of InternalRow to the Flink's internal data + structure object.
        +
        +
        Parameters:
        +
        flussField - A single field of a InternalRow
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToFlinkRowConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToFlinkRowConverter.html new file mode 100644 index 0000000000..d115d3317e --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToFlinkRowConverter.html @@ -0,0 +1,381 @@ + + + + + +FlussRowToFlinkRowConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlussRowToFlinkRowConverter

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter
    • +
    +
  • +
+
+
    +
  • +
    +
    public class FlussRowToFlinkRowConverter
    +extends Object
    +
    A converter to convert Fluss's InternalRow to Flink's RowData. + +

    Note: fluss-datalake-tiering also contains the same class, we need to keep them in sync if we + modify this class.

    +
  • +
+
+
+ +
+
+ +
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToJsonConverters.FlussRowToJsonConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToJsonConverters.FlussRowToJsonConverter.html new file mode 100644 index 0000000000..cfc8c61e0c --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToJsonConverters.FlussRowToJsonConverter.html @@ -0,0 +1,269 @@ + + + + + +FlussRowToJsonConverters.FlussRowToJsonConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Interface FlussRowToJsonConverters.FlussRowToJsonConverter

+
+
+
+
    +
  • +
    +
    All Superinterfaces:
    +
    Serializable
    +
    +
    +
    Enclosing class:
    +
    FlussRowToJsonConverters
    +
    +
    +
    Functional Interface:
    +
    This is a functional interface and can therefore be used as the assignment target for a lambda expression or method reference.
    +
    +
    +
    @FunctionalInterface
    +public static interface FlussRowToJsonConverters.FlussRowToJsonConverter
    +extends Serializable
    +
    Runtime converter that converts objects of Fluss data structures to corresponding JsonNodes.
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Method Summary

      + + + + + + + + + + + + +
      All Methods Instance Methods Abstract Methods 
      Modifier and TypeMethodDescription
      org.apache.fluss.shaded.jackson2.com.fasterxml.jackson.databind.JsonNodeconvert​(org.apache.fluss.shaded.jackson2.com.fasterxml.jackson.databind.ObjectMapper mapper, + org.apache.fluss.shaded.jackson2.com.fasterxml.jackson.databind.JsonNode reuse, + Object value) 
      +
    • +
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        convert

        +
        org.apache.fluss.shaded.jackson2.com.fasterxml.jackson.databind.JsonNode convert​(org.apache.fluss.shaded.jackson2.com.fasterxml.jackson.databind.ObjectMapper mapper,
        +                                                                                 org.apache.fluss.shaded.jackson2.com.fasterxml.jackson.databind.JsonNode reuse,
        +                                                                                 Object value)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToJsonConverters.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToJsonConverters.html new file mode 100644 index 0000000000..9a46068d72 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToJsonConverters.html @@ -0,0 +1,332 @@ + + + + + +FlussRowToJsonConverters (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class FlussRowToJsonConverters

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.utils.FlussRowToJsonConverters
    • +
    +
  • +
+
+
    +
  • +
    +
    public class FlussRowToJsonConverters
    +extends Object
    +
    A converter to convert Fluss's InternalRow to JsonNode.
    +
  • +
+
+
+ +
+
+ +
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/LakeSourceUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/LakeSourceUtils.html new file mode 100644 index 0000000000..b4b71f1834 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/LakeSourceUtils.html @@ -0,0 +1,356 @@ + + + + + +LakeSourceUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class LakeSourceUtils

+
+
+ +
+
    +
  • +
    +
    public class LakeSourceUtils
    +extends Object
    +
    Utils for create lake source.
    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + +
        +
      • +

        LOG

        +
        public static final org.slf4j.Logger LOG
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        LakeSourceUtils

        +
        public LakeSourceUtils()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        createLakeSource

        +
        @Nullable
        +public static org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> createLakeSource​(org.apache.fluss.metadata.TablePath tablePath,
        +                                                                                                               Map<String,​String> properties)
        +
        Return the lake source of the given table. Return null when the lake storage doesn't support + create lake source.
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PojoToRowConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PojoToRowConverter.html new file mode 100644 index 0000000000..70060c64d3 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PojoToRowConverter.html @@ -0,0 +1,349 @@ + + + + + +PojoToRowConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class PojoToRowConverter<T>

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.utils.PojoToRowConverter<T>
    • +
    +
  • +
+
+
    +
  • +
    +
    Type Parameters:
    +
    T - The POJO type to convert
    +
    +
    +
    public class PojoToRowConverter<T>
    +extends Object
    +
    Utility class for converting Java POJOs to Fluss's InternalRow format. + +

    This utility uses Flink's POJO type information to map fields from POJOs to InternalRow based + on a given schema. + +

    Example usage: + +

    
    + // Create a converter
    + PojoToRowConverter<Order> converter =
    +     new PojoToRowConverter<>(Order.class, rowType);
    +
    + // Convert a POJO to GenericRow
    + Order order = new Order(1001L, 5001L, 10, "123 Mumbai");
    + GenericRow row = converter.convert(order);
    + 
    + +

    Note: Nested POJO fields are not supported in the current implementation.

    +
  • +
+
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        PojoToRowConverter

        +
        public PojoToRowConverter​(Class<T> pojoClass,
        +                          org.apache.fluss.types.RowType rowType)
        +
        Creates a new converter for the specified POJO class and row type.
        +
        +
        Parameters:
        +
        pojoClass - The class of POJOs to convert
        +
        rowType - The row schema to use for conversion
        +
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + + + +
        +
      • +

        convert

        +
        public org.apache.fluss.row.GenericRow convert​(T pojo)
        +
        Converts a POJO to a GenericRow object according to the schema.
        +
        +
        Parameters:
        +
        pojo - The POJO to convert
        +
        Returns:
        +
        The converted GenericRow, or null if the input is null
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ +
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.FieldEqual.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.FieldEqual.html new file mode 100644 index 0000000000..2ffa75b64b --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.FieldEqual.html @@ -0,0 +1,340 @@ + + + + + +PushdownUtils.FieldEqual (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class PushdownUtils.FieldEqual

+
+
+
    +
  • java.lang.Object
  • +
  • +
      +
    • org.apache.fluss.flink.utils.PushdownUtils.FieldEqual
    • +
    +
  • +
+
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Field Detail

      + + + +
        +
      • +

        fieldIndex

        +
        public final int fieldIndex
        +
      • +
      + + + +
        +
      • +

        equalValue

        +
        public final Object equalValue
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        FieldEqual

        +
        public FieldEqual​(int fieldIndex,
        +                  Object equalValue)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.ValueConversion.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.ValueConversion.html new file mode 100644 index 0000000000..53f6d6aeeb --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.ValueConversion.html @@ -0,0 +1,383 @@ + + + + + +PushdownUtils.ValueConversion (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Enum PushdownUtils.ValueConversion

+
+
+ +
+ +
+
+ +
+
+
    +
  • + +
    + +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        values

        +
        public static PushdownUtils.ValueConversion[] values()
        +
        Returns an array containing the constants of this enum type, in +the order they are declared. This method may be used to iterate +over the constants as follows: +
        +for (PushdownUtils.ValueConversion c : PushdownUtils.ValueConversion.values())
        +    System.out.println(c);
        +
        +
        +
        Returns:
        +
        an array containing the constants of this enum type, in the order they are declared
        +
        +
      • +
      + + + +
        +
      • +

        valueOf

        +
        public static PushdownUtils.ValueConversion valueOf​(String name)
        +
        Returns the enum constant of this type with the specified name. +The string must match exactly an identifier used to declare an +enum constant in this type. (Extraneous whitespace characters are +not permitted.)
        +
        +
        Parameters:
        +
        name - the name of the enum constant to be returned.
        +
        Returns:
        +
        the enum constant with the specified name
        +
        Throws:
        +
        IllegalArgumentException - if this enum type has no constant with the specified name
        +
        NullPointerException - if the argument is null
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.html new file mode 100644 index 0000000000..bb5dcd7710 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.html @@ -0,0 +1,434 @@ + + + + + +PushdownUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Class PushdownUtils

+
+
+ +
+
    +
  • +
    +
    public class PushdownUtils
    +extends Object
    +
    Utilities for pushdown abilities.
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Nested Class Summary

      + + + + + + + + + + + + + + + + + +
      Nested Classes 
      Modifier and TypeClassDescription
      static class PushdownUtils.FieldEqual +
      A structure represents a source field equal literal expression.
      +
      static class PushdownUtils.ValueConversion +
      The value conversion type between Flink internal value and Fluss internal value.
      +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Constructor Summary

      + + + + + + + + + + +
      Constructors 
      ConstructorDescription
      PushdownUtils() 
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Summary

      + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
      All Methods Static Methods Concrete Methods 
      Modifier and TypeMethodDescription
      static longcountLogTable​(org.apache.fluss.metadata.TablePath tablePath, + org.apache.fluss.config.Configuration flussConfig) 
      static voiddeleteSingleRow​(org.apache.fluss.row.GenericRow deleteRow, + org.apache.fluss.metadata.TablePath tablePath, + org.apache.fluss.config.Configuration flussConfig) 
      static List<PushdownUtils.FieldEqual>extractFieldEquals​(List<org.apache.flink.table.expressions.ResolvedExpression> expressions, + Map<Integer,​org.apache.flink.table.types.logical.LogicalType> fieldIndexToType, + List<org.apache.flink.table.expressions.ResolvedExpression> acceptedFiltersResult, + List<org.apache.flink.table.expressions.ResolvedExpression> remainingFiltersResult, + PushdownUtils.ValueConversion valueConversion) +
      Extract field equality information from expressions.
      +
      static Collection<org.apache.flink.table.data.RowData>limitScan​(org.apache.fluss.metadata.TablePath tablePath, + org.apache.fluss.config.Configuration flussConfig, + org.apache.flink.table.types.logical.RowType sourceOutputType, + int[] projectedFields, + long limitRowNum) 
      static Collection<org.apache.flink.table.data.RowData>querySingleRow​(org.apache.flink.table.data.GenericRowData lookupRow, + org.apache.fluss.metadata.TablePath tablePath, + org.apache.fluss.config.Configuration flussConfig, + org.apache.flink.table.types.logical.RowType sourceOutputType, + int[] primaryKeyIndexes, + int lookupMaxRetryTimes, + int[] projectedFields) 
      + +
    • +
    +
    +
  • +
+
+
+
    +
  • + +
    +
      +
    • + + +

      Constructor Detail

      + + + +
        +
      • +

        PushdownUtils

        +
        public PushdownUtils()
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        extractFieldEquals

        +
        public static List<PushdownUtils.FieldEqual> extractFieldEquals​(List<org.apache.flink.table.expressions.ResolvedExpression> expressions,
        +                                                                Map<Integer,​org.apache.flink.table.types.logical.LogicalType> fieldIndexToType,
        +                                                                List<org.apache.flink.table.expressions.ResolvedExpression> acceptedFiltersResult,
        +                                                                List<org.apache.flink.table.expressions.ResolvedExpression> remainingFiltersResult,
        +                                                                PushdownUtils.ValueConversion valueConversion)
        +
        Extract field equality information from expressions.
        +
      • +
      + + + +
        +
      • +

        querySingleRow

        +
        public static Collection<org.apache.flink.table.data.RowData> querySingleRow​(org.apache.flink.table.data.GenericRowData lookupRow,
        +                                                                             org.apache.fluss.metadata.TablePath tablePath,
        +                                                                             org.apache.fluss.config.Configuration flussConfig,
        +                                                                             org.apache.flink.table.types.logical.RowType sourceOutputType,
        +                                                                             int[] primaryKeyIndexes,
        +                                                                             int lookupMaxRetryTimes,
        +                                                                             @Nullable
        +                                                                             int[] projectedFields)
        +
      • +
      + + + +
        +
      • +

        deleteSingleRow

        +
        public static void deleteSingleRow​(org.apache.fluss.row.GenericRow deleteRow,
        +                                   org.apache.fluss.metadata.TablePath tablePath,
        +                                   org.apache.fluss.config.Configuration flussConfig)
        +
      • +
      + + + +
        +
      • +

        limitScan

        +
        public static Collection<org.apache.flink.table.data.RowData> limitScan​(org.apache.fluss.metadata.TablePath tablePath,
        +                                                                        org.apache.fluss.config.Configuration flussConfig,
        +                                                                        org.apache.flink.table.types.logical.RowType sourceOutputType,
        +                                                                        @Nullable
        +                                                                        int[] projectedFields,
        +                                                                        long limitRowNum)
        +
      • +
      + + + +
        +
      • +

        countLogTable

        +
        public static long countLogTable​(org.apache.fluss.metadata.TablePath tablePath,
        +                                 org.apache.fluss.config.Configuration flussConfig)
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/TimestampFormat.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/TimestampFormat.html new file mode 100644 index 0000000000..042594e097 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/TimestampFormat.html @@ -0,0 +1,390 @@ + + + + + +TimestampFormat (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+ +
+
+ +

Enum TimestampFormat

+
+
+ +
+ +
+
+ +
+
+
    +
  • + +
    +
      +
    • + + +

      Enum Constant Detail

      + + + +
        +
      • +

        SQL

        +
        public static final TimestampFormat SQL
        +
        Options to specify TIMESTAMP/TIMESTAMP_WITH_LOCAL_ZONE format. It will parse TIMESTAMP in + "yyyy-MM-dd HH:mm:ss.s{precision}" format, TIMESTAMP_WITH_LOCAL_TIMEZONE in "yyyy-MM-dd + HH:mm:ss.s{precision}'Z'" and output in the same format.
        +
      • +
      + + + +
        +
      • +

        ISO_8601

        +
        public static final TimestampFormat ISO_8601
        +
        Options to specify TIMESTAMP/TIMESTAMP_WITH_LOCAL_ZONE format. It will parse TIMESTAMP in + "yyyy-MM-ddTHH:mm:ss.s{precision}" format, TIMESTAMP_WITH_LOCAL_TIMEZONE in + "yyyy-MM-ddTHH:mm:ss.s{precision}'Z'" and output in the same format.
        +
      • +
      +
    • +
    +
    + +
    +
      +
    • + + +

      Method Detail

      + + + +
        +
      • +

        values

        +
        public static TimestampFormat[] values()
        +
        Returns an array containing the constants of this enum type, in +the order they are declared. This method may be used to iterate +over the constants as follows: +
        +for (TimestampFormat c : TimestampFormat.values())
        +    System.out.println(c);
        +
        +
        +
        Returns:
        +
        an array containing the constants of this enum type, in the order they are declared
        +
        +
      • +
      + + + +
        +
      • +

        valueOf

        +
        public static TimestampFormat valueOf​(String name)
        +
        Returns the enum constant of this type with the specified name. +The string must match exactly an identifier used to declare an +enum constant in this type. (Extraneous whitespace characters are +not permitted.)
        +
        +
        Parameters:
        +
        name - the name of the enum constant to be returned.
        +
        Returns:
        +
        the enum constant with the specified name
        +
        Throws:
        +
        IllegalArgumentException - if this enum type has no constant with the specified name
        +
        NullPointerException - if the argument is null
        +
        +
      • +
      +
    • +
    +
    +
  • +
+
+
+
+ + + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/CatalogExceptionUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/CatalogExceptionUtils.html new file mode 100644 index 0000000000..13e558da9b --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/CatalogExceptionUtils.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.utils.CatalogExceptionUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.utils.CatalogExceptionUtils

+
+
No usage of org.apache.fluss.flink.utils.CatalogExceptionUtils
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/CatalogPropertiesUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/CatalogPropertiesUtils.html new file mode 100644 index 0000000000..3e4c6b3afc --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/CatalogPropertiesUtils.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.utils.CatalogPropertiesUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.utils.CatalogPropertiesUtils

+
+
No usage of org.apache.fluss.flink.utils.CatalogPropertiesUtils
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/DataLakeUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/DataLakeUtils.html new file mode 100644 index 0000000000..3473ba3ab3 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/DataLakeUtils.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.utils.DataLakeUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.utils.DataLakeUtils

+
+
No usage of org.apache.fluss.flink.utils.DataLakeUtils
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConnectorOptionsUtils.StartupOptions.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConnectorOptionsUtils.StartupOptions.html new file mode 100644 index 0000000000..89bc5031e9 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConnectorOptionsUtils.StartupOptions.html @@ -0,0 +1,231 @@ + + + + + +Uses of Class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils.StartupOptions (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils.StartupOptions

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConnectorOptionsUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConnectorOptionsUtils.html new file mode 100644 index 0000000000..6526d73f12 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConnectorOptionsUtils.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils

+
+
No usage of org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConversions.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConversions.html new file mode 100644 index 0000000000..e57305db49 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConversions.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.utils.FlinkConversions (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.utils.FlinkConversions

+
+
No usage of org.apache.fluss.flink.utils.FlinkConversions
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkRowToFlussRowConverter.FlussSerializationConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkRowToFlussRowConverter.FlussSerializationConverter.html new file mode 100644 index 0000000000..1f1d9df228 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkRowToFlussRowConverter.FlussSerializationConverter.html @@ -0,0 +1,148 @@ + + + + + +Uses of Interface org.apache.fluss.flink.utils.FlinkRowToFlussRowConverter.FlussSerializationConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.flink.utils.FlinkRowToFlussRowConverter.FlussSerializationConverter

+
+
No usage of org.apache.fluss.flink.utils.FlinkRowToFlussRowConverter.FlussSerializationConverter
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkRowToFlussRowConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkRowToFlussRowConverter.html new file mode 100644 index 0000000000..ae401c9c5f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkRowToFlussRowConverter.html @@ -0,0 +1,204 @@ + + + + + +Uses of Class org.apache.fluss.flink.utils.FlinkRowToFlussRowConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.utils.FlinkRowToFlussRowConverter

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkUtils.html new file mode 100644 index 0000000000..e66391c0b1 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkUtils.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.utils.FlinkUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.utils.FlinkUtils

+
+
No usage of org.apache.fluss.flink.utils.FlinkUtils
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToFlinkRowConverter.FlussDeserializationConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToFlinkRowConverter.FlussDeserializationConverter.html new file mode 100644 index 0000000000..f6c68bd6a7 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToFlinkRowConverter.FlussDeserializationConverter.html @@ -0,0 +1,214 @@ + + + + + +Uses of Interface org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter.FlussDeserializationConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter.FlussDeserializationConverter

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToFlinkRowConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToFlinkRowConverter.html new file mode 100644 index 0000000000..52d36a7010 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToFlinkRowConverter.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter

+
+
No usage of org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToJsonConverters.FlussRowToJsonConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToJsonConverters.FlussRowToJsonConverter.html new file mode 100644 index 0000000000..6ca52dd26f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToJsonConverters.FlussRowToJsonConverter.html @@ -0,0 +1,192 @@ + + + + + +Uses of Interface org.apache.fluss.flink.utils.FlussRowToJsonConverters.FlussRowToJsonConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Interface
org.apache.fluss.flink.utils.FlussRowToJsonConverters.FlussRowToJsonConverter

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToJsonConverters.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToJsonConverters.html new file mode 100644 index 0000000000..7041db0bc6 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToJsonConverters.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.utils.FlussRowToJsonConverters (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.utils.FlussRowToJsonConverters

+
+
No usage of org.apache.fluss.flink.utils.FlussRowToJsonConverters
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/LakeSourceUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/LakeSourceUtils.html new file mode 100644 index 0000000000..7fc2e1ba97 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/LakeSourceUtils.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.utils.LakeSourceUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.utils.LakeSourceUtils

+
+
No usage of org.apache.fluss.flink.utils.LakeSourceUtils
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PojoToRowConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PojoToRowConverter.html new file mode 100644 index 0000000000..fe9bf965b7 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PojoToRowConverter.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.utils.PojoToRowConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.utils.PojoToRowConverter

+
+
No usage of org.apache.fluss.flink.utils.PojoToRowConverter
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.FieldEqual.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.FieldEqual.html new file mode 100644 index 0000000000..b4b1e60743 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.FieldEqual.html @@ -0,0 +1,292 @@ + + + + + +Uses of Class org.apache.fluss.flink.utils.PushdownUtils.FieldEqual (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.utils.PushdownUtils.FieldEqual

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.ValueConversion.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.ValueConversion.html new file mode 100644 index 0000000000..eb102e9728 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.ValueConversion.html @@ -0,0 +1,223 @@ + + + + + +Uses of Class org.apache.fluss.flink.utils.PushdownUtils.ValueConversion (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.utils.PushdownUtils.ValueConversion

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.html new file mode 100644 index 0000000000..7e9d3b75f9 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.html @@ -0,0 +1,148 @@ + + + + + +Uses of Class org.apache.fluss.flink.utils.PushdownUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.utils.PushdownUtils

+
+
No usage of org.apache.fluss.flink.utils.PushdownUtils
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/TimestampFormat.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/TimestampFormat.html new file mode 100644 index 0000000000..2d3fcaf69c --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/TimestampFormat.html @@ -0,0 +1,215 @@ + + + + + +Uses of Class org.apache.fluss.flink.utils.TimestampFormat (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Class
org.apache.fluss.flink.utils.TimestampFormat

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-summary.html new file mode 100644 index 0000000000..e847293228 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-summary.html @@ -0,0 +1,300 @@ + + + + + +org.apache.fluss.flink.utils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Package org.apache.fluss.flink.utils

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-tree.html new file mode 100644 index 0000000000..1a9951226e --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-tree.html @@ -0,0 +1,203 @@ + + + + + +org.apache.fluss.flink.utils Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For Package org.apache.fluss.flink.utils

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+

Interface Hierarchy

+ +
+
+

Enum Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-use.html new file mode 100644 index 0000000000..c401e778b1 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-use.html @@ -0,0 +1,276 @@ + + + + + +Uses of Package org.apache.fluss.flink.utils (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Uses of Package
org.apache.fluss.flink.utils

+
+
+ +
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/overview-summary.html b/website/static/javadoc/0.8-SNAPSHOT/overview-summary.html new file mode 100644 index 0000000000..f78f7d818f --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/overview-summary.html @@ -0,0 +1,23 @@ + + + + + +fluss-javadoc-temp 0.8-SNAPSHOT API + + + + + + + +
+ +

index.html

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/overview-tree.html b/website/static/javadoc/0.8-SNAPSHOT/overview-tree.html new file mode 100644 index 0000000000..430dcc2918 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/overview-tree.html @@ -0,0 +1,581 @@ + + + + + +Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
+ +
+
+
+

Hierarchy For All Packages

+Package Hierarchies: + +
+
+
+

Class Hierarchy

+ +
+
+

Interface Hierarchy

+ +
+
+

Enum Hierarchy

+ +
+
+
+
+ +

Copyright © 2025. All rights reserved.

+
+ + diff --git a/website/static/javadoc/0.8-SNAPSHOT/package-search-index.js b/website/static/javadoc/0.8-SNAPSHOT/package-search-index.js new file mode 100644 index 0000000000..9c696ae856 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/package-search-index.js @@ -0,0 +1 @@ +packageSearchIndex = [{"l":"All Packages","url":"allpackages-index.html"},{"l":"org.apache.fluss.client"},{"l":"org.apache.fluss.client.admin"},{"l":"org.apache.fluss.client.lookup"},{"l":"org.apache.fluss.client.metadata"},{"l":"org.apache.fluss.client.metrics"},{"l":"org.apache.fluss.client.table"},{"l":"org.apache.fluss.client.table.getter"},{"l":"org.apache.fluss.client.table.scanner"},{"l":"org.apache.fluss.client.table.scanner.batch"},{"l":"org.apache.fluss.client.table.scanner.log"},{"l":"org.apache.fluss.client.table.writer"},{"l":"org.apache.fluss.client.token"},{"l":"org.apache.fluss.client.utils"},{"l":"org.apache.fluss.client.write"},{"l":"org.apache.fluss.flink"},{"l":"org.apache.fluss.flink.adapter"},{"l":"org.apache.fluss.flink.catalog"},{"l":"org.apache.fluss.flink.lake"},{"l":"org.apache.fluss.flink.lake.reader"},{"l":"org.apache.fluss.flink.lake.split"},{"l":"org.apache.fluss.flink.lake.state"},{"l":"org.apache.fluss.flink.metrics"},{"l":"org.apache.fluss.flink.procedure"},{"l":"org.apache.fluss.flink.row"},{"l":"org.apache.fluss.flink.sink"},{"l":"org.apache.fluss.flink.sink.serializer"},{"l":"org.apache.fluss.flink.sink.writer"},{"l":"org.apache.fluss.flink.source"},{"l":"org.apache.fluss.flink.source.deserializer"},{"l":"org.apache.fluss.flink.source.emitter"},{"l":"org.apache.fluss.flink.source.enumerator"},{"l":"org.apache.fluss.flink.source.enumerator.initializer"},{"l":"org.apache.fluss.flink.source.event"},{"l":"org.apache.fluss.flink.source.lookup"},{"l":"org.apache.fluss.flink.source.metrics"},{"l":"org.apache.fluss.flink.source.reader"},{"l":"org.apache.fluss.flink.source.reader.fetcher"},{"l":"org.apache.fluss.flink.source.split"},{"l":"org.apache.fluss.flink.source.state"},{"l":"org.apache.fluss.flink.tiering"},{"l":"org.apache.fluss.flink.tiering.committer"},{"l":"org.apache.fluss.flink.tiering.event"},{"l":"org.apache.fluss.flink.tiering.source"},{"l":"org.apache.fluss.flink.tiering.source.enumerator"},{"l":"org.apache.fluss.flink.tiering.source.split"},{"l":"org.apache.fluss.flink.tiering.source.state"},{"l":"org.apache.fluss.flink.utils"}] \ No newline at end of file diff --git a/website/static/javadoc/0.8-SNAPSHOT/package-search-index.zip b/website/static/javadoc/0.8-SNAPSHOT/package-search-index.zip new file mode 100644 index 0000000000000000000000000000000000000000..605b6f06024075e1f995de828ea2f7451e1493e0 GIT binary patch literal 489 zcmWIWW@Zs#;Nak3Sj%l3&42_Xf$W0BR3-ORj{)C#?<;{3d+Q+)l7 z81S^NZ`vhS#87sp!8Ylk=iO_oW`}ehRuCv({>6Qfa^UXSdXwIpo;$rce!tebd8^W& zwTpkS-7;gfcj%?}N(TL}Zz;D**~h3ehu@#>DjfVefyt_bclw;U>4y|6y)@FMJruqP3r&EEvk6! z@7(@B^Rs3+t{J-)pW25oWL8EG+2E4~J&oAmpz&;2l6B;fCb!@Cob zyX0O^y7GT%#pMY(740_l27(5)ug;y>A+}SrSGDr16!+6Co4~_L2OImM%l+Gp5|}!l zv-UBaJbCLL|MC}oSC>8EGh3FiG&l49!F2~Sqh`7;Q@a`AJ)6VVk|XQz`Wc@KHylk6 zo)MSUoGWypvl3&5a7+oB*K7*4P-e` aY@h<1aTDOp$_7%(2!!rHx(8@H0|Nl=0=Q5B literal 0 HcmV?d00001 diff --git a/website/static/javadoc/0.8-SNAPSHOT/packages b/website/static/javadoc/0.8-SNAPSHOT/packages new file mode 100644 index 0000000000..07a069f6f9 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/packages @@ -0,0 +1,185 @@ +org.apache.fluss.metrics +org.apache.fluss.metrics.reporter +org.apache.fluss.metrics.utils +org.apache.fluss.metrics.groups +org.apache.fluss.metrics.registry +org.apache.fluss.rocksdb +org.apache.fluss.cluster +org.apache.fluss.predicate +org.apache.fluss.types +org.apache.fluss.record +org.apache.fluss.record.bytesview +org.apache.fluss.record.send +org.apache.fluss.memory +org.apache.fluss.config +org.apache.fluss.security.auth +org.apache.fluss.security.auth.sasl.plain +org.apache.fluss.security.auth.sasl.authenticator +org.apache.fluss.security.auth.sasl.jaas +org.apache.fluss.security.acl +org.apache.fluss.plugin +org.apache.fluss.annotation +org.apache.fluss.utils +org.apache.fluss.utils.types +org.apache.fluss.utils.crc +org.apache.fluss.utils.function +org.apache.fluss.utils.json +org.apache.fluss.utils.concurrent +org.apache.fluss.utils.log +org.apache.fluss.utils.clock +org.apache.fluss.classloading +org.apache.fluss.row +org.apache.fluss.row.columnar +org.apache.fluss.row.compacted +org.apache.fluss.row.arrow.writers +org.apache.fluss.row.arrow.vectors +org.apache.fluss.row.arrow +org.apache.fluss.row.encode +org.apache.fluss.row.encode.paimon +org.apache.fluss.row.encode.iceberg +org.apache.fluss.row.indexed +org.apache.fluss.row.decode +org.apache.fluss.lake.lakestorage +org.apache.fluss.lake.serializer +org.apache.fluss.lake.source +org.apache.fluss.lake.writer +org.apache.fluss.lake.committer +org.apache.fluss.lake.batch +org.apache.fluss.compression +org.apache.fluss.bucketing +org.apache.fluss.rpc.messages +org.apache.fluss.fs.token +org.apache.fluss.fs +org.apache.fluss.fs.utils +org.apache.fluss.fs.local +org.apache.fluss.exception +org.apache.fluss.metadata +org.apache.fluss.remote +org.apache.fluss.metrics.prometheus +org.apache.fluss.metrics.jmx +org.apache.fluss.client.write +org.apache.fluss.client.token +org.apache.fluss.client.metrics +org.apache.fluss.client +org.apache.fluss.client.admin +org.apache.fluss.client.lookup +org.apache.fluss.client.utils +org.apache.fluss.client.table.getter +org.apache.fluss.client.table +org.apache.fluss.client.table.writer +org.apache.fluss.client.table.scanner +org.apache.fluss.client.table.scanner.batch +org.apache.fluss.client.table.scanner.log +org.apache.fluss.client.metadata +org.apache.fluss.rpc.metrics +org.apache.fluss.rpc +org.apache.fluss.rpc.util +org.apache.fluss.rpc.entity +org.apache.fluss.rpc.protocol +org.apache.fluss.rpc.netty +org.apache.fluss.rpc.netty.server +org.apache.fluss.rpc.netty.client +org.apache.fluss.rpc.gateway +org.apache.fluss.dist +org.apache.fluss.fs.hdfs +org.apache.fluss.fs.hdfs.utils +org.apache.fluss.fs.oss.token +org.apache.fluss.fs.oss +org.apache.fluss.fs.s3.token +org.apache.fluss.fs.s3 +com.amazonaws.services.s3.model.transform +org.apache.fluss.fs.gs +org.apache.fluss.fs.obs.token +org.apache.fluss.fs.obs +org.apache.fluss.testutils.junit.parameterized +org.apache.fluss.testutils.common +org.apache.fluss.server.metrics.group +org.apache.fluss.server.metrics +org.apache.fluss.server.tablet +org.apache.fluss.server.zk +org.apache.fluss.server.zk.data +org.apache.fluss.server.coordinator +org.apache.fluss.server.coordinator.statemachine +org.apache.fluss.server.coordinator.event +org.apache.fluss.server.coordinator.event.watcher +org.apache.fluss.server.kv.rocksdb +org.apache.fluss.server.kv.snapshot +org.apache.fluss.server.kv.prewrite +org.apache.fluss.server.kv +org.apache.fluss.server.kv.wal +org.apache.fluss.server.kv.partialupdate +org.apache.fluss.server.kv.rowmerger +org.apache.fluss.server +org.apache.fluss.server.entity +org.apache.fluss.server.utils +org.apache.fluss.server.utils.timer +org.apache.fluss.server.cli +org.apache.fluss.server.authorizer +org.apache.fluss.server.log +org.apache.fluss.server.log.checkpoint +org.apache.fluss.server.log.remote +org.apache.fluss.server.replica +org.apache.fluss.server.replica.fetcher +org.apache.fluss.server.replica.delay +org.apache.fluss.server.exception +org.apache.fluss.server.metadata +org.apache.fluss.flink.metrics +org.apache.fluss.flink.catalog +org.apache.fluss.flink.source.reader +org.apache.fluss.flink.source.reader.fetcher +org.apache.fluss.flink.source.split +org.apache.fluss.flink.source.metrics +org.apache.fluss.flink.source +org.apache.fluss.flink.source.deserializer +org.apache.fluss.flink.source.emitter +org.apache.fluss.flink.source.lookup +org.apache.fluss.flink.source.state +org.apache.fluss.flink.source.event +org.apache.fluss.flink.source.enumerator.initializer +org.apache.fluss.flink.source.enumerator +org.apache.fluss.flink.tiering.source.split +org.apache.fluss.flink.tiering.source +org.apache.fluss.flink.tiering.source.state +org.apache.fluss.flink.tiering.source.enumerator +org.apache.fluss.flink.tiering +org.apache.fluss.flink.tiering.committer +org.apache.fluss.flink.tiering.event +org.apache.fluss.flink.sink +org.apache.fluss.flink.sink.serializer +org.apache.fluss.flink.sink.writer +org.apache.fluss.flink.utils +org.apache.fluss.flink +org.apache.fluss.flink.adapter +org.apache.fluss.flink.procedure +org.apache.fluss.flink.row +org.apache.fluss.flink.lake.reader +org.apache.fluss.flink.lake.split +org.apache.fluss.flink.lake +org.apache.fluss.flink.lake.state +org.apache.flink.streaming.api.connector.sink2 +org.apache.flink.api.connector.sink2 +org.apache.fluss.lake.paimon +org.apache.fluss.lake.paimon.source +org.apache.fluss.lake.paimon.tiering +org.apache.fluss.lake.paimon.tiering.append +org.apache.fluss.lake.paimon.tiering.mergetree +org.apache.fluss.lake.paimon.utils +org.apache.fluss.lake.iceberg +org.apache.fluss.lake.iceberg.source +org.apache.fluss.lake.iceberg.tiering +org.apache.fluss.lake.iceberg.tiering.writer +org.apache.fluss.lake.iceberg.utils +org.apache.fluss.lake.iceberg.maintenance +org.apache.fluss.lake.iceberg.conf +org.apache.iceberg.data +org.apache.fluss.lake.lance +org.apache.fluss.lake.lance.writers +org.apache.fluss.lake.lance.tiering +org.apache.fluss.lake.lance.utils +org.apache.fluss.kafka +org.apache.fluss.tools.ci.utils.deploy +org.apache.fluss.tools.ci.utils.dependency +org.apache.fluss.tools.ci.utils.shared +org.apache.fluss.tools.ci.utils.notice +org.apache.fluss.tools.ci.utils.shade +org.apache.fluss.tools.ci.licensecheck \ No newline at end of file diff --git a/website/static/javadoc/0.8-SNAPSHOT/resources/glass.png b/website/static/javadoc/0.8-SNAPSHOT/resources/glass.png new file mode 100644 index 0000000000000000000000000000000000000000..a7f591f467a1c0c949bbc510156a0c1afb860a6e GIT binary patch literal 499 zcmVJoRsvExf%rEN>jUL}qZ_~k#FbE+Q;{`;0FZwVNX2n-^JoI; zP;4#$8DIy*Yk-P>VN(DUKmPse7mx+ExD4O|;?E5D0Z5($mjO3`*anwQU^s{ZDK#Lz zj>~{qyaIx5K!t%=G&2IJNzg!ChRpyLkO7}Ry!QaotAHAMpbB3AF(}|_f!G-oI|uK6 z`id_dumai5K%C3Y$;tKS_iqMPHg<*|-@e`liWLAggVM!zAP#@l;=c>S03;{#04Z~5 zN_+ss=Yg6*hTr59mzMwZ@+l~q!+?ft!fF66AXT#wWavHt30bZWFCK%!BNk}LN?0Hg z1VF_nfs`Lm^DjYZ1(1uD0u4CSIr)XAaqW6IT{!St5~1{i=i}zAy76p%_|w8rh@@c0Axr!ns=D-X+|*sY6!@wacG9%)Qn*O zl0sa739kT-&_?#oVxXF6tOnqTD)cZ}2vi$`ZU8RLAlo8=_z#*P3xI~i!lEh+Pdu-L zx{d*wgjtXbnGX_Yf@Tc7Q3YhLhPvc8noGJs2DA~1DySiA&6V{5JzFt ojAY1KXm~va;tU{v7C?Xj0BHw!K;2aXV*mgE07*qoM6N<$f;4TDA^-pY literal 0 HcmV?d00001 diff --git a/website/static/javadoc/0.8-SNAPSHOT/script.js b/website/static/javadoc/0.8-SNAPSHOT/script.js new file mode 100644 index 0000000000..7dc93c48e3 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/script.js @@ -0,0 +1,149 @@ +/* + * Copyright (c) 2013, 2018, Oracle and/or its affiliates. All rights reserved. + * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. + * + * This code is free software; you can redistribute it and/or modify it + * under the terms of the GNU General Public License version 2 only, as + * published by the Free Software Foundation. Oracle designates this + * particular file as subject to the "Classpath" exception as provided + * by Oracle in the LICENSE file that accompanied this code. + * + * This code is distributed in the hope that it will be useful, but WITHOUT + * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or + * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License + * version 2 for more details (a copy is included in the LICENSE file that + * accompanied this code). + * + * You should have received a copy of the GNU General Public License version + * 2 along with this work; if not, write to the Free Software Foundation, + * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. + * + * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA + * or visit www.oracle.com if you need additional information or have any + * questions. + */ + +var moduleSearchIndex; +var packageSearchIndex; +var typeSearchIndex; +var memberSearchIndex; +var tagSearchIndex; +function loadScripts(doc, tag) { + createElem(doc, tag, 'jquery/jszip/dist/jszip.js'); + createElem(doc, tag, 'jquery/jszip-utils/dist/jszip-utils.js'); + if (window.navigator.userAgent.indexOf('MSIE ') > 0 || window.navigator.userAgent.indexOf('Trident/') > 0 || + window.navigator.userAgent.indexOf('Edge/') > 0) { + createElem(doc, tag, 'jquery/jszip-utils/dist/jszip-utils-ie.js'); + } + createElem(doc, tag, 'search.js'); + + $.get(pathtoroot + "module-search-index.zip") + .done(function() { + JSZipUtils.getBinaryContent(pathtoroot + "module-search-index.zip", function(e, data) { + JSZip.loadAsync(data).then(function(zip){ + zip.file("module-search-index.json").async("text").then(function(content){ + moduleSearchIndex = JSON.parse(content); + }); + }); + }); + }); + $.get(pathtoroot + "package-search-index.zip") + .done(function() { + JSZipUtils.getBinaryContent(pathtoroot + "package-search-index.zip", function(e, data) { + JSZip.loadAsync(data).then(function(zip){ + zip.file("package-search-index.json").async("text").then(function(content){ + packageSearchIndex = JSON.parse(content); + }); + }); + }); + }); + $.get(pathtoroot + "type-search-index.zip") + .done(function() { + JSZipUtils.getBinaryContent(pathtoroot + "type-search-index.zip", function(e, data) { + JSZip.loadAsync(data).then(function(zip){ + zip.file("type-search-index.json").async("text").then(function(content){ + typeSearchIndex = JSON.parse(content); + }); + }); + }); + }); + $.get(pathtoroot + "member-search-index.zip") + .done(function() { + JSZipUtils.getBinaryContent(pathtoroot + "member-search-index.zip", function(e, data) { + JSZip.loadAsync(data).then(function(zip){ + zip.file("member-search-index.json").async("text").then(function(content){ + memberSearchIndex = JSON.parse(content); + }); + }); + }); + }); + $.get(pathtoroot + "tag-search-index.zip") + .done(function() { + JSZipUtils.getBinaryContent(pathtoroot + "tag-search-index.zip", function(e, data) { + JSZip.loadAsync(data).then(function(zip){ + zip.file("tag-search-index.json").async("text").then(function(content){ + tagSearchIndex = JSON.parse(content); + }); + }); + }); + }); + if (!moduleSearchIndex) { + createElem(doc, tag, 'module-search-index.js'); + } + if (!packageSearchIndex) { + createElem(doc, tag, 'package-search-index.js'); + } + if (!typeSearchIndex) { + createElem(doc, tag, 'type-search-index.js'); + } + if (!memberSearchIndex) { + createElem(doc, tag, 'member-search-index.js'); + } + if (!tagSearchIndex) { + createElem(doc, tag, 'tag-search-index.js'); + } + $(window).resize(function() { + $('.navPadding').css('padding-top', $('.fixedNav').css("height")); + }); +} + +function createElem(doc, tag, path) { + var script = doc.createElement(tag); + var scriptElement = doc.getElementsByTagName(tag)[0]; + script.src = pathtoroot + path; + scriptElement.parentNode.insertBefore(script, scriptElement); +} + +function show(type) { + count = 0; + for (var key in data) { + var row = document.getElementById(key); + if ((data[key] & type) !== 0) { + row.style.display = ''; + row.className = (count++ % 2) ? rowColor : altColor; + } + else + row.style.display = 'none'; + } + updateTabs(type); +} + +function updateTabs(type) { + for (var value in tabs) { + var sNode = document.getElementById(tabs[value][0]); + var spanNode = sNode.firstChild; + if (value == type) { + sNode.className = activeTableTab; + spanNode.innerHTML = tabs[value][1]; + } + else { + sNode.className = tableTab; + spanNode.innerHTML = "" + tabs[value][1] + ""; + } + } +} + +function updateModuleFrame(pFrame, cFrame) { + top.packageFrame.location = pFrame; + top.classFrame.location = cFrame; +} diff --git a/website/static/javadoc/0.8-SNAPSHOT/search.js b/website/static/javadoc/0.8-SNAPSHOT/search.js new file mode 100644 index 0000000000..8492271e71 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/search.js @@ -0,0 +1,326 @@ +/* + * Copyright (c) 2015, 2018, Oracle and/or its affiliates. All rights reserved. + * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. + * + * This code is free software; you can redistribute it and/or modify it + * under the terms of the GNU General Public License version 2 only, as + * published by the Free Software Foundation. Oracle designates this + * particular file as subject to the "Classpath" exception as provided + * by Oracle in the LICENSE file that accompanied this code. + * + * This code is distributed in the hope that it will be useful, but WITHOUT + * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or + * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License + * version 2 for more details (a copy is included in the LICENSE file that + * accompanied this code). + * + * You should have received a copy of the GNU General Public License version + * 2 along with this work; if not, write to the Free Software Foundation, + * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. + * + * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA + * or visit www.oracle.com if you need additional information or have any + * questions. + */ + +var noResult = {l: "No results found"}; +var catModules = "Modules"; +var catPackages = "Packages"; +var catTypes = "Types"; +var catMembers = "Members"; +var catSearchTags = "SearchTags"; +var highlight = "$&"; +var camelCaseRegexp = ""; +var secondaryMatcher = ""; +function getHighlightedText(item) { + var ccMatcher = new RegExp(camelCaseRegexp); + var label = item.replace(ccMatcher, highlight); + if (label === item) { + label = item.replace(secondaryMatcher, highlight); + } + return label; +} +function getURLPrefix(ui) { + var urlPrefix=""; + if (useModuleDirectories) { + var slash = "/"; + if (ui.item.category === catModules) { + return ui.item.l + slash; + } else if (ui.item.category === catPackages && ui.item.m) { + return ui.item.m + slash; + } else if ((ui.item.category === catTypes && ui.item.p) || ui.item.category === catMembers) { + $.each(packageSearchIndex, function(index, item) { + if (item.m && ui.item.p == item.l) { + urlPrefix = item.m + slash; + } + }); + return urlPrefix; + } else { + return urlPrefix; + } + } + return urlPrefix; +} +var watermark = 'Search'; +$(function() { + $("#search").val(''); + $("#search").prop("disabled", false); + $("#reset").prop("disabled", false); + $("#search").val(watermark).addClass('watermark'); + $("#search").blur(function() { + if ($(this).val().length == 0) { + $(this).val(watermark).addClass('watermark'); + } + }); + $("#search").on('click keydown', function() { + if ($(this).val() == watermark) { + $(this).val('').removeClass('watermark'); + } + }); + $("#reset").click(function() { + $("#search").val(''); + $("#search").focus(); + }); + $("#search").focus(); + $("#search")[0].setSelectionRange(0, 0); +}); +$.widget("custom.catcomplete", $.ui.autocomplete, { + _create: function() { + this._super(); + this.widget().menu("option", "items", "> :not(.ui-autocomplete-category)"); + }, + _renderMenu: function(ul, items) { + var rMenu = this, + currentCategory = ""; + rMenu.menu.bindings = $(); + $.each(items, function(index, item) { + var li; + if (item.l !== noResult.l && item.category !== currentCategory) { + ul.append("
  • " + item.category + "
  • "); + currentCategory = item.category; + } + li = rMenu._renderItemData(ul, item); + if (item.category) { + li.attr("aria-label", item.category + " : " + item.l); + li.attr("class", "resultItem"); + } else { + li.attr("aria-label", item.l); + li.attr("class", "resultItem"); + } + }); + }, + _renderItem: function(ul, item) { + var label = ""; + if (item.category === catModules) { + label = getHighlightedText(item.l); + } else if (item.category === catPackages) { + label = (item.m) + ? getHighlightedText(item.m + "/" + item.l) + : getHighlightedText(item.l); + } else if (item.category === catTypes) { + label = (item.p) + ? getHighlightedText(item.p + "." + item.l) + : getHighlightedText(item.l); + } else if (item.category === catMembers) { + label = getHighlightedText(item.p + "." + (item.c + "." + item.l)); + } else if (item.category === catSearchTags) { + label = getHighlightedText(item.l); + } else { + label = item.l; + } + var li = $("
  • ").appendTo(ul); + var div = $("
    ").appendTo(li); + if (item.category === catSearchTags) { + if (item.d) { + div.html(label + " (" + item.h + ")
    " + + item.d + "
    "); + } else { + div.html(label + " (" + item.h + ")"); + } + } else { + div.html(label); + } + return li; + } +}); +$(function() { + $("#search").catcomplete({ + minLength: 1, + delay: 100, + source: function(request, response) { + var result = new Array(); + var presult = new Array(); + var tresult = new Array(); + var mresult = new Array(); + var tgresult = new Array(); + var secondaryresult = new Array(); + var displayCount = 0; + var exactMatcher = new RegExp("^" + $.ui.autocomplete.escapeRegex(request.term) + "$", "i"); + camelCaseRegexp = ($.ui.autocomplete.escapeRegex(request.term)).split(/(?=[A-Z])/).join("([a-z0-9_$]*?)"); + var camelCaseMatcher = new RegExp("^" + camelCaseRegexp); + secondaryMatcher = new RegExp($.ui.autocomplete.escapeRegex(request.term), "i"); + + // Return the nested innermost name from the specified object + function nestedName(e) { + return e.l.substring(e.l.lastIndexOf(".") + 1); + } + + function concatResults(a1, a2) { + a1 = a1.concat(a2); + a2.length = 0; + return a1; + } + + if (moduleSearchIndex) { + var mdleCount = 0; + $.each(moduleSearchIndex, function(index, item) { + item.category = catModules; + if (exactMatcher.test(item.l)) { + result.push(item); + mdleCount++; + } else if (camelCaseMatcher.test(item.l)) { + result.push(item); + } else if (secondaryMatcher.test(item.l)) { + secondaryresult.push(item); + } + }); + displayCount = mdleCount; + result = concatResults(result, secondaryresult); + } + if (packageSearchIndex) { + var pCount = 0; + var pkg = ""; + $.each(packageSearchIndex, function(index, item) { + item.category = catPackages; + pkg = (item.m) + ? (item.m + "/" + item.l) + : item.l; + if (exactMatcher.test(item.l)) { + presult.push(item); + pCount++; + } else if (camelCaseMatcher.test(pkg)) { + presult.push(item); + } else if (secondaryMatcher.test(pkg)) { + secondaryresult.push(item); + } + }); + result = result.concat(concatResults(presult, secondaryresult)); + displayCount = (pCount > displayCount) ? pCount : displayCount; + } + if (typeSearchIndex) { + var tCount = 0; + $.each(typeSearchIndex, function(index, item) { + item.category = catTypes; + var s = nestedName(item); + if (exactMatcher.test(s)) { + tresult.push(item); + tCount++; + } else if (camelCaseMatcher.test(s)) { + tresult.push(item); + } else if (secondaryMatcher.test(item.p + "." + item.l)) { + secondaryresult.push(item); + } + }); + result = result.concat(concatResults(tresult, secondaryresult)); + displayCount = (tCount > displayCount) ? tCount : displayCount; + } + if (memberSearchIndex) { + var mCount = 0; + $.each(memberSearchIndex, function(index, item) { + item.category = catMembers; + var s = nestedName(item); + if (exactMatcher.test(s)) { + mresult.push(item); + mCount++; + } else if (camelCaseMatcher.test(s)) { + mresult.push(item); + } else if (secondaryMatcher.test(item.c + "." + item.l)) { + secondaryresult.push(item); + } + }); + result = result.concat(concatResults(mresult, secondaryresult)); + displayCount = (mCount > displayCount) ? mCount : displayCount; + } + if (tagSearchIndex) { + var tgCount = 0; + $.each(tagSearchIndex, function(index, item) { + item.category = catSearchTags; + if (exactMatcher.test(item.l)) { + tgresult.push(item); + tgCount++; + } else if (secondaryMatcher.test(item.l)) { + secondaryresult.push(item); + } + }); + result = result.concat(concatResults(tgresult, secondaryresult)); + displayCount = (tgCount > displayCount) ? tgCount : displayCount; + } + displayCount = (displayCount > 500) ? displayCount : 500; + var counter = function() { + var count = {Modules: 0, Packages: 0, Types: 0, Members: 0, SearchTags: 0}; + var f = function(item) { + count[item.category] += 1; + return (count[item.category] <= displayCount); + }; + return f; + }(); + response(result.filter(counter)); + }, + response: function(event, ui) { + if (!ui.content.length) { + ui.content.push(noResult); + } else { + $("#search").empty(); + } + }, + autoFocus: true, + position: { + collision: "flip" + }, + select: function(event, ui) { + if (ui.item.l !== noResult.l) { + var url = getURLPrefix(ui); + if (ui.item.category === catModules) { + if (useModuleDirectories) { + url += "module-summary.html"; + } else { + url = ui.item.l + "-summary.html"; + } + } else if (ui.item.category === catPackages) { + if (ui.item.url) { + url = ui.item.url; + } else { + url += ui.item.l.replace(/\./g, '/') + "/package-summary.html"; + } + } else if (ui.item.category === catTypes) { + if (ui.item.url) { + url = ui.item.url; + } else if (ui.item.p === "") { + url += ui.item.l + ".html"; + } else { + url += ui.item.p.replace(/\./g, '/') + "/" + ui.item.l + ".html"; + } + } else if (ui.item.category === catMembers) { + if (ui.item.p === "") { + url += ui.item.c + ".html" + "#"; + } else { + url += ui.item.p.replace(/\./g, '/') + "/" + ui.item.c + ".html" + "#"; + } + if (ui.item.url) { + url += ui.item.url; + } else { + url += ui.item.l; + } + } else if (ui.item.category === catSearchTags) { + url += ui.item.u; + } + if (top !== window) { + parent.classFrame.location = pathtoroot + url; + } else { + window.location.href = pathtoroot + url; + } + $("#search").focus(); + } + } + }); +}); diff --git a/website/static/javadoc/0.8-SNAPSHOT/serialized-form.html b/website/static/javadoc/0.8-SNAPSHOT/serialized-form.html new file mode 100644 index 0000000000..691102e404 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/serialized-form.html @@ -0,0 +1,899 @@ + + + + + +Serialized Form (fluss-javadoc-temp 0.8-SNAPSHOT API) + + + + + + + + + + + + + + + +
    + +
    +
    +
    +

    Serialized Form

    +
    +
    + +
    +
    +
    + +

    Copyright © 2025. All rights reserved.

    +
    + + diff --git a/website/static/javadoc/0.8-SNAPSHOT/stylesheet.css b/website/static/javadoc/0.8-SNAPSHOT/stylesheet.css new file mode 100644 index 0000000000..de945eda26 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/stylesheet.css @@ -0,0 +1,910 @@ +/* + * Javadoc style sheet + */ + +@import url('resources/fonts/dejavu.css'); + +/* + * Styles for individual HTML elements. + * + * These are styles that are specific to individual HTML elements. Changing them affects the style of a particular + * HTML element throughout the page. + */ + +body { + background-color:#ffffff; + color:#353833; + font-family:'DejaVu Sans', Arial, Helvetica, sans-serif; + font-size:14px; + margin:0; + padding:0; + height:100%; + width:100%; +} +iframe { + margin:0; + padding:0; + height:100%; + width:100%; + overflow-y:scroll; + border:none; +} +a:link, a:visited { + text-decoration:none; + color:#4A6782; +} +a[href]:hover, a[href]:focus { + text-decoration:none; + color:#bb7a2a; +} +a[name] { + color:#353833; +} +a[name]:before, a[name]:target, a[id]:before, a[id]:target { + content:""; + display:inline-block; + position:relative; + padding-top:129px; + margin-top:-129px; +} +pre { + font-family:'DejaVu Sans Mono', monospace; + font-size:14px; +} +h1 { + font-size:20px; +} +h2 { + font-size:18px; +} +h3 { + font-size:16px; + font-style:italic; +} +h4 { + font-size:13px; +} +h5 { + font-size:12px; +} +h6 { + font-size:11px; +} +ul { + list-style-type:disc; +} +code, tt { + font-family:'DejaVu Sans Mono', monospace; + font-size:14px; + padding-top:4px; + margin-top:8px; + line-height:1.4em; +} +dt code { + font-family:'DejaVu Sans Mono', monospace; + font-size:14px; + padding-top:4px; +} +table tr td dt code { + font-family:'DejaVu Sans Mono', monospace; + font-size:14px; + vertical-align:top; + padding-top:4px; +} +sup { + font-size:8px; +} + +/* + * Styles for HTML generated by javadoc. + * + * These are style classes that are used by the standard doclet to generate HTML documentation. + */ + +/* + * Styles for document title and copyright. + */ +.clear { + clear:both; + height:0px; + overflow:hidden; +} +.aboutLanguage { + float:right; + padding:0px 21px; + font-size:11px; + z-index:200; + margin-top:-9px; +} +.legalCopy { + margin-left:.5em; +} +.bar a, .bar a:link, .bar a:visited, .bar a:active { + color:#FFFFFF; + text-decoration:none; +} +.bar a:hover, .bar a:focus { + color:#bb7a2a; +} +.tab { + background-color:#0066FF; + color:#ffffff; + padding:8px; + width:5em; + font-weight:bold; +} +/* + * Styles for navigation bar. + */ +.bar { + background-color:#4D7A97; + color:#FFFFFF; + padding:.8em .5em .4em .8em; + height:auto;/*height:1.8em;*/ + font-size:11px; + margin:0; +} +.navPadding { + padding-top: 107px; +} +.fixedNav { + position:fixed; + width:100%; + z-index:999; + background-color:#ffffff; +} +.topNav { + background-color:#4D7A97; + color:#FFFFFF; + float:left; + padding:0; + width:100%; + clear:right; + height:2.8em; + padding-top:10px; + overflow:hidden; + font-size:12px; +} +.bottomNav { + margin-top:10px; + background-color:#4D7A97; + color:#FFFFFF; + float:left; + padding:0; + width:100%; + clear:right; + height:2.8em; + padding-top:10px; + overflow:hidden; + font-size:12px; +} +.subNav { + background-color:#dee3e9; + float:left; + width:100%; + overflow:hidden; + font-size:12px; +} +.subNav div { + clear:left; + float:left; + padding:0 0 5px 6px; + text-transform:uppercase; +} +ul.navList, ul.subNavList { + float:left; + margin:0 25px 0 0; + padding:0; +} +ul.navList li{ + list-style:none; + float:left; + padding: 5px 6px; + text-transform:uppercase; +} +ul.navListSearch { + float:right; + margin:0 0 0 0; + padding:0; +} +ul.navListSearch li { + list-style:none; + float:right; + padding: 5px 6px; + text-transform:uppercase; +} +ul.navListSearch li label { + position:relative; + right:-16px; +} +ul.subNavList li { + list-style:none; + float:left; +} +.topNav a:link, .topNav a:active, .topNav a:visited, .bottomNav a:link, .bottomNav a:active, .bottomNav a:visited { + color:#FFFFFF; + text-decoration:none; + text-transform:uppercase; +} +.topNav a:hover, .bottomNav a:hover { + text-decoration:none; + color:#bb7a2a; + text-transform:uppercase; +} +.navBarCell1Rev { + background-color:#F8981D; + color:#253441; + margin: auto 5px; +} +.skipNav { + position:absolute; + top:auto; + left:-9999px; + overflow:hidden; +} +/* + * Styles for page header and footer. + */ +.header, .footer { + clear:both; + margin:0 20px; + padding:5px 0 0 0; +} +.indexNav { + position:relative; + font-size:12px; + background-color:#dee3e9; +} +.indexNav ul { + margin-top:0; + padding:5px; +} +.indexNav ul li { + display:inline; + list-style-type:none; + padding-right:10px; + text-transform:uppercase; +} +.indexNav h1 { + font-size:13px; +} +.title { + color:#2c4557; + margin:10px 0; +} +.subTitle { + margin:5px 0 0 0; +} +.header ul { + margin:0 0 15px 0; + padding:0; +} +.footer ul { + margin:20px 0 5px 0; +} +.header ul li, .footer ul li { + list-style:none; + font-size:13px; +} +/* + * Styles for headings. + */ +div.details ul.blockList ul.blockList ul.blockList li.blockList h4, div.details ul.blockList ul.blockList ul.blockListLast li.blockList h4 { + background-color:#dee3e9; + border:1px solid #d0d9e0; + margin:0 0 6px -8px; + padding:7px 5px; +} +ul.blockList ul.blockList ul.blockList li.blockList h3 { + background-color:#dee3e9; + border:1px solid #d0d9e0; + margin:0 0 6px -8px; + padding:7px 5px; +} +ul.blockList ul.blockList li.blockList h3 { + padding:0; + margin:15px 0; +} +ul.blockList li.blockList h2 { + padding:0px 0 20px 0; +} +/* + * Styles for page layout containers. + */ +.contentContainer, .sourceContainer, .classUseContainer, .serializedFormContainer, .constantValuesContainer, +.allClassesContainer, .allPackagesContainer { + clear:both; + padding:10px 20px; + position:relative; +} +.indexContainer { + margin:10px; + position:relative; + font-size:12px; +} +.indexContainer h2 { + font-size:13px; + padding:0 0 3px 0; +} +.indexContainer ul { + margin:0; + padding:0; +} +.indexContainer ul li { + list-style:none; + padding-top:2px; +} +.contentContainer .description dl dt, .contentContainer .details dl dt, .serializedFormContainer dl dt { + font-size:12px; + font-weight:bold; + margin:10px 0 0 0; + color:#4E4E4E; +} +.contentContainer .description dl dd, .contentContainer .details dl dd, .serializedFormContainer dl dd { + margin:5px 0 10px 0px; + font-size:14px; + font-family:'DejaVu Serif', Georgia, "Times New Roman", Times, serif; +} +.serializedFormContainer dl.nameValue dt { + margin-left:1px; + font-size:1.1em; + display:inline; + font-weight:bold; +} +.serializedFormContainer dl.nameValue dd { + margin:0 0 0 1px; + font-size:1.1em; + display:inline; +} +/* + * Styles for lists. + */ +li.circle { + list-style:circle; +} +ul.horizontal li { + display:inline; + font-size:0.9em; +} +ul.inheritance { + margin:0; + padding:0; +} +ul.inheritance li { + display:inline; + list-style:none; +} +ul.inheritance li ul.inheritance { + margin-left:15px; + padding-left:15px; + padding-top:1px; +} +ul.blockList, ul.blockListLast { + margin:10px 0 10px 0; + padding:0; +} +ul.blockList li.blockList, ul.blockListLast li.blockList { + list-style:none; + margin-bottom:15px; + line-height:1.4; +} +ul.blockList ul.blockList li.blockList, ul.blockList ul.blockListLast li.blockList { + padding:0px 20px 5px 10px; + border:1px solid #ededed; + background-color:#f8f8f8; +} +ul.blockList ul.blockList ul.blockList li.blockList, ul.blockList ul.blockList ul.blockListLast li.blockList { + padding:0 0 5px 8px; + background-color:#ffffff; + border:none; +} +ul.blockList ul.blockList ul.blockList ul.blockList li.blockList { + margin-left:0; + padding-left:0; + padding-bottom:15px; + border:none; +} +ul.blockList ul.blockList ul.blockList ul.blockList li.blockListLast { + list-style:none; + border-bottom:none; + padding-bottom:0; +} +table tr td dl, table tr td dl dt, table tr td dl dd { + margin-top:0; + margin-bottom:1px; +} +/* + * Styles for tables. + */ +.overviewSummary, .memberSummary, .typeSummary, .useSummary, .constantsSummary, .deprecatedSummary, +.requiresSummary, .packagesSummary, .providesSummary, .usesSummary { + width:100%; + border-spacing:0; + border-left:1px solid #EEE; + border-right:1px solid #EEE; + border-bottom:1px solid #EEE; +} +.overviewSummary, .memberSummary, .requiresSummary, .packagesSummary, .providesSummary, .usesSummary { + padding:0px; +} +.overviewSummary caption, .memberSummary caption, .typeSummary caption, +.useSummary caption, .constantsSummary caption, .deprecatedSummary caption, +.requiresSummary caption, .packagesSummary caption, .providesSummary caption, .usesSummary caption { + position:relative; + text-align:left; + background-repeat:no-repeat; + color:#253441; + font-weight:bold; + clear:none; + overflow:hidden; + padding:0px; + padding-top:10px; + padding-left:1px; + margin:0px; + white-space:pre; +} +.constantsSummary caption a:link, .constantsSummary caption a:visited, +.useSummary caption a:link, .useSummary caption a:visited { + color:#1f389c; +} +.overviewSummary caption a:link, .memberSummary caption a:link, .typeSummary caption a:link, +.deprecatedSummary caption a:link, +.requiresSummary caption a:link, .packagesSummary caption a:link, .providesSummary caption a:link, +.usesSummary caption a:link, +.overviewSummary caption a:hover, .memberSummary caption a:hover, .typeSummary caption a:hover, +.useSummary caption a:hover, .constantsSummary caption a:hover, .deprecatedSummary caption a:hover, +.requiresSummary caption a:hover, .packagesSummary caption a:hover, .providesSummary caption a:hover, +.usesSummary caption a:hover, +.overviewSummary caption a:active, .memberSummary caption a:active, .typeSummary caption a:active, +.useSummary caption a:active, .constantsSummary caption a:active, .deprecatedSummary caption a:active, +.requiresSummary caption a:active, .packagesSummary caption a:active, .providesSummary caption a:active, +.usesSummary caption a:active, +.overviewSummary caption a:visited, .memberSummary caption a:visited, .typeSummary caption a:visited, +.deprecatedSummary caption a:visited, +.requiresSummary caption a:visited, .packagesSummary caption a:visited, .providesSummary caption a:visited, +.usesSummary caption a:visited { + color:#FFFFFF; +} +.overviewSummary caption span, .memberSummary caption span, .typeSummary caption span, +.useSummary caption span, .constantsSummary caption span, .deprecatedSummary caption span, +.requiresSummary caption span, .packagesSummary caption span, .providesSummary caption span, +.usesSummary caption span { + white-space:nowrap; + padding-top:5px; + padding-left:12px; + padding-right:12px; + padding-bottom:7px; + display:inline-block; + float:left; + background-color:#F8981D; + border: none; + height:16px; +} +.memberSummary caption span.activeTableTab span, .packagesSummary caption span.activeTableTab span, +.overviewSummary caption span.activeTableTab span, .typeSummary caption span.activeTableTab span { + white-space:nowrap; + padding-top:5px; + padding-left:12px; + padding-right:12px; + margin-right:3px; + display:inline-block; + float:left; + background-color:#F8981D; + height:16px; +} +.memberSummary caption span.tableTab span, .packagesSummary caption span.tableTab span, +.overviewSummary caption span.tableTab span, .typeSummary caption span.tableTab span { + white-space:nowrap; + padding-top:5px; + padding-left:12px; + padding-right:12px; + margin-right:3px; + display:inline-block; + float:left; + background-color:#4D7A97; + height:16px; +} +.memberSummary caption span.tableTab, .memberSummary caption span.activeTableTab, +.packagesSummary caption span.tableTab, .packagesSummary caption span.activeTableTab, +.overviewSummary caption span.tableTab, .overviewSummary caption span.activeTableTab, +.typeSummary caption span.tableTab, .typeSummary caption span.activeTableTab { + padding-top:0px; + padding-left:0px; + padding-right:0px; + background-image:none; + float:none; + display:inline; +} +.overviewSummary .tabEnd, .memberSummary .tabEnd, .typeSummary .tabEnd, +.useSummary .tabEnd, .constantsSummary .tabEnd, .deprecatedSummary .tabEnd, +.requiresSummary .tabEnd, .packagesSummary .tabEnd, .providesSummary .tabEnd, .usesSummary .tabEnd { + display:none; + width:5px; + position:relative; + float:left; + background-color:#F8981D; +} +.memberSummary .activeTableTab .tabEnd, .packagesSummary .activeTableTab .tabEnd, +.overviewSummary .activeTableTab .tabEnd, .typeSummary .activeTableTab .tabEnd { + display:none; + width:5px; + margin-right:3px; + position:relative; + float:left; + background-color:#F8981D; +} +.memberSummary .tableTab .tabEnd, .packagesSummary .tableTab .tabEnd, +.overviewSummary .tableTab .tabEnd, .typeSummary .tableTab .tabEnd { + display:none; + width:5px; + margin-right:3px; + position:relative; + background-color:#4D7A97; + float:left; +} +.rowColor th, .altColor th { + font-weight:normal; +} +.overviewSummary td, .memberSummary td, .typeSummary td, +.useSummary td, .constantsSummary td, .deprecatedSummary td, +.requiresSummary td, .packagesSummary td, .providesSummary td, .usesSummary td { + text-align:left; + padding:0px 0px 12px 10px; +} +th.colFirst, th.colSecond, th.colLast, th.colConstructorName, th.colDeprecatedItemName, .useSummary th, +.constantsSummary th, .packagesSummary th, td.colFirst, td.colSecond, td.colLast, .useSummary td, +.constantsSummary td { + vertical-align:top; + padding-right:0px; + padding-top:8px; + padding-bottom:3px; +} +th.colFirst, th.colSecond, th.colLast, th.colConstructorName, th.colDeprecatedItemName, .constantsSummary th, +.packagesSummary th { + background:#dee3e9; + text-align:left; + padding:8px 3px 3px 7px; +} +td.colFirst, th.colFirst { + font-size:13px; +} +td.colSecond, th.colSecond, td.colLast, th.colConstructorName, th.colDeprecatedItemName, th.colLast { + font-size:13px; +} +.constantsSummary th, .packagesSummary th { + font-size:13px; +} +.providesSummary th.colFirst, .providesSummary th.colLast, .providesSummary td.colFirst, +.providesSummary td.colLast { + white-space:normal; + font-size:13px; +} +.overviewSummary td.colFirst, .overviewSummary th.colFirst, +.requiresSummary td.colFirst, .requiresSummary th.colFirst, +.packagesSummary td.colFirst, .packagesSummary td.colSecond, .packagesSummary th.colFirst, .packagesSummary th, +.usesSummary td.colFirst, .usesSummary th.colFirst, +.providesSummary td.colFirst, .providesSummary th.colFirst, +.memberSummary td.colFirst, .memberSummary th.colFirst, +.memberSummary td.colSecond, .memberSummary th.colSecond, .memberSummary th.colConstructorName, +.typeSummary td.colFirst, .typeSummary th.colFirst { + vertical-align:top; +} +.packagesSummary th.colLast, .packagesSummary td.colLast { + white-space:normal; +} +td.colFirst a:link, td.colFirst a:visited, +td.colSecond a:link, td.colSecond a:visited, +th.colFirst a:link, th.colFirst a:visited, +th.colSecond a:link, th.colSecond a:visited, +th.colConstructorName a:link, th.colConstructorName a:visited, +th.colDeprecatedItemName a:link, th.colDeprecatedItemName a:visited, +.constantValuesContainer td a:link, .constantValuesContainer td a:visited, +.allClassesContainer td a:link, .allClassesContainer td a:visited, +.allPackagesContainer td a:link, .allPackagesContainer td a:visited { + font-weight:bold; +} +.tableSubHeadingColor { + background-color:#EEEEFF; +} +.altColor, .altColor th { + background-color:#FFFFFF; +} +.rowColor, .rowColor th { + background-color:#EEEEEF; +} +/* + * Styles for contents. + */ +.description pre { + margin-top:0; +} +.deprecatedContent { + margin:0; + padding:10px 0; +} +.docSummary { + padding:0; +} +ul.blockList ul.blockList ul.blockList li.blockList h3 { + font-style:normal; +} +div.block { + font-size:14px; + font-family:'DejaVu Serif', Georgia, "Times New Roman", Times, serif; +} +td.colLast div { + padding-top:0px; +} +td.colLast a { + padding-bottom:3px; +} +/* + * Styles for formatting effect. + */ +.sourceLineNo { + color:green; + padding:0 30px 0 0; +} +h1.hidden { + visibility:hidden; + overflow:hidden; + font-size:10px; +} +.block { + display:block; + margin:3px 10px 2px 0px; + color:#474747; +} +.deprecatedLabel, .descfrmTypeLabel, .implementationLabel, .memberNameLabel, .memberNameLink, +.moduleLabelInPackage, .moduleLabelInType, .overrideSpecifyLabel, .packageLabelInType, +.packageHierarchyLabel, .paramLabel, .returnLabel, .seeLabel, .simpleTagLabel, +.throwsLabel, .typeNameLabel, .typeNameLink, .searchTagLink { + font-weight:bold; +} +.deprecationComment, .emphasizedPhrase, .interfaceName { + font-style:italic; +} +.deprecationBlock { + font-size:14px; + font-family:'DejaVu Serif', Georgia, "Times New Roman", Times, serif; + border-style:solid; + border-width:thin; + border-radius:10px; + padding:10px; + margin-bottom:10px; + margin-right:10px; + display:inline-block; +} +div.block div.deprecationComment, div.block div.block span.emphasizedPhrase, +div.block div.block span.interfaceName { + font-style:normal; +} +div.contentContainer ul.blockList li.blockList h2 { + padding-bottom:0px; +} +/* + * Styles for IFRAME. + */ +.mainContainer { + margin:0 auto; + padding:0; + height:100%; + width:100%; + position:fixed; + top:0; + left:0; +} +.leftContainer { + height:100%; + position:fixed; + width:320px; +} +.leftTop { + position:relative; + float:left; + width:315px; + top:0; + left:0; + height:30%; + border-right:6px solid #ccc; + border-bottom:6px solid #ccc; +} +.leftBottom { + position:relative; + float:left; + width:315px; + bottom:0; + left:0; + height:70%; + border-right:6px solid #ccc; + border-top:1px solid #000; +} +.rightContainer { + position:absolute; + left:320px; + top:0; + bottom:0; + height:100%; + right:0; + border-left:1px solid #000; +} +.rightIframe { + margin:0; + padding:0; + height:100%; + right:30px; + width:100%; + overflow:visible; + margin-bottom:30px; +} +/* + * Styles specific to HTML5 elements. + */ +main, nav, header, footer, section { + display:block; +} +/* + * Styles for javadoc search. + */ +.ui-autocomplete-category { + font-weight:bold; + font-size:15px; + padding:7px 0 7px 3px; + background-color:#4D7A97; + color:#FFFFFF; +} +.resultItem { + font-size:13px; +} +.ui-autocomplete { + max-height:85%; + max-width:65%; + overflow-y:scroll; + overflow-x:scroll; + white-space:nowrap; + box-shadow: 0 3px 6px rgba(0,0,0,0.16), 0 3px 6px rgba(0,0,0,0.23); +} +ul.ui-autocomplete { + position:fixed; + z-index:999999; + background-color: #FFFFFF; +} +ul.ui-autocomplete li { + float:left; + clear:both; + width:100%; +} +.resultHighlight { + font-weight:bold; +} +.ui-autocomplete .result-item { + font-size: inherit; +} +#search { + background-image:url('resources/glass.png'); + background-size:13px; + background-repeat:no-repeat; + background-position:2px 3px; + padding-left:20px; + position:relative; + right:-18px; +} +#reset { + background-color: rgb(255,255,255); + background-image:url('resources/x.png'); + background-position:center; + background-repeat:no-repeat; + background-size:12px; + border:0 none; + width:16px; + height:17px; + position:relative; + left:-4px; + top:-4px; + font-size:0px; +} +.watermark { + color:#545454; +} +.searchTagDescResult { + font-style:italic; + font-size:11px; +} +.searchTagHolderResult { + font-style:italic; + font-size:12px; +} +.searchTagResult:before, .searchTagResult:target { + color:red; +} +.moduleGraph span { + display:none; + position:absolute; +} +.moduleGraph:hover span { + display:block; + margin: -100px 0 0 100px; + z-index: 1; +} +.methodSignature { + white-space:normal; +} + +/* + * Styles for user-provided tables. + * + * borderless: + * No borders, vertical margins, styled caption. + * This style is provided for use with existing doc comments. + * In general, borderless tables should not be used for layout purposes. + * + * plain: + * Plain borders around table and cells, vertical margins, styled caption. + * Best for small tables or for complex tables for tables with cells that span + * rows and columns, when the "striped" style does not work well. + * + * striped: + * Borders around the table and vertical borders between cells, striped rows, + * vertical margins, styled caption. + * Best for tables that have a header row, and a body containing a series of simple rows. + */ + +table.borderless, +table.plain, +table.striped { + margin-top: 10px; + margin-bottom: 10px; +} +table.borderless > caption, +table.plain > caption, +table.striped > caption { + font-weight: bold; + font-size: smaller; +} +table.borderless th, table.borderless td, +table.plain th, table.plain td, +table.striped th, table.striped td { + padding: 2px 5px; +} +table.borderless, +table.borderless > thead > tr > th, table.borderless > tbody > tr > th, table.borderless > tr > th, +table.borderless > thead > tr > td, table.borderless > tbody > tr > td, table.borderless > tr > td { + border: none; +} +table.borderless > thead > tr, table.borderless > tbody > tr, table.borderless > tr { + background-color: transparent; +} +table.plain { + border-collapse: collapse; + border: 1px solid black; +} +table.plain > thead > tr, table.plain > tbody tr, table.plain > tr { + background-color: transparent; +} +table.plain > thead > tr > th, table.plain > tbody > tr > th, table.plain > tr > th, +table.plain > thead > tr > td, table.plain > tbody > tr > td, table.plain > tr > td { + border: 1px solid black; +} +table.striped { + border-collapse: collapse; + border: 1px solid black; +} +table.striped > thead { + background-color: #E3E3E3; +} +table.striped > thead > tr > th, table.striped > thead > tr > td { + border: 1px solid black; +} +table.striped > tbody > tr:nth-child(even) { + background-color: #EEE +} +table.striped > tbody > tr:nth-child(odd) { + background-color: #FFF +} +table.striped > tbody > tr > th, table.striped > tbody > tr > td { + border-left: 1px solid black; + border-right: 1px solid black; +} +table.striped > tbody > tr > th { + font-weight: normal; +} diff --git a/website/static/javadoc/0.8-SNAPSHOT/type-search-index.js b/website/static/javadoc/0.8-SNAPSHOT/type-search-index.js new file mode 100644 index 0000000000..f0ec13f7c8 --- /dev/null +++ b/website/static/javadoc/0.8-SNAPSHOT/type-search-index.js @@ -0,0 +1 @@ +typeSearchIndex = [{"p":"org.apache.fluss.flink.procedure","l":"AbstractAclProcedure"},{"p":"org.apache.fluss.client.lookup","l":"AbstractLookupBatch"},{"p":"org.apache.fluss.client.lookup","l":"AbstractLookupQuery"},{"p":"org.apache.fluss.client.table.writer","l":"AbstractTableWriter"},{"p":"org.apache.fluss.flink.procedure","l":"AddAclProcedure"},{"p":"org.apache.fluss.client.admin","l":"Admin"},{"l":"All Classes","url":"allclasses-index.html"},{"p":"org.apache.fluss.client.table.writer","l":"Append"},{"p":"org.apache.fluss.client.table.writer","l":"AppendResult"},{"p":"org.apache.fluss.flink.sink.writer","l":"AppendSinkWriter"},{"p":"org.apache.fluss.client.table.writer","l":"AppendWriter"},{"p":"org.apache.fluss.client.write","l":"ArrowLogWriteBatch"},{"p":"org.apache.fluss.client.table.scanner.batch","l":"BatchScanner"},{"p":"org.apache.fluss.client.table.scanner.batch","l":"BatchScanUtils"},{"p":"org.apache.fluss.flink.source.reader","l":"BoundedSplitReader"},{"p":"org.apache.fluss.flink.source.enumerator.initializer","l":"OffsetsInitializer.BucketOffsetsRetriever"},{"p":"org.apache.fluss.flink.source.enumerator.initializer","l":"BucketOffsetsRetrieverImpl"},{"p":"org.apache.fluss.flink.tiering.source","l":"TieringSource.Builder"},{"p":"org.apache.fluss.flink.utils","l":"CatalogExceptionUtils"},{"p":"org.apache.fluss.flink.utils","l":"CatalogPropertiesUtils"},{"p":"org.apache.fluss.flink.adapter","l":"CatalogTableAdapter"},{"p":"org.apache.fluss.flink.sink","l":"ChannelComputer"},{"p":"org.apache.fluss.client.utils","l":"ClientRpcMessageUtils"},{"p":"org.apache.fluss.client.utils","l":"ClientUtils"},{"p":"org.apache.fluss.flink.tiering.committer","l":"CommittableMessage"},{"p":"org.apache.fluss.flink.tiering.committer","l":"CommittableMessageTypeInfo"},{"p":"org.apache.fluss.client","l":"Connection"},{"p":"org.apache.fluss.client","l":"ConnectionFactory"},{"p":"org.apache.fluss.client.admin","l":"CreateAclsResult"},{"p":"org.apache.fluss.flink.utils","l":"DataLakeUtils"},{"p":"org.apache.fluss.client.token","l":"DefaultSecurityTokenManager"},{"p":"org.apache.fluss.client.token","l":"DefaultSecurityTokenProvider"},{"p":"org.apache.fluss.client.table.writer","l":"DeleteResult"},{"p":"org.apache.fluss.flink.source.deserializer","l":"DeserializerInitContextImpl"},{"p":"org.apache.fluss.flink.procedure","l":"DropAclProcedure"},{"p":"org.apache.fluss.client.admin","l":"DropAclsResult"},{"p":"org.apache.fluss.client.write","l":"DynamicPartitionCreator"},{"p":"org.apache.fluss.client.write","l":"DynamicWriteBatchSizeEstimator"},{"p":"org.apache.fluss.client.admin","l":"OffsetSpec.EarliestSpec"},{"p":"org.apache.fluss.flink.tiering.event","l":"FailedTieringEvent"},{"p":"org.apache.fluss.flink.utils","l":"PushdownUtils.FieldEqual"},{"p":"org.apache.fluss.client.admin","l":"DropAclsResult.FilterResult"},{"p":"org.apache.fluss.client.admin","l":"DropAclsResult.FilterResults"},{"p":"org.apache.fluss.flink.tiering.event","l":"FinishedTieringEvent"},{"p":"org.apache.fluss.flink.row","l":"FlinkAsFlussRow"},{"p":"org.apache.fluss.flink.source.lookup","l":"FlinkAsyncLookupFunction"},{"p":"org.apache.fluss.flink.catalog","l":"FlinkCatalog"},{"p":"org.apache.fluss.flink.catalog","l":"FlinkCatalogFactory"},{"p":"org.apache.fluss.flink.catalog","l":"FlinkCatalogOptions"},{"p":"org.apache.fluss.flink","l":"FlinkConnectorOptions"},{"p":"org.apache.fluss.flink.utils","l":"FlinkConnectorOptionsUtils"},{"p":"org.apache.fluss.flink.utils","l":"FlinkConversions"},{"p":"org.apache.fluss.flink.metrics","l":"FlinkCounter"},{"p":"org.apache.fluss.flink.metrics","l":"FlinkGauge"},{"p":"org.apache.fluss.flink.metrics","l":"FlinkHistogram"},{"p":"org.apache.fluss.flink.source.lookup","l":"FlinkLookupFunction"},{"p":"org.apache.fluss.flink.metrics","l":"FlinkMeter"},{"p":"org.apache.fluss.flink.metrics","l":"FlinkMetricRegistry"},{"p":"org.apache.fluss.flink.source.emitter","l":"FlinkRecordEmitter"},{"p":"org.apache.fluss.flink.source.reader","l":"FlinkRecordsWithSplitIds"},{"p":"org.apache.fluss.flink.sink","l":"FlinkRowDataChannelComputer"},{"p":"org.apache.fluss.flink.utils","l":"FlinkRowToFlussRowConverter"},{"p":"org.apache.fluss.flink.sink.writer","l":"FlinkSinkWriter"},{"p":"org.apache.fluss.flink.source","l":"FlinkSource"},{"p":"org.apache.fluss.flink.source.enumerator","l":"FlinkSourceEnumerator"},{"p":"org.apache.fluss.flink.source.reader.fetcher","l":"FlinkSourceFetcherManager"},{"p":"org.apache.fluss.flink.source.reader","l":"FlinkSourceReader"},{"p":"org.apache.fluss.flink.source.metrics","l":"FlinkSourceReaderMetrics"},{"p":"org.apache.fluss.flink.source.reader","l":"FlinkSourceSplitReader"},{"p":"org.apache.fluss.flink.sink","l":"FlinkStreamPartitioner"},{"p":"org.apache.fluss.flink.catalog","l":"FlinkTableFactory"},{"p":"org.apache.fluss.flink.sink","l":"FlinkTableSink"},{"p":"org.apache.fluss.flink.source","l":"FlinkTableSource"},{"p":"org.apache.fluss.flink.utils","l":"FlinkUtils"},{"p":"org.apache.fluss.client.admin","l":"FlussAdmin"},{"p":"org.apache.fluss.client","l":"FlussConnection"},{"p":"org.apache.fluss.flink.utils","l":"FlussRowToFlinkRowConverter.FlussDeserializationConverter"},{"p":"org.apache.fluss.flink.source.deserializer","l":"FlussDeserializationSchema"},{"p":"org.apache.fluss.flink.utils","l":"FlussRowToFlinkRowConverter"},{"p":"org.apache.fluss.flink.utils","l":"FlussRowToJsonConverters.FlussRowToJsonConverter"},{"p":"org.apache.fluss.flink.utils","l":"FlussRowToJsonConverters"},{"p":"org.apache.fluss.flink.utils","l":"FlinkRowToFlussRowConverter.FlussSerializationConverter"},{"p":"org.apache.fluss.flink.sink.serializer","l":"FlussSerializationSchema"},{"p":"org.apache.fluss.flink.sink","l":"FlussSink"},{"p":"org.apache.fluss.flink.sink","l":"FlussSinkBuilder"},{"p":"org.apache.fluss.flink.source","l":"FlussSource"},{"p":"org.apache.fluss.flink.source","l":"FlussSourceBuilder"},{"p":"org.apache.fluss.flink.source.state","l":"FlussSourceEnumeratorStateSerializer"},{"p":"org.apache.fluss.client.table","l":"FlussTable"},{"p":"org.apache.fluss.flink.tiering.committer","l":"FlussTableLakeSnapshotCommitter"},{"p":"org.apache.fluss.client.write","l":"HashBucketAssigner"},{"p":"org.apache.fluss.flink.source.split","l":"HybridSnapshotLogSplit"},{"p":"org.apache.fluss.flink.source.split","l":"HybridSnapshotLogSplitState"},{"p":"org.apache.fluss.client.write","l":"IdempotenceBucketEntry"},{"p":"org.apache.fluss.client.write","l":"IdempotenceBucketMap"},{"p":"org.apache.fluss.client.write","l":"IdempotenceManager"},{"p":"org.apache.fluss.flink.lake.reader","l":"IndexedLakeSplitRecordIterator"},{"p":"org.apache.fluss.client.write","l":"IndexedLogWriteBatch"},{"p":"org.apache.fluss.flink.sink.serializer","l":"FlussSerializationSchema.InitializationContext"},{"p":"org.apache.fluss.flink.source.deserializer","l":"FlussDeserializationSchema.InitializationContext"},{"p":"org.apache.fluss.flink.source.deserializer","l":"JsonStringDeserializationSchema"},{"p":"org.apache.fluss.flink.lake.reader","l":"KeyValueRow"},{"p":"org.apache.fluss.client.table.scanner.batch","l":"KvSnapshotBatchScanner"},{"p":"org.apache.fluss.client.metadata","l":"KvSnapshotMetadata"},{"p":"org.apache.fluss.client.metadata","l":"KvSnapshots"},{"p":"org.apache.fluss.client.write","l":"KvWriteBatch"},{"p":"org.apache.fluss.flink.lake","l":"LakeCatalog"},{"p":"org.apache.fluss.flink.lake","l":"LakeRecordRecordEmitter"},{"p":"org.apache.fluss.client.metadata","l":"LakeSnapshot"},{"p":"org.apache.fluss.flink.lake.split","l":"LakeSnapshotAndFlussLogSplit"},{"p":"org.apache.fluss.flink.lake.state","l":"LakeSnapshotAndFlussLogSplitState"},{"p":"org.apache.fluss.flink.lake.reader","l":"LakeSnapshotAndLogSplitScanner"},{"p":"org.apache.fluss.flink.lake.reader","l":"LakeSnapshotScanner"},{"p":"org.apache.fluss.flink.lake.split","l":"LakeSnapshotSplit"},{"p":"org.apache.fluss.flink.lake.state","l":"LakeSnapshotSplitState"},{"p":"org.apache.fluss.flink.utils","l":"LakeSourceUtils"},{"p":"org.apache.fluss.flink.lake","l":"LakeSplitGenerator"},{"p":"org.apache.fluss.flink.lake","l":"LakeSplitReaderGenerator"},{"p":"org.apache.fluss.flink.lake","l":"LakeSplitSerializer"},{"p":"org.apache.fluss.flink.lake","l":"LakeSplitStateInitializer"},{"p":"org.apache.fluss.flink.lake","l":"LakeTableFactory"},{"p":"org.apache.fluss.flink.tiering","l":"LakeTieringJobBuilder"},{"p":"org.apache.fluss.flink.source.enumerator.initializer","l":"LatestOffsetsInitializer"},{"p":"org.apache.fluss.client.admin","l":"OffsetSpec.LatestSpec"},{"p":"org.apache.fluss.client.table.scanner.batch","l":"LimitBatchScanner"},{"p":"org.apache.fluss.flink.procedure","l":"ListAclProcedure"},{"p":"org.apache.fluss.client.admin","l":"ListOffsetsResult"},{"p":"org.apache.fluss.client.table.scanner.log","l":"LogFetchBuffer"},{"p":"org.apache.fluss.client.table.scanner.log","l":"LogFetchCollector"},{"p":"org.apache.fluss.client.table.scanner.log","l":"LogFetcher"},{"p":"org.apache.fluss.client.table.scanner.log","l":"LogScan"},{"p":"org.apache.fluss.client.table.scanner.log","l":"LogScanner"},{"p":"org.apache.fluss.client.table.scanner.log","l":"LogScannerImpl"},{"p":"org.apache.fluss.client.table.scanner.log","l":"LogScannerStatus"},{"p":"org.apache.fluss.flink.source.split","l":"LogSplit"},{"p":"org.apache.fluss.flink.source.split","l":"LogSplitState"},{"p":"org.apache.fluss.client.lookup","l":"Lookup"},{"p":"org.apache.fluss.client.lookup","l":"LookupBatch"},{"p":"org.apache.fluss.client.lookup","l":"LookupClient"},{"p":"org.apache.fluss.client.lookup","l":"Lookuper"},{"p":"org.apache.fluss.flink.source.lookup","l":"LookupNormalizer"},{"p":"org.apache.fluss.client.lookup","l":"LookupQuery"},{"p":"org.apache.fluss.client.lookup","l":"LookupResult"},{"p":"org.apache.fluss.client.lookup","l":"LookupType"},{"p":"org.apache.fluss.client.metadata","l":"MetadataUpdater"},{"p":"org.apache.fluss.client.utils","l":"MetadataUtils"},{"p":"org.apache.fluss.flink.source.reader","l":"MutableRecordAndPos"},{"p":"org.apache.fluss.flink.source.enumerator.initializer","l":"NoStoppingOffsetsInitializer"},{"p":"org.apache.fluss.flink.source.enumerator.initializer","l":"OffsetsInitializer"},{"p":"org.apache.fluss.client.admin","l":"OffsetSpec"},{"p":"org.apache.fluss.flink.row","l":"OperationType"},{"p":"org.apache.fluss.flink.source.event","l":"PartitionBucketsUnsubscribedEvent"},{"p":"org.apache.fluss.client.table.getter","l":"PartitionGetter"},{"p":"org.apache.fluss.flink.source.event","l":"PartitionsRemovedEvent"},{"p":"org.apache.fluss.flink.utils","l":"PojoToRowConverter"},{"p":"org.apache.fluss.client.lookup","l":"PrefixLookup"},{"p":"org.apache.fluss.client.lookup","l":"PrefixLookupBatch"},{"p":"org.apache.fluss.client.lookup","l":"PrefixLookupQuery"},{"p":"org.apache.fluss.flink.procedure","l":"ProcedureBase"},{"p":"org.apache.fluss.flink.procedure","l":"ProcedureManager"},{"p":"org.apache.fluss.flink.utils","l":"PushdownUtils"},{"p":"org.apache.fluss.client.write","l":"RecordAccumulator.ReadyCheckResult"},{"p":"org.apache.fluss.client.write","l":"ReadyWriteBatch"},{"p":"org.apache.fluss.client.write","l":"RecordAccumulator"},{"p":"org.apache.fluss.flink.source.reader","l":"RecordAndPos"},{"p":"org.apache.fluss.client.write","l":"RecordAccumulator.RecordAppendResult"},{"p":"org.apache.fluss.flink.source.lookup","l":"LookupNormalizer.RemainingFilter"},{"p":"org.apache.fluss.client.table.scanner","l":"RemoteFileDownloader"},{"p":"org.apache.fluss.client.table.scanner.log","l":"RemoteLogDownloader"},{"p":"org.apache.fluss.client.table.scanner.log","l":"RemoteLogDownloadFuture"},{"p":"org.apache.fluss.client.write","l":"WriteBatch.RequestFuture"},{"p":"org.apache.fluss.client.write","l":"RoundRobinBucketAssigner"},{"p":"org.apache.fluss.flink.source.deserializer","l":"RowDataDeserializationSchema"},{"p":"org.apache.fluss.flink.sink.serializer","l":"RowDataSerializationSchema"},{"p":"org.apache.fluss.flink.row","l":"RowWithOp"},{"p":"org.apache.fluss.client.table.scanner","l":"Scan"},{"p":"org.apache.fluss.client.metrics","l":"ScannerMetricGroup"},{"p":"org.apache.fluss.client.table.scanner","l":"ScanRecord"},{"p":"org.apache.fluss.client.table.scanner.log","l":"ScanRecords"},{"p":"org.apache.fluss.flink","l":"FlinkConnectorOptions.ScanStartupMode"},{"p":"org.apache.fluss.client.token","l":"SecurityTokenManager"},{"p":"org.apache.fluss.client.token","l":"SecurityTokenProvider"},{"p":"org.apache.fluss.flink.lake.reader","l":"SeekableLakeSnapshotSplitScanner"},{"p":"org.apache.fluss.client.write","l":"Sender"},{"p":"org.apache.fluss.flink.sink.serializer","l":"SerializerInitContextImpl"},{"p":"org.apache.fluss.flink.adapter","l":"SingleThreadFetcherManagerAdapter"},{"p":"org.apache.fluss.flink.adapter","l":"SingleThreadMultiplexSourceReaderBaseAdapter"},{"p":"org.apache.fluss.flink.source.enumerator.initializer","l":"SnapshotOffsetsInitializer"},{"p":"org.apache.fluss.flink.source.split","l":"SnapshotSplit"},{"p":"org.apache.fluss.flink.source.state","l":"SourceEnumeratorState"},{"p":"org.apache.fluss.flink.source.split","l":"SourceSplitBase"},{"p":"org.apache.fluss.flink.source.split","l":"SourceSplitSerializer"},{"p":"org.apache.fluss.flink.source.split","l":"SourceSplitState"},{"p":"org.apache.fluss.flink.source.reader","l":"SplitScanner"},{"p":"org.apache.fluss.flink.utils","l":"FlinkConnectorOptionsUtils.StartupOptions"},{"p":"org.apache.fluss.client.write","l":"StickyBucketAssigner"},{"p":"org.apache.fluss.client.table","l":"Table"},{"p":"org.apache.fluss.client.table.writer","l":"TableAppend"},{"p":"org.apache.fluss.flink.tiering.source","l":"TableBucketWriteResult"},{"p":"org.apache.fluss.flink.tiering.source","l":"TableBucketWriteResultEmitter"},{"p":"org.apache.fluss.flink.tiering.source","l":"TableBucketWriteResultSerializer"},{"p":"org.apache.fluss.flink.tiering.source","l":"TableBucketWriteResultTypeInfo"},{"p":"org.apache.fluss.client.lookup","l":"TableLookup"},{"p":"org.apache.fluss.client.table.scanner","l":"TableScan"},{"p":"org.apache.fluss.client.table.writer","l":"TableUpsert"},{"p":"org.apache.fluss.client.table.writer","l":"TableWriter"},{"p":"org.apache.fluss.flink.tiering.committer","l":"TieringCommitOperator"},{"p":"org.apache.fluss.flink.tiering.committer","l":"TieringCommitOperatorFactory"},{"p":"org.apache.fluss.flink.tiering.committer","l":"TieringCommitterInitContext"},{"p":"org.apache.fluss.flink.tiering.source.split","l":"TieringLogSplit"},{"p":"org.apache.fluss.flink.tiering.event","l":"TieringRestoreEvent"},{"p":"org.apache.fluss.flink.tiering.source.split","l":"TieringSnapshotSplit"},{"p":"org.apache.fluss.flink.tiering.source","l":"TieringSource"},{"p":"org.apache.fluss.flink.tiering.source.enumerator","l":"TieringSourceEnumerator"},{"p":"org.apache.fluss.flink.tiering.source.state","l":"TieringSourceEnumeratorState"},{"p":"org.apache.fluss.flink.tiering.source.state","l":"TieringSourceEnumeratorStateSerializer"},{"p":"org.apache.fluss.flink.tiering.source","l":"TieringSourceOptions"},{"p":"org.apache.fluss.flink.tiering.source","l":"TieringSourceReader"},{"p":"org.apache.fluss.flink.tiering.source.split","l":"TieringSplit"},{"p":"org.apache.fluss.flink.tiering.source.split","l":"TieringSplitGenerator"},{"p":"org.apache.fluss.flink.tiering.source","l":"TieringSplitReader"},{"p":"org.apache.fluss.flink.tiering.source.split","l":"TieringSplitSerializer"},{"p":"org.apache.fluss.flink.tiering.source.state","l":"TieringSplitState"},{"p":"org.apache.fluss.flink.tiering.source","l":"TieringWriterInitContext"},{"p":"org.apache.fluss.flink.utils","l":"TimestampFormat"},{"p":"org.apache.fluss.flink.source.enumerator.initializer","l":"TimestampOffsetsInitializer"},{"p":"org.apache.fluss.client.admin","l":"OffsetSpec.TimestampSpec"},{"p":"org.apache.fluss.client.table.writer","l":"Upsert"},{"p":"org.apache.fluss.client.table.writer","l":"UpsertResult"},{"p":"org.apache.fluss.flink.sink.writer","l":"UpsertSinkWriter"},{"p":"org.apache.fluss.client.table.writer","l":"UpsertWrite"},{"p":"org.apache.fluss.client.table.writer","l":"UpsertWriter"},{"p":"org.apache.fluss.flink.utils","l":"PushdownUtils.ValueConversion"},{"p":"org.apache.fluss.client.write","l":"WriteBatch"},{"p":"org.apache.fluss.client.write","l":"WriteCallback"},{"p":"org.apache.fluss.client.write","l":"WriteFormat"},{"p":"org.apache.fluss.client.write","l":"WriterClient"},{"p":"org.apache.fluss.client.write","l":"WriteRecord"},{"p":"org.apache.fluss.client.metrics","l":"WriterMetricGroup"}] \ No newline at end of file diff --git a/website/static/javadoc/0.8-SNAPSHOT/type-search-index.zip b/website/static/javadoc/0.8-SNAPSHOT/type-search-index.zip new file mode 100644 index 0000000000000000000000000000000000000000..a787845eeaab0ee8dba81b769c96537531e7a2fc GIT binary patch literal 2356 zcmZveYd8}O8^_0tAq)|jIb=3ujE3gLa+=dLLdu?yQ%nwXjwdE^$Z0~(G|7@gnL~3f zhhZW)7db{HA|>VV^n7}s_qyKS{o(#!*Z;om&p)Cykmo1>3~a9e$8i0wXWmAzX{TG=4MVM^K8%yW zB{j@6uCeV0(#^a>59M>Bb#pLbH*X)WXr0f{pRjQ3A4?Uh*8*xDCFy`_X>MW0R*GQy zav5yb_vwQ3R?-rVGU*w8a)v4)iRlWQj4F`%ahIVSI$JInPqq)3ZE6V8SatWCiio%}UXx-|oh#+N7@(#+FHodpl+qKvC1K z1b1-@N)x2yrKa88lsPVL(q?)?gEKO*8o8^Y8&tgV6f7dJtxrbDHE*$o_W8>e{bn8` zhXo1>gh74z0zF^x_O%@t+;?_!Yeh?zd7v=v@7zB>!Vgrj#l`miz>$W06FTmg%gQdg zxNowt^vQ#}rXY4HEfi_RY$&X&15aIs0^L+~J%>RmIN~>}v?g1%Vt9zdU~p{*F5LmU zH%F38CKfkcimv73cKEAhfmM82Bbl z6X6*y+xA8m7aoz~T_>^O`ts1@K8e~+?{~uG<0oB^zhp%930rcML)it>#L5LVCxCLo z300)~CFl2+X@=a=>zbux&3@3xTmiKvO6=N(UQ)hXpK%6a+D%%o9vI}#+gig8S>-*s zruu%wL&j@b*j1+@OJ+4i_`Y$xq(Rshpk7RKJDAI)bNwf?#{OAQm^NL$ME|L@lwi}3 z8#;9!u>RQb5$-I(RB~!>e(nVs;3fw<>Vbcic4~n~<^U-a(RJUx6?eWc5<27)@F)Xu z0yx>(o;`~>XR~CBkJUS?EM#sHlU0yt-Xagw4?Yp+Jkub>kSHSCppW+v=TO5z+VPrhF9VNu^EY z?fyn=0!heXGkl~PMj<)ayt6N9z4)fJSSbpU46nY-6=y>o7=xOx@a6vMYccc|&y^s= zKn$@nq>sx)zS@+H=Mr2r3Ub=1$Y57)^vzhnWZ04fF5Z0(kV{eO(E>zl@7ER`L zox@s#RgEpOlPTBIb zV1AazvjmRUM@2ALx**F*-+LSYL>t%ot3Rl(@TG#*`#6W8%z-)$nh7{yB9g$GIooG zDoVrVSg$;1@sJ>6yQ-RLS|51}g^%m4L<-26&+zfP-xkcErCfn)#ov#UJ;5J-PI&rp z4Jz()zk(1sngZs2G^g-DVEwL5wMWt8++RmN*{pea6nW*PHwv48;Nwx=#C$emJGV04 zF*nUgfwqzzP~0C0dMs<`=%=$?tNmy1bBuHDpy4z*fc=ZP%qjH2Tv&Q@t{1vjuimHy zHCV$#iB7?`*1gZ-+oIktGIXDIObgzC%l?tKLs}Aw%KYNn=$YX2w8x;mo1=B@T5wYH zdANBZ)iQHbEOx~4Ddy`*a_UXLCHvxi&DjNixzQVU6M-{)h$^wA4z~=!X3>313w?xq zkmVG&C7)9OVbes)WgX$7NreO2?^M}NhqBow(MR&{$r%C3qhRpc# zt|#_{GPOo!FNf?2+DE^^7j`MA2*WzFG^8rDBI^Z_xrykg3SAdXZWbX|shj9XjAdeg zdu#Q&GQe9C%o8u*LTV}Tv|^Sdf$7=eI{4{Q`>i4_APdm9ZK=}xbL!w+10$X5i>ap~cB2q$bK+MSs3 z(k>|^@7?9q;7`GEBSpSEkrSL7;&u6Vdw7-u-3`c=A5B}+gHX4{m1gD2X#;jLyE$fk ztR%ws*7tp==hYv!s(7sCY)E}qrR)UaxOI|2^4(z?fTJ^LBc-JKl0vNYW;$cNjp{Q$ zocR|s+Yowf{6EK%b{Xm8OS3$UDxF0?8_IV1sEI2RRp=ghM&kN#XDpL3Lc05Kqa`ux z8RAaTne(BC3Gs^Qhe>8saqLG^Xk1sc=-El%#BNbjA|xfD?e+B3NT(#(2vUO=E!SK~ z)e-=r76~YkIVyj#)2AP&C3uFM?@wGI1;inqtPcvR+#AH1F^T;A$dY0umcjC_-%SHg z+;kA6XtAP@mUggPOKbl$$>KE`bVb6-jBP^2hri$@Xg`XA>ZR$S=}%IG*=Wl$<4pE= zr`1G(yk<0ds5Jk)a5f_LK!j+`2X5TZ=gR{F00bNWK(yxKJ_`7++x)xx$8G+X|8F`( UYY^{0cU-^6^!J|m{Tl%I8?JFQb^rhX literal 0 HcmV?d00001 diff --git a/website/static/javadoc/index.html b/website/static/javadoc/index.html new file mode 100644 index 0000000000..b9c9f98d2b --- /dev/null +++ b/website/static/javadoc/index.html @@ -0,0 +1,3 @@ + + +Apache Fluss 0.8-SNAPSHOT API From 4496bbae035012c8894acf57e48f3965d113c486 Mon Sep 17 00:00:00 2001 From: Mehul Batra Date: Fri, 19 Sep 2025 15:20:52 +0530 Subject: [PATCH 2/4] remove unwanted comments --- pom.xml | 77 --------------------------------------------------------- 1 file changed, 77 deletions(-) diff --git a/pom.xml b/pom.xml index 32b95848fb..67df7a52ca 100644 --- a/pom.xml +++ b/pom.xml @@ -1001,84 +1001,7 @@ - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - From d511f90ae40f9a6a959d1bb47954ef7aeeeec325 Mon Sep 17 00:00:00 2001 From: Mehul Batra Date: Sun, 28 Sep 2025 19:08:38 +0530 Subject: [PATCH 3/4] address Jark commennts --- tools/refresh_javadoc.sh | 62 - website/build_javadocs.sh | 181 + website/docusaurus.config.ts | 1 + .../0.8-SNAPSHOT/allclasses-index.html | 1627 --- .../javadoc/0.8-SNAPSHOT/allclasses.html | 268 - .../0.8-SNAPSHOT/allpackages-index.html | 350 - website/static/javadoc/0.8-SNAPSHOT/argfile | 1401 -- .../javadoc/0.8-SNAPSHOT/constant-values.html | 572 - .../javadoc/0.8-SNAPSHOT/deprecated-list.html | 175 - .../static/javadoc/0.8-SNAPSHOT/element-list | 47 - .../0.8-SNAPSHOT/index-files/index-1.html | 369 - .../0.8-SNAPSHOT/index-files/index-10.html | 157 - .../0.8-SNAPSHOT/index-files/index-11.html | 187 - .../0.8-SNAPSHOT/index-files/index-12.html | 524 - .../0.8-SNAPSHOT/index-files/index-13.html | 195 - .../0.8-SNAPSHOT/index-files/index-14.html | 219 - .../0.8-SNAPSHOT/index-files/index-15.html | 340 - .../0.8-SNAPSHOT/index-files/index-16.html | 314 - .../0.8-SNAPSHOT/index-files/index-17.html | 153 - .../0.8-SNAPSHOT/index-files/index-18.html | 333 - .../0.8-SNAPSHOT/index-files/index-19.html | 536 - .../0.8-SNAPSHOT/index-files/index-2.html | 248 - .../0.8-SNAPSHOT/index-files/index-20.html | 547 - .../0.8-SNAPSHOT/index-files/index-21.html | 222 - .../0.8-SNAPSHOT/index-files/index-22.html | 222 - .../0.8-SNAPSHOT/index-files/index-23.html | 232 - .../0.8-SNAPSHOT/index-files/index-3.html | 561 - .../0.8-SNAPSHOT/index-files/index-4.html | 334 - .../0.8-SNAPSHOT/index-files/index-5.html | 236 - .../0.8-SNAPSHOT/index-files/index-6.html | 548 - .../0.8-SNAPSHOT/index-files/index-7.html | 659 - .../0.8-SNAPSHOT/index-files/index-8.html | 238 - .../0.8-SNAPSHOT/index-files/index-9.html | 327 - .../static/javadoc/0.8-SNAPSHOT/index.html | 346 - .../static/javadoc/0.8-SNAPSHOT/javadoc.sh | 1 - .../0.8-SNAPSHOT/jquery-ui.overrides.css | 35 - .../jquery/external/jquery/jquery.js | 10872 --------------- .../0.8-SNAPSHOT/jquery/jquery-3.7.1.min.js | 2 - .../0.8-SNAPSHOT/jquery/jquery-ui.min.css | 6 - .../0.8-SNAPSHOT/jquery/jquery-ui.min.js | 6 - .../jquery/jszip-utils/dist/jszip-utils-ie.js | 56 - .../jszip-utils/dist/jszip-utils-ie.min.js | 10 - .../jquery/jszip-utils/dist/jszip-utils.js | 118 - .../jszip-utils/dist/jszip-utils.min.js | 10 - .../0.8-SNAPSHOT/jquery/jszip/dist/jszip.js | 11370 ---------------- .../jquery/jszip/dist/jszip.min.js | 13 - .../legal/ADDITIONAL_LICENSE_INFO | 37 - .../0.8-SNAPSHOT/legal/ASSEMBLY_EXCEPTION | 27 - .../static/javadoc/0.8-SNAPSHOT/legal/LICENSE | 347 - .../javadoc/0.8-SNAPSHOT/legal/jquery.md | 26 - .../javadoc/0.8-SNAPSHOT/legal/jqueryUI.md | 49 - .../javadoc/0.8-SNAPSHOT/legal/jszip.md | 653 - .../static/javadoc/0.8-SNAPSHOT/legal/pako.md | 45 - .../0.8-SNAPSHOT/member-search-index.js | 1 - .../0.8-SNAPSHOT/member-search-index.zip | Bin 20501 -> 0 bytes website/static/javadoc/0.8-SNAPSHOT/options | 35 - .../org/apache/fluss/client/Connection.html | 340 - .../fluss/client/ConnectionFactory.html | 314 - .../apache/fluss/client/FlussConnection.html | 435 - .../org/apache/fluss/client/admin/Admin.html | 1040 -- .../fluss/client/admin/CreateAclsResult.html | 385 - .../admin/DropAclsResult.FilterResult.html | 292 - .../admin/DropAclsResult.FilterResults.html | 274 - .../fluss/client/admin/DropAclsResult.html | 352 - .../apache/fluss/client/admin/FlussAdmin.html | 1173 -- .../fluss/client/admin/ListOffsetsResult.html | 326 - .../client/admin/OffsetSpec.EarliestSpec.html | 312 - .../client/admin/OffsetSpec.LatestSpec.html | 312 - .../admin/OffsetSpec.TimestampSpec.html | 350 - .../apache/fluss/client/admin/OffsetSpec.html | 399 - .../fluss/client/admin/class-use/Admin.html | 413 - .../admin/class-use/CreateAclsResult.html | 199 - .../DropAclsResult.FilterResult.html | 194 - .../DropAclsResult.FilterResults.html | 195 - .../admin/class-use/DropAclsResult.html | 199 - .../client/admin/class-use/FlussAdmin.html | 148 - .../admin/class-use/ListOffsetsResult.html | 221 - .../class-use/OffsetSpec.EarliestSpec.html | 148 - .../class-use/OffsetSpec.LatestSpec.html | 148 - .../class-use/OffsetSpec.TimestampSpec.html | 148 - .../client/admin/class-use/OffsetSpec.html | 281 - .../fluss/client/admin/package-summary.html | 241 - .../fluss/client/admin/package-tree.html | 185 - .../fluss/client/admin/package-use.html | 414 - .../fluss/client/class-use/Connection.html | 246 - .../client/class-use/ConnectionFactory.html | 148 - .../client/class-use/FlussConnection.html | 217 - .../client/lookup/AbstractLookupBatch.html | 412 - .../client/lookup/AbstractLookupQuery.html | 356 - .../apache/fluss/client/lookup/Lookup.html | 354 - .../fluss/client/lookup/LookupBatch.html | 370 - .../fluss/client/lookup/LookupClient.html | 397 - .../fluss/client/lookup/LookupQuery.html | 303 - .../fluss/client/lookup/LookupResult.html | 395 - .../fluss/client/lookup/LookupType.html | 380 - .../apache/fluss/client/lookup/Lookuper.html | 269 - .../fluss/client/lookup/PrefixLookup.html | 312 - .../client/lookup/PrefixLookupBatch.html | 368 - .../client/lookup/PrefixLookupQuery.html | 303 - .../fluss/client/lookup/TableLookup.html | 377 - .../lookup/class-use/AbstractLookupBatch.html | 148 - .../lookup/class-use/AbstractLookupQuery.html | 248 - .../fluss/client/lookup/class-use/Lookup.html | 255 - .../client/lookup/class-use/LookupBatch.html | 193 - .../client/lookup/class-use/LookupClient.html | 218 - .../client/lookup/class-use/LookupQuery.html | 207 - .../client/lookup/class-use/LookupResult.html | 194 - .../client/lookup/class-use/LookupType.html | 246 - .../client/lookup/class-use/Lookuper.html | 200 - .../client/lookup/class-use/PrefixLookup.html | 148 - .../lookup/class-use/PrefixLookupBatch.html | 193 - .../lookup/class-use/PrefixLookupQuery.html | 207 - .../client/lookup/class-use/TableLookup.html | 148 - .../fluss/client/lookup/package-summary.html | 265 - .../fluss/client/lookup/package-tree.html | 196 - .../fluss/client/lookup/package-use.html | 328 - .../client/metadata/KvSnapshotMetadata.html | 354 - .../fluss/client/metadata/KvSnapshots.html | 385 - .../fluss/client/metadata/LakeSnapshot.html | 363 - .../client/metadata/MetadataUpdater.html | 721 - .../class-use/KvSnapshotMetadata.html | 227 - .../metadata/class-use/KvSnapshots.html | 239 - .../metadata/class-use/LakeSnapshot.html | 225 - .../metadata/class-use/MetadataUpdater.html | 398 - .../client/metadata/package-summary.html | 186 - .../fluss/client/metadata/package-tree.html | 166 - .../fluss/client/metadata/package-use.html | 359 - .../client/metrics/ScannerMetricGroup.html | 494 - .../client/metrics/WriterMetricGroup.html | 460 - .../metrics/class-use/ScannerMetricGroup.html | 203 - .../metrics/class-use/WriterMetricGroup.html | 197 - .../fluss/client/metrics/package-summary.html | 174 - .../fluss/client/metrics/package-tree.html | 168 - .../fluss/client/metrics/package-use.html | 208 - .../apache/fluss/client/package-summary.html | 191 - .../org/apache/fluss/client/package-tree.html | 174 - .../org/apache/fluss/client/package-use.html | 254 - .../apache/fluss/client/table/FlussTable.html | 451 - .../org/apache/fluss/client/table/Table.html | 358 - .../client/table/class-use/FlussTable.html | 148 - .../fluss/client/table/class-use/Table.html | 306 - .../client/table/getter/PartitionGetter.html | 309 - .../getter/class-use/PartitionGetter.html | 197 - .../client/table/getter/package-summary.html | 168 - .../client/table/getter/package-tree.html | 163 - .../client/table/getter/package-use.html | 185 - .../fluss/client/table/package-summary.html | 185 - .../fluss/client/table/package-tree.html | 173 - .../fluss/client/table/package-use.html | 277 - .../table/scanner/RemoteFileDownloader.html | 435 - .../fluss/client/table/scanner/Scan.html | 374 - .../client/table/scanner/ScanRecord.html | 447 - .../fluss/client/table/scanner/TableScan.html | 449 - .../table/scanner/batch/BatchScanUtils.html | 367 - .../table/scanner/batch/BatchScanner.html | 305 - .../scanner/batch/KvSnapshotBatchScanner.html | 423 - .../scanner/batch/LimitBatchScanner.html | 365 - .../batch/class-use/BatchScanUtils.html | 148 - .../scanner/batch/class-use/BatchScanner.html | 358 - .../class-use/KvSnapshotBatchScanner.html | 148 - .../batch/class-use/LimitBatchScanner.html | 148 - .../table/scanner/batch/package-summary.html | 197 - .../table/scanner/batch/package-tree.html | 179 - .../table/scanner/batch/package-use.html | 254 - .../class-use/RemoteFileDownloader.html | 268 - .../client/table/scanner/class-use/Scan.html | 274 - .../table/scanner/class-use/ScanRecord.html | 351 - .../table/scanner/class-use/TableScan.html | 148 - .../table/scanner/log/LogFetchBuffer.html | 340 - .../table/scanner/log/LogFetchCollector.html | 329 - .../client/table/scanner/log/LogFetcher.html | 423 - .../client/table/scanner/log/LogScan.html | 334 - .../client/table/scanner/log/LogScanner.html | 503 - .../table/scanner/log/LogScannerImpl.html | 502 - .../table/scanner/log/LogScannerStatus.html | 271 - .../scanner/log/RemoteLogDownloadFuture.html | 351 - .../scanner/log/RemoteLogDownloader.html | 378 - .../client/table/scanner/log/ScanRecords.html | 445 - .../scanner/log/class-use/LogFetchBuffer.html | 194 - .../log/class-use/LogFetchCollector.html | 148 - .../scanner/log/class-use/LogFetcher.html | 148 - .../table/scanner/log/class-use/LogScan.html | 194 - .../scanner/log/class-use/LogScanner.html | 227 - .../scanner/log/class-use/LogScannerImpl.html | 148 - .../log/class-use/LogScannerStatus.html | 203 - .../class-use/RemoteLogDownloadFuture.html | 195 - .../log/class-use/RemoteLogDownloader.html | 148 - .../scanner/log/class-use/ScanRecords.html | 214 - .../table/scanner/log/package-summary.html | 235 - .../table/scanner/log/package-tree.html | 181 - .../client/table/scanner/log/package-use.html | 239 - .../client/table/scanner/package-summary.html | 198 - .../client/table/scanner/package-tree.html | 171 - .../client/table/scanner/package-use.html | 332 - .../table/writer/AbstractTableWriter.html | 443 - .../fluss/client/table/writer/Append.html | 265 - .../client/table/writer/AppendResult.html | 274 - .../client/table/writer/AppendWriter.html | 277 - .../client/table/writer/DeleteResult.html | 274 - .../client/table/writer/TableAppend.html | 324 - .../client/table/writer/TableUpsert.html | 384 - .../client/table/writer/TableWriter.html | 275 - .../fluss/client/table/writer/Upsert.html | 322 - .../client/table/writer/UpsertResult.html | 274 - .../client/table/writer/UpsertWrite.html | 352 - .../client/table/writer/UpsertWriter.html | 302 - .../writer/class-use/AbstractTableWriter.html | 148 - .../client/table/writer/class-use/Append.html | 228 - .../table/writer/class-use/AppendResult.html | 194 - .../table/writer/class-use/AppendWriter.html | 199 - .../table/writer/class-use/DeleteResult.html | 195 - .../table/writer/class-use/TableAppend.html | 148 - .../table/writer/class-use/TableUpsert.html | 148 - .../table/writer/class-use/TableWriter.html | 218 - .../client/table/writer/class-use/Upsert.html | 260 - .../table/writer/class-use/UpsertResult.html | 194 - .../table/writer/class-use/UpsertWrite.html | 194 - .../table/writer/class-use/UpsertWriter.html | 200 - .../client/table/writer/package-summary.html | 246 - .../client/table/writer/package-tree.html | 182 - .../client/table/writer/package-use.html | 264 - .../token/DefaultSecurityTokenManager.html | 360 - .../token/DefaultSecurityTokenProvider.html | 325 - .../client/token/SecurityTokenManager.html | 280 - .../client/token/SecurityTokenProvider.html | 265 - .../DefaultSecurityTokenManager.html | 148 - .../DefaultSecurityTokenProvider.html | 148 - .../token/class-use/SecurityTokenManager.html | 194 - .../class-use/SecurityTokenProvider.html | 208 - .../fluss/client/token/package-summary.html | 197 - .../fluss/client/token/package-tree.html | 171 - .../fluss/client/token/package-use.html | 191 - .../client/utils/ClientRpcMessageUtils.html | 556 - .../fluss/client/utils/ClientUtils.html | 343 - .../fluss/client/utils/MetadataUtils.html | 420 - .../class-use/ClientRpcMessageUtils.html | 148 - .../client/utils/class-use/ClientUtils.html | 148 - .../client/utils/class-use/MetadataUtils.html | 148 - .../fluss/client/utils/package-summary.html | 181 - .../fluss/client/utils/package-tree.html | 165 - .../fluss/client/utils/package-use.html | 148 - .../client/write/ArrowLogWriteBatch.html | 597 - .../client/write/DynamicPartitionCreator.html | 314 - .../write/DynamicWriteBatchSizeEstimator.html | 329 - .../client/write/HashBucketAssigner.html | 370 - .../client/write/IdempotenceBucketEntry.html | 224 - .../client/write/IdempotenceBucketMap.html | 270 - .../client/write/IdempotenceManager.html | 276 - .../client/write/IndexedLogWriteBatch.html | 597 - .../fluss/client/write/KvWriteBatch.html | 617 - .../fluss/client/write/ReadyWriteBatch.html | 325 - .../RecordAccumulator.ReadyCheckResult.html | 347 - .../RecordAccumulator.RecordAppendResult.html | 350 - .../fluss/client/write/RecordAccumulator.html | 541 - .../write/RoundRobinBucketAssigner.html | 355 - .../org/apache/fluss/client/write/Sender.html | 402 - .../client/write/StickyBucketAssigner.html | 356 - .../write/WriteBatch.RequestFuture.html | 336 - .../apache/fluss/client/write/WriteBatch.html | 763 -- .../fluss/client/write/WriteCallback.html | 255 - .../fluss/client/write/WriteFormat.html | 393 - .../fluss/client/write/WriteRecord.html | 461 - .../fluss/client/write/WriterClient.html | 419 - .../write/class-use/ArrowLogWriteBatch.html | 148 - .../class-use/DynamicPartitionCreator.html | 148 - .../DynamicWriteBatchSizeEstimator.html | 148 - .../write/class-use/HashBucketAssigner.html | 148 - .../class-use/IdempotenceBucketEntry.html | 148 - .../write/class-use/IdempotenceBucketMap.html | 148 - .../write/class-use/IdempotenceManager.html | 197 - .../write/class-use/IndexedLogWriteBatch.html | 148 - .../client/write/class-use/KvWriteBatch.html | 148 - .../write/class-use/ReadyWriteBatch.html | 249 - .../RecordAccumulator.ReadyCheckResult.html | 194 - .../RecordAccumulator.RecordAppendResult.html | 198 - .../write/class-use/RecordAccumulator.html | 197 - .../class-use/RoundRobinBucketAssigner.html | 148 - .../fluss/client/write/class-use/Sender.html | 148 - .../write/class-use/StickyBucketAssigner.html | 148 - .../class-use/WriteBatch.RequestFuture.html | 192 - .../client/write/class-use/WriteBatch.html | 254 - .../client/write/class-use/WriteCallback.html | 248 - .../client/write/class-use/WriteFormat.html | 207 - .../client/write/class-use/WriteRecord.html | 308 - .../client/write/class-use/WriterClient.html | 245 - .../fluss/client/write/package-summary.html | 319 - .../fluss/client/write/package-tree.html | 205 - .../fluss/client/write/package-use.html | 317 - ...FlinkConnectorOptions.ScanStartupMode.html | 446 - .../fluss/flink/FlinkConnectorOptions.html | 476 - .../flink/adapter/CatalogTableAdapter.html | 315 - .../SingleThreadFetcherManagerAdapter.html | 320 - ...hreadMultiplexSourceReaderBaseAdapter.html | 334 - .../class-use/CatalogTableAdapter.html | 148 - .../SingleThreadFetcherManagerAdapter.html | 194 - ...hreadMultiplexSourceReaderBaseAdapter.html | 194 - .../fluss/flink/adapter/package-summary.html | 181 - .../fluss/flink/adapter/package-tree.html | 181 - .../fluss/flink/adapter/package-use.html | 209 - .../fluss/flink/catalog/FlinkCatalog.html | 1463 -- .../flink/catalog/FlinkCatalogFactory.html | 446 - .../flink/catalog/FlinkCatalogOptions.html | 271 - .../flink/catalog/FlinkTableFactory.html | 412 - .../flink/catalog/class-use/FlinkCatalog.html | 192 - .../class-use/FlinkCatalogFactory.html | 148 - .../class-use/FlinkCatalogOptions.html | 148 - .../catalog/class-use/FlinkTableFactory.html | 148 - .../fluss/flink/catalog/package-summary.html | 186 - .../fluss/flink/catalog/package-tree.html | 170 - .../fluss/flink/catalog/package-use.html | 185 - ...FlinkConnectorOptions.ScanStartupMode.html | 243 - .../class-use/FlinkConnectorOptions.html | 148 - .../apache/fluss/flink/lake/LakeCatalog.html | 318 - .../flink/lake/LakeRecordRecordEmitter.html | 311 - .../fluss/flink/lake/LakeSplitGenerator.html | 330 - .../flink/lake/LakeSplitReaderGenerator.html | 329 - .../fluss/flink/lake/LakeSplitSerializer.html | 340 - .../flink/lake/LakeSplitStateInitializer.html | 307 - .../fluss/flink/lake/LakeTableFactory.html | 309 - .../flink/lake/class-use/LakeCatalog.html | 148 - .../class-use/LakeRecordRecordEmitter.html | 148 - .../lake/class-use/LakeSplitGenerator.html | 148 - .../class-use/LakeSplitReaderGenerator.html | 148 - .../lake/class-use/LakeSplitSerializer.html | 148 - .../class-use/LakeSplitStateInitializer.html | 148 - .../lake/class-use/LakeTableFactory.html | 148 - .../fluss/flink/lake/package-summary.html | 204 - .../apache/fluss/flink/lake/package-tree.html | 169 - .../apache/fluss/flink/lake/package-use.html | 148 - .../IndexedLakeSplitRecordIterator.html | 393 - .../fluss/flink/lake/reader/KeyValueRow.html | 339 - .../LakeSnapshotAndLogSplitScanner.html | 363 - .../lake/reader/LakeSnapshotScanner.html | 358 - .../SeekableLakeSnapshotSplitScanner.html | 373 - .../IndexedLakeSplitRecordIterator.html | 148 - .../lake/reader/class-use/KeyValueRow.html | 148 - .../LakeSnapshotAndLogSplitScanner.html | 148 - .../reader/class-use/LakeSnapshotScanner.html | 148 - .../SeekableLakeSnapshotSplitScanner.html | 148 - .../flink/lake/reader/package-summary.html | 193 - .../fluss/flink/lake/reader/package-tree.html | 167 - .../fluss/flink/lake/reader/package-use.html | 148 - .../split/LakeSnapshotAndFlussLogSplit.html | 624 - .../flink/lake/split/LakeSnapshotSplit.html | 501 - .../LakeSnapshotAndFlussLogSplit.html | 258 - .../split/class-use/LakeSnapshotSplit.html | 215 - .../flink/lake/split/package-summary.html | 174 - .../fluss/flink/lake/split/package-tree.html | 168 - .../fluss/flink/lake/split/package-use.html | 243 - .../LakeSnapshotAndFlussLogSplitState.html | 365 - .../lake/state/LakeSnapshotSplitState.html | 337 - .../LakeSnapshotAndFlussLogSplitState.html | 148 - .../class-use/LakeSnapshotSplitState.html | 148 - .../flink/lake/state/package-summary.html | 174 - .../fluss/flink/lake/state/package-tree.html | 168 - .../fluss/flink/lake/state/package-use.html | 148 - .../fluss/flink/metrics/FlinkCounter.html | 395 - .../fluss/flink/metrics/FlinkGauge.html | 323 - .../fluss/flink/metrics/FlinkHistogram.html | 359 - .../fluss/flink/metrics/FlinkMeter.html | 377 - .../flink/metrics/FlinkMetricRegistry.html | 477 - .../flink/metrics/class-use/FlinkCounter.html | 148 - .../flink/metrics/class-use/FlinkGauge.html | 148 - .../metrics/class-use/FlinkHistogram.html | 148 - .../flink/metrics/class-use/FlinkMeter.html | 148 - .../class-use/FlinkMetricRegistry.html | 192 - .../fluss/flink/metrics/package-summary.html | 193 - .../fluss/flink/metrics/package-tree.html | 167 - .../fluss/flink/metrics/package-use.html | 186 - .../apache/fluss/flink/package-summary.html | 185 - .../org/apache/fluss/flink/package-tree.html | 177 - .../org/apache/fluss/flink/package-use.html | 208 - .../flink/procedure/AbstractAclProcedure.html | 419 - .../flink/procedure/AddAclProcedure.html | 397 - .../flink/procedure/DropAclProcedure.html | 401 - .../flink/procedure/ListAclProcedure.html | 401 - .../fluss/flink/procedure/ProcedureBase.html | 358 - .../flink/procedure/ProcedureManager.html | 323 - .../class-use/AbstractAclProcedure.html | 208 - .../procedure/class-use/AddAclProcedure.html | 148 - .../procedure/class-use/DropAclProcedure.html | 148 - .../procedure/class-use/ListAclProcedure.html | 148 - .../procedure/class-use/ProcedureBase.html | 215 - .../procedure/class-use/ProcedureManager.html | 148 - .../flink/procedure/package-summary.html | 198 - .../fluss/flink/procedure/package-tree.html | 174 - .../fluss/flink/procedure/package-use.html | 191 - .../fluss/flink/row/FlinkAsFlussRow.html | 643 - .../apache/fluss/flink/row/OperationType.html | 432 - .../org/apache/fluss/flink/row/RowWithOp.html | 406 - .../flink/row/class-use/FlinkAsFlussRow.html | 192 - .../flink/row/class-use/OperationType.html | 225 - .../fluss/flink/row/class-use/RowWithOp.html | 201 - .../fluss/flink/row/package-summary.html | 192 - .../apache/fluss/flink/row/package-tree.html | 178 - .../apache/fluss/flink/row/package-use.html | 215 - .../fluss/flink/sink/ChannelComputer.html | 315 - .../sink/FlinkRowDataChannelComputer.html | 369 - .../flink/sink/FlinkStreamPartitioner.html | 454 - .../fluss/flink/sink/FlinkTableSink.html | 541 - .../apache/fluss/flink/sink/FlussSink.html | 376 - .../fluss/flink/sink/FlussSinkBuilder.html | 455 - .../flink/sink/class-use/ChannelComputer.html | 224 - .../FlinkRowDataChannelComputer.html | 148 - .../class-use/FlinkStreamPartitioner.html | 148 - .../flink/sink/class-use/FlinkTableSink.html | 148 - .../fluss/flink/sink/class-use/FlussSink.html | 194 - .../sink/class-use/FlussSinkBuilder.html | 244 - .../fluss/flink/sink/package-summary.html | 210 - .../apache/fluss/flink/sink/package-tree.html | 181 - .../apache/fluss/flink/sink/package-use.html | 198 - ...alizationSchema.InitializationContext.html | 272 - .../serializer/FlussSerializationSchema.html | 331 - .../RowDataSerializationSchema.html | 384 - .../serializer/SerializerInitContextImpl.html | 322 - ...alizationSchema.InitializationContext.html | 218 - .../class-use/FlussSerializationSchema.html | 294 - .../class-use/RowDataSerializationSchema.html | 148 - .../class-use/SerializerInitContextImpl.html | 148 - .../sink/serializer/package-summary.html | 197 - .../flink/sink/serializer/package-tree.html | 175 - .../flink/sink/serializer/package-use.html | 237 - .../flink/sink/writer/AppendSinkWriter.html | 399 - .../flink/sink/writer/FlinkSinkWriter.html | 593 - .../flink/sink/writer/UpsertSinkWriter.html | 404 - .../writer/class-use/AppendSinkWriter.html | 148 - .../writer/class-use/FlinkSinkWriter.html | 201 - .../writer/class-use/UpsertSinkWriter.html | 148 - .../flink/sink/writer/package-summary.html | 180 - .../fluss/flink/sink/writer/package-tree.html | 168 - .../fluss/flink/sink/writer/package-use.html | 185 - .../fluss/flink/source/FlinkSource.html | 550 - .../fluss/flink/source/FlinkTableSource.html | 704 - .../fluss/flink/source/FlussSource.html | 338 - .../flink/source/FlussSourceBuilder.html | 547 - .../flink/source/class-use/FlinkSource.html | 194 - .../source/class-use/FlinkTableSource.html | 148 - .../flink/source/class-use/FlussSource.html | 194 - .../source/class-use/FlussSourceBuilder.html | 250 - .../DeserializerInitContextImpl.html | 380 - ...alizationSchema.InitializationContext.html | 322 - .../FlussDeserializationSchema.html | 357 - .../JsonStringDeserializationSchema.html | 435 - .../RowDataDeserializationSchema.html | 414 - .../DeserializerInitContextImpl.html | 148 - ...alizationSchema.InitializationContext.html | 225 - .../class-use/FlussDeserializationSchema.html | 292 - .../JsonStringDeserializationSchema.html | 148 - .../RowDataDeserializationSchema.html | 148 - .../source/deserializer/package-summary.html | 205 - .../source/deserializer/package-tree.html | 176 - .../source/deserializer/package-use.html | 240 - .../source/emitter/FlinkRecordEmitter.html | 327 - .../emitter/class-use/FlinkRecordEmitter.html | 198 - .../flink/source/emitter/package-summary.html | 168 - .../flink/source/emitter/package-tree.html | 163 - .../flink/source/emitter/package-use.html | 185 - .../enumerator/FlinkSourceEnumerator.html | 580 - .../class-use/FlinkSourceEnumerator.html | 148 - .../BucketOffsetsRetrieverImpl.html | 365 - .../initializer/LatestOffsetsInitializer.html | 357 - .../NoStoppingOffsetsInitializer.html | 356 - ...etsInitializer.BucketOffsetsRetriever.html | 299 - .../initializer/OffsetsInitializer.html | 410 - .../SnapshotOffsetsInitializer.html | 357 - .../TimestampOffsetsInitializer.html | 356 - .../class-use/BucketOffsetsRetrieverImpl.html | 148 - .../class-use/LatestOffsetsInitializer.html | 148 - .../NoStoppingOffsetsInitializer.html | 148 - ...etsInitializer.BucketOffsetsRetriever.html | 273 - .../class-use/OffsetsInitializer.html | 431 - .../class-use/SnapshotOffsetsInitializer.html | 148 - .../TimestampOffsetsInitializer.html | 148 - .../initializer/package-summary.html | 217 - .../enumerator/initializer/package-tree.html | 178 - .../enumerator/initializer/package-use.html | 268 - .../source/enumerator/package-summary.html | 168 - .../flink/source/enumerator/package-tree.html | 163 - .../flink/source/enumerator/package-use.html | 148 - .../PartitionBucketsUnsubscribedEvent.html | 371 - .../source/event/PartitionsRemovedEvent.html | 337 - .../PartitionBucketsUnsubscribedEvent.html | 148 - .../class-use/PartitionsRemovedEvent.html | 148 - .../flink/source/event/package-summary.html | 175 - .../flink/source/event/package-tree.html | 164 - .../fluss/flink/source/event/package-use.html | 148 - .../lookup/FlinkAsyncLookupFunction.html | 417 - .../source/lookup/FlinkLookupFunction.html | 418 - .../LookupNormalizer.RemainingFilter.html | 271 - .../flink/source/lookup/LookupNormalizer.html | 418 - .../class-use/FlinkAsyncLookupFunction.html | 148 - .../lookup/class-use/FlinkLookupFunction.html | 148 - .../LookupNormalizer.RemainingFilter.html | 192 - .../lookup/class-use/LookupNormalizer.html | 234 - .../flink/source/lookup/package-summary.html | 187 - .../flink/source/lookup/package-tree.html | 186 - .../flink/source/lookup/package-use.html | 192 - .../metrics/FlinkSourceReaderMetrics.html | 523 - .../class-use/FlinkSourceReaderMetrics.html | 242 - .../flink/source/metrics/package-summary.html | 168 - .../flink/source/metrics/package-tree.html | 163 - .../flink/source/metrics/package-use.html | 185 - .../fluss/flink/source/package-summary.html | 186 - .../fluss/flink/source/package-tree.html | 169 - .../fluss/flink/source/package-use.html | 197 - .../source/reader/BoundedSplitReader.html | 350 - .../reader/FlinkRecordsWithSplitIds.html | 444 - .../source/reader/FlinkSourceReader.html | 443 - .../source/reader/FlinkSourceSplitReader.html | 411 - .../source/reader/MutableRecordAndPos.html | 359 - .../flink/source/reader/RecordAndPos.html | 544 - .../flink/source/reader/SplitScanner.html | 269 - .../reader/class-use/BoundedSplitReader.html | 192 - .../class-use/FlinkRecordsWithSplitIds.html | 192 - .../reader/class-use/FlinkSourceReader.html | 148 - .../class-use/FlinkSourceSplitReader.html | 148 - .../reader/class-use/MutableRecordAndPos.html | 148 - .../source/reader/class-use/RecordAndPos.html | 359 - .../source/reader/class-use/SplitScanner.html | 148 - .../fetcher/FlinkSourceFetcherManager.html | 373 - .../class-use/FlinkSourceFetcherManager.html | 148 - .../reader/fetcher/package-summary.html | 168 - .../source/reader/fetcher/package-tree.html | 175 - .../source/reader/fetcher/package-use.html | 148 - .../flink/source/reader/package-summary.html | 217 - .../flink/source/reader/package-tree.html | 193 - .../flink/source/reader/package-use.html | 274 - .../source/split/HybridSnapshotLogSplit.html | 476 - .../split/HybridSnapshotLogSplitState.html | 368 - .../fluss/flink/source/split/LogSplit.html | 501 - .../flink/source/split/LogSplitState.html | 354 - .../flink/source/split/SnapshotSplit.html | 496 - .../flink/source/split/SourceSplitBase.html | 582 - .../source/split/SourceSplitSerializer.html | 362 - .../flink/source/split/SourceSplitState.html | 435 - .../class-use/HybridSnapshotLogSplit.html | 212 - .../HybridSnapshotLogSplitState.html | 194 - .../source/split/class-use/LogSplit.html | 212 - .../source/split/class-use/LogSplitState.html | 194 - .../source/split/class-use/SnapshotSplit.html | 195 - .../split/class-use/SourceSplitBase.html | 733 - .../class-use/SourceSplitSerializer.html | 148 - .../split/class-use/SourceSplitState.html | 364 - .../flink/source/split/package-summary.html | 211 - .../flink/source/split/package-tree.html | 179 - .../fluss/flink/source/split/package-use.html | 447 - .../FlussSourceEnumeratorStateSerializer.html | 362 - .../source/state/SourceEnumeratorState.html | 395 - .../FlussSourceEnumeratorStateSerializer.html | 148 - .../class-use/SourceEnumeratorState.html | 287 - .../flink/source/state/package-summary.html | 174 - .../flink/source/state/package-tree.html | 164 - .../fluss/flink/source/state/package-use.html | 231 - .../flink/tiering/LakeTieringJobBuilder.html | 292 - .../class-use/LakeTieringJobBuilder.html | 195 - .../tiering/committer/CommittableMessage.html | 318 - .../committer/CommittableMessageTypeInfo.html | 485 - .../FlussTableLakeSnapshotCommitter.html | 354 - .../committer/TieringCommitOperator.html | 499 - .../TieringCommitOperatorFactory.html | 376 - .../TieringCommitterInitContext.html | 316 - .../class-use/CommittableMessage.html | 254 - .../class-use/CommittableMessageTypeInfo.html | 148 - .../FlussTableLakeSnapshotCommitter.html | 148 - .../class-use/TieringCommitOperator.html | 148 - .../TieringCommitOperatorFactory.html | 148 - .../TieringCommitterInitContext.html | 148 - .../tiering/committer/package-summary.html | 199 - .../flink/tiering/committer/package-tree.html | 180 - .../flink/tiering/committer/package-use.html | 185 - .../tiering/event/FailedTieringEvent.html | 332 - .../tiering/event/FinishedTieringEvent.html | 316 - .../tiering/event/TieringRestoreEvent.html | 278 - .../event/class-use/FailedTieringEvent.html | 148 - .../event/class-use/FinishedTieringEvent.html | 148 - .../event/class-use/TieringRestoreEvent.html | 148 - .../flink/tiering/event/package-summary.html | 180 - .../flink/tiering/event/package-tree.html | 165 - .../flink/tiering/event/package-use.html | 148 - .../fluss/flink/tiering/package-summary.html | 168 - .../fluss/flink/tiering/package-tree.html | 163 - .../fluss/flink/tiering/package-use.html | 185 - .../source/TableBucketWriteResult.html | 405 - .../source/TableBucketWriteResultEmitter.html | 320 - .../TableBucketWriteResultSerializer.html | 362 - .../TableBucketWriteResultTypeInfo.html | 485 - .../tiering/source/TieringSource.Builder.html | 327 - .../flink/tiering/source/TieringSource.html | 514 - .../tiering/source/TieringSourceOptions.html | 329 - .../tiering/source/TieringSourceReader.html | 446 - .../tiering/source/TieringSplitReader.html | 385 - .../source/TieringWriterInitContext.html | 378 - .../class-use/TableBucketWriteResult.html | 293 - .../TableBucketWriteResultEmitter.html | 148 - .../TableBucketWriteResultSerializer.html | 148 - .../TableBucketWriteResultTypeInfo.html | 148 - .../class-use/TieringSource.Builder.html | 192 - .../source/class-use/TieringSource.html | 192 - .../class-use/TieringSourceOptions.html | 148 - .../source/class-use/TieringSourceReader.html | 148 - .../source/class-use/TieringSplitReader.html | 148 - .../class-use/TieringWriterInitContext.html | 148 - .../enumerator/TieringSourceEnumerator.html | 468 - .../class-use/TieringSourceEnumerator.html | 148 - .../source/enumerator/package-summary.html | 169 - .../source/enumerator/package-tree.html | 163 - .../source/enumerator/package-use.html | 148 - .../flink/tiering/source/package-summary.html | 224 - .../flink/tiering/source/package-tree.html | 184 - .../flink/tiering/source/package-use.html | 224 - .../tiering/source/split/TieringLogSplit.html | 473 - .../source/split/TieringSnapshotSplit.html | 474 - .../tiering/source/split/TieringSplit.html | 660 - .../source/split/TieringSplitGenerator.html | 312 - .../source/split/TieringSplitSerializer.html | 404 - .../split/class-use/TieringLogSplit.html | 199 - .../split/class-use/TieringSnapshotSplit.html | 199 - .../source/split/class-use/TieringSplit.html | 452 - .../class-use/TieringSplitGenerator.html | 148 - .../class-use/TieringSplitSerializer.html | 192 - .../tiering/source/split/package-summary.html | 192 - .../tiering/source/split/package-tree.html | 170 - .../tiering/source/split/package-use.html | 272 - .../state/TieringSourceEnumeratorState.html | 329 - ...ieringSourceEnumeratorStateSerializer.html | 404 - .../source/state/TieringSplitState.html | 353 - .../TieringSourceEnumeratorState.html | 287 - ...ieringSourceEnumeratorStateSerializer.html | 192 - .../state/class-use/TieringSplitState.html | 230 - .../tiering/source/state/package-summary.html | 180 - .../tiering/source/state/package-tree.html | 165 - .../tiering/source/state/package-use.html | 243 - .../flink/utils/CatalogExceptionUtils.html | 407 - .../flink/utils/CatalogPropertiesUtils.html | 379 - .../fluss/flink/utils/DataLakeUtils.html | 321 - ...kConnectorOptionsUtils.StartupOptions.html | 329 - .../utils/FlinkConnectorOptionsUtils.html | 424 - .../fluss/flink/utils/FlinkConversions.html | 422 - ...Converter.FlussSerializationConverter.html | 275 - .../utils/FlinkRowToFlussRowConverter.html | 371 - .../apache/fluss/flink/utils/FlinkUtils.html | 317 - ...nverter.FlussDeserializationConverter.html | 275 - .../utils/FlussRowToFlinkRowConverter.html | 381 - ...sonConverters.FlussRowToJsonConverter.html | 269 - .../flink/utils/FlussRowToJsonConverters.html | 332 - .../fluss/flink/utils/LakeSourceUtils.html | 356 - .../fluss/flink/utils/PojoToRowConverter.html | 349 - .../flink/utils/PushdownUtils.FieldEqual.html | 340 - .../utils/PushdownUtils.ValueConversion.html | 383 - .../fluss/flink/utils/PushdownUtils.html | 434 - .../fluss/flink/utils/TimestampFormat.html | 390 - .../class-use/CatalogExceptionUtils.html | 148 - .../class-use/CatalogPropertiesUtils.html | 148 - .../flink/utils/class-use/DataLakeUtils.html | 148 - ...kConnectorOptionsUtils.StartupOptions.html | 231 - .../class-use/FlinkConnectorOptionsUtils.html | 148 - .../utils/class-use/FlinkConversions.html | 148 - ...Converter.FlussSerializationConverter.html | 148 - .../FlinkRowToFlussRowConverter.html | 204 - .../flink/utils/class-use/FlinkUtils.html | 148 - ...nverter.FlussDeserializationConverter.html | 214 - .../FlussRowToFlinkRowConverter.html | 148 - ...sonConverters.FlussRowToJsonConverter.html | 192 - .../class-use/FlussRowToJsonConverters.html | 148 - .../utils/class-use/LakeSourceUtils.html | 148 - .../utils/class-use/PojoToRowConverter.html | 148 - .../class-use/PushdownUtils.FieldEqual.html | 292 - .../PushdownUtils.ValueConversion.html | 223 - .../flink/utils/class-use/PushdownUtils.html | 148 - .../utils/class-use/TimestampFormat.html | 215 - .../fluss/flink/utils/package-summary.html | 300 - .../fluss/flink/utils/package-tree.html | 203 - .../apache/fluss/flink/utils/package-use.html | 276 - .../0.8-SNAPSHOT/overview-summary.html | 23 - .../javadoc/0.8-SNAPSHOT/overview-tree.html | 581 - .../0.8-SNAPSHOT/package-search-index.js | 1 - .../0.8-SNAPSHOT/package-search-index.zip | Bin 489 -> 0 bytes website/static/javadoc/0.8-SNAPSHOT/packages | 185 - .../javadoc/0.8-SNAPSHOT/resources/glass.png | Bin 499 -> 0 bytes .../javadoc/0.8-SNAPSHOT/resources/x.png | Bin 394 -> 0 bytes website/static/javadoc/0.8-SNAPSHOT/script.js | 149 - website/static/javadoc/0.8-SNAPSHOT/search.js | 326 - .../javadoc/0.8-SNAPSHOT/serialized-form.html | 899 -- .../javadoc/0.8-SNAPSHOT/stylesheet.css | 910 -- .../javadoc/0.8-SNAPSHOT/type-search-index.js | 1 - .../0.8-SNAPSHOT/type-search-index.zip | Bin 2356 -> 0 bytes website/static/javadoc/index.html | 3 - 687 files changed, 182 insertions(+), 207230 deletions(-) delete mode 100755 tools/refresh_javadoc.sh create mode 100755 website/build_javadocs.sh delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/allclasses-index.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/allclasses.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/allpackages-index.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/argfile delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/constant-values.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/deprecated-list.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/element-list delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-1.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-10.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-11.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-12.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-13.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-14.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-15.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-16.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-17.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-18.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-19.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-2.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-20.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-21.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-22.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-23.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-3.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-4.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-5.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-6.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-7.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-8.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index-files/index-9.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/index.html delete mode 100755 website/static/javadoc/0.8-SNAPSHOT/javadoc.sh delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/jquery-ui.overrides.css delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/jquery/external/jquery/jquery.js delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/jquery/jquery-3.7.1.min.js delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/jquery/jquery-ui.min.css delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/jquery/jquery-ui.min.js delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/jquery/jszip-utils/dist/jszip-utils-ie.js delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/jquery/jszip-utils/dist/jszip-utils-ie.min.js delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/jquery/jszip-utils/dist/jszip-utils.js delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/jquery/jszip-utils/dist/jszip-utils.min.js delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/jquery/jszip/dist/jszip.js delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/jquery/jszip/dist/jszip.min.js delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/legal/ADDITIONAL_LICENSE_INFO delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/legal/ASSEMBLY_EXCEPTION delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/legal/LICENSE delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/legal/jquery.md delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/legal/jqueryUI.md delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/legal/jszip.md delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/legal/pako.md delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/member-search-index.js delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/member-search-index.zip delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/options delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/Connection.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/ConnectionFactory.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/FlussConnection.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/Admin.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/CreateAclsResult.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.FilterResult.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.FilterResults.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/FlussAdmin.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/ListOffsetsResult.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.EarliestSpec.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.LatestSpec.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.TimestampSpec.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/Admin.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/CreateAclsResult.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.FilterResult.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.FilterResults.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/FlussAdmin.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/ListOffsetsResult.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.EarliestSpec.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.LatestSpec.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.TimestampSpec.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/Connection.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/ConnectionFactory.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/FlussConnection.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/AbstractLookupBatch.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/AbstractLookupQuery.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/Lookup.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupBatch.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupClient.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupQuery.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupResult.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupType.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/Lookuper.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookup.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookupBatch.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookupQuery.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/TableLookup.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/AbstractLookupBatch.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/AbstractLookupQuery.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/Lookup.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupBatch.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupClient.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupQuery.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupResult.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupType.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/Lookuper.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookup.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookupBatch.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookupQuery.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/TableLookup.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/KvSnapshotMetadata.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/KvSnapshots.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/LakeSnapshot.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/MetadataUpdater.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/KvSnapshotMetadata.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/KvSnapshots.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/LakeSnapshot.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/MetadataUpdater.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/ScannerMetricGroup.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/WriterMetricGroup.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/class-use/ScannerMetricGroup.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/class-use/WriterMetricGroup.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/FlussTable.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/Table.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/class-use/FlussTable.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/class-use/Table.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/PartitionGetter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/class-use/PartitionGetter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/RemoteFileDownloader.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/Scan.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/ScanRecord.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/TableScan.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/BatchScanUtils.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/BatchScanner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/KvSnapshotBatchScanner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/LimitBatchScanner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/BatchScanUtils.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/BatchScanner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/KvSnapshotBatchScanner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/LimitBatchScanner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/RemoteFileDownloader.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/Scan.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/ScanRecord.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/TableScan.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetchBuffer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetchCollector.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetcher.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScan.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScanner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScannerImpl.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScannerStatus.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/RemoteLogDownloadFuture.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/RemoteLogDownloader.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/ScanRecords.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetchBuffer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetchCollector.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetcher.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScan.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScanner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScannerImpl.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScannerStatus.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/RemoteLogDownloadFuture.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/RemoteLogDownloader.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/ScanRecords.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AbstractTableWriter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/Append.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AppendResult.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AppendWriter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/DeleteResult.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableAppend.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableUpsert.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableWriter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/Upsert.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertResult.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertWrite.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertWriter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AbstractTableWriter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/Append.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AppendResult.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AppendWriter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/DeleteResult.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableAppend.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableUpsert.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableWriter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/Upsert.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertResult.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertWrite.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertWriter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/DefaultSecurityTokenManager.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/DefaultSecurityTokenProvider.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/SecurityTokenManager.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/SecurityTokenProvider.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/DefaultSecurityTokenManager.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/DefaultSecurityTokenProvider.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/SecurityTokenManager.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/SecurityTokenProvider.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/ClientRpcMessageUtils.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/ClientUtils.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/MetadataUtils.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/ClientRpcMessageUtils.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/ClientUtils.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/MetadataUtils.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/ArrowLogWriteBatch.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/DynamicPartitionCreator.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/DynamicWriteBatchSizeEstimator.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/HashBucketAssigner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceBucketEntry.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceBucketMap.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceManager.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IndexedLogWriteBatch.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/KvWriteBatch.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/ReadyWriteBatch.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.ReadyCheckResult.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.RecordAppendResult.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RoundRobinBucketAssigner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/Sender.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/StickyBucketAssigner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteBatch.RequestFuture.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteBatch.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteCallback.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteFormat.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteRecord.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriterClient.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/ArrowLogWriteBatch.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/DynamicPartitionCreator.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/DynamicWriteBatchSizeEstimator.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/HashBucketAssigner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceBucketEntry.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceBucketMap.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceManager.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IndexedLogWriteBatch.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/KvWriteBatch.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/ReadyWriteBatch.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.ReadyCheckResult.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.RecordAppendResult.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RoundRobinBucketAssigner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/Sender.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/StickyBucketAssigner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteBatch.RequestFuture.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteBatch.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteCallback.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteFormat.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteRecord.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriterClient.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/FlinkConnectorOptions.ScanStartupMode.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/FlinkConnectorOptions.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/CatalogTableAdapter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/SingleThreadFetcherManagerAdapter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/SingleThreadMultiplexSourceReaderBaseAdapter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/CatalogTableAdapter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/SingleThreadFetcherManagerAdapter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/SingleThreadMultiplexSourceReaderBaseAdapter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalog.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalogFactory.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalogOptions.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkTableFactory.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalog.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalogFactory.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalogOptions.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkTableFactory.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/class-use/FlinkConnectorOptions.ScanStartupMode.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/class-use/FlinkConnectorOptions.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeCatalog.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeRecordRecordEmitter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitGenerator.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitReaderGenerator.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitSerializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitStateInitializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeTableFactory.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeCatalog.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeRecordRecordEmitter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitGenerator.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitReaderGenerator.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitSerializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitStateInitializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeTableFactory.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/IndexedLakeSplitRecordIterator.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/KeyValueRow.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/LakeSnapshotAndLogSplitScanner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/LakeSnapshotScanner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/SeekableLakeSnapshotSplitScanner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/IndexedLakeSplitRecordIterator.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/KeyValueRow.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/LakeSnapshotAndLogSplitScanner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/LakeSnapshotScanner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/SeekableLakeSnapshotSplitScanner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/LakeSnapshotAndFlussLogSplit.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/LakeSnapshotSplit.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/class-use/LakeSnapshotAndFlussLogSplit.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/class-use/LakeSnapshotSplit.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/LakeSnapshotAndFlussLogSplitState.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/LakeSnapshotSplitState.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/class-use/LakeSnapshotAndFlussLogSplitState.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/class-use/LakeSnapshotSplitState.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkCounter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkGauge.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkHistogram.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkMeter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkMetricRegistry.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkCounter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkGauge.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkHistogram.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkMeter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkMetricRegistry.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/AbstractAclProcedure.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/AddAclProcedure.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/DropAclProcedure.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ListAclProcedure.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ProcedureBase.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ProcedureManager.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/AbstractAclProcedure.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/AddAclProcedure.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/DropAclProcedure.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ListAclProcedure.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ProcedureBase.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ProcedureManager.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/FlinkAsFlussRow.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/OperationType.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/RowWithOp.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/FlinkAsFlussRow.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/OperationType.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/RowWithOp.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/ChannelComputer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkRowDataChannelComputer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkStreamPartitioner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkTableSink.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlussSink.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlussSinkBuilder.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/ChannelComputer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkRowDataChannelComputer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkStreamPartitioner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkTableSink.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlussSink.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlussSinkBuilder.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/FlussSerializationSchema.InitializationContext.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/FlussSerializationSchema.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/RowDataSerializationSchema.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/SerializerInitContextImpl.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/FlussSerializationSchema.InitializationContext.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/FlussSerializationSchema.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/RowDataSerializationSchema.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/SerializerInitContextImpl.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/AppendSinkWriter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/FlinkSinkWriter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/UpsertSinkWriter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/AppendSinkWriter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/FlinkSinkWriter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/UpsertSinkWriter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlinkSource.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlinkTableSource.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlussSource.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlussSourceBuilder.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlinkSource.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlinkTableSource.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlussSource.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlussSourceBuilder.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/DeserializerInitContextImpl.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/FlussDeserializationSchema.InitializationContext.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/FlussDeserializationSchema.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/JsonStringDeserializationSchema.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/RowDataDeserializationSchema.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/DeserializerInitContextImpl.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/FlussDeserializationSchema.InitializationContext.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/FlussDeserializationSchema.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/JsonStringDeserializationSchema.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/RowDataDeserializationSchema.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/FlinkRecordEmitter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/class-use/FlinkRecordEmitter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/FlinkSourceEnumerator.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/class-use/FlinkSourceEnumerator.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/BucketOffsetsRetrieverImpl.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/LatestOffsetsInitializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/NoStoppingOffsetsInitializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/OffsetsInitializer.BucketOffsetsRetriever.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/OffsetsInitializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/SnapshotOffsetsInitializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/TimestampOffsetsInitializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/BucketOffsetsRetrieverImpl.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/LatestOffsetsInitializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/NoStoppingOffsetsInitializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/OffsetsInitializer.BucketOffsetsRetriever.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/OffsetsInitializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/SnapshotOffsetsInitializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/TimestampOffsetsInitializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/PartitionBucketsUnsubscribedEvent.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/PartitionsRemovedEvent.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/class-use/PartitionBucketsUnsubscribedEvent.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/class-use/PartitionsRemovedEvent.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/FlinkAsyncLookupFunction.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/FlinkLookupFunction.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/LookupNormalizer.RemainingFilter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/LookupNormalizer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/FlinkAsyncLookupFunction.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/FlinkLookupFunction.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/LookupNormalizer.RemainingFilter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/LookupNormalizer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/FlinkSourceReaderMetrics.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/class-use/FlinkSourceReaderMetrics.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/BoundedSplitReader.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkRecordsWithSplitIds.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkSourceReader.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkSourceSplitReader.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/MutableRecordAndPos.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/RecordAndPos.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/SplitScanner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/BoundedSplitReader.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkRecordsWithSplitIds.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkSourceReader.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkSourceSplitReader.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/MutableRecordAndPos.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/RecordAndPos.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/SplitScanner.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/FlinkSourceFetcherManager.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/class-use/FlinkSourceFetcherManager.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/HybridSnapshotLogSplit.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/HybridSnapshotLogSplitState.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/LogSplit.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/LogSplitState.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SnapshotSplit.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitBase.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitSerializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitState.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/HybridSnapshotLogSplit.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/HybridSnapshotLogSplitState.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/LogSplit.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/LogSplitState.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SnapshotSplit.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitBase.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitSerializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitState.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/FlussSourceEnumeratorStateSerializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/SourceEnumeratorState.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/class-use/FlussSourceEnumeratorStateSerializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/class-use/SourceEnumeratorState.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/LakeTieringJobBuilder.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/class-use/LakeTieringJobBuilder.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/CommittableMessage.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/CommittableMessageTypeInfo.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/FlussTableLakeSnapshotCommitter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitOperator.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitOperatorFactory.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitterInitContext.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/CommittableMessage.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/CommittableMessageTypeInfo.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/FlussTableLakeSnapshotCommitter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitOperator.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitOperatorFactory.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitterInitContext.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/FailedTieringEvent.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/FinishedTieringEvent.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/TieringRestoreEvent.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/FailedTieringEvent.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/FinishedTieringEvent.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/TieringRestoreEvent.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResult.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultEmitter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultSerializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultTypeInfo.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSource.Builder.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSource.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSourceOptions.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSourceReader.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSplitReader.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringWriterInitContext.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResult.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultEmitter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultSerializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultTypeInfo.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSource.Builder.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSource.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSourceOptions.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSourceReader.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSplitReader.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringWriterInitContext.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/TieringSourceEnumerator.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/class-use/TieringSourceEnumerator.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringLogSplit.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSnapshotSplit.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplit.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplitGenerator.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplitSerializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringLogSplit.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSnapshotSplit.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplit.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplitGenerator.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplitSerializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSourceEnumeratorState.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSourceEnumeratorStateSerializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSplitState.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSourceEnumeratorState.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSourceEnumeratorStateSerializer.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSplitState.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/CatalogExceptionUtils.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/CatalogPropertiesUtils.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/DataLakeUtils.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConnectorOptionsUtils.StartupOptions.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConnectorOptionsUtils.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConversions.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkRowToFlussRowConverter.FlussSerializationConverter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkRowToFlussRowConverter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkUtils.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToFlinkRowConverter.FlussDeserializationConverter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToFlinkRowConverter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToJsonConverters.FlussRowToJsonConverter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToJsonConverters.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/LakeSourceUtils.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PojoToRowConverter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.FieldEqual.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.ValueConversion.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/TimestampFormat.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/CatalogExceptionUtils.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/CatalogPropertiesUtils.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/DataLakeUtils.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConnectorOptionsUtils.StartupOptions.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConnectorOptionsUtils.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConversions.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkRowToFlussRowConverter.FlussSerializationConverter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkRowToFlussRowConverter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkUtils.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToFlinkRowConverter.FlussDeserializationConverter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToFlinkRowConverter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToJsonConverters.FlussRowToJsonConverter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToJsonConverters.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/LakeSourceUtils.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PojoToRowConverter.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.FieldEqual.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.ValueConversion.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/TimestampFormat.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-use.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/overview-summary.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/overview-tree.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/package-search-index.js delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/package-search-index.zip delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/packages delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/resources/glass.png delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/resources/x.png delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/script.js delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/search.js delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/serialized-form.html delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/stylesheet.css delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/type-search-index.js delete mode 100644 website/static/javadoc/0.8-SNAPSHOT/type-search-index.zip delete mode 100644 website/static/javadoc/index.html diff --git a/tools/refresh_javadoc.sh b/tools/refresh_javadoc.sh deleted file mode 100755 index 3610a8c78b..0000000000 --- a/tools/refresh_javadoc.sh +++ /dev/null @@ -1,62 +0,0 @@ -#!/bin/bash - -# Licensed to the Apache Software Foundation (ASF) under one or more -# contributor license agreements. See the NOTICE file distributed with -# this work for additional information regarding copyright ownership. -# The ASF licenses this file to You under the Apache License, Version 2.0 -# (the "License"); you may not use this file except in compliance with -# the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -set -e - -# Parse arguments -while [[ $# -gt 0 ]]; do - case $1 in - -v|--version) VERSION_OVERRIDE="$2"; shift 2 ;; - -h|--help) echo "Usage: $0 [-v VERSION] [-h|--help]"; exit 0 ;; - *) echo "Unknown option: $1"; exit 1 ;; - esac -done - -# Find project root and get version -cd "$(dirname "$0")" -while [[ ! -f "pom.xml" && "$(pwd)" != "/" ]]; do cd ..; done -[[ ! -f "pom.xml" ]] && { echo "Error: pom.xml not found"; exit 1; } - -VERSION=${VERSION_OVERRIDE:-$(mvn help:evaluate -Dexpression=project.version -q -DforceStdout)} -SHORT_VERSION=$(echo "$VERSION" | cut -d. -f1-2) -OUTPUT_DIR="website/static/javadoc/$SHORT_VERSION" - -echo "Generating Javadoc $VERSION" - -# Setup output -rm -rf "$OUTPUT_DIR" -mkdir -p "$OUTPUT_DIR" - -# Generate javadoc -export MAVEN_OPTS="--add-exports=java.rmi/sun.rmi.registry=ALL-UNNAMED --add-opens=java.base/java.lang=ALL-UNNAMED" - -if mvn javadoc:aggregate -Pjavadoc-aggregate -Djavadoc.aggregate=true -q 2>/dev/null; then - [[ -d "target/site/apidocs" ]] && cp -r target/site/apidocs/* "$OUTPUT_DIR/" || { echo "No javadoc generated"; exit 1; } -else - echo "Javadoc generation failed"; exit 1 -fi - -# Create redirect -mkdir -p "website/static/javadoc" -cat > "website/static/javadoc/index.html" << EOF - - -Apache Fluss ${SHORT_VERSION} API -EOF - -echo "Javadoc generated: $OUTPUT_DIR" -echo "URL: http://localhost:3000/javadoc/$SHORT_VERSION/" \ No newline at end of file diff --git a/website/build_javadocs.sh b/website/build_javadocs.sh new file mode 100755 index 0000000000..59ec37ce64 --- /dev/null +++ b/website/build_javadocs.sh @@ -0,0 +1,181 @@ +#!/usr/bin/env bash +# +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# + +set -e + +# Find project root directory +SCRIPT_PATH=$(cd "$(dirname "$0")" && pwd) +cd "$SCRIPT_PATH" + +# Navigate to project root by looking for pom.xml +while [[ ! -f "pom.xml" && "$(pwd)" != "/" ]]; do + cd .. +done + +# Validate we found the project root +if [[ ! -f "pom.xml" ]]; then + echo "Error: Could not find project root with pom.xml" + exit 1 +fi + +echo "Found project root: $(pwd)" + +# Extract version from Maven project +VERSION=$(./mvnw help:evaluate -Dexpression=project.version -q -DforceStdout 2>/dev/null || mvn help:evaluate -Dexpression=project.version -q -DforceStdout) +if [[ -z "$VERSION" ]]; then + echo "Error: Could not extract project version" + exit 1 +fi + +# Get current branch name for version determination +CURRENT_BRANCH=$(git rev-parse --abbrev-ref HEAD 2>/dev/null || echo "unknown") +echo "Current branch: $CURRENT_BRANCH" + +# Determine javadoc version based on branch type +if [[ "$CURRENT_BRANCH" == "main" ]]; then + JAVADOC_VERSION="main" +elif [[ "$CURRENT_BRANCH" =~ ^release-([0-9]+\.[0-9]+)$ ]]; then + JAVADOC_VERSION="${BASH_REMATCH[1]}" +else + # For feature branches or other branches, use sanitized branch name + JAVADOC_VERSION=$(echo "$CURRENT_BRANCH" | sed 's/[^a-zA-Z0-9._-]/_/g') +fi + +# Setup output directory +OUTPUT_DIR="website/static/javadoc/$JAVADOC_VERSION" +echo "Generating Javadoc for version: $JAVADOC_VERSION" +echo "Output directory: $OUTPUT_DIR" + +# Create output directory +rm -rf "$OUTPUT_DIR" +mkdir -p "$OUTPUT_DIR" + +# Create temporary POM for targeted javadoc generation +TEMP_POM="temp-javadoc-pom.xml" +cat > "$TEMP_POM" << EOF + + + 4.0.0 + + org.apache.fluss + fluss-javadoc-temp + $VERSION + pom + + + 11 + 11 + UTF-8 + + + + fluss-client + fluss-flink/fluss-flink-common + + + + + + org.apache.maven.plugins + maven-javadoc-plugin + 3.6.3 + + Apache Fluss \${project.version} API + Apache Fluss \${project.version} API + none + false + true + + --add-opens=java.base/java.lang=ALL-UNNAMED + -J--add-exports=java.rmi/sun.rmi.registry=ALL-UNNAMED + + + + + + +EOF + +# Cleanup function for temporary files +cleanup() { + if [[ -f "$TEMP_POM" ]]; then + rm -f "$TEMP_POM" + echo "Cleaned up temporary POM file" + fi +} + +# Set trap for cleanup on script exit +trap cleanup EXIT + +# Set Maven options for Java 11+ compatibility +export MAVEN_OPTS="--add-exports=java.rmi/sun.rmi.registry=ALL-UNNAMED --add-opens=java.base/java.lang=ALL-UNNAMED" + +echo "Building and generating Javadoc..." + +# Build and generate javadoc using temporary POM +if (./mvnw -f "$TEMP_POM" clean compile javadoc:aggregate -q 2>/dev/null || mvn -f "$TEMP_POM" clean compile javadoc:aggregate -q); then + echo "Javadoc generation completed successfully" +else + echo "Error: Javadoc generation failed" + exit 1 +fi + +# Copy generated javadoc to output directory +if [[ -d "target/site/apidocs" ]]; then + cp -r target/site/apidocs/* "$OUTPUT_DIR/" + echo "Javadoc copied to: $OUTPUT_DIR" +else + echo "Error: No javadoc files generated in target/site/apidocs" + exit 1 +fi + +# Create or update redirect page +JAVADOC_INDEX="website/static/javadoc/index.html" +mkdir -p "website/static/javadoc" +cat > "$JAVADOC_INDEX" << EOF + + + + + + Apache Fluss API Documentation + + +

    Redirecting to Apache Fluss ${JAVADOC_VERSION} API

    + + +EOF + +echo "Created redirect page: $JAVADOC_INDEX" + +# Verify generation success +PACKAGE_COUNT=$(find "$OUTPUT_DIR" -name "package-summary.html" 2>/dev/null | wc -l | tr -d ' ') +CLASS_COUNT=$(find "$OUTPUT_DIR" -name "*.html" -not -name "index*.html" -not -name "*-summary.html" 2>/dev/null | wc -l | tr -d ' ') + +echo "" +echo "Javadoc generation summary:" +echo " Version: $JAVADOC_VERSION" +echo " Packages: $PACKAGE_COUNT" +echo " Classes/Interfaces: $CLASS_COUNT" +echo " Location: $OUTPUT_DIR" +echo " Local URL: http://localhost:3000/javadoc/$JAVADOC_VERSION/" +echo "" +echo "Javadoc generation completed successfully!" \ No newline at end of file diff --git a/website/docusaurus.config.ts b/website/docusaurus.config.ts index 08c66381d6..b410492d11 100644 --- a/website/docusaurus.config.ts +++ b/website/docusaurus.config.ts @@ -156,6 +156,7 @@ const config: Config = { {to: '/community/welcome', label: 'Community', position: 'left'}, {to: '/roadmap', label: 'Roadmap', position: 'left'}, {to: '/downloads', label: 'Downloads', position: 'left'}, + {to: '/javadoc', label: 'Javadocs', position: 'left', target: '_blank'}, { label: 'ASF', position: 'right', items: [ {to: 'https://www.apache.org/', label: 'Foundation'}, diff --git a/website/static/javadoc/0.8-SNAPSHOT/allclasses-index.html b/website/static/javadoc/0.8-SNAPSHOT/allclasses-index.html deleted file mode 100644 index 4789b53dd7..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/allclasses-index.html +++ /dev/null @@ -1,1627 +0,0 @@ - - - - - -All Classes (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    All Classes

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/allclasses.html b/website/static/javadoc/0.8-SNAPSHOT/allclasses.html deleted file mode 100644 index 8594df6d9e..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/allclasses.html +++ /dev/null @@ -1,268 +0,0 @@ - - - - - -All Classes (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - -
    -

    All Classes

    -
    - -
    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/allpackages-index.html b/website/static/javadoc/0.8-SNAPSHOT/allpackages-index.html deleted file mode 100644 index 0a66089a38..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/allpackages-index.html +++ /dev/null @@ -1,350 +0,0 @@ - - - - - -All Packages (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    All Packages

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/argfile b/website/static/javadoc/0.8-SNAPSHOT/argfile deleted file mode 100644 index d90e6ba99d..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/argfile +++ /dev/null @@ -1,1401 +0,0 @@ -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/HistogramStatistics.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/NoOpCounter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/MetricView.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/Gauge.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/reporter/ScheduledMetricReporter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/reporter/ReporterSetup.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/reporter/MetricReporter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/reporter/MetricReporterPlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/CharacterFilter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/Metric.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/DescriptiveStatisticsHistogramStatistics.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/DescriptiveStatisticsHistogram.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/utils/MetricGroupUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/groups/AbstractMetricGroup.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/groups/GenericValueMetricGroup.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/groups/GenericKeyMetricGroup.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/groups/ProxyMetricGroup.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/groups/FrontMetricGroup.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/groups/ReporterScopedSettings.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/groups/GenericMetricGroup.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/groups/MetricGroup.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/Counter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/SimpleCounter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/Meter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/MetricType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/registry/MetricRegistry.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/registry/MetricRegistryImpl.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/registry/NOPMetricRegistry.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/ThreadSafeSimpleCounter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/MetricViewUpdater.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/Histogram.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/MetricNames.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metrics/MeterView.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/rocksdb/RocksDBHandle.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/rocksdb/RocksDBOperationUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/rocksdb/RocksIteratorWrapper.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/cluster/TabletServerInfo.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/cluster/ServerType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/cluster/Cluster.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/cluster/ServerNode.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/cluster/Endpoint.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/cluster/BucketLocation.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/PredicateBuilder.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/LessThan.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/Contains.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/CompareUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/NotIn.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/EndsWith.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/LeafPredicate.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/CompoundPredicate.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/NullFalseLeafBinaryFunction.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/IsNotNull.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/FieldRef.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/In.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/PredicateVisitor.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/UnsupportedExpression.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/Predicate.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/LessOrEqual.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/And.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/NotEqual.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/IsNull.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/PartitionPredicateVisitor.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/LeafFunction.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/GreaterOrEqual.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/FunctionVisitor.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/GreaterThan.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/StartsWith.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/Equal.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/Or.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/predicate/LeafUnaryFunction.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/TinyIntType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/DataType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/LocalZonedTimestampType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/DoubleType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/StringType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/DataTypeVisitor.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/BinaryType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/DecimalType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/RowType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/CharType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/MapType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/DataTypeDefaultVisitor.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/DataTypeFamily.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/DataTypeChecks.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/BytesType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/DataTypeRoot.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/DateType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/BigIntType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/SmallIntType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/IntType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/DataTypes.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/TimeType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/TimestampType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/FloatType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/DataField.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/BooleanType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/types/ArrayType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/FileLogProjection.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/TimestampAndOffset.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/KvRecord.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/ChangeType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/GenericRecord.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/FileLogInputStream.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/DefaultValueRecord.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/IndexedLogRecord.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/KvRecordReadContext.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/bytesview/MultiBytesView.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/bytesview/BytesView.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/bytesview/MemorySegmentBytesView.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/bytesview/FlussFileRegion.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/bytesview/FileRegionBytesView.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/bytesview/ByteBufBytesView.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/LogRecord.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/MemoryLogRecords.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/ChangeTypeVectorWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/ValueRecordReadContext.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/DefaultLogRecordBatch.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/LogRecords.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/LogRecordBatchFormat.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/KvRecordBatch.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/FileLogRecords.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/MemoryLogRecordsIndexedBuilder.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/FileChannelChunk.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/DefaultValueRecordBatch.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/BytesViewLogRecords.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/MemoryLogRecordsArrowBuilder.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/ValueRecord.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/DefaultKvRecord.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/MemorySegmentLogInputStream.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/KvRecordBatchBuilder.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/DefaultKvRecordBatch.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/LogInputStream.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/LogRecordReadContext.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/LogRecordBatchIterator.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/send/FileRegionSend.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/send/MultiSend.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/send/ByteBufSend.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/send/WritableOutput.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/send/Send.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/send/SendWritableOutput.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/send/ByteBufWritableOutput.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/ChangeTypeVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/ValueRecordBatch.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/record/LogRecordBatch.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/OutputView.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/MemorySegmentPool.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/MemorySegmentInputView.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/MemorySegmentOutputView.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/AbstractPagedOutputView.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/MemorySegmentWritable.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/LazyMemorySegmentPool.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/MemoryUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/ManagedPagedOutputView.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/MemorySegment.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/UnmanagedPagedOutputView.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/PreAllocatedPagedOutputView.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/memory/InputView.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/MemorySize.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/StructuredOptionsSplitter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/GlobalConfiguration.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/ConfigBuilder.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/Configuration.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/AutoPartitionTimeUnit.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/TableConfig.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/FallbackKey.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/ReadableConfig.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/ConfigOption.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/ConfigurationUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/FlussConfigUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/ConfigOptions.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/config/Password.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/AuthenticationFactory.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/plain/PlainSaslServer.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/plain/PlainLoginModule.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/plain/PlainAuthenticateCallback.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/plain/PlainSaslServerProvider.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/plain/PlainServerCallbackHandler.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/authenticator/SaslServerAuthenticator.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/authenticator/SaslClientAuthenticator.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/authenticator/SaslAuthenticationPlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/jaas/DefaultLogin.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/jaas/SaslClientCallbackHandler.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/jaas/JaasConfig.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/jaas/JaasContext.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/jaas/AuthenticateCallbackHandler.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/jaas/Login.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/jaas/LoginManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/sasl/jaas/SaslServerFactory.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/AuthenticationPlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/ClientAuthenticationPlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/ServerAuthenticator.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/PlainTextAuthenticationPlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/ClientAuthenticator.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/auth/ServerAuthenticationPlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/acl/PermissionType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/acl/OperationType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/acl/Resource.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/acl/ResourceFilter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/acl/AclBinding.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/acl/AccessControlEntry.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/acl/AccessControlEntryFilter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/acl/FlussPrincipal.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/acl/ResourceType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/security/acl/AclBindingFilter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/plugin/PluginUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/plugin/PluginLoader.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/plugin/PluginManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/plugin/Plugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/plugin/PluginDescriptor.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/plugin/DefaultPluginManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/plugin/PluginConfig.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/plugin/DirectoryBasedPluginFinder.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/plugin/PluginFinder.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/annotation/Internal.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/annotation/VisibleForTesting.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/annotation/PublicEvolving.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/annotation/PublicStable.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/annotation/PublicUnstable.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/Preconditions.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/CloseableRegistry.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/AbstractAutoCloseableRegistry.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/IOUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/StringUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/PartitionUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/ExecutorUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/ByteBufferReadableChannel.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/types/Tuple.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/types/Either.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/types/Tuple2.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/OperatingSystem.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/InstantiationUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/TypeUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/UnionIterator.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/MapUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/FileUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/SerializedValue.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/ArrayUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/CollectionUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/crc/Checksums.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/crc/Crc32C.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/crc/PureJavaCrc32C.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/crc/Java.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/MathUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/VarLengthUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/NetUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/CloseableIterator.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/function/FunctionUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/function/SupplierWithException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/function/QuadConsumer.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/function/ThrowingConsumer.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/function/RunnableWithException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/function/ThrowingRunnable.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/function/CheckedSupplier.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/function/BiFunction.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/function/FunctionWithException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/ExponentialBackoff.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/FatalExitExceptionHandler.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/FlussPaths.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/ArrowUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/PagedMemorySegmentWritableChannel.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/DateTimeUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/PropertiesUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/SerializableObject.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/json/TablePartitionJsonSerde.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/json/BucketOffsetJsonSerde.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/json/JsonSerdeUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/json/JsonDeserializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/json/DatabaseDescriptorJsonSerde.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/json/JsonSerializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/json/ColumnJsonSerde.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/json/SchemaJsonSerde.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/json/TableDescriptorJsonSerde.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/json/DataTypeJsonSerde.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/UnsafeUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/DecimalUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/FileLock.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/TimeUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/WrappingProxy.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/BytesUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/ExceptionUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/Projection.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/AutoCloseableAsync.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/concurrent/Scheduler.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/concurrent/FlussScheduler.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/concurrent/ExecutorThreadFactory.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/concurrent/ShutdownableThread.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/concurrent/FutureUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/concurrent/LockUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/concurrent/ThreadUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/concurrent/Executors.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/BinaryStringUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/TemporaryClassLoaderContext.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/log/FairBucketStatusMap.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/log/ByteBufferUnmapper.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/AbstractIterator.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/MurmurHashUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/AutoPartitionStrategy.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/WrappingProxyUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/ProtoCodecUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/MemorySegmentWritableChannel.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/EncodingUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/clock/SystemClock.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/clock/ManualClock.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/clock/Clock.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/utils/CopyOnWriteMap.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/classloading/ComponentClassLoader.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/BinarySegmentUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/DoubleColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/FloatColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/DecimalColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/ColumnarRow.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/ColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/BytesColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/TimestampLtzColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/BooleanColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/TimestampNtzColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/VectorizedColumnBatch.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/ShortColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/ByteColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/IntColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/columnar/LongColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/InternalRow.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/compacted/CompactedRowDeserializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/compacted/CompactedRowReader.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/compacted/CompactedRowWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/compacted/CompactedRow.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/compacted/CompactedKeyWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowDoubleWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowFieldWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowTimestampLtzWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowBinaryWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowSmallIntWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowTimeWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowIntWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowDateWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowVarCharWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowBigIntWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowBooleanWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowDecimalWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowFloatWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowTimestampNtzWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowVarBinaryWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/writers/ArrowTinyIntWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowTinyIntColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowDateColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowBinaryColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowTimeColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowVarCharColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowTimestampNtzColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowFloatColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowDoubleColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowVarBinaryColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowIntColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowDecimalColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowSmallIntColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowBigIntColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowTimestampLtzColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/vectors/ArrowBooleanColumnVector.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/ArrowWriterProvider.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/ArrowWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/ArrowWriterPool.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/arrow/ArrowReader.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/GenericRow.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/Decimal.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/MemoryAwareGetters.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/BinaryString.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/encode/IndexedRowEncoder.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/encode/ValueEncoder.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/encode/paimon/PaimonBinaryRowWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/encode/paimon/PaimonKeyEncoder.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/encode/KeyEncoder.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/encode/RowEncoder.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/encode/iceberg/IcebergBinaryRowWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/encode/iceberg/IcebergKeyEncoder.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/encode/ValueDecoder.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/encode/CompactedKeyEncoder.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/encode/CompactedRowEncoder.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/indexed/IndexedRowReader.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/indexed/IndexedRowWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/indexed/IndexedRow.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/decode/CompactedRowDecoder.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/decode/IndexedRowDecoder.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/decode/RowDecoder.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/ProjectedRow.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/BinaryRow.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/TimestampNtz.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/TimestampLtz.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/NullAwareGetters.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/row/BinarySection.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/lakestorage/LakeCatalog.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/lakestorage/LakeStoragePlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/lakestorage/PluginLakeStorageWrapper.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/lakestorage/LakeStorage.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/lakestorage/LakeStoragePluginSetUp.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/serializer/SimpleVersionedSerializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/source/Planner.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/source/SortedRecordReader.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/source/LakeSplit.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/source/LakeSource.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/source/RecordReader.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/writer/LakeWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/writer/WriterInitContext.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/writer/LakeTieringFactory.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/writer/SupportsRecordBatchWrite.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/committer/LakeCommitter.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/committer/CommitterInitContext.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/committer/BucketOffset.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/committer/CommittedLakeSnapshot.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/batch/ArrowRecordBatch.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/lake/batch/RecordBatch.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/compression/FlussLZ4BlockOutputStream.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/compression/ZstdArrowCompressionCodec.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/compression/Lz4ArrowCompressionCodec.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/compression/ArrowCompressionInfo.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/compression/FlussLZ4BlockInputStream.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/compression/ArrowCompressionFactory.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/compression/ArrowCompressionRatioEstimator.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/compression/ArrowCompressionType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/bucketing/PaimonBucketingFunction.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/bucketing/FlussBucketingFunction.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/bucketing/IcebergBucketingFunction.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/bucketing/BucketingFunction.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/rpc/messages/ErrorMessage.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/rpc/messages/ApiMessage.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/token/SecurityTokenReceiver.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/token/ObtainedSecurityToken.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/token/Credentials.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/token/CredentialsJsonSerde.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/FileStatus.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/WrappingProxyCloseable.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/FSDataInputStreamWrapper.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/FileSystem.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/UnsupportedFileSystemSchemeException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/FileSystemSafetyNet.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/FSDataInputStream.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/ClosingFSDataInputStream.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/FsPathAndFileName.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/utils/FileDownloadUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/utils/FileDownloadSpec.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/SafetyNetCloseableRegistry.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/local/LocalFileSystem.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/local/LocalSecurityTokenReceiver.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/local/LocalFileStatus.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/local/LocalFileSystemPlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/local/LocalDataInputStream.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/local/LocalDataOutputStream.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/ClosingFSDataOutputStream.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/FSDataOutputStreamWrapper.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/FsPath.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/PluginFileSystemWrapper.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/FileSystemPlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/FSDataOutputStream.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/fs/SafetyNetWrapperFileSystem.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/ApiException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/FlussException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidMetadataException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/BufferExhaustedException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/DatabaseNotExistException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/TableNotExistException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/DisconnectException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidColumnProjectionException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidServerRackInfoException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/IllegalConfigurationException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/LeaderNotAvailableException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/FetchException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/PartitionAlreadyExistsException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/SecurityTokenException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/FlussRuntimeException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/RecordTooLargeException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/DatabaseAlreadyExistException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/TooManyPartitionsException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/UnknownServerException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/KvSnapshotNotExistException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidTargetColumnException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidTableException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/LakeStorageNotConfiguredException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/StorageException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/OutOfOrderSequenceException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidRecordException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/KvStorageException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/RemoteResourceNotFoundException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/IndexOffsetOverflowException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/CorruptSnapshotException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/LakeTableSnapshotNotExistException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/LakeTableAlreadyExistException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/NotLeaderOrFollowerException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/StaleMetadataException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/TooManyBucketsException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidDatabaseException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/PartitionNotExistException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/CorruptRecordException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/UnknownTableOrBucketException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/TableAlreadyExistException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidPartitionException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/RetriableAuthenticationException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidConfigException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidUpdateVersionException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/FencedLeaderEpochException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/AuthorizationException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/CorruptMessageException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/TableNotPartitionedException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidTimestampException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidCoordinatorException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidRequiredAcksException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/UnknownWriterIdException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/LogOffsetOutOfRangeException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/DatabaseNotEmptyException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/TimeoutException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/LogSegmentOffsetOverflowException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/WakeupException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidBucketsException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/AuthenticationException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/RemoteStorageException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/NonPrimaryKeyTableException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidOffsetException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/OperationNotAttemptedException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/LogStorageException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/UnsupportedVersionException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/InvalidReplicationFactorException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/NetworkException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/TraversableOnceException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/NotEnoughReplicasException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/DuplicateSequenceException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/LogRecordAppendFailedException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/NotEnoughReplicasAfterAppendException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/RetriableException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/SecurityDisabledException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/FencedTieringEpochException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/exception/SchemaNotExistException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/ValidationException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/PartitionInfo.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/TableException.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/TablePath.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/ResolvedPartitionSpec.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/TableBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/DataLakeFormat.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/MergeEngineType.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/Schema.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/TableBucketReplica.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/TablePartition.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/LogFormat.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/SchemaInfo.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/DatabaseDescriptor.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/TableInfo.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/PartitionSpec.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/DatabaseInfo.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/KvFormat.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/TableDescriptor.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/metadata/PhysicalTablePath.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/remote/RemoteLogFetchInfo.java' -'/Users/mehulbatra/Documents/fluss/fluss-common/src/main/java/org/apache/fluss/remote/RemoteLogSegment.java' -'/Users/mehulbatra/Documents/fluss/fluss-metrics/fluss-metrics-prometheus/src/main/java/org/apache/fluss/metrics/prometheus/PrometheusReporter.java' -'/Users/mehulbatra/Documents/fluss/fluss-metrics/fluss-metrics-prometheus/src/main/java/org/apache/fluss/metrics/prometheus/PrometheusReporterPlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/ReadyWriteBatch.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/IdempotenceBucketEntry.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/WriteRecord.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/BucketAssigner.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/IdempotenceManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/RoundRobinBucketAssigner.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/StickyBucketAssigner.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/ArrowLogWriteBatch.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/IdempotenceBucketMap.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/IncompleteBatches.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/HashBucketAssigner.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/WriteBatch.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/Sender.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/DynamicWriteBatchSizeEstimator.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/StaticBucketAssigner.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/RecordAccumulator.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/KvWriteBatch.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/DynamicPartitionCreator.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/WriterClient.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/WriteCallback.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/IndexedLogWriteBatch.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/WriteFormat.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/write/DynamicBucketAssigner.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/token/SecurityTokenProvider.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/token/DefaultSecurityTokenManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/token/DefaultSecurityTokenProvider.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/token/SecurityTokenManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/token/SecurityTokenReceiverRepository.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/metrics/WriterMetricGroup.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/metrics/ScannerMetricGroup.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/FlussConnection.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/admin/Admin.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/admin/FlussAdmin.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/admin/ListOffsetsResult.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/admin/OffsetSpec.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/admin/DropAclsResult.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/admin/CreateAclsResult.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/PrefixLookupBatch.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/LookupType.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/LookupSender.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/AbstractLookupQuery.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/Lookup.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/Lookuper.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/LookupQueue.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/LookupBatch.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/PrefixLookup.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/LookupResult.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/LookupQuery.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/PrimaryKeyLookuper.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/LookupClient.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/PrefixLookupQuery.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/TableLookup.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/PrefixKeyLookuper.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/lookup/AbstractLookupBatch.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/utils/ClientRpcMessageUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/utils/ClientUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/utils/MetadataUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/getter/PartitionGetter.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/FlussTable.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/Table.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/AppendResult.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/AbstractTableWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/UpsertWriterImpl.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/TableUpsert.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/Append.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/UpsertWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/TableAppend.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/TableWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/AppendWriterImpl.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/Upsert.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/AppendWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/DeleteResult.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/UpsertWrite.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/writer/UpsertResult.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/Scan.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/RemoteFileDownloader.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/batch/KvSnapshotBatchScanner.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/batch/SnapshotFilesReader.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/batch/BatchScanUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/batch/LimitBatchScanner.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/batch/BatchScanner.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/CompletedPendingFetch.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/LogFetcher.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/RemoteCompletedFetch.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/RemoteLogDownloader.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/DefaultCompletedFetch.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/PendingFetch.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/LogFetchCollector.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/LogScannerImpl.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/LogScanner.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/CompletedFetch.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/LogScan.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/LogFetchBuffer.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/RemotePendingFetch.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/LogScannerStatus.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/ScanRecords.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/RemoteLogDownloadFuture.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/log/BucketScanStatus.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/TableScan.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/table/scanner/ScanRecord.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/Connection.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/ConnectionFactory.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/metadata/KvSnapshots.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/metadata/KvSnapshotMetadata.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/metadata/LakeSnapshot.java' -'/Users/mehulbatra/Documents/fluss/fluss-client/src/main/java/org/apache/fluss/client/metadata/MetadataUpdater.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/metrics/ConnectionMetricGroup.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/metrics/ClientMetricGroup.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/RpcGateway.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/RpcClient.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/util/CommonRpcMessageUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/GatewayClientProxy.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/entity/ProduceLogResultForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/entity/LookupResultForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/entity/ListOffsetsResultForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/entity/PutKvResultForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/entity/LimitScanResultForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/entity/FetchLogResultForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/entity/ResultForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/entity/PrefixLookupResultForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/entity/WriteResultForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/RpcGatewayService.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/protocol/ResponseType.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/protocol/MessageCodec.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/protocol/ApiMethod.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/protocol/RPC.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/protocol/ApiKeys.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/protocol/ApiError.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/protocol/NetworkProtocolPlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/protocol/Errors.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/protocol/ApiManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/protocol/RequestType.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/NettyLogger.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/FlussRequestHandler.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/RequestProcessor.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/FlussRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/RequestsMetrics.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/RequestHandler.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/ShutdownRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/RequestProcessorPool.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/NettyServerHandler.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/RpcRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/RequestChannel.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/ServerChannelInitializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/FlussProtocolPlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/NettyServer.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/server/Session.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/NettyChannelInitializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/NettyMetrics.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/NettyUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/client/ClientChannelInitializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/client/NettyClient.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/client/NettyClientHandler.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/client/ClientHandlerCallback.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/client/ServerConnection.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/netty/client/ServerApiVersions.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/gateway/CoordinatorGateway.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/gateway/TabletServerGateway.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/gateway/AdminReadOnlyGateway.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/gateway/AdminGateway.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/src/main/java/org/apache/fluss/rpc/RpcServer.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CreatePartitionResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/DropTableResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/DatabaseExistsRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbCommitLakeTableSnapshotRespForTable.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/DropTableRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PrefixLookupRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbStopReplicaRespForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbRemotePathAndLocalFile.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/DropAclsResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbFetchLogReqForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetTableSchemaResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/DropDatabaseRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbLakeTableOffsetForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbProduceLogRespForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/AdjustIsrResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbPartitionSpec.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ListAclsRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/StopReplicaResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/DropAclsRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbLookupReqForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbAdjustIsrRespForTable.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbPutKvReqForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbNotifyLeaderAndIsrReqForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbBucketMetadata.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetTableSchemaRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbCreateAclRespInfo.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbFetchLogReqForTable.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/NotifyRemoteLogOffsetsResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbHeartbeatRespForTable.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ListPartitionInfosRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/TableExistsRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbValueList.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PutKvResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbKvSnapshot.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbAclFilter.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ListTablesResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/NotifyKvSnapshotOffsetResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbTableBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/NotifyRemoteLogOffsetsRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CreateTableResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CreateAclsRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ListOffsetsResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/NotifyKvSnapshotOffsetRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetKvSnapshotMetadataResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/LimitScanResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetLatestKvSnapshotsRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/NotifyLeaderAndIsrResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/LookupResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbProduceLogReqForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/AdjustIsrRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetDatabaseInfoResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbLakeSnapshotForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/DropPartitionResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbLakeTableSnapshotInfo.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbPutKvRespForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/NotifyLakeTableOffsetRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/NotifyLeaderAndIsrRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/DatabaseExistsResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbRemoteLogSegment.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ApiVersionsRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CommitKvSnapshotResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbPhysicalTablePath.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbApiVersion.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ListPartitionInfosResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CreatePartitionRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbRemoteLogFetchInfo.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CommitKvSnapshotRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbPartitionInfo.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ProduceLogRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbNotifyLeaderAndIsrRespForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/MetadataResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ListOffsetsRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ListTablesRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetLatestKvSnapshotsResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbDropAclsFilterResult.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbLookupRespForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CommitLakeTableSnapshotRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbFetchLogRespForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CreateDatabaseRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CreateTableRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbKeyValue.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbLakeTieringTableInfo.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CommitRemoteLogManifestRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/AuthenticateRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbHeartbeatReqForTable.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/NotifyLakeTableOffsetResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbListOffsetsRespForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetLatestLakeSnapshotResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbAdjustIsrRespForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PrefixLookupResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetKvSnapshotMetadataRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CreateAclsResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/UpdateMetadataResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbTablePath.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/InitWriterRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ApiVersionsResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/DropPartitionRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/DropDatabaseResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/LakeTieringHeartbeatResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/FetchLogRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbAclInfo.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/LakeTieringHeartbeatRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbValue.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbNotifyLakeTableOffsetReqForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbServerNode.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CommitRemoteLogManifestResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetFileSystemSecurityTokenRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/LimitScanRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ListAclsResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ErrorResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/FetchLogResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/LookupRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetTableInfoRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbPrefixLookupRespForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbAdjustIsrReqForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/UpdateMetadataRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbTableMetadata.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetFileSystemSecurityTokenResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbDropAclsMatchingAcl.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbPrefixLookupReqForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbFetchLogRespForTable.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbPartitionMetadata.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ListDatabasesResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbStopReplicaReqForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CommitLakeTableSnapshotResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PbAdjustIsrReqForTable.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/TableExistsResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ProduceLogResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/ListDatabasesRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetLatestLakeSnapshotRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetDatabaseInfoRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/PutKvRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/CreateDatabaseResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/StopReplicaRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/InitWriterResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/MetadataRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/AuthenticateResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-rpc/target/generated-sources/protobuf/java/org/apache/fluss/rpc/messages/GetTableInfoResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-dist/src/main/java/org/apache/fluss/dist/DummyClass.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-hadoop/src/main/java/org/apache/fluss/fs/hdfs/HadoopSecurityTokenReceiver.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-hadoop/src/main/java/org/apache/fluss/fs/hdfs/HadoopDataInputStream.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-hadoop/src/main/java/org/apache/fluss/fs/hdfs/HadoopFileSystem.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-hadoop/src/main/java/org/apache/fluss/fs/hdfs/utils/HadoopUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-hadoop/src/main/java/org/apache/fluss/fs/hdfs/HadoopFileStatus.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-hadoop/src/main/java/org/apache/fluss/fs/hdfs/HadoopFsPlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-hadoop/src/main/java/org/apache/fluss/fs/hdfs/HadoopDataOutputStream.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-oss/src/main/java/org/apache/fluss/fs/oss/token/OSSSecurityTokenReceiver.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-oss/src/main/java/org/apache/fluss/fs/oss/token/OSSSecurityTokenProvider.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-oss/src/main/java/org/apache/fluss/fs/oss/token/DynamicTemporaryOssCredentialsProvider.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-oss/src/main/java/org/apache/fluss/fs/oss/OSSFileSystem.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-oss/src/main/java/org/apache/fluss/fs/oss/OSSFileSystemPlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-s3/src/main/java/org/apache/fluss/fs/s3/token/S3ADelegationTokenReceiver.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-s3/src/main/java/org/apache/fluss/fs/s3/token/S3DelegationTokenProvider.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-s3/src/main/java/org/apache/fluss/fs/s3/token/S3DelegationTokenReceiver.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-s3/src/main/java/org/apache/fluss/fs/s3/token/DynamicTemporaryAWSCredentialsProvider.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-s3/src/main/java/org/apache/fluss/fs/s3/S3FileSystemPlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-s3/src/main/java/org/apache/fluss/fs/s3/S3FileSystem.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-s3/src/main/java/org/apache/fluss/fs/s3/S3AFileSystemPlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-s3/src/main/java/com/amazonaws/services/s3/model/transform/XmlResponsesSaxParser.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-gs/src/main/java/org/apache/fluss/fs/gs/GSFileSystemPlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-gs/src/main/java/org/apache/fluss/fs/gs/GSFileSystem.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-obs/src/main/java/org/apache/fluss/fs/obs/token/DynamicTemporaryOBSCredentialsProvider.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-obs/src/main/java/org/apache/fluss/fs/obs/token/OBSSecurityTokenReceiver.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-obs/src/main/java/org/apache/fluss/fs/obs/token/OBSSecurityTokenProvider.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-obs/src/main/java/org/apache/fluss/fs/obs/OBSFileSystemPlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-filesystems/fluss-fs-obs/src/main/java/org/apache/fluss/fs/obs/OBSFileSystem.java' -'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/junit/parameterized/Parameter.java' -'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/junit/parameterized/ParameterizedTestExtension.java' -'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/junit/parameterized/Parameters.java' -'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/AllCallbackWrapper.java' -'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/FlussCompletableFutureAssert.java' -'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/CheckedThread.java' -'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/CustomExtension.java' -'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/OneShotLatch.java' -'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/TestLoggerExtension.java' -'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/ManuallyTriggeredScheduledExecutorService.java' -'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/ClassLoaderUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/ContextClassLoaderExtension.java' -'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/ScheduledTask.java' -'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/CommonTestUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/TestExecutorExtension.java' -'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/FlussAssertions.java' -'/Users/mehulbatra/Documents/fluss/fluss-test-utils/src/main/java/org/apache/fluss/testutils/common/EachCallbackWrapper.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metrics/group/TableMetricGroup.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metrics/group/TabletServerMetricGroup.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metrics/group/CoordinatorMetricGroup.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metrics/group/BucketMetricGroup.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metrics/ServerMetricUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/tablet/TabletServer.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/tablet/TabletService.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/ZkSequenceIDCounter.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/CuratorFrameworkWithUnhandledErrorListener.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/ZooKeeperClient.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/ZkAsyncResponse.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/ZkAsyncRequest.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/LakeTableSnapshot.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/CoordinatorAddress.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/BucketSnapshotJsonSerde.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/PartitionAssignment.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/DatabaseRegistrationJsonSerde.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/DatabaseRegistration.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/TabletServerRegistration.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/TableAssignmentJsonSerde.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/TabletServerRegistrationJsonSerde.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/PartitionAssignmentJsonSerde.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/RemoteLogManifestHandle.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/TableAssignment.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/RemoteLogManifestHandleJsonSerde.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/BucketAssignment.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/LakeTableSnapshotJsonSerde.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/TableRegistrationJsonSerde.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/BucketSnapshot.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/LeaderAndIsr.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/ResourceAcl.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/CoordinatorAddressJsonSerde.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/ResourceAclJsonSerde.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/TableRegistration.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/ZkData.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/data/LeaderAndIsrJsonSerde.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/zk/ZooKeeperUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/LakeTableTieringManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/CoordinatorEventProcessor.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/CoordinatorServer.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/CoordinatorContext.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/MetadataManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/TableManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/CompletedSnapshotStoreManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/statemachine/BaseState.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/statemachine/BucketState.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/statemachine/ReplicaState.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/statemachine/ReplicaLeaderElectionAlgorithms.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/statemachine/ReplicaStateMachine.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/statemachine/TableBucketStateMachine.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/CoordinatorChannelManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/CoordinatorRequestBatch.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/AutoPartitionManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/RemoteStorageCleaner.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/CoordinatorService.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/AdjustIsrReceivedEvent.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/CommitKvSnapshotEvent.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/CoordinatorEventManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/DeadTabletServerEvent.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/DropTableEvent.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/watcher/TableChangeWatcher.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/watcher/TabletServerChangeWatcher.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/FencedCoordinatorEvent.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/AccessContextEvent.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/NotifyLeaderAndIsrResponseReceivedEvent.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/NotifyKvSnapshotOffsetEvent.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/EventManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/DeleteReplicaResponseReceivedEvent.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/CreateTableEvent.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/CommitRemoteLogManifestEvent.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/ShutdownEventThreadEvent.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/CreatePartitionEvent.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/EventProcessor.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/DropPartitionEvent.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/CoordinatorEvent.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/CommitLakeTableSnapshotEvent.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/coordinator/event/NewTabletServerEvent.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/rocksdb/RocksDBKv.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/rocksdb/RocksDBWriteBatchWrapper.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/rocksdb/RocksDBKvBuilder.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/rocksdb/RocksDBResourceContainer.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/SnapshotRunner.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/KvSnapshotDownloadSpec.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/CompletedSnapshotHandleStore.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/CompletedSnapshotHandle.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/SharedKvFileRegistryKey.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/SnapshotLocation.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/ZooKeeperCompletedSnapshotHandleStore.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/CompletedSnapshot.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/KvSnapshotDataTransfer.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/PeriodicSnapshotManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/DefaultCompletedKvSnapshotCommitter.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/SnapshotsCleaner.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/KvSnapshotDataUploader.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/KvSnapshotHandle.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/AsyncSnapshotCallable.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/SnapshotFileScope.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/SnapshotResources.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/DefaultSnapshotContext.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/KvSnapshotDataDownloader.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/SnapshotResult.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/RocksIncrementalSnapshot.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/CompletedKvSnapshotCommitter.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/CompletedSnapshotStore.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/KvFileHandleAndLocalPath.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/SnapshotContext.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/CompletedSnapshotJsonSerde.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/SnapshotResultSupplier.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/KvFileHandle.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/KvTabletSnapshotTarget.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/PlaceholderKvFileHandler.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/snapshot/SharedKvFileRegistry.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/prewrite/KvPreWriteBuffer.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/KvTablet.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/wal/WalBuilder.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/wal/IndexWalBuilder.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/wal/ArrowWalBuilder.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/KvManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/partialupdate/PartialUpdater.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/partialupdate/PartialUpdaterCache.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/KvBatchWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/KvRecoverHelper.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/KvSnapshotResource.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/rowmerger/FirstRowRowMerger.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/rowmerger/VersionedRowMerger.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/rowmerger/DefaultRowMerger.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/kv/rowmerger/RowMerger.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/TabletManagerBase.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/NotifyKvSnapshotOffsetData.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/NotifyLeaderAndIsrData.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/FetchReqInfo.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/LakeBucketOffset.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/CommitRemoteLogManifestData.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/DeleteReplicaResultForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/NotifyRemoteLogOffsetsData.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/StopReplicaData.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/AdjustIsrResultForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/CommitKvSnapshotData.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/NotifyLakeTableOffsetData.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/NotifyLeaderAndIsrResultForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/BatchRegisterLeadAndIsr.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/StopReplicaResultForBucket.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/RegisterTableBucketLeadAndIsrInfo.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/CommitLakeTableSnapshotData.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/entity/LakeTieringTableInfo.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/ServerBase.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/ResourceGuard.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/ConfigurationParserUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/timer/TimerTaskList.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/timer/TimerTaskEntry.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/timer/Timer.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/timer/TimerTask.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/timer/TimingWheel.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/timer/DefaultTimer.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/ServerRpcMessageUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/ShutdownHookUtil.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/FatalErrorHandler.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/TableDescriptorValidation.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/TableAssignmentUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/RpcGatewayManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/SnapshotUtil.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/utils/LakeStorageUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/cli/ParserResultFactory.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/cli/ServerConfigurationParserFactory.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/cli/ServerConfiguration.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/cli/CommandLineOptions.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/cli/CommandLineParser.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/authorizer/AclCreateResult.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/authorizer/AuthorizationPlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/authorizer/AclDeleteResult.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/authorizer/AuthorizerLoader.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/authorizer/ZkNodeChangeNotificationWatcher.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/authorizer/DefaultAuthorizer.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/authorizer/Authorizer.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/authorizer/DefaultAuthorizationPlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/authorizer/AbstractAuthorizer.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/authorizer/Action.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/OffsetIndex.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/StorageAction.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LocalLog.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/checkpoint/OffsetCheckpointFile.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/checkpoint/CheckpointFile.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LogAppendInfo.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LogSegments.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LazyIndex.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/WriterStateEntry.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LogManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/FetchIsolation.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LogSegment.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/FetchParams.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/SnapshotFile.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LogTablet.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/IndexEntry.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/TimestampOffset.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/WriterStateManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/FetchDataInfo.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/ListOffsetsParam.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LogReadInfo.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LogLoader.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/AbstractIndex.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/IndexSearchType.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/RollParams.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/WriterAppendInfo.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/TimeIndex.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LoadedLogOffsets.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LogOffsetMetadata.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/OffsetPosition.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/LogOffsetSnapshot.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/remote/RemoteLogManifestJsonSerde.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/remote/RemoteLogIndexCache.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/remote/DefaultRemoteLogStorage.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/remote/LogSegmentFiles.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/remote/RemoteLogManifest.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/remote/RemoteLogTablet.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/remote/RemoteLogManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/remote/RemoteLogStorage.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/remote/FsRemoteLogOutputStream.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/log/remote/LogTieringTask.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/IsrState.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/fetcher/ReplicaFetcherManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/fetcher/ReplicaFetcherThread.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/fetcher/InitialFetchStatus.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/fetcher/RemoteLeaderEndpoint.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/fetcher/FetchLogContext.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/fetcher/BucketFetchStatus.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/fetcher/LeaderEndpoint.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/fetcher/DelayedItem.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/ReplicaManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/Replica.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/FollowerReplica.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/AdjustIsrManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/delay/DelayedFetchLog.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/delay/DelayedWrite.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/delay/DelayedOperationKey.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/delay/DelayedOperation.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/delay/DelayedTableBucketKey.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/replica/delay/DelayedOperationManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/SequenceIDCounter.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/exception/FlussParseException.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/exception/KvBuildingException.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/exception/EndpointNotAvailableException.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/exception/CorruptIndexException.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metadata/ServerInfo.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metadata/TableMetadata.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metadata/ServerMetadataCache.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metadata/CoordinatorMetadataCache.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metadata/ServerMetadataSnapshot.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metadata/PartitionMetadata.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metadata/BucketMetadata.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metadata/TabletServerMetadataCache.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/metadata/ClusterMetadata.java' -'/Users/mehulbatra/Documents/fluss/fluss-server/src/main/java/org/apache/fluss/server/RpcServiceBase.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/metrics/FlinkMeter.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/metrics/FlinkMetricRegistry.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/metrics/FlinkCounter.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/metrics/FlinkHistogram.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/metrics/FlinkGauge.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/catalog/FlinkCatalogFactory.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/catalog/FlinkCatalog.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/catalog/FlinkTableFactory.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/catalog/FlinkCatalogOptions.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/reader/MutableRecordAndPos.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/reader/FlinkSourceSplitReader.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/reader/fetcher/FlinkSourceFetcherManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/reader/FlinkRecordsWithSplitIds.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/reader/BoundedSplitReader.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/reader/SplitScanner.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/reader/RecordAndPos.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/reader/FlinkSourceReader.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/split/HybridSnapshotLogSplitState.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/split/SourceSplitBase.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/split/SourceSplitSerializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/split/LogSplitState.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/split/SnapshotSplit.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/split/LogSplit.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/split/SourceSplitState.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/split/HybridSnapshotLogSplit.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/metrics/FlinkSourceReaderMetrics.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/FlussSource.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/deserializer/JsonStringDeserializationSchema.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/deserializer/FlussDeserializationSchema.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/deserializer/DeserializerInitContextImpl.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/deserializer/RowDataDeserializationSchema.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/emitter/FlinkRecordEmitter.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/lookup/LookupNormalizer.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/lookup/FlinkLookupFunction.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/lookup/FlinkAsyncLookupFunction.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/state/FlussSourceEnumeratorStateSerializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/state/SourceEnumeratorState.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/FlinkSource.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/FlinkTableSource.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/event/PartitionsRemovedEvent.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/event/PartitionBucketsUnsubscribedEvent.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/enumerator/initializer/LatestOffsetsInitializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/enumerator/initializer/BucketOffsetsRetrieverImpl.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/enumerator/initializer/TimestampOffsetsInitializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/enumerator/initializer/SnapshotOffsetsInitializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/enumerator/initializer/NoStoppingOffsetsInitializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/enumerator/initializer/OffsetsInitializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/enumerator/initializer/EarliestOffsetsInitializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/enumerator/FlinkSourceEnumerator.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/source/FlussSourceBuilder.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/split/TieringSplitGenerator.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/split/TieringSplitSerializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/split/TieringSplit.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/split/TieringLogSplit.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/split/TieringSnapshotSplit.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/TieringSourceOptions.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/TableBucketWriteResultTypeInfo.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/TieringSource.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/TieringSourceReader.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/TableBucketWriteResultEmitter.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/TableBucketWriteResult.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/state/TieringSourceEnumeratorState.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/state/TieringSplitState.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/state/TieringSourceEnumeratorStateSerializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/TieringSplitReader.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/TieringWriterInitContext.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/TableBucketWriteResultSerializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/source/enumerator/TieringSourceEnumerator.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/LakeTieringJobBuilder.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/committer/FlussTableLakeSnapshot.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/committer/CommittableMessageTypeInfo.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/committer/TieringCommitOperator.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/committer/CommittableMessage.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/committer/TieringCommitterInitContext.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/committer/TieringCommitOperatorFactory.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/committer/FlussTableLakeSnapshotCommitter.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/event/FinishedTieringEvent.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/event/TieringRestoreEvent.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/tiering/event/FailedTieringEvent.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/FlussSink.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/FlinkStreamPartitioner.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/serializer/RowDataSerializationSchema.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/serializer/SerializerInitContextImpl.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/serializer/FlussSerializationSchema.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/FlinkTableSink.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/writer/AppendSinkWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/writer/FlinkSinkWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/writer/UpsertSinkWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/FlinkRowDataChannelComputer.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/ChannelComputer.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/FlussSinkBuilder.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/sink/FlinkSink.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/FlinkTypeToFlussType.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/PojoToRowConverter.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/FlinkConversions.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/TimestampFormat.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/CatalogExceptionUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/FlussRowToJsonConverters.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/CatalogPropertiesUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/FlussRowToFlinkRowConverter.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/LakeSourceUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/FlussTypeToFlinkType.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/PushdownUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/FlinkConnectorOptionsUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/DataLakeUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/FlinkUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/FlinkRowToFlussRowConverter.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/utils/TimeFormats.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/FlinkConnectorOptions.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/adapter/SingleThreadFetcherManagerAdapter.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/adapter/CatalogTableAdapter.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/adapter/SingleThreadMultiplexSourceReaderBaseAdapter.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/procedure/ProcedureManager.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/procedure/ListAclProcedure.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/procedure/AddAclProcedure.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/procedure/AbstractAclProcedure.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/procedure/ProcedureBase.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/procedure/DropAclProcedure.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/row/OperationType.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/row/RowWithOp.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/row/FlinkAsFlussRow.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/reader/KeyValueRow.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/reader/LakeSnapshotAndLogSplitScanner.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/reader/SortMergeReader.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/reader/LakeSnapshotScanner.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/reader/SeekableLakeSnapshotSplitScanner.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/reader/IndexedLakeSplitRecordIterator.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/split/LakeSnapshotAndFlussLogSplit.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/split/LakeSnapshotSplit.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/LakeTableFactory.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/LakeCatalog.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/LakeSplitStateInitializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/LakeSplitReaderGenerator.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/state/LakeSnapshotSplitState.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/state/LakeSnapshotAndFlussLogSplitState.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/LakeSplitGenerator.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/LakeSplitSerializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-common/src/main/java/org/apache/fluss/flink/lake/LakeRecordRecordEmitter.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-1.20/src/main/java/org/apache/fluss/flink/DummyClass120.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-1.19/src/main/java/org/apache/fluss/flink/adapter/CatalogTableAdapter.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-1.18/src/main/java/org/apache/fluss/flink/adapter/SingleThreadFetcherManagerAdapter.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-1.18/src/main/java/org/apache/fluss/flink/adapter/CatalogTableAdapter.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-1.18/src/main/java/org/apache/fluss/flink/adapter/SingleThreadMultiplexSourceReaderBaseAdapter.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-1.18/src/main/java/org/apache/flink/api/connector/sink2/WriterInitContext.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-2.1/src/main/java/org/apache/flink/api/connector/sink2/Sink.java' -'/Users/mehulbatra/Documents/fluss/fluss-flink/fluss-flink-tiering/src/main/java/org/apache/fluss/flink/tiering/FlussLakeTieringEntrypoint.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/PaimonLakeCatalog.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/PaimonLakeStorage.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/source/PaimonLakeSource.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/source/PaimonSplit.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/source/PaimonSplitSerializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/source/PaimonSortedRecordReader.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/source/PaimonRecordReader.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/source/FlussRowAsPaimonRow.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/source/PaimonSplitPlanner.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonWriteResult.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonCatalogProvider.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonCommittable.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/append/AppendOnlyWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonLakeWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonCommittableSerializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/mergetree/MergeTreeWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonWriteResultSerializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/RecordWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonLakeCommitter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonLakeTieringFactory.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/FlussRecordAsPaimonRow.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/utils/PaimonRowAsFlussRow.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/utils/FlussToPaimonPredicateConverter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/utils/PaimonConversions.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/PaimonLakeStoragePlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/FlussDataTypeToPaimonDataType.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/IcebergLakeCatalog.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/IcebergLakeStoragePlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/IcebergLakeStorage.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/source/IcebergSplitSerializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/source/IcebergSplit.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/source/IcebergRecordAsFlussRow.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/source/IcebergLakeSource.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/source/IcebergSplitPlanner.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/source/IcebergRecordReader.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/IcebergWriteResult.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/IcebergCommittable.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/FlussRecordAsIcebergRecord.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/IcebergCommittableSerializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/IcebergWriteResultSerializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/IcebergCatalogProvider.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/IcebergLakeCommitter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/IcebergLakeTieringFactory.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/writer/GenericRecordAppendOnlyWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/writer/DeltaTaskWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/writer/TaskWriterFactory.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/writer/AppendOnlyTaskWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/writer/GenericRecordDeltaWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/RecordWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/tiering/IcebergLakeWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/utils/IcebergConversions.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/utils/IcebergCatalogUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/maintenance/IcebergRewriteDataFiles.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/maintenance/RewriteDataFileResult.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/conf/IcebergConfiguration.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/conf/HadoopUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/conf/HadoopConfSerde.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/fluss/lake/iceberg/FlussDataTypeToIcebergDataType.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-iceberg/src/main/java/org/apache/iceberg/data/IcebergGenericReader.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/LanceLakeStorage.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/LanceLakeCatalog.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowDoubleWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowFieldWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowTimestampLtzWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowBinaryWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowSmallIntWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowTimeWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowIntWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowDateWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowVarCharWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowBigIntWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowBooleanWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowDecimalWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowFloatWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowTimestampNtzWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowVarBinaryWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/writers/ArrowTinyIntWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/tiering/LanceWriteResultSerializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/tiering/LanceWriteResult.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/tiering/LanceCommittable.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/tiering/LanceLakeTieringFactory.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/tiering/LanceCommittableSerializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/tiering/LanceLakeWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/tiering/LanceLakeCommitter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/tiering/ArrowWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/tiering/LanceArrowWriter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/utils/LanceArrowUtils.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/utils/LanceDatasetAdapter.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/LanceLakeStoragePlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-lake/fluss-lake-lance/src/main/java/org/apache/fluss/lake/lance/LanceConfig.java' -'/Users/mehulbatra/Documents/fluss/fluss-kafka/src/main/java/org/apache/fluss/kafka/KafkaRequestHandler.java' -'/Users/mehulbatra/Documents/fluss/fluss-kafka/src/main/java/org/apache/fluss/kafka/KafkaChannelInitializer.java' -'/Users/mehulbatra/Documents/fluss/fluss-kafka/src/main/java/org/apache/fluss/kafka/KafkaCommandDecoder.java' -'/Users/mehulbatra/Documents/fluss/fluss-kafka/src/main/java/org/apache/fluss/kafka/KafkaProtocolPlugin.java' -'/Users/mehulbatra/Documents/fluss/fluss-kafka/src/main/java/org/apache/fluss/kafka/KafkaRequest.java' -'/Users/mehulbatra/Documents/fluss/tools/ci/fluss-ci-tools/src/main/java/org/apache/fluss/tools/ci/utils/deploy/DeployParser.java' -'/Users/mehulbatra/Documents/fluss/tools/ci/fluss-ci-tools/src/main/java/org/apache/fluss/tools/ci/utils/dependency/DependencyParser.java' -'/Users/mehulbatra/Documents/fluss/tools/ci/fluss-ci-tools/src/main/java/org/apache/fluss/tools/ci/utils/shared/DependencyTree.java' -'/Users/mehulbatra/Documents/fluss/tools/ci/fluss-ci-tools/src/main/java/org/apache/fluss/tools/ci/utils/shared/ParserUtils.java' -'/Users/mehulbatra/Documents/fluss/tools/ci/fluss-ci-tools/src/main/java/org/apache/fluss/tools/ci/utils/shared/Dependency.java' -'/Users/mehulbatra/Documents/fluss/tools/ci/fluss-ci-tools/src/main/java/org/apache/fluss/tools/ci/utils/notice/NoticeContents.java' -'/Users/mehulbatra/Documents/fluss/tools/ci/fluss-ci-tools/src/main/java/org/apache/fluss/tools/ci/utils/notice/NoticeParser.java' -'/Users/mehulbatra/Documents/fluss/tools/ci/fluss-ci-tools/src/main/java/org/apache/fluss/tools/ci/utils/shade/ShadeParser.java' -'/Users/mehulbatra/Documents/fluss/tools/ci/fluss-ci-tools/src/main/java/org/apache/fluss/tools/ci/licensecheck/LicenseChecker.java' -'/Users/mehulbatra/Documents/fluss/tools/ci/fluss-ci-tools/src/main/java/org/apache/fluss/tools/ci/licensecheck/JarFileChecker.java' -'/Users/mehulbatra/Documents/fluss/tools/ci/fluss-ci-tools/src/main/java/org/apache/fluss/tools/ci/licensecheck/NoticeFileChecker.java' diff --git a/website/static/javadoc/0.8-SNAPSHOT/constant-values.html b/website/static/javadoc/0.8-SNAPSHOT/constant-values.html deleted file mode 100644 index 7b69325fe7..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/constant-values.html +++ /dev/null @@ -1,572 +0,0 @@ - - - - - -Constant Field Values (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Constant Field Values

    -
    -

    Contents

    - -
    -
    -
    - - -
    -

    org.apache.*

    - - -
      -
    • - - - - - - - - - - - - - - -
      org.apache.fluss.client.table.scanner.log.LogScanner 
      Modifier and TypeConstant FieldValue
      - -public static final longEARLIEST_OFFSET-2L
      -
    • -
    - - - - - - - - - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/deprecated-list.html b/website/static/javadoc/0.8-SNAPSHOT/deprecated-list.html deleted file mode 100644 index 3876b76067..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/deprecated-list.html +++ /dev/null @@ -1,175 +0,0 @@ - - - - - -Deprecated List (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Deprecated API

    -

    Contents

    - -
    -
    - - - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/element-list b/website/static/javadoc/0.8-SNAPSHOT/element-list deleted file mode 100644 index f5b65b0576..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/element-list +++ /dev/null @@ -1,47 +0,0 @@ -org.apache.fluss.client -org.apache.fluss.client.admin -org.apache.fluss.client.lookup -org.apache.fluss.client.metadata -org.apache.fluss.client.metrics -org.apache.fluss.client.table -org.apache.fluss.client.table.getter -org.apache.fluss.client.table.scanner -org.apache.fluss.client.table.scanner.batch -org.apache.fluss.client.table.scanner.log -org.apache.fluss.client.table.writer -org.apache.fluss.client.token -org.apache.fluss.client.utils -org.apache.fluss.client.write -org.apache.fluss.flink -org.apache.fluss.flink.adapter -org.apache.fluss.flink.catalog -org.apache.fluss.flink.lake -org.apache.fluss.flink.lake.reader -org.apache.fluss.flink.lake.split -org.apache.fluss.flink.lake.state -org.apache.fluss.flink.metrics -org.apache.fluss.flink.procedure -org.apache.fluss.flink.row -org.apache.fluss.flink.sink -org.apache.fluss.flink.sink.serializer -org.apache.fluss.flink.sink.writer -org.apache.fluss.flink.source -org.apache.fluss.flink.source.deserializer -org.apache.fluss.flink.source.emitter -org.apache.fluss.flink.source.enumerator -org.apache.fluss.flink.source.enumerator.initializer -org.apache.fluss.flink.source.event -org.apache.fluss.flink.source.lookup -org.apache.fluss.flink.source.metrics -org.apache.fluss.flink.source.reader -org.apache.fluss.flink.source.reader.fetcher -org.apache.fluss.flink.source.split -org.apache.fluss.flink.source.state -org.apache.fluss.flink.tiering -org.apache.fluss.flink.tiering.committer -org.apache.fluss.flink.tiering.event -org.apache.fluss.flink.tiering.source -org.apache.fluss.flink.tiering.source.enumerator -org.apache.fluss.flink.tiering.source.split -org.apache.fluss.flink.tiering.source.state -org.apache.fluss.flink.utils diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-1.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-1.html deleted file mode 100644 index 272a9558bd..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-1.html +++ /dev/null @@ -1,369 +0,0 @@ - - - - - -A-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages - - -

    A

    -
    -
    abort(Exception) - Method in class org.apache.fluss.client.write.WriteBatch
    -
    -
    Abort the batch and complete the future and callbacks.
    -
    -
    abortBatches(Exception) - Method in class org.apache.fluss.client.write.RecordAccumulator
    -
    -
    Abort all incomplete batches (whether they have been sent or not).
    -
    -
    abortIfBatchFull() - Method in class org.apache.fluss.client.write.RoundRobinBucketAssigner
    -
     
    -
    abortIfBatchFull() - Method in class org.apache.fluss.client.write.StickyBucketAssigner
    -
     
    -
    abortRecordAppends() - Method in class org.apache.fluss.client.write.ArrowLogWriteBatch
    -
     
    -
    abortRecordAppends() - Method in class org.apache.fluss.client.write.IndexedLogWriteBatch
    -
     
    -
    abortRecordAppends() - Method in class org.apache.fluss.client.write.KvWriteBatch
    -
     
    -
    abortRecordAppends() - Method in class org.apache.fluss.client.write.WriteBatch
    -
     
    -
    abortRecordForNewBatch - Variable in class org.apache.fluss.client.write.RecordAccumulator.RecordAppendResult
    -
    -
    Whether this record was abort because the new batch created in record accumulator.
    -
    -
    AbstractAclProcedure - Class in org.apache.fluss.flink.procedure
    -
    -
    Procedure to operate acl.
    -
    -
    AbstractAclProcedure() - Constructor for class org.apache.fluss.flink.procedure.AbstractAclProcedure
    -
     
    -
    AbstractLookupBatch<T> - Class in org.apache.fluss.client.lookup
    -
    -
    An abstract lookup batch.
    -
    -
    AbstractLookupBatch(TableBucket) - Constructor for class org.apache.fluss.client.lookup.AbstractLookupBatch
    -
     
    -
    AbstractLookupQuery<T> - Class in org.apache.fluss.client.lookup
    -
    -
    Abstract Class to represent a lookup operation.
    -
    -
    AbstractLookupQuery(TableBucket, byte[]) - Constructor for class org.apache.fluss.client.lookup.AbstractLookupQuery
    -
     
    -
    AbstractTableWriter - Class in org.apache.fluss.client.table.writer
    -
    -
    A base class for AppendWriter and UpsertWriter to write data to table.
    -
    -
    AbstractTableWriter(TablePath, TableInfo, WriterClient) - Constructor for class org.apache.fluss.client.table.writer.AbstractTableWriter
    -
     
    -
    aclOperation(Resource, PermissionType, FlussPrincipal, OperationType, String) - Method in class org.apache.fluss.flink.procedure.AbstractAclProcedure
    -
     
    -
    aclOperation(Resource, PermissionType, FlussPrincipal, OperationType, String) - Method in class org.apache.fluss.flink.procedure.AddAclProcedure
    -
     
    -
    aclOperation(Resource, PermissionType, FlussPrincipal, OperationType, String) - Method in class org.apache.fluss.flink.procedure.DropAclProcedure
    -
     
    -
    aclOperation(Resource, PermissionType, FlussPrincipal, OperationType, String) - Method in class org.apache.fluss.flink.procedure.ListAclProcedure
    -
     
    -
    AddAclProcedure - Class in org.apache.fluss.flink.procedure
    -
    -
    Procedure to add acl.
    -
    -
    AddAclProcedure() - Constructor for class org.apache.fluss.flink.procedure.AddAclProcedure
    -
     
    -
    addLookup(AbstractLookupQuery<T>) - Method in class org.apache.fluss.client.lookup.AbstractLookupBatch
    -
     
    -
    addLookup(LookupQuery) - Method in class org.apache.fluss.client.lookup.LookupBatch
    -
     
    -
    addLookup(PrefixLookupQuery) - Method in class org.apache.fluss.client.lookup.PrefixLookupBatch
    -
     
    -
    addReader(int) - Method in class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
    -
     
    -
    addReader(int) - Method in class org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator
    -
     
    -
    addSplit(SourceSplitBase, Queue<SourceSplitBase>) - Method in class org.apache.fluss.flink.lake.LakeSplitReaderGenerator
    -
     
    -
    addSplitsBack(List<SourceSplitBase>, int) - Method in class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
    -
     
    -
    addSplitsBack(List<TieringSplit>, int) - Method in class org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator
    -
     
    -
    admin - Variable in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    admin - Variable in class org.apache.fluss.flink.procedure.ProcedureBase
    -
     
    -
    Admin - Interface in org.apache.fluss.client.admin
    -
    -
    The administrative client for Fluss, which supports managing and inspecting tables, servers, - configurations and ACLs.
    -
    -
    all() - Method in class org.apache.fluss.client.admin.CreateAclsResult
    -
     
    -
    all() - Method in class org.apache.fluss.client.admin.DropAclsResult
    -
    -
    Return a future which succeeds only if all the ACLs deletions succeed, and which contains all - the deleted ACLs.
    -
    -
    all() - Method in class org.apache.fluss.client.admin.ListOffsetsResult
    -
     
    -
    alterDatabase(String, CatalogDatabase, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    alterFunction(ObjectPath, CatalogFunction, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    alterPartition(ObjectPath, CatalogPartitionSpec, CatalogPartition, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    alterPartitionColumnStatistics(ObjectPath, CatalogPartitionSpec, CatalogColumnStatistics, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    alterPartitionStatistics(ObjectPath, CatalogPartitionSpec, CatalogTableStatistics, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    alterTable(ObjectPath, CatalogBaseTable, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    alterTableColumnStatistics(ObjectPath, CatalogColumnStatistics, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    alterTableStatistics(ObjectPath, CatalogTableStatistics, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    append(WriteRecord, WriteCallback, Cluster, int, boolean) - Method in class org.apache.fluss.client.write.RecordAccumulator
    -
    -
    Add a record to the accumulator, return to append result.
    -
    -
    append(InternalRow) - Method in interface org.apache.fluss.client.table.writer.AppendWriter
    -
    -
    Append row into a Log Table.
    -
    -
    Append - Interface in org.apache.fluss.client.table.writer
    -
    -
    Used to configure and create a AppendWriter to write data to a Log Table.
    -
    -
    APPEND - org.apache.fluss.flink.row.OperationType
    -
    -
    Represents an append-only (insert) operation.
    -
    -
    AppendResult - Class in org.apache.fluss.client.table.writer
    -
    -
    The result of appending a record (AppendWriter.append(InternalRow)).
    -
    -
    AppendResult() - Constructor for class org.apache.fluss.client.table.writer.AppendResult
    -
     
    -
    AppendSinkWriter<InputT> - Class in org.apache.fluss.flink.sink.writer
    -
    -
    An append only sink writer for fluss log table.
    -
    -
    AppendSinkWriter(TablePath, Configuration, RowType, MailboxExecutor, FlussSerializationSchema<InputT>) - Constructor for class org.apache.fluss.flink.sink.writer.AppendSinkWriter
    -
     
    -
    AppendWriter - Interface in org.apache.fluss.client.table.writer
    -
    -
    The writer to write data to the log table.
    -
    -
    applyAggregates(List<int[]>, List<AggregateExpression>, DataType) - Method in class org.apache.fluss.flink.source.FlinkTableSource
    -
     
    -
    applyDeleteFilters(List<ResolvedExpression>) - Method in class org.apache.fluss.flink.sink.FlinkTableSink
    -
     
    -
    applyFilters(List<ResolvedExpression>) - Method in class org.apache.fluss.flink.source.FlinkTableSource
    -
     
    -
    applyLimit(long) - Method in class org.apache.fluss.flink.source.FlinkTableSource
    -
     
    -
    applyProjection(int[][], DataType) - Method in class org.apache.fluss.flink.source.FlinkTableSource
    -
     
    -
    applyRowLevelDelete(RowLevelModificationScanContext) - Method in class org.apache.fluss.flink.sink.FlinkTableSink
    -
     
    -
    applyRowLevelModificationScan(SupportsRowLevelModificationScan.RowLevelModificationType, RowLevelModificationScanContext) - Method in class org.apache.fluss.flink.source.FlinkTableSource
    -
     
    -
    applyRowLevelUpdate(List<Column>, RowLevelModificationScanContext) - Method in class org.apache.fluss.flink.sink.FlinkTableSink
    -
     
    -
    applyStaticPartition(Map<String, String>) - Method in class org.apache.fluss.flink.sink.FlinkTableSink
    -
     
    -
    ARROW_LOG - org.apache.fluss.client.write.WriteFormat
    -
     
    -
    ArrowLogWriteBatch - Class in org.apache.fluss.client.write
    -
    -
    A batch of log records managed in ARROW format that is or will be sent to server by ProduceLogRequest.
    -
    -
    ArrowLogWriteBatch(int, PhysicalTablePath, int, ArrowWriter, AbstractPagedOutputView, long) - Constructor for class org.apache.fluss.client.write.ArrowLogWriteBatch
    -
     
    -
    asHybridSnapshotLogSplit() - Method in class org.apache.fluss.flink.source.split.SourceSplitBase
    -
    -
    Casts this split into a HybridSnapshotLogSplit.
    -
    -
    asHybridSnapshotLogSplitState() - Method in class org.apache.fluss.flink.source.split.SourceSplitState
    -
    -
    Casts this split state into a HybridSnapshotLogSplitState.
    -
    -
    asLogSplit() - Method in class org.apache.fluss.flink.source.split.SourceSplitBase
    -
    -
    Casts this split into a LogSplit.
    -
    -
    asLogSplitState() - Method in class org.apache.fluss.flink.source.split.SourceSplitState
    -
    -
    Casts this split state into a LogSplitState.
    -
    -
    assignBucket(byte[]) - Method in class org.apache.fluss.client.write.HashBucketAssigner
    -
     
    -
    assignBucket(Cluster) - Method in class org.apache.fluss.client.write.RoundRobinBucketAssigner
    -
     
    -
    assignBucket(Cluster) - Method in class org.apache.fluss.client.write.StickyBucketAssigner
    -
     
    -
    asSummaryString() - Method in class org.apache.fluss.flink.sink.FlinkTableSink
    -
     
    -
    asSummaryString() - Method in class org.apache.fluss.flink.source.FlinkTableSource
    -
     
    -
    asTieringLogSplit() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
    -
    -
    Casts this split into a TieringLogSplit.
    -
    -
    asTieringSnapshotSplit() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
    -
    -
    Casts this split into a TieringSnapshotSplit.
    -
    -
    asyncLookup(RowData) - Method in class org.apache.fluss.flink.source.lookup.FlinkAsyncLookupFunction
    -
    -
    The invoke entry point of lookup function.
    -
    -
    await() - Method in class org.apache.fluss.client.write.WriteBatch.RequestFuture
    -
    -
    Await the completion of this request.
    -
    -
    awaitFlushCompletion() - Method in class org.apache.fluss.client.write.RecordAccumulator
    -
    -
    Mark all buckets as ready to send and block until to send is complete.
    -
    -
    awaitNotEmpty(long) - Method in class org.apache.fluss.client.table.scanner.log.LogFetcher
    -
     
    -
    -A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-10.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-10.html deleted file mode 100644 index 3ef1223edc..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-10.html +++ /dev/null @@ -1,157 +0,0 @@ - - - - - -J-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages - - -

    J

    -
    -
    JsonStringDeserializationSchema - Class in org.apache.fluss.flink.source.deserializer
    -
    -
    A deserialization schema that converts LogRecord objects to JSON strings.
    -
    -
    JsonStringDeserializationSchema() - Constructor for class org.apache.fluss.flink.source.deserializer.JsonStringDeserializationSchema
    -
     
    -
    -A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-11.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-11.html deleted file mode 100644 index 9e79166a86..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-11.html +++ /dev/null @@ -1,187 +0,0 @@ - - - - - -K-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages - - -

    K

    -
    -
    key() - Method in class org.apache.fluss.client.lookup.AbstractLookupQuery
    -
     
    -
    keyRow() - Method in class org.apache.fluss.flink.lake.reader.KeyValueRow
    -
     
    -
    KeyValueRow - Class in org.apache.fluss.flink.lake.reader
    -
    -
    An InternalRow with the key part.
    -
    -
    KeyValueRow(int[], InternalRow, boolean) - Constructor for class org.apache.fluss.flink.lake.reader.KeyValueRow
    -
     
    -
    KV - org.apache.fluss.client.write.WriteFormat
    -
     
    -
    KvSnapshotBatchScanner - Class in org.apache.fluss.client.table.scanner.batch
    -
    -
    A scanner to scan the snapshot data of a kv bucket.
    -
    -
    KvSnapshotBatchScanner(RowType, TableBucket, List<FsPathAndFileName>, int[], String, KvFormat, RemoteFileDownloader) - Constructor for class org.apache.fluss.client.table.scanner.batch.KvSnapshotBatchScanner
    -
     
    -
    KvSnapshotMetadata - Class in org.apache.fluss.client.metadata
    -
    -
    A class to represent the snapshot info of a kv tablet (bucket).
    -
    -
    KvSnapshotMetadata(List<FsPathAndFileName>, long) - Constructor for class org.apache.fluss.client.metadata.KvSnapshotMetadata
    -
     
    -
    KvSnapshots - Class in org.apache.fluss.client.metadata
    -
    -
    A class representing the kv snapshots of a table or a partition.
    -
    -
    KvSnapshots(long, Long, Map<Integer, Long>, Map<Integer, Long>) - Constructor for class org.apache.fluss.client.metadata.KvSnapshots
    -
     
    -
    KvWriteBatch - Class in org.apache.fluss.client.write
    -
    -
    A batch of kv records that is or will be sent to server by PutKvRequest.
    -
    -
    KvWriteBatch(int, PhysicalTablePath, int, KvFormat, int, AbstractPagedOutputView, int[], long) - Constructor for class org.apache.fluss.client.write.KvWriteBatch
    -
     
    -
    -A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-12.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-12.html deleted file mode 100644 index 585244194a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-12.html +++ /dev/null @@ -1,524 +0,0 @@ - - - - - -L-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages - - -

    L

    -
    -
    LAKE_SNAPSHOT_FLUSS_LOG_SPLIT_KIND - Static variable in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
    -
     
    -
    LAKE_SNAPSHOT_SPLIT_KIND - Static variable in class org.apache.fluss.flink.lake.split.LakeSnapshotSplit
    -
     
    -
    LAKE_TABLE_SPLITTER - Static variable in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    LakeCatalog - Class in org.apache.fluss.flink.lake
    -
    -
    A lake catalog to delegate the operations on lake table.
    -
    -
    LakeCatalog(String, Map<String, String>, ClassLoader) - Constructor for class org.apache.fluss.flink.lake.LakeCatalog
    -
     
    -
    LakeRecordRecordEmitter<OUT> - Class in org.apache.fluss.flink.lake
    -
    -
    The emitter to emit record from lake split.
    -
    -
    LakeRecordRecordEmitter(BiConsumer<ScanRecord, SourceOutput<OUT>>) - Constructor for class org.apache.fluss.flink.lake.LakeRecordRecordEmitter
    -
     
    -
    LakeSnapshot - Class in org.apache.fluss.client.metadata
    -
    -
    A class representing the lake snapshot information of a table.
    -
    -
    LakeSnapshot(long, Map<TableBucket, Long>, Map<Long, String>) - Constructor for class org.apache.fluss.client.metadata.LakeSnapshot
    -
     
    -
    LakeSnapshotAndFlussLogSplit - Class in org.apache.fluss.flink.lake.split
    -
    -
    A split mixing Lake snapshot and Fluss log.
    -
    -
    LakeSnapshotAndFlussLogSplit(TableBucket, String, List<LakeSplit>, long, long) - Constructor for class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
    -
     
    -
    LakeSnapshotAndFlussLogSplit(TableBucket, String, List<LakeSplit>, long, long, long, int, boolean) - Constructor for class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
    -
     
    -
    LakeSnapshotAndFlussLogSplitState - Class in org.apache.fluss.flink.lake.state
    -
    - -
    -
    LakeSnapshotAndFlussLogSplitState(LakeSnapshotAndFlussLogSplit) - Constructor for class org.apache.fluss.flink.lake.state.LakeSnapshotAndFlussLogSplitState
    -
     
    -
    LakeSnapshotAndLogSplitScanner - Class in org.apache.fluss.flink.lake.reader
    -
    -
    A scanner to merge the lakehouse's snapshot and change log.
    -
    -
    LakeSnapshotAndLogSplitScanner(Table, LakeSource<LakeSplit>, LakeSnapshotAndFlussLogSplit, int[]) - Constructor for class org.apache.fluss.flink.lake.reader.LakeSnapshotAndLogSplitScanner
    -
     
    -
    LakeSnapshotScanner - Class in org.apache.fluss.flink.lake.reader
    -
    -
    A scanner for reading lake split LakeSnapshotSplit.
    -
    -
    LakeSnapshotScanner(LakeSource<LakeSplit>, LakeSnapshotSplit) - Constructor for class org.apache.fluss.flink.lake.reader.LakeSnapshotScanner
    -
     
    -
    LakeSnapshotSplit - Class in org.apache.fluss.flink.lake.split
    -
    -
    A split for reading a snapshot of lake.
    -
    -
    LakeSnapshotSplit(TableBucket, String, LakeSplit, int) - Constructor for class org.apache.fluss.flink.lake.split.LakeSnapshotSplit
    -
     
    -
    LakeSnapshotSplit(TableBucket, String, LakeSplit, int, long) - Constructor for class org.apache.fluss.flink.lake.split.LakeSnapshotSplit
    -
     
    -
    LakeSnapshotSplitState - Class in org.apache.fluss.flink.lake.state
    -
    -
    The state of LakeSnapshotSplit.
    -
    -
    LakeSnapshotSplitState(LakeSnapshotSplit) - Constructor for class org.apache.fluss.flink.lake.state.LakeSnapshotSplitState
    -
     
    -
    LakeSourceUtils - Class in org.apache.fluss.flink.utils
    -
    -
    Utils for create lake source.
    -
    -
    LakeSourceUtils() - Constructor for class org.apache.fluss.flink.utils.LakeSourceUtils
    -
     
    -
    LakeSplitGenerator - Class in org.apache.fluss.flink.lake
    -
    -
    A generator for lake splits.
    -
    -
    LakeSplitGenerator(TableInfo, Admin, LakeSource<LakeSplit>, OffsetsInitializer.BucketOffsetsRetriever, OffsetsInitializer, int, Supplier<Set<PartitionInfo>>) - Constructor for class org.apache.fluss.flink.lake.LakeSplitGenerator
    -
     
    -
    LakeSplitReaderGenerator - Class in org.apache.fluss.flink.lake
    -
    -
    A generator to generate reader for lake split.
    -
    -
    LakeSplitReaderGenerator(Table, int[], LakeSource<LakeSplit>) - Constructor for class org.apache.fluss.flink.lake.LakeSplitReaderGenerator
    -
     
    -
    LakeSplitSerializer - Class in org.apache.fluss.flink.lake
    -
    -
    A serializer for lake split.
    -
    -
    LakeSplitSerializer(SimpleVersionedSerializer<LakeSplit>) - Constructor for class org.apache.fluss.flink.lake.LakeSplitSerializer
    -
     
    -
    LakeSplitStateInitializer - Class in org.apache.fluss.flink.lake
    -
    -
    The state initializer for lake split.
    -
    -
    LakeSplitStateInitializer() - Constructor for class org.apache.fluss.flink.lake.LakeSplitStateInitializer
    -
     
    -
    LakeTableFactory - Class in org.apache.fluss.flink.lake
    -
    -
    A factory to create DynamicTableSource for lake table.
    -
    -
    LakeTableFactory() - Constructor for class org.apache.fluss.flink.lake.LakeTableFactory
    -
     
    -
    LakeTieringJobBuilder - Class in org.apache.fluss.flink.tiering
    -
    -
    The builder to build Flink lake tiering job.
    -
    -
    latest() - Static method in interface org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer
    -
    -
    Get an OffsetsInitializer which initializes the offsets to the latest offsets of each - bucket.
    -
    -
    LATEST - org.apache.fluss.flink.FlinkConnectorOptions.ScanStartupMode
    -
     
    -
    latestOffsets(String, Collection<Integer>) - Method in class org.apache.fluss.flink.source.enumerator.initializer.BucketOffsetsRetrieverImpl
    -
     
    -
    latestOffsets(String, Collection<Integer>) - Method in interface org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer.BucketOffsetsRetriever
    -
     
    -
    LatestOffsetsInitializer - Class in org.apache.fluss.flink.source.enumerator.initializer
    -
    -
    An implementation of OffsetsInitializer to initialize the offsets based on a - latest-offset.
    -
    -
    LatestOffsetsInitializer() - Constructor for class org.apache.fluss.flink.source.enumerator.initializer.LatestOffsetsInitializer
    -
     
    -
    LatestSpec() - Constructor for class org.apache.fluss.client.admin.OffsetSpec.LatestSpec
    -
     
    -
    leaderFor(TableBucket) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
    -
     
    -
    limit(int) - Method in interface org.apache.fluss.client.table.scanner.Scan
    -
    -
    Returns a new scan from this that will read the given limited row number.
    -
    -
    limit(int) - Method in class org.apache.fluss.client.table.scanner.TableScan
    -
     
    -
    LimitBatchScanner - Class in org.apache.fluss.client.table.scanner.batch
    -
    -
    A BatchScanner implementation that scans a limited number of records from a table.
    -
    -
    LimitBatchScanner(TableInfo, TableBucket, MetadataUpdater, int[], int) - Constructor for class org.apache.fluss.client.table.scanner.batch.LimitBatchScanner
    -
     
    -
    limitScan(TablePath, Configuration, RowType, int[], long) - Static method in class org.apache.fluss.flink.utils.PushdownUtils
    -
     
    -
    LIST_EARLIEST_OFFSET - Static variable in class org.apache.fluss.client.admin.OffsetSpec
    -
     
    -
    LIST_LATEST_OFFSET - Static variable in class org.apache.fluss.client.admin.OffsetSpec
    -
     
    -
    LIST_OFFSET_FROM_TIMESTAMP - Static variable in class org.apache.fluss.client.admin.OffsetSpec
    -
     
    -
    ListAclProcedure - Class in org.apache.fluss.flink.procedure
    -
    -
    Procedure to list acl.
    -
    -
    ListAclProcedure() - Constructor for class org.apache.fluss.flink.procedure.ListAclProcedure
    -
     
    -
    listAcls(AclBindingFilter) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    Retrieves ACL entries filtered by principal for the specified resource.
    -
    -
    listAcls(AclBindingFilter) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    listDatabases() - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    List all databases in fluss cluster asynchronously.
    -
    -
    listDatabases() - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    listDatabases() - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    listFunctions(String) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    listOffsets(TablePath, String, Collection<Integer>, OffsetSpec) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    List offset for the specified buckets.
    -
    -
    listOffsets(TablePath, String, Collection<Integer>, OffsetSpec) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    listOffsets(TablePath, Collection<Integer>, OffsetSpec) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    List offset for the specified buckets.
    -
    -
    listOffsets(TablePath, Collection<Integer>, OffsetSpec) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    ListOffsetsResult - Class in org.apache.fluss.client.admin
    -
    -
    Result of list offsets request.
    -
    -
    ListOffsetsResult(Map<Integer, CompletableFuture<Long>>) - Constructor for class org.apache.fluss.client.admin.ListOffsetsResult
    -
     
    -
    listPartitionInfos(TablePath) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    List all partitions in the given table in fluss cluster asynchronously.
    -
    -
    listPartitionInfos(TablePath) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    listPartitionInfos(TablePath, PartitionSpec) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    List all partitions in fluss cluster that are under the given table and the given partial - PartitionSpec asynchronously.
    -
    -
    listPartitionInfos(TablePath, PartitionSpec) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    listPartitions(ObjectPath) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    listPartitions(ObjectPath, CatalogPartitionSpec) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    listPartitionsByFilter(ObjectPath, List<Expression>) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    listProcedures() - Static method in class org.apache.fluss.flink.procedure.ProcedureManager
    -
     
    -
    listProcedures(String) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    listTables(String) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    List all tables in the given database in fluss cluster asynchronously.
    -
    -
    listTables(String) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    listTables(String) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    listViews(String) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    LOG - Static variable in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
    -
     
    -
    LOG - Static variable in class org.apache.fluss.flink.source.FlinkTableSource
    -
     
    -
    LOG - Static variable in class org.apache.fluss.flink.utils.LakeSourceUtils
    -
     
    -
    LOG_SPLIT_FLAG - Static variable in class org.apache.fluss.flink.source.split.SourceSplitBase
    -
     
    -
    logEndOffset() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResult
    -
     
    -
    LogFetchBuffer - Class in org.apache.fluss.client.table.scanner.log
    -
    -
    LogFetchBuffer buffers up the results from the tablet server - responses as they are received.
    -
    -
    LogFetchBuffer() - Constructor for class org.apache.fluss.client.table.scanner.log.LogFetchBuffer
    -
     
    -
    LogFetchCollector - Class in org.apache.fluss.client.table.scanner.log
    -
    -
    LogFetchCollector operates at the LogRecordBatch level, as that is what is stored - in the LogFetchBuffer.
    -
    -
    LogFetchCollector(TablePath, LogScannerStatus, Configuration, MetadataUpdater) - Constructor for class org.apache.fluss.client.table.scanner.log.LogFetchCollector
    -
     
    -
    LogFetcher - Class in org.apache.fluss.client.table.scanner.log
    -
    -
    fetcher to fetch log.
    -
    -
    LogFetcher(TableInfo, Projection, LogScannerStatus, Configuration, MetadataUpdater, ScannerMetricGroup, RemoteFileDownloader) - Constructor for class org.apache.fluss.client.table.scanner.log.LogFetcher
    -
     
    -
    logOffset() - Method in class org.apache.fluss.client.table.scanner.ScanRecord
    -
    -
    The position of this record in the corresponding fluss table bucket.
    -
    -
    LogScan - Class in org.apache.fluss.client.table.scanner.log
    -
    -
    Used to describe the operation to scan log data by LogScanner to a table.
    -
    -
    LogScan() - Constructor for class org.apache.fluss.client.table.scanner.log.LogScan
    -
     
    -
    LogScanner - Interface in org.apache.fluss.client.table.scanner.log
    -
    -
    The scanner is used to scan log data of specify table from Fluss.
    -
    -
    LogScannerImpl - Class in org.apache.fluss.client.table.scanner.log
    -
    -
    The default impl of LogScanner.
    -
    -
    LogScannerImpl(Configuration, TableInfo, MetadataUpdater, ClientMetricGroup, RemoteFileDownloader, int[]) - Constructor for class org.apache.fluss.client.table.scanner.log.LogScannerImpl
    -
     
    -
    LogScannerStatus - Class in org.apache.fluss.client.table.scanner.log
    -
    -
    The status of a LogScanner.
    -
    -
    LogScannerStatus() - Constructor for class org.apache.fluss.client.table.scanner.log.LogScannerStatus
    -
     
    -
    LogSplit - Class in org.apache.fluss.flink.source.split
    -
    -
    The split for log.
    -
    -
    LogSplit(TableBucket, String, long) - Constructor for class org.apache.fluss.flink.source.split.LogSplit
    -
     
    -
    LogSplit(TableBucket, String, long, long) - Constructor for class org.apache.fluss.flink.source.split.LogSplit
    -
     
    -
    LogSplitState - Class in org.apache.fluss.flink.source.split
    -
    -
    The state of LogSplit.
    -
    -
    LogSplitState(LogSplit) - Constructor for class org.apache.fluss.flink.source.split.LogSplitState
    -
     
    -
    lookup(RowData) - Method in class org.apache.fluss.flink.source.lookup.FlinkLookupFunction
    -
    -
    The invoke entry point of lookup function.
    -
    -
    lookup(TableBucket, byte[]) - Method in class org.apache.fluss.client.lookup.LookupClient
    -
     
    -
    lookup(InternalRow) - Method in interface org.apache.fluss.client.lookup.Lookuper
    -
    -
    Lookups certain row from the given lookup key.
    -
    -
    Lookup - Interface in org.apache.fluss.client.lookup
    -
    -
    Used to configure and create a Lookuper to lookup rows of a primary key table.
    -
    -
    LOOKUP - org.apache.fluss.client.lookup.LookupType
    -
     
    -
    LOOKUP_ASYNC - Static variable in class org.apache.fluss.flink.FlinkConnectorOptions
    -
     
    -
    LOOKUP_THREAD_PREFIX - Static variable in class org.apache.fluss.client.lookup.LookupClient
    -
     
    -
    LookupBatch - Class in org.apache.fluss.client.lookup
    -
    -
    A batch that contains the lookup operations that send to same tablet bucket together.
    -
    -
    LookupBatch(TableBucket) - Constructor for class org.apache.fluss.client.lookup.LookupBatch
    -
     
    -
    lookupBy(String...) - Method in interface org.apache.fluss.client.lookup.Lookup
    -
     
    -
    lookupBy(List<String>) - Method in interface org.apache.fluss.client.lookup.Lookup
    -
    -
    Returns a new Lookup instance with the given lookup columns.
    -
    -
    lookupBy(List<String>) - Method in class org.apache.fluss.client.lookup.TableLookup
    -
     
    -
    LookupClient - Class in org.apache.fluss.client.lookup
    -
    -
    A client that lookups value of keys from server.
    -
    -
    LookupClient(Configuration, MetadataUpdater) - Constructor for class org.apache.fluss.client.lookup.LookupClient
    -
     
    -
    Lookuper - Interface in org.apache.fluss.client.lookup
    -
    -
    The lookup-er is used to lookup row of a primary key table by primary key or prefix key.
    -
    -
    LookupNormalizer - Class in org.apache.fluss.flink.source.lookup
    -
    -
    A utility class to normalize the lookup key row to match the Fluss key fields order and drop the - lookup result that doesn't match remaining conditions.
    -
    -
    LookupNormalizer.RemainingFilter - Class in org.apache.fluss.flink.source.lookup
    -
    -
    A filter to check if the lookup result matches the remaining conditions.
    -
    -
    LookupQuery - Class in org.apache.fluss.client.lookup
    -
    -
    Class to represent a Lookup operation, it contains the table bucket that the key should lookup - from, the bytes of the key, and a future for the lookup operation.
    -
    -
    LookupResult - Class in org.apache.fluss.client.lookup
    -
    - -
    -
    LookupResult(List<InternalRow>) - Constructor for class org.apache.fluss.client.lookup.LookupResult
    -
     
    -
    LookupResult(InternalRow) - Constructor for class org.apache.fluss.client.lookup.LookupResult
    -
     
    -
    lookups - Variable in class org.apache.fluss.client.lookup.AbstractLookupBatch
    -
     
    -
    lookups() - Method in class org.apache.fluss.client.lookup.AbstractLookupBatch
    -
     
    -
    lookups() - Method in class org.apache.fluss.client.lookup.LookupBatch
    -
     
    -
    lookups() - Method in class org.apache.fluss.client.lookup.PrefixLookupBatch
    -
     
    -
    lookupType() - Method in class org.apache.fluss.client.lookup.AbstractLookupQuery
    -
     
    -
    lookupType() - Method in class org.apache.fluss.client.lookup.LookupQuery
    -
     
    -
    lookupType() - Method in class org.apache.fluss.client.lookup.PrefixLookupQuery
    -
     
    -
    LookupType - Enum in org.apache.fluss.client.lookup
    -
    -
    Enum to represent the type of lookup operation.
    -
    -
    -A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-13.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-13.html deleted file mode 100644 index 8ec98c2ae9..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-13.html +++ /dev/null @@ -1,195 +0,0 @@ - - - - - -M-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages - - -

    M

    -
    -
    makeCreatePartitionRequest(TablePath, PartitionSpec, boolean) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
    -
     
    -
    makeDropPartitionRequest(TablePath, PartitionSpec, boolean) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
    -
     
    -
    makeListOffsetsRequest(long, Long, List<Integer>, OffsetSpec) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
    -
     
    -
    makeLookupRequest(long, Collection<LookupBatch>) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
    -
     
    -
    makeMetadataRequest(Set<TablePath>, Collection<PhysicalTablePath>, Collection<Long>) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
    -
     
    -
    makePbPartitionSpec(PartitionSpec) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
    -
     
    -
    makePrefixLookupRequest(long, Collection<PrefixLookupBatch>) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
    -
     
    -
    makeProduceLogRequest(long, int, int, List<ReadyWriteBatch>) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
    -
     
    -
    makePutKvRequest(long, int, int, List<ReadyWriteBatch>) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
    -
     
    -
    markEvent() - Method in class org.apache.fluss.flink.metrics.FlinkMeter
    -
     
    -
    markEvent(long) - Method in class org.apache.fluss.flink.metrics.FlinkMeter
    -
     
    -
    MetadataUpdater - Class in org.apache.fluss.client.metadata
    -
    -
    The updater to initialize and update client metadata.
    -
    -
    MetadataUpdater(Configuration, RpcClient) - Constructor for class org.apache.fluss.client.metadata.MetadataUpdater
    -
     
    -
    MetadataUpdater(RpcClient, Cluster) - Constructor for class org.apache.fluss.client.metadata.MetadataUpdater
    -
     
    -
    MetadataUtils - Class in org.apache.fluss.client.utils
    -
    -
    Utils for metadata for client.
    -
    -
    MetadataUtils() - Constructor for class org.apache.fluss.client.utils.MetadataUtils
    -
     
    -
    metricGroup - Variable in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
    -
     
    -
    MutableRecordAndPos - Class in org.apache.fluss.flink.source.reader
    -
    -
    A mutable version of the RecordAndPos.
    -
    -
    MutableRecordAndPos() - Constructor for class org.apache.fluss.flink.source.reader.MutableRecordAndPos
    -
     
    -
    -A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-14.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-14.html deleted file mode 100644 index 804dcc3a73..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-14.html +++ /dev/null @@ -1,219 +0,0 @@ - - - - - -N-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages - - -

    N

    -
    -
    newAppend() - Method in class org.apache.fluss.client.table.FlussTable
    -
     
    -
    newAppend() - Method in interface org.apache.fluss.client.table.Table
    -
    -
    Creates a new Append to build a AppendWriter to append data to this table - (requires to be a Log Table).
    -
    -
    newBatchCreated - Variable in class org.apache.fluss.client.write.RecordAccumulator.RecordAppendResult
    -
     
    -
    newBuilder(StreamExecutionEnvironment, Configuration, Configuration, String) - Static method in class org.apache.fluss.flink.tiering.LakeTieringJobBuilder
    -
     
    -
    newCoordinatorServerClient() - Method in class org.apache.fluss.client.metadata.MetadataUpdater
    -
     
    -
    newLookup() - Method in class org.apache.fluss.client.table.FlussTable
    -
     
    -
    newLookup() - Method in interface org.apache.fluss.client.table.Table
    -
    -
    Creates a new Lookup for this table to configure and create a Lookuper to - lookup data for this table by primary key or a prefix of primary key.
    -
    -
    newRandomTabletServerClient() - Method in class org.apache.fluss.client.metadata.MetadataUpdater
    -
     
    -
    newScan() - Method in class org.apache.fluss.client.table.FlussTable
    -
     
    -
    newScan() - Method in interface org.apache.fluss.client.table.Table
    -
    -
    Creates a new Scan for this table to configure and create a scanner to scan data for - this table.
    -
    -
    newTabletServerClientForNode(int) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
    -
     
    -
    newUpsert() - Method in class org.apache.fluss.client.table.FlussTable
    -
     
    -
    newUpsert() - Method in interface org.apache.fluss.client.table.Table
    -
    -
    Creates a new Upsert to build a UpsertWriter to upsert and delete data to - this table (requires to be a Primary Key Table).
    -
    -
    next() - Method in class org.apache.fluss.flink.lake.reader.IndexedLakeSplitRecordIterator
    -
     
    -
    nextReadyCheckDelayMs - Variable in class org.apache.fluss.client.write.RecordAccumulator.ReadyCheckResult
    -
     
    -
    nextRecordFromSplit() - Method in class org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds
    -
     
    -
    nextSplit() - Method in class org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds
    -
     
    -
    NO_DATA_AVAILABLE - Static variable in class org.apache.fluss.client.table.scanner.batch.KvSnapshotBatchScanner
    -
     
    -
    NO_READ_RECORDS_COUNT - Static variable in class org.apache.fluss.flink.source.reader.RecordAndPos
    -
     
    -
    NO_STOPPING_OFFSET - Static variable in class org.apache.fluss.flink.source.split.LogSplit
    -
     
    -
    nonPhysicalColumnsCount(Map<String, String>, List<String>) - Static method in class org.apache.fluss.flink.utils.CatalogPropertiesUtils
    -
     
    -
    normalizeLookupKey(RowData) - Method in class org.apache.fluss.flink.source.lookup.LookupNormalizer
    -
    -
    Normalize the lookup key row to match the request key and the key fields order.
    -
    -
    NoStoppingOffsetsInitializer - Class in org.apache.fluss.flink.source.enumerator.initializer
    -
    -
    An implementation of OffsetsInitializer which does not initialize anything.
    -
    -
    NoStoppingOffsetsInitializer() - Constructor for class org.apache.fluss.flink.source.enumerator.initializer.NoStoppingOffsetsInitializer
    -
     
    -
    numberOfSplits - Variable in class org.apache.fluss.flink.tiering.source.split.TieringSplit
    -
     
    -
    numberOfWriteResults() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResult
    -
     
    -
    -A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-15.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-15.html deleted file mode 100644 index 4069f8b68f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-15.html +++ /dev/null @@ -1,340 +0,0 @@ - - - - - -O-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages - - -

    O

    -
    -
    obtainSecurityToken() - Method in class org.apache.fluss.client.token.DefaultSecurityTokenProvider
    -
     
    -
    obtainSecurityToken() - Method in interface org.apache.fluss.client.token.SecurityTokenProvider
    -
    -
    Obtain security token.
    -
    -
    obtainSecurityTokensAndGetNextRenewal(AtomicReference<ObtainedSecurityToken>) - Method in class org.apache.fluss.client.token.DefaultSecurityTokenManager
    -
     
    -
    of(SerializableSupplier<SimpleVersionedSerializer<Committable>>) - Static method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
    -
     
    -
    of(SerializableSupplier<SimpleVersionedSerializer<WriteResult>>) - Static method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
    -
     
    -
    offsetsFromTimestamp(String, Collection<Integer>, long) - Method in class org.apache.fluss.flink.source.enumerator.initializer.BucketOffsetsRetrieverImpl
    -
     
    -
    offsetsFromTimestamp(String, Collection<Integer>, long) - Method in interface org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer.BucketOffsetsRetriever
    -
     
    -
    offsetsInitializer - Variable in class org.apache.fluss.flink.source.FlinkSource
    -
     
    -
    OffsetsInitializer - Interface in org.apache.fluss.flink.source.enumerator.initializer
    -
    -
    An interface for users to specify the starting offset of a SourceSplitBase.
    -
    -
    OffsetsInitializer.BucketOffsetsRetriever - Interface in org.apache.fluss.flink.source.enumerator.initializer
    -
    -
    An interface that provides necessary information to the OffsetsInitializer to get the - initial offsets of the fluss buckets.
    -
    -
    OffsetSpec - Class in org.apache.fluss.client.admin
    -
    -
    Offset spec.
    -
    -
    OffsetSpec() - Constructor for class org.apache.fluss.client.admin.OffsetSpec
    -
     
    -
    OffsetSpec.EarliestSpec - Class in org.apache.fluss.client.admin
    -
    -
    Earliest offset spec.
    -
    -
    OffsetSpec.LatestSpec - Class in org.apache.fluss.client.admin
    -
    -
    latest offset spec.
    -
    -
    OffsetSpec.TimestampSpec - Class in org.apache.fluss.client.admin
    -
    -
    timestamp offset spec.
    -
    -
    onComplete(Runnable) - Method in class org.apache.fluss.client.table.scanner.log.RemoteLogDownloadFuture
    -
     
    -
    onCompletion(Exception) - Method in interface org.apache.fluss.client.write.WriteCallback
    -
     
    -
    onNewBatch(Cluster, int) - Method in class org.apache.fluss.client.write.RoundRobinBucketAssigner
    -
     
    -
    onNewBatch(Cluster, int) - Method in class org.apache.fluss.client.write.StickyBucketAssigner
    -
     
    -
    onSplitFinished(Map<String, SourceSplitState>) - Method in class org.apache.fluss.flink.source.reader.FlinkSourceReader
    -
     
    -
    onSplitFinished(Map<String, TieringSplitState>) - Method in class org.apache.fluss.flink.tiering.source.TieringSourceReader
    -
     
    -
    open() - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    open() - Method in class org.apache.fluss.flink.tiering.committer.FlussTableLakeSnapshotCommitter
    -
     
    -
    open() - Method in class org.apache.fluss.flink.tiering.committer.TieringCommitOperator
    -
     
    -
    open(FunctionContext) - Method in class org.apache.fluss.flink.source.lookup.FlinkAsyncLookupFunction
    -
     
    -
    open(FunctionContext) - Method in class org.apache.fluss.flink.source.lookup.FlinkLookupFunction
    -
     
    -
    open(FlussSerializationSchema.InitializationContext) - Method in interface org.apache.fluss.flink.sink.serializer.FlussSerializationSchema
    -
    -
    Initialization method for the schema.
    -
    -
    open(FlussSerializationSchema.InitializationContext) - Method in class org.apache.fluss.flink.sink.serializer.RowDataSerializationSchema
    -
    -
    Initializes the schema and its internal converter.
    -
    -
    open(FlussDeserializationSchema.InitializationContext) - Method in interface org.apache.fluss.flink.source.deserializer.FlussDeserializationSchema
    -
    -
    Initialization method for the schema.
    -
    -
    open(FlussDeserializationSchema.InitializationContext) - Method in class org.apache.fluss.flink.source.deserializer.JsonStringDeserializationSchema
    -
    -
    Initializes the JSON serialization mechanism.
    -
    -
    open(FlussDeserializationSchema.InitializationContext) - Method in class org.apache.fluss.flink.source.deserializer.RowDataDeserializationSchema
    -
    -
    Initializes the deserialization schema.
    -
    -
    OperationType - Enum in org.apache.fluss.flink.row
    -
    -
    Enumeration of row operation types used in Fluss-Flink data processing.
    -
    -
    optionalOptions() - Method in class org.apache.fluss.flink.catalog.FlinkCatalogFactory
    -
     
    -
    optionalOptions() - Method in class org.apache.fluss.flink.catalog.FlinkTableFactory
    -
     
    -
    org.apache.fluss.client - package org.apache.fluss.client
    -
     
    -
    org.apache.fluss.client.admin - package org.apache.fluss.client.admin
    -
     
    -
    org.apache.fluss.client.lookup - package org.apache.fluss.client.lookup
    -
     
    -
    org.apache.fluss.client.metadata - package org.apache.fluss.client.metadata
    -
     
    -
    org.apache.fluss.client.metrics - package org.apache.fluss.client.metrics
    -
     
    -
    org.apache.fluss.client.table - package org.apache.fluss.client.table
    -
     
    -
    org.apache.fluss.client.table.getter - package org.apache.fluss.client.table.getter
    -
     
    -
    org.apache.fluss.client.table.scanner - package org.apache.fluss.client.table.scanner
    -
     
    -
    org.apache.fluss.client.table.scanner.batch - package org.apache.fluss.client.table.scanner.batch
    -
     
    -
    org.apache.fluss.client.table.scanner.log - package org.apache.fluss.client.table.scanner.log
    -
     
    -
    org.apache.fluss.client.table.writer - package org.apache.fluss.client.table.writer
    -
     
    -
    org.apache.fluss.client.token - package org.apache.fluss.client.token
    -
     
    -
    org.apache.fluss.client.utils - package org.apache.fluss.client.utils
    -
     
    -
    org.apache.fluss.client.write - package org.apache.fluss.client.write
    -
     
    -
    org.apache.fluss.flink - package org.apache.fluss.flink
    -
     
    -
    org.apache.fluss.flink.adapter - package org.apache.fluss.flink.adapter
    -
     
    -
    org.apache.fluss.flink.catalog - package org.apache.fluss.flink.catalog
    -
     
    -
    org.apache.fluss.flink.lake - package org.apache.fluss.flink.lake
    -
     
    -
    org.apache.fluss.flink.lake.reader - package org.apache.fluss.flink.lake.reader
    -
     
    -
    org.apache.fluss.flink.lake.split - package org.apache.fluss.flink.lake.split
    -
     
    -
    org.apache.fluss.flink.lake.state - package org.apache.fluss.flink.lake.state
    -
     
    -
    org.apache.fluss.flink.metrics - package org.apache.fluss.flink.metrics
    -
     
    -
    org.apache.fluss.flink.procedure - package org.apache.fluss.flink.procedure
    -
     
    -
    org.apache.fluss.flink.row - package org.apache.fluss.flink.row
    -
     
    -
    org.apache.fluss.flink.sink - package org.apache.fluss.flink.sink
    -
     
    -
    org.apache.fluss.flink.sink.serializer - package org.apache.fluss.flink.sink.serializer
    -
     
    -
    org.apache.fluss.flink.sink.writer - package org.apache.fluss.flink.sink.writer
    -
     
    -
    org.apache.fluss.flink.source - package org.apache.fluss.flink.source
    -
     
    -
    org.apache.fluss.flink.source.deserializer - package org.apache.fluss.flink.source.deserializer
    -
     
    -
    org.apache.fluss.flink.source.emitter - package org.apache.fluss.flink.source.emitter
    -
     
    -
    org.apache.fluss.flink.source.enumerator - package org.apache.fluss.flink.source.enumerator
    -
     
    -
    org.apache.fluss.flink.source.enumerator.initializer - package org.apache.fluss.flink.source.enumerator.initializer
    -
     
    -
    org.apache.fluss.flink.source.event - package org.apache.fluss.flink.source.event
    -
     
    -
    org.apache.fluss.flink.source.lookup - package org.apache.fluss.flink.source.lookup
    -
     
    -
    org.apache.fluss.flink.source.metrics - package org.apache.fluss.flink.source.metrics
    -
     
    -
    org.apache.fluss.flink.source.reader - package org.apache.fluss.flink.source.reader
    -
     
    -
    org.apache.fluss.flink.source.reader.fetcher - package org.apache.fluss.flink.source.reader.fetcher
    -
     
    -
    org.apache.fluss.flink.source.split - package org.apache.fluss.flink.source.split
    -
     
    -
    org.apache.fluss.flink.source.state - package org.apache.fluss.flink.source.state
    -
     
    -
    org.apache.fluss.flink.tiering - package org.apache.fluss.flink.tiering
    -
     
    -
    org.apache.fluss.flink.tiering.committer - package org.apache.fluss.flink.tiering.committer
    -
     
    -
    org.apache.fluss.flink.tiering.event - package org.apache.fluss.flink.tiering.event
    -
     
    -
    org.apache.fluss.flink.tiering.source - package org.apache.fluss.flink.tiering.source
    -
     
    -
    org.apache.fluss.flink.tiering.source.enumerator - package org.apache.fluss.flink.tiering.source.enumerator
    -
     
    -
    org.apache.fluss.flink.tiering.source.split - package org.apache.fluss.flink.tiering.source.split
    -
     
    -
    org.apache.fluss.flink.tiering.source.state - package org.apache.fluss.flink.tiering.source.state
    -
     
    -
    org.apache.fluss.flink.utils - package org.apache.fluss.flink.utils
    -
     
    -
    -A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-16.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-16.html deleted file mode 100644 index a8757d8f7b..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-16.html +++ /dev/null @@ -1,314 +0,0 @@ - - - - - -P-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages - - -

    P

    -
    -
    parseAndValidateAddresses(List<String>) - Static method in class org.apache.fluss.client.utils.ClientUtils
    -
     
    -
    parseTimestamp(String, String, ZoneId) - Static method in class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils
    -
    -
    Parses timestamp String to Long.
    -
    -
    partialUpdate(int[]) - Method in class org.apache.fluss.client.table.writer.TableUpsert
    -
     
    -
    partialUpdate(int[]) - Method in interface org.apache.fluss.client.table.writer.Upsert
    -
    -
    Apply partial update columns and returns a new Upsert instance.
    -
    -
    partialUpdate(String...) - Method in class org.apache.fluss.client.table.writer.TableUpsert
    -
     
    -
    partialUpdate(String...) - Method in interface org.apache.fluss.client.table.writer.Upsert
    -
     
    -
    partition() - Method in class org.apache.fluss.flink.tiering.source.TieringWriterInitContext
    -
     
    -
    partition(DataStream<T>, ChannelComputer<T>, Integer) - Static method in class org.apache.fluss.flink.sink.FlinkStreamPartitioner
    -
     
    -
    PARTITION_GROUP - Static variable in class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
    -
     
    -
    PartitionBucketsUnsubscribedEvent - Class in org.apache.fluss.flink.source.event
    -
    -
    An event send from reader to enumerator to indicate the splits of the partition buckets have been - removed by the reader.
    -
    -
    PartitionBucketsUnsubscribedEvent(Collection<TableBucket>) - Constructor for class org.apache.fluss.flink.source.event.PartitionBucketsUnsubscribedEvent
    -
     
    -
    partitionExists(ObjectPath, CatalogPartitionSpec) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    PartitionGetter - Class in org.apache.fluss.client.table.getter
    -
    -
    A getter to get partition name from a row.
    -
    -
    PartitionGetter(RowType, List<String>) - Constructor for class org.apache.fluss.client.table.getter.PartitionGetter
    -
     
    -
    partitionName - Variable in class org.apache.fluss.flink.source.split.SourceSplitBase
    -
     
    -
    partitionName - Variable in class org.apache.fluss.flink.tiering.source.split.TieringSplit
    -
     
    -
    partitionName() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResult
    -
     
    -
    PartitionsRemovedEvent - Class in org.apache.fluss.flink.source.event
    -
    -
    A source event to represent partitions is removed to send from enumerator to reader.
    -
    -
    PartitionsRemovedEvent(Map<Long, String>) - Constructor for class org.apache.fluss.flink.source.event.PartitionsRemovedEvent
    -
     
    -
    physicalTablePath() - Method in class org.apache.fluss.client.write.WriteBatch
    -
     
    -
    PojoToRowConverter<T> - Class in org.apache.fluss.flink.utils
    -
    -
    Utility class for converting Java POJOs to Fluss's InternalRow format.
    -
    -
    PojoToRowConverter(Class<T>, RowType) - Constructor for class org.apache.fluss.flink.utils.PojoToRowConverter
    -
    -
    Creates a new converter for the specified POJO class and row type.
    -
    -
    poll(Duration) - Method in interface org.apache.fluss.client.table.scanner.log.LogScanner
    -
    -
    Poll log data from tablet server.
    -
    -
    poll(Duration) - Method in class org.apache.fluss.client.table.scanner.log.LogScannerImpl
    -
     
    -
    poll(Duration) - Method in interface org.apache.fluss.flink.source.reader.SplitScanner
    -
     
    -
    POLL_TIERING_TABLE_INTERVAL - Static variable in class org.apache.fluss.flink.tiering.source.TieringSourceOptions
    -
     
    -
    pollBatch(Duration) - Method in interface org.apache.fluss.client.table.scanner.batch.BatchScanner
    -
    -
    Poll one batch records.
    -
    -
    pollBatch(Duration) - Method in class org.apache.fluss.client.table.scanner.batch.KvSnapshotBatchScanner
    -
    -
    Fetch data from snapshot.
    -
    -
    pollBatch(Duration) - Method in class org.apache.fluss.client.table.scanner.batch.LimitBatchScanner
    -
     
    -
    pollBatch(Duration) - Method in class org.apache.fluss.flink.lake.reader.LakeSnapshotAndLogSplitScanner
    -
     
    -
    pollBatch(Duration) - Method in class org.apache.fluss.flink.lake.reader.LakeSnapshotScanner
    -
     
    -
    pollBatch(Duration) - Method in class org.apache.fluss.flink.lake.reader.SeekableLakeSnapshotSplitScanner
    -
     
    -
    pooledMemorySegments() - Method in class org.apache.fluss.client.write.ArrowLogWriteBatch
    -
     
    -
    pooledMemorySegments() - Method in class org.apache.fluss.client.write.IndexedLogWriteBatch
    -
     
    -
    pooledMemorySegments() - Method in class org.apache.fluss.client.write.KvWriteBatch
    -
     
    -
    pooledMemorySegments() - Method in class org.apache.fluss.client.write.WriteBatch
    -
    -
    get pooled memory segments to de-allocate.
    -
    -
    PREFIX_LOOKUP - org.apache.fluss.client.lookup.LookupType
    -
     
    -
    prefixLookup(TableBucket, byte[]) - Method in class org.apache.fluss.client.lookup.LookupClient
    -
     
    -
    PrefixLookup - Class in org.apache.fluss.client.lookup
    -
    -
    Used to describe the operation to prefix lookup by PrefixLookuper to a primary key table.
    -
    -
    PrefixLookup(List<String>) - Constructor for class org.apache.fluss.client.lookup.PrefixLookup
    -
     
    -
    PrefixLookupBatch - Class in org.apache.fluss.client.lookup
    -
    -
    A batch that contains the prefix lookup operations that send to same destination and some table - together.
    -
    -
    PrefixLookupBatch(TableBucket) - Constructor for class org.apache.fluss.client.lookup.PrefixLookupBatch
    -
     
    -
    PrefixLookupQuery - Class in org.apache.fluss.client.lookup
    -
    -
    Class to represent a prefix lookup operation, it contains the table id, bucketNums and related - CompletableFuture.
    -
    -
    ProcedureBase - Class in org.apache.fluss.flink.procedure
    -
    -
    Base implementation for flink Procedure.
    -
    -
    ProcedureBase() - Constructor for class org.apache.fluss.flink.procedure.ProcedureBase
    -
     
    -
    ProcedureManager - Class in org.apache.fluss.flink.procedure
    -
    -
    ProcedureUtil to load procedure.
    -
    -
    ProcedureManager() - Constructor for class org.apache.fluss.flink.procedure.ProcedureManager
    -
     
    -
    processElement(StreamRecord<TableBucketWriteResult<WriteResult>>) - Method in class org.apache.fluss.flink.tiering.committer.TieringCommitOperator
    -
     
    -
    project(int[]) - Method in interface org.apache.fluss.client.table.scanner.Scan
    -
    -
    Returns a new scan from this that will read the given data columns.
    -
    -
    project(int[]) - Method in class org.apache.fluss.client.table.scanner.TableScan
    -
     
    -
    project(List<String>) - Method in interface org.apache.fluss.client.table.scanner.Scan
    -
    -
    Returns a new scan from this that will read the given data columns.
    -
    -
    project(List<String>) - Method in class org.apache.fluss.client.table.scanner.TableScan
    -
     
    -
    projectRowType(RowType, int[]) - Static method in class org.apache.fluss.flink.utils.FlinkUtils
    -
    -
    Returns projected RowType by given projection indexes over original RowType.
    -
    -
    PushdownUtils - Class in org.apache.fluss.flink.utils
    -
    -
    Utilities for pushdown abilities.
    -
    -
    PushdownUtils() - Constructor for class org.apache.fluss.flink.utils.PushdownUtils
    -
     
    -
    PushdownUtils.FieldEqual - Class in org.apache.fluss.flink.utils
    -
    -
    A structure represents a source field equal literal expression.
    -
    -
    PushdownUtils.ValueConversion - Enum in org.apache.fluss.flink.utils
    -
    -
    The value conversion type between Flink internal value and Fluss internal value.
    -
    -
    putVariables(Map<String, String>) - Method in class org.apache.fluss.client.metrics.ScannerMetricGroup
    -
     
    -
    -A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-17.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-17.html deleted file mode 100644 index 6810314938..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-17.html +++ /dev/null @@ -1,153 +0,0 @@ - - - - - -Q-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages - - -

    Q

    -
    -
    querySingleRow(GenericRowData, TablePath, Configuration, RowType, int[], int, int[]) - Static method in class org.apache.fluss.flink.utils.PushdownUtils
    -
     
    -
    -A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-18.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-18.html deleted file mode 100644 index 12c8a5a49d..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-18.html +++ /dev/null @@ -1,333 +0,0 @@ - - - - - -R-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages - - -

    R

    -
    -
    readBatch() - Method in class org.apache.fluss.flink.source.reader.BoundedSplitReader
    -
    -
    Read next batch of data.
    -
    -
    READER_METRIC_GROUP - Static variable in class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
    -
     
    -
    readRecordsCount - Variable in class org.apache.fluss.flink.source.reader.RecordAndPos
    -
     
    -
    readRecordsCount() - Method in class org.apache.fluss.flink.source.reader.RecordAndPos
    -
     
    -
    ready(Cluster) - Method in class org.apache.fluss.client.write.RecordAccumulator
    -
    -
    Get a list of nodes whose buckets are ready to be sent.
    -
    -
    ReadyCheckResult(Set<Integer>, long, Set<PhysicalTablePath>) - Constructor for class org.apache.fluss.client.write.RecordAccumulator.ReadyCheckResult
    -
     
    -
    readyNodes - Variable in class org.apache.fluss.client.write.RecordAccumulator.ReadyCheckResult
    -
     
    -
    ReadyWriteBatch - Class in org.apache.fluss.client.write
    -
    -
    The WriteBatch already ready in sender.
    -
    -
    ReadyWriteBatch(TableBucket, WriteBatch) - Constructor for class org.apache.fluss.client.write.ReadyWriteBatch
    -
     
    -
    record() - Method in class org.apache.fluss.flink.source.reader.RecordAndPos
    -
     
    -
    RecordAccumulator - Class in org.apache.fluss.client.write
    -
    -
    This class act as a queue that accumulates records into WriteBatch instances to be sent - to tablet servers.
    -
    -
    RecordAccumulator.ReadyCheckResult - Class in org.apache.fluss.client.write
    -
    -
    The set of nodes that have at leader one complete record batch in the accumulator.
    -
    -
    RecordAccumulator.RecordAppendResult - Class in org.apache.fluss.client.write
    -
    -
    Metadata about a record just appended to the record accumulator.
    -
    -
    RecordAndPos - Class in org.apache.fluss.flink.source.reader
    -
    -
    A record wrapping a Fluss ScanRecord, the readRecordsCount when the record is - from reading snapshot, the code currentSplitIndex where the record is from when the Flink - source split read by split reader contains multiple splits(splittable unit), like LakeSnapshotAndFlussLogSplit contains multiple LakeSplits.
    -
    -
    RecordAndPos(ScanRecord) - Constructor for class org.apache.fluss.flink.source.reader.RecordAndPos
    -
     
    -
    RecordAndPos(ScanRecord, long) - Constructor for class org.apache.fluss.flink.source.reader.RecordAndPos
    -
     
    -
    RecordAndPos(ScanRecord, long, int) - Constructor for class org.apache.fluss.flink.source.reader.RecordAndPos
    -
     
    -
    RecordAppendResult(boolean, boolean, boolean) - Constructor for class org.apache.fluss.client.write.RecordAccumulator.RecordAppendResult
    -
     
    -
    recordCount - Variable in class org.apache.fluss.client.write.WriteBatch
    -
     
    -
    recordCurrentOffset(TableBucket, long) - Method in class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
    -
    -
    Update current consuming offset of the given TableBucket.
    -
    -
    recordPerBatch() - Method in class org.apache.fluss.client.metrics.WriterMetricGroup
    -
     
    -
    recordPollEnd(long) - Method in class org.apache.fluss.client.metrics.ScannerMetricGroup
    -
     
    -
    recordPollStart(long) - Method in class org.apache.fluss.client.metrics.ScannerMetricGroup
    -
     
    -
    records(TableBucket) - Method in class org.apache.fluss.client.table.scanner.log.ScanRecords
    -
    -
    Get just the records for the given bucketId.
    -
    -
    recordsRetryTotal() - Method in class org.apache.fluss.client.metrics.WriterMetricGroup
    -
     
    -
    recordsSendTotal() - Method in class org.apache.fluss.client.metrics.WriterMetricGroup
    -
     
    -
    recordsToSkip - Variable in class org.apache.fluss.flink.source.split.SnapshotSplit
    -
    -
    The records to skip when reading the snapshot.
    -
    -
    recordsToSkip() - Method in class org.apache.fluss.flink.source.split.SnapshotSplit
    -
     
    -
    recycle() - Method in class org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds
    -
     
    -
    reEnqueue(ReadyWriteBatch) - Method in class org.apache.fluss.client.write.RecordAccumulator
    -
     
    -
    register(Metric, String, AbstractMetricGroup) - Method in class org.apache.fluss.flink.metrics.FlinkMetricRegistry
    -
     
    -
    registerTableBucket(TableBucket) - Method in class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
    -
     
    -
    remoteFetchBytes() - Method in class org.apache.fluss.client.metrics.ScannerMetricGroup
    -
     
    -
    remoteFetchErrorCount() - Method in class org.apache.fluss.client.metrics.ScannerMetricGroup
    -
     
    -
    remoteFetchRequestCount() - Method in class org.apache.fluss.client.metrics.ScannerMetricGroup
    -
     
    -
    RemoteFileDownloader - Class in org.apache.fluss.client.table.scanner
    -
    -
    The downloader that has a IO thread pool to download the remote files (like kv snapshots files, - log segment files).
    -
    -
    RemoteFileDownloader(int) - Constructor for class org.apache.fluss.client.table.scanner.RemoteFileDownloader
    -
     
    -
    RemoteLogDownloader - Class in org.apache.fluss.client.table.scanner.log
    -
    -
    Downloader to read remote log files to local disk.
    -
    -
    RemoteLogDownloader(TablePath, Configuration, RemoteFileDownloader, ScannerMetricGroup) - Constructor for class org.apache.fluss.client.table.scanner.log.RemoteLogDownloader
    -
     
    -
    RemoteLogDownloadFuture - Class in org.apache.fluss.client.table.scanner.log
    -
    -
    Represents the future of a remote log download request.
    -
    -
    RemoteLogDownloadFuture(CompletableFuture<File>, Runnable) - Constructor for class org.apache.fluss.client.table.scanner.log.RemoteLogDownloadFuture
    -
     
    -
    removePartitions(Map<Long, String>) - Method in class org.apache.fluss.flink.source.reader.FlinkSourceSplitReader
    -
     
    -
    removePartitions(Map<Long, String>, Consumer<Set<TableBucket>>) - Method in class org.apache.fluss.flink.source.reader.fetcher.FlinkSourceFetcherManager
    -
     
    -
    renameTable(ObjectPath, String, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    reopened - Variable in class org.apache.fluss.client.write.WriteBatch
    -
     
    -
    replace(RowData) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
    -
     
    -
    reportRecordEventTime(long) - Method in class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
    -
     
    -
    RequestFuture() - Constructor for class org.apache.fluss.client.write.WriteBatch.RequestFuture
    -
     
    -
    requestRemoteLog(FsPath, RemoteLogSegment) - Method in class org.apache.fluss.client.table.scanner.log.RemoteLogDownloader
    -
    -
    Request to fetch remote log segment to local.
    -
    -
    requiredOptions() - Method in class org.apache.fluss.flink.catalog.FlinkCatalogFactory
    -
     
    -
    requiredOptions() - Method in class org.apache.fluss.flink.catalog.FlinkTableFactory
    -
     
    -
    resetWriterState(long, int) - Method in class org.apache.fluss.client.write.ArrowLogWriteBatch
    -
     
    -
    resetWriterState(long, int) - Method in class org.apache.fluss.client.write.IndexedLogWriteBatch
    -
     
    -
    resetWriterState(long, int) - Method in class org.apache.fluss.client.write.KvWriteBatch
    -
     
    -
    resetWriterState(long, int) - Method in class org.apache.fluss.client.write.WriteBatch
    -
     
    -
    restoreEnumerator(SplitEnumeratorContext<SourceSplitBase>, SourceEnumeratorState) - Method in class org.apache.fluss.flink.source.FlinkSource
    -
     
    -
    restoreEnumerator(SplitEnumeratorContext<TieringSplit>, TieringSourceEnumeratorState) - Method in class org.apache.fluss.flink.tiering.source.TieringSource
    -
     
    -
    RoundRobinBucketAssigner - Class in org.apache.fluss.client.write
    -
    -
    The bucket assigner use round-robin strategy.
    -
    -
    RoundRobinBucketAssigner(PhysicalTablePath) - Constructor for class org.apache.fluss.client.write.RoundRobinBucketAssigner
    -
     
    -
    RowDataDeserializationSchema - Class in org.apache.fluss.flink.source.deserializer
    -
    -
    A deserialization schema that converts LogRecord objects to Flink's RowData - format.
    -
    -
    RowDataDeserializationSchema() - Constructor for class org.apache.fluss.flink.source.deserializer.RowDataDeserializationSchema
    -
     
    -
    RowDataSerializationSchema - Class in org.apache.fluss.flink.sink.serializer
    -
    -
    Default implementation of RowDataConverter for RowData.
    -
    -
    RowDataSerializationSchema(boolean, boolean) - Constructor for class org.apache.fluss.flink.sink.serializer.RowDataSerializationSchema
    -
    -
    Constructs a new RowSerializationSchema.
    -
    -
    RowWithOp - Class in org.apache.fluss.flink.row
    -
    -
    A wrapper class that associates an InternalRow with an OperationType for use in - Fluss-Flink data processing.
    -
    -
    RowWithOp(InternalRow, OperationType) - Constructor for class org.apache.fluss.flink.row.RowWithOp
    -
    -
    Constructs a RowWithOp with the specified internal row and operation type.
    -
    -
    run() - Method in class org.apache.fluss.client.write.Sender
    -
     
    -
    runOnce() - Method in class org.apache.fluss.client.write.Sender
    -
    -
    Run a single iteration of sending.
    -
    -
    -A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-19.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-19.html deleted file mode 100644 index 1eb01fce2f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-19.html +++ /dev/null @@ -1,536 +0,0 @@ - - - - - -S-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages - - -

    S

    -
    -
    Scan - Interface in org.apache.fluss.client.table.scanner
    -
    -
    Used to configure and create a scanner to scan data for a table.
    -
    -
    SCAN_PARTITION_DISCOVERY_INTERVAL - Static variable in class org.apache.fluss.flink.FlinkConnectorOptions
    -
     
    -
    SCAN_STARTUP_MODE - Static variable in class org.apache.fluss.flink.FlinkConnectorOptions
    -
     
    -
    SCAN_STARTUP_TIMESTAMP - Static variable in class org.apache.fluss.flink.FlinkConnectorOptions
    -
     
    -
    ScannerMetricGroup - Class in org.apache.fluss.client.metrics
    -
    -
    The metric group for scanner, including LogScanner and BatchScanner.
    -
    -
    ScannerMetricGroup(ClientMetricGroup, TablePath) - Constructor for class org.apache.fluss.client.metrics.ScannerMetricGroup
    -
     
    -
    scanPartitionDiscoveryIntervalMs - Variable in class org.apache.fluss.flink.source.FlinkSource
    -
     
    -
    scanRecord - Variable in class org.apache.fluss.flink.source.reader.RecordAndPos
    -
     
    -
    ScanRecord - Class in org.apache.fluss.client.table.scanner
    -
    -
    one scan record.
    -
    -
    ScanRecord(long, long, ChangeType, InternalRow) - Constructor for class org.apache.fluss.client.table.scanner.ScanRecord
    -
     
    -
    ScanRecord(InternalRow) - Constructor for class org.apache.fluss.client.table.scanner.ScanRecord
    -
     
    -
    ScanRecords - Class in org.apache.fluss.client.table.scanner.log
    -
    -
    A container that holds the list ScanRecord per bucket for a particular table.
    -
    -
    ScanRecords(Map<TableBucket, List<ScanRecord>>) - Constructor for class org.apache.fluss.client.table.scanner.log.ScanRecords
    -
     
    -
    SecurityTokenManager - Interface in org.apache.fluss.client.token
    -
    -
    API for security token manager.
    -
    -
    SecurityTokenProvider - Interface in org.apache.fluss.client.token
    -
    -
    Security token provider API.
    -
    -
    SeekableLakeSnapshotSplitScanner - Class in org.apache.fluss.flink.lake.reader
    -
    -
    A scanner that supports seeking to a specific LakeSplit and reading from that point.
    -
    -
    SeekableLakeSnapshotSplitScanner(LakeSource<LakeSplit>, List<LakeSplit>, int) - Constructor for class org.apache.fluss.flink.lake.reader.SeekableLakeSnapshotSplitScanner
    -
     
    -
    select(int, int) - Static method in interface org.apache.fluss.flink.sink.ChannelComputer
    -
     
    -
    select(String, int, int) - Static method in interface org.apache.fluss.flink.sink.ChannelComputer
    -
     
    -
    selectChannel(SerializationDelegate<StreamRecord<T>>) - Method in class org.apache.fluss.flink.sink.FlinkStreamPartitioner
    -
     
    -
    selectRowCount - Variable in class org.apache.fluss.flink.source.FlinkTableSource
    -
     
    -
    send(WriteRecord) - Method in class org.apache.fluss.client.table.writer.AbstractTableWriter
    -
     
    -
    send(WriteRecord, WriteCallback) - Method in class org.apache.fluss.client.write.WriterClient
    -
    -
    Asynchronously send a record to a table and invoke the provided callback when to send has - been acknowledged.
    -
    -
    Sender - Class in org.apache.fluss.client.write
    -
    -
    This background thread handles the sending of produce requests to the tablet server.
    -
    -
    Sender(RecordAccumulator, int, int, short, int, MetadataUpdater, IdempotenceManager, WriterMetricGroup) - Constructor for class org.apache.fluss.client.write.Sender
    -
     
    -
    SENDER_THREAD_PREFIX - Static variable in class org.apache.fluss.client.write.WriterClient
    -
     
    -
    sendFetches() - Method in class org.apache.fluss.client.table.scanner.log.LogFetcher
    -
    -
    Set up a fetch request for any node that we have assigned buckets for which doesn't already - have an in-flight fetch or pending fetch data.
    -
    -
    sendMetadataRequestAndRebuildCluster(Cluster, RpcClient, Set<TablePath>, Collection<PhysicalTablePath>, Collection<Long>) - Static method in class org.apache.fluss.client.utils.MetadataUtils
    -
    -
    Partial update cluster, means we will rebuild the cluster by sending metadata request to - request the input tables/partitions in physicalTablePaths, after that add those - tables/partitions into cluster.
    -
    -
    sendMetadataRequestAndRebuildCluster(AdminReadOnlyGateway, boolean, Cluster, Set<TablePath>, Collection<PhysicalTablePath>, Collection<Long>) - Static method in class org.apache.fluss.client.utils.MetadataUtils
    -
    -
    maybe partial update cluster.
    -
    -
    sendMetadataRequestAndRebuildCluster(AdminReadOnlyGateway, Set<TablePath>) - Static method in class org.apache.fluss.client.utils.MetadataUtils
    -
    -
    full update cluster, means we will rebuild the cluster by clearing all cached table in - cluster, and then send metadata request to request the input tables in tablePaths, after that - add those table into cluster.
    -
    -
    sequenceHasBeenReset() - Method in class org.apache.fluss.client.write.WriteBatch
    -
     
    -
    serialize(Object) - Method in interface org.apache.fluss.flink.utils.FlinkRowToFlussRowConverter.FlussSerializationConverter
    -
    -
    Deprecated.
    -
    Convert a Flink field object of RowData to the Fluss's internal data structure - object.
    -
    -
    serialize(DataOutputSerializer, SourceSplitBase) - Method in class org.apache.fluss.flink.lake.LakeSplitSerializer
    -
     
    -
    serialize(RowData) - Method in class org.apache.fluss.flink.sink.serializer.RowDataSerializationSchema
    -
    -
    Serializes a Flink RowData into a Fluss RowWithOp containing an InternalRow and the corresponding OperationType.
    -
    -
    serialize(SourceSplitBase) - Method in class org.apache.fluss.flink.source.split.SourceSplitSerializer
    -
     
    -
    serialize(SourceEnumeratorState) - Method in class org.apache.fluss.flink.source.state.FlussSourceEnumeratorStateSerializer
    -
     
    -
    serialize(TieringSplit) - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplitSerializer
    -
     
    -
    serialize(TieringSourceEnumeratorState) - Method in class org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorStateSerializer
    -
     
    -
    serialize(TableBucketWriteResult<WriteResult>) - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultSerializer
    -
     
    -
    serialize(T) - Method in interface org.apache.fluss.flink.sink.serializer.FlussSerializationSchema
    -
    -
    Serializes an object of type T into a RowWithOp.
    -
    -
    serializeComputedColumns(List<Column>) - Static method in class org.apache.fluss.flink.utils.CatalogPropertiesUtils
    -
     
    -
    serializeComputedColumns(Map<String, String>, List<Column>) - Static method in class org.apache.fluss.flink.utils.CatalogPropertiesUtils
    -
     
    -
    SerializerInitContextImpl - Class in org.apache.fluss.flink.sink.serializer
    -
    -
    Contextual information that can be used during initialization of FlussSerializationSchema.
    -
    -
    SerializerInitContextImpl(RowType) - Constructor for class org.apache.fluss.flink.sink.serializer.SerializerInitContextImpl
    -
     
    -
    serializeWatermarkSpecs(Map<String, String>, List<WatermarkSpec>) - Static method in class org.apache.fluss.flink.utils.CatalogPropertiesUtils
    -
     
    -
    setBatchQueueTimeMs(long) - Method in class org.apache.fluss.client.metrics.WriterMetricGroup
    -
     
    -
    setBootstrapServers(String) - Method in class org.apache.fluss.flink.sink.FlussSinkBuilder
    -
    -
    Set the bootstrap server for the sink.
    -
    -
    setBootstrapServers(String) - Method in class org.apache.fluss.flink.source.FlussSourceBuilder
    -
    -
    Sets the bootstrap servers for the Fluss source connection.
    -
    -
    setCurrentLakeSplitIndex(int) - Method in class org.apache.fluss.flink.lake.state.LakeSnapshotAndFlussLogSplitState
    -
     
    -
    setDatabase(String) - Method in class org.apache.fluss.flink.sink.FlussSinkBuilder
    -
    -
    Set the database for the sink.
    -
    -
    setDatabase(String) - Method in class org.apache.fluss.flink.source.FlussSourceBuilder
    -
    -
    Sets the database name for the Fluss source.
    -
    -
    setDeserializationSchema(FlussDeserializationSchema<OUT>) - Method in class org.apache.fluss.flink.source.FlussSourceBuilder
    -
    -
    Sets the deserialization schema for converting Fluss records to output records.
    -
    -
    setFlussConfig(Configuration) - Method in class org.apache.fluss.flink.source.FlussSourceBuilder
    -
    -
    Sets custom Fluss configuration properties for the source connector.
    -
    -
    setNextLogOffset(long) - Method in class org.apache.fluss.flink.lake.state.LakeSnapshotAndFlussLogSplitState
    -
     
    -
    setNextOffset(long) - Method in class org.apache.fluss.flink.source.split.HybridSnapshotLogSplitState
    -
     
    -
    setNextOffset(long) - Method in class org.apache.fluss.flink.source.split.LogSplitState
    -
     
    -
    setOption(String, String) - Method in class org.apache.fluss.flink.sink.FlussSinkBuilder
    -
    -
    Set a configuration option.
    -
    -
    setOptions(Map<String, String>) - Method in class org.apache.fluss.flink.sink.FlussSinkBuilder
    -
    -
    Set multiple configuration options.
    -
    -
    setProjectedFields(String...) - Method in class org.apache.fluss.flink.source.FlussSourceBuilder
    -
    -
    Sets the projected fields for this source using field names.
    -
    -
    setRecord(ScanRecord, long) - Method in class org.apache.fluss.flink.source.reader.MutableRecordAndPos
    -
     
    -
    setRecord(ScanRecord, long, int) - Method in class org.apache.fluss.flink.source.reader.MutableRecordAndPos
    -
     
    -
    setRecordsToSkip(long) - Method in class org.apache.fluss.flink.lake.state.LakeSnapshotAndFlussLogSplitState
    -
     
    -
    setRecordsToSkip(long) - Method in class org.apache.fluss.flink.lake.state.LakeSnapshotSplitState
    -
     
    -
    setRecordsToSkip(long) - Method in class org.apache.fluss.flink.source.split.HybridSnapshotLogSplitState
    -
     
    -
    setScanPartitionDiscoveryIntervalMs(long) - Method in class org.apache.fluss.flink.source.FlussSourceBuilder
    -
    -
    Sets the scan partition discovery interval in milliseconds.
    -
    -
    setSendLatencyInMs(long) - Method in class org.apache.fluss.client.metrics.WriterMetricGroup
    -
     
    -
    setSerializationSchema(FlussSerializationSchema<InputT>) - Method in class org.apache.fluss.flink.sink.FlussSinkBuilder
    -
    -
    Set a FlussSerializationSchema.
    -
    -
    setShuffleByBucketId(boolean) - Method in class org.apache.fluss.flink.sink.FlussSinkBuilder
    -
    -
    Set shuffle by bucket id.
    -
    -
    setStartingOffsets(OffsetsInitializer) - Method in class org.apache.fluss.flink.source.FlussSourceBuilder
    -
    -
    Sets the starting offsets strategy for the Fluss source.
    -
    -
    setTable(String) - Method in class org.apache.fluss.flink.sink.FlussSinkBuilder
    -
    -
    Set the table name for the sink.
    -
    -
    setTable(String) - Method in class org.apache.fluss.flink.source.FlussSourceBuilder
    -
    -
    Sets the table name for the Fluss source.
    -
    -
    setTableBucketStoppingOffset(TableBucket, long) - Method in class org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds
    -
     
    -
    setup(int) - Method in interface org.apache.fluss.flink.sink.ChannelComputer
    -
     
    -
    setup(int) - Method in class org.apache.fluss.flink.sink.FlinkRowDataChannelComputer
    -
     
    -
    setup(int) - Method in class org.apache.fluss.flink.sink.FlinkStreamPartitioner
    -
     
    -
    setup(StreamTask<?, ?>, StreamConfig, Output<StreamRecord<CommittableMessage<Committable>>>) - Method in class org.apache.fluss.flink.tiering.committer.TieringCommitOperator
    -
     
    -
    setWriterState(long, int) - Method in class org.apache.fluss.client.write.ArrowLogWriteBatch
    -
     
    -
    setWriterState(long, int) - Method in class org.apache.fluss.client.write.IndexedLogWriteBatch
    -
     
    -
    setWriterState(long, int) - Method in class org.apache.fluss.client.write.KvWriteBatch
    -
     
    -
    setWriterState(long, int) - Method in class org.apache.fluss.client.write.WriteBatch
    -
     
    -
    SingleThreadFetcherManagerAdapter<E,​SplitT extends org.apache.flink.api.connector.source.SourceSplit> - Class in org.apache.fluss.flink.adapter
    -
    -
    Adapter for SingleThreadFetcherManager.TODO: remove it until not supported in flink 1.18.
    -
    -
    SingleThreadFetcherManagerAdapter(FutureCompletingBlockingQueue<RecordsWithSplitIds<E>>, Supplier<SplitReader<E, SplitT>>, Configuration, Consumer<Collection<String>>) - Constructor for class org.apache.fluss.flink.adapter.SingleThreadFetcherManagerAdapter
    -
     
    -
    SingleThreadMultiplexSourceReaderBaseAdapter<E,​T,​SplitT extends org.apache.flink.api.connector.source.SourceSplit,​SplitStateT> - Class in org.apache.fluss.flink.adapter
    -
    -
    Adapter for SingleThreadMultiplexSourceReaderBase.TODO: remove it until not supported in - flink 1.18.
    -
    -
    SingleThreadMultiplexSourceReaderBaseAdapter(FutureCompletingBlockingQueue<RecordsWithSplitIds<E>>, SingleThreadFetcherManager<E, SplitT>, RecordEmitter<E, T, SplitStateT>, Configuration, SourceReaderContext) - Constructor for class org.apache.fluss.flink.adapter.SingleThreadMultiplexSourceReaderBaseAdapter
    -
     
    -
    SINK_BUCKET_SHUFFLE - Static variable in class org.apache.fluss.flink.FlinkConnectorOptions
    -
     
    -
    SINK_IGNORE_DELETE - Static variable in class org.apache.fluss.flink.FlinkConnectorOptions
    -
     
    -
    snapshotId - Variable in class org.apache.fluss.flink.source.split.SnapshotSplit
    -
    -
    The snapshot id.
    -
    -
    SnapshotOffsetsInitializer - Class in org.apache.fluss.flink.source.enumerator.initializer
    -
    -
    An implementation of OffsetsInitializer which initializes the offsets.
    -
    -
    SnapshotOffsetsInitializer() - Constructor for class org.apache.fluss.flink.source.enumerator.initializer.SnapshotOffsetsInitializer
    -
     
    -
    SnapshotSplit - Class in org.apache.fluss.flink.source.split
    -
    -
    The split for snapshot.
    -
    -
    SnapshotSplit(TableBucket, String, long) - Constructor for class org.apache.fluss.flink.source.split.SnapshotSplit
    -
     
    -
    SnapshotSplit(TableBucket, String, long, long) - Constructor for class org.apache.fluss.flink.source.split.SnapshotSplit
    -
     
    -
    snapshotState(long) - Method in class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
    -
     
    -
    snapshotState(long) - Method in class org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator
    -
     
    -
    snapshotState(long) - Method in class org.apache.fluss.flink.tiering.source.TieringSourceReader
    -
     
    -
    SourceEnumeratorState - Class in org.apache.fluss.flink.source.state
    -
    -
    A checkpoint of the current state of the containing the buckets that is already assigned.
    -
    -
    SourceEnumeratorState(Set<TableBucket>, Map<Long, String>, List<SourceSplitBase>) - Constructor for class org.apache.fluss.flink.source.state.SourceEnumeratorState
    -
     
    -
    SourceSplitBase - Class in org.apache.fluss.flink.source.split
    -
    -
    A base source split for SnapshotSplit and LogSplit.
    -
    -
    SourceSplitBase(TableBucket, String) - Constructor for class org.apache.fluss.flink.source.split.SourceSplitBase
    -
     
    -
    SourceSplitSerializer - Class in org.apache.fluss.flink.source.split
    -
    -
    A serializer for the SourceSplitBase.
    -
    -
    SourceSplitSerializer(LakeSource<LakeSplit>) - Constructor for class org.apache.fluss.flink.source.split.SourceSplitSerializer
    -
     
    -
    SourceSplitState - Class in org.apache.fluss.flink.source.split
    -
    -
    State of the reader, essentially a mutable version of the SourceSplitBase.
    -
    -
    SourceSplitState(SourceSplitBase) - Constructor for class org.apache.fluss.flink.source.split.SourceSplitState
    -
     
    -
    split - Variable in class org.apache.fluss.flink.source.split.SourceSplitState
    -
     
    -
    splitId() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
    -
     
    -
    splitId() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotSplit
    -
     
    -
    splitId() - Method in class org.apache.fluss.flink.source.split.HybridSnapshotLogSplit
    -
     
    -
    splitId() - Method in class org.apache.fluss.flink.source.split.LogSplit
    -
     
    -
    splitId() - Method in class org.apache.fluss.flink.tiering.source.split.TieringLogSplit
    -
     
    -
    splitId() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSnapshotSplit
    -
     
    -
    splitKind() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
    -
     
    -
    splitKind() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotSplit
    -
     
    -
    splitKind() - Method in class org.apache.fluss.flink.source.split.LogSplit
    -
     
    -
    splitKind() - Method in class org.apache.fluss.flink.source.split.SourceSplitBase
    -
     
    -
    splitKind() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
    -
     
    -
    SplitScanner - Interface in org.apache.fluss.flink.source.reader
    -
    -
    A scanner to scan split to records .
    -
    -
    SQL - org.apache.fluss.flink.utils.TimestampFormat
    -
    -
    Options to specify TIMESTAMP/TIMESTAMP_WITH_LOCAL_ZONE format.
    -
    -
    start() - Method in class org.apache.fluss.client.table.scanner.log.RemoteLogDownloader
    -
     
    -
    start() - Method in class org.apache.fluss.client.token.DefaultSecurityTokenManager
    -
     
    -
    start() - Method in interface org.apache.fluss.client.token.SecurityTokenManager
    -
    -
    Creates a re-occurring task which obtains new tokens.
    -
    -
    start() - Method in class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
    -
     
    -
    start() - Method in class org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator
    -
     
    -
    start() - Method in class org.apache.fluss.flink.tiering.source.TieringSourceReader
    -
     
    -
    startupMode - Variable in class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils.StartupOptions
    -
     
    -
    StartupOptions() - Constructor for class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils.StartupOptions
    -
     
    -
    startupTimestampMs - Variable in class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils.StartupOptions
    -
     
    -
    StickyBucketAssigner - Class in org.apache.fluss.client.write
    -
    -
    The bucket assigner with sticky strategy.
    -
    -
    StickyBucketAssigner(PhysicalTablePath) - Constructor for class org.apache.fluss.client.write.StickyBucketAssigner
    -
     
    -
    stop() - Method in class org.apache.fluss.client.token.DefaultSecurityTokenManager
    -
    -
    Stops re-occurring token obtain task.
    -
    -
    stop() - Method in interface org.apache.fluss.client.token.SecurityTokenManager
    -
    -
    Stops re-occurring token obtain task.
    -
    -
    subscribe(int, long) - Method in interface org.apache.fluss.client.table.scanner.log.LogScanner
    -
    -
    Subscribe to the given table bucket in given offset dynamically.
    -
    -
    subscribe(int, long) - Method in class org.apache.fluss.client.table.scanner.log.LogScannerImpl
    -
     
    -
    subscribe(long, int, long) - Method in interface org.apache.fluss.client.table.scanner.log.LogScanner
    -
    -
    Subscribe to the given partitioned table bucket in given offset dynamically.
    -
    -
    subscribe(long, int, long) - Method in class org.apache.fluss.client.table.scanner.log.LogScannerImpl
    -
     
    -
    subscribeFromBeginning(int) - Method in interface org.apache.fluss.client.table.scanner.log.LogScanner
    -
    -
    Subscribe to the given table buckets from beginning dynamically.
    -
    -
    subscribeFromBeginning(long, int) - Method in interface org.apache.fluss.client.table.scanner.log.LogScanner
    -
    -
    Subscribe to the given partitioned table bucket from beginning dynamically.
    -
    -
    supportsNestedProjection() - Method in class org.apache.fluss.flink.source.FlinkTableSource
    -
     
    -
    -A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-2.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-2.html deleted file mode 100644 index 425fda4652..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-2.html +++ /dev/null @@ -1,248 +0,0 @@ - - - - - -B-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages - - -

    B

    -
    -
    batchIsFull - Variable in class org.apache.fluss.client.write.RecordAccumulator.RecordAppendResult
    -
     
    -
    BatchScanner - Interface in org.apache.fluss.client.table.scanner.batch
    -
    -
    The scanner that reads records form a table in a batch fashion.
    -
    -
    BatchScanUtils - Class in org.apache.fluss.client.table.scanner.batch
    -
    -
    Utility class for batch scan.
    -
    -
    BatchScanUtils() - Constructor for class org.apache.fluss.client.table.scanner.batch.BatchScanUtils
    -
     
    -
    batchSequence() - Method in class org.apache.fluss.client.write.ArrowLogWriteBatch
    -
     
    -
    batchSequence() - Method in class org.apache.fluss.client.write.IndexedLogWriteBatch
    -
     
    -
    batchSequence() - Method in class org.apache.fluss.client.write.KvWriteBatch
    -
     
    -
    batchSequence() - Method in class org.apache.fluss.client.write.WriteBatch
    -
     
    -
    beginFlush() - Method in class org.apache.fluss.client.write.RecordAccumulator
    -
    -
    Initiate the flushing of data from the accumulator...this makes all requests immediately - ready.
    -
    -
    binding() - Method in class org.apache.fluss.client.admin.DropAclsResult.FilterResult
    -
    -
    Return the deleted ACL binding or null if there was an error.
    -
    -
    BOOTSTRAP_SERVERS - Static variable in class org.apache.fluss.flink.FlinkConnectorOptions
    -
     
    -
    bootstrapServers - Variable in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    BoundedSplitReader - Class in org.apache.fluss.flink.source.reader
    -
    -
    A bounded reader to reading Fluss's bounded split (e.g., SnapshotSplit) into RecordAndPoss.
    -
    -
    BoundedSplitReader(BatchScanner, long) - Constructor for class org.apache.fluss.flink.source.reader.BoundedSplitReader
    -
     
    -
    BUCKET_GROUP - Static variable in class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
    -
     
    -
    BUCKET_KEY - Static variable in class org.apache.fluss.flink.FlinkConnectorOptions
    -
     
    -
    BUCKET_NUMBER - Static variable in class org.apache.fluss.flink.FlinkConnectorOptions
    -
     
    -
    bucketId() - Method in class org.apache.fluss.client.write.WriteBatch
    -
     
    -
    BucketOffsetsRetrieverImpl - Class in org.apache.fluss.flink.source.enumerator.initializer
    -
    -
    The default implementation for offsets retriever.
    -
    -
    BucketOffsetsRetrieverImpl(Admin, TablePath) - Constructor for class org.apache.fluss.flink.source.enumerator.initializer.BucketOffsetsRetrieverImpl
    -
     
    -
    bucketResult(int) - Method in class org.apache.fluss.client.admin.ListOffsetsResult
    -
     
    -
    buckets() - Method in class org.apache.fluss.client.table.scanner.log.ScanRecords
    -
    -
    Get the bucket ids which have records contained in this record set.
    -
    -
    build() - Method in class org.apache.fluss.client.write.ArrowLogWriteBatch
    -
     
    -
    build() - Method in class org.apache.fluss.client.write.IndexedLogWriteBatch
    -
     
    -
    build() - Method in class org.apache.fluss.client.write.KvWriteBatch
    -
     
    -
    build() - Method in class org.apache.fluss.client.write.WriteBatch
    -
    -
    Gets the memory segment bytes view of the batch.
    -
    -
    build() - Method in class org.apache.fluss.flink.sink.FlussSinkBuilder
    -
    -
    Build the FlussSink.
    -
    -
    build() - Method in class org.apache.fluss.flink.source.FlussSourceBuilder
    -
    -
    Builds and returns a new FlussSource instance with the configured properties.
    -
    -
    build() - Method in class org.apache.fluss.flink.tiering.LakeTieringJobBuilder
    -
     
    -
    build() - Method in class org.apache.fluss.flink.tiering.source.TieringSource.Builder
    -
     
    -
    builder() - Static method in class org.apache.fluss.flink.sink.FlussSink
    -
    -
    Creates a new FlussSinkBuilder instance for building a FlussSink.
    -
    -
    builder() - Static method in class org.apache.fluss.flink.source.FlussSource
    -
    -
    Get a FlussSourceBuilder to build a FlussSource.
    -
    -
    Builder(Configuration, LakeTieringFactory<WriteResult, ?>) - Constructor for class org.apache.fluss.flink.tiering.source.TieringSource.Builder
    -
     
    -
    bytesPerBatch() - Method in class org.apache.fluss.client.metrics.WriterMetricGroup
    -
     
    -
    bytesPerRequest() - Method in class org.apache.fluss.client.metrics.ScannerMetricGroup
    -
     
    -
    bytesSendTotal() - Method in class org.apache.fluss.client.metrics.WriterMetricGroup
    -
     
    -
    -A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-20.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-20.html deleted file mode 100644 index 03b3777a58..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-20.html +++ /dev/null @@ -1,547 +0,0 @@ - - - - - -T-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages - - -

    T

    -
    -
    table - Variable in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
    -
     
    -
    Table - Interface in org.apache.fluss.client.table
    -
    -
    Used to communicate with a single Fluss table.
    -
    -
    TABLE_OPTIONS - Static variable in class org.apache.fluss.flink.FlinkConnectorOptions
    -
     
    -
    TableAppend - Class in org.apache.fluss.client.table.writer
    -
    -
    API for configuring and creating AppendWriter.
    -
    -
    TableAppend(TablePath, TableInfo, WriterClient) - Constructor for class org.apache.fluss.client.table.writer.TableAppend
    -
     
    -
    tableBucket - Variable in class org.apache.fluss.flink.source.split.SourceSplitBase
    -
     
    -
    tableBucket - Variable in class org.apache.fluss.flink.tiering.source.split.TieringSplit
    -
     
    -
    tableBucket() - Method in class org.apache.fluss.client.lookup.AbstractLookupBatch
    -
     
    -
    tableBucket() - Method in class org.apache.fluss.client.lookup.AbstractLookupQuery
    -
     
    -
    tableBucket() - Method in class org.apache.fluss.client.lookup.LookupBatch
    -
     
    -
    tableBucket() - Method in class org.apache.fluss.client.lookup.PrefixLookupBatch
    -
     
    -
    tableBucket() - Method in class org.apache.fluss.client.write.ReadyWriteBatch
    -
     
    -
    tableBucket() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResult
    -
     
    -
    tableBucket() - Method in class org.apache.fluss.flink.tiering.source.TieringWriterInitContext
    -
     
    -
    TableBucketWriteResult<WriteResult> - Class in org.apache.fluss.flink.tiering.source
    -
    -
    This class contains the TableBucketWriteResult of LakeWriter, the table path and the bucket - that the write result is for, the end log offset of tiering, the total number of write results in - one round of tiering.
    -
    -
    TableBucketWriteResult(TablePath, TableBucket, String, WriteResult, long, int) - Constructor for class org.apache.fluss.flink.tiering.source.TableBucketWriteResult
    -
     
    -
    TableBucketWriteResultEmitter<WriteResult> - Class in org.apache.fluss.flink.tiering.source
    -
    -
    The emitter to emit TableBucketWriteResult to downstream LakeCommitter.
    -
    -
    TableBucketWriteResultEmitter() - Constructor for class org.apache.fluss.flink.tiering.source.TableBucketWriteResultEmitter
    -
     
    -
    TableBucketWriteResultSerializer<WriteResult> - Class in org.apache.fluss.flink.tiering.source
    -
    -
    The serializer for TableBucketWriteResult.
    -
    -
    TableBucketWriteResultSerializer(SimpleVersionedSerializer<WriteResult>) - Constructor for class org.apache.fluss.flink.tiering.source.TableBucketWriteResultSerializer
    -
     
    -
    TableBucketWriteResultTypeInfo<WriteResult> - Class in org.apache.fluss.flink.tiering.source
    -
    -
    A TypeInformation for TableBucketWriteResult .
    -
    -
    tableExists(ObjectPath) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    tableExists(TablePath) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    Get whether table exists asynchronously.
    -
    -
    tableExists(TablePath) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    tableInfo() - Method in class org.apache.fluss.flink.tiering.source.TieringWriterInitContext
    -
     
    -
    TableLookup - Class in org.apache.fluss.client.lookup
    -
    -
    API for configuring and creating Lookuper.
    -
    -
    TableLookup(TableInfo, MetadataUpdater, LookupClient) - Constructor for class org.apache.fluss.client.lookup.TableLookup
    -
     
    -
    tablePath - Variable in class org.apache.fluss.client.table.writer.AbstractTableWriter
    -
     
    -
    tablePath - Variable in class org.apache.fluss.flink.tiering.source.split.TieringSplit
    -
     
    -
    tablePath() - Method in class org.apache.fluss.flink.tiering.committer.TieringCommitterInitContext
    -
     
    -
    tablePath() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResult
    -
     
    -
    tablePath() - Method in class org.apache.fluss.flink.tiering.source.TieringWriterInitContext
    -
     
    -
    tableRowType - Variable in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
    -
     
    -
    TableScan - Class in org.apache.fluss.client.table.scanner
    -
    -
    API for configuring and creating LogScanner and BatchScanner.
    -
    -
    TableScan(FlussConnection, TableInfo) - Constructor for class org.apache.fluss.client.table.scanner.TableScan
    -
     
    -
    TableUpsert - Class in org.apache.fluss.client.table.writer
    -
    -
    API for configuring and creating UpsertWriter.
    -
    -
    TableUpsert(TablePath, TableInfo, WriterClient) - Constructor for class org.apache.fluss.client.table.writer.TableUpsert
    -
     
    -
    TableWriter - Interface in org.apache.fluss.client.table.writer
    -
    -
    A writer to write data to a Fluss table.
    -
    -
    targetColumnIndexes - Variable in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
    -
     
    -
    TIERING_LOG_SPLIT_FLAG - Static variable in class org.apache.fluss.flink.tiering.source.split.TieringSplit
    -
     
    -
    TIERING_SNAPSHOT_SPLIT_FLAG - Static variable in class org.apache.fluss.flink.tiering.source.split.TieringSplit
    -
     
    -
    TIERING_SOURCE_OPERATOR_UID - Static variable in class org.apache.fluss.flink.tiering.source.TieringSource
    -
     
    -
    TIERING_SOURCE_TRANSFORMATION_UID - Static variable in class org.apache.fluss.flink.tiering.source.TieringSource
    -
     
    -
    TieringCommitOperator<WriteResult,​Committable> - Class in org.apache.fluss.flink.tiering.committer
    -
    -
    A Flink operator to aggregate TieringCommitOperators by table to TieringCommitOperator which will - then be committed to lake & Fluss cluster.
    -
    -
    TieringCommitOperator(StreamOperatorParameters<CommittableMessage<Committable>>, Configuration, LakeTieringFactory<WriteResult, Committable>) - Constructor for class org.apache.fluss.flink.tiering.committer.TieringCommitOperator
    -
     
    -
    TieringCommitOperatorFactory<WriteResult,​Committable> - Class in org.apache.fluss.flink.tiering.committer
    -
    -
    The factory to create TieringCommitOperator.
    -
    -
    TieringCommitOperatorFactory(Configuration, LakeTieringFactory<WriteResult, Committable>) - Constructor for class org.apache.fluss.flink.tiering.committer.TieringCommitOperatorFactory
    -
     
    -
    TieringCommitterInitContext - Class in org.apache.fluss.flink.tiering.committer
    -
    -
    The CommitterInitContext implementation for LakeCommitter.
    -
    -
    TieringCommitterInitContext(TablePath) - Constructor for class org.apache.fluss.flink.tiering.committer.TieringCommitterInitContext
    -
     
    -
    TieringLogSplit - Class in org.apache.fluss.flink.tiering.source.split
    -
    -
    The table split for tiering service.
    -
    -
    TieringLogSplit(TablePath, TableBucket, String, long, long) - Constructor for class org.apache.fluss.flink.tiering.source.split.TieringLogSplit
    -
     
    -
    TieringLogSplit(TablePath, TableBucket, String, long, long, int) - Constructor for class org.apache.fluss.flink.tiering.source.split.TieringLogSplit
    -
     
    -
    TieringRestoreEvent - Class in org.apache.fluss.flink.tiering.event
    -
    -
    SourceEvent used to represent tiering is restoring.
    -
    -
    TieringRestoreEvent() - Constructor for class org.apache.fluss.flink.tiering.event.TieringRestoreEvent
    -
     
    -
    TieringSnapshotSplit - Class in org.apache.fluss.flink.tiering.source.split
    -
    -
    The table split for tiering service.
    -
    -
    TieringSnapshotSplit(TablePath, TableBucket, String, long, long) - Constructor for class org.apache.fluss.flink.tiering.source.split.TieringSnapshotSplit
    -
     
    -
    TieringSnapshotSplit(TablePath, TableBucket, String, long, long, int) - Constructor for class org.apache.fluss.flink.tiering.source.split.TieringSnapshotSplit
    -
     
    -
    TieringSource<WriteResult> - Class in org.apache.fluss.flink.tiering.source
    -
    -
    The flink source implementation for tiering data from Fluss to downstream lake.
    -
    -
    TieringSource(Configuration, LakeTieringFactory<WriteResult, ?>, long) - Constructor for class org.apache.fluss.flink.tiering.source.TieringSource
    -
     
    -
    TieringSource.Builder<WriteResult> - Class in org.apache.fluss.flink.tiering.source
    -
    -
    Builder for TieringSource.
    -
    -
    TieringSourceEnumerator - Class in org.apache.fluss.flink.tiering.source.enumerator
    -
    -
    An implementation of SplitEnumerator used to request TieringSplit from Fluss - Cluster.
    -
    -
    TieringSourceEnumerator(Configuration, SplitEnumeratorContext<TieringSplit>, long) - Constructor for class org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator
    -
     
    -
    TieringSourceEnumeratorState - Class in org.apache.fluss.flink.tiering.source.state
    -
    -
    The marker class of stateless component TieringSourceEnumerator.
    -
    -
    TieringSourceEnumeratorState() - Constructor for class org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorState
    -
     
    -
    TieringSourceEnumeratorStateSerializer - Class in org.apache.fluss.flink.tiering.source.state
    -
    - -
    -
    TieringSourceEnumeratorStateSerializer() - Constructor for class org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorStateSerializer
    -
     
    -
    TieringSourceOptions - Class in org.apache.fluss.flink.tiering.source
    -
    -
    Configuration options for the TieringSource.
    -
    -
    TieringSourceOptions() - Constructor for class org.apache.fluss.flink.tiering.source.TieringSourceOptions
    -
     
    -
    TieringSourceReader<WriteResult> - Class in org.apache.fluss.flink.tiering.source
    -
    -
    A SourceReader that read records from Fluss and write to lake.
    -
    -
    TieringSourceReader(SourceReaderContext, Configuration, LakeTieringFactory<WriteResult, ?>) - Constructor for class org.apache.fluss.flink.tiering.source.TieringSourceReader
    -
     
    -
    tieringSplit - Variable in class org.apache.fluss.flink.tiering.source.state.TieringSplitState
    -
     
    -
    TieringSplit - Class in org.apache.fluss.flink.tiering.source.split
    -
    -
    The base table split for tiering service.
    -
    -
    TieringSplit(TablePath, TableBucket, String, int) - Constructor for class org.apache.fluss.flink.tiering.source.split.TieringSplit
    -
     
    -
    TieringSplitGenerator - Class in org.apache.fluss.flink.tiering.source.split
    -
    -
    A generator for lake splits.
    -
    -
    TieringSplitGenerator(Admin) - Constructor for class org.apache.fluss.flink.tiering.source.split.TieringSplitGenerator
    -
     
    -
    TieringSplitReader<WriteResult> - Class in org.apache.fluss.flink.tiering.source
    -
    -
    The SplitReader implementation which will read Fluss and write to lake.
    -
    -
    TieringSplitReader(Configuration, LakeTieringFactory<WriteResult, ?>) - Constructor for class org.apache.fluss.flink.tiering.source.TieringSplitReader
    -
     
    -
    TieringSplitSerializer - Class in org.apache.fluss.flink.tiering.source.split
    -
    -
    A serializer for the TieringSplit.
    -
    -
    TieringSplitSerializer() - Constructor for class org.apache.fluss.flink.tiering.source.split.TieringSplitSerializer
    -
     
    -
    TieringSplitState - Class in org.apache.fluss.flink.tiering.source.state
    -
    -
    The state of a TieringSplit.
    -
    -
    TieringSplitState(TieringSplit) - Constructor for class org.apache.fluss.flink.tiering.source.state.TieringSplitState
    -
     
    -
    TieringWriterInitContext - Class in org.apache.fluss.flink.tiering.source
    -
    -
    The implementation of WriterInitContext.
    -
    -
    TieringWriterInitContext(TablePath, TableBucket, String, TableInfo) - Constructor for class org.apache.fluss.flink.tiering.source.TieringWriterInitContext
    -
     
    -
    timestamp() - Method in class org.apache.fluss.client.table.scanner.ScanRecord
    -
     
    -
    timestamp(long) - Static method in interface org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer
    -
    -
    Get an OffsetsInitializer which initializes the offsets in each bucket so that the - initialized offset is the offset of the first record batch whose commit timestamp is greater - than or equals the given timestamp (milliseconds).
    -
    -
    TIMESTAMP - org.apache.fluss.flink.FlinkConnectorOptions.ScanStartupMode
    -
     
    -
    TimestampFormat - Enum in org.apache.fluss.flink.utils
    -
    -
    Timestamp format Enums.
    -
    -
    TimestampOffsetsInitializer - Class in org.apache.fluss.flink.source.enumerator.initializer
    -
    -
    An implementation of OffsetsInitializer to initialize the offsets based on a timestamp.
    -
    -
    TimestampOffsetsInitializer(long) - Constructor for class org.apache.fluss.flink.source.enumerator.initializer.TimestampOffsetsInitializer
    -
     
    -
    TimestampSpec(long) - Constructor for class org.apache.fluss.client.admin.OffsetSpec.TimestampSpec
    -
     
    -
    toBucketOffsetsProperty(Map<TableBucket, Long>, Map<Long, String>, List<String>) - Static method in class org.apache.fluss.flink.tiering.committer.TieringCommitOperator
    -
     
    -
    toCatalogTable(Schema, String, List<String>, Map<String, String>) - Static method in class org.apache.fluss.flink.adapter.CatalogTableAdapter
    -
     
    -
    toFlinkOption(ConfigOption<T>) - Static method in class org.apache.fluss.flink.utils.FlinkConversions
    -
    -
    Convert Fluss's ConfigOption to Flink's ConfigOption.
    -
    -
    toFlinkOptions(Collection<ConfigOption<?>>) - Static method in class org.apache.fluss.flink.utils.FlinkConversions
    -
    -
    Convert Fluss's ConfigOptions to Flink's ConfigOptions.
    -
    -
    toFlinkRowData(LogRecord) - Method in class org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter
    -
     
    -
    toFlinkRowData(InternalRow) - Method in class org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter
    -
     
    -
    toFlinkRowKind(ChangeType) - Static method in class org.apache.fluss.flink.utils.FlinkConversions
    -
     
    -
    toFlinkRowType(RowType) - Static method in class org.apache.fluss.flink.utils.FlinkConversions
    -
    -
    Convert Fluss's RowType to Flink's RowType.
    -
    -
    toFlinkTable(TableInfo) - Static method in class org.apache.fluss.flink.utils.FlinkConversions
    -
    -
    Convert Fluss's table to Flink's table.
    -
    -
    toFlinkType(DataType) - Static method in class org.apache.fluss.flink.utils.FlinkConversions
    -
    -
    Convert Fluss's type to Flink's type.
    -
    -
    toFlussDatabase(CatalogDatabase) - Static method in class org.apache.fluss.flink.utils.FlinkConversions
    -
    -
    Convert Flink's table to Fluss's database.
    -
    -
    toFlussRowType(RowType) - Static method in class org.apache.fluss.flink.utils.FlinkConversions
    -
    -
    Convert Flink's RowType to Fluss' RowType.
    -
    -
    toFlussTable(ResolvedCatalogTable) - Static method in class org.apache.fluss.flink.utils.FlinkConversions
    -
    -
    Convert Flink's table to Fluss's table.
    -
    -
    toFlussType(DataType) - Static method in class org.apache.fluss.flink.utils.FlinkConversions
    -
    -
    Convert Flink's physical type to Fluss' type.
    -
    -
    toFsPathAndFileName(List<PbRemotePathAndLocalFile>) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
    -
     
    -
    toInternalRow(RowData) - Method in class org.apache.fluss.flink.utils.FlinkRowToFlussRowConverter
    -
    -
    Deprecated.
    -
    toKeyValueMap(List<PbKeyValue>) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
    -
     
    -
    toKvSnapshotMetadata(GetKvSnapshotMetadataResponse) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
    -
     
    -
    toKvSnapshots(GetLatestKvSnapshotsResponse) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
    -
     
    -
    toLakeTableSnapshotInfo(GetLatestLakeSnapshotResponse) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
    -
     
    -
    toList(CloseableIterator<T>) - Static method in class org.apache.fluss.client.table.scanner.batch.BatchScanUtils
    -
    -
    Collect all rows from the iterator and close it.
    -
    -
    toPartitionInfos(ListPartitionInfosResponse) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
    -
     
    -
    toSecurityToken(GetFileSystemSecurityTokenResponse) - Static method in class org.apache.fluss.client.utils.ClientRpcMessageUtils
    -
     
    -
    toSourceSplit() - Method in class org.apache.fluss.flink.lake.state.LakeSnapshotAndFlussLogSplitState
    -
     
    -
    toSourceSplit() - Method in class org.apache.fluss.flink.lake.state.LakeSnapshotSplitState
    -
     
    -
    toSourceSplit() - Method in class org.apache.fluss.flink.source.split.HybridSnapshotLogSplitState
    -
     
    -
    toSourceSplit() - Method in class org.apache.fluss.flink.source.split.LogSplitState
    -
     
    -
    toSourceSplit() - Method in class org.apache.fluss.flink.source.split.SourceSplitState
    -
     
    -
    toSourceSplit() - Method in class org.apache.fluss.flink.tiering.source.state.TieringSplitState
    -
     
    -
    toSplitId(String, TableBucket) - Static method in class org.apache.fluss.flink.source.split.SourceSplitBase
    -
     
    -
    toSplitId(String, TableBucket) - Static method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
    -
     
    -
    toSplitType(String, SourceSplitState) - Method in class org.apache.fluss.flink.source.reader.FlinkSourceReader
    -
     
    -
    toSplitType(String, TieringSplitState) - Method in class org.apache.fluss.flink.tiering.source.TieringSourceReader
    -
     
    -
    toString() - Method in class org.apache.fluss.client.lookup.LookupResult
    -
     
    -
    toString() - Method in class org.apache.fluss.client.metadata.KvSnapshotMetadata
    -
     
    -
    toString() - Method in class org.apache.fluss.client.metadata.LakeSnapshot
    -
     
    -
    toString() - Method in class org.apache.fluss.client.table.scanner.ScanRecord
    -
     
    -
    toString() - Method in enum org.apache.fluss.flink.FlinkConnectorOptions.ScanStartupMode
    -
     
    -
    toString() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
    -
     
    -
    toString() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotSplit
    -
     
    -
    toString() - Method in class org.apache.fluss.flink.sink.FlinkRowDataChannelComputer
    -
     
    -
    toString() - Method in class org.apache.fluss.flink.sink.FlinkStreamPartitioner
    -
     
    -
    toString() - Method in class org.apache.fluss.flink.source.event.PartitionBucketsUnsubscribedEvent
    -
     
    -
    toString() - Method in class org.apache.fluss.flink.source.event.PartitionsRemovedEvent
    -
     
    -
    toString() - Method in class org.apache.fluss.flink.source.reader.RecordAndPos
    -
     
    -
    toString() - Method in class org.apache.fluss.flink.source.split.HybridSnapshotLogSplit
    -
     
    -
    toString() - Method in class org.apache.fluss.flink.source.split.LogSplit
    -
     
    -
    toString() - Method in class org.apache.fluss.flink.source.split.SnapshotSplit
    -
     
    -
    toString() - Method in class org.apache.fluss.flink.source.state.SourceEnumeratorState
    -
     
    -
    toString() - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
    -
     
    -
    toString() - Method in class org.apache.fluss.flink.tiering.source.split.TieringLogSplit
    -
     
    -
    toString() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSnapshotSplit
    -
     
    -
    toString() - Method in class org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorState
    -
     
    -
    toString() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
    -
     
    -
    toTablePath(ObjectPath) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    transferAllToDirectory(List<FsPathAndFileName>, Path, CloseableRegistry) - Method in class org.apache.fluss.client.table.scanner.RemoteFileDownloader
    -
     
    -
    tryAppend(WriteRecord, WriteCallback) - Method in class org.apache.fluss.client.write.ArrowLogWriteBatch
    -
     
    -
    tryAppend(WriteRecord, WriteCallback) - Method in class org.apache.fluss.client.write.IndexedLogWriteBatch
    -
     
    -
    tryAppend(WriteRecord, WriteCallback) - Method in class org.apache.fluss.client.write.KvWriteBatch
    -
     
    -
    tryAppend(WriteRecord, WriteCallback) - Method in class org.apache.fluss.client.write.WriteBatch
    -
    -
    try to append one write record to the record batch.
    -
    -
    -A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-21.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-21.html deleted file mode 100644 index e8b47bdece..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-21.html +++ /dev/null @@ -1,222 +0,0 @@ - - - - - -U-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages - - -

    U

    -
    -
    UNINITIALIZED - Static variable in class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
    -
     
    -
    UNKNOWN_NUMBER_OF_SPLITS - Static variable in class org.apache.fluss.flink.tiering.source.split.TieringSplit
    -
     
    -
    unknownLeaderTables - Variable in class org.apache.fluss.client.write.RecordAccumulator.ReadyCheckResult
    -
     
    -
    unregister(Metric, String, AbstractMetricGroup) - Method in class org.apache.fluss.flink.metrics.FlinkMetricRegistry
    -
     
    -
    unsubscribe(long, int) - Method in interface org.apache.fluss.client.table.scanner.log.LogScanner
    -
    -
    Unsubscribe from the given bucket of given partition dynamically.
    -
    -
    unsubscribe(long, int) - Method in class org.apache.fluss.client.table.scanner.log.LogScannerImpl
    -
     
    -
    update(long) - Method in class org.apache.fluss.flink.metrics.FlinkHistogram
    -
     
    -
    updateEstimation(PhysicalTablePath, int) - Method in class org.apache.fluss.client.write.DynamicWriteBatchSizeEstimator
    -
     
    -
    updateFetchLatency(long) - Method in class org.apache.fluss.client.metrics.ScannerMetricGroup
    -
     
    -
    updateMetadata(Set<TablePath>, Collection<PhysicalTablePath>, Collection<Long>) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
    -
     
    -
    updatePhysicalTableMetadata(Set<PhysicalTablePath>) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
    -
    -
    Update the table or partition metadata info.
    -
    -
    updateTableOrPartitionMetadata(TablePath, Long) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
    -
     
    -
    updateWithCurrentLakeSplitIndex(int) - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
    -
     
    -
    updateWithLakeSplitFinished(boolean) - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
    -
     
    -
    updateWithRecordsToSkip(long) - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
    -
     
    -
    updateWithStartingOffset(long) - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
    -
     
    -
    upsert(InternalRow) - Method in interface org.apache.fluss.client.table.writer.UpsertWriter
    -
    -
    Inserts row into Fluss table if they do not already exist, or updates them if they do exist.
    -
    -
    Upsert - Interface in org.apache.fluss.client.table.writer
    -
    -
    Used to configure and create UpsertWriter to upsert and delete data to a Primary Key - Table.
    -
    -
    UPSERT - org.apache.fluss.flink.row.OperationType
    -
    -
    Represents an upsert operation (update or insert).
    -
    -
    UpsertResult - Class in org.apache.fluss.client.table.writer
    -
    -
    The result of upserting a record (UpsertWriter.upsert(InternalRow)).
    -
    -
    UpsertResult() - Constructor for class org.apache.fluss.client.table.writer.UpsertResult
    -
     
    -
    UpsertSinkWriter<InputT> - Class in org.apache.fluss.flink.sink.writer
    -
    -
    An upsert sink writer or fluss primary key table.
    -
    -
    UpsertSinkWriter(TablePath, Configuration, RowType, int[], MailboxExecutor, FlussSerializationSchema<InputT>) - Constructor for class org.apache.fluss.flink.sink.writer.UpsertSinkWriter
    -
     
    -
    UpsertWrite - Class in org.apache.fluss.client.table.writer
    -
    -
    Used to describe the operation to write data by UpsertWriter to a table.
    -
    -
    UpsertWrite() - Constructor for class org.apache.fluss.client.table.writer.UpsertWrite
    -
     
    -
    UpsertWriter - Interface in org.apache.fluss.client.table.writer
    -
    -
    The writer to write data to the primary key table.
    -
    -
    -A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-22.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-22.html deleted file mode 100644 index fffe6643a3..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-22.html +++ /dev/null @@ -1,222 +0,0 @@ - - - - - -V-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages - - -

    V

    -
    -
    validateAndCreateLookupNormalizer(int[][], int[], int[], int[], RowType, int[]) - Static method in class org.apache.fluss.flink.source.lookup.LookupNormalizer
    -
    -
    Validate the lookup key indexes and primary keys, and create a LookupNormalizer.
    -
    -
    validateTableSourceOptions(ReadableConfig) - Static method in class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils
    -
     
    -
    valueOf(String) - Static method in enum org.apache.fluss.client.lookup.LookupType
    -
    -
    Returns the enum constant of this type with the specified name.
    -
    -
    valueOf(String) - Static method in enum org.apache.fluss.client.write.WriteFormat
    -
    -
    Returns the enum constant of this type with the specified name.
    -
    -
    valueOf(String) - Static method in enum org.apache.fluss.flink.FlinkConnectorOptions.ScanStartupMode
    -
    -
    Returns the enum constant of this type with the specified name.
    -
    -
    valueOf(String) - Static method in enum org.apache.fluss.flink.row.OperationType
    -
    -
    Returns the enum constant of this type with the specified name.
    -
    -
    valueOf(String) - Static method in enum org.apache.fluss.flink.utils.PushdownUtils.ValueConversion
    -
    -
    Returns the enum constant of this type with the specified name.
    -
    -
    valueOf(String) - Static method in enum org.apache.fluss.flink.utils.TimestampFormat
    -
    -
    Returns the enum constant of this type with the specified name.
    -
    -
    valueRow() - Method in class org.apache.fluss.flink.lake.reader.KeyValueRow
    -
     
    -
    values() - Method in class org.apache.fluss.client.admin.DropAclsResult.FilterResults
    -
    -
    Return a list of delete ACLs results for a given filter.
    -
    -
    values() - Method in class org.apache.fluss.client.admin.DropAclsResult
    -
    -
    Return a map from acl filters to futures which can be used to check the status of the - deletions by each filter.
    -
    -
    values() - Static method in enum org.apache.fluss.client.lookup.LookupType
    -
    -
    Returns an array containing the constants of this enum type, in -the order they are declared.
    -
    -
    values() - Static method in enum org.apache.fluss.client.write.WriteFormat
    -
    -
    Returns an array containing the constants of this enum type, in -the order they are declared.
    -
    -
    values() - Static method in enum org.apache.fluss.flink.FlinkConnectorOptions.ScanStartupMode
    -
    -
    Returns an array containing the constants of this enum type, in -the order they are declared.
    -
    -
    values() - Static method in enum org.apache.fluss.flink.row.OperationType
    -
    -
    Returns an array containing the constants of this enum type, in -the order they are declared.
    -
    -
    values() - Static method in enum org.apache.fluss.flink.utils.PushdownUtils.ValueConversion
    -
    -
    Returns an array containing the constants of this enum type, in -the order they are declared.
    -
    -
    values() - Static method in enum org.apache.fluss.flink.utils.TimestampFormat
    -
    -
    Returns an array containing the constants of this enum type, in -the order they are declared.
    -
    -
    -A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-23.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-23.html deleted file mode 100644 index 1c5ac6f1d0..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-23.html +++ /dev/null @@ -1,232 +0,0 @@ - - - - - -W-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages - - -

    W

    -
    -
    waitedTimeMs(long) - Method in class org.apache.fluss.client.write.WriteBatch
    -
     
    -
    wakeup() - Method in class org.apache.fluss.client.table.scanner.log.LogFetchBuffer
    -
     
    -
    wakeup() - Method in class org.apache.fluss.client.table.scanner.log.LogFetcher
    -
     
    -
    wakeup() - Method in interface org.apache.fluss.client.table.scanner.log.LogScanner
    -
    -
    Wake up the log scanner in case the fetcher thread in log scanner is blocking in LogScanner.poll(Duration timeout).
    -
    -
    wakeup() - Method in class org.apache.fluss.client.table.scanner.log.LogScannerImpl
    -
     
    -
    wakeUp() - Method in class org.apache.fluss.flink.source.reader.FlinkSourceSplitReader
    -
     
    -
    wakeUp() - Method in class org.apache.fluss.flink.tiering.source.TieringSplitReader
    -
     
    -
    withAdmin(Admin) - Method in class org.apache.fluss.flink.procedure.ProcedureBase
    -
     
    -
    withPartialUpdate(int[]) - Method in class org.apache.fluss.client.table.writer.UpsertWrite
    -
    -
    Returns a new instance of UpsertWrite description with partial update the specified columns.
    -
    -
    withPollTieringTableIntervalMs(long) - Method in class org.apache.fluss.flink.tiering.source.TieringSource.Builder
    -
     
    -
    withProjectedFields(int[]) - Method in class org.apache.fluss.client.table.scanner.log.LogScan
    -
    -
    Returns a new instance of LogScan description with column projection.
    -
    -
    wrapIntoNullableInternalConverter(FlussRowToFlinkRowConverter.FlussDeserializationConverter) - Method in class org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter
    -
     
    -
    write(InputT, SinkWriter.Context) - Method in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
    -
     
    -
    writeBatch() - Method in class org.apache.fluss.client.write.ReadyWriteBatch
    -
     
    -
    WriteBatch - Class in org.apache.fluss.client.write
    -
    -
    The abstract write batch contains write callback object to wait write request feedback.
    -
    -
    WriteBatch(int, PhysicalTablePath, long) - Constructor for class org.apache.fluss.client.write.WriteBatch
    -
     
    -
    WriteBatch.RequestFuture - Class in org.apache.fluss.client.write
    -
    -
    The future for this batch.
    -
    -
    WriteCallback - Interface in org.apache.fluss.client.write
    -
    -
    A callback interface that the user can implement to allow code to execute when the write request - is complete.
    -
    -
    WriteFormat - Enum in org.apache.fluss.client.write
    -
    -
    The format of the write record.
    -
    -
    writerClient - Variable in class org.apache.fluss.client.table.writer.AbstractTableWriter
    -
     
    -
    WriterClient - Class in org.apache.fluss.client.write
    -
    -
    A client that write records to server.
    -
    -
    WriterClient(Configuration, MetadataUpdater, ClientMetricGroup, Admin) - Constructor for class org.apache.fluss.client.write.WriterClient
    -
     
    -
    WriteRecord - Class in org.apache.fluss.client.write
    -
    -
    A record to write to a table.
    -
    -
    writeResult() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResult
    -
     
    -
    writerId() - Method in class org.apache.fluss.client.write.ArrowLogWriteBatch
    -
     
    -
    writerId() - Method in class org.apache.fluss.client.write.IndexedLogWriteBatch
    -
     
    -
    writerId() - Method in class org.apache.fluss.client.write.KvWriteBatch
    -
     
    -
    writerId() - Method in class org.apache.fluss.client.write.WriteBatch
    -
     
    -
    WriterMetricGroup - Class in org.apache.fluss.client.metrics
    -
    -
    Metrics for WriterClient.
    -
    -
    WriterMetricGroup(ClientMetricGroup) - Constructor for class org.apache.fluss.client.metrics.WriterMetricGroup
    -
     
    -
    -A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-3.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-3.html deleted file mode 100644 index 3b152c3dac..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-3.html +++ /dev/null @@ -1,561 +0,0 @@ - - - - - -C-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages - - -

    C

    -
    -
    call(ProcedureContext, String, String, String, String, String) - Method in class org.apache.fluss.flink.procedure.AddAclProcedure
    -
     
    -
    call(ProcedureContext, String, String, String, String, String) - Method in class org.apache.fluss.flink.procedure.DropAclProcedure
    -
     
    -
    call(ProcedureContext, String, String, String, String, String) - Method in class org.apache.fluss.flink.procedure.ListAclProcedure
    -
     
    -
    callbacks - Variable in class org.apache.fluss.client.write.WriteBatch
    -
     
    -
    canEqual(Object) - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
    -
     
    -
    canEqual(Object) - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
    -
     
    -
    CatalogExceptionUtils - Class in org.apache.fluss.flink.utils
    -
    -
    Utility class for catalog exceptions.
    -
    -
    catalogName - Variable in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    CatalogPropertiesUtils - Class in org.apache.fluss.flink.utils
    -
    -
    Utilities for de/serializing Catalog objects into a map of string properties.
    -
    -
    CatalogTableAdapter - Class in org.apache.fluss.flink.adapter
    -
    -
    A adapter for CatalogTable constructor.
    -
    -
    CatalogTableAdapter() - Constructor for class org.apache.fluss.flink.adapter.CatalogTableAdapter
    -
     
    -
    channel(InputT) - Method in class org.apache.fluss.flink.sink.FlinkRowDataChannelComputer
    -
     
    -
    channel(T) - Method in interface org.apache.fluss.flink.sink.ChannelComputer
    -
     
    -
    ChannelComputer<T> - Interface in org.apache.fluss.flink.sink
    -
    -
    A utility class to compute which downstream channel a given record should be sent to before flink - sink.
    -
    -
    checkAndCreatePartitionAsync(PhysicalTablePath) - Method in class org.apache.fluss.client.write.DynamicPartitionCreator
    -
     
    -
    checkAndUpdateMetadata(TablePath, TableBucket) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
    -
    -
    Check the table/partition info for the given table bucket exist in metadata cache, if not, - try to update the metadata cache.
    -
    -
    checkAndUpdatePartitionMetadata(PhysicalTablePath) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
    -
    -
    Check the partition exists in metadata cache, if not, try to update the metadata cache, if - not exist yet, throw exception.
    -
    -
    checkAndUpdatePartitionMetadata(TablePath, Collection<Long>) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
    -
    -
    Check the partitions info for the given partition ids exist in metadata cache, if not, try to - update the metadata cache.
    -
    -
    checkAndUpdateTableMetadata(Set<TablePath>) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
    -
     
    -
    checkAsyncException() - Method in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
    -
    -
    This method should only be invoked in the mailbox thread since the counter is not volatile.
    -
    -
    checkFieldCount(InternalRow) - Method in class org.apache.fluss.client.table.writer.AbstractTableWriter
    -
     
    -
    classLoader - Variable in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    CLIENT_OPTIONS - Static variable in class org.apache.fluss.flink.FlinkConnectorOptions
    -
     
    -
    ClientRpcMessageUtils - Class in org.apache.fluss.client.utils
    -
    -
    Utils for making rpc request/response from inner object or convert inner class to rpc - request/response for client.
    -
    -
    ClientRpcMessageUtils() - Constructor for class org.apache.fluss.client.utils.ClientRpcMessageUtils
    -
     
    -
    ClientUtils - Class in org.apache.fluss.client.utils
    -
    -
    Utils for Fluss Client.
    -
    -
    close() - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    close() - Method in interface org.apache.fluss.client.Connection
    -
    -
    Close the connection and release all resources.
    -
    -
    close() - Method in class org.apache.fluss.client.FlussConnection
    -
     
    -
    close() - Method in class org.apache.fluss.client.table.FlussTable
    -
     
    -
    close() - Method in interface org.apache.fluss.client.table.scanner.batch.BatchScanner
    -
    -
    Closes the scanner and should release all resources.
    -
    -
    close() - Method in class org.apache.fluss.client.table.scanner.batch.KvSnapshotBatchScanner
    -
     
    -
    close() - Method in class org.apache.fluss.client.table.scanner.batch.LimitBatchScanner
    -
     
    -
    close() - Method in class org.apache.fluss.client.table.scanner.log.LogFetchBuffer
    -
     
    -
    close() - Method in class org.apache.fluss.client.table.scanner.log.LogFetcher
    -
     
    -
    close() - Method in class org.apache.fluss.client.table.scanner.log.LogScannerImpl
    -
     
    -
    close() - Method in class org.apache.fluss.client.table.scanner.log.RemoteLogDownloader
    -
     
    -
    close() - Method in class org.apache.fluss.client.table.scanner.RemoteFileDownloader
    -
     
    -
    close() - Method in class org.apache.fluss.client.write.ArrowLogWriteBatch
    -
     
    -
    close() - Method in class org.apache.fluss.client.write.IndexedLogWriteBatch
    -
     
    -
    close() - Method in class org.apache.fluss.client.write.KvWriteBatch
    -
     
    -
    close() - Method in class org.apache.fluss.client.write.RecordAccumulator
    -
    -
    Close this accumulator and force all the record buffers to be drained.
    -
    -
    close() - Method in class org.apache.fluss.client.write.WriteBatch
    -
    -
    close the batch to not append new records.
    -
    -
    close() - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    close() - Method in class org.apache.fluss.flink.lake.reader.IndexedLakeSplitRecordIterator
    -
     
    -
    close() - Method in class org.apache.fluss.flink.lake.reader.LakeSnapshotAndLogSplitScanner
    -
     
    -
    close() - Method in class org.apache.fluss.flink.lake.reader.LakeSnapshotScanner
    -
     
    -
    close() - Method in class org.apache.fluss.flink.lake.reader.SeekableLakeSnapshotSplitScanner
    -
     
    -
    close() - Method in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
    -
     
    -
    close() - Method in class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
    -
     
    -
    close() - Method in class org.apache.fluss.flink.source.lookup.FlinkAsyncLookupFunction
    -
     
    -
    close() - Method in class org.apache.fluss.flink.source.lookup.FlinkLookupFunction
    -
     
    -
    close() - Method in class org.apache.fluss.flink.source.reader.BoundedSplitReader
    -
     
    -
    close() - Method in class org.apache.fluss.flink.source.reader.FlinkSourceSplitReader
    -
     
    -
    close() - Method in class org.apache.fluss.flink.tiering.committer.FlussTableLakeSnapshotCommitter
    -
     
    -
    close() - Method in class org.apache.fluss.flink.tiering.committer.TieringCommitOperator
    -
     
    -
    close() - Method in class org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator
    -
     
    -
    close() - Method in class org.apache.fluss.flink.tiering.source.TieringSplitReader
    -
     
    -
    close() - Method in class org.apache.fluss.flink.utils.FlinkRowToFlussRowConverter
    -
    -
    Deprecated.
    -
    close(Duration) - Method in class org.apache.fluss.client.lookup.LookupClient
    -
     
    -
    close(Duration) - Method in class org.apache.fluss.client.write.WriterClient
    -
     
    -
    closeAsync() - Method in class org.apache.fluss.flink.metrics.FlinkMetricRegistry
    -
     
    -
    cluster - Variable in class org.apache.fluss.client.metadata.MetadataUpdater
    -
     
    -
    collectAllRows(List<BatchScanner>) - Static method in class org.apache.fluss.client.table.scanner.batch.BatchScanUtils
    -
    -
    Collect all rows from the scanners.
    -
    -
    collectFetch() - Method in class org.apache.fluss.client.table.scanner.log.LogFetcher
    -
     
    -
    collectFetch(LogFetchBuffer) - Method in class org.apache.fluss.client.table.scanner.log.LogFetchCollector
    -
    -
    Return the fetched log records, empty the record buffer and update the consumed position.
    -
    -
    collectLimitedRows(List<BatchScanner>, int) - Static method in class org.apache.fluss.client.table.scanner.batch.BatchScanUtils
    -
    -
    Collect rows from the scanners until the number of rows reaches the limit.
    -
    -
    collectRows(BatchScanner) - Static method in class org.apache.fluss.client.table.scanner.batch.BatchScanUtils
    -
    -
    Collect all rows from the scanner.
    -
    -
    columnKey(int) - Static method in class org.apache.fluss.flink.utils.CatalogPropertiesUtils
    -
     
    -
    commit(long, CommittedLakeSnapshot) - Method in class org.apache.fluss.flink.tiering.committer.FlussTableLakeSnapshotCommitter
    -
     
    -
    committable() - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessage
    -
     
    -
    CommittableMessage<Committable> - Class in org.apache.fluss.flink.tiering.committer
    -
    -
    A class wrapping CommittableMessage to commit to lake.
    -
    -
    CommittableMessage(Committable) - Constructor for class org.apache.fluss.flink.tiering.committer.CommittableMessage
    -
     
    -
    CommittableMessageTypeInfo<Committable> - Class in org.apache.fluss.flink.tiering.committer
    -
    -
    A TypeInformation for CommittableMessage.
    -
    -
    complete() - Method in class org.apache.fluss.client.write.WriteBatch
    -
    -
    Complete the batch successfully.
    -
    -
    complete(List<byte[]>) - Method in class org.apache.fluss.client.lookup.LookupBatch
    -
    -
    Complete the lookup operations using given values .
    -
    -
    complete(List<List<byte[]>>) - Method in class org.apache.fluss.client.lookup.PrefixLookupBatch
    -
     
    -
    complete(List<PbCreateAclRespInfo>) - Method in class org.apache.fluss.client.admin.CreateAclsResult
    -
    -
    Completes individual futures based on RPC response information.
    -
    -
    complete(List<PbDropAclsFilterResult>) - Method in class org.apache.fluss.client.admin.DropAclsResult
    -
     
    -
    complete(List<T>) - Method in class org.apache.fluss.client.lookup.AbstractLookupBatch
    -
    -
    Complete the lookup operations using given values .
    -
    -
    completeExceptionally(Exception) - Method in class org.apache.fluss.client.lookup.AbstractLookupBatch
    -
    -
    Complete the get operations with given exception.
    -
    -
    completeExceptionally(Exception) - Method in class org.apache.fluss.client.lookup.LookupBatch
    -
    -
    Complete the lookup operations with given exception.
    -
    -
    completeExceptionally(Exception) - Method in class org.apache.fluss.client.lookup.PrefixLookupBatch
    -
    -
    Complete the get operations with given exception.
    -
    -
    completeExceptionally(Exception) - Method in class org.apache.fluss.client.write.WriteBatch
    -
    -
    Complete the batch exceptionally.
    -
    -
    completeExceptionally(Throwable) - Method in class org.apache.fluss.client.admin.CreateAclsResult
    -
    -
    Marks all futures as exceptionally completed with the provided throwable.
    -
    -
    completeExceptionally(Throwable) - Method in class org.apache.fluss.client.admin.DropAclsResult
    -
     
    -
    connection - Variable in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    Connection - Interface in org.apache.fluss.client
    -
    -
    A cluster connection encapsulating lower level individual connections to actual Fluss servers.
    -
    -
    ConnectionFactory - Class in org.apache.fluss.client
    -
    -
    A non-instantiable class that manages creation of Connections.
    -
    -
    convert(ObjectMapper, JsonNode, Object) - Method in interface org.apache.fluss.flink.utils.FlussRowToJsonConverters.FlussRowToJsonConverter
    -
     
    -
    convert(T) - Method in class org.apache.fluss.flink.utils.PojoToRowConverter
    -
    -
    Converts a POJO to a GenericRow object according to the schema.
    -
    -
    copy() - Method in class org.apache.fluss.flink.sink.FlinkStreamPartitioner
    -
     
    -
    copy() - Method in class org.apache.fluss.flink.sink.FlinkTableSink
    -
     
    -
    copy() - Method in class org.apache.fluss.flink.source.FlinkTableSource
    -
     
    -
    copy(int) - Method in class org.apache.fluss.flink.tiering.source.split.TieringLogSplit
    -
     
    -
    copy(int) - Method in class org.apache.fluss.flink.tiering.source.split.TieringSnapshotSplit
    -
     
    -
    copy(int) - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
    -
     
    -
    count() - Method in class org.apache.fluss.client.table.scanner.log.ScanRecords
    -
    -
    The number of records for all buckets.
    -
    -
    countLogTable(TablePath, Configuration) - Static method in class org.apache.fluss.flink.utils.PushdownUtils
    -
     
    -
    create(RowType) - Static method in class org.apache.fluss.flink.utils.FlinkRowToFlussRowConverter
    -
    -
    Deprecated.
    -
    Create a FlinkRowToFlussRowConverter which will convert to IndexedRow.
    -
    -
    create(RowType, KvFormat) - Static method in class org.apache.fluss.flink.utils.FlinkRowToFlussRowConverter
    -
    -
    Deprecated.
    -
    Create a FlinkRowToFlussRowConverter according to the given KvFormat.
    -
    -
    createAcls(Collection<AclBinding>) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    Creates multiple ACL entries in a single atomic operation.
    -
    -
    createAcls(Collection<AclBinding>) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    CreateAclsResult - Class in org.apache.fluss.client.admin
    -
    -
    Represents the result of a batch ACL operation, managing asynchronous completion of individual - ACL operations.
    -
    -
    CreateAclsResult(Collection<AclBinding>) - Constructor for class org.apache.fluss.client.admin.CreateAclsResult
    -
     
    -
    createBatchScanner(TableBucket) - Method in interface org.apache.fluss.client.table.scanner.Scan
    -
    -
    Creates a BatchScanner to read current data in the given table bucket for this scan.
    -
    -
    createBatchScanner(TableBucket) - Method in class org.apache.fluss.client.table.scanner.TableScan
    -
     
    -
    createBatchScanner(TableBucket, long) - Method in interface org.apache.fluss.client.table.scanner.Scan
    -
    -
    Creates a BatchScanner to read given snapshot data in the given table bucket for this - scan.
    -
    -
    createBatchScanner(TableBucket, long) - Method in class org.apache.fluss.client.table.scanner.TableScan
    -
     
    -
    createCatalog(CatalogFactory.Context) - Method in class org.apache.fluss.flink.catalog.FlinkCatalogFactory
    -
     
    -
    createConnection(Configuration) - Static method in class org.apache.fluss.client.ConnectionFactory
    -
    -
    Creates a new Connection to the Fluss cluster.
    -
    -
    createConnection(Configuration, MetricRegistry) - Static method in class org.apache.fluss.client.ConnectionFactory
    -
    -
    Create a new Connection to the Fluss cluster with registering metrics to the given - metricRegistry.
    -
    -
    createDatabase(String, CatalogDatabase, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    createDatabase(String, DatabaseDescriptor, boolean) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    Create a new database asynchronously.
    -
    -
    createDatabase(String, DatabaseDescriptor, boolean) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    createDynamicTableSink(DynamicTableFactory.Context) - Method in class org.apache.fluss.flink.catalog.FlinkTableFactory
    -
     
    -
    createDynamicTableSource(DynamicTableFactory.Context) - Method in class org.apache.fluss.flink.catalog.FlinkTableFactory
    -
     
    -
    createDynamicTableSource(DynamicTableFactory.Context, String) - Method in class org.apache.fluss.flink.lake.LakeTableFactory
    -
     
    -
    createEnumerator(SplitEnumeratorContext<SourceSplitBase>) - Method in class org.apache.fluss.flink.source.FlinkSource
    -
     
    -
    createEnumerator(SplitEnumeratorContext<TieringSplit>) - Method in class org.apache.fluss.flink.tiering.source.TieringSource
    -
     
    -
    createFunction(ObjectPath, CatalogFunction, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    createLakeSource(TablePath, Map<String, String>) - Static method in class org.apache.fluss.flink.utils.LakeSourceUtils
    -
    -
    Return the lake source of the given table.
    -
    -
    createLogScanner() - Method in interface org.apache.fluss.client.table.scanner.Scan
    -
    -
    Creates a LogScanner to continuously read log data for this scan.
    -
    -
    createLogScanner() - Method in class org.apache.fluss.client.table.scanner.TableScan
    -
     
    -
    createLookuper() - Method in interface org.apache.fluss.client.lookup.Lookup
    -
    -
    Creates a Lookuper instance to lookup rows of a primary key table by the specified - lookup columns.
    -
    -
    createLookuper() - Method in class org.apache.fluss.client.lookup.TableLookup
    -
     
    -
    createNullableConverter(DataType) - Method in class org.apache.fluss.flink.utils.FlussRowToJsonConverters
    -
     
    -
    createNullableInternalConverter(DataType) - Method in class org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter
    -
    -
    Create a nullable runtime FlussRowToFlinkRowConverter.FlussDeserializationConverter from given DataType.
    -
    -
    createPartition(ObjectPath, CatalogPartitionSpec, CatalogPartition, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    createPartition(TablePath, PartitionSpec, boolean) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    Create a new partition for a partitioned table.
    -
    -
    createPartition(TablePath, PartitionSpec, boolean) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    createPrimaryKeyLookupNormalizer(int[], RowType) - Static method in class org.apache.fluss.flink.source.lookup.LookupNormalizer
    -
    -
    Create a LookupNormalizer for primary key lookup.
    -
    -
    createReader(SourceReaderContext) - Method in class org.apache.fluss.flink.source.FlinkSource
    -
     
    -
    createReader(SourceReaderContext) - Method in class org.apache.fluss.flink.tiering.source.TieringSource
    -
     
    -
    createRemainingFilter(RowData) - Method in class org.apache.fluss.flink.source.lookup.LookupNormalizer
    -
     
    -
    createSerializer(ExecutionConfig) - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
    -
     
    -
    createSerializer(ExecutionConfig) - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
    -
     
    -
    createStreamOperator(StreamOperatorParameters<CommittableMessage<Committable>>) - Method in class org.apache.fluss.flink.tiering.committer.TieringCommitOperatorFactory
    -
     
    -
    createTable(ObjectPath, CatalogBaseTable, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    createTable(TablePath, TableDescriptor, boolean) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    Create a new table asynchronously.
    -
    -
    createTable(TablePath, TableDescriptor, boolean) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    createWriter() - Method in interface org.apache.fluss.client.table.writer.Append
    -
    -
    Create a new AppendWriter to write data to a Log Table.
    -
    -
    createWriter() - Method in class org.apache.fluss.client.table.writer.TableAppend
    -
     
    -
    createWriter() - Method in class org.apache.fluss.client.table.writer.TableUpsert
    -
     
    -
    createWriter() - Method in interface org.apache.fluss.client.table.writer.Upsert
    -
    -
    Create a new UpsertWriter with the optional Upsert.partialUpdate(String...) - information to upsert and delete data to a Primary Key Table.
    -
    -
    CURRENT_OFFSET_METRIC_GAUGE - Static variable in class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
    -
     
    -
    currentSplitIndex - Variable in class org.apache.fluss.flink.source.reader.RecordAndPos
    -
     
    -
    -A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-4.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-4.html deleted file mode 100644 index 16d40b4acc..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-4.html +++ /dev/null @@ -1,334 +0,0 @@ - - - - - -D-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages - - -

    D

    -
    -
    DATA_LAKE_CONFIG_PREFIX - Static variable in class org.apache.fluss.flink.tiering.source.TieringSourceOptions
    -
     
    -
    databaseExists(String) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    Get whether database exists asynchronously.
    -
    -
    databaseExists(String) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    databaseExists(String) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    DataLakeUtils - Class in org.apache.fluss.flink.utils
    -
    -
    Utility class for accessing data lake related configurations.
    -
    -
    DataLakeUtils() - Constructor for class org.apache.fluss.flink.utils.DataLakeUtils
    -
     
    -
    deallocate(WriteBatch) - Method in class org.apache.fluss.client.write.RecordAccumulator
    -
    -
    Deallocate the record batch.
    -
    -
    dec() - Method in class org.apache.fluss.flink.metrics.FlinkCounter
    -
     
    -
    dec(long) - Method in class org.apache.fluss.flink.metrics.FlinkCounter
    -
     
    -
    DEFAULT_DATABASE - Static variable in class org.apache.fluss.flink.catalog.FlinkCatalogOptions
    -
     
    -
    DEFAULT_SPLIT_INDEX - Static variable in class org.apache.fluss.flink.source.reader.RecordAndPos
    -
     
    -
    defaultDatabase - Variable in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    DefaultSecurityTokenManager - Class in org.apache.fluss.client.token
    -
    -
    Manager for security tokens to access files in fluss client.
    -
    -
    DefaultSecurityTokenManager(Configuration, SecurityTokenProvider) - Constructor for class org.apache.fluss.client.token.DefaultSecurityTokenManager
    -
     
    -
    DefaultSecurityTokenProvider - Class in org.apache.fluss.client.token
    -
    -
    A default implementation of SecurityTokenProvider to get token from server.
    -
    -
    DefaultSecurityTokenProvider(AdminReadOnlyGateway) - Constructor for class org.apache.fluss.client.token.DefaultSecurityTokenProvider
    -
     
    -
    delete(InternalRow) - Method in interface org.apache.fluss.client.table.writer.UpsertWriter
    -
    -
    Delete certain row by the input row in Fluss table, the input row must contain the primary - key.
    -
    -
    DELETE - org.apache.fluss.flink.row.OperationType
    -
    -
    Represents a delete operation.
    -
    -
    DeleteResult - Class in org.apache.fluss.client.table.writer
    -
    -
    The result of deleting a record (UpsertWriter.delete(InternalRow).
    -
    -
    DeleteResult() - Constructor for class org.apache.fluss.client.table.writer.DeleteResult
    -
     
    -
    deleteSingleRow(GenericRow, TablePath, Configuration) - Static method in class org.apache.fluss.flink.utils.PushdownUtils
    -
     
    -
    deserialize(byte, TableBucket, String, DataInputDeserializer) - Method in class org.apache.fluss.flink.lake.LakeSplitSerializer
    -
     
    -
    deserialize(int, byte[]) - Method in class org.apache.fluss.flink.source.split.SourceSplitSerializer
    -
     
    -
    deserialize(int, byte[]) - Method in class org.apache.fluss.flink.source.state.FlussSourceEnumeratorStateSerializer
    -
     
    -
    deserialize(int, byte[]) - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplitSerializer
    -
     
    -
    deserialize(int, byte[]) - Method in class org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorStateSerializer
    -
     
    -
    deserialize(int, byte[]) - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultSerializer
    -
     
    -
    deserialize(Object) - Method in interface org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter.FlussDeserializationConverter
    -
    -
    Convert a Fluss field object of InternalRow to the Flink's internal data - structure object.
    -
    -
    deserialize(LogRecord) - Method in interface org.apache.fluss.flink.source.deserializer.FlussDeserializationSchema
    -
    -
    Deserializes a LogRecord into an object of type T.
    -
    -
    deserialize(LogRecord) - Method in class org.apache.fluss.flink.source.deserializer.JsonStringDeserializationSchema
    -
    -
    Deserializes a LogRecord into a JSON String.
    -
    -
    deserialize(LogRecord) - Method in class org.apache.fluss.flink.source.deserializer.RowDataDeserializationSchema
    -
    -
    Deserializes a LogRecord into a Flink RowData object.
    -
    -
    deserializeComputedColumn(Map<String, String>, int, Schema.Builder) - Static method in class org.apache.fluss.flink.utils.CatalogPropertiesUtils
    -
     
    -
    deserializeOptions(Map<String, String>) - Static method in class org.apache.fluss.flink.utils.CatalogPropertiesUtils
    -
     
    -
    DeserializerInitContextImpl - Class in org.apache.fluss.flink.source.deserializer
    -
    -
    Contextual information that can be used during initialization of FlussDeserializationSchema.
    -
    -
    DeserializerInitContextImpl(MetricGroup, UserCodeClassLoader, RowType) - Constructor for class org.apache.fluss.flink.source.deserializer.DeserializerInitContextImpl
    -
     
    -
    deserializeWatermark(Map<String, String>, Schema.Builder) - Static method in class org.apache.fluss.flink.utils.CatalogPropertiesUtils
    -
     
    -
    done() - Method in class org.apache.fluss.client.write.WriteBatch.RequestFuture
    -
    -
    Mark this request as complete and unblock any threads waiting on its completion.
    -
    -
    downloadFile(Path, FsPath) - Method in class org.apache.fluss.client.table.scanner.RemoteFileDownloader
    -
    -
    Copies the file from a remote file path to the given target file path, returns the number of - downloaded bytes.
    -
    -
    downloadFileAsync(FsPathAndFileName, Path) - Method in class org.apache.fluss.client.table.scanner.RemoteFileDownloader
    -
    -
    Downloads the file from the given remote file path to the target directory asynchronously, - returns a Future object of the number of downloaded bytes.
    -
    -
    downloadThreadPool - Variable in class org.apache.fluss.client.table.scanner.RemoteFileDownloader
    -
     
    -
    drain(Cluster, Set<Integer>, int) - Method in class org.apache.fluss.client.write.RecordAccumulator
    -
    -
    Drain all the data for the given nodes and collate them into a list of batches that will fit - within the specified size on a per-node basis.
    -
    -
    DropAclProcedure - Class in org.apache.fluss.flink.procedure
    -
    -
    Procedure to drop acl.
    -
    -
    DropAclProcedure() - Constructor for class org.apache.fluss.flink.procedure.DropAclProcedure
    -
     
    -
    dropAcls(Collection<AclBindingFilter>) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    Removes multiple ACL entries in a single atomic operation.
    -
    -
    dropAcls(Collection<AclBindingFilter>) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    DropAclsResult - Class in org.apache.fluss.client.admin
    -
    -
    Represents the result of a drop ACLs operation.
    -
    -
    DropAclsResult.FilterResult - Class in org.apache.fluss.client.admin
    -
    -
    A class containing either the deleted ACL binding or an exception if the delete failed.
    -
    -
    DropAclsResult.FilterResults - Class in org.apache.fluss.client.admin
    -
    -
    A class containing the results of the delete ACLs operation.
    -
    -
    dropDatabase(String, boolean, boolean) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    Drop the database with the given name asynchronously.
    -
    -
    dropDatabase(String, boolean, boolean) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    dropDatabase(String, boolean, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    dropFunction(ObjectPath, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    dropPartition(ObjectPath, CatalogPartitionSpec, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    dropPartition(TablePath, PartitionSpec, boolean) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    Drop a partition from a partitioned table.
    -
    -
    dropPartition(TablePath, PartitionSpec, boolean) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    dropTable(ObjectPath, boolean) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    dropTable(TablePath, boolean) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    Drop the table with the given table path asynchronously.
    -
    -
    dropTable(TablePath, boolean) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    DynamicPartitionCreator - Class in org.apache.fluss.client.write
    -
    -
    A creator to create partition when dynamic partition create enable for table.
    -
    -
    DynamicPartitionCreator(MetadataUpdater, Admin, boolean, Consumer<Throwable>) - Constructor for class org.apache.fluss.client.write.DynamicPartitionCreator
    -
     
    -
    DynamicWriteBatchSizeEstimator - Class in org.apache.fluss.client.write
    -
    -
    An estimator to estimate the buffer usage of a writeBatch.
    -
    -
    DynamicWriteBatchSizeEstimator(boolean, int, int) - Constructor for class org.apache.fluss.client.write.DynamicWriteBatchSizeEstimator
    -
     
    -
    -A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-5.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-5.html deleted file mode 100644 index 7cff70d6a2..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-5.html +++ /dev/null @@ -1,236 +0,0 @@ - - - - - -E-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages - - -

    E

    -
    -
    earliest() - Static method in interface org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer
    -
    -
    Get an OffsetsInitializer which initializes the offsets to the earliest available - offsets of each bucket.
    -
    -
    EARLIEST - org.apache.fluss.flink.FlinkConnectorOptions.ScanStartupMode
    -
     
    -
    EARLIEST_OFFSET - Static variable in interface org.apache.fluss.client.table.scanner.log.LogScanner
    -
    -
    The earliest offset to fetch from.
    -
    -
    earliestOffsets(String, Collection<Integer>) - Method in class org.apache.fluss.flink.source.enumerator.initializer.BucketOffsetsRetrieverImpl
    -
     
    -
    earliestOffsets(String, Collection<Integer>) - Method in interface org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer.BucketOffsetsRetriever
    -
     
    -
    EarliestSpec() - Constructor for class org.apache.fluss.client.admin.OffsetSpec.EarliestSpec
    -
     
    -
    emitRecord(RecordAndPos, SourceOutput<OUT>, SourceSplitState) - Method in class org.apache.fluss.flink.source.emitter.FlinkRecordEmitter
    -
     
    -
    emitRecord(SourceSplitState, SourceOutput<OUT>, RecordAndPos) - Method in class org.apache.fluss.flink.lake.LakeRecordRecordEmitter
    -
     
    -
    emitRecord(TableBucketWriteResult<WriteResult>, SourceOutput<TableBucketWriteResult<WriteResult>>, TieringSplitState) - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultEmitter
    -
     
    -
    EMPTY - Static variable in class org.apache.fluss.client.table.scanner.log.ScanRecords
    -
     
    -
    emptyRecords(FlinkSourceReaderMetrics) - Static method in class org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds
    -
     
    -
    equals(Object) - Method in class org.apache.fluss.client.lookup.LookupResult
    -
     
    -
    equals(Object) - Method in class org.apache.fluss.client.table.scanner.ScanRecord
    -
     
    -
    equals(Object) - Method in class org.apache.fluss.flink.row.RowWithOp
    -
    -
    Indicates whether some other object is "equal to" this one.
    -
    -
    equals(Object) - Method in class org.apache.fluss.flink.source.event.PartitionBucketsUnsubscribedEvent
    -
     
    -
    equals(Object) - Method in class org.apache.fluss.flink.source.reader.RecordAndPos
    -
     
    -
    equals(Object) - Method in class org.apache.fluss.flink.source.split.HybridSnapshotLogSplit
    -
     
    -
    equals(Object) - Method in class org.apache.fluss.flink.source.split.LogSplit
    -
     
    -
    equals(Object) - Method in class org.apache.fluss.flink.source.split.SnapshotSplit
    -
     
    -
    equals(Object) - Method in class org.apache.fluss.flink.source.split.SourceSplitBase
    -
     
    -
    equals(Object) - Method in class org.apache.fluss.flink.source.state.SourceEnumeratorState
    -
     
    -
    equals(Object) - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
    -
     
    -
    equals(Object) - Method in class org.apache.fluss.flink.tiering.source.split.TieringLogSplit
    -
     
    -
    equals(Object) - Method in class org.apache.fluss.flink.tiering.source.split.TieringSnapshotSplit
    -
     
    -
    equals(Object) - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
    -
     
    -
    equals(Object) - Method in class org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorState
    -
     
    -
    equals(Object) - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
    -
     
    -
    equalValue - Variable in class org.apache.fluss.flink.utils.PushdownUtils.FieldEqual
    -
     
    -
    estimatedSizeInBytes() - Method in class org.apache.fluss.client.write.ArrowLogWriteBatch
    -
     
    -
    estimatedSizeInBytes() - Method in class org.apache.fluss.client.write.IndexedLogWriteBatch
    -
     
    -
    estimatedSizeInBytes() - Method in class org.apache.fluss.client.write.KvWriteBatch
    -
     
    -
    estimatedSizeInBytes() - Method in class org.apache.fluss.client.write.WriteBatch
    -
    -
    Get an estimate of the number of bytes written to the underlying buffer.
    -
    -
    exception() - Method in class org.apache.fluss.client.admin.DropAclsResult.FilterResult
    -
    -
    Return an exception if the ACL delete was not successful or null if it was.
    -
    -
    executeDeletion() - Method in class org.apache.fluss.flink.sink.FlinkTableSink
    -
     
    -
    extractFieldEquals(List<ResolvedExpression>, Map<Integer, LogicalType>, List<ResolvedExpression>, List<ResolvedExpression>, PushdownUtils.ValueConversion) - Static method in class org.apache.fluss.flink.utils.PushdownUtils
    -
    -
    Extract field equality information from expressions.
    -
    -
    extractLakeCatalogProperties(Configuration) - Static method in class org.apache.fluss.flink.utils.DataLakeUtils
    -
     
    -
    -A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-6.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-6.html deleted file mode 100644 index 5f272d66d3..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-6.html +++ /dev/null @@ -1,548 +0,0 @@ - - - - - -F-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages - - -

    F

    -
    -
    factoryIdentifier() - Method in class org.apache.fluss.flink.catalog.FlinkCatalogFactory
    -
     
    -
    factoryIdentifier() - Method in class org.apache.fluss.flink.catalog.FlinkTableFactory
    -
     
    -
    FailedTieringEvent - Class in org.apache.fluss.flink.tiering.event
    -
    -
    SourceEvent used to represent a Fluss table is failed during tiering.
    -
    -
    FailedTieringEvent(long, String) - Constructor for class org.apache.fluss.flink.tiering.event.FailedTieringEvent
    -
     
    -
    failReason() - Method in class org.apache.fluss.flink.tiering.event.FailedTieringEvent
    -
     
    -
    fetch() - Method in class org.apache.fluss.flink.source.reader.FlinkSourceSplitReader
    -
     
    -
    fetch() - Method in class org.apache.fluss.flink.tiering.source.TieringSplitReader
    -
     
    -
    fetchRequestCount() - Method in class org.apache.fluss.client.metrics.ScannerMetricGroup
    -
     
    -
    fieldCount - Variable in class org.apache.fluss.client.table.writer.AbstractTableWriter
    -
     
    -
    FieldEqual(int, Object) - Constructor for class org.apache.fluss.flink.utils.PushdownUtils.FieldEqual
    -
     
    -
    fieldIndex - Variable in class org.apache.fluss.flink.utils.PushdownUtils.FieldEqual
    -
     
    -
    finishedSplits() - Method in class org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds
    -
     
    -
    FinishedTieringEvent - Class in org.apache.fluss.flink.tiering.event
    -
    -
    SourceEvent used to represent a Fluss table has been tiered finished.
    -
    -
    FinishedTieringEvent(long) - Constructor for class org.apache.fluss.flink.tiering.event.FinishedTieringEvent
    -
     
    -
    FLINK_INTERNAL_VALUE - org.apache.fluss.flink.utils.PushdownUtils.ValueConversion
    -
     
    -
    FlinkAsFlussRow - Class in org.apache.fluss.flink.row
    -
    -
    Wraps a Flink RowData as a Fluss InternalRow.
    -
    -
    FlinkAsFlussRow() - Constructor for class org.apache.fluss.flink.row.FlinkAsFlussRow
    -
     
    -
    FlinkAsyncLookupFunction - Class in org.apache.fluss.flink.source.lookup
    -
    -
    A flink async lookup function for fluss.
    -
    -
    FlinkAsyncLookupFunction(Configuration, TablePath, RowType, int, LookupNormalizer, int[]) - Constructor for class org.apache.fluss.flink.source.lookup.FlinkAsyncLookupFunction
    -
     
    -
    FlinkCatalog - Class in org.apache.fluss.flink.catalog
    -
    -
    A Flink Catalog for fluss.
    -
    -
    FlinkCatalog(String, String, String, ClassLoader, Map<String, String>) - Constructor for class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    FlinkCatalogFactory - Class in org.apache.fluss.flink.catalog
    -
    -
    Factory for FlinkCatalog.
    -
    -
    FlinkCatalogFactory() - Constructor for class org.apache.fluss.flink.catalog.FlinkCatalogFactory
    -
     
    -
    FlinkCatalogOptions - Class in org.apache.fluss.flink.catalog
    -
    -
    Options for flink catalog.
    -
    -
    FlinkConnectorOptions - Class in org.apache.fluss.flink
    -
    -
    Options for flink connector.
    -
    -
    FlinkConnectorOptions() - Constructor for class org.apache.fluss.flink.FlinkConnectorOptions
    -
     
    -
    FlinkConnectorOptions.ScanStartupMode - Enum in org.apache.fluss.flink
    -
    -
    Startup mode for the fluss scanner, see FlinkConnectorOptions.SCAN_STARTUP_MODE.
    -
    -
    FlinkConnectorOptionsUtils - Class in org.apache.fluss.flink.utils
    -
    -
    Utility class for FlinkConnectorOptions.
    -
    -
    FlinkConnectorOptionsUtils() - Constructor for class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils
    -
     
    -
    FlinkConnectorOptionsUtils.StartupOptions - Class in org.apache.fluss.flink.utils
    -
    -
    Fluss startup options.
    -
    -
    FlinkConversions - Class in org.apache.fluss.flink.utils
    -
    -
    Utils for conversion between Flink and Fluss.
    -
    -
    FlinkCounter - Class in org.apache.fluss.flink.metrics
    -
    -
    An implementation of Flink's Counter which wraps Fluss's Counter.
    -
    -
    FlinkCounter(Counter) - Constructor for class org.apache.fluss.flink.metrics.FlinkCounter
    -
     
    -
    FlinkGauge<T> - Class in org.apache.fluss.flink.metrics
    -
    -
    An implementation of Flink's Gauge which wraps Fluss's Gauge.
    -
    -
    FlinkGauge(Gauge<T>) - Constructor for class org.apache.fluss.flink.metrics.FlinkGauge
    -
     
    -
    FlinkHistogram - Class in org.apache.fluss.flink.metrics
    -
    -
    An implementation of Flink's Histogram which wraps Fluss's Histogram.
    -
    -
    FlinkHistogram(Histogram) - Constructor for class org.apache.fluss.flink.metrics.FlinkHistogram
    -
     
    -
    FlinkLookupFunction - Class in org.apache.fluss.flink.source.lookup
    -
    -
    A flink lookup function for fluss.
    -
    -
    FlinkLookupFunction(Configuration, TablePath, RowType, int, LookupNormalizer, int[]) - Constructor for class org.apache.fluss.flink.source.lookup.FlinkLookupFunction
    -
     
    -
    FlinkMeter - Class in org.apache.fluss.flink.metrics
    -
    -
    An implementation of Flink's Meter which wraps Fluss's Meter.
    -
    -
    FlinkMeter(Meter) - Constructor for class org.apache.fluss.flink.metrics.FlinkMeter
    -
     
    -
    flinkMetricRegistry - Variable in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
    -
     
    -
    FlinkMetricRegistry - Class in org.apache.fluss.flink.metrics
    -
    -
    An implementation of MetricRegistry which registers all metrics into Flink's metric - system.
    -
    -
    FlinkMetricRegistry(MetricGroup) - Constructor for class org.apache.fluss.flink.metrics.FlinkMetricRegistry
    -
     
    -
    FlinkMetricRegistry(MetricGroup, Set<String>) - Constructor for class org.apache.fluss.flink.metrics.FlinkMetricRegistry
    -
     
    -
    FlinkRecordEmitter<OUT> - Class in org.apache.fluss.flink.source.emitter
    -
    -
    The RecordEmitter implementation for FlinkSourceReader.
    -
    -
    FlinkRecordEmitter(FlussDeserializationSchema<OUT>) - Constructor for class org.apache.fluss.flink.source.emitter.FlinkRecordEmitter
    -
     
    -
    FlinkRecordsWithSplitIds - Class in org.apache.fluss.flink.source.reader
    -
    -
    An implementation of RecordsWithSplitIds which contains records from multiple splits.
    -
    -
    FlinkRecordsWithSplitIds(String, TableBucket, CloseableIterator<RecordAndPos>, FlinkSourceReaderMetrics) - Constructor for class org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds
    -
     
    -
    FlinkRecordsWithSplitIds(Map<String, CloseableIterator<RecordAndPos>>, Iterator<String>, Iterator<TableBucket>, Set<String>, FlinkSourceReaderMetrics) - Constructor for class org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds
    -
     
    -
    FlinkRecordsWithSplitIds(Set<String>, FlinkSourceReaderMetrics) - Constructor for class org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds
    -
     
    -
    FlinkRowDataChannelComputer<InputT> - Class in org.apache.fluss.flink.sink
    -
    -
    ChannelComputer for flink RowData.
    -
    -
    FlinkRowDataChannelComputer(RowType, List<String>, List<String>, DataLakeFormat, int, FlussSerializationSchema<InputT>) - Constructor for class org.apache.fluss.flink.sink.FlinkRowDataChannelComputer
    -
     
    -
    FlinkRowToFlussRowConverter - Class in org.apache.fluss.flink.utils
    -
    -
    Deprecated. -
    This class is deprecated and will be removed in future versions. It is recommended to - use the FlinkAsFlussRow to wraps Flink's RowData as Fluss's InternalRow.
    -
    -
    -
    FlinkRowToFlussRowConverter.FlussSerializationConverter - Interface in org.apache.fluss.flink.utils
    -
    -
    Deprecated.
    -
    Runtime converter to convert field in Flink's RowData to Fluss's InternalRow - type object.
    -
    -
    FlinkSinkWriter<InputT> - Class in org.apache.fluss.flink.sink.writer
    -
    -
    Base class for Flink SinkWriter implementations in Fluss.
    -
    -
    FlinkSinkWriter(TablePath, Configuration, RowType, int[], MailboxExecutor, FlussSerializationSchema<InputT>) - Constructor for class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
    -
     
    -
    FlinkSinkWriter(TablePath, Configuration, RowType, MailboxExecutor, FlussSerializationSchema<InputT>) - Constructor for class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
    -
     
    -
    FlinkSource<OUT> - Class in org.apache.fluss.flink.source
    -
    -
    Flink source for Fluss.
    -
    -
    FlinkSource(Configuration, TablePath, boolean, boolean, RowType, int[], OffsetsInitializer, long, FlussDeserializationSchema<OUT>, boolean, List<PushdownUtils.FieldEqual>) - Constructor for class org.apache.fluss.flink.source.FlinkSource
    -
     
    -
    FlinkSource(Configuration, TablePath, boolean, boolean, RowType, int[], OffsetsInitializer, long, FlussDeserializationSchema<OUT>, boolean, List<PushdownUtils.FieldEqual>, LakeSource<LakeSplit>) - Constructor for class org.apache.fluss.flink.source.FlinkSource
    -
     
    -
    FlinkSourceEnumerator - Class in org.apache.fluss.flink.source.enumerator
    -
    -
    An implementation of SplitEnumerator for the data of Fluss.
    -
    -
    FlinkSourceEnumerator(TablePath, Configuration, boolean, boolean, SplitEnumeratorContext<SourceSplitBase>, Set<TableBucket>, Map<Long, String>, List<SourceSplitBase>, OffsetsInitializer, long, boolean, List<PushdownUtils.FieldEqual>, LakeSource<LakeSplit>) - Constructor for class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
    -
     
    -
    FlinkSourceEnumerator(TablePath, Configuration, boolean, boolean, SplitEnumeratorContext<SourceSplitBase>, OffsetsInitializer, long, boolean, List<PushdownUtils.FieldEqual>) - Constructor for class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
    -
     
    -
    FlinkSourceEnumerator(TablePath, Configuration, boolean, boolean, SplitEnumeratorContext<SourceSplitBase>, OffsetsInitializer, long, boolean, List<PushdownUtils.FieldEqual>, LakeSource<LakeSplit>) - Constructor for class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
    -
     
    -
    FlinkSourceFetcherManager - Class in org.apache.fluss.flink.source.reader.fetcher
    -
    -
    The SplitFetcherManager for Fluss source.
    -
    -
    FlinkSourceFetcherManager(FutureCompletingBlockingQueue<RecordsWithSplitIds<RecordAndPos>>, Supplier<SplitReader<RecordAndPos, SourceSplitBase>>, Consumer<Collection<String>>) - Constructor for class org.apache.fluss.flink.source.reader.fetcher.FlinkSourceFetcherManager
    -
    -
    Creates a new SplitFetcherManager with a single I/O threads.
    -
    -
    FlinkSourceReader<OUT> - Class in org.apache.fluss.flink.source.reader
    -
    -
    The source reader for Fluss.
    -
    -
    FlinkSourceReader(FutureCompletingBlockingQueue<RecordsWithSplitIds<RecordAndPos>>, Configuration, TablePath, RowType, SourceReaderContext, int[], FlinkSourceReaderMetrics, FlinkRecordEmitter<OUT>, LakeSource<LakeSplit>) - Constructor for class org.apache.fluss.flink.source.reader.FlinkSourceReader
    -
     
    -
    FlinkSourceReaderMetrics - Class in org.apache.fluss.flink.source.metrics
    -
    -
    A collection class for handling metrics in FlinkSourceReader of Fluss.
    -
    -
    FlinkSourceReaderMetrics(SourceReaderMetricGroup) - Constructor for class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
    -
     
    -
    FlinkSourceSplitReader - Class in org.apache.fluss.flink.source.reader
    -
    -
    An implementation of SplitReader for reading splits into RecordAndPos.
    -
    -
    FlinkSourceSplitReader(Configuration, TablePath, RowType, int[], FlinkSourceReaderMetrics, LakeSource<LakeSplit>) - Constructor for class org.apache.fluss.flink.source.reader.FlinkSourceSplitReader
    -
     
    -
    FlinkStreamPartitioner<T> - Class in org.apache.fluss.flink.sink
    -
    -
    A StreamPartitioner which wraps a ChannelComputer.
    -
    -
    FlinkStreamPartitioner(ChannelComputer<T>) - Constructor for class org.apache.fluss.flink.sink.FlinkStreamPartitioner
    -
     
    -
    FlinkTableFactory - Class in org.apache.fluss.flink.catalog
    -
    -
    Factory to create table source and table sink for Fluss.
    -
    -
    FlinkTableFactory() - Constructor for class org.apache.fluss.flink.catalog.FlinkTableFactory
    -
     
    -
    FlinkTableSink - Class in org.apache.fluss.flink.sink
    -
    -
    A Flink DynamicTableSink.
    -
    -
    FlinkTableSink(TablePath, Configuration, RowType, int[], List<String>, boolean, MergeEngineType, DataLakeFormat, boolean, int, List<String>, boolean) - Constructor for class org.apache.fluss.flink.sink.FlinkTableSink
    -
     
    -
    FlinkTableSource - Class in org.apache.fluss.flink.source
    -
    -
    Flink table source to scan Fluss data.
    -
    -
    FlinkTableSource(TablePath, Configuration, RowType, int[], int[], int[], boolean, FlinkConnectorOptionsUtils.StartupOptions, int, boolean, LookupCache, long, boolean, MergeEngineType, Map<String, String>) - Constructor for class org.apache.fluss.flink.source.FlinkTableSource
    -
     
    -
    FlinkUtils - Class in org.apache.fluss.flink.utils
    -
    -
    Utils for Flink classes.
    -
    -
    FlinkUtils() - Constructor for class org.apache.fluss.flink.utils.FlinkUtils
    -
     
    -
    flush() - Method in class org.apache.fluss.client.table.writer.AbstractTableWriter
    -
    -
    Flush data written that have not yet been sent to the server, forcing the client to send the - requests to server and blocks on the completion of the requests associated with these - records.
    -
    -
    flush() - Method in interface org.apache.fluss.client.table.writer.TableWriter
    -
    -
    Flush data written that have not yet been sent to the server, forcing the client to send the - requests to server and blocks on the completion of the requests associated with these - records.
    -
    -
    flush() - Method in class org.apache.fluss.client.write.WriterClient
    -
    -
    Invoking this method makes all buffered records immediately available to send (even if - linger.ms is greater than 0) and blocks on the completion of the requests associated - with these records.
    -
    -
    flush(boolean) - Method in class org.apache.fluss.flink.sink.writer.AppendSinkWriter
    -
     
    -
    flush(boolean) - Method in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
    -
     
    -
    flush(boolean) - Method in class org.apache.fluss.flink.sink.writer.UpsertSinkWriter
    -
     
    -
    FLUSS_GROUP_NAME - Static variable in class org.apache.fluss.flink.metrics.FlinkMetricRegistry
    -
     
    -
    FLUSS_INTERNAL_VALUE - org.apache.fluss.flink.utils.PushdownUtils.ValueConversion
    -
     
    -
    FLUSS_METRIC_GROUP - Static variable in class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
    -
     
    -
    FlussAdmin - Class in org.apache.fluss.client.admin
    -
    -
    The default implementation of Admin.
    -
    -
    FlussAdmin(RpcClient, MetadataUpdater) - Constructor for class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    FlussConnection - Class in org.apache.fluss.client
    -
    -
    A connection to Fluss cluster, and holds the client session resources.
    -
    -
    FlussDeserializationSchema<T> - Interface in org.apache.fluss.flink.source.deserializer
    -
    -
    Interface for deserialization schema used to deserialize LogRecord objects into specific - data types.
    -
    -
    FlussDeserializationSchema.InitializationContext - Interface in org.apache.fluss.flink.source.deserializer
    -
    -
    A contextual information provided for FlussDeserializationSchema.open(InitializationContext) method.
    -
    -
    FlussRowToFlinkRowConverter - Class in org.apache.fluss.flink.utils
    -
    -
    A converter to convert Fluss's InternalRow to Flink's RowData.
    -
    -
    FlussRowToFlinkRowConverter(RowType) - Constructor for class org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter
    -
     
    -
    FlussRowToFlinkRowConverter.FlussDeserializationConverter - Interface in org.apache.fluss.flink.utils
    -
    -
    Runtime converter to convert field in Fluss's InternalRow to Flink's RowData - type object.
    -
    -
    FlussRowToJsonConverters - Class in org.apache.fluss.flink.utils
    -
    -
    A converter to convert Fluss's InternalRow to JsonNode.
    -
    -
    FlussRowToJsonConverters(TimestampFormat) - Constructor for class org.apache.fluss.flink.utils.FlussRowToJsonConverters
    -
     
    -
    FlussRowToJsonConverters.FlussRowToJsonConverter - Interface in org.apache.fluss.flink.utils
    -
    -
    Runtime converter that converts objects of Fluss data structures to corresponding JsonNodes.
    -
    -
    FlussSerializationSchema<T> - Interface in org.apache.fluss.flink.sink.serializer
    -
    -
    A serialization schema for Fluss.
    -
    -
    FlussSerializationSchema.InitializationContext - Interface in org.apache.fluss.flink.sink.serializer
    -
    -
    A contextual information provided for FlussSerializationSchema.open(InitializationContext) method.
    -
    -
    FlussSink<InputT> - Class in org.apache.fluss.flink.sink
    -
    -
    FlussSink is a specialized Flink sink for writing data to Fluss.
    -
    -
    FlussSinkBuilder<InputT> - Class in org.apache.fluss.flink.sink
    -
    -
    Builder for creating and configuring Fluss sink connectors for Apache Flink.
    -
    -
    FlussSinkBuilder() - Constructor for class org.apache.fluss.flink.sink.FlussSinkBuilder
    -
     
    -
    FlussSource<OUT> - Class in org.apache.fluss.flink.source
    -
    -
    A Flink DataStream source implementation for reading data from Fluss tables.
    -
    -
    FlussSourceBuilder<OUT> - Class in org.apache.fluss.flink.source
    -
    -
    Builder class for creating FlussSource instances.
    -
    -
    FlussSourceBuilder() - Constructor for class org.apache.fluss.flink.source.FlussSourceBuilder
    -
     
    -
    FlussSourceEnumeratorStateSerializer - Class in org.apache.fluss.flink.source.state
    -
    -
    A serializer for SourceEnumeratorState.
    -
    -
    FlussSourceEnumeratorStateSerializer(LakeSource<LakeSplit>) - Constructor for class org.apache.fluss.flink.source.state.FlussSourceEnumeratorStateSerializer
    -
     
    -
    FlussTable - Class in org.apache.fluss.client.table
    -
    -
    The base impl of Table.
    -
    -
    FlussTable(FlussConnection, TablePath, TableInfo) - Constructor for class org.apache.fluss.client.table.FlussTable
    -
     
    -
    FlussTableLakeSnapshotCommitter - Class in org.apache.fluss.flink.tiering.committer
    -
    -
    Committer to commit FlussTableLakeSnapshot of lake to Fluss.
    -
    -
    FlussTableLakeSnapshotCommitter(Configuration) - Constructor for class org.apache.fluss.flink.tiering.committer.FlussTableLakeSnapshotCommitter
    -
     
    -
    forArrowAppend(PhysicalTablePath, InternalRow, byte[]) - Static method in class org.apache.fluss.client.write.WriteRecord
    -
    -
    Creates a write record for append operation for Arrow format.
    -
    -
    forceClose() - Method in class org.apache.fluss.client.write.Sender
    -
    -
    Closes the sender without sending out any pending messages.
    -
    -
    forDelete(PhysicalTablePath, byte[], byte[], int[]) - Static method in class org.apache.fluss.client.write.WriteRecord
    -
    -
    Create a write record for delete operation and partial-delete update.
    -
    -
    forIndexedAppend(PhysicalTablePath, IndexedRow, byte[]) - Static method in class org.apache.fluss.client.write.WriteRecord
    -
    -
    Create a write record for append operation for indexed format.
    -
    -
    forUpsert(PhysicalTablePath, BinaryRow, byte[], byte[], int[]) - Static method in class org.apache.fluss.client.write.WriteRecord
    -
    -
    Create a write record for upsert operation and partial-upsert operation.
    -
    -
    fromFlinkDecimal(DecimalData) - Static method in class org.apache.fluss.flink.row.FlinkAsFlussRow
    -
     
    -
    full() - Static method in interface org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer
    -
    -
    Get an OffsetsInitializer which performs a full snapshot on the table upon first - startup, and continue to read log with the offset to the snapshot.
    -
    -
    FULL - org.apache.fluss.flink.FlinkConnectorOptions.ScanStartupMode
    -
     
    -
    functionExists(ObjectPath) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    future() - Method in class org.apache.fluss.client.lookup.AbstractLookupQuery
    -
     
    -
    future() - Method in class org.apache.fluss.client.lookup.LookupQuery
    -
     
    -
    future() - Method in class org.apache.fluss.client.lookup.PrefixLookupQuery
    -
     
    -
    -A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-7.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-7.html deleted file mode 100644 index da9f8f0acc..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-7.html +++ /dev/null @@ -1,659 +0,0 @@ - - - - - -G-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages - - -

    G

    -
    -
    generateHybridLakeFlussSplits() - Method in class org.apache.fluss.flink.lake.LakeSplitGenerator
    -
    -
    Return A list of hybrid lake snapshot LakeSnapshotSplit, LakeSnapshotAndFlussLogSplit and the corresponding Fluss LogSplit based on the lake - snapshot.
    -
    -
    generateTableSplits(TablePath) - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplitGenerator
    -
     
    -
    getAdmin() - Method in interface org.apache.fluss.client.Connection
    -
    -
    Retrieve a new Admin client to administer a Fluss cluster.
    -
    -
    getAdmin() - Method in class org.apache.fluss.client.FlussConnection
    -
     
    -
    getArity() - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
    -
     
    -
    getArity() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
    -
     
    -
    getAssignedBuckets() - Method in class org.apache.fluss.flink.source.state.SourceEnumeratorState
    -
     
    -
    getAssignedPartitions() - Method in class org.apache.fluss.flink.source.state.SourceEnumeratorState
    -
     
    -
    getBinary(int, int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
    -
     
    -
    getBoolean(int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
    -
     
    -
    getBoundedness() - Method in class org.apache.fluss.flink.source.FlinkSource
    -
     
    -
    getBoundedness() - Method in class org.apache.fluss.flink.tiering.source.TieringSource
    -
     
    -
    getBoundedSplitScanner(SourceSplitBase) - Method in class org.apache.fluss.flink.lake.LakeSplitReaderGenerator
    -
     
    -
    getBucketIds() - Method in class org.apache.fluss.client.metadata.KvSnapshots
    -
     
    -
    getBucketKey() - Method in class org.apache.fluss.client.write.WriteRecord
    -
     
    -
    getBucketKeyIndexes() - Method in class org.apache.fluss.flink.source.FlinkTableSource
    -
     
    -
    getBucketKeyIndexes(ReadableConfig, RowType) - Static method in class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils
    -
     
    -
    getBucketKeys() - Method in class org.apache.fluss.flink.sink.FlinkTableSink
    -
     
    -
    getBucketKeys(ReadableConfig) - Static method in class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils
    -
     
    -
    getBucketLocation(TableBucket) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
    -
     
    -
    getBucketOffsets(String, Collection<Integer>, OffsetsInitializer.BucketOffsetsRetriever) - Method in class org.apache.fluss.flink.source.enumerator.initializer.LatestOffsetsInitializer
    -
     
    -
    getBucketOffsets(String, Collection<Integer>, OffsetsInitializer.BucketOffsetsRetriever) - Method in class org.apache.fluss.flink.source.enumerator.initializer.NoStoppingOffsetsInitializer
    -
     
    -
    getBucketOffsets(String, Collection<Integer>, OffsetsInitializer.BucketOffsetsRetriever) - Method in interface org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer
    -
    -
    Get the initial offsets for the given fluss buckets.
    -
    -
    getBucketOffsets(String, Collection<Integer>, OffsetsInitializer.BucketOffsetsRetriever) - Method in class org.apache.fluss.flink.source.enumerator.initializer.SnapshotOffsetsInitializer
    -
    -
    For table with primary key.
    -
    -
    getBucketOffsets(String, Collection<Integer>, OffsetsInitializer.BucketOffsetsRetriever) - Method in class org.apache.fluss.flink.source.enumerator.initializer.TimestampOffsetsInitializer
    -
     
    -
    getByte(int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
    -
     
    -
    getBytes(int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
    -
     
    -
    getCache() - Method in class org.apache.fluss.flink.source.FlinkTableSource
    -
     
    -
    getChangelogMode() - Method in class org.apache.fluss.flink.source.FlinkTableSource
    -
     
    -
    getChangelogMode(ChangelogMode) - Method in class org.apache.fluss.flink.sink.FlinkTableSink
    -
     
    -
    getChangeType() - Method in class org.apache.fluss.client.table.scanner.ScanRecord
    -
     
    -
    getChar(int, int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
    -
     
    -
    getClientMetricGroup() - Method in class org.apache.fluss.client.FlussConnection
    -
     
    -
    getCluster() - Method in class org.apache.fluss.client.metadata.MetadataUpdater
    -
     
    -
    getConfiguration() - Method in interface org.apache.fluss.client.Connection
    -
    -
    Retrieve the configuration used to create this connection.
    -
    -
    getConfiguration() - Method in class org.apache.fluss.client.FlussConnection
    -
     
    -
    getCoordinatorServer() - Method in class org.apache.fluss.client.metadata.MetadataUpdater
    -
     
    -
    getCount() - Method in class org.apache.fluss.flink.metrics.FlinkCounter
    -
     
    -
    getCount() - Method in class org.apache.fluss.flink.metrics.FlinkHistogram
    -
     
    -
    getCount() - Method in class org.apache.fluss.flink.metrics.FlinkMeter
    -
     
    -
    getCurrentLakeSplitIndex() - Method in class org.apache.fluss.flink.lake.reader.IndexedLakeSplitRecordIterator
    -
     
    -
    getCurrentLakeSplitIndex() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
    -
     
    -
    getCurrentSplitIndex() - Method in class org.apache.fluss.flink.source.reader.RecordAndPos
    -
     
    -
    getDatabase(String) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    getDatabaseInfo(String) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    Get the database with the given database name asynchronously.
    -
    -
    getDatabaseInfo(String) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    getDatalakeFormat(ReadableConfig) - Static method in class org.apache.fluss.flink.utils.DataLakeUtils
    -
     
    -
    getDecimal(int, int, int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
    -
     
    -
    getDescription() - Method in enum org.apache.fluss.flink.FlinkConnectorOptions.ScanStartupMode
    -
     
    -
    getDouble(int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
    -
     
    -
    getDownstreamSubtaskStateMapper() - Method in class org.apache.fluss.flink.sink.FlinkStreamPartitioner
    -
     
    -
    getEnumeratorCheckpointSerializer() - Method in class org.apache.fluss.flink.source.FlinkSource
    -
     
    -
    getEnumeratorCheckpointSerializer() - Method in class org.apache.fluss.flink.tiering.source.TieringSource
    -
     
    -
    getEstimatedBatchSize(PhysicalTablePath) - Method in class org.apache.fluss.client.write.DynamicWriteBatchSizeEstimator
    -
     
    -
    getEstimatedSizeInBytes() - Method in class org.apache.fluss.client.write.WriteRecord
    -
    -
    Get the estimated size in bytes of the record with batch header.
    -
    -
    getFactory() - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    getFieldCount() - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
    -
     
    -
    getFileLogRecords(int) - Method in class org.apache.fluss.client.table.scanner.log.RemoteLogDownloadFuture
    -
     
    -
    getFloat(int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
    -
     
    -
    getFlussMetric(String) - Method in class org.apache.fluss.flink.metrics.FlinkMetricRegistry
    -
    -
    Exposes the metrics of Fluss metics group for flink.
    -
    -
    getFsPathAndFileName(FsPath, RemoteLogSegment) - Static method in class org.apache.fluss.client.table.scanner.log.RemoteLogDownloader
    -
     
    -
    getFunction(ObjectPath) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    getFutures() - Method in class org.apache.fluss.client.admin.CreateAclsResult
    -
    -
    Gets the map of ACL bindings to their associated futures.
    -
    -
    getGroupName(CharacterFilter) - Method in class org.apache.fluss.client.metrics.ScannerMetricGroup
    -
     
    -
    getGroupName(CharacterFilter) - Method in class org.apache.fluss.client.metrics.WriterMetricGroup
    -
     
    -
    getHost(String) - Static method in class org.apache.fluss.client.utils.ClientUtils
    -
    -
    Extracts the hostname from a "host:port" address string.
    -
    -
    getInt(int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
    -
     
    -
    getKey() - Method in class org.apache.fluss.client.write.WriteRecord
    -
     
    -
    getKvSnapshotMetadata(TableBucket, long) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    Get the kv snapshot metadata of the given kv snapshot asynchronously.
    -
    -
    getKvSnapshotMetadata(TableBucket, long) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    getLakeSplit() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotSplit
    -
     
    -
    getLakeSplits() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
    -
     
    -
    getLakeTable(String, String, Configuration) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    getLatestKvSnapshots(TablePath) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    Get the latest kv snapshots of the given table asynchronously.
    -
    -
    getLatestKvSnapshots(TablePath) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    getLatestKvSnapshots(TablePath, String) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    Get the latest kv snapshots of the given table partition asynchronously.
    -
    -
    getLatestKvSnapshots(TablePath, String) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    getLatestLakeSnapshot(TablePath) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    Get table lake snapshot info of the given table asynchronously.
    -
    -
    getLatestLakeSnapshot(TablePath) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    getLocalTimeZone(String) - Static method in class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils
    -
     
    -
    getLogOffset() - Method in class org.apache.fluss.client.metadata.KvSnapshotMetadata
    -
     
    -
    getLogOffset(int) - Method in class org.apache.fluss.client.metadata.KvSnapshots
    -
    -
    Get the log offset to read after the snapshot for this kv tablet (bucket), or empty if there - are no snapshots.
    -
    -
    getLogOffsetOfSnapshot() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSnapshotSplit
    -
     
    -
    getLogStartingOffset() - Method in class org.apache.fluss.flink.source.split.HybridSnapshotLogSplit
    -
     
    -
    getLong(int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
    -
     
    -
    getLookupColumnNames() - Method in class org.apache.fluss.client.lookup.PrefixLookup
    -
     
    -
    getLookupKeyIndexes() - Method in class org.apache.fluss.flink.source.lookup.LookupNormalizer
    -
    -
    Returns the indexes of the normalized lookup keys.
    -
    -
    getLookupRuntimeProvider(LookupTableSource.LookupContext) - Method in class org.apache.fluss.flink.source.FlinkTableSource
    -
     
    -
    getLookupType() - Method in class org.apache.fluss.flink.source.lookup.LookupNormalizer
    -
    -
    Returns the lookup type (primary key lookup, or prefix key lookup) requested from Flink to - Fluss.
    -
    -
    getMetadataUpdater() - Method in class org.apache.fluss.client.FlussConnection
    -
     
    -
    getMetricGroup() - Method in class org.apache.fluss.flink.source.deserializer.DeserializerInitContextImpl
    -
     
    -
    getMetricGroup() - Method in interface org.apache.fluss.flink.source.deserializer.FlussDeserializationSchema.InitializationContext
    -
    -
    Returns the metric group for the parallel subtask of the source that runs this FlussDeserializationSchema.
    -
    -
    getNumberOfSplits() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
    -
     
    -
    getNumberReporters() - Method in class org.apache.fluss.flink.metrics.FlinkMetricRegistry
    -
     
    -
    getOneAvailableTabletServerNode(Cluster) - Static method in class org.apache.fluss.client.utils.MetadataUtils
    -
     
    -
    getOperationType() - Method in class org.apache.fluss.flink.row.RowWithOp
    -
    -
    Returns the operation type associated with this row.
    -
    -
    getOrCreateLookupClient() - Method in class org.apache.fluss.client.FlussConnection
    -
     
    -
    getOrCreateRemoteFileDownloader() - Method in class org.apache.fluss.client.FlussConnection
    -
     
    -
    getOrCreateWriterClient() - Method in class org.apache.fluss.client.FlussConnection
    -
     
    -
    getPartialUpdateColumns() - Method in class org.apache.fluss.client.table.writer.UpsertWrite
    -
    -
    Returns the columns to partial update.
    -
    -
    getPartition(ObjectPath, CatalogPartitionSpec) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    getPartition(InternalRow) - Method in class org.apache.fluss.client.table.getter.PartitionGetter
    -
     
    -
    getPartitionColumnStatistics(ObjectPath, CatalogPartitionSpec) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    getPartitionId() - Method in class org.apache.fluss.client.metadata.KvSnapshots
    -
     
    -
    getPartitionId(PhysicalTablePath) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
    -
     
    -
    getPartitionId(InternalRow, PartitionGetter, TablePath, MetadataUpdater) - Static method in class org.apache.fluss.client.utils.ClientUtils
    -
    -
    Return the id of the partition the row belongs to.
    -
    -
    getPartitionIdOrElseThrow(PhysicalTablePath) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
    -
     
    -
    getPartitionKeyIndexes() - Method in class org.apache.fluss.flink.source.FlinkTableSource
    -
     
    -
    getPartitionName() - Method in class org.apache.fluss.flink.source.split.SourceSplitBase
    -
     
    -
    getPartitionName() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
    -
     
    -
    getPartitionNameById() - Method in class org.apache.fluss.client.metadata.LakeSnapshot
    -
     
    -
    getPartitionStatistics(ObjectPath, CatalogPartitionSpec) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    getPhysicalPath(InternalRow) - Method in class org.apache.fluss.client.table.writer.AbstractTableWriter
    -
     
    -
    getPhysicalTablePath() - Method in class org.apache.fluss.client.write.WriteRecord
    -
     
    -
    getPhysicalTablePathByIds(Collection<Long>, Collection<TablePartition>) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
    -
    -
    Get the table physical paths by table ids and partition ids.
    -
    -
    getPhysicalTablePathsInBatches() - Method in class org.apache.fluss.client.write.RecordAccumulator
    -
     
    -
    getPort(String) - Static method in class org.apache.fluss.client.utils.ClientUtils
    -
    -
    Extracts the port number from a "host:port" address string.
    -
    -
    getPrimaryKeyIndexes() - Method in class org.apache.fluss.flink.source.FlinkTableSource
    -
     
    -
    getProcedure(ObjectPath) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    getProcedure(Admin, ObjectPath) - Static method in class org.apache.fluss.flink.procedure.ProcedureManager
    -
     
    -
    getProducedType() - Method in class org.apache.fluss.flink.source.FlinkSource
    -
     
    -
    getProducedType(RowType) - Method in interface org.apache.fluss.flink.source.deserializer.FlussDeserializationSchema
    -
    -
    Gets the data type (as a TypeInformation) produced by this deserializer.
    -
    -
    getProducedType(RowType) - Method in class org.apache.fluss.flink.source.deserializer.JsonStringDeserializationSchema
    -
    -
    Returns the TypeInformation for the produced String type.
    -
    -
    getProducedType(RowType) - Method in class org.apache.fluss.flink.source.deserializer.RowDataDeserializationSchema
    -
    -
    Returns the TypeInformation for the produced RowData type.
    -
    -
    getProjectedFields() - Method in class org.apache.fluss.client.table.scanner.log.LogScan
    -
     
    -
    getQueueTimeMs() - Method in class org.apache.fluss.client.write.WriteBatch
    -
     
    -
    getRandomTabletServer() - Method in class org.apache.fluss.client.metadata.MetadataUpdater
    -
     
    -
    getRate() - Method in class org.apache.fluss.flink.metrics.FlinkMeter
    -
     
    -
    getRecordCount() - Method in class org.apache.fluss.client.write.WriteBatch
    -
     
    -
    getRecordsToSkip() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
    -
     
    -
    getRecordsToSkip() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotSplit
    -
     
    -
    getRecycleCallback() - Method in class org.apache.fluss.client.table.scanner.log.RemoteLogDownloadFuture
    -
     
    -
    getRemainingHybridLakeFlussSplits() - Method in class org.apache.fluss.flink.source.state.SourceEnumeratorState
    -
     
    -
    getRemovedPartitions() - Method in class org.apache.fluss.flink.source.event.PartitionsRemovedEvent
    -
     
    -
    getRemovedTableBuckets() - Method in class org.apache.fluss.flink.source.event.PartitionBucketsUnsubscribedEvent
    -
     
    -
    getRequestFuture() - Method in class org.apache.fluss.client.write.WriteBatch
    -
     
    -
    getRow() - Method in class org.apache.fluss.client.table.scanner.ScanRecord
    -
     
    -
    getRow() - Method in class org.apache.fluss.client.write.WriteRecord
    -
     
    -
    getRow() - Method in class org.apache.fluss.flink.row.RowWithOp
    -
    -
    Returns the internal row data.
    -
    -
    getRowList() - Method in class org.apache.fluss.client.lookup.LookupResult
    -
     
    -
    getRowSchema() - Method in interface org.apache.fluss.flink.sink.serializer.FlussSerializationSchema.InitializationContext
    -
    -
    Returns the target row schema.
    -
    -
    getRowSchema() - Method in class org.apache.fluss.flink.sink.serializer.SerializerInitContextImpl
    -
     
    -
    getRowSchema() - Method in class org.apache.fluss.flink.source.deserializer.DeserializerInitContextImpl
    -
     
    -
    getRowSchema() - Method in interface org.apache.fluss.flink.source.deserializer.FlussDeserializationSchema.InitializationContext
    -
    -
    Returns the schema of the LogRecord.getRow().
    -
    -
    getRpcClient() - Method in class org.apache.fluss.client.FlussConnection
    -
     
    -
    getScanRuntimeProvider(ScanTableSource.ScanContext) - Method in class org.apache.fluss.flink.source.FlinkTableSource
    -
     
    -
    getSecurityConfigs() - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    getServerNodes() - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    Get the current server node information.
    -
    -
    getServerNodes() - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    getShort(int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
    -
     
    -
    getSingletonRow() - Method in class org.apache.fluss.client.lookup.LookupResult
    -
     
    -
    getSinkRuntimeProvider(DynamicTableSink.Context) - Method in class org.apache.fluss.flink.sink.FlinkTableSink
    -
     
    -
    getSnapshotFiles() - Method in class org.apache.fluss.client.metadata.KvSnapshotMetadata
    -
     
    -
    getSnapshotId() - Method in class org.apache.fluss.client.metadata.LakeSnapshot
    -
     
    -
    getSnapshotId() - Method in class org.apache.fluss.flink.source.split.SnapshotSplit
    -
     
    -
    getSnapshotId() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSnapshotSplit
    -
     
    -
    getSnapshotId(int) - Method in class org.apache.fluss.client.metadata.KvSnapshots
    -
    -
    Get the latest snapshot id for this kv tablet (bucket), or empty if there are no snapshots.
    -
    -
    getSourceReaderMetricGroup() - Method in class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
    -
     
    -
    getSplitIndex() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotSplit
    -
     
    -
    getSplitOwner(SourceSplitBase) - Method in class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
    -
    -
    Returns the index of the target subtask that a specific split should be assigned to.
    -
    -
    getSplitSerializer() - Method in class org.apache.fluss.flink.source.FlinkSource
    -
     
    -
    getSplitSerializer() - Method in class org.apache.fluss.flink.tiering.source.TieringSource
    -
     
    -
    getStartingOffset() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
    -
     
    -
    getStartingOffset() - Method in class org.apache.fluss.flink.source.split.LogSplit
    -
     
    -
    getStartingOffset() - Method in class org.apache.fluss.flink.tiering.source.split.TieringLogSplit
    -
     
    -
    getStartupOptions(ReadableConfig, ZoneId) - Static method in class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils
    -
     
    -
    getStatistics() - Method in class org.apache.fluss.flink.metrics.FlinkHistogram
    -
     
    -
    getStoppingOffset() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
    -
     
    -
    getStoppingOffset() - Method in class org.apache.fluss.flink.source.split.LogSplit
    -
     
    -
    getStoppingOffset() - Method in class org.apache.fluss.flink.tiering.source.split.TieringLogSplit
    -
     
    -
    getStreamOperatorClass(ClassLoader) - Method in class org.apache.fluss.flink.tiering.committer.TieringCommitOperatorFactory
    -
     
    -
    getString(int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
    -
     
    -
    getTable(ObjectPath) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    getTable(ObjectPath) - Method in class org.apache.fluss.flink.lake.LakeCatalog
    -
     
    -
    getTable(TablePath) - Method in interface org.apache.fluss.client.Connection
    -
    -
    Retrieve a new Table client to operate data in table.
    -
    -
    getTable(TablePath) - Method in class org.apache.fluss.client.FlussConnection
    -
     
    -
    getTableBucket() - Method in class org.apache.fluss.flink.source.split.SourceSplitBase
    -
     
    -
    getTableBucket() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
    -
     
    -
    getTableBucketsOffset() - Method in class org.apache.fluss.client.metadata.LakeSnapshot
    -
     
    -
    getTableColumnStatistics(ObjectPath) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    getTableId() - Method in class org.apache.fluss.client.metadata.KvSnapshots
    -
     
    -
    getTableId() - Method in class org.apache.fluss.flink.tiering.event.FailedTieringEvent
    -
     
    -
    getTableId() - Method in class org.apache.fluss.flink.tiering.event.FinishedTieringEvent
    -
     
    -
    getTableId(TablePath) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
    -
     
    -
    getTableInfo() - Method in class org.apache.fluss.client.table.FlussTable
    -
     
    -
    getTableInfo() - Method in interface org.apache.fluss.client.table.Table
    -
    -
    Get the TableInfo for this table.
    -
    -
    getTableInfo(TablePath) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    Get the table with the given table path asynchronously.
    -
    -
    getTableInfo(TablePath) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    getTableInfoOrElseThrow(long) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
    -
     
    -
    getTableInfoOrElseThrow(TablePath) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
    -
     
    -
    getTablePath() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
    -
     
    -
    getTableSchema(TablePath) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    Get the latest table schema of the given table asynchronously.
    -
    -
    getTableSchema(TablePath) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    getTableSchema(TablePath, int) - Method in interface org.apache.fluss.client.admin.Admin
    -
    -
    Get the specific table schema of the given table by schema id asynchronously.
    -
    -
    getTableSchema(TablePath, int) - Method in class org.apache.fluss.client.admin.FlussAdmin
    -
     
    -
    getTableStatistics(ObjectPath) - Method in class org.apache.fluss.flink.catalog.FlinkCatalog
    -
     
    -
    getTargetColumns() - Method in class org.apache.fluss.client.write.KvWriteBatch
    -
     
    -
    getTargetColumns() - Method in class org.apache.fluss.client.write.WriteRecord
    -
     
    -
    getTimestamp() - Method in class org.apache.fluss.client.admin.OffsetSpec.TimestampSpec
    -
     
    -
    getTimestampLtz(int, int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
    -
     
    -
    getTimestampNtz(int, int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
    -
     
    -
    getTotalFields() - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
    -
     
    -
    getTotalFields() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
    -
     
    -
    getTypeClass() - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
    -
     
    -
    getTypeClass() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
    -
     
    -
    getUserCodeClassLoader() - Method in class org.apache.fluss.flink.source.deserializer.DeserializerInitContextImpl
    -
     
    -
    getUserCodeClassLoader() - Method in interface org.apache.fluss.flink.source.deserializer.FlussDeserializationSchema.InitializationContext
    -
    -
    Gets the UserCodeClassLoader to load classes that are not in system's classpath, - but are part of the jar file of a user job.
    -
    -
    getValue() - Method in class org.apache.fluss.flink.metrics.FlinkGauge
    -
     
    -
    getVersion() - Method in class org.apache.fluss.flink.source.split.SourceSplitSerializer
    -
     
    -
    getVersion() - Method in class org.apache.fluss.flink.source.state.FlussSourceEnumeratorStateSerializer
    -
     
    -
    getVersion() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplitSerializer
    -
     
    -
    getVersion() - Method in class org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorStateSerializer
    -
     
    -
    getVersion() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultSerializer
    -
     
    -
    getWriteFormat() - Method in class org.apache.fluss.client.write.WriteRecord
    -
     
    -
    -A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-8.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-8.html deleted file mode 100644 index 9212594a68..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-8.html +++ /dev/null @@ -1,238 +0,0 @@ - - - - - -H-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages - - -

    H

    -
    -
    handleSourceEvent(int, SourceEvent) - Method in class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
    -
     
    -
    handleSourceEvent(int, SourceEvent) - Method in class org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator
    -
     
    -
    handleSourceEvents(SourceEvent) - Method in class org.apache.fluss.flink.source.reader.FlinkSourceReader
    -
     
    -
    handleSplitRequest(int, String) - Method in class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
    -
     
    -
    handleSplitRequest(int, String) - Method in class org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator
    -
     
    -
    handleSplitsChanges(SplitsChange<SourceSplitBase>) - Method in class org.apache.fluss.flink.source.reader.FlinkSourceSplitReader
    -
     
    -
    handleSplitsChanges(SplitsChange<TieringSplit>) - Method in class org.apache.fluss.flink.tiering.source.TieringSplitReader
    -
     
    -
    hasAvailableFetches() - Method in class org.apache.fluss.client.table.scanner.log.LogFetcher
    -
    -
    Return whether we have any completed fetches that are fetch-able.
    -
    -
    hasBatchSequence() - Method in class org.apache.fluss.client.write.WriteBatch
    -
     
    -
    HashBucketAssigner - Class in org.apache.fluss.client.write
    -
    -
    Hash bucket assigner.
    -
    -
    HashBucketAssigner(int) - Constructor for class org.apache.fluss.client.write.HashBucketAssigner
    -
     
    -
    HashBucketAssigner(int, BucketingFunction) - Constructor for class org.apache.fluss.client.write.HashBucketAssigner
    -
     
    -
    hashCode() - Method in class org.apache.fluss.client.lookup.LookupResult
    -
     
    -
    hashCode() - Method in class org.apache.fluss.client.table.scanner.ScanRecord
    -
     
    -
    hashCode() - Method in class org.apache.fluss.flink.row.RowWithOp
    -
    -
    Returns a hash code value for the object.
    -
    -
    hashCode() - Method in class org.apache.fluss.flink.source.event.PartitionBucketsUnsubscribedEvent
    -
     
    -
    hashCode() - Method in class org.apache.fluss.flink.source.reader.RecordAndPos
    -
     
    -
    hashCode() - Method in class org.apache.fluss.flink.source.split.HybridSnapshotLogSplit
    -
     
    -
    hashCode() - Method in class org.apache.fluss.flink.source.split.LogSplit
    -
     
    -
    hashCode() - Method in class org.apache.fluss.flink.source.split.SnapshotSplit
    -
     
    -
    hashCode() - Method in class org.apache.fluss.flink.source.split.SourceSplitBase
    -
     
    -
    hashCode() - Method in class org.apache.fluss.flink.source.state.SourceEnumeratorState
    -
     
    -
    hashCode() - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
    -
     
    -
    hashCode() - Method in class org.apache.fluss.flink.tiering.source.split.TieringLogSplit
    -
     
    -
    hashCode() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSnapshotSplit
    -
     
    -
    hashCode() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
    -
     
    -
    hashCode() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
    -
     
    -
    hasIncomplete() - Method in class org.apache.fluss.client.write.RecordAccumulator
    -
    -
    Check whether there are any pending batches (whether sent or unsent).
    -
    -
    hasNext() - Method in class org.apache.fluss.flink.lake.reader.IndexedLakeSplitRecordIterator
    -
     
    -
    hasUnDrained() - Method in class org.apache.fluss.client.write.RecordAccumulator
    -
    -
    Check whether there are any batches which haven't been drained.
    -
    -
    HYBRID_SNAPSHOT_SPLIT_FLAG - Static variable in class org.apache.fluss.flink.source.split.SourceSplitBase
    -
     
    -
    HybridSnapshotLogSplit - Class in org.apache.fluss.flink.source.split
    -
    -
    The hybrid split for first reading the snapshot files and then switch to read the cdc log from a - specified offset.
    -
    -
    HybridSnapshotLogSplit(TableBucket, String, long, long) - Constructor for class org.apache.fluss.flink.source.split.HybridSnapshotLogSplit
    -
     
    -
    HybridSnapshotLogSplit(TableBucket, String, long, long, boolean, long) - Constructor for class org.apache.fluss.flink.source.split.HybridSnapshotLogSplit
    -
     
    -
    HybridSnapshotLogSplitState - Class in org.apache.fluss.flink.source.split
    -
    - -
    -
    HybridSnapshotLogSplitState(HybridSnapshotLogSplit) - Constructor for class org.apache.fluss.flink.source.split.HybridSnapshotLogSplitState
    -
     
    -
    -A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-9.html b/website/static/javadoc/0.8-SNAPSHOT/index-files/index-9.html deleted file mode 100644 index f3d2fc54cf..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index-files/index-9.html +++ /dev/null @@ -1,327 +0,0 @@ - - - - - -I-Index (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages - - -

    I

    -
    -
    IdempotenceBucketEntry - Class in org.apache.fluss.client.write
    -
    -
    Entry to store the idempotence information of each table-bucket.
    -
    -
    IdempotenceBucketMap - Class in org.apache.fluss.client.write
    -
    -
    Map to manage IdempotenceBucketEntry of all table-bucket.
    -
    -
    IdempotenceBucketMap() - Constructor for class org.apache.fluss.client.write.IdempotenceBucketMap
    -
     
    -
    IdempotenceManager - Class in org.apache.fluss.client.write
    -
    -
    A class which manages the idempotence in writer keeps the state necessary to ensure idempotent of - writer.
    -
    -
    IdempotenceManager(boolean, int, TabletServerGateway) - Constructor for class org.apache.fluss.client.write.IdempotenceManager
    -
     
    -
    IDENTIFIER - Static variable in class org.apache.fluss.flink.catalog.FlinkCatalogFactory
    -
     
    -
    IGNORE - org.apache.fluss.flink.row.OperationType
    -
    -
    Represents an operation that should be ignored.
    -
    -
    inc() - Method in class org.apache.fluss.flink.metrics.FlinkCounter
    -
     
    -
    inc(long) - Method in class org.apache.fluss.flink.metrics.FlinkCounter
    -
     
    -
    INDEXED_LOG - org.apache.fluss.client.write.WriteFormat
    -
     
    -
    IndexedLakeSplitRecordIterator - Class in org.apache.fluss.flink.lake.reader
    -
    -
    An iterator wrapper that converts LogRecord objects to InternalRow objects while tracking the - current LakeSplit index being processed.
    -
    -
    IndexedLakeSplitRecordIterator(CloseableIterator<LogRecord>, int) - Constructor for class org.apache.fluss.flink.lake.reader.IndexedLakeSplitRecordIterator
    -
     
    -
    IndexedLogWriteBatch - Class in org.apache.fluss.client.write
    -
    -
    A batch of log records managed in INDEXED format that is or will be sent to server by ProduceLogRequest.
    -
    -
    IndexedLogWriteBatch(int, PhysicalTablePath, int, int, AbstractPagedOutputView, long) - Constructor for class org.apache.fluss.client.write.IndexedLogWriteBatch
    -
     
    -
    INITIAL_OFFSET - Static variable in class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
    -
     
    -
    initialize(SinkWriterMetricGroup) - Method in class org.apache.fluss.flink.sink.writer.AppendSinkWriter
    -
     
    -
    initialize(SinkWriterMetricGroup) - Method in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
    -
     
    -
    initialize(SinkWriterMetricGroup) - Method in class org.apache.fluss.flink.sink.writer.UpsertSinkWriter
    -
     
    -
    initializedState(SourceSplitBase) - Static method in class org.apache.fluss.flink.lake.LakeSplitStateInitializer
    -
     
    -
    initializedState(SourceSplitBase) - Method in class org.apache.fluss.flink.source.reader.FlinkSourceReader
    -
     
    -
    initializedState(TieringSplit) - Method in class org.apache.fluss.flink.tiering.source.TieringSourceReader
    -
     
    -
    initiateClose() - Method in class org.apache.fluss.client.write.Sender
    -
    -
    Start closing the sender (won't actually complete until all data is sent out).
    -
    -
    initMetrics() - Method in class org.apache.fluss.flink.sink.writer.FlinkSinkWriter
    -
     
    -
    INSTANCE - Static variable in class org.apache.fluss.flink.tiering.source.split.TieringSplitSerializer
    -
     
    -
    INSTANCE - Static variable in class org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorStateSerializer
    -
     
    -
    internalCall(String, String, String, String, String) - Method in class org.apache.fluss.flink.procedure.AbstractAclProcedure
    -
    -
    Invokes the ACL operation as a Flink table procedure.
    -
    -
    invalidPhysicalTableBucketMeta(Set<PhysicalTablePath>) - Method in class org.apache.fluss.client.metadata.MetadataUpdater
    -
    -
    Invalid the bucket metadata for the given physical table paths.
    -
    -
    isBasicType() - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
    -
     
    -
    isBasicType() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
    -
     
    -
    isClosed() - Method in class org.apache.fluss.client.write.ArrowLogWriteBatch
    -
     
    -
    isClosed() - Method in class org.apache.fluss.client.write.IndexedLogWriteBatch
    -
     
    -
    isClosed() - Method in class org.apache.fluss.client.write.KvWriteBatch
    -
     
    -
    isClosed() - Method in class org.apache.fluss.client.write.WriteBatch
    -
    -
    check if the batch is closed.
    -
    -
    isDatabaseAlreadyExist(Throwable) - Static method in class org.apache.fluss.flink.utils.CatalogExceptionUtils
    -
     
    -
    isDatabaseNotEmpty(Throwable) - Static method in class org.apache.fluss.flink.utils.CatalogExceptionUtils
    -
     
    -
    isDatabaseNotExist(Throwable) - Static method in class org.apache.fluss.flink.utils.CatalogExceptionUtils
    -
     
    -
    isDelete() - Method in class org.apache.fluss.flink.lake.reader.KeyValueRow
    -
     
    -
    isDone() - Method in class org.apache.fluss.client.table.scanner.log.RemoteLogDownloadFuture
    -
     
    -
    isEmpty() - Method in class org.apache.fluss.client.table.scanner.log.ScanRecords
    -
     
    -
    isHybridSnapshotLogSplit() - Method in class org.apache.fluss.flink.source.split.SourceSplitBase
    -
    -
    Checks whether this split is a hybrid snapshot and log split.
    -
    -
    isHybridSnapshotLogSplitState() - Method in class org.apache.fluss.flink.source.split.SourceSplitState
    -
    -
    Checks whether this split state is a hybrid snapshot log split state.
    -
    -
    isKeyType() - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
    -
     
    -
    isKeyType() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
    -
     
    -
    isLakeSplit() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
    -
     
    -
    isLakeSplit() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotSplit
    -
     
    -
    isLakeSplit() - Method in class org.apache.fluss.flink.source.split.SourceSplitBase
    -
     
    -
    isLakeSplit() - Method in class org.apache.fluss.flink.source.split.SourceSplitState
    -
     
    -
    isLakeSplitFinished() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
    -
     
    -
    isLakeTableAlreadyExist(Throwable) - Static method in class org.apache.fluss.flink.utils.CatalogExceptionUtils
    -
     
    -
    isLogSplit() - Method in class org.apache.fluss.flink.source.split.SourceSplitBase
    -
    -
    Checks whether this split is a log split.
    -
    -
    isLogSplitState() - Method in class org.apache.fluss.flink.source.split.SourceSplitState
    -
    -
    Checks whether this split state is a log split state.
    -
    -
    isMatch(RowData) - Method in class org.apache.fluss.flink.source.lookup.LookupNormalizer.RemainingFilter
    -
     
    -
    isNullAt(int) - Method in class org.apache.fluss.flink.row.FlinkAsFlussRow
    -
     
    -
    ISO_8601 - org.apache.fluss.flink.utils.TimestampFormat
    -
    -
    Options to specify TIMESTAMP/TIMESTAMP_WITH_LOCAL_ZONE format.
    -
    -
    isPartitionAlreadyExists(Throwable) - Static method in class org.apache.fluss.flink.utils.CatalogExceptionUtils
    -
     
    -
    isPartitionInvalid(Throwable) - Static method in class org.apache.fluss.flink.utils.CatalogExceptionUtils
    -
     
    -
    isPartitionNotExist(Throwable) - Static method in class org.apache.fluss.flink.utils.CatalogExceptionUtils
    -
     
    -
    isPointwise() - Method in class org.apache.fluss.flink.sink.FlinkStreamPartitioner
    -
     
    -
    isRunning() - Method in class org.apache.fluss.client.write.Sender
    -
     
    -
    isSnapshotFinished() - Method in class org.apache.fluss.flink.source.split.HybridSnapshotLogSplit
    -
     
    -
    isStreaming() - Method in class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit
    -
     
    -
    isTableAlreadyExist(Throwable) - Static method in class org.apache.fluss.flink.utils.CatalogExceptionUtils
    -
     
    -
    isTableInvalid(Throwable) - Static method in class org.apache.fluss.flink.utils.CatalogExceptionUtils
    -
     
    -
    isTableNotExist(Throwable) - Static method in class org.apache.fluss.flink.utils.CatalogExceptionUtils
    -
     
    -
    isTableNotPartitioned(Throwable) - Static method in class org.apache.fluss.flink.utils.CatalogExceptionUtils
    -
     
    -
    isTieringLogSplit() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
    -
    -
    Checks whether this split is a log split to tier.
    -
    -
    isTieringSnapshotSplit() - Method in class org.apache.fluss.flink.tiering.source.split.TieringSplit
    -
    -
    Checks whether this split is a primary key table split to tier.
    -
    -
    isTupleType() - Method in class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
    -
     
    -
    isTupleType() - Method in class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
    -
     
    -
    iterator() - Method in class org.apache.fluss.client.table.scanner.log.ScanRecords
    -
     
    -
    -A B C D E F G H I J K L M N O P Q R S T U V W 
    All Classes All Packages
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/index.html b/website/static/javadoc/0.8-SNAPSHOT/index.html deleted file mode 100644 index 3009a01bc3..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/index.html +++ /dev/null @@ -1,346 +0,0 @@ - - - - - -Overview (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Apache Fluss 0.8-SNAPSHOT API

    -
    -
    - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
    Packages 
    PackageDescription
    org.apache.fluss.client 
    org.apache.fluss.client.admin 
    org.apache.fluss.client.lookup 
    org.apache.fluss.client.metadata 
    org.apache.fluss.client.metrics 
    org.apache.fluss.client.table 
    org.apache.fluss.client.table.getter 
    org.apache.fluss.client.table.scanner 
    org.apache.fluss.client.table.scanner.batch 
    org.apache.fluss.client.table.scanner.log 
    org.apache.fluss.client.table.writer 
    org.apache.fluss.client.token 
    org.apache.fluss.client.utils 
    org.apache.fluss.client.write 
    org.apache.fluss.flink 
    org.apache.fluss.flink.adapter 
    org.apache.fluss.flink.catalog 
    org.apache.fluss.flink.lake 
    org.apache.fluss.flink.lake.reader 
    org.apache.fluss.flink.lake.split 
    org.apache.fluss.flink.lake.state 
    org.apache.fluss.flink.metrics 
    org.apache.fluss.flink.procedure 
    org.apache.fluss.flink.row 
    org.apache.fluss.flink.sink 
    org.apache.fluss.flink.sink.serializer 
    org.apache.fluss.flink.sink.writer 
    org.apache.fluss.flink.source 
    org.apache.fluss.flink.source.deserializer 
    org.apache.fluss.flink.source.emitter 
    org.apache.fluss.flink.source.enumerator 
    org.apache.fluss.flink.source.enumerator.initializer 
    org.apache.fluss.flink.source.event 
    org.apache.fluss.flink.source.lookup 
    org.apache.fluss.flink.source.metrics 
    org.apache.fluss.flink.source.reader 
    org.apache.fluss.flink.source.reader.fetcher 
    org.apache.fluss.flink.source.split 
    org.apache.fluss.flink.source.state 
    org.apache.fluss.flink.tiering 
    org.apache.fluss.flink.tiering.committer 
    org.apache.fluss.flink.tiering.event 
    org.apache.fluss.flink.tiering.source 
    org.apache.fluss.flink.tiering.source.enumerator 
    org.apache.fluss.flink.tiering.source.split 
    org.apache.fluss.flink.tiering.source.state 
    org.apache.fluss.flink.utils 
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/javadoc.sh b/website/static/javadoc/0.8-SNAPSHOT/javadoc.sh deleted file mode 100755 index 136f87af54..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/javadoc.sh +++ /dev/null @@ -1 +0,0 @@ -/Users/mehulbatra/.sdkman/candidates/java/11.0.28-zulu/zulu-11.jdk/Contents/Home/bin/javadoc -Xdoclint:none -J-Xmx2g -J--illegal-access=permit @options @packages \ No newline at end of file diff --git a/website/static/javadoc/0.8-SNAPSHOT/jquery-ui.overrides.css b/website/static/javadoc/0.8-SNAPSHOT/jquery-ui.overrides.css deleted file mode 100644 index facf852c27..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/jquery-ui.overrides.css +++ /dev/null @@ -1,35 +0,0 @@ -/* - * Copyright (c) 2020, 2022, Oracle and/or its affiliates. All rights reserved. - * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. - * - * This code is free software; you can redistribute it and/or modify it - * under the terms of the GNU General Public License version 2 only, as - * published by the Free Software Foundation. Oracle designates this - * particular file as subject to the "Classpath" exception as provided - * by Oracle in the LICENSE file that accompanied this code. - * - * This code is distributed in the hope that it will be useful, but WITHOUT - * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or - * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License - * version 2 for more details (a copy is included in the LICENSE file that - * accompanied this code). - * - * You should have received a copy of the GNU General Public License version - * 2 along with this work; if not, write to the Free Software Foundation, - * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. - * - * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA - * or visit www.oracle.com if you need additional information or have any - * questions. - */ - -.ui-state-active, -.ui-widget-content .ui-state-active, -.ui-widget-header .ui-state-active, -a.ui-button:active, -.ui-button:active, -.ui-button.ui-state-active:hover { - /* Overrides the color of selection used in jQuery UI */ - background: #F8981D; - border: 1px solid #F8981D; -} diff --git a/website/static/javadoc/0.8-SNAPSHOT/jquery/external/jquery/jquery.js b/website/static/javadoc/0.8-SNAPSHOT/jquery/external/jquery/jquery.js deleted file mode 100644 index 50937333b9..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/jquery/external/jquery/jquery.js +++ /dev/null @@ -1,10872 +0,0 @@ -/*! - * jQuery JavaScript Library v3.5.1 - * https://jquery.com/ - * - * Includes Sizzle.js - * https://sizzlejs.com/ - * - * Copyright JS Foundation and other contributors - * Released under the MIT license - * https://jquery.org/license - * - * Date: 2020-05-04T22:49Z - */ -( function( global, factory ) { - - "use strict"; - - if ( typeof module === "object" && typeof module.exports === "object" ) { - - // For CommonJS and CommonJS-like environments where a proper `window` - // is present, execute the factory and get jQuery. - // For environments that do not have a `window` with a `document` - // (such as Node.js), expose a factory as module.exports. - // This accentuates the need for the creation of a real `window`. - // e.g. var jQuery = require("jquery")(window); - // See ticket #14549 for more info. - module.exports = global.document ? - factory( global, true ) : - function( w ) { - if ( !w.document ) { - throw new Error( "jQuery requires a window with a document" ); - } - return factory( w ); - }; - } else { - factory( global ); - } - -// Pass this if window is not defined yet -} )( typeof window !== "undefined" ? window : this, function( window, noGlobal ) { - -// Edge <= 12 - 13+, Firefox <=18 - 45+, IE 10 - 11, Safari 5.1 - 9+, iOS 6 - 9.1 -// throw exceptions when non-strict code (e.g., ASP.NET 4.5) accesses strict mode -// arguments.callee.caller (trac-13335). But as of jQuery 3.0 (2016), strict mode should be common -// enough that all such attempts are guarded in a try block. -"use strict"; - -var arr = []; - -var getProto = Object.getPrototypeOf; - -var slice = arr.slice; - -var flat = arr.flat ? function( array ) { - return arr.flat.call( array ); -} : function( array ) { - return arr.concat.apply( [], array ); -}; - - -var push = arr.push; - -var indexOf = arr.indexOf; - -var class2type = {}; - -var toString = class2type.toString; - -var hasOwn = class2type.hasOwnProperty; - -var fnToString = hasOwn.toString; - -var ObjectFunctionString = fnToString.call( Object ); - -var support = {}; - -var isFunction = function isFunction( obj ) { - - // Support: Chrome <=57, Firefox <=52 - // In some browsers, typeof returns "function" for HTML elements - // (i.e., `typeof document.createElement( "object" ) === "function"`). - // We don't want to classify *any* DOM node as a function. - return typeof obj === "function" && typeof obj.nodeType !== "number"; - }; - - -var isWindow = function isWindow( obj ) { - return obj != null && obj === obj.window; - }; - - -var document = window.document; - - - - var preservedScriptAttributes = { - type: true, - src: true, - nonce: true, - noModule: true - }; - - function DOMEval( code, node, doc ) { - doc = doc || document; - - var i, val, - script = doc.createElement( "script" ); - - script.text = code; - if ( node ) { - for ( i in preservedScriptAttributes ) { - - // Support: Firefox 64+, Edge 18+ - // Some browsers don't support the "nonce" property on scripts. - // On the other hand, just using `getAttribute` is not enough as - // the `nonce` attribute is reset to an empty string whenever it - // becomes browsing-context connected. - // See https://github.com/whatwg/html/issues/2369 - // See https://html.spec.whatwg.org/#nonce-attributes - // The `node.getAttribute` check was added for the sake of - // `jQuery.globalEval` so that it can fake a nonce-containing node - // via an object. - val = node[ i ] || node.getAttribute && node.getAttribute( i ); - if ( val ) { - script.setAttribute( i, val ); - } - } - } - doc.head.appendChild( script ).parentNode.removeChild( script ); - } - - -function toType( obj ) { - if ( obj == null ) { - return obj + ""; - } - - // Support: Android <=2.3 only (functionish RegExp) - return typeof obj === "object" || typeof obj === "function" ? - class2type[ toString.call( obj ) ] || "object" : - typeof obj; -} -/* global Symbol */ -// Defining this global in .eslintrc.json would create a danger of using the global -// unguarded in another place, it seems safer to define global only for this module - - - -var - version = "3.5.1", - - // Define a local copy of jQuery - jQuery = function( selector, context ) { - - // The jQuery object is actually just the init constructor 'enhanced' - // Need init if jQuery is called (just allow error to be thrown if not included) - return new jQuery.fn.init( selector, context ); - }; - -jQuery.fn = jQuery.prototype = { - - // The current version of jQuery being used - jquery: version, - - constructor: jQuery, - - // The default length of a jQuery object is 0 - length: 0, - - toArray: function() { - return slice.call( this ); - }, - - // Get the Nth element in the matched element set OR - // Get the whole matched element set as a clean array - get: function( num ) { - - // Return all the elements in a clean array - if ( num == null ) { - return slice.call( this ); - } - - // Return just the one element from the set - return num < 0 ? this[ num + this.length ] : this[ num ]; - }, - - // Take an array of elements and push it onto the stack - // (returning the new matched element set) - pushStack: function( elems ) { - - // Build a new jQuery matched element set - var ret = jQuery.merge( this.constructor(), elems ); - - // Add the old object onto the stack (as a reference) - ret.prevObject = this; - - // Return the newly-formed element set - return ret; - }, - - // Execute a callback for every element in the matched set. - each: function( callback ) { - return jQuery.each( this, callback ); - }, - - map: function( callback ) { - return this.pushStack( jQuery.map( this, function( elem, i ) { - return callback.call( elem, i, elem ); - } ) ); - }, - - slice: function() { - return this.pushStack( slice.apply( this, arguments ) ); - }, - - first: function() { - return this.eq( 0 ); - }, - - last: function() { - return this.eq( -1 ); - }, - - even: function() { - return this.pushStack( jQuery.grep( this, function( _elem, i ) { - return ( i + 1 ) % 2; - } ) ); - }, - - odd: function() { - return this.pushStack( jQuery.grep( this, function( _elem, i ) { - return i % 2; - } ) ); - }, - - eq: function( i ) { - var len = this.length, - j = +i + ( i < 0 ? len : 0 ); - return this.pushStack( j >= 0 && j < len ? [ this[ j ] ] : [] ); - }, - - end: function() { - return this.prevObject || this.constructor(); - }, - - // For internal use only. - // Behaves like an Array's method, not like a jQuery method. - push: push, - sort: arr.sort, - splice: arr.splice -}; - -jQuery.extend = jQuery.fn.extend = function() { - var options, name, src, copy, copyIsArray, clone, - target = arguments[ 0 ] || {}, - i = 1, - length = arguments.length, - deep = false; - - // Handle a deep copy situation - if ( typeof target === "boolean" ) { - deep = target; - - // Skip the boolean and the target - target = arguments[ i ] || {}; - i++; - } - - // Handle case when target is a string or something (possible in deep copy) - if ( typeof target !== "object" && !isFunction( target ) ) { - target = {}; - } - - // Extend jQuery itself if only one argument is passed - if ( i === length ) { - target = this; - i--; - } - - for ( ; i < length; i++ ) { - - // Only deal with non-null/undefined values - if ( ( options = arguments[ i ] ) != null ) { - - // Extend the base object - for ( name in options ) { - copy = options[ name ]; - - // Prevent Object.prototype pollution - // Prevent never-ending loop - if ( name === "__proto__" || target === copy ) { - continue; - } - - // Recurse if we're merging plain objects or arrays - if ( deep && copy && ( jQuery.isPlainObject( copy ) || - ( copyIsArray = Array.isArray( copy ) ) ) ) { - src = target[ name ]; - - // Ensure proper type for the source value - if ( copyIsArray && !Array.isArray( src ) ) { - clone = []; - } else if ( !copyIsArray && !jQuery.isPlainObject( src ) ) { - clone = {}; - } else { - clone = src; - } - copyIsArray = false; - - // Never move original objects, clone them - target[ name ] = jQuery.extend( deep, clone, copy ); - - // Don't bring in undefined values - } else if ( copy !== undefined ) { - target[ name ] = copy; - } - } - } - } - - // Return the modified object - return target; -}; - -jQuery.extend( { - - // Unique for each copy of jQuery on the page - expando: "jQuery" + ( version + Math.random() ).replace( /\D/g, "" ), - - // Assume jQuery is ready without the ready module - isReady: true, - - error: function( msg ) { - throw new Error( msg ); - }, - - noop: function() {}, - - isPlainObject: function( obj ) { - var proto, Ctor; - - // Detect obvious negatives - // Use toString instead of jQuery.type to catch host objects - if ( !obj || toString.call( obj ) !== "[object Object]" ) { - return false; - } - - proto = getProto( obj ); - - // Objects with no prototype (e.g., `Object.create( null )`) are plain - if ( !proto ) { - return true; - } - - // Objects with prototype are plain iff they were constructed by a global Object function - Ctor = hasOwn.call( proto, "constructor" ) && proto.constructor; - return typeof Ctor === "function" && fnToString.call( Ctor ) === ObjectFunctionString; - }, - - isEmptyObject: function( obj ) { - var name; - - for ( name in obj ) { - return false; - } - return true; - }, - - // Evaluates a script in a provided context; falls back to the global one - // if not specified. - globalEval: function( code, options, doc ) { - DOMEval( code, { nonce: options && options.nonce }, doc ); - }, - - each: function( obj, callback ) { - var length, i = 0; - - if ( isArrayLike( obj ) ) { - length = obj.length; - for ( ; i < length; i++ ) { - if ( callback.call( obj[ i ], i, obj[ i ] ) === false ) { - break; - } - } - } else { - for ( i in obj ) { - if ( callback.call( obj[ i ], i, obj[ i ] ) === false ) { - break; - } - } - } - - return obj; - }, - - // results is for internal usage only - makeArray: function( arr, results ) { - var ret = results || []; - - if ( arr != null ) { - if ( isArrayLike( Object( arr ) ) ) { - jQuery.merge( ret, - typeof arr === "string" ? - [ arr ] : arr - ); - } else { - push.call( ret, arr ); - } - } - - return ret; - }, - - inArray: function( elem, arr, i ) { - return arr == null ? -1 : indexOf.call( arr, elem, i ); - }, - - // Support: Android <=4.0 only, PhantomJS 1 only - // push.apply(_, arraylike) throws on ancient WebKit - merge: function( first, second ) { - var len = +second.length, - j = 0, - i = first.length; - - for ( ; j < len; j++ ) { - first[ i++ ] = second[ j ]; - } - - first.length = i; - - return first; - }, - - grep: function( elems, callback, invert ) { - var callbackInverse, - matches = [], - i = 0, - length = elems.length, - callbackExpect = !invert; - - // Go through the array, only saving the items - // that pass the validator function - for ( ; i < length; i++ ) { - callbackInverse = !callback( elems[ i ], i ); - if ( callbackInverse !== callbackExpect ) { - matches.push( elems[ i ] ); - } - } - - return matches; - }, - - // arg is for internal usage only - map: function( elems, callback, arg ) { - var length, value, - i = 0, - ret = []; - - // Go through the array, translating each of the items to their new values - if ( isArrayLike( elems ) ) { - length = elems.length; - for ( ; i < length; i++ ) { - value = callback( elems[ i ], i, arg ); - - if ( value != null ) { - ret.push( value ); - } - } - - // Go through every key on the object, - } else { - for ( i in elems ) { - value = callback( elems[ i ], i, arg ); - - if ( value != null ) { - ret.push( value ); - } - } - } - - // Flatten any nested arrays - return flat( ret ); - }, - - // A global GUID counter for objects - guid: 1, - - // jQuery.support is not used in Core but other projects attach their - // properties to it so it needs to exist. - support: support -} ); - -if ( typeof Symbol === "function" ) { - jQuery.fn[ Symbol.iterator ] = arr[ Symbol.iterator ]; -} - -// Populate the class2type map -jQuery.each( "Boolean Number String Function Array Date RegExp Object Error Symbol".split( " " ), -function( _i, name ) { - class2type[ "[object " + name + "]" ] = name.toLowerCase(); -} ); - -function isArrayLike( obj ) { - - // Support: real iOS 8.2 only (not reproducible in simulator) - // `in` check used to prevent JIT error (gh-2145) - // hasOwn isn't used here due to false negatives - // regarding Nodelist length in IE - var length = !!obj && "length" in obj && obj.length, - type = toType( obj ); - - if ( isFunction( obj ) || isWindow( obj ) ) { - return false; - } - - return type === "array" || length === 0 || - typeof length === "number" && length > 0 && ( length - 1 ) in obj; -} -var Sizzle = -/*! - * Sizzle CSS Selector Engine v2.3.5 - * https://sizzlejs.com/ - * - * Copyright JS Foundation and other contributors - * Released under the MIT license - * https://js.foundation/ - * - * Date: 2020-03-14 - */ -( function( window ) { -var i, - support, - Expr, - getText, - isXML, - tokenize, - compile, - select, - outermostContext, - sortInput, - hasDuplicate, - - // Local document vars - setDocument, - document, - docElem, - documentIsHTML, - rbuggyQSA, - rbuggyMatches, - matches, - contains, - - // Instance-specific data - expando = "sizzle" + 1 * new Date(), - preferredDoc = window.document, - dirruns = 0, - done = 0, - classCache = createCache(), - tokenCache = createCache(), - compilerCache = createCache(), - nonnativeSelectorCache = createCache(), - sortOrder = function( a, b ) { - if ( a === b ) { - hasDuplicate = true; - } - return 0; - }, - - // Instance methods - hasOwn = ( {} ).hasOwnProperty, - arr = [], - pop = arr.pop, - pushNative = arr.push, - push = arr.push, - slice = arr.slice, - - // Use a stripped-down indexOf as it's faster than native - // https://jsperf.com/thor-indexof-vs-for/5 - indexOf = function( list, elem ) { - var i = 0, - len = list.length; - for ( ; i < len; i++ ) { - if ( list[ i ] === elem ) { - return i; - } - } - return -1; - }, - - booleans = "checked|selected|async|autofocus|autoplay|controls|defer|disabled|hidden|" + - "ismap|loop|multiple|open|readonly|required|scoped", - - // Regular expressions - - // http://www.w3.org/TR/css3-selectors/#whitespace - whitespace = "[\\x20\\t\\r\\n\\f]", - - // https://www.w3.org/TR/css-syntax-3/#ident-token-diagram - identifier = "(?:\\\\[\\da-fA-F]{1,6}" + whitespace + - "?|\\\\[^\\r\\n\\f]|[\\w-]|[^\0-\\x7f])+", - - // Attribute selectors: http://www.w3.org/TR/selectors/#attribute-selectors - attributes = "\\[" + whitespace + "*(" + identifier + ")(?:" + whitespace + - - // Operator (capture 2) - "*([*^$|!~]?=)" + whitespace + - - // "Attribute values must be CSS identifiers [capture 5] - // or strings [capture 3 or capture 4]" - "*(?:'((?:\\\\.|[^\\\\'])*)'|\"((?:\\\\.|[^\\\\\"])*)\"|(" + identifier + "))|)" + - whitespace + "*\\]", - - pseudos = ":(" + identifier + ")(?:\\((" + - - // To reduce the number of selectors needing tokenize in the preFilter, prefer arguments: - // 1. quoted (capture 3; capture 4 or capture 5) - "('((?:\\\\.|[^\\\\'])*)'|\"((?:\\\\.|[^\\\\\"])*)\")|" + - - // 2. simple (capture 6) - "((?:\\\\.|[^\\\\()[\\]]|" + attributes + ")*)|" + - - // 3. anything else (capture 2) - ".*" + - ")\\)|)", - - // Leading and non-escaped trailing whitespace, capturing some non-whitespace characters preceding the latter - rwhitespace = new RegExp( whitespace + "+", "g" ), - rtrim = new RegExp( "^" + whitespace + "+|((?:^|[^\\\\])(?:\\\\.)*)" + - whitespace + "+$", "g" ), - - rcomma = new RegExp( "^" + whitespace + "*," + whitespace + "*" ), - rcombinators = new RegExp( "^" + whitespace + "*([>+~]|" + whitespace + ")" + whitespace + - "*" ), - rdescend = new RegExp( whitespace + "|>" ), - - rpseudo = new RegExp( pseudos ), - ridentifier = new RegExp( "^" + identifier + "$" ), - - matchExpr = { - "ID": new RegExp( "^#(" + identifier + ")" ), - "CLASS": new RegExp( "^\\.(" + identifier + ")" ), - "TAG": new RegExp( "^(" + identifier + "|[*])" ), - "ATTR": new RegExp( "^" + attributes ), - "PSEUDO": new RegExp( "^" + pseudos ), - "CHILD": new RegExp( "^:(only|first|last|nth|nth-last)-(child|of-type)(?:\\(" + - whitespace + "*(even|odd|(([+-]|)(\\d*)n|)" + whitespace + "*(?:([+-]|)" + - whitespace + "*(\\d+)|))" + whitespace + "*\\)|)", "i" ), - "bool": new RegExp( "^(?:" + booleans + ")$", "i" ), - - // For use in libraries implementing .is() - // We use this for POS matching in `select` - "needsContext": new RegExp( "^" + whitespace + - "*[>+~]|:(even|odd|eq|gt|lt|nth|first|last)(?:\\(" + whitespace + - "*((?:-\\d)?\\d*)" + whitespace + "*\\)|)(?=[^-]|$)", "i" ) - }, - - rhtml = /HTML$/i, - rinputs = /^(?:input|select|textarea|button)$/i, - rheader = /^h\d$/i, - - rnative = /^[^{]+\{\s*\[native \w/, - - // Easily-parseable/retrievable ID or TAG or CLASS selectors - rquickExpr = /^(?:#([\w-]+)|(\w+)|\.([\w-]+))$/, - - rsibling = /[+~]/, - - // CSS escapes - // http://www.w3.org/TR/CSS21/syndata.html#escaped-characters - runescape = new RegExp( "\\\\[\\da-fA-F]{1,6}" + whitespace + "?|\\\\([^\\r\\n\\f])", "g" ), - funescape = function( escape, nonHex ) { - var high = "0x" + escape.slice( 1 ) - 0x10000; - - return nonHex ? - - // Strip the backslash prefix from a non-hex escape sequence - nonHex : - - // Replace a hexadecimal escape sequence with the encoded Unicode code point - // Support: IE <=11+ - // For values outside the Basic Multilingual Plane (BMP), manually construct a - // surrogate pair - high < 0 ? - String.fromCharCode( high + 0x10000 ) : - String.fromCharCode( high >> 10 | 0xD800, high & 0x3FF | 0xDC00 ); - }, - - // CSS string/identifier serialization - // https://drafts.csswg.org/cssom/#common-serializing-idioms - rcssescape = /([\0-\x1f\x7f]|^-?\d)|^-$|[^\0-\x1f\x7f-\uFFFF\w-]/g, - fcssescape = function( ch, asCodePoint ) { - if ( asCodePoint ) { - - // U+0000 NULL becomes U+FFFD REPLACEMENT CHARACTER - if ( ch === "\0" ) { - return "\uFFFD"; - } - - // Control characters and (dependent upon position) numbers get escaped as code points - return ch.slice( 0, -1 ) + "\\" + - ch.charCodeAt( ch.length - 1 ).toString( 16 ) + " "; - } - - // Other potentially-special ASCII characters get backslash-escaped - return "\\" + ch; - }, - - // Used for iframes - // See setDocument() - // Removing the function wrapper causes a "Permission Denied" - // error in IE - unloadHandler = function() { - setDocument(); - }, - - inDisabledFieldset = addCombinator( - function( elem ) { - return elem.disabled === true && elem.nodeName.toLowerCase() === "fieldset"; - }, - { dir: "parentNode", next: "legend" } - ); - -// Optimize for push.apply( _, NodeList ) -try { - push.apply( - ( arr = slice.call( preferredDoc.childNodes ) ), - preferredDoc.childNodes - ); - - // Support: Android<4.0 - // Detect silently failing push.apply - // eslint-disable-next-line no-unused-expressions - arr[ preferredDoc.childNodes.length ].nodeType; -} catch ( e ) { - push = { apply: arr.length ? - - // Leverage slice if possible - function( target, els ) { - pushNative.apply( target, slice.call( els ) ); - } : - - // Support: IE<9 - // Otherwise append directly - function( target, els ) { - var j = target.length, - i = 0; - - // Can't trust NodeList.length - while ( ( target[ j++ ] = els[ i++ ] ) ) {} - target.length = j - 1; - } - }; -} - -function Sizzle( selector, context, results, seed ) { - var m, i, elem, nid, match, groups, newSelector, - newContext = context && context.ownerDocument, - - // nodeType defaults to 9, since context defaults to document - nodeType = context ? context.nodeType : 9; - - results = results || []; - - // Return early from calls with invalid selector or context - if ( typeof selector !== "string" || !selector || - nodeType !== 1 && nodeType !== 9 && nodeType !== 11 ) { - - return results; - } - - // Try to shortcut find operations (as opposed to filters) in HTML documents - if ( !seed ) { - setDocument( context ); - context = context || document; - - if ( documentIsHTML ) { - - // If the selector is sufficiently simple, try using a "get*By*" DOM method - // (excepting DocumentFragment context, where the methods don't exist) - if ( nodeType !== 11 && ( match = rquickExpr.exec( selector ) ) ) { - - // ID selector - if ( ( m = match[ 1 ] ) ) { - - // Document context - if ( nodeType === 9 ) { - if ( ( elem = context.getElementById( m ) ) ) { - - // Support: IE, Opera, Webkit - // TODO: identify versions - // getElementById can match elements by name instead of ID - if ( elem.id === m ) { - results.push( elem ); - return results; - } - } else { - return results; - } - - // Element context - } else { - - // Support: IE, Opera, Webkit - // TODO: identify versions - // getElementById can match elements by name instead of ID - if ( newContext && ( elem = newContext.getElementById( m ) ) && - contains( context, elem ) && - elem.id === m ) { - - results.push( elem ); - return results; - } - } - - // Type selector - } else if ( match[ 2 ] ) { - push.apply( results, context.getElementsByTagName( selector ) ); - return results; - - // Class selector - } else if ( ( m = match[ 3 ] ) && support.getElementsByClassName && - context.getElementsByClassName ) { - - push.apply( results, context.getElementsByClassName( m ) ); - return results; - } - } - - // Take advantage of querySelectorAll - if ( support.qsa && - !nonnativeSelectorCache[ selector + " " ] && - ( !rbuggyQSA || !rbuggyQSA.test( selector ) ) && - - // Support: IE 8 only - // Exclude object elements - ( nodeType !== 1 || context.nodeName.toLowerCase() !== "object" ) ) { - - newSelector = selector; - newContext = context; - - // qSA considers elements outside a scoping root when evaluating child or - // descendant combinators, which is not what we want. - // In such cases, we work around the behavior by prefixing every selector in the - // list with an ID selector referencing the scope context. - // The technique has to be used as well when a leading combinator is used - // as such selectors are not recognized by querySelectorAll. - // Thanks to Andrew Dupont for this technique. - if ( nodeType === 1 && - ( rdescend.test( selector ) || rcombinators.test( selector ) ) ) { - - // Expand context for sibling selectors - newContext = rsibling.test( selector ) && testContext( context.parentNode ) || - context; - - // We can use :scope instead of the ID hack if the browser - // supports it & if we're not changing the context. - if ( newContext !== context || !support.scope ) { - - // Capture the context ID, setting it first if necessary - if ( ( nid = context.getAttribute( "id" ) ) ) { - nid = nid.replace( rcssescape, fcssescape ); - } else { - context.setAttribute( "id", ( nid = expando ) ); - } - } - - // Prefix every selector in the list - groups = tokenize( selector ); - i = groups.length; - while ( i-- ) { - groups[ i ] = ( nid ? "#" + nid : ":scope" ) + " " + - toSelector( groups[ i ] ); - } - newSelector = groups.join( "," ); - } - - try { - push.apply( results, - newContext.querySelectorAll( newSelector ) - ); - return results; - } catch ( qsaError ) { - nonnativeSelectorCache( selector, true ); - } finally { - if ( nid === expando ) { - context.removeAttribute( "id" ); - } - } - } - } - } - - // All others - return select( selector.replace( rtrim, "$1" ), context, results, seed ); -} - -/** - * Create key-value caches of limited size - * @returns {function(string, object)} Returns the Object data after storing it on itself with - * property name the (space-suffixed) string and (if the cache is larger than Expr.cacheLength) - * deleting the oldest entry - */ -function createCache() { - var keys = []; - - function cache( key, value ) { - - // Use (key + " ") to avoid collision with native prototype properties (see Issue #157) - if ( keys.push( key + " " ) > Expr.cacheLength ) { - - // Only keep the most recent entries - delete cache[ keys.shift() ]; - } - return ( cache[ key + " " ] = value ); - } - return cache; -} - -/** - * Mark a function for special use by Sizzle - * @param {Function} fn The function to mark - */ -function markFunction( fn ) { - fn[ expando ] = true; - return fn; -} - -/** - * Support testing using an element - * @param {Function} fn Passed the created element and returns a boolean result - */ -function assert( fn ) { - var el = document.createElement( "fieldset" ); - - try { - return !!fn( el ); - } catch ( e ) { - return false; - } finally { - - // Remove from its parent by default - if ( el.parentNode ) { - el.parentNode.removeChild( el ); - } - - // release memory in IE - el = null; - } -} - -/** - * Adds the same handler for all of the specified attrs - * @param {String} attrs Pipe-separated list of attributes - * @param {Function} handler The method that will be applied - */ -function addHandle( attrs, handler ) { - var arr = attrs.split( "|" ), - i = arr.length; - - while ( i-- ) { - Expr.attrHandle[ arr[ i ] ] = handler; - } -} - -/** - * Checks document order of two siblings - * @param {Element} a - * @param {Element} b - * @returns {Number} Returns less than 0 if a precedes b, greater than 0 if a follows b - */ -function siblingCheck( a, b ) { - var cur = b && a, - diff = cur && a.nodeType === 1 && b.nodeType === 1 && - a.sourceIndex - b.sourceIndex; - - // Use IE sourceIndex if available on both nodes - if ( diff ) { - return diff; - } - - // Check if b follows a - if ( cur ) { - while ( ( cur = cur.nextSibling ) ) { - if ( cur === b ) { - return -1; - } - } - } - - return a ? 1 : -1; -} - -/** - * Returns a function to use in pseudos for input types - * @param {String} type - */ -function createInputPseudo( type ) { - return function( elem ) { - var name = elem.nodeName.toLowerCase(); - return name === "input" && elem.type === type; - }; -} - -/** - * Returns a function to use in pseudos for buttons - * @param {String} type - */ -function createButtonPseudo( type ) { - return function( elem ) { - var name = elem.nodeName.toLowerCase(); - return ( name === "input" || name === "button" ) && elem.type === type; - }; -} - -/** - * Returns a function to use in pseudos for :enabled/:disabled - * @param {Boolean} disabled true for :disabled; false for :enabled - */ -function createDisabledPseudo( disabled ) { - - // Known :disabled false positives: fieldset[disabled] > legend:nth-of-type(n+2) :can-disable - return function( elem ) { - - // Only certain elements can match :enabled or :disabled - // https://html.spec.whatwg.org/multipage/scripting.html#selector-enabled - // https://html.spec.whatwg.org/multipage/scripting.html#selector-disabled - if ( "form" in elem ) { - - // Check for inherited disabledness on relevant non-disabled elements: - // * listed form-associated elements in a disabled fieldset - // https://html.spec.whatwg.org/multipage/forms.html#category-listed - // https://html.spec.whatwg.org/multipage/forms.html#concept-fe-disabled - // * option elements in a disabled optgroup - // https://html.spec.whatwg.org/multipage/forms.html#concept-option-disabled - // All such elements have a "form" property. - if ( elem.parentNode && elem.disabled === false ) { - - // Option elements defer to a parent optgroup if present - if ( "label" in elem ) { - if ( "label" in elem.parentNode ) { - return elem.parentNode.disabled === disabled; - } else { - return elem.disabled === disabled; - } - } - - // Support: IE 6 - 11 - // Use the isDisabled shortcut property to check for disabled fieldset ancestors - return elem.isDisabled === disabled || - - // Where there is no isDisabled, check manually - /* jshint -W018 */ - elem.isDisabled !== !disabled && - inDisabledFieldset( elem ) === disabled; - } - - return elem.disabled === disabled; - - // Try to winnow out elements that can't be disabled before trusting the disabled property. - // Some victims get caught in our net (label, legend, menu, track), but it shouldn't - // even exist on them, let alone have a boolean value. - } else if ( "label" in elem ) { - return elem.disabled === disabled; - } - - // Remaining elements are neither :enabled nor :disabled - return false; - }; -} - -/** - * Returns a function to use in pseudos for positionals - * @param {Function} fn - */ -function createPositionalPseudo( fn ) { - return markFunction( function( argument ) { - argument = +argument; - return markFunction( function( seed, matches ) { - var j, - matchIndexes = fn( [], seed.length, argument ), - i = matchIndexes.length; - - // Match elements found at the specified indexes - while ( i-- ) { - if ( seed[ ( j = matchIndexes[ i ] ) ] ) { - seed[ j ] = !( matches[ j ] = seed[ j ] ); - } - } - } ); - } ); -} - -/** - * Checks a node for validity as a Sizzle context - * @param {Element|Object=} context - * @returns {Element|Object|Boolean} The input node if acceptable, otherwise a falsy value - */ -function testContext( context ) { - return context && typeof context.getElementsByTagName !== "undefined" && context; -} - -// Expose support vars for convenience -support = Sizzle.support = {}; - -/** - * Detects XML nodes - * @param {Element|Object} elem An element or a document - * @returns {Boolean} True iff elem is a non-HTML XML node - */ -isXML = Sizzle.isXML = function( elem ) { - var namespace = elem.namespaceURI, - docElem = ( elem.ownerDocument || elem ).documentElement; - - // Support: IE <=8 - // Assume HTML when documentElement doesn't yet exist, such as inside loading iframes - // https://bugs.jquery.com/ticket/4833 - return !rhtml.test( namespace || docElem && docElem.nodeName || "HTML" ); -}; - -/** - * Sets document-related variables once based on the current document - * @param {Element|Object} [doc] An element or document object to use to set the document - * @returns {Object} Returns the current document - */ -setDocument = Sizzle.setDocument = function( node ) { - var hasCompare, subWindow, - doc = node ? node.ownerDocument || node : preferredDoc; - - // Return early if doc is invalid or already selected - // Support: IE 11+, Edge 17 - 18+ - // IE/Edge sometimes throw a "Permission denied" error when strict-comparing - // two documents; shallow comparisons work. - // eslint-disable-next-line eqeqeq - if ( doc == document || doc.nodeType !== 9 || !doc.documentElement ) { - return document; - } - - // Update global variables - document = doc; - docElem = document.documentElement; - documentIsHTML = !isXML( document ); - - // Support: IE 9 - 11+, Edge 12 - 18+ - // Accessing iframe documents after unload throws "permission denied" errors (jQuery #13936) - // Support: IE 11+, Edge 17 - 18+ - // IE/Edge sometimes throw a "Permission denied" error when strict-comparing - // two documents; shallow comparisons work. - // eslint-disable-next-line eqeqeq - if ( preferredDoc != document && - ( subWindow = document.defaultView ) && subWindow.top !== subWindow ) { - - // Support: IE 11, Edge - if ( subWindow.addEventListener ) { - subWindow.addEventListener( "unload", unloadHandler, false ); - - // Support: IE 9 - 10 only - } else if ( subWindow.attachEvent ) { - subWindow.attachEvent( "onunload", unloadHandler ); - } - } - - // Support: IE 8 - 11+, Edge 12 - 18+, Chrome <=16 - 25 only, Firefox <=3.6 - 31 only, - // Safari 4 - 5 only, Opera <=11.6 - 12.x only - // IE/Edge & older browsers don't support the :scope pseudo-class. - // Support: Safari 6.0 only - // Safari 6.0 supports :scope but it's an alias of :root there. - support.scope = assert( function( el ) { - docElem.appendChild( el ).appendChild( document.createElement( "div" ) ); - return typeof el.querySelectorAll !== "undefined" && - !el.querySelectorAll( ":scope fieldset div" ).length; - } ); - - /* Attributes - ---------------------------------------------------------------------- */ - - // Support: IE<8 - // Verify that getAttribute really returns attributes and not properties - // (excepting IE8 booleans) - support.attributes = assert( function( el ) { - el.className = "i"; - return !el.getAttribute( "className" ); - } ); - - /* getElement(s)By* - ---------------------------------------------------------------------- */ - - // Check if getElementsByTagName("*") returns only elements - support.getElementsByTagName = assert( function( el ) { - el.appendChild( document.createComment( "" ) ); - return !el.getElementsByTagName( "*" ).length; - } ); - - // Support: IE<9 - support.getElementsByClassName = rnative.test( document.getElementsByClassName ); - - // Support: IE<10 - // Check if getElementById returns elements by name - // The broken getElementById methods don't pick up programmatically-set names, - // so use a roundabout getElementsByName test - support.getById = assert( function( el ) { - docElem.appendChild( el ).id = expando; - return !document.getElementsByName || !document.getElementsByName( expando ).length; - } ); - - // ID filter and find - if ( support.getById ) { - Expr.filter[ "ID" ] = function( id ) { - var attrId = id.replace( runescape, funescape ); - return function( elem ) { - return elem.getAttribute( "id" ) === attrId; - }; - }; - Expr.find[ "ID" ] = function( id, context ) { - if ( typeof context.getElementById !== "undefined" && documentIsHTML ) { - var elem = context.getElementById( id ); - return elem ? [ elem ] : []; - } - }; - } else { - Expr.filter[ "ID" ] = function( id ) { - var attrId = id.replace( runescape, funescape ); - return function( elem ) { - var node = typeof elem.getAttributeNode !== "undefined" && - elem.getAttributeNode( "id" ); - return node && node.value === attrId; - }; - }; - - // Support: IE 6 - 7 only - // getElementById is not reliable as a find shortcut - Expr.find[ "ID" ] = function( id, context ) { - if ( typeof context.getElementById !== "undefined" && documentIsHTML ) { - var node, i, elems, - elem = context.getElementById( id ); - - if ( elem ) { - - // Verify the id attribute - node = elem.getAttributeNode( "id" ); - if ( node && node.value === id ) { - return [ elem ]; - } - - // Fall back on getElementsByName - elems = context.getElementsByName( id ); - i = 0; - while ( ( elem = elems[ i++ ] ) ) { - node = elem.getAttributeNode( "id" ); - if ( node && node.value === id ) { - return [ elem ]; - } - } - } - - return []; - } - }; - } - - // Tag - Expr.find[ "TAG" ] = support.getElementsByTagName ? - function( tag, context ) { - if ( typeof context.getElementsByTagName !== "undefined" ) { - return context.getElementsByTagName( tag ); - - // DocumentFragment nodes don't have gEBTN - } else if ( support.qsa ) { - return context.querySelectorAll( tag ); - } - } : - - function( tag, context ) { - var elem, - tmp = [], - i = 0, - - // By happy coincidence, a (broken) gEBTN appears on DocumentFragment nodes too - results = context.getElementsByTagName( tag ); - - // Filter out possible comments - if ( tag === "*" ) { - while ( ( elem = results[ i++ ] ) ) { - if ( elem.nodeType === 1 ) { - tmp.push( elem ); - } - } - - return tmp; - } - return results; - }; - - // Class - Expr.find[ "CLASS" ] = support.getElementsByClassName && function( className, context ) { - if ( typeof context.getElementsByClassName !== "undefined" && documentIsHTML ) { - return context.getElementsByClassName( className ); - } - }; - - /* QSA/matchesSelector - ---------------------------------------------------------------------- */ - - // QSA and matchesSelector support - - // matchesSelector(:active) reports false when true (IE9/Opera 11.5) - rbuggyMatches = []; - - // qSa(:focus) reports false when true (Chrome 21) - // We allow this because of a bug in IE8/9 that throws an error - // whenever `document.activeElement` is accessed on an iframe - // So, we allow :focus to pass through QSA all the time to avoid the IE error - // See https://bugs.jquery.com/ticket/13378 - rbuggyQSA = []; - - if ( ( support.qsa = rnative.test( document.querySelectorAll ) ) ) { - - // Build QSA regex - // Regex strategy adopted from Diego Perini - assert( function( el ) { - - var input; - - // Select is set to empty string on purpose - // This is to test IE's treatment of not explicitly - // setting a boolean content attribute, - // since its presence should be enough - // https://bugs.jquery.com/ticket/12359 - docElem.appendChild( el ).innerHTML = "" + - ""; - - // Support: IE8, Opera 11-12.16 - // Nothing should be selected when empty strings follow ^= or $= or *= - // The test attribute must be unknown in Opera but "safe" for WinRT - // https://msdn.microsoft.com/en-us/library/ie/hh465388.aspx#attribute_section - if ( el.querySelectorAll( "[msallowcapture^='']" ).length ) { - rbuggyQSA.push( "[*^$]=" + whitespace + "*(?:''|\"\")" ); - } - - // Support: IE8 - // Boolean attributes and "value" are not treated correctly - if ( !el.querySelectorAll( "[selected]" ).length ) { - rbuggyQSA.push( "\\[" + whitespace + "*(?:value|" + booleans + ")" ); - } - - // Support: Chrome<29, Android<4.4, Safari<7.0+, iOS<7.0+, PhantomJS<1.9.8+ - if ( !el.querySelectorAll( "[id~=" + expando + "-]" ).length ) { - rbuggyQSA.push( "~=" ); - } - - // Support: IE 11+, Edge 15 - 18+ - // IE 11/Edge don't find elements on a `[name='']` query in some cases. - // Adding a temporary attribute to the document before the selection works - // around the issue. - // Interestingly, IE 10 & older don't seem to have the issue. - input = document.createElement( "input" ); - input.setAttribute( "name", "" ); - el.appendChild( input ); - if ( !el.querySelectorAll( "[name='']" ).length ) { - rbuggyQSA.push( "\\[" + whitespace + "*name" + whitespace + "*=" + - whitespace + "*(?:''|\"\")" ); - } - - // Webkit/Opera - :checked should return selected option elements - // http://www.w3.org/TR/2011/REC-css3-selectors-20110929/#checked - // IE8 throws error here and will not see later tests - if ( !el.querySelectorAll( ":checked" ).length ) { - rbuggyQSA.push( ":checked" ); - } - - // Support: Safari 8+, iOS 8+ - // https://bugs.webkit.org/show_bug.cgi?id=136851 - // In-page `selector#id sibling-combinator selector` fails - if ( !el.querySelectorAll( "a#" + expando + "+*" ).length ) { - rbuggyQSA.push( ".#.+[+~]" ); - } - - // Support: Firefox <=3.6 - 5 only - // Old Firefox doesn't throw on a badly-escaped identifier. - el.querySelectorAll( "\\\f" ); - rbuggyQSA.push( "[\\r\\n\\f]" ); - } ); - - assert( function( el ) { - el.innerHTML = "" + - ""; - - // Support: Windows 8 Native Apps - // The type and name attributes are restricted during .innerHTML assignment - var input = document.createElement( "input" ); - input.setAttribute( "type", "hidden" ); - el.appendChild( input ).setAttribute( "name", "D" ); - - // Support: IE8 - // Enforce case-sensitivity of name attribute - if ( el.querySelectorAll( "[name=d]" ).length ) { - rbuggyQSA.push( "name" + whitespace + "*[*^$|!~]?=" ); - } - - // FF 3.5 - :enabled/:disabled and hidden elements (hidden elements are still enabled) - // IE8 throws error here and will not see later tests - if ( el.querySelectorAll( ":enabled" ).length !== 2 ) { - rbuggyQSA.push( ":enabled", ":disabled" ); - } - - // Support: IE9-11+ - // IE's :disabled selector does not pick up the children of disabled fieldsets - docElem.appendChild( el ).disabled = true; - if ( el.querySelectorAll( ":disabled" ).length !== 2 ) { - rbuggyQSA.push( ":enabled", ":disabled" ); - } - - // Support: Opera 10 - 11 only - // Opera 10-11 does not throw on post-comma invalid pseudos - el.querySelectorAll( "*,:x" ); - rbuggyQSA.push( ",.*:" ); - } ); - } - - if ( ( support.matchesSelector = rnative.test( ( matches = docElem.matches || - docElem.webkitMatchesSelector || - docElem.mozMatchesSelector || - docElem.oMatchesSelector || - docElem.msMatchesSelector ) ) ) ) { - - assert( function( el ) { - - // Check to see if it's possible to do matchesSelector - // on a disconnected node (IE 9) - support.disconnectedMatch = matches.call( el, "*" ); - - // This should fail with an exception - // Gecko does not error, returns false instead - matches.call( el, "[s!='']:x" ); - rbuggyMatches.push( "!=", pseudos ); - } ); - } - - rbuggyQSA = rbuggyQSA.length && new RegExp( rbuggyQSA.join( "|" ) ); - rbuggyMatches = rbuggyMatches.length && new RegExp( rbuggyMatches.join( "|" ) ); - - /* Contains - ---------------------------------------------------------------------- */ - hasCompare = rnative.test( docElem.compareDocumentPosition ); - - // Element contains another - // Purposefully self-exclusive - // As in, an element does not contain itself - contains = hasCompare || rnative.test( docElem.contains ) ? - function( a, b ) { - var adown = a.nodeType === 9 ? a.documentElement : a, - bup = b && b.parentNode; - return a === bup || !!( bup && bup.nodeType === 1 && ( - adown.contains ? - adown.contains( bup ) : - a.compareDocumentPosition && a.compareDocumentPosition( bup ) & 16 - ) ); - } : - function( a, b ) { - if ( b ) { - while ( ( b = b.parentNode ) ) { - if ( b === a ) { - return true; - } - } - } - return false; - }; - - /* Sorting - ---------------------------------------------------------------------- */ - - // Document order sorting - sortOrder = hasCompare ? - function( a, b ) { - - // Flag for duplicate removal - if ( a === b ) { - hasDuplicate = true; - return 0; - } - - // Sort on method existence if only one input has compareDocumentPosition - var compare = !a.compareDocumentPosition - !b.compareDocumentPosition; - if ( compare ) { - return compare; - } - - // Calculate position if both inputs belong to the same document - // Support: IE 11+, Edge 17 - 18+ - // IE/Edge sometimes throw a "Permission denied" error when strict-comparing - // two documents; shallow comparisons work. - // eslint-disable-next-line eqeqeq - compare = ( a.ownerDocument || a ) == ( b.ownerDocument || b ) ? - a.compareDocumentPosition( b ) : - - // Otherwise we know they are disconnected - 1; - - // Disconnected nodes - if ( compare & 1 || - ( !support.sortDetached && b.compareDocumentPosition( a ) === compare ) ) { - - // Choose the first element that is related to our preferred document - // Support: IE 11+, Edge 17 - 18+ - // IE/Edge sometimes throw a "Permission denied" error when strict-comparing - // two documents; shallow comparisons work. - // eslint-disable-next-line eqeqeq - if ( a == document || a.ownerDocument == preferredDoc && - contains( preferredDoc, a ) ) { - return -1; - } - - // Support: IE 11+, Edge 17 - 18+ - // IE/Edge sometimes throw a "Permission denied" error when strict-comparing - // two documents; shallow comparisons work. - // eslint-disable-next-line eqeqeq - if ( b == document || b.ownerDocument == preferredDoc && - contains( preferredDoc, b ) ) { - return 1; - } - - // Maintain original order - return sortInput ? - ( indexOf( sortInput, a ) - indexOf( sortInput, b ) ) : - 0; - } - - return compare & 4 ? -1 : 1; - } : - function( a, b ) { - - // Exit early if the nodes are identical - if ( a === b ) { - hasDuplicate = true; - return 0; - } - - var cur, - i = 0, - aup = a.parentNode, - bup = b.parentNode, - ap = [ a ], - bp = [ b ]; - - // Parentless nodes are either documents or disconnected - if ( !aup || !bup ) { - - // Support: IE 11+, Edge 17 - 18+ - // IE/Edge sometimes throw a "Permission denied" error when strict-comparing - // two documents; shallow comparisons work. - /* eslint-disable eqeqeq */ - return a == document ? -1 : - b == document ? 1 : - /* eslint-enable eqeqeq */ - aup ? -1 : - bup ? 1 : - sortInput ? - ( indexOf( sortInput, a ) - indexOf( sortInput, b ) ) : - 0; - - // If the nodes are siblings, we can do a quick check - } else if ( aup === bup ) { - return siblingCheck( a, b ); - } - - // Otherwise we need full lists of their ancestors for comparison - cur = a; - while ( ( cur = cur.parentNode ) ) { - ap.unshift( cur ); - } - cur = b; - while ( ( cur = cur.parentNode ) ) { - bp.unshift( cur ); - } - - // Walk down the tree looking for a discrepancy - while ( ap[ i ] === bp[ i ] ) { - i++; - } - - return i ? - - // Do a sibling check if the nodes have a common ancestor - siblingCheck( ap[ i ], bp[ i ] ) : - - // Otherwise nodes in our document sort first - // Support: IE 11+, Edge 17 - 18+ - // IE/Edge sometimes throw a "Permission denied" error when strict-comparing - // two documents; shallow comparisons work. - /* eslint-disable eqeqeq */ - ap[ i ] == preferredDoc ? -1 : - bp[ i ] == preferredDoc ? 1 : - /* eslint-enable eqeqeq */ - 0; - }; - - return document; -}; - -Sizzle.matches = function( expr, elements ) { - return Sizzle( expr, null, null, elements ); -}; - -Sizzle.matchesSelector = function( elem, expr ) { - setDocument( elem ); - - if ( support.matchesSelector && documentIsHTML && - !nonnativeSelectorCache[ expr + " " ] && - ( !rbuggyMatches || !rbuggyMatches.test( expr ) ) && - ( !rbuggyQSA || !rbuggyQSA.test( expr ) ) ) { - - try { - var ret = matches.call( elem, expr ); - - // IE 9's matchesSelector returns false on disconnected nodes - if ( ret || support.disconnectedMatch || - - // As well, disconnected nodes are said to be in a document - // fragment in IE 9 - elem.document && elem.document.nodeType !== 11 ) { - return ret; - } - } catch ( e ) { - nonnativeSelectorCache( expr, true ); - } - } - - return Sizzle( expr, document, null, [ elem ] ).length > 0; -}; - -Sizzle.contains = function( context, elem ) { - - // Set document vars if needed - // Support: IE 11+, Edge 17 - 18+ - // IE/Edge sometimes throw a "Permission denied" error when strict-comparing - // two documents; shallow comparisons work. - // eslint-disable-next-line eqeqeq - if ( ( context.ownerDocument || context ) != document ) { - setDocument( context ); - } - return contains( context, elem ); -}; - -Sizzle.attr = function( elem, name ) { - - // Set document vars if needed - // Support: IE 11+, Edge 17 - 18+ - // IE/Edge sometimes throw a "Permission denied" error when strict-comparing - // two documents; shallow comparisons work. - // eslint-disable-next-line eqeqeq - if ( ( elem.ownerDocument || elem ) != document ) { - setDocument( elem ); - } - - var fn = Expr.attrHandle[ name.toLowerCase() ], - - // Don't get fooled by Object.prototype properties (jQuery #13807) - val = fn && hasOwn.call( Expr.attrHandle, name.toLowerCase() ) ? - fn( elem, name, !documentIsHTML ) : - undefined; - - return val !== undefined ? - val : - support.attributes || !documentIsHTML ? - elem.getAttribute( name ) : - ( val = elem.getAttributeNode( name ) ) && val.specified ? - val.value : - null; -}; - -Sizzle.escape = function( sel ) { - return ( sel + "" ).replace( rcssescape, fcssescape ); -}; - -Sizzle.error = function( msg ) { - throw new Error( "Syntax error, unrecognized expression: " + msg ); -}; - -/** - * Document sorting and removing duplicates - * @param {ArrayLike} results - */ -Sizzle.uniqueSort = function( results ) { - var elem, - duplicates = [], - j = 0, - i = 0; - - // Unless we *know* we can detect duplicates, assume their presence - hasDuplicate = !support.detectDuplicates; - sortInput = !support.sortStable && results.slice( 0 ); - results.sort( sortOrder ); - - if ( hasDuplicate ) { - while ( ( elem = results[ i++ ] ) ) { - if ( elem === results[ i ] ) { - j = duplicates.push( i ); - } - } - while ( j-- ) { - results.splice( duplicates[ j ], 1 ); - } - } - - // Clear input after sorting to release objects - // See https://github.com/jquery/sizzle/pull/225 - sortInput = null; - - return results; -}; - -/** - * Utility function for retrieving the text value of an array of DOM nodes - * @param {Array|Element} elem - */ -getText = Sizzle.getText = function( elem ) { - var node, - ret = "", - i = 0, - nodeType = elem.nodeType; - - if ( !nodeType ) { - - // If no nodeType, this is expected to be an array - while ( ( node = elem[ i++ ] ) ) { - - // Do not traverse comment nodes - ret += getText( node ); - } - } else if ( nodeType === 1 || nodeType === 9 || nodeType === 11 ) { - - // Use textContent for elements - // innerText usage removed for consistency of new lines (jQuery #11153) - if ( typeof elem.textContent === "string" ) { - return elem.textContent; - } else { - - // Traverse its children - for ( elem = elem.firstChild; elem; elem = elem.nextSibling ) { - ret += getText( elem ); - } - } - } else if ( nodeType === 3 || nodeType === 4 ) { - return elem.nodeValue; - } - - // Do not include comment or processing instruction nodes - - return ret; -}; - -Expr = Sizzle.selectors = { - - // Can be adjusted by the user - cacheLength: 50, - - createPseudo: markFunction, - - match: matchExpr, - - attrHandle: {}, - - find: {}, - - relative: { - ">": { dir: "parentNode", first: true }, - " ": { dir: "parentNode" }, - "+": { dir: "previousSibling", first: true }, - "~": { dir: "previousSibling" } - }, - - preFilter: { - "ATTR": function( match ) { - match[ 1 ] = match[ 1 ].replace( runescape, funescape ); - - // Move the given value to match[3] whether quoted or unquoted - match[ 3 ] = ( match[ 3 ] || match[ 4 ] || - match[ 5 ] || "" ).replace( runescape, funescape ); - - if ( match[ 2 ] === "~=" ) { - match[ 3 ] = " " + match[ 3 ] + " "; - } - - return match.slice( 0, 4 ); - }, - - "CHILD": function( match ) { - - /* matches from matchExpr["CHILD"] - 1 type (only|nth|...) - 2 what (child|of-type) - 3 argument (even|odd|\d*|\d*n([+-]\d+)?|...) - 4 xn-component of xn+y argument ([+-]?\d*n|) - 5 sign of xn-component - 6 x of xn-component - 7 sign of y-component - 8 y of y-component - */ - match[ 1 ] = match[ 1 ].toLowerCase(); - - if ( match[ 1 ].slice( 0, 3 ) === "nth" ) { - - // nth-* requires argument - if ( !match[ 3 ] ) { - Sizzle.error( match[ 0 ] ); - } - - // numeric x and y parameters for Expr.filter.CHILD - // remember that false/true cast respectively to 0/1 - match[ 4 ] = +( match[ 4 ] ? - match[ 5 ] + ( match[ 6 ] || 1 ) : - 2 * ( match[ 3 ] === "even" || match[ 3 ] === "odd" ) ); - match[ 5 ] = +( ( match[ 7 ] + match[ 8 ] ) || match[ 3 ] === "odd" ); - - // other types prohibit arguments - } else if ( match[ 3 ] ) { - Sizzle.error( match[ 0 ] ); - } - - return match; - }, - - "PSEUDO": function( match ) { - var excess, - unquoted = !match[ 6 ] && match[ 2 ]; - - if ( matchExpr[ "CHILD" ].test( match[ 0 ] ) ) { - return null; - } - - // Accept quoted arguments as-is - if ( match[ 3 ] ) { - match[ 2 ] = match[ 4 ] || match[ 5 ] || ""; - - // Strip excess characters from unquoted arguments - } else if ( unquoted && rpseudo.test( unquoted ) && - - // Get excess from tokenize (recursively) - ( excess = tokenize( unquoted, true ) ) && - - // advance to the next closing parenthesis - ( excess = unquoted.indexOf( ")", unquoted.length - excess ) - unquoted.length ) ) { - - // excess is a negative index - match[ 0 ] = match[ 0 ].slice( 0, excess ); - match[ 2 ] = unquoted.slice( 0, excess ); - } - - // Return only captures needed by the pseudo filter method (type and argument) - return match.slice( 0, 3 ); - } - }, - - filter: { - - "TAG": function( nodeNameSelector ) { - var nodeName = nodeNameSelector.replace( runescape, funescape ).toLowerCase(); - return nodeNameSelector === "*" ? - function() { - return true; - } : - function( elem ) { - return elem.nodeName && elem.nodeName.toLowerCase() === nodeName; - }; - }, - - "CLASS": function( className ) { - var pattern = classCache[ className + " " ]; - - return pattern || - ( pattern = new RegExp( "(^|" + whitespace + - ")" + className + "(" + whitespace + "|$)" ) ) && classCache( - className, function( elem ) { - return pattern.test( - typeof elem.className === "string" && elem.className || - typeof elem.getAttribute !== "undefined" && - elem.getAttribute( "class" ) || - "" - ); - } ); - }, - - "ATTR": function( name, operator, check ) { - return function( elem ) { - var result = Sizzle.attr( elem, name ); - - if ( result == null ) { - return operator === "!="; - } - if ( !operator ) { - return true; - } - - result += ""; - - /* eslint-disable max-len */ - - return operator === "=" ? result === check : - operator === "!=" ? result !== check : - operator === "^=" ? check && result.indexOf( check ) === 0 : - operator === "*=" ? check && result.indexOf( check ) > -1 : - operator === "$=" ? check && result.slice( -check.length ) === check : - operator === "~=" ? ( " " + result.replace( rwhitespace, " " ) + " " ).indexOf( check ) > -1 : - operator === "|=" ? result === check || result.slice( 0, check.length + 1 ) === check + "-" : - false; - /* eslint-enable max-len */ - - }; - }, - - "CHILD": function( type, what, _argument, first, last ) { - var simple = type.slice( 0, 3 ) !== "nth", - forward = type.slice( -4 ) !== "last", - ofType = what === "of-type"; - - return first === 1 && last === 0 ? - - // Shortcut for :nth-*(n) - function( elem ) { - return !!elem.parentNode; - } : - - function( elem, _context, xml ) { - var cache, uniqueCache, outerCache, node, nodeIndex, start, - dir = simple !== forward ? "nextSibling" : "previousSibling", - parent = elem.parentNode, - name = ofType && elem.nodeName.toLowerCase(), - useCache = !xml && !ofType, - diff = false; - - if ( parent ) { - - // :(first|last|only)-(child|of-type) - if ( simple ) { - while ( dir ) { - node = elem; - while ( ( node = node[ dir ] ) ) { - if ( ofType ? - node.nodeName.toLowerCase() === name : - node.nodeType === 1 ) { - - return false; - } - } - - // Reverse direction for :only-* (if we haven't yet done so) - start = dir = type === "only" && !start && "nextSibling"; - } - return true; - } - - start = [ forward ? parent.firstChild : parent.lastChild ]; - - // non-xml :nth-child(...) stores cache data on `parent` - if ( forward && useCache ) { - - // Seek `elem` from a previously-cached index - - // ...in a gzip-friendly way - node = parent; - outerCache = node[ expando ] || ( node[ expando ] = {} ); - - // Support: IE <9 only - // Defend against cloned attroperties (jQuery gh-1709) - uniqueCache = outerCache[ node.uniqueID ] || - ( outerCache[ node.uniqueID ] = {} ); - - cache = uniqueCache[ type ] || []; - nodeIndex = cache[ 0 ] === dirruns && cache[ 1 ]; - diff = nodeIndex && cache[ 2 ]; - node = nodeIndex && parent.childNodes[ nodeIndex ]; - - while ( ( node = ++nodeIndex && node && node[ dir ] || - - // Fallback to seeking `elem` from the start - ( diff = nodeIndex = 0 ) || start.pop() ) ) { - - // When found, cache indexes on `parent` and break - if ( node.nodeType === 1 && ++diff && node === elem ) { - uniqueCache[ type ] = [ dirruns, nodeIndex, diff ]; - break; - } - } - - } else { - - // Use previously-cached element index if available - if ( useCache ) { - - // ...in a gzip-friendly way - node = elem; - outerCache = node[ expando ] || ( node[ expando ] = {} ); - - // Support: IE <9 only - // Defend against cloned attroperties (jQuery gh-1709) - uniqueCache = outerCache[ node.uniqueID ] || - ( outerCache[ node.uniqueID ] = {} ); - - cache = uniqueCache[ type ] || []; - nodeIndex = cache[ 0 ] === dirruns && cache[ 1 ]; - diff = nodeIndex; - } - - // xml :nth-child(...) - // or :nth-last-child(...) or :nth(-last)?-of-type(...) - if ( diff === false ) { - - // Use the same loop as above to seek `elem` from the start - while ( ( node = ++nodeIndex && node && node[ dir ] || - ( diff = nodeIndex = 0 ) || start.pop() ) ) { - - if ( ( ofType ? - node.nodeName.toLowerCase() === name : - node.nodeType === 1 ) && - ++diff ) { - - // Cache the index of each encountered element - if ( useCache ) { - outerCache = node[ expando ] || - ( node[ expando ] = {} ); - - // Support: IE <9 only - // Defend against cloned attroperties (jQuery gh-1709) - uniqueCache = outerCache[ node.uniqueID ] || - ( outerCache[ node.uniqueID ] = {} ); - - uniqueCache[ type ] = [ dirruns, diff ]; - } - - if ( node === elem ) { - break; - } - } - } - } - } - - // Incorporate the offset, then check against cycle size - diff -= last; - return diff === first || ( diff % first === 0 && diff / first >= 0 ); - } - }; - }, - - "PSEUDO": function( pseudo, argument ) { - - // pseudo-class names are case-insensitive - // http://www.w3.org/TR/selectors/#pseudo-classes - // Prioritize by case sensitivity in case custom pseudos are added with uppercase letters - // Remember that setFilters inherits from pseudos - var args, - fn = Expr.pseudos[ pseudo ] || Expr.setFilters[ pseudo.toLowerCase() ] || - Sizzle.error( "unsupported pseudo: " + pseudo ); - - // The user may use createPseudo to indicate that - // arguments are needed to create the filter function - // just as Sizzle does - if ( fn[ expando ] ) { - return fn( argument ); - } - - // But maintain support for old signatures - if ( fn.length > 1 ) { - args = [ pseudo, pseudo, "", argument ]; - return Expr.setFilters.hasOwnProperty( pseudo.toLowerCase() ) ? - markFunction( function( seed, matches ) { - var idx, - matched = fn( seed, argument ), - i = matched.length; - while ( i-- ) { - idx = indexOf( seed, matched[ i ] ); - seed[ idx ] = !( matches[ idx ] = matched[ i ] ); - } - } ) : - function( elem ) { - return fn( elem, 0, args ); - }; - } - - return fn; - } - }, - - pseudos: { - - // Potentially complex pseudos - "not": markFunction( function( selector ) { - - // Trim the selector passed to compile - // to avoid treating leading and trailing - // spaces as combinators - var input = [], - results = [], - matcher = compile( selector.replace( rtrim, "$1" ) ); - - return matcher[ expando ] ? - markFunction( function( seed, matches, _context, xml ) { - var elem, - unmatched = matcher( seed, null, xml, [] ), - i = seed.length; - - // Match elements unmatched by `matcher` - while ( i-- ) { - if ( ( elem = unmatched[ i ] ) ) { - seed[ i ] = !( matches[ i ] = elem ); - } - } - } ) : - function( elem, _context, xml ) { - input[ 0 ] = elem; - matcher( input, null, xml, results ); - - // Don't keep the element (issue #299) - input[ 0 ] = null; - return !results.pop(); - }; - } ), - - "has": markFunction( function( selector ) { - return function( elem ) { - return Sizzle( selector, elem ).length > 0; - }; - } ), - - "contains": markFunction( function( text ) { - text = text.replace( runescape, funescape ); - return function( elem ) { - return ( elem.textContent || getText( elem ) ).indexOf( text ) > -1; - }; - } ), - - // "Whether an element is represented by a :lang() selector - // is based solely on the element's language value - // being equal to the identifier C, - // or beginning with the identifier C immediately followed by "-". - // The matching of C against the element's language value is performed case-insensitively. - // The identifier C does not have to be a valid language name." - // http://www.w3.org/TR/selectors/#lang-pseudo - "lang": markFunction( function( lang ) { - - // lang value must be a valid identifier - if ( !ridentifier.test( lang || "" ) ) { - Sizzle.error( "unsupported lang: " + lang ); - } - lang = lang.replace( runescape, funescape ).toLowerCase(); - return function( elem ) { - var elemLang; - do { - if ( ( elemLang = documentIsHTML ? - elem.lang : - elem.getAttribute( "xml:lang" ) || elem.getAttribute( "lang" ) ) ) { - - elemLang = elemLang.toLowerCase(); - return elemLang === lang || elemLang.indexOf( lang + "-" ) === 0; - } - } while ( ( elem = elem.parentNode ) && elem.nodeType === 1 ); - return false; - }; - } ), - - // Miscellaneous - "target": function( elem ) { - var hash = window.location && window.location.hash; - return hash && hash.slice( 1 ) === elem.id; - }, - - "root": function( elem ) { - return elem === docElem; - }, - - "focus": function( elem ) { - return elem === document.activeElement && - ( !document.hasFocus || document.hasFocus() ) && - !!( elem.type || elem.href || ~elem.tabIndex ); - }, - - // Boolean properties - "enabled": createDisabledPseudo( false ), - "disabled": createDisabledPseudo( true ), - - "checked": function( elem ) { - - // In CSS3, :checked should return both checked and selected elements - // http://www.w3.org/TR/2011/REC-css3-selectors-20110929/#checked - var nodeName = elem.nodeName.toLowerCase(); - return ( nodeName === "input" && !!elem.checked ) || - ( nodeName === "option" && !!elem.selected ); - }, - - "selected": function( elem ) { - - // Accessing this property makes selected-by-default - // options in Safari work properly - if ( elem.parentNode ) { - // eslint-disable-next-line no-unused-expressions - elem.parentNode.selectedIndex; - } - - return elem.selected === true; - }, - - // Contents - "empty": function( elem ) { - - // http://www.w3.org/TR/selectors/#empty-pseudo - // :empty is negated by element (1) or content nodes (text: 3; cdata: 4; entity ref: 5), - // but not by others (comment: 8; processing instruction: 7; etc.) - // nodeType < 6 works because attributes (2) do not appear as children - for ( elem = elem.firstChild; elem; elem = elem.nextSibling ) { - if ( elem.nodeType < 6 ) { - return false; - } - } - return true; - }, - - "parent": function( elem ) { - return !Expr.pseudos[ "empty" ]( elem ); - }, - - // Element/input types - "header": function( elem ) { - return rheader.test( elem.nodeName ); - }, - - "input": function( elem ) { - return rinputs.test( elem.nodeName ); - }, - - "button": function( elem ) { - var name = elem.nodeName.toLowerCase(); - return name === "input" && elem.type === "button" || name === "button"; - }, - - "text": function( elem ) { - var attr; - return elem.nodeName.toLowerCase() === "input" && - elem.type === "text" && - - // Support: IE<8 - // New HTML5 attribute values (e.g., "search") appear with elem.type === "text" - ( ( attr = elem.getAttribute( "type" ) ) == null || - attr.toLowerCase() === "text" ); - }, - - // Position-in-collection - "first": createPositionalPseudo( function() { - return [ 0 ]; - } ), - - "last": createPositionalPseudo( function( _matchIndexes, length ) { - return [ length - 1 ]; - } ), - - "eq": createPositionalPseudo( function( _matchIndexes, length, argument ) { - return [ argument < 0 ? argument + length : argument ]; - } ), - - "even": createPositionalPseudo( function( matchIndexes, length ) { - var i = 0; - for ( ; i < length; i += 2 ) { - matchIndexes.push( i ); - } - return matchIndexes; - } ), - - "odd": createPositionalPseudo( function( matchIndexes, length ) { - var i = 1; - for ( ; i < length; i += 2 ) { - matchIndexes.push( i ); - } - return matchIndexes; - } ), - - "lt": createPositionalPseudo( function( matchIndexes, length, argument ) { - var i = argument < 0 ? - argument + length : - argument > length ? - length : - argument; - for ( ; --i >= 0; ) { - matchIndexes.push( i ); - } - return matchIndexes; - } ), - - "gt": createPositionalPseudo( function( matchIndexes, length, argument ) { - var i = argument < 0 ? argument + length : argument; - for ( ; ++i < length; ) { - matchIndexes.push( i ); - } - return matchIndexes; - } ) - } -}; - -Expr.pseudos[ "nth" ] = Expr.pseudos[ "eq" ]; - -// Add button/input type pseudos -for ( i in { radio: true, checkbox: true, file: true, password: true, image: true } ) { - Expr.pseudos[ i ] = createInputPseudo( i ); -} -for ( i in { submit: true, reset: true } ) { - Expr.pseudos[ i ] = createButtonPseudo( i ); -} - -// Easy API for creating new setFilters -function setFilters() {} -setFilters.prototype = Expr.filters = Expr.pseudos; -Expr.setFilters = new setFilters(); - -tokenize = Sizzle.tokenize = function( selector, parseOnly ) { - var matched, match, tokens, type, - soFar, groups, preFilters, - cached = tokenCache[ selector + " " ]; - - if ( cached ) { - return parseOnly ? 0 : cached.slice( 0 ); - } - - soFar = selector; - groups = []; - preFilters = Expr.preFilter; - - while ( soFar ) { - - // Comma and first run - if ( !matched || ( match = rcomma.exec( soFar ) ) ) { - if ( match ) { - - // Don't consume trailing commas as valid - soFar = soFar.slice( match[ 0 ].length ) || soFar; - } - groups.push( ( tokens = [] ) ); - } - - matched = false; - - // Combinators - if ( ( match = rcombinators.exec( soFar ) ) ) { - matched = match.shift(); - tokens.push( { - value: matched, - - // Cast descendant combinators to space - type: match[ 0 ].replace( rtrim, " " ) - } ); - soFar = soFar.slice( matched.length ); - } - - // Filters - for ( type in Expr.filter ) { - if ( ( match = matchExpr[ type ].exec( soFar ) ) && ( !preFilters[ type ] || - ( match = preFilters[ type ]( match ) ) ) ) { - matched = match.shift(); - tokens.push( { - value: matched, - type: type, - matches: match - } ); - soFar = soFar.slice( matched.length ); - } - } - - if ( !matched ) { - break; - } - } - - // Return the length of the invalid excess - // if we're just parsing - // Otherwise, throw an error or return tokens - return parseOnly ? - soFar.length : - soFar ? - Sizzle.error( selector ) : - - // Cache the tokens - tokenCache( selector, groups ).slice( 0 ); -}; - -function toSelector( tokens ) { - var i = 0, - len = tokens.length, - selector = ""; - for ( ; i < len; i++ ) { - selector += tokens[ i ].value; - } - return selector; -} - -function addCombinator( matcher, combinator, base ) { - var dir = combinator.dir, - skip = combinator.next, - key = skip || dir, - checkNonElements = base && key === "parentNode", - doneName = done++; - - return combinator.first ? - - // Check against closest ancestor/preceding element - function( elem, context, xml ) { - while ( ( elem = elem[ dir ] ) ) { - if ( elem.nodeType === 1 || checkNonElements ) { - return matcher( elem, context, xml ); - } - } - return false; - } : - - // Check against all ancestor/preceding elements - function( elem, context, xml ) { - var oldCache, uniqueCache, outerCache, - newCache = [ dirruns, doneName ]; - - // We can't set arbitrary data on XML nodes, so they don't benefit from combinator caching - if ( xml ) { - while ( ( elem = elem[ dir ] ) ) { - if ( elem.nodeType === 1 || checkNonElements ) { - if ( matcher( elem, context, xml ) ) { - return true; - } - } - } - } else { - while ( ( elem = elem[ dir ] ) ) { - if ( elem.nodeType === 1 || checkNonElements ) { - outerCache = elem[ expando ] || ( elem[ expando ] = {} ); - - // Support: IE <9 only - // Defend against cloned attroperties (jQuery gh-1709) - uniqueCache = outerCache[ elem.uniqueID ] || - ( outerCache[ elem.uniqueID ] = {} ); - - if ( skip && skip === elem.nodeName.toLowerCase() ) { - elem = elem[ dir ] || elem; - } else if ( ( oldCache = uniqueCache[ key ] ) && - oldCache[ 0 ] === dirruns && oldCache[ 1 ] === doneName ) { - - // Assign to newCache so results back-propagate to previous elements - return ( newCache[ 2 ] = oldCache[ 2 ] ); - } else { - - // Reuse newcache so results back-propagate to previous elements - uniqueCache[ key ] = newCache; - - // A match means we're done; a fail means we have to keep checking - if ( ( newCache[ 2 ] = matcher( elem, context, xml ) ) ) { - return true; - } - } - } - } - } - return false; - }; -} - -function elementMatcher( matchers ) { - return matchers.length > 1 ? - function( elem, context, xml ) { - var i = matchers.length; - while ( i-- ) { - if ( !matchers[ i ]( elem, context, xml ) ) { - return false; - } - } - return true; - } : - matchers[ 0 ]; -} - -function multipleContexts( selector, contexts, results ) { - var i = 0, - len = contexts.length; - for ( ; i < len; i++ ) { - Sizzle( selector, contexts[ i ], results ); - } - return results; -} - -function condense( unmatched, map, filter, context, xml ) { - var elem, - newUnmatched = [], - i = 0, - len = unmatched.length, - mapped = map != null; - - for ( ; i < len; i++ ) { - if ( ( elem = unmatched[ i ] ) ) { - if ( !filter || filter( elem, context, xml ) ) { - newUnmatched.push( elem ); - if ( mapped ) { - map.push( i ); - } - } - } - } - - return newUnmatched; -} - -function setMatcher( preFilter, selector, matcher, postFilter, postFinder, postSelector ) { - if ( postFilter && !postFilter[ expando ] ) { - postFilter = setMatcher( postFilter ); - } - if ( postFinder && !postFinder[ expando ] ) { - postFinder = setMatcher( postFinder, postSelector ); - } - return markFunction( function( seed, results, context, xml ) { - var temp, i, elem, - preMap = [], - postMap = [], - preexisting = results.length, - - // Get initial elements from seed or context - elems = seed || multipleContexts( - selector || "*", - context.nodeType ? [ context ] : context, - [] - ), - - // Prefilter to get matcher input, preserving a map for seed-results synchronization - matcherIn = preFilter && ( seed || !selector ) ? - condense( elems, preMap, preFilter, context, xml ) : - elems, - - matcherOut = matcher ? - - // If we have a postFinder, or filtered seed, or non-seed postFilter or preexisting results, - postFinder || ( seed ? preFilter : preexisting || postFilter ) ? - - // ...intermediate processing is necessary - [] : - - // ...otherwise use results directly - results : - matcherIn; - - // Find primary matches - if ( matcher ) { - matcher( matcherIn, matcherOut, context, xml ); - } - - // Apply postFilter - if ( postFilter ) { - temp = condense( matcherOut, postMap ); - postFilter( temp, [], context, xml ); - - // Un-match failing elements by moving them back to matcherIn - i = temp.length; - while ( i-- ) { - if ( ( elem = temp[ i ] ) ) { - matcherOut[ postMap[ i ] ] = !( matcherIn[ postMap[ i ] ] = elem ); - } - } - } - - if ( seed ) { - if ( postFinder || preFilter ) { - if ( postFinder ) { - - // Get the final matcherOut by condensing this intermediate into postFinder contexts - temp = []; - i = matcherOut.length; - while ( i-- ) { - if ( ( elem = matcherOut[ i ] ) ) { - - // Restore matcherIn since elem is not yet a final match - temp.push( ( matcherIn[ i ] = elem ) ); - } - } - postFinder( null, ( matcherOut = [] ), temp, xml ); - } - - // Move matched elements from seed to results to keep them synchronized - i = matcherOut.length; - while ( i-- ) { - if ( ( elem = matcherOut[ i ] ) && - ( temp = postFinder ? indexOf( seed, elem ) : preMap[ i ] ) > -1 ) { - - seed[ temp ] = !( results[ temp ] = elem ); - } - } - } - - // Add elements to results, through postFinder if defined - } else { - matcherOut = condense( - matcherOut === results ? - matcherOut.splice( preexisting, matcherOut.length ) : - matcherOut - ); - if ( postFinder ) { - postFinder( null, results, matcherOut, xml ); - } else { - push.apply( results, matcherOut ); - } - } - } ); -} - -function matcherFromTokens( tokens ) { - var checkContext, matcher, j, - len = tokens.length, - leadingRelative = Expr.relative[ tokens[ 0 ].type ], - implicitRelative = leadingRelative || Expr.relative[ " " ], - i = leadingRelative ? 1 : 0, - - // The foundational matcher ensures that elements are reachable from top-level context(s) - matchContext = addCombinator( function( elem ) { - return elem === checkContext; - }, implicitRelative, true ), - matchAnyContext = addCombinator( function( elem ) { - return indexOf( checkContext, elem ) > -1; - }, implicitRelative, true ), - matchers = [ function( elem, context, xml ) { - var ret = ( !leadingRelative && ( xml || context !== outermostContext ) ) || ( - ( checkContext = context ).nodeType ? - matchContext( elem, context, xml ) : - matchAnyContext( elem, context, xml ) ); - - // Avoid hanging onto element (issue #299) - checkContext = null; - return ret; - } ]; - - for ( ; i < len; i++ ) { - if ( ( matcher = Expr.relative[ tokens[ i ].type ] ) ) { - matchers = [ addCombinator( elementMatcher( matchers ), matcher ) ]; - } else { - matcher = Expr.filter[ tokens[ i ].type ].apply( null, tokens[ i ].matches ); - - // Return special upon seeing a positional matcher - if ( matcher[ expando ] ) { - - // Find the next relative operator (if any) for proper handling - j = ++i; - for ( ; j < len; j++ ) { - if ( Expr.relative[ tokens[ j ].type ] ) { - break; - } - } - return setMatcher( - i > 1 && elementMatcher( matchers ), - i > 1 && toSelector( - - // If the preceding token was a descendant combinator, insert an implicit any-element `*` - tokens - .slice( 0, i - 1 ) - .concat( { value: tokens[ i - 2 ].type === " " ? "*" : "" } ) - ).replace( rtrim, "$1" ), - matcher, - i < j && matcherFromTokens( tokens.slice( i, j ) ), - j < len && matcherFromTokens( ( tokens = tokens.slice( j ) ) ), - j < len && toSelector( tokens ) - ); - } - matchers.push( matcher ); - } - } - - return elementMatcher( matchers ); -} - -function matcherFromGroupMatchers( elementMatchers, setMatchers ) { - var bySet = setMatchers.length > 0, - byElement = elementMatchers.length > 0, - superMatcher = function( seed, context, xml, results, outermost ) { - var elem, j, matcher, - matchedCount = 0, - i = "0", - unmatched = seed && [], - setMatched = [], - contextBackup = outermostContext, - - // We must always have either seed elements or outermost context - elems = seed || byElement && Expr.find[ "TAG" ]( "*", outermost ), - - // Use integer dirruns iff this is the outermost matcher - dirrunsUnique = ( dirruns += contextBackup == null ? 1 : Math.random() || 0.1 ), - len = elems.length; - - if ( outermost ) { - - // Support: IE 11+, Edge 17 - 18+ - // IE/Edge sometimes throw a "Permission denied" error when strict-comparing - // two documents; shallow comparisons work. - // eslint-disable-next-line eqeqeq - outermostContext = context == document || context || outermost; - } - - // Add elements passing elementMatchers directly to results - // Support: IE<9, Safari - // Tolerate NodeList properties (IE: "length"; Safari: ) matching elements by id - for ( ; i !== len && ( elem = elems[ i ] ) != null; i++ ) { - if ( byElement && elem ) { - j = 0; - - // Support: IE 11+, Edge 17 - 18+ - // IE/Edge sometimes throw a "Permission denied" error when strict-comparing - // two documents; shallow comparisons work. - // eslint-disable-next-line eqeqeq - if ( !context && elem.ownerDocument != document ) { - setDocument( elem ); - xml = !documentIsHTML; - } - while ( ( matcher = elementMatchers[ j++ ] ) ) { - if ( matcher( elem, context || document, xml ) ) { - results.push( elem ); - break; - } - } - if ( outermost ) { - dirruns = dirrunsUnique; - } - } - - // Track unmatched elements for set filters - if ( bySet ) { - - // They will have gone through all possible matchers - if ( ( elem = !matcher && elem ) ) { - matchedCount--; - } - - // Lengthen the array for every element, matched or not - if ( seed ) { - unmatched.push( elem ); - } - } - } - - // `i` is now the count of elements visited above, and adding it to `matchedCount` - // makes the latter nonnegative. - matchedCount += i; - - // Apply set filters to unmatched elements - // NOTE: This can be skipped if there are no unmatched elements (i.e., `matchedCount` - // equals `i`), unless we didn't visit _any_ elements in the above loop because we have - // no element matchers and no seed. - // Incrementing an initially-string "0" `i` allows `i` to remain a string only in that - // case, which will result in a "00" `matchedCount` that differs from `i` but is also - // numerically zero. - if ( bySet && i !== matchedCount ) { - j = 0; - while ( ( matcher = setMatchers[ j++ ] ) ) { - matcher( unmatched, setMatched, context, xml ); - } - - if ( seed ) { - - // Reintegrate element matches to eliminate the need for sorting - if ( matchedCount > 0 ) { - while ( i-- ) { - if ( !( unmatched[ i ] || setMatched[ i ] ) ) { - setMatched[ i ] = pop.call( results ); - } - } - } - - // Discard index placeholder values to get only actual matches - setMatched = condense( setMatched ); - } - - // Add matches to results - push.apply( results, setMatched ); - - // Seedless set matches succeeding multiple successful matchers stipulate sorting - if ( outermost && !seed && setMatched.length > 0 && - ( matchedCount + setMatchers.length ) > 1 ) { - - Sizzle.uniqueSort( results ); - } - } - - // Override manipulation of globals by nested matchers - if ( outermost ) { - dirruns = dirrunsUnique; - outermostContext = contextBackup; - } - - return unmatched; - }; - - return bySet ? - markFunction( superMatcher ) : - superMatcher; -} - -compile = Sizzle.compile = function( selector, match /* Internal Use Only */ ) { - var i, - setMatchers = [], - elementMatchers = [], - cached = compilerCache[ selector + " " ]; - - if ( !cached ) { - - // Generate a function of recursive functions that can be used to check each element - if ( !match ) { - match = tokenize( selector ); - } - i = match.length; - while ( i-- ) { - cached = matcherFromTokens( match[ i ] ); - if ( cached[ expando ] ) { - setMatchers.push( cached ); - } else { - elementMatchers.push( cached ); - } - } - - // Cache the compiled function - cached = compilerCache( - selector, - matcherFromGroupMatchers( elementMatchers, setMatchers ) - ); - - // Save selector and tokenization - cached.selector = selector; - } - return cached; -}; - -/** - * A low-level selection function that works with Sizzle's compiled - * selector functions - * @param {String|Function} selector A selector or a pre-compiled - * selector function built with Sizzle.compile - * @param {Element} context - * @param {Array} [results] - * @param {Array} [seed] A set of elements to match against - */ -select = Sizzle.select = function( selector, context, results, seed ) { - var i, tokens, token, type, find, - compiled = typeof selector === "function" && selector, - match = !seed && tokenize( ( selector = compiled.selector || selector ) ); - - results = results || []; - - // Try to minimize operations if there is only one selector in the list and no seed - // (the latter of which guarantees us context) - if ( match.length === 1 ) { - - // Reduce context if the leading compound selector is an ID - tokens = match[ 0 ] = match[ 0 ].slice( 0 ); - if ( tokens.length > 2 && ( token = tokens[ 0 ] ).type === "ID" && - context.nodeType === 9 && documentIsHTML && Expr.relative[ tokens[ 1 ].type ] ) { - - context = ( Expr.find[ "ID" ]( token.matches[ 0 ] - .replace( runescape, funescape ), context ) || [] )[ 0 ]; - if ( !context ) { - return results; - - // Precompiled matchers will still verify ancestry, so step up a level - } else if ( compiled ) { - context = context.parentNode; - } - - selector = selector.slice( tokens.shift().value.length ); - } - - // Fetch a seed set for right-to-left matching - i = matchExpr[ "needsContext" ].test( selector ) ? 0 : tokens.length; - while ( i-- ) { - token = tokens[ i ]; - - // Abort if we hit a combinator - if ( Expr.relative[ ( type = token.type ) ] ) { - break; - } - if ( ( find = Expr.find[ type ] ) ) { - - // Search, expanding context for leading sibling combinators - if ( ( seed = find( - token.matches[ 0 ].replace( runescape, funescape ), - rsibling.test( tokens[ 0 ].type ) && testContext( context.parentNode ) || - context - ) ) ) { - - // If seed is empty or no tokens remain, we can return early - tokens.splice( i, 1 ); - selector = seed.length && toSelector( tokens ); - if ( !selector ) { - push.apply( results, seed ); - return results; - } - - break; - } - } - } - } - - // Compile and execute a filtering function if one is not provided - // Provide `match` to avoid retokenization if we modified the selector above - ( compiled || compile( selector, match ) )( - seed, - context, - !documentIsHTML, - results, - !context || rsibling.test( selector ) && testContext( context.parentNode ) || context - ); - return results; -}; - -// One-time assignments - -// Sort stability -support.sortStable = expando.split( "" ).sort( sortOrder ).join( "" ) === expando; - -// Support: Chrome 14-35+ -// Always assume duplicates if they aren't passed to the comparison function -support.detectDuplicates = !!hasDuplicate; - -// Initialize against the default document -setDocument(); - -// Support: Webkit<537.32 - Safari 6.0.3/Chrome 25 (fixed in Chrome 27) -// Detached nodes confoundingly follow *each other* -support.sortDetached = assert( function( el ) { - - // Should return 1, but returns 4 (following) - return el.compareDocumentPosition( document.createElement( "fieldset" ) ) & 1; -} ); - -// Support: IE<8 -// Prevent attribute/property "interpolation" -// https://msdn.microsoft.com/en-us/library/ms536429%28VS.85%29.aspx -if ( !assert( function( el ) { - el.innerHTML = ""; - return el.firstChild.getAttribute( "href" ) === "#"; -} ) ) { - addHandle( "type|href|height|width", function( elem, name, isXML ) { - if ( !isXML ) { - return elem.getAttribute( name, name.toLowerCase() === "type" ? 1 : 2 ); - } - } ); -} - -// Support: IE<9 -// Use defaultValue in place of getAttribute("value") -if ( !support.attributes || !assert( function( el ) { - el.innerHTML = ""; - el.firstChild.setAttribute( "value", "" ); - return el.firstChild.getAttribute( "value" ) === ""; -} ) ) { - addHandle( "value", function( elem, _name, isXML ) { - if ( !isXML && elem.nodeName.toLowerCase() === "input" ) { - return elem.defaultValue; - } - } ); -} - -// Support: IE<9 -// Use getAttributeNode to fetch booleans when getAttribute lies -if ( !assert( function( el ) { - return el.getAttribute( "disabled" ) == null; -} ) ) { - addHandle( booleans, function( elem, name, isXML ) { - var val; - if ( !isXML ) { - return elem[ name ] === true ? name.toLowerCase() : - ( val = elem.getAttributeNode( name ) ) && val.specified ? - val.value : - null; - } - } ); -} - -return Sizzle; - -} )( window ); - - - -jQuery.find = Sizzle; -jQuery.expr = Sizzle.selectors; - -// Deprecated -jQuery.expr[ ":" ] = jQuery.expr.pseudos; -jQuery.uniqueSort = jQuery.unique = Sizzle.uniqueSort; -jQuery.text = Sizzle.getText; -jQuery.isXMLDoc = Sizzle.isXML; -jQuery.contains = Sizzle.contains; -jQuery.escapeSelector = Sizzle.escape; - - - - -var dir = function( elem, dir, until ) { - var matched = [], - truncate = until !== undefined; - - while ( ( elem = elem[ dir ] ) && elem.nodeType !== 9 ) { - if ( elem.nodeType === 1 ) { - if ( truncate && jQuery( elem ).is( until ) ) { - break; - } - matched.push( elem ); - } - } - return matched; -}; - - -var siblings = function( n, elem ) { - var matched = []; - - for ( ; n; n = n.nextSibling ) { - if ( n.nodeType === 1 && n !== elem ) { - matched.push( n ); - } - } - - return matched; -}; - - -var rneedsContext = jQuery.expr.match.needsContext; - - - -function nodeName( elem, name ) { - - return elem.nodeName && elem.nodeName.toLowerCase() === name.toLowerCase(); - -}; -var rsingleTag = ( /^<([a-z][^\/\0>:\x20\t\r\n\f]*)[\x20\t\r\n\f]*\/?>(?:<\/\1>|)$/i ); - - - -// Implement the identical functionality for filter and not -function winnow( elements, qualifier, not ) { - if ( isFunction( qualifier ) ) { - return jQuery.grep( elements, function( elem, i ) { - return !!qualifier.call( elem, i, elem ) !== not; - } ); - } - - // Single element - if ( qualifier.nodeType ) { - return jQuery.grep( elements, function( elem ) { - return ( elem === qualifier ) !== not; - } ); - } - - // Arraylike of elements (jQuery, arguments, Array) - if ( typeof qualifier !== "string" ) { - return jQuery.grep( elements, function( elem ) { - return ( indexOf.call( qualifier, elem ) > -1 ) !== not; - } ); - } - - // Filtered directly for both simple and complex selectors - return jQuery.filter( qualifier, elements, not ); -} - -jQuery.filter = function( expr, elems, not ) { - var elem = elems[ 0 ]; - - if ( not ) { - expr = ":not(" + expr + ")"; - } - - if ( elems.length === 1 && elem.nodeType === 1 ) { - return jQuery.find.matchesSelector( elem, expr ) ? [ elem ] : []; - } - - return jQuery.find.matches( expr, jQuery.grep( elems, function( elem ) { - return elem.nodeType === 1; - } ) ); -}; - -jQuery.fn.extend( { - find: function( selector ) { - var i, ret, - len = this.length, - self = this; - - if ( typeof selector !== "string" ) { - return this.pushStack( jQuery( selector ).filter( function() { - for ( i = 0; i < len; i++ ) { - if ( jQuery.contains( self[ i ], this ) ) { - return true; - } - } - } ) ); - } - - ret = this.pushStack( [] ); - - for ( i = 0; i < len; i++ ) { - jQuery.find( selector, self[ i ], ret ); - } - - return len > 1 ? jQuery.uniqueSort( ret ) : ret; - }, - filter: function( selector ) { - return this.pushStack( winnow( this, selector || [], false ) ); - }, - not: function( selector ) { - return this.pushStack( winnow( this, selector || [], true ) ); - }, - is: function( selector ) { - return !!winnow( - this, - - // If this is a positional/relative selector, check membership in the returned set - // so $("p:first").is("p:last") won't return true for a doc with two "p". - typeof selector === "string" && rneedsContext.test( selector ) ? - jQuery( selector ) : - selector || [], - false - ).length; - } -} ); - - -// Initialize a jQuery object - - -// A central reference to the root jQuery(document) -var rootjQuery, - - // A simple way to check for HTML strings - // Prioritize #id over to avoid XSS via location.hash (#9521) - // Strict HTML recognition (#11290: must start with <) - // Shortcut simple #id case for speed - rquickExpr = /^(?:\s*(<[\w\W]+>)[^>]*|#([\w-]+))$/, - - init = jQuery.fn.init = function( selector, context, root ) { - var match, elem; - - // HANDLE: $(""), $(null), $(undefined), $(false) - if ( !selector ) { - return this; - } - - // Method init() accepts an alternate rootjQuery - // so migrate can support jQuery.sub (gh-2101) - root = root || rootjQuery; - - // Handle HTML strings - if ( typeof selector === "string" ) { - if ( selector[ 0 ] === "<" && - selector[ selector.length - 1 ] === ">" && - selector.length >= 3 ) { - - // Assume that strings that start and end with <> are HTML and skip the regex check - match = [ null, selector, null ]; - - } else { - match = rquickExpr.exec( selector ); - } - - // Match html or make sure no context is specified for #id - if ( match && ( match[ 1 ] || !context ) ) { - - // HANDLE: $(html) -> $(array) - if ( match[ 1 ] ) { - context = context instanceof jQuery ? context[ 0 ] : context; - - // Option to run scripts is true for back-compat - // Intentionally let the error be thrown if parseHTML is not present - jQuery.merge( this, jQuery.parseHTML( - match[ 1 ], - context && context.nodeType ? context.ownerDocument || context : document, - true - ) ); - - // HANDLE: $(html, props) - if ( rsingleTag.test( match[ 1 ] ) && jQuery.isPlainObject( context ) ) { - for ( match in context ) { - - // Properties of context are called as methods if possible - if ( isFunction( this[ match ] ) ) { - this[ match ]( context[ match ] ); - - // ...and otherwise set as attributes - } else { - this.attr( match, context[ match ] ); - } - } - } - - return this; - - // HANDLE: $(#id) - } else { - elem = document.getElementById( match[ 2 ] ); - - if ( elem ) { - - // Inject the element directly into the jQuery object - this[ 0 ] = elem; - this.length = 1; - } - return this; - } - - // HANDLE: $(expr, $(...)) - } else if ( !context || context.jquery ) { - return ( context || root ).find( selector ); - - // HANDLE: $(expr, context) - // (which is just equivalent to: $(context).find(expr) - } else { - return this.constructor( context ).find( selector ); - } - - // HANDLE: $(DOMElement) - } else if ( selector.nodeType ) { - this[ 0 ] = selector; - this.length = 1; - return this; - - // HANDLE: $(function) - // Shortcut for document ready - } else if ( isFunction( selector ) ) { - return root.ready !== undefined ? - root.ready( selector ) : - - // Execute immediately if ready is not present - selector( jQuery ); - } - - return jQuery.makeArray( selector, this ); - }; - -// Give the init function the jQuery prototype for later instantiation -init.prototype = jQuery.fn; - -// Initialize central reference -rootjQuery = jQuery( document ); - - -var rparentsprev = /^(?:parents|prev(?:Until|All))/, - - // Methods guaranteed to produce a unique set when starting from a unique set - guaranteedUnique = { - children: true, - contents: true, - next: true, - prev: true - }; - -jQuery.fn.extend( { - has: function( target ) { - var targets = jQuery( target, this ), - l = targets.length; - - return this.filter( function() { - var i = 0; - for ( ; i < l; i++ ) { - if ( jQuery.contains( this, targets[ i ] ) ) { - return true; - } - } - } ); - }, - - closest: function( selectors, context ) { - var cur, - i = 0, - l = this.length, - matched = [], - targets = typeof selectors !== "string" && jQuery( selectors ); - - // Positional selectors never match, since there's no _selection_ context - if ( !rneedsContext.test( selectors ) ) { - for ( ; i < l; i++ ) { - for ( cur = this[ i ]; cur && cur !== context; cur = cur.parentNode ) { - - // Always skip document fragments - if ( cur.nodeType < 11 && ( targets ? - targets.index( cur ) > -1 : - - // Don't pass non-elements to Sizzle - cur.nodeType === 1 && - jQuery.find.matchesSelector( cur, selectors ) ) ) { - - matched.push( cur ); - break; - } - } - } - } - - return this.pushStack( matched.length > 1 ? jQuery.uniqueSort( matched ) : matched ); - }, - - // Determine the position of an element within the set - index: function( elem ) { - - // No argument, return index in parent - if ( !elem ) { - return ( this[ 0 ] && this[ 0 ].parentNode ) ? this.first().prevAll().length : -1; - } - - // Index in selector - if ( typeof elem === "string" ) { - return indexOf.call( jQuery( elem ), this[ 0 ] ); - } - - // Locate the position of the desired element - return indexOf.call( this, - - // If it receives a jQuery object, the first element is used - elem.jquery ? elem[ 0 ] : elem - ); - }, - - add: function( selector, context ) { - return this.pushStack( - jQuery.uniqueSort( - jQuery.merge( this.get(), jQuery( selector, context ) ) - ) - ); - }, - - addBack: function( selector ) { - return this.add( selector == null ? - this.prevObject : this.prevObject.filter( selector ) - ); - } -} ); - -function sibling( cur, dir ) { - while ( ( cur = cur[ dir ] ) && cur.nodeType !== 1 ) {} - return cur; -} - -jQuery.each( { - parent: function( elem ) { - var parent = elem.parentNode; - return parent && parent.nodeType !== 11 ? parent : null; - }, - parents: function( elem ) { - return dir( elem, "parentNode" ); - }, - parentsUntil: function( elem, _i, until ) { - return dir( elem, "parentNode", until ); - }, - next: function( elem ) { - return sibling( elem, "nextSibling" ); - }, - prev: function( elem ) { - return sibling( elem, "previousSibling" ); - }, - nextAll: function( elem ) { - return dir( elem, "nextSibling" ); - }, - prevAll: function( elem ) { - return dir( elem, "previousSibling" ); - }, - nextUntil: function( elem, _i, until ) { - return dir( elem, "nextSibling", until ); - }, - prevUntil: function( elem, _i, until ) { - return dir( elem, "previousSibling", until ); - }, - siblings: function( elem ) { - return siblings( ( elem.parentNode || {} ).firstChild, elem ); - }, - children: function( elem ) { - return siblings( elem.firstChild ); - }, - contents: function( elem ) { - if ( elem.contentDocument != null && - - // Support: IE 11+ - // elements with no `data` attribute has an object - // `contentDocument` with a `null` prototype. - getProto( elem.contentDocument ) ) { - - return elem.contentDocument; - } - - // Support: IE 9 - 11 only, iOS 7 only, Android Browser <=4.3 only - // Treat the template element as a regular one in browsers that - // don't support it. - if ( nodeName( elem, "template" ) ) { - elem = elem.content || elem; - } - - return jQuery.merge( [], elem.childNodes ); - } -}, function( name, fn ) { - jQuery.fn[ name ] = function( until, selector ) { - var matched = jQuery.map( this, fn, until ); - - if ( name.slice( -5 ) !== "Until" ) { - selector = until; - } - - if ( selector && typeof selector === "string" ) { - matched = jQuery.filter( selector, matched ); - } - - if ( this.length > 1 ) { - - // Remove duplicates - if ( !guaranteedUnique[ name ] ) { - jQuery.uniqueSort( matched ); - } - - // Reverse order for parents* and prev-derivatives - if ( rparentsprev.test( name ) ) { - matched.reverse(); - } - } - - return this.pushStack( matched ); - }; -} ); -var rnothtmlwhite = ( /[^\x20\t\r\n\f]+/g ); - - - -// Convert String-formatted options into Object-formatted ones -function createOptions( options ) { - var object = {}; - jQuery.each( options.match( rnothtmlwhite ) || [], function( _, flag ) { - object[ flag ] = true; - } ); - return object; -} - -/* - * Create a callback list using the following parameters: - * - * options: an optional list of space-separated options that will change how - * the callback list behaves or a more traditional option object - * - * By default a callback list will act like an event callback list and can be - * "fired" multiple times. - * - * Possible options: - * - * once: will ensure the callback list can only be fired once (like a Deferred) - * - * memory: will keep track of previous values and will call any callback added - * after the list has been fired right away with the latest "memorized" - * values (like a Deferred) - * - * unique: will ensure a callback can only be added once (no duplicate in the list) - * - * stopOnFalse: interrupt callings when a callback returns false - * - */ -jQuery.Callbacks = function( options ) { - - // Convert options from String-formatted to Object-formatted if needed - // (we check in cache first) - options = typeof options === "string" ? - createOptions( options ) : - jQuery.extend( {}, options ); - - var // Flag to know if list is currently firing - firing, - - // Last fire value for non-forgettable lists - memory, - - // Flag to know if list was already fired - fired, - - // Flag to prevent firing - locked, - - // Actual callback list - list = [], - - // Queue of execution data for repeatable lists - queue = [], - - // Index of currently firing callback (modified by add/remove as needed) - firingIndex = -1, - - // Fire callbacks - fire = function() { - - // Enforce single-firing - locked = locked || options.once; - - // Execute callbacks for all pending executions, - // respecting firingIndex overrides and runtime changes - fired = firing = true; - for ( ; queue.length; firingIndex = -1 ) { - memory = queue.shift(); - while ( ++firingIndex < list.length ) { - - // Run callback and check for early termination - if ( list[ firingIndex ].apply( memory[ 0 ], memory[ 1 ] ) === false && - options.stopOnFalse ) { - - // Jump to end and forget the data so .add doesn't re-fire - firingIndex = list.length; - memory = false; - } - } - } - - // Forget the data if we're done with it - if ( !options.memory ) { - memory = false; - } - - firing = false; - - // Clean up if we're done firing for good - if ( locked ) { - - // Keep an empty list if we have data for future add calls - if ( memory ) { - list = []; - - // Otherwise, this object is spent - } else { - list = ""; - } - } - }, - - // Actual Callbacks object - self = { - - // Add a callback or a collection of callbacks to the list - add: function() { - if ( list ) { - - // If we have memory from a past run, we should fire after adding - if ( memory && !firing ) { - firingIndex = list.length - 1; - queue.push( memory ); - } - - ( function add( args ) { - jQuery.each( args, function( _, arg ) { - if ( isFunction( arg ) ) { - if ( !options.unique || !self.has( arg ) ) { - list.push( arg ); - } - } else if ( arg && arg.length && toType( arg ) !== "string" ) { - - // Inspect recursively - add( arg ); - } - } ); - } )( arguments ); - - if ( memory && !firing ) { - fire(); - } - } - return this; - }, - - // Remove a callback from the list - remove: function() { - jQuery.each( arguments, function( _, arg ) { - var index; - while ( ( index = jQuery.inArray( arg, list, index ) ) > -1 ) { - list.splice( index, 1 ); - - // Handle firing indexes - if ( index <= firingIndex ) { - firingIndex--; - } - } - } ); - return this; - }, - - // Check if a given callback is in the list. - // If no argument is given, return whether or not list has callbacks attached. - has: function( fn ) { - return fn ? - jQuery.inArray( fn, list ) > -1 : - list.length > 0; - }, - - // Remove all callbacks from the list - empty: function() { - if ( list ) { - list = []; - } - return this; - }, - - // Disable .fire and .add - // Abort any current/pending executions - // Clear all callbacks and values - disable: function() { - locked = queue = []; - list = memory = ""; - return this; - }, - disabled: function() { - return !list; - }, - - // Disable .fire - // Also disable .add unless we have memory (since it would have no effect) - // Abort any pending executions - lock: function() { - locked = queue = []; - if ( !memory && !firing ) { - list = memory = ""; - } - return this; - }, - locked: function() { - return !!locked; - }, - - // Call all callbacks with the given context and arguments - fireWith: function( context, args ) { - if ( !locked ) { - args = args || []; - args = [ context, args.slice ? args.slice() : args ]; - queue.push( args ); - if ( !firing ) { - fire(); - } - } - return this; - }, - - // Call all the callbacks with the given arguments - fire: function() { - self.fireWith( this, arguments ); - return this; - }, - - // To know if the callbacks have already been called at least once - fired: function() { - return !!fired; - } - }; - - return self; -}; - - -function Identity( v ) { - return v; -} -function Thrower( ex ) { - throw ex; -} - -function adoptValue( value, resolve, reject, noValue ) { - var method; - - try { - - // Check for promise aspect first to privilege synchronous behavior - if ( value && isFunction( ( method = value.promise ) ) ) { - method.call( value ).done( resolve ).fail( reject ); - - // Other thenables - } else if ( value && isFunction( ( method = value.then ) ) ) { - method.call( value, resolve, reject ); - - // Other non-thenables - } else { - - // Control `resolve` arguments by letting Array#slice cast boolean `noValue` to integer: - // * false: [ value ].slice( 0 ) => resolve( value ) - // * true: [ value ].slice( 1 ) => resolve() - resolve.apply( undefined, [ value ].slice( noValue ) ); - } - - // For Promises/A+, convert exceptions into rejections - // Since jQuery.when doesn't unwrap thenables, we can skip the extra checks appearing in - // Deferred#then to conditionally suppress rejection. - } catch ( value ) { - - // Support: Android 4.0 only - // Strict mode functions invoked without .call/.apply get global-object context - reject.apply( undefined, [ value ] ); - } -} - -jQuery.extend( { - - Deferred: function( func ) { - var tuples = [ - - // action, add listener, callbacks, - // ... .then handlers, argument index, [final state] - [ "notify", "progress", jQuery.Callbacks( "memory" ), - jQuery.Callbacks( "memory" ), 2 ], - [ "resolve", "done", jQuery.Callbacks( "once memory" ), - jQuery.Callbacks( "once memory" ), 0, "resolved" ], - [ "reject", "fail", jQuery.Callbacks( "once memory" ), - jQuery.Callbacks( "once memory" ), 1, "rejected" ] - ], - state = "pending", - promise = { - state: function() { - return state; - }, - always: function() { - deferred.done( arguments ).fail( arguments ); - return this; - }, - "catch": function( fn ) { - return promise.then( null, fn ); - }, - - // Keep pipe for back-compat - pipe: function( /* fnDone, fnFail, fnProgress */ ) { - var fns = arguments; - - return jQuery.Deferred( function( newDefer ) { - jQuery.each( tuples, function( _i, tuple ) { - - // Map tuples (progress, done, fail) to arguments (done, fail, progress) - var fn = isFunction( fns[ tuple[ 4 ] ] ) && fns[ tuple[ 4 ] ]; - - // deferred.progress(function() { bind to newDefer or newDefer.notify }) - // deferred.done(function() { bind to newDefer or newDefer.resolve }) - // deferred.fail(function() { bind to newDefer or newDefer.reject }) - deferred[ tuple[ 1 ] ]( function() { - var returned = fn && fn.apply( this, arguments ); - if ( returned && isFunction( returned.promise ) ) { - returned.promise() - .progress( newDefer.notify ) - .done( newDefer.resolve ) - .fail( newDefer.reject ); - } else { - newDefer[ tuple[ 0 ] + "With" ]( - this, - fn ? [ returned ] : arguments - ); - } - } ); - } ); - fns = null; - } ).promise(); - }, - then: function( onFulfilled, onRejected, onProgress ) { - var maxDepth = 0; - function resolve( depth, deferred, handler, special ) { - return function() { - var that = this, - args = arguments, - mightThrow = function() { - var returned, then; - - // Support: Promises/A+ section 2.3.3.3.3 - // https://promisesaplus.com/#point-59 - // Ignore double-resolution attempts - if ( depth < maxDepth ) { - return; - } - - returned = handler.apply( that, args ); - - // Support: Promises/A+ section 2.3.1 - // https://promisesaplus.com/#point-48 - if ( returned === deferred.promise() ) { - throw new TypeError( "Thenable self-resolution" ); - } - - // Support: Promises/A+ sections 2.3.3.1, 3.5 - // https://promisesaplus.com/#point-54 - // https://promisesaplus.com/#point-75 - // Retrieve `then` only once - then = returned && - - // Support: Promises/A+ section 2.3.4 - // https://promisesaplus.com/#point-64 - // Only check objects and functions for thenability - ( typeof returned === "object" || - typeof returned === "function" ) && - returned.then; - - // Handle a returned thenable - if ( isFunction( then ) ) { - - // Special processors (notify) just wait for resolution - if ( special ) { - then.call( - returned, - resolve( maxDepth, deferred, Identity, special ), - resolve( maxDepth, deferred, Thrower, special ) - ); - - // Normal processors (resolve) also hook into progress - } else { - - // ...and disregard older resolution values - maxDepth++; - - then.call( - returned, - resolve( maxDepth, deferred, Identity, special ), - resolve( maxDepth, deferred, Thrower, special ), - resolve( maxDepth, deferred, Identity, - deferred.notifyWith ) - ); - } - - // Handle all other returned values - } else { - - // Only substitute handlers pass on context - // and multiple values (non-spec behavior) - if ( handler !== Identity ) { - that = undefined; - args = [ returned ]; - } - - // Process the value(s) - // Default process is resolve - ( special || deferred.resolveWith )( that, args ); - } - }, - - // Only normal processors (resolve) catch and reject exceptions - process = special ? - mightThrow : - function() { - try { - mightThrow(); - } catch ( e ) { - - if ( jQuery.Deferred.exceptionHook ) { - jQuery.Deferred.exceptionHook( e, - process.stackTrace ); - } - - // Support: Promises/A+ section 2.3.3.3.4.1 - // https://promisesaplus.com/#point-61 - // Ignore post-resolution exceptions - if ( depth + 1 >= maxDepth ) { - - // Only substitute handlers pass on context - // and multiple values (non-spec behavior) - if ( handler !== Thrower ) { - that = undefined; - args = [ e ]; - } - - deferred.rejectWith( that, args ); - } - } - }; - - // Support: Promises/A+ section 2.3.3.3.1 - // https://promisesaplus.com/#point-57 - // Re-resolve promises immediately to dodge false rejection from - // subsequent errors - if ( depth ) { - process(); - } else { - - // Call an optional hook to record the stack, in case of exception - // since it's otherwise lost when execution goes async - if ( jQuery.Deferred.getStackHook ) { - process.stackTrace = jQuery.Deferred.getStackHook(); - } - window.setTimeout( process ); - } - }; - } - - return jQuery.Deferred( function( newDefer ) { - - // progress_handlers.add( ... ) - tuples[ 0 ][ 3 ].add( - resolve( - 0, - newDefer, - isFunction( onProgress ) ? - onProgress : - Identity, - newDefer.notifyWith - ) - ); - - // fulfilled_handlers.add( ... ) - tuples[ 1 ][ 3 ].add( - resolve( - 0, - newDefer, - isFunction( onFulfilled ) ? - onFulfilled : - Identity - ) - ); - - // rejected_handlers.add( ... ) - tuples[ 2 ][ 3 ].add( - resolve( - 0, - newDefer, - isFunction( onRejected ) ? - onRejected : - Thrower - ) - ); - } ).promise(); - }, - - // Get a promise for this deferred - // If obj is provided, the promise aspect is added to the object - promise: function( obj ) { - return obj != null ? jQuery.extend( obj, promise ) : promise; - } - }, - deferred = {}; - - // Add list-specific methods - jQuery.each( tuples, function( i, tuple ) { - var list = tuple[ 2 ], - stateString = tuple[ 5 ]; - - // promise.progress = list.add - // promise.done = list.add - // promise.fail = list.add - promise[ tuple[ 1 ] ] = list.add; - - // Handle state - if ( stateString ) { - list.add( - function() { - - // state = "resolved" (i.e., fulfilled) - // state = "rejected" - state = stateString; - }, - - // rejected_callbacks.disable - // fulfilled_callbacks.disable - tuples[ 3 - i ][ 2 ].disable, - - // rejected_handlers.disable - // fulfilled_handlers.disable - tuples[ 3 - i ][ 3 ].disable, - - // progress_callbacks.lock - tuples[ 0 ][ 2 ].lock, - - // progress_handlers.lock - tuples[ 0 ][ 3 ].lock - ); - } - - // progress_handlers.fire - // fulfilled_handlers.fire - // rejected_handlers.fire - list.add( tuple[ 3 ].fire ); - - // deferred.notify = function() { deferred.notifyWith(...) } - // deferred.resolve = function() { deferred.resolveWith(...) } - // deferred.reject = function() { deferred.rejectWith(...) } - deferred[ tuple[ 0 ] ] = function() { - deferred[ tuple[ 0 ] + "With" ]( this === deferred ? undefined : this, arguments ); - return this; - }; - - // deferred.notifyWith = list.fireWith - // deferred.resolveWith = list.fireWith - // deferred.rejectWith = list.fireWith - deferred[ tuple[ 0 ] + "With" ] = list.fireWith; - } ); - - // Make the deferred a promise - promise.promise( deferred ); - - // Call given func if any - if ( func ) { - func.call( deferred, deferred ); - } - - // All done! - return deferred; - }, - - // Deferred helper - when: function( singleValue ) { - var - - // count of uncompleted subordinates - remaining = arguments.length, - - // count of unprocessed arguments - i = remaining, - - // subordinate fulfillment data - resolveContexts = Array( i ), - resolveValues = slice.call( arguments ), - - // the master Deferred - master = jQuery.Deferred(), - - // subordinate callback factory - updateFunc = function( i ) { - return function( value ) { - resolveContexts[ i ] = this; - resolveValues[ i ] = arguments.length > 1 ? slice.call( arguments ) : value; - if ( !( --remaining ) ) { - master.resolveWith( resolveContexts, resolveValues ); - } - }; - }; - - // Single- and empty arguments are adopted like Promise.resolve - if ( remaining <= 1 ) { - adoptValue( singleValue, master.done( updateFunc( i ) ).resolve, master.reject, - !remaining ); - - // Use .then() to unwrap secondary thenables (cf. gh-3000) - if ( master.state() === "pending" || - isFunction( resolveValues[ i ] && resolveValues[ i ].then ) ) { - - return master.then(); - } - } - - // Multiple arguments are aggregated like Promise.all array elements - while ( i-- ) { - adoptValue( resolveValues[ i ], updateFunc( i ), master.reject ); - } - - return master.promise(); - } -} ); - - -// These usually indicate a programmer mistake during development, -// warn about them ASAP rather than swallowing them by default. -var rerrorNames = /^(Eval|Internal|Range|Reference|Syntax|Type|URI)Error$/; - -jQuery.Deferred.exceptionHook = function( error, stack ) { - - // Support: IE 8 - 9 only - // Console exists when dev tools are open, which can happen at any time - if ( window.console && window.console.warn && error && rerrorNames.test( error.name ) ) { - window.console.warn( "jQuery.Deferred exception: " + error.message, error.stack, stack ); - } -}; - - - - -jQuery.readyException = function( error ) { - window.setTimeout( function() { - throw error; - } ); -}; - - - - -// The deferred used on DOM ready -var readyList = jQuery.Deferred(); - -jQuery.fn.ready = function( fn ) { - - readyList - .then( fn ) - - // Wrap jQuery.readyException in a function so that the lookup - // happens at the time of error handling instead of callback - // registration. - .catch( function( error ) { - jQuery.readyException( error ); - } ); - - return this; -}; - -jQuery.extend( { - - // Is the DOM ready to be used? Set to true once it occurs. - isReady: false, - - // A counter to track how many items to wait for before - // the ready event fires. See #6781 - readyWait: 1, - - // Handle when the DOM is ready - ready: function( wait ) { - - // Abort if there are pending holds or we're already ready - if ( wait === true ? --jQuery.readyWait : jQuery.isReady ) { - return; - } - - // Remember that the DOM is ready - jQuery.isReady = true; - - // If a normal DOM Ready event fired, decrement, and wait if need be - if ( wait !== true && --jQuery.readyWait > 0 ) { - return; - } - - // If there are functions bound, to execute - readyList.resolveWith( document, [ jQuery ] ); - } -} ); - -jQuery.ready.then = readyList.then; - -// The ready event handler and self cleanup method -function completed() { - document.removeEventListener( "DOMContentLoaded", completed ); - window.removeEventListener( "load", completed ); - jQuery.ready(); -} - -// Catch cases where $(document).ready() is called -// after the browser event has already occurred. -// Support: IE <=9 - 10 only -// Older IE sometimes signals "interactive" too soon -if ( document.readyState === "complete" || - ( document.readyState !== "loading" && !document.documentElement.doScroll ) ) { - - // Handle it asynchronously to allow scripts the opportunity to delay ready - window.setTimeout( jQuery.ready ); - -} else { - - // Use the handy event callback - document.addEventListener( "DOMContentLoaded", completed ); - - // A fallback to window.onload, that will always work - window.addEventListener( "load", completed ); -} - - - - -// Multifunctional method to get and set values of a collection -// The value/s can optionally be executed if it's a function -var access = function( elems, fn, key, value, chainable, emptyGet, raw ) { - var i = 0, - len = elems.length, - bulk = key == null; - - // Sets many values - if ( toType( key ) === "object" ) { - chainable = true; - for ( i in key ) { - access( elems, fn, i, key[ i ], true, emptyGet, raw ); - } - - // Sets one value - } else if ( value !== undefined ) { - chainable = true; - - if ( !isFunction( value ) ) { - raw = true; - } - - if ( bulk ) { - - // Bulk operations run against the entire set - if ( raw ) { - fn.call( elems, value ); - fn = null; - - // ...except when executing function values - } else { - bulk = fn; - fn = function( elem, _key, value ) { - return bulk.call( jQuery( elem ), value ); - }; - } - } - - if ( fn ) { - for ( ; i < len; i++ ) { - fn( - elems[ i ], key, raw ? - value : - value.call( elems[ i ], i, fn( elems[ i ], key ) ) - ); - } - } - } - - if ( chainable ) { - return elems; - } - - // Gets - if ( bulk ) { - return fn.call( elems ); - } - - return len ? fn( elems[ 0 ], key ) : emptyGet; -}; - - -// Matches dashed string for camelizing -var rmsPrefix = /^-ms-/, - rdashAlpha = /-([a-z])/g; - -// Used by camelCase as callback to replace() -function fcamelCase( _all, letter ) { - return letter.toUpperCase(); -} - -// Convert dashed to camelCase; used by the css and data modules -// Support: IE <=9 - 11, Edge 12 - 15 -// Microsoft forgot to hump their vendor prefix (#9572) -function camelCase( string ) { - return string.replace( rmsPrefix, "ms-" ).replace( rdashAlpha, fcamelCase ); -} -var acceptData = function( owner ) { - - // Accepts only: - // - Node - // - Node.ELEMENT_NODE - // - Node.DOCUMENT_NODE - // - Object - // - Any - return owner.nodeType === 1 || owner.nodeType === 9 || !( +owner.nodeType ); -}; - - - - -function Data() { - this.expando = jQuery.expando + Data.uid++; -} - -Data.uid = 1; - -Data.prototype = { - - cache: function( owner ) { - - // Check if the owner object already has a cache - var value = owner[ this.expando ]; - - // If not, create one - if ( !value ) { - value = {}; - - // We can accept data for non-element nodes in modern browsers, - // but we should not, see #8335. - // Always return an empty object. - if ( acceptData( owner ) ) { - - // If it is a node unlikely to be stringify-ed or looped over - // use plain assignment - if ( owner.nodeType ) { - owner[ this.expando ] = value; - - // Otherwise secure it in a non-enumerable property - // configurable must be true to allow the property to be - // deleted when data is removed - } else { - Object.defineProperty( owner, this.expando, { - value: value, - configurable: true - } ); - } - } - } - - return value; - }, - set: function( owner, data, value ) { - var prop, - cache = this.cache( owner ); - - // Handle: [ owner, key, value ] args - // Always use camelCase key (gh-2257) - if ( typeof data === "string" ) { - cache[ camelCase( data ) ] = value; - - // Handle: [ owner, { properties } ] args - } else { - - // Copy the properties one-by-one to the cache object - for ( prop in data ) { - cache[ camelCase( prop ) ] = data[ prop ]; - } - } - return cache; - }, - get: function( owner, key ) { - return key === undefined ? - this.cache( owner ) : - - // Always use camelCase key (gh-2257) - owner[ this.expando ] && owner[ this.expando ][ camelCase( key ) ]; - }, - access: function( owner, key, value ) { - - // In cases where either: - // - // 1. No key was specified - // 2. A string key was specified, but no value provided - // - // Take the "read" path and allow the get method to determine - // which value to return, respectively either: - // - // 1. The entire cache object - // 2. The data stored at the key - // - if ( key === undefined || - ( ( key && typeof key === "string" ) && value === undefined ) ) { - - return this.get( owner, key ); - } - - // When the key is not a string, or both a key and value - // are specified, set or extend (existing objects) with either: - // - // 1. An object of properties - // 2. A key and value - // - this.set( owner, key, value ); - - // Since the "set" path can have two possible entry points - // return the expected data based on which path was taken[*] - return value !== undefined ? value : key; - }, - remove: function( owner, key ) { - var i, - cache = owner[ this.expando ]; - - if ( cache === undefined ) { - return; - } - - if ( key !== undefined ) { - - // Support array or space separated string of keys - if ( Array.isArray( key ) ) { - - // If key is an array of keys... - // We always set camelCase keys, so remove that. - key = key.map( camelCase ); - } else { - key = camelCase( key ); - - // If a key with the spaces exists, use it. - // Otherwise, create an array by matching non-whitespace - key = key in cache ? - [ key ] : - ( key.match( rnothtmlwhite ) || [] ); - } - - i = key.length; - - while ( i-- ) { - delete cache[ key[ i ] ]; - } - } - - // Remove the expando if there's no more data - if ( key === undefined || jQuery.isEmptyObject( cache ) ) { - - // Support: Chrome <=35 - 45 - // Webkit & Blink performance suffers when deleting properties - // from DOM nodes, so set to undefined instead - // https://bugs.chromium.org/p/chromium/issues/detail?id=378607 (bug restricted) - if ( owner.nodeType ) { - owner[ this.expando ] = undefined; - } else { - delete owner[ this.expando ]; - } - } - }, - hasData: function( owner ) { - var cache = owner[ this.expando ]; - return cache !== undefined && !jQuery.isEmptyObject( cache ); - } -}; -var dataPriv = new Data(); - -var dataUser = new Data(); - - - -// Implementation Summary -// -// 1. Enforce API surface and semantic compatibility with 1.9.x branch -// 2. Improve the module's maintainability by reducing the storage -// paths to a single mechanism. -// 3. Use the same single mechanism to support "private" and "user" data. -// 4. _Never_ expose "private" data to user code (TODO: Drop _data, _removeData) -// 5. Avoid exposing implementation details on user objects (eg. expando properties) -// 6. Provide a clear path for implementation upgrade to WeakMap in 2014 - -var rbrace = /^(?:\{[\w\W]*\}|\[[\w\W]*\])$/, - rmultiDash = /[A-Z]/g; - -function getData( data ) { - if ( data === "true" ) { - return true; - } - - if ( data === "false" ) { - return false; - } - - if ( data === "null" ) { - return null; - } - - // Only convert to a number if it doesn't change the string - if ( data === +data + "" ) { - return +data; - } - - if ( rbrace.test( data ) ) { - return JSON.parse( data ); - } - - return data; -} - -function dataAttr( elem, key, data ) { - var name; - - // If nothing was found internally, try to fetch any - // data from the HTML5 data-* attribute - if ( data === undefined && elem.nodeType === 1 ) { - name = "data-" + key.replace( rmultiDash, "-$&" ).toLowerCase(); - data = elem.getAttribute( name ); - - if ( typeof data === "string" ) { - try { - data = getData( data ); - } catch ( e ) {} - - // Make sure we set the data so it isn't changed later - dataUser.set( elem, key, data ); - } else { - data = undefined; - } - } - return data; -} - -jQuery.extend( { - hasData: function( elem ) { - return dataUser.hasData( elem ) || dataPriv.hasData( elem ); - }, - - data: function( elem, name, data ) { - return dataUser.access( elem, name, data ); - }, - - removeData: function( elem, name ) { - dataUser.remove( elem, name ); - }, - - // TODO: Now that all calls to _data and _removeData have been replaced - // with direct calls to dataPriv methods, these can be deprecated. - _data: function( elem, name, data ) { - return dataPriv.access( elem, name, data ); - }, - - _removeData: function( elem, name ) { - dataPriv.remove( elem, name ); - } -} ); - -jQuery.fn.extend( { - data: function( key, value ) { - var i, name, data, - elem = this[ 0 ], - attrs = elem && elem.attributes; - - // Gets all values - if ( key === undefined ) { - if ( this.length ) { - data = dataUser.get( elem ); - - if ( elem.nodeType === 1 && !dataPriv.get( elem, "hasDataAttrs" ) ) { - i = attrs.length; - while ( i-- ) { - - // Support: IE 11 only - // The attrs elements can be null (#14894) - if ( attrs[ i ] ) { - name = attrs[ i ].name; - if ( name.indexOf( "data-" ) === 0 ) { - name = camelCase( name.slice( 5 ) ); - dataAttr( elem, name, data[ name ] ); - } - } - } - dataPriv.set( elem, "hasDataAttrs", true ); - } - } - - return data; - } - - // Sets multiple values - if ( typeof key === "object" ) { - return this.each( function() { - dataUser.set( this, key ); - } ); - } - - return access( this, function( value ) { - var data; - - // The calling jQuery object (element matches) is not empty - // (and therefore has an element appears at this[ 0 ]) and the - // `value` parameter was not undefined. An empty jQuery object - // will result in `undefined` for elem = this[ 0 ] which will - // throw an exception if an attempt to read a data cache is made. - if ( elem && value === undefined ) { - - // Attempt to get data from the cache - // The key will always be camelCased in Data - data = dataUser.get( elem, key ); - if ( data !== undefined ) { - return data; - } - - // Attempt to "discover" the data in - // HTML5 custom data-* attrs - data = dataAttr( elem, key ); - if ( data !== undefined ) { - return data; - } - - // We tried really hard, but the data doesn't exist. - return; - } - - // Set the data... - this.each( function() { - - // We always store the camelCased key - dataUser.set( this, key, value ); - } ); - }, null, value, arguments.length > 1, null, true ); - }, - - removeData: function( key ) { - return this.each( function() { - dataUser.remove( this, key ); - } ); - } -} ); - - -jQuery.extend( { - queue: function( elem, type, data ) { - var queue; - - if ( elem ) { - type = ( type || "fx" ) + "queue"; - queue = dataPriv.get( elem, type ); - - // Speed up dequeue by getting out quickly if this is just a lookup - if ( data ) { - if ( !queue || Array.isArray( data ) ) { - queue = dataPriv.access( elem, type, jQuery.makeArray( data ) ); - } else { - queue.push( data ); - } - } - return queue || []; - } - }, - - dequeue: function( elem, type ) { - type = type || "fx"; - - var queue = jQuery.queue( elem, type ), - startLength = queue.length, - fn = queue.shift(), - hooks = jQuery._queueHooks( elem, type ), - next = function() { - jQuery.dequeue( elem, type ); - }; - - // If the fx queue is dequeued, always remove the progress sentinel - if ( fn === "inprogress" ) { - fn = queue.shift(); - startLength--; - } - - if ( fn ) { - - // Add a progress sentinel to prevent the fx queue from being - // automatically dequeued - if ( type === "fx" ) { - queue.unshift( "inprogress" ); - } - - // Clear up the last queue stop function - delete hooks.stop; - fn.call( elem, next, hooks ); - } - - if ( !startLength && hooks ) { - hooks.empty.fire(); - } - }, - - // Not public - generate a queueHooks object, or return the current one - _queueHooks: function( elem, type ) { - var key = type + "queueHooks"; - return dataPriv.get( elem, key ) || dataPriv.access( elem, key, { - empty: jQuery.Callbacks( "once memory" ).add( function() { - dataPriv.remove( elem, [ type + "queue", key ] ); - } ) - } ); - } -} ); - -jQuery.fn.extend( { - queue: function( type, data ) { - var setter = 2; - - if ( typeof type !== "string" ) { - data = type; - type = "fx"; - setter--; - } - - if ( arguments.length < setter ) { - return jQuery.queue( this[ 0 ], type ); - } - - return data === undefined ? - this : - this.each( function() { - var queue = jQuery.queue( this, type, data ); - - // Ensure a hooks for this queue - jQuery._queueHooks( this, type ); - - if ( type === "fx" && queue[ 0 ] !== "inprogress" ) { - jQuery.dequeue( this, type ); - } - } ); - }, - dequeue: function( type ) { - return this.each( function() { - jQuery.dequeue( this, type ); - } ); - }, - clearQueue: function( type ) { - return this.queue( type || "fx", [] ); - }, - - // Get a promise resolved when queues of a certain type - // are emptied (fx is the type by default) - promise: function( type, obj ) { - var tmp, - count = 1, - defer = jQuery.Deferred(), - elements = this, - i = this.length, - resolve = function() { - if ( !( --count ) ) { - defer.resolveWith( elements, [ elements ] ); - } - }; - - if ( typeof type !== "string" ) { - obj = type; - type = undefined; - } - type = type || "fx"; - - while ( i-- ) { - tmp = dataPriv.get( elements[ i ], type + "queueHooks" ); - if ( tmp && tmp.empty ) { - count++; - tmp.empty.add( resolve ); - } - } - resolve(); - return defer.promise( obj ); - } -} ); -var pnum = ( /[+-]?(?:\d*\.|)\d+(?:[eE][+-]?\d+|)/ ).source; - -var rcssNum = new RegExp( "^(?:([+-])=|)(" + pnum + ")([a-z%]*)$", "i" ); - - -var cssExpand = [ "Top", "Right", "Bottom", "Left" ]; - -var documentElement = document.documentElement; - - - - var isAttached = function( elem ) { - return jQuery.contains( elem.ownerDocument, elem ); - }, - composed = { composed: true }; - - // Support: IE 9 - 11+, Edge 12 - 18+, iOS 10.0 - 10.2 only - // Check attachment across shadow DOM boundaries when possible (gh-3504) - // Support: iOS 10.0-10.2 only - // Early iOS 10 versions support `attachShadow` but not `getRootNode`, - // leading to errors. We need to check for `getRootNode`. - if ( documentElement.getRootNode ) { - isAttached = function( elem ) { - return jQuery.contains( elem.ownerDocument, elem ) || - elem.getRootNode( composed ) === elem.ownerDocument; - }; - } -var isHiddenWithinTree = function( elem, el ) { - - // isHiddenWithinTree might be called from jQuery#filter function; - // in that case, element will be second argument - elem = el || elem; - - // Inline style trumps all - return elem.style.display === "none" || - elem.style.display === "" && - - // Otherwise, check computed style - // Support: Firefox <=43 - 45 - // Disconnected elements can have computed display: none, so first confirm that elem is - // in the document. - isAttached( elem ) && - - jQuery.css( elem, "display" ) === "none"; - }; - - - -function adjustCSS( elem, prop, valueParts, tween ) { - var adjusted, scale, - maxIterations = 20, - currentValue = tween ? - function() { - return tween.cur(); - } : - function() { - return jQuery.css( elem, prop, "" ); - }, - initial = currentValue(), - unit = valueParts && valueParts[ 3 ] || ( jQuery.cssNumber[ prop ] ? "" : "px" ), - - // Starting value computation is required for potential unit mismatches - initialInUnit = elem.nodeType && - ( jQuery.cssNumber[ prop ] || unit !== "px" && +initial ) && - rcssNum.exec( jQuery.css( elem, prop ) ); - - if ( initialInUnit && initialInUnit[ 3 ] !== unit ) { - - // Support: Firefox <=54 - // Halve the iteration target value to prevent interference from CSS upper bounds (gh-2144) - initial = initial / 2; - - // Trust units reported by jQuery.css - unit = unit || initialInUnit[ 3 ]; - - // Iteratively approximate from a nonzero starting point - initialInUnit = +initial || 1; - - while ( maxIterations-- ) { - - // Evaluate and update our best guess (doubling guesses that zero out). - // Finish if the scale equals or crosses 1 (making the old*new product non-positive). - jQuery.style( elem, prop, initialInUnit + unit ); - if ( ( 1 - scale ) * ( 1 - ( scale = currentValue() / initial || 0.5 ) ) <= 0 ) { - maxIterations = 0; - } - initialInUnit = initialInUnit / scale; - - } - - initialInUnit = initialInUnit * 2; - jQuery.style( elem, prop, initialInUnit + unit ); - - // Make sure we update the tween properties later on - valueParts = valueParts || []; - } - - if ( valueParts ) { - initialInUnit = +initialInUnit || +initial || 0; - - // Apply relative offset (+=/-=) if specified - adjusted = valueParts[ 1 ] ? - initialInUnit + ( valueParts[ 1 ] + 1 ) * valueParts[ 2 ] : - +valueParts[ 2 ]; - if ( tween ) { - tween.unit = unit; - tween.start = initialInUnit; - tween.end = adjusted; - } - } - return adjusted; -} - - -var defaultDisplayMap = {}; - -function getDefaultDisplay( elem ) { - var temp, - doc = elem.ownerDocument, - nodeName = elem.nodeName, - display = defaultDisplayMap[ nodeName ]; - - if ( display ) { - return display; - } - - temp = doc.body.appendChild( doc.createElement( nodeName ) ); - display = jQuery.css( temp, "display" ); - - temp.parentNode.removeChild( temp ); - - if ( display === "none" ) { - display = "block"; - } - defaultDisplayMap[ nodeName ] = display; - - return display; -} - -function showHide( elements, show ) { - var display, elem, - values = [], - index = 0, - length = elements.length; - - // Determine new display value for elements that need to change - for ( ; index < length; index++ ) { - elem = elements[ index ]; - if ( !elem.style ) { - continue; - } - - display = elem.style.display; - if ( show ) { - - // Since we force visibility upon cascade-hidden elements, an immediate (and slow) - // check is required in this first loop unless we have a nonempty display value (either - // inline or about-to-be-restored) - if ( display === "none" ) { - values[ index ] = dataPriv.get( elem, "display" ) || null; - if ( !values[ index ] ) { - elem.style.display = ""; - } - } - if ( elem.style.display === "" && isHiddenWithinTree( elem ) ) { - values[ index ] = getDefaultDisplay( elem ); - } - } else { - if ( display !== "none" ) { - values[ index ] = "none"; - - // Remember what we're overwriting - dataPriv.set( elem, "display", display ); - } - } - } - - // Set the display of the elements in a second loop to avoid constant reflow - for ( index = 0; index < length; index++ ) { - if ( values[ index ] != null ) { - elements[ index ].style.display = values[ index ]; - } - } - - return elements; -} - -jQuery.fn.extend( { - show: function() { - return showHide( this, true ); - }, - hide: function() { - return showHide( this ); - }, - toggle: function( state ) { - if ( typeof state === "boolean" ) { - return state ? this.show() : this.hide(); - } - - return this.each( function() { - if ( isHiddenWithinTree( this ) ) { - jQuery( this ).show(); - } else { - jQuery( this ).hide(); - } - } ); - } -} ); -var rcheckableType = ( /^(?:checkbox|radio)$/i ); - -var rtagName = ( /<([a-z][^\/\0>\x20\t\r\n\f]*)/i ); - -var rscriptType = ( /^$|^module$|\/(?:java|ecma)script/i ); - - - -( function() { - var fragment = document.createDocumentFragment(), - div = fragment.appendChild( document.createElement( "div" ) ), - input = document.createElement( "input" ); - - // Support: Android 4.0 - 4.3 only - // Check state lost if the name is set (#11217) - // Support: Windows Web Apps (WWA) - // `name` and `type` must use .setAttribute for WWA (#14901) - input.setAttribute( "type", "radio" ); - input.setAttribute( "checked", "checked" ); - input.setAttribute( "name", "t" ); - - div.appendChild( input ); - - // Support: Android <=4.1 only - // Older WebKit doesn't clone checked state correctly in fragments - support.checkClone = div.cloneNode( true ).cloneNode( true ).lastChild.checked; - - // Support: IE <=11 only - // Make sure textarea (and checkbox) defaultValue is properly cloned - div.innerHTML = ""; - support.noCloneChecked = !!div.cloneNode( true ).lastChild.defaultValue; - - // Support: IE <=9 only - // IE <=9 replaces "; - support.option = !!div.lastChild; -} )(); - - -// We have to close these tags to support XHTML (#13200) -var wrapMap = { - - // XHTML parsers do not magically insert elements in the - // same way that tag soup parsers do. So we cannot shorten - // this by omitting or other required elements. - thead: [ 1, "", "
    " ], - col: [ 2, "", "
    " ], - tr: [ 2, "", "
    " ], - td: [ 3, "", "
    " ], - - _default: [ 0, "", "" ] -}; - -wrapMap.tbody = wrapMap.tfoot = wrapMap.colgroup = wrapMap.caption = wrapMap.thead; -wrapMap.th = wrapMap.td; - -// Support: IE <=9 only -if ( !support.option ) { - wrapMap.optgroup = wrapMap.option = [ 1, "" ]; -} - - -function getAll( context, tag ) { - - // Support: IE <=9 - 11 only - // Use typeof to avoid zero-argument method invocation on host objects (#15151) - var ret; - - if ( typeof context.getElementsByTagName !== "undefined" ) { - ret = context.getElementsByTagName( tag || "*" ); - - } else if ( typeof context.querySelectorAll !== "undefined" ) { - ret = context.querySelectorAll( tag || "*" ); - - } else { - ret = []; - } - - if ( tag === undefined || tag && nodeName( context, tag ) ) { - return jQuery.merge( [ context ], ret ); - } - - return ret; -} - - -// Mark scripts as having already been evaluated -function setGlobalEval( elems, refElements ) { - var i = 0, - l = elems.length; - - for ( ; i < l; i++ ) { - dataPriv.set( - elems[ i ], - "globalEval", - !refElements || dataPriv.get( refElements[ i ], "globalEval" ) - ); - } -} - - -var rhtml = /<|&#?\w+;/; - -function buildFragment( elems, context, scripts, selection, ignored ) { - var elem, tmp, tag, wrap, attached, j, - fragment = context.createDocumentFragment(), - nodes = [], - i = 0, - l = elems.length; - - for ( ; i < l; i++ ) { - elem = elems[ i ]; - - if ( elem || elem === 0 ) { - - // Add nodes directly - if ( toType( elem ) === "object" ) { - - // Support: Android <=4.0 only, PhantomJS 1 only - // push.apply(_, arraylike) throws on ancient WebKit - jQuery.merge( nodes, elem.nodeType ? [ elem ] : elem ); - - // Convert non-html into a text node - } else if ( !rhtml.test( elem ) ) { - nodes.push( context.createTextNode( elem ) ); - - // Convert html into DOM nodes - } else { - tmp = tmp || fragment.appendChild( context.createElement( "div" ) ); - - // Deserialize a standard representation - tag = ( rtagName.exec( elem ) || [ "", "" ] )[ 1 ].toLowerCase(); - wrap = wrapMap[ tag ] || wrapMap._default; - tmp.innerHTML = wrap[ 1 ] + jQuery.htmlPrefilter( elem ) + wrap[ 2 ]; - - // Descend through wrappers to the right content - j = wrap[ 0 ]; - while ( j-- ) { - tmp = tmp.lastChild; - } - - // Support: Android <=4.0 only, PhantomJS 1 only - // push.apply(_, arraylike) throws on ancient WebKit - jQuery.merge( nodes, tmp.childNodes ); - - // Remember the top-level container - tmp = fragment.firstChild; - - // Ensure the created nodes are orphaned (#12392) - tmp.textContent = ""; - } - } - } - - // Remove wrapper from fragment - fragment.textContent = ""; - - i = 0; - while ( ( elem = nodes[ i++ ] ) ) { - - // Skip elements already in the context collection (trac-4087) - if ( selection && jQuery.inArray( elem, selection ) > -1 ) { - if ( ignored ) { - ignored.push( elem ); - } - continue; - } - - attached = isAttached( elem ); - - // Append to fragment - tmp = getAll( fragment.appendChild( elem ), "script" ); - - // Preserve script evaluation history - if ( attached ) { - setGlobalEval( tmp ); - } - - // Capture executables - if ( scripts ) { - j = 0; - while ( ( elem = tmp[ j++ ] ) ) { - if ( rscriptType.test( elem.type || "" ) ) { - scripts.push( elem ); - } - } - } - } - - return fragment; -} - - -var - rkeyEvent = /^key/, - rmouseEvent = /^(?:mouse|pointer|contextmenu|drag|drop)|click/, - rtypenamespace = /^([^.]*)(?:\.(.+)|)/; - -function returnTrue() { - return true; -} - -function returnFalse() { - return false; -} - -// Support: IE <=9 - 11+ -// focus() and blur() are asynchronous, except when they are no-op. -// So expect focus to be synchronous when the element is already active, -// and blur to be synchronous when the element is not already active. -// (focus and blur are always synchronous in other supported browsers, -// this just defines when we can count on it). -function expectSync( elem, type ) { - return ( elem === safeActiveElement() ) === ( type === "focus" ); -} - -// Support: IE <=9 only -// Accessing document.activeElement can throw unexpectedly -// https://bugs.jquery.com/ticket/13393 -function safeActiveElement() { - try { - return document.activeElement; - } catch ( err ) { } -} - -function on( elem, types, selector, data, fn, one ) { - var origFn, type; - - // Types can be a map of types/handlers - if ( typeof types === "object" ) { - - // ( types-Object, selector, data ) - if ( typeof selector !== "string" ) { - - // ( types-Object, data ) - data = data || selector; - selector = undefined; - } - for ( type in types ) { - on( elem, type, selector, data, types[ type ], one ); - } - return elem; - } - - if ( data == null && fn == null ) { - - // ( types, fn ) - fn = selector; - data = selector = undefined; - } else if ( fn == null ) { - if ( typeof selector === "string" ) { - - // ( types, selector, fn ) - fn = data; - data = undefined; - } else { - - // ( types, data, fn ) - fn = data; - data = selector; - selector = undefined; - } - } - if ( fn === false ) { - fn = returnFalse; - } else if ( !fn ) { - return elem; - } - - if ( one === 1 ) { - origFn = fn; - fn = function( event ) { - - // Can use an empty set, since event contains the info - jQuery().off( event ); - return origFn.apply( this, arguments ); - }; - - // Use same guid so caller can remove using origFn - fn.guid = origFn.guid || ( origFn.guid = jQuery.guid++ ); - } - return elem.each( function() { - jQuery.event.add( this, types, fn, data, selector ); - } ); -} - -/* - * Helper functions for managing events -- not part of the public interface. - * Props to Dean Edwards' addEvent library for many of the ideas. - */ -jQuery.event = { - - global: {}, - - add: function( elem, types, handler, data, selector ) { - - var handleObjIn, eventHandle, tmp, - events, t, handleObj, - special, handlers, type, namespaces, origType, - elemData = dataPriv.get( elem ); - - // Only attach events to objects that accept data - if ( !acceptData( elem ) ) { - return; - } - - // Caller can pass in an object of custom data in lieu of the handler - if ( handler.handler ) { - handleObjIn = handler; - handler = handleObjIn.handler; - selector = handleObjIn.selector; - } - - // Ensure that invalid selectors throw exceptions at attach time - // Evaluate against documentElement in case elem is a non-element node (e.g., document) - if ( selector ) { - jQuery.find.matchesSelector( documentElement, selector ); - } - - // Make sure that the handler has a unique ID, used to find/remove it later - if ( !handler.guid ) { - handler.guid = jQuery.guid++; - } - - // Init the element's event structure and main handler, if this is the first - if ( !( events = elemData.events ) ) { - events = elemData.events = Object.create( null ); - } - if ( !( eventHandle = elemData.handle ) ) { - eventHandle = elemData.handle = function( e ) { - - // Discard the second event of a jQuery.event.trigger() and - // when an event is called after a page has unloaded - return typeof jQuery !== "undefined" && jQuery.event.triggered !== e.type ? - jQuery.event.dispatch.apply( elem, arguments ) : undefined; - }; - } - - // Handle multiple events separated by a space - types = ( types || "" ).match( rnothtmlwhite ) || [ "" ]; - t = types.length; - while ( t-- ) { - tmp = rtypenamespace.exec( types[ t ] ) || []; - type = origType = tmp[ 1 ]; - namespaces = ( tmp[ 2 ] || "" ).split( "." ).sort(); - - // There *must* be a type, no attaching namespace-only handlers - if ( !type ) { - continue; - } - - // If event changes its type, use the special event handlers for the changed type - special = jQuery.event.special[ type ] || {}; - - // If selector defined, determine special event api type, otherwise given type - type = ( selector ? special.delegateType : special.bindType ) || type; - - // Update special based on newly reset type - special = jQuery.event.special[ type ] || {}; - - // handleObj is passed to all event handlers - handleObj = jQuery.extend( { - type: type, - origType: origType, - data: data, - handler: handler, - guid: handler.guid, - selector: selector, - needsContext: selector && jQuery.expr.match.needsContext.test( selector ), - namespace: namespaces.join( "." ) - }, handleObjIn ); - - // Init the event handler queue if we're the first - if ( !( handlers = events[ type ] ) ) { - handlers = events[ type ] = []; - handlers.delegateCount = 0; - - // Only use addEventListener if the special events handler returns false - if ( !special.setup || - special.setup.call( elem, data, namespaces, eventHandle ) === false ) { - - if ( elem.addEventListener ) { - elem.addEventListener( type, eventHandle ); - } - } - } - - if ( special.add ) { - special.add.call( elem, handleObj ); - - if ( !handleObj.handler.guid ) { - handleObj.handler.guid = handler.guid; - } - } - - // Add to the element's handler list, delegates in front - if ( selector ) { - handlers.splice( handlers.delegateCount++, 0, handleObj ); - } else { - handlers.push( handleObj ); - } - - // Keep track of which events have ever been used, for event optimization - jQuery.event.global[ type ] = true; - } - - }, - - // Detach an event or set of events from an element - remove: function( elem, types, handler, selector, mappedTypes ) { - - var j, origCount, tmp, - events, t, handleObj, - special, handlers, type, namespaces, origType, - elemData = dataPriv.hasData( elem ) && dataPriv.get( elem ); - - if ( !elemData || !( events = elemData.events ) ) { - return; - } - - // Once for each type.namespace in types; type may be omitted - types = ( types || "" ).match( rnothtmlwhite ) || [ "" ]; - t = types.length; - while ( t-- ) { - tmp = rtypenamespace.exec( types[ t ] ) || []; - type = origType = tmp[ 1 ]; - namespaces = ( tmp[ 2 ] || "" ).split( "." ).sort(); - - // Unbind all events (on this namespace, if provided) for the element - if ( !type ) { - for ( type in events ) { - jQuery.event.remove( elem, type + types[ t ], handler, selector, true ); - } - continue; - } - - special = jQuery.event.special[ type ] || {}; - type = ( selector ? special.delegateType : special.bindType ) || type; - handlers = events[ type ] || []; - tmp = tmp[ 2 ] && - new RegExp( "(^|\\.)" + namespaces.join( "\\.(?:.*\\.|)" ) + "(\\.|$)" ); - - // Remove matching events - origCount = j = handlers.length; - while ( j-- ) { - handleObj = handlers[ j ]; - - if ( ( mappedTypes || origType === handleObj.origType ) && - ( !handler || handler.guid === handleObj.guid ) && - ( !tmp || tmp.test( handleObj.namespace ) ) && - ( !selector || selector === handleObj.selector || - selector === "**" && handleObj.selector ) ) { - handlers.splice( j, 1 ); - - if ( handleObj.selector ) { - handlers.delegateCount--; - } - if ( special.remove ) { - special.remove.call( elem, handleObj ); - } - } - } - - // Remove generic event handler if we removed something and no more handlers exist - // (avoids potential for endless recursion during removal of special event handlers) - if ( origCount && !handlers.length ) { - if ( !special.teardown || - special.teardown.call( elem, namespaces, elemData.handle ) === false ) { - - jQuery.removeEvent( elem, type, elemData.handle ); - } - - delete events[ type ]; - } - } - - // Remove data and the expando if it's no longer used - if ( jQuery.isEmptyObject( events ) ) { - dataPriv.remove( elem, "handle events" ); - } - }, - - dispatch: function( nativeEvent ) { - - var i, j, ret, matched, handleObj, handlerQueue, - args = new Array( arguments.length ), - - // Make a writable jQuery.Event from the native event object - event = jQuery.event.fix( nativeEvent ), - - handlers = ( - dataPriv.get( this, "events" ) || Object.create( null ) - )[ event.type ] || [], - special = jQuery.event.special[ event.type ] || {}; - - // Use the fix-ed jQuery.Event rather than the (read-only) native event - args[ 0 ] = event; - - for ( i = 1; i < arguments.length; i++ ) { - args[ i ] = arguments[ i ]; - } - - event.delegateTarget = this; - - // Call the preDispatch hook for the mapped type, and let it bail if desired - if ( special.preDispatch && special.preDispatch.call( this, event ) === false ) { - return; - } - - // Determine handlers - handlerQueue = jQuery.event.handlers.call( this, event, handlers ); - - // Run delegates first; they may want to stop propagation beneath us - i = 0; - while ( ( matched = handlerQueue[ i++ ] ) && !event.isPropagationStopped() ) { - event.currentTarget = matched.elem; - - j = 0; - while ( ( handleObj = matched.handlers[ j++ ] ) && - !event.isImmediatePropagationStopped() ) { - - // If the event is namespaced, then each handler is only invoked if it is - // specially universal or its namespaces are a superset of the event's. - if ( !event.rnamespace || handleObj.namespace === false || - event.rnamespace.test( handleObj.namespace ) ) { - - event.handleObj = handleObj; - event.data = handleObj.data; - - ret = ( ( jQuery.event.special[ handleObj.origType ] || {} ).handle || - handleObj.handler ).apply( matched.elem, args ); - - if ( ret !== undefined ) { - if ( ( event.result = ret ) === false ) { - event.preventDefault(); - event.stopPropagation(); - } - } - } - } - } - - // Call the postDispatch hook for the mapped type - if ( special.postDispatch ) { - special.postDispatch.call( this, event ); - } - - return event.result; - }, - - handlers: function( event, handlers ) { - var i, handleObj, sel, matchedHandlers, matchedSelectors, - handlerQueue = [], - delegateCount = handlers.delegateCount, - cur = event.target; - - // Find delegate handlers - if ( delegateCount && - - // Support: IE <=9 - // Black-hole SVG instance trees (trac-13180) - cur.nodeType && - - // Support: Firefox <=42 - // Suppress spec-violating clicks indicating a non-primary pointer button (trac-3861) - // https://www.w3.org/TR/DOM-Level-3-Events/#event-type-click - // Support: IE 11 only - // ...but not arrow key "clicks" of radio inputs, which can have `button` -1 (gh-2343) - !( event.type === "click" && event.button >= 1 ) ) { - - for ( ; cur !== this; cur = cur.parentNode || this ) { - - // Don't check non-elements (#13208) - // Don't process clicks on disabled elements (#6911, #8165, #11382, #11764) - if ( cur.nodeType === 1 && !( event.type === "click" && cur.disabled === true ) ) { - matchedHandlers = []; - matchedSelectors = {}; - for ( i = 0; i < delegateCount; i++ ) { - handleObj = handlers[ i ]; - - // Don't conflict with Object.prototype properties (#13203) - sel = handleObj.selector + " "; - - if ( matchedSelectors[ sel ] === undefined ) { - matchedSelectors[ sel ] = handleObj.needsContext ? - jQuery( sel, this ).index( cur ) > -1 : - jQuery.find( sel, this, null, [ cur ] ).length; - } - if ( matchedSelectors[ sel ] ) { - matchedHandlers.push( handleObj ); - } - } - if ( matchedHandlers.length ) { - handlerQueue.push( { elem: cur, handlers: matchedHandlers } ); - } - } - } - } - - // Add the remaining (directly-bound) handlers - cur = this; - if ( delegateCount < handlers.length ) { - handlerQueue.push( { elem: cur, handlers: handlers.slice( delegateCount ) } ); - } - - return handlerQueue; - }, - - addProp: function( name, hook ) { - Object.defineProperty( jQuery.Event.prototype, name, { - enumerable: true, - configurable: true, - - get: isFunction( hook ) ? - function() { - if ( this.originalEvent ) { - return hook( this.originalEvent ); - } - } : - function() { - if ( this.originalEvent ) { - return this.originalEvent[ name ]; - } - }, - - set: function( value ) { - Object.defineProperty( this, name, { - enumerable: true, - configurable: true, - writable: true, - value: value - } ); - } - } ); - }, - - fix: function( originalEvent ) { - return originalEvent[ jQuery.expando ] ? - originalEvent : - new jQuery.Event( originalEvent ); - }, - - special: { - load: { - - // Prevent triggered image.load events from bubbling to window.load - noBubble: true - }, - click: { - - // Utilize native event to ensure correct state for checkable inputs - setup: function( data ) { - - // For mutual compressibility with _default, replace `this` access with a local var. - // `|| data` is dead code meant only to preserve the variable through minification. - var el = this || data; - - // Claim the first handler - if ( rcheckableType.test( el.type ) && - el.click && nodeName( el, "input" ) ) { - - // dataPriv.set( el, "click", ... ) - leverageNative( el, "click", returnTrue ); - } - - // Return false to allow normal processing in the caller - return false; - }, - trigger: function( data ) { - - // For mutual compressibility with _default, replace `this` access with a local var. - // `|| data` is dead code meant only to preserve the variable through minification. - var el = this || data; - - // Force setup before triggering a click - if ( rcheckableType.test( el.type ) && - el.click && nodeName( el, "input" ) ) { - - leverageNative( el, "click" ); - } - - // Return non-false to allow normal event-path propagation - return true; - }, - - // For cross-browser consistency, suppress native .click() on links - // Also prevent it if we're currently inside a leveraged native-event stack - _default: function( event ) { - var target = event.target; - return rcheckableType.test( target.type ) && - target.click && nodeName( target, "input" ) && - dataPriv.get( target, "click" ) || - nodeName( target, "a" ); - } - }, - - beforeunload: { - postDispatch: function( event ) { - - // Support: Firefox 20+ - // Firefox doesn't alert if the returnValue field is not set. - if ( event.result !== undefined && event.originalEvent ) { - event.originalEvent.returnValue = event.result; - } - } - } - } -}; - -// Ensure the presence of an event listener that handles manually-triggered -// synthetic events by interrupting progress until reinvoked in response to -// *native* events that it fires directly, ensuring that state changes have -// already occurred before other listeners are invoked. -function leverageNative( el, type, expectSync ) { - - // Missing expectSync indicates a trigger call, which must force setup through jQuery.event.add - if ( !expectSync ) { - if ( dataPriv.get( el, type ) === undefined ) { - jQuery.event.add( el, type, returnTrue ); - } - return; - } - - // Register the controller as a special universal handler for all event namespaces - dataPriv.set( el, type, false ); - jQuery.event.add( el, type, { - namespace: false, - handler: function( event ) { - var notAsync, result, - saved = dataPriv.get( this, type ); - - if ( ( event.isTrigger & 1 ) && this[ type ] ) { - - // Interrupt processing of the outer synthetic .trigger()ed event - // Saved data should be false in such cases, but might be a leftover capture object - // from an async native handler (gh-4350) - if ( !saved.length ) { - - // Store arguments for use when handling the inner native event - // There will always be at least one argument (an event object), so this array - // will not be confused with a leftover capture object. - saved = slice.call( arguments ); - dataPriv.set( this, type, saved ); - - // Trigger the native event and capture its result - // Support: IE <=9 - 11+ - // focus() and blur() are asynchronous - notAsync = expectSync( this, type ); - this[ type ](); - result = dataPriv.get( this, type ); - if ( saved !== result || notAsync ) { - dataPriv.set( this, type, false ); - } else { - result = {}; - } - if ( saved !== result ) { - - // Cancel the outer synthetic event - event.stopImmediatePropagation(); - event.preventDefault(); - return result.value; - } - - // If this is an inner synthetic event for an event with a bubbling surrogate - // (focus or blur), assume that the surrogate already propagated from triggering the - // native event and prevent that from happening again here. - // This technically gets the ordering wrong w.r.t. to `.trigger()` (in which the - // bubbling surrogate propagates *after* the non-bubbling base), but that seems - // less bad than duplication. - } else if ( ( jQuery.event.special[ type ] || {} ).delegateType ) { - event.stopPropagation(); - } - - // If this is a native event triggered above, everything is now in order - // Fire an inner synthetic event with the original arguments - } else if ( saved.length ) { - - // ...and capture the result - dataPriv.set( this, type, { - value: jQuery.event.trigger( - - // Support: IE <=9 - 11+ - // Extend with the prototype to reset the above stopImmediatePropagation() - jQuery.extend( saved[ 0 ], jQuery.Event.prototype ), - saved.slice( 1 ), - this - ) - } ); - - // Abort handling of the native event - event.stopImmediatePropagation(); - } - } - } ); -} - -jQuery.removeEvent = function( elem, type, handle ) { - - // This "if" is needed for plain objects - if ( elem.removeEventListener ) { - elem.removeEventListener( type, handle ); - } -}; - -jQuery.Event = function( src, props ) { - - // Allow instantiation without the 'new' keyword - if ( !( this instanceof jQuery.Event ) ) { - return new jQuery.Event( src, props ); - } - - // Event object - if ( src && src.type ) { - this.originalEvent = src; - this.type = src.type; - - // Events bubbling up the document may have been marked as prevented - // by a handler lower down the tree; reflect the correct value. - this.isDefaultPrevented = src.defaultPrevented || - src.defaultPrevented === undefined && - - // Support: Android <=2.3 only - src.returnValue === false ? - returnTrue : - returnFalse; - - // Create target properties - // Support: Safari <=6 - 7 only - // Target should not be a text node (#504, #13143) - this.target = ( src.target && src.target.nodeType === 3 ) ? - src.target.parentNode : - src.target; - - this.currentTarget = src.currentTarget; - this.relatedTarget = src.relatedTarget; - - // Event type - } else { - this.type = src; - } - - // Put explicitly provided properties onto the event object - if ( props ) { - jQuery.extend( this, props ); - } - - // Create a timestamp if incoming event doesn't have one - this.timeStamp = src && src.timeStamp || Date.now(); - - // Mark it as fixed - this[ jQuery.expando ] = true; -}; - -// jQuery.Event is based on DOM3 Events as specified by the ECMAScript Language Binding -// https://www.w3.org/TR/2003/WD-DOM-Level-3-Events-20030331/ecma-script-binding.html -jQuery.Event.prototype = { - constructor: jQuery.Event, - isDefaultPrevented: returnFalse, - isPropagationStopped: returnFalse, - isImmediatePropagationStopped: returnFalse, - isSimulated: false, - - preventDefault: function() { - var e = this.originalEvent; - - this.isDefaultPrevented = returnTrue; - - if ( e && !this.isSimulated ) { - e.preventDefault(); - } - }, - stopPropagation: function() { - var e = this.originalEvent; - - this.isPropagationStopped = returnTrue; - - if ( e && !this.isSimulated ) { - e.stopPropagation(); - } - }, - stopImmediatePropagation: function() { - var e = this.originalEvent; - - this.isImmediatePropagationStopped = returnTrue; - - if ( e && !this.isSimulated ) { - e.stopImmediatePropagation(); - } - - this.stopPropagation(); - } -}; - -// Includes all common event props including KeyEvent and MouseEvent specific props -jQuery.each( { - altKey: true, - bubbles: true, - cancelable: true, - changedTouches: true, - ctrlKey: true, - detail: true, - eventPhase: true, - metaKey: true, - pageX: true, - pageY: true, - shiftKey: true, - view: true, - "char": true, - code: true, - charCode: true, - key: true, - keyCode: true, - button: true, - buttons: true, - clientX: true, - clientY: true, - offsetX: true, - offsetY: true, - pointerId: true, - pointerType: true, - screenX: true, - screenY: true, - targetTouches: true, - toElement: true, - touches: true, - - which: function( event ) { - var button = event.button; - - // Add which for key events - if ( event.which == null && rkeyEvent.test( event.type ) ) { - return event.charCode != null ? event.charCode : event.keyCode; - } - - // Add which for click: 1 === left; 2 === middle; 3 === right - if ( !event.which && button !== undefined && rmouseEvent.test( event.type ) ) { - if ( button & 1 ) { - return 1; - } - - if ( button & 2 ) { - return 3; - } - - if ( button & 4 ) { - return 2; - } - - return 0; - } - - return event.which; - } -}, jQuery.event.addProp ); - -jQuery.each( { focus: "focusin", blur: "focusout" }, function( type, delegateType ) { - jQuery.event.special[ type ] = { - - // Utilize native event if possible so blur/focus sequence is correct - setup: function() { - - // Claim the first handler - // dataPriv.set( this, "focus", ... ) - // dataPriv.set( this, "blur", ... ) - leverageNative( this, type, expectSync ); - - // Return false to allow normal processing in the caller - return false; - }, - trigger: function() { - - // Force setup before trigger - leverageNative( this, type ); - - // Return non-false to allow normal event-path propagation - return true; - }, - - delegateType: delegateType - }; -} ); - -// Create mouseenter/leave events using mouseover/out and event-time checks -// so that event delegation works in jQuery. -// Do the same for pointerenter/pointerleave and pointerover/pointerout -// -// Support: Safari 7 only -// Safari sends mouseenter too often; see: -// https://bugs.chromium.org/p/chromium/issues/detail?id=470258 -// for the description of the bug (it existed in older Chrome versions as well). -jQuery.each( { - mouseenter: "mouseover", - mouseleave: "mouseout", - pointerenter: "pointerover", - pointerleave: "pointerout" -}, function( orig, fix ) { - jQuery.event.special[ orig ] = { - delegateType: fix, - bindType: fix, - - handle: function( event ) { - var ret, - target = this, - related = event.relatedTarget, - handleObj = event.handleObj; - - // For mouseenter/leave call the handler if related is outside the target. - // NB: No relatedTarget if the mouse left/entered the browser window - if ( !related || ( related !== target && !jQuery.contains( target, related ) ) ) { - event.type = handleObj.origType; - ret = handleObj.handler.apply( this, arguments ); - event.type = fix; - } - return ret; - } - }; -} ); - -jQuery.fn.extend( { - - on: function( types, selector, data, fn ) { - return on( this, types, selector, data, fn ); - }, - one: function( types, selector, data, fn ) { - return on( this, types, selector, data, fn, 1 ); - }, - off: function( types, selector, fn ) { - var handleObj, type; - if ( types && types.preventDefault && types.handleObj ) { - - // ( event ) dispatched jQuery.Event - handleObj = types.handleObj; - jQuery( types.delegateTarget ).off( - handleObj.namespace ? - handleObj.origType + "." + handleObj.namespace : - handleObj.origType, - handleObj.selector, - handleObj.handler - ); - return this; - } - if ( typeof types === "object" ) { - - // ( types-object [, selector] ) - for ( type in types ) { - this.off( type, selector, types[ type ] ); - } - return this; - } - if ( selector === false || typeof selector === "function" ) { - - // ( types [, fn] ) - fn = selector; - selector = undefined; - } - if ( fn === false ) { - fn = returnFalse; - } - return this.each( function() { - jQuery.event.remove( this, types, fn, selector ); - } ); - } -} ); - - -var - - // Support: IE <=10 - 11, Edge 12 - 13 only - // In IE/Edge using regex groups here causes severe slowdowns. - // See https://connect.microsoft.com/IE/feedback/details/1736512/ - rnoInnerhtml = /\s*$/g; - -// Prefer a tbody over its parent table for containing new rows -function manipulationTarget( elem, content ) { - if ( nodeName( elem, "table" ) && - nodeName( content.nodeType !== 11 ? content : content.firstChild, "tr" ) ) { - - return jQuery( elem ).children( "tbody" )[ 0 ] || elem; - } - - return elem; -} - -// Replace/restore the type attribute of script elements for safe DOM manipulation -function disableScript( elem ) { - elem.type = ( elem.getAttribute( "type" ) !== null ) + "/" + elem.type; - return elem; -} -function restoreScript( elem ) { - if ( ( elem.type || "" ).slice( 0, 5 ) === "true/" ) { - elem.type = elem.type.slice( 5 ); - } else { - elem.removeAttribute( "type" ); - } - - return elem; -} - -function cloneCopyEvent( src, dest ) { - var i, l, type, pdataOld, udataOld, udataCur, events; - - if ( dest.nodeType !== 1 ) { - return; - } - - // 1. Copy private data: events, handlers, etc. - if ( dataPriv.hasData( src ) ) { - pdataOld = dataPriv.get( src ); - events = pdataOld.events; - - if ( events ) { - dataPriv.remove( dest, "handle events" ); - - for ( type in events ) { - for ( i = 0, l = events[ type ].length; i < l; i++ ) { - jQuery.event.add( dest, type, events[ type ][ i ] ); - } - } - } - } - - // 2. Copy user data - if ( dataUser.hasData( src ) ) { - udataOld = dataUser.access( src ); - udataCur = jQuery.extend( {}, udataOld ); - - dataUser.set( dest, udataCur ); - } -} - -// Fix IE bugs, see support tests -function fixInput( src, dest ) { - var nodeName = dest.nodeName.toLowerCase(); - - // Fails to persist the checked state of a cloned checkbox or radio button. - if ( nodeName === "input" && rcheckableType.test( src.type ) ) { - dest.checked = src.checked; - - // Fails to return the selected option to the default selected state when cloning options - } else if ( nodeName === "input" || nodeName === "textarea" ) { - dest.defaultValue = src.defaultValue; - } -} - -function domManip( collection, args, callback, ignored ) { - - // Flatten any nested arrays - args = flat( args ); - - var fragment, first, scripts, hasScripts, node, doc, - i = 0, - l = collection.length, - iNoClone = l - 1, - value = args[ 0 ], - valueIsFunction = isFunction( value ); - - // We can't cloneNode fragments that contain checked, in WebKit - if ( valueIsFunction || - ( l > 1 && typeof value === "string" && - !support.checkClone && rchecked.test( value ) ) ) { - return collection.each( function( index ) { - var self = collection.eq( index ); - if ( valueIsFunction ) { - args[ 0 ] = value.call( this, index, self.html() ); - } - domManip( self, args, callback, ignored ); - } ); - } - - if ( l ) { - fragment = buildFragment( args, collection[ 0 ].ownerDocument, false, collection, ignored ); - first = fragment.firstChild; - - if ( fragment.childNodes.length === 1 ) { - fragment = first; - } - - // Require either new content or an interest in ignored elements to invoke the callback - if ( first || ignored ) { - scripts = jQuery.map( getAll( fragment, "script" ), disableScript ); - hasScripts = scripts.length; - - // Use the original fragment for the last item - // instead of the first because it can end up - // being emptied incorrectly in certain situations (#8070). - for ( ; i < l; i++ ) { - node = fragment; - - if ( i !== iNoClone ) { - node = jQuery.clone( node, true, true ); - - // Keep references to cloned scripts for later restoration - if ( hasScripts ) { - - // Support: Android <=4.0 only, PhantomJS 1 only - // push.apply(_, arraylike) throws on ancient WebKit - jQuery.merge( scripts, getAll( node, "script" ) ); - } - } - - callback.call( collection[ i ], node, i ); - } - - if ( hasScripts ) { - doc = scripts[ scripts.length - 1 ].ownerDocument; - - // Reenable scripts - jQuery.map( scripts, restoreScript ); - - // Evaluate executable scripts on first document insertion - for ( i = 0; i < hasScripts; i++ ) { - node = scripts[ i ]; - if ( rscriptType.test( node.type || "" ) && - !dataPriv.access( node, "globalEval" ) && - jQuery.contains( doc, node ) ) { - - if ( node.src && ( node.type || "" ).toLowerCase() !== "module" ) { - - // Optional AJAX dependency, but won't run scripts if not present - if ( jQuery._evalUrl && !node.noModule ) { - jQuery._evalUrl( node.src, { - nonce: node.nonce || node.getAttribute( "nonce" ) - }, doc ); - } - } else { - DOMEval( node.textContent.replace( rcleanScript, "" ), node, doc ); - } - } - } - } - } - } - - return collection; -} - -function remove( elem, selector, keepData ) { - var node, - nodes = selector ? jQuery.filter( selector, elem ) : elem, - i = 0; - - for ( ; ( node = nodes[ i ] ) != null; i++ ) { - if ( !keepData && node.nodeType === 1 ) { - jQuery.cleanData( getAll( node ) ); - } - - if ( node.parentNode ) { - if ( keepData && isAttached( node ) ) { - setGlobalEval( getAll( node, "script" ) ); - } - node.parentNode.removeChild( node ); - } - } - - return elem; -} - -jQuery.extend( { - htmlPrefilter: function( html ) { - return html; - }, - - clone: function( elem, dataAndEvents, deepDataAndEvents ) { - var i, l, srcElements, destElements, - clone = elem.cloneNode( true ), - inPage = isAttached( elem ); - - // Fix IE cloning issues - if ( !support.noCloneChecked && ( elem.nodeType === 1 || elem.nodeType === 11 ) && - !jQuery.isXMLDoc( elem ) ) { - - // We eschew Sizzle here for performance reasons: https://jsperf.com/getall-vs-sizzle/2 - destElements = getAll( clone ); - srcElements = getAll( elem ); - - for ( i = 0, l = srcElements.length; i < l; i++ ) { - fixInput( srcElements[ i ], destElements[ i ] ); - } - } - - // Copy the events from the original to the clone - if ( dataAndEvents ) { - if ( deepDataAndEvents ) { - srcElements = srcElements || getAll( elem ); - destElements = destElements || getAll( clone ); - - for ( i = 0, l = srcElements.length; i < l; i++ ) { - cloneCopyEvent( srcElements[ i ], destElements[ i ] ); - } - } else { - cloneCopyEvent( elem, clone ); - } - } - - // Preserve script evaluation history - destElements = getAll( clone, "script" ); - if ( destElements.length > 0 ) { - setGlobalEval( destElements, !inPage && getAll( elem, "script" ) ); - } - - // Return the cloned set - return clone; - }, - - cleanData: function( elems ) { - var data, elem, type, - special = jQuery.event.special, - i = 0; - - for ( ; ( elem = elems[ i ] ) !== undefined; i++ ) { - if ( acceptData( elem ) ) { - if ( ( data = elem[ dataPriv.expando ] ) ) { - if ( data.events ) { - for ( type in data.events ) { - if ( special[ type ] ) { - jQuery.event.remove( elem, type ); - - // This is a shortcut to avoid jQuery.event.remove's overhead - } else { - jQuery.removeEvent( elem, type, data.handle ); - } - } - } - - // Support: Chrome <=35 - 45+ - // Assign undefined instead of using delete, see Data#remove - elem[ dataPriv.expando ] = undefined; - } - if ( elem[ dataUser.expando ] ) { - - // Support: Chrome <=35 - 45+ - // Assign undefined instead of using delete, see Data#remove - elem[ dataUser.expando ] = undefined; - } - } - } - } -} ); - -jQuery.fn.extend( { - detach: function( selector ) { - return remove( this, selector, true ); - }, - - remove: function( selector ) { - return remove( this, selector ); - }, - - text: function( value ) { - return access( this, function( value ) { - return value === undefined ? - jQuery.text( this ) : - this.empty().each( function() { - if ( this.nodeType === 1 || this.nodeType === 11 || this.nodeType === 9 ) { - this.textContent = value; - } - } ); - }, null, value, arguments.length ); - }, - - append: function() { - return domManip( this, arguments, function( elem ) { - if ( this.nodeType === 1 || this.nodeType === 11 || this.nodeType === 9 ) { - var target = manipulationTarget( this, elem ); - target.appendChild( elem ); - } - } ); - }, - - prepend: function() { - return domManip( this, arguments, function( elem ) { - if ( this.nodeType === 1 || this.nodeType === 11 || this.nodeType === 9 ) { - var target = manipulationTarget( this, elem ); - target.insertBefore( elem, target.firstChild ); - } - } ); - }, - - before: function() { - return domManip( this, arguments, function( elem ) { - if ( this.parentNode ) { - this.parentNode.insertBefore( elem, this ); - } - } ); - }, - - after: function() { - return domManip( this, arguments, function( elem ) { - if ( this.parentNode ) { - this.parentNode.insertBefore( elem, this.nextSibling ); - } - } ); - }, - - empty: function() { - var elem, - i = 0; - - for ( ; ( elem = this[ i ] ) != null; i++ ) { - if ( elem.nodeType === 1 ) { - - // Prevent memory leaks - jQuery.cleanData( getAll( elem, false ) ); - - // Remove any remaining nodes - elem.textContent = ""; - } - } - - return this; - }, - - clone: function( dataAndEvents, deepDataAndEvents ) { - dataAndEvents = dataAndEvents == null ? false : dataAndEvents; - deepDataAndEvents = deepDataAndEvents == null ? dataAndEvents : deepDataAndEvents; - - return this.map( function() { - return jQuery.clone( this, dataAndEvents, deepDataAndEvents ); - } ); - }, - - html: function( value ) { - return access( this, function( value ) { - var elem = this[ 0 ] || {}, - i = 0, - l = this.length; - - if ( value === undefined && elem.nodeType === 1 ) { - return elem.innerHTML; - } - - // See if we can take a shortcut and just use innerHTML - if ( typeof value === "string" && !rnoInnerhtml.test( value ) && - !wrapMap[ ( rtagName.exec( value ) || [ "", "" ] )[ 1 ].toLowerCase() ] ) { - - value = jQuery.htmlPrefilter( value ); - - try { - for ( ; i < l; i++ ) { - elem = this[ i ] || {}; - - // Remove element nodes and prevent memory leaks - if ( elem.nodeType === 1 ) { - jQuery.cleanData( getAll( elem, false ) ); - elem.innerHTML = value; - } - } - - elem = 0; - - // If using innerHTML throws an exception, use the fallback method - } catch ( e ) {} - } - - if ( elem ) { - this.empty().append( value ); - } - }, null, value, arguments.length ); - }, - - replaceWith: function() { - var ignored = []; - - // Make the changes, replacing each non-ignored context element with the new content - return domManip( this, arguments, function( elem ) { - var parent = this.parentNode; - - if ( jQuery.inArray( this, ignored ) < 0 ) { - jQuery.cleanData( getAll( this ) ); - if ( parent ) { - parent.replaceChild( elem, this ); - } - } - - // Force callback invocation - }, ignored ); - } -} ); - -jQuery.each( { - appendTo: "append", - prependTo: "prepend", - insertBefore: "before", - insertAfter: "after", - replaceAll: "replaceWith" -}, function( name, original ) { - jQuery.fn[ name ] = function( selector ) { - var elems, - ret = [], - insert = jQuery( selector ), - last = insert.length - 1, - i = 0; - - for ( ; i <= last; i++ ) { - elems = i === last ? this : this.clone( true ); - jQuery( insert[ i ] )[ original ]( elems ); - - // Support: Android <=4.0 only, PhantomJS 1 only - // .get() because push.apply(_, arraylike) throws on ancient WebKit - push.apply( ret, elems.get() ); - } - - return this.pushStack( ret ); - }; -} ); -var rnumnonpx = new RegExp( "^(" + pnum + ")(?!px)[a-z%]+$", "i" ); - -var getStyles = function( elem ) { - - // Support: IE <=11 only, Firefox <=30 (#15098, #14150) - // IE throws on elements created in popups - // FF meanwhile throws on frame elements through "defaultView.getComputedStyle" - var view = elem.ownerDocument.defaultView; - - if ( !view || !view.opener ) { - view = window; - } - - return view.getComputedStyle( elem ); - }; - -var swap = function( elem, options, callback ) { - var ret, name, - old = {}; - - // Remember the old values, and insert the new ones - for ( name in options ) { - old[ name ] = elem.style[ name ]; - elem.style[ name ] = options[ name ]; - } - - ret = callback.call( elem ); - - // Revert the old values - for ( name in options ) { - elem.style[ name ] = old[ name ]; - } - - return ret; -}; - - -var rboxStyle = new RegExp( cssExpand.join( "|" ), "i" ); - - - -( function() { - - // Executing both pixelPosition & boxSizingReliable tests require only one layout - // so they're executed at the same time to save the second computation. - function computeStyleTests() { - - // This is a singleton, we need to execute it only once - if ( !div ) { - return; - } - - container.style.cssText = "position:absolute;left:-11111px;width:60px;" + - "margin-top:1px;padding:0;border:0"; - div.style.cssText = - "position:relative;display:block;box-sizing:border-box;overflow:scroll;" + - "margin:auto;border:1px;padding:1px;" + - "width:60%;top:1%"; - documentElement.appendChild( container ).appendChild( div ); - - var divStyle = window.getComputedStyle( div ); - pixelPositionVal = divStyle.top !== "1%"; - - // Support: Android 4.0 - 4.3 only, Firefox <=3 - 44 - reliableMarginLeftVal = roundPixelMeasures( divStyle.marginLeft ) === 12; - - // Support: Android 4.0 - 4.3 only, Safari <=9.1 - 10.1, iOS <=7.0 - 9.3 - // Some styles come back with percentage values, even though they shouldn't - div.style.right = "60%"; - pixelBoxStylesVal = roundPixelMeasures( divStyle.right ) === 36; - - // Support: IE 9 - 11 only - // Detect misreporting of content dimensions for box-sizing:border-box elements - boxSizingReliableVal = roundPixelMeasures( divStyle.width ) === 36; - - // Support: IE 9 only - // Detect overflow:scroll screwiness (gh-3699) - // Support: Chrome <=64 - // Don't get tricked when zoom affects offsetWidth (gh-4029) - div.style.position = "absolute"; - scrollboxSizeVal = roundPixelMeasures( div.offsetWidth / 3 ) === 12; - - documentElement.removeChild( container ); - - // Nullify the div so it wouldn't be stored in the memory and - // it will also be a sign that checks already performed - div = null; - } - - function roundPixelMeasures( measure ) { - return Math.round( parseFloat( measure ) ); - } - - var pixelPositionVal, boxSizingReliableVal, scrollboxSizeVal, pixelBoxStylesVal, - reliableTrDimensionsVal, reliableMarginLeftVal, - container = document.createElement( "div" ), - div = document.createElement( "div" ); - - // Finish early in limited (non-browser) environments - if ( !div.style ) { - return; - } - - // Support: IE <=9 - 11 only - // Style of cloned element affects source element cloned (#8908) - div.style.backgroundClip = "content-box"; - div.cloneNode( true ).style.backgroundClip = ""; - support.clearCloneStyle = div.style.backgroundClip === "content-box"; - - jQuery.extend( support, { - boxSizingReliable: function() { - computeStyleTests(); - return boxSizingReliableVal; - }, - pixelBoxStyles: function() { - computeStyleTests(); - return pixelBoxStylesVal; - }, - pixelPosition: function() { - computeStyleTests(); - return pixelPositionVal; - }, - reliableMarginLeft: function() { - computeStyleTests(); - return reliableMarginLeftVal; - }, - scrollboxSize: function() { - computeStyleTests(); - return scrollboxSizeVal; - }, - - // Support: IE 9 - 11+, Edge 15 - 18+ - // IE/Edge misreport `getComputedStyle` of table rows with width/height - // set in CSS while `offset*` properties report correct values. - // Behavior in IE 9 is more subtle than in newer versions & it passes - // some versions of this test; make sure not to make it pass there! - reliableTrDimensions: function() { - var table, tr, trChild, trStyle; - if ( reliableTrDimensionsVal == null ) { - table = document.createElement( "table" ); - tr = document.createElement( "tr" ); - trChild = document.createElement( "div" ); - - table.style.cssText = "position:absolute;left:-11111px"; - tr.style.height = "1px"; - trChild.style.height = "9px"; - - documentElement - .appendChild( table ) - .appendChild( tr ) - .appendChild( trChild ); - - trStyle = window.getComputedStyle( tr ); - reliableTrDimensionsVal = parseInt( trStyle.height ) > 3; - - documentElement.removeChild( table ); - } - return reliableTrDimensionsVal; - } - } ); -} )(); - - -function curCSS( elem, name, computed ) { - var width, minWidth, maxWidth, ret, - - // Support: Firefox 51+ - // Retrieving style before computed somehow - // fixes an issue with getting wrong values - // on detached elements - style = elem.style; - - computed = computed || getStyles( elem ); - - // getPropertyValue is needed for: - // .css('filter') (IE 9 only, #12537) - // .css('--customProperty) (#3144) - if ( computed ) { - ret = computed.getPropertyValue( name ) || computed[ name ]; - - if ( ret === "" && !isAttached( elem ) ) { - ret = jQuery.style( elem, name ); - } - - // A tribute to the "awesome hack by Dean Edwards" - // Android Browser returns percentage for some values, - // but width seems to be reliably pixels. - // This is against the CSSOM draft spec: - // https://drafts.csswg.org/cssom/#resolved-values - if ( !support.pixelBoxStyles() && rnumnonpx.test( ret ) && rboxStyle.test( name ) ) { - - // Remember the original values - width = style.width; - minWidth = style.minWidth; - maxWidth = style.maxWidth; - - // Put in the new values to get a computed value out - style.minWidth = style.maxWidth = style.width = ret; - ret = computed.width; - - // Revert the changed values - style.width = width; - style.minWidth = minWidth; - style.maxWidth = maxWidth; - } - } - - return ret !== undefined ? - - // Support: IE <=9 - 11 only - // IE returns zIndex value as an integer. - ret + "" : - ret; -} - - -function addGetHookIf( conditionFn, hookFn ) { - - // Define the hook, we'll check on the first run if it's really needed. - return { - get: function() { - if ( conditionFn() ) { - - // Hook not needed (or it's not possible to use it due - // to missing dependency), remove it. - delete this.get; - return; - } - - // Hook needed; redefine it so that the support test is not executed again. - return ( this.get = hookFn ).apply( this, arguments ); - } - }; -} - - -var cssPrefixes = [ "Webkit", "Moz", "ms" ], - emptyStyle = document.createElement( "div" ).style, - vendorProps = {}; - -// Return a vendor-prefixed property or undefined -function vendorPropName( name ) { - - // Check for vendor prefixed names - var capName = name[ 0 ].toUpperCase() + name.slice( 1 ), - i = cssPrefixes.length; - - while ( i-- ) { - name = cssPrefixes[ i ] + capName; - if ( name in emptyStyle ) { - return name; - } - } -} - -// Return a potentially-mapped jQuery.cssProps or vendor prefixed property -function finalPropName( name ) { - var final = jQuery.cssProps[ name ] || vendorProps[ name ]; - - if ( final ) { - return final; - } - if ( name in emptyStyle ) { - return name; - } - return vendorProps[ name ] = vendorPropName( name ) || name; -} - - -var - - // Swappable if display is none or starts with table - // except "table", "table-cell", or "table-caption" - // See here for display values: https://developer.mozilla.org/en-US/docs/CSS/display - rdisplayswap = /^(none|table(?!-c[ea]).+)/, - rcustomProp = /^--/, - cssShow = { position: "absolute", visibility: "hidden", display: "block" }, - cssNormalTransform = { - letterSpacing: "0", - fontWeight: "400" - }; - -function setPositiveNumber( _elem, value, subtract ) { - - // Any relative (+/-) values have already been - // normalized at this point - var matches = rcssNum.exec( value ); - return matches ? - - // Guard against undefined "subtract", e.g., when used as in cssHooks - Math.max( 0, matches[ 2 ] - ( subtract || 0 ) ) + ( matches[ 3 ] || "px" ) : - value; -} - -function boxModelAdjustment( elem, dimension, box, isBorderBox, styles, computedVal ) { - var i = dimension === "width" ? 1 : 0, - extra = 0, - delta = 0; - - // Adjustment may not be necessary - if ( box === ( isBorderBox ? "border" : "content" ) ) { - return 0; - } - - for ( ; i < 4; i += 2 ) { - - // Both box models exclude margin - if ( box === "margin" ) { - delta += jQuery.css( elem, box + cssExpand[ i ], true, styles ); - } - - // If we get here with a content-box, we're seeking "padding" or "border" or "margin" - if ( !isBorderBox ) { - - // Add padding - delta += jQuery.css( elem, "padding" + cssExpand[ i ], true, styles ); - - // For "border" or "margin", add border - if ( box !== "padding" ) { - delta += jQuery.css( elem, "border" + cssExpand[ i ] + "Width", true, styles ); - - // But still keep track of it otherwise - } else { - extra += jQuery.css( elem, "border" + cssExpand[ i ] + "Width", true, styles ); - } - - // If we get here with a border-box (content + padding + border), we're seeking "content" or - // "padding" or "margin" - } else { - - // For "content", subtract padding - if ( box === "content" ) { - delta -= jQuery.css( elem, "padding" + cssExpand[ i ], true, styles ); - } - - // For "content" or "padding", subtract border - if ( box !== "margin" ) { - delta -= jQuery.css( elem, "border" + cssExpand[ i ] + "Width", true, styles ); - } - } - } - - // Account for positive content-box scroll gutter when requested by providing computedVal - if ( !isBorderBox && computedVal >= 0 ) { - - // offsetWidth/offsetHeight is a rounded sum of content, padding, scroll gutter, and border - // Assuming integer scroll gutter, subtract the rest and round down - delta += Math.max( 0, Math.ceil( - elem[ "offset" + dimension[ 0 ].toUpperCase() + dimension.slice( 1 ) ] - - computedVal - - delta - - extra - - 0.5 - - // If offsetWidth/offsetHeight is unknown, then we can't determine content-box scroll gutter - // Use an explicit zero to avoid NaN (gh-3964) - ) ) || 0; - } - - return delta; -} - -function getWidthOrHeight( elem, dimension, extra ) { - - // Start with computed style - var styles = getStyles( elem ), - - // To avoid forcing a reflow, only fetch boxSizing if we need it (gh-4322). - // Fake content-box until we know it's needed to know the true value. - boxSizingNeeded = !support.boxSizingReliable() || extra, - isBorderBox = boxSizingNeeded && - jQuery.css( elem, "boxSizing", false, styles ) === "border-box", - valueIsBorderBox = isBorderBox, - - val = curCSS( elem, dimension, styles ), - offsetProp = "offset" + dimension[ 0 ].toUpperCase() + dimension.slice( 1 ); - - // Support: Firefox <=54 - // Return a confounding non-pixel value or feign ignorance, as appropriate. - if ( rnumnonpx.test( val ) ) { - if ( !extra ) { - return val; - } - val = "auto"; - } - - - // Support: IE 9 - 11 only - // Use offsetWidth/offsetHeight for when box sizing is unreliable. - // In those cases, the computed value can be trusted to be border-box. - if ( ( !support.boxSizingReliable() && isBorderBox || - - // Support: IE 10 - 11+, Edge 15 - 18+ - // IE/Edge misreport `getComputedStyle` of table rows with width/height - // set in CSS while `offset*` properties report correct values. - // Interestingly, in some cases IE 9 doesn't suffer from this issue. - !support.reliableTrDimensions() && nodeName( elem, "tr" ) || - - // Fall back to offsetWidth/offsetHeight when value is "auto" - // This happens for inline elements with no explicit setting (gh-3571) - val === "auto" || - - // Support: Android <=4.1 - 4.3 only - // Also use offsetWidth/offsetHeight for misreported inline dimensions (gh-3602) - !parseFloat( val ) && jQuery.css( elem, "display", false, styles ) === "inline" ) && - - // Make sure the element is visible & connected - elem.getClientRects().length ) { - - isBorderBox = jQuery.css( elem, "boxSizing", false, styles ) === "border-box"; - - // Where available, offsetWidth/offsetHeight approximate border box dimensions. - // Where not available (e.g., SVG), assume unreliable box-sizing and interpret the - // retrieved value as a content box dimension. - valueIsBorderBox = offsetProp in elem; - if ( valueIsBorderBox ) { - val = elem[ offsetProp ]; - } - } - - // Normalize "" and auto - val = parseFloat( val ) || 0; - - // Adjust for the element's box model - return ( val + - boxModelAdjustment( - elem, - dimension, - extra || ( isBorderBox ? "border" : "content" ), - valueIsBorderBox, - styles, - - // Provide the current computed size to request scroll gutter calculation (gh-3589) - val - ) - ) + "px"; -} - -jQuery.extend( { - - // Add in style property hooks for overriding the default - // behavior of getting and setting a style property - cssHooks: { - opacity: { - get: function( elem, computed ) { - if ( computed ) { - - // We should always get a number back from opacity - var ret = curCSS( elem, "opacity" ); - return ret === "" ? "1" : ret; - } - } - } - }, - - // Don't automatically add "px" to these possibly-unitless properties - cssNumber: { - "animationIterationCount": true, - "columnCount": true, - "fillOpacity": true, - "flexGrow": true, - "flexShrink": true, - "fontWeight": true, - "gridArea": true, - "gridColumn": true, - "gridColumnEnd": true, - "gridColumnStart": true, - "gridRow": true, - "gridRowEnd": true, - "gridRowStart": true, - "lineHeight": true, - "opacity": true, - "order": true, - "orphans": true, - "widows": true, - "zIndex": true, - "zoom": true - }, - - // Add in properties whose names you wish to fix before - // setting or getting the value - cssProps: {}, - - // Get and set the style property on a DOM Node - style: function( elem, name, value, extra ) { - - // Don't set styles on text and comment nodes - if ( !elem || elem.nodeType === 3 || elem.nodeType === 8 || !elem.style ) { - return; - } - - // Make sure that we're working with the right name - var ret, type, hooks, - origName = camelCase( name ), - isCustomProp = rcustomProp.test( name ), - style = elem.style; - - // Make sure that we're working with the right name. We don't - // want to query the value if it is a CSS custom property - // since they are user-defined. - if ( !isCustomProp ) { - name = finalPropName( origName ); - } - - // Gets hook for the prefixed version, then unprefixed version - hooks = jQuery.cssHooks[ name ] || jQuery.cssHooks[ origName ]; - - // Check if we're setting a value - if ( value !== undefined ) { - type = typeof value; - - // Convert "+=" or "-=" to relative numbers (#7345) - if ( type === "string" && ( ret = rcssNum.exec( value ) ) && ret[ 1 ] ) { - value = adjustCSS( elem, name, ret ); - - // Fixes bug #9237 - type = "number"; - } - - // Make sure that null and NaN values aren't set (#7116) - if ( value == null || value !== value ) { - return; - } - - // If a number was passed in, add the unit (except for certain CSS properties) - // The isCustomProp check can be removed in jQuery 4.0 when we only auto-append - // "px" to a few hardcoded values. - if ( type === "number" && !isCustomProp ) { - value += ret && ret[ 3 ] || ( jQuery.cssNumber[ origName ] ? "" : "px" ); - } - - // background-* props affect original clone's values - if ( !support.clearCloneStyle && value === "" && name.indexOf( "background" ) === 0 ) { - style[ name ] = "inherit"; - } - - // If a hook was provided, use that value, otherwise just set the specified value - if ( !hooks || !( "set" in hooks ) || - ( value = hooks.set( elem, value, extra ) ) !== undefined ) { - - if ( isCustomProp ) { - style.setProperty( name, value ); - } else { - style[ name ] = value; - } - } - - } else { - - // If a hook was provided get the non-computed value from there - if ( hooks && "get" in hooks && - ( ret = hooks.get( elem, false, extra ) ) !== undefined ) { - - return ret; - } - - // Otherwise just get the value from the style object - return style[ name ]; - } - }, - - css: function( elem, name, extra, styles ) { - var val, num, hooks, - origName = camelCase( name ), - isCustomProp = rcustomProp.test( name ); - - // Make sure that we're working with the right name. We don't - // want to modify the value if it is a CSS custom property - // since they are user-defined. - if ( !isCustomProp ) { - name = finalPropName( origName ); - } - - // Try prefixed name followed by the unprefixed name - hooks = jQuery.cssHooks[ name ] || jQuery.cssHooks[ origName ]; - - // If a hook was provided get the computed value from there - if ( hooks && "get" in hooks ) { - val = hooks.get( elem, true, extra ); - } - - // Otherwise, if a way to get the computed value exists, use that - if ( val === undefined ) { - val = curCSS( elem, name, styles ); - } - - // Convert "normal" to computed value - if ( val === "normal" && name in cssNormalTransform ) { - val = cssNormalTransform[ name ]; - } - - // Make numeric if forced or a qualifier was provided and val looks numeric - if ( extra === "" || extra ) { - num = parseFloat( val ); - return extra === true || isFinite( num ) ? num || 0 : val; - } - - return val; - } -} ); - -jQuery.each( [ "height", "width" ], function( _i, dimension ) { - jQuery.cssHooks[ dimension ] = { - get: function( elem, computed, extra ) { - if ( computed ) { - - // Certain elements can have dimension info if we invisibly show them - // but it must have a current display style that would benefit - return rdisplayswap.test( jQuery.css( elem, "display" ) ) && - - // Support: Safari 8+ - // Table columns in Safari have non-zero offsetWidth & zero - // getBoundingClientRect().width unless display is changed. - // Support: IE <=11 only - // Running getBoundingClientRect on a disconnected node - // in IE throws an error. - ( !elem.getClientRects().length || !elem.getBoundingClientRect().width ) ? - swap( elem, cssShow, function() { - return getWidthOrHeight( elem, dimension, extra ); - } ) : - getWidthOrHeight( elem, dimension, extra ); - } - }, - - set: function( elem, value, extra ) { - var matches, - styles = getStyles( elem ), - - // Only read styles.position if the test has a chance to fail - // to avoid forcing a reflow. - scrollboxSizeBuggy = !support.scrollboxSize() && - styles.position === "absolute", - - // To avoid forcing a reflow, only fetch boxSizing if we need it (gh-3991) - boxSizingNeeded = scrollboxSizeBuggy || extra, - isBorderBox = boxSizingNeeded && - jQuery.css( elem, "boxSizing", false, styles ) === "border-box", - subtract = extra ? - boxModelAdjustment( - elem, - dimension, - extra, - isBorderBox, - styles - ) : - 0; - - // Account for unreliable border-box dimensions by comparing offset* to computed and - // faking a content-box to get border and padding (gh-3699) - if ( isBorderBox && scrollboxSizeBuggy ) { - subtract -= Math.ceil( - elem[ "offset" + dimension[ 0 ].toUpperCase() + dimension.slice( 1 ) ] - - parseFloat( styles[ dimension ] ) - - boxModelAdjustment( elem, dimension, "border", false, styles ) - - 0.5 - ); - } - - // Convert to pixels if value adjustment is needed - if ( subtract && ( matches = rcssNum.exec( value ) ) && - ( matches[ 3 ] || "px" ) !== "px" ) { - - elem.style[ dimension ] = value; - value = jQuery.css( elem, dimension ); - } - - return setPositiveNumber( elem, value, subtract ); - } - }; -} ); - -jQuery.cssHooks.marginLeft = addGetHookIf( support.reliableMarginLeft, - function( elem, computed ) { - if ( computed ) { - return ( parseFloat( curCSS( elem, "marginLeft" ) ) || - elem.getBoundingClientRect().left - - swap( elem, { marginLeft: 0 }, function() { - return elem.getBoundingClientRect().left; - } ) - ) + "px"; - } - } -); - -// These hooks are used by animate to expand properties -jQuery.each( { - margin: "", - padding: "", - border: "Width" -}, function( prefix, suffix ) { - jQuery.cssHooks[ prefix + suffix ] = { - expand: function( value ) { - var i = 0, - expanded = {}, - - // Assumes a single number if not a string - parts = typeof value === "string" ? value.split( " " ) : [ value ]; - - for ( ; i < 4; i++ ) { - expanded[ prefix + cssExpand[ i ] + suffix ] = - parts[ i ] || parts[ i - 2 ] || parts[ 0 ]; - } - - return expanded; - } - }; - - if ( prefix !== "margin" ) { - jQuery.cssHooks[ prefix + suffix ].set = setPositiveNumber; - } -} ); - -jQuery.fn.extend( { - css: function( name, value ) { - return access( this, function( elem, name, value ) { - var styles, len, - map = {}, - i = 0; - - if ( Array.isArray( name ) ) { - styles = getStyles( elem ); - len = name.length; - - for ( ; i < len; i++ ) { - map[ name[ i ] ] = jQuery.css( elem, name[ i ], false, styles ); - } - - return map; - } - - return value !== undefined ? - jQuery.style( elem, name, value ) : - jQuery.css( elem, name ); - }, name, value, arguments.length > 1 ); - } -} ); - - -function Tween( elem, options, prop, end, easing ) { - return new Tween.prototype.init( elem, options, prop, end, easing ); -} -jQuery.Tween = Tween; - -Tween.prototype = { - constructor: Tween, - init: function( elem, options, prop, end, easing, unit ) { - this.elem = elem; - this.prop = prop; - this.easing = easing || jQuery.easing._default; - this.options = options; - this.start = this.now = this.cur(); - this.end = end; - this.unit = unit || ( jQuery.cssNumber[ prop ] ? "" : "px" ); - }, - cur: function() { - var hooks = Tween.propHooks[ this.prop ]; - - return hooks && hooks.get ? - hooks.get( this ) : - Tween.propHooks._default.get( this ); - }, - run: function( percent ) { - var eased, - hooks = Tween.propHooks[ this.prop ]; - - if ( this.options.duration ) { - this.pos = eased = jQuery.easing[ this.easing ]( - percent, this.options.duration * percent, 0, 1, this.options.duration - ); - } else { - this.pos = eased = percent; - } - this.now = ( this.end - this.start ) * eased + this.start; - - if ( this.options.step ) { - this.options.step.call( this.elem, this.now, this ); - } - - if ( hooks && hooks.set ) { - hooks.set( this ); - } else { - Tween.propHooks._default.set( this ); - } - return this; - } -}; - -Tween.prototype.init.prototype = Tween.prototype; - -Tween.propHooks = { - _default: { - get: function( tween ) { - var result; - - // Use a property on the element directly when it is not a DOM element, - // or when there is no matching style property that exists. - if ( tween.elem.nodeType !== 1 || - tween.elem[ tween.prop ] != null && tween.elem.style[ tween.prop ] == null ) { - return tween.elem[ tween.prop ]; - } - - // Passing an empty string as a 3rd parameter to .css will automatically - // attempt a parseFloat and fallback to a string if the parse fails. - // Simple values such as "10px" are parsed to Float; - // complex values such as "rotate(1rad)" are returned as-is. - result = jQuery.css( tween.elem, tween.prop, "" ); - - // Empty strings, null, undefined and "auto" are converted to 0. - return !result || result === "auto" ? 0 : result; - }, - set: function( tween ) { - - // Use step hook for back compat. - // Use cssHook if its there. - // Use .style if available and use plain properties where available. - if ( jQuery.fx.step[ tween.prop ] ) { - jQuery.fx.step[ tween.prop ]( tween ); - } else if ( tween.elem.nodeType === 1 && ( - jQuery.cssHooks[ tween.prop ] || - tween.elem.style[ finalPropName( tween.prop ) ] != null ) ) { - jQuery.style( tween.elem, tween.prop, tween.now + tween.unit ); - } else { - tween.elem[ tween.prop ] = tween.now; - } - } - } -}; - -// Support: IE <=9 only -// Panic based approach to setting things on disconnected nodes -Tween.propHooks.scrollTop = Tween.propHooks.scrollLeft = { - set: function( tween ) { - if ( tween.elem.nodeType && tween.elem.parentNode ) { - tween.elem[ tween.prop ] = tween.now; - } - } -}; - -jQuery.easing = { - linear: function( p ) { - return p; - }, - swing: function( p ) { - return 0.5 - Math.cos( p * Math.PI ) / 2; - }, - _default: "swing" -}; - -jQuery.fx = Tween.prototype.init; - -// Back compat <1.8 extension point -jQuery.fx.step = {}; - - - - -var - fxNow, inProgress, - rfxtypes = /^(?:toggle|show|hide)$/, - rrun = /queueHooks$/; - -function schedule() { - if ( inProgress ) { - if ( document.hidden === false && window.requestAnimationFrame ) { - window.requestAnimationFrame( schedule ); - } else { - window.setTimeout( schedule, jQuery.fx.interval ); - } - - jQuery.fx.tick(); - } -} - -// Animations created synchronously will run synchronously -function createFxNow() { - window.setTimeout( function() { - fxNow = undefined; - } ); - return ( fxNow = Date.now() ); -} - -// Generate parameters to create a standard animation -function genFx( type, includeWidth ) { - var which, - i = 0, - attrs = { height: type }; - - // If we include width, step value is 1 to do all cssExpand values, - // otherwise step value is 2 to skip over Left and Right - includeWidth = includeWidth ? 1 : 0; - for ( ; i < 4; i += 2 - includeWidth ) { - which = cssExpand[ i ]; - attrs[ "margin" + which ] = attrs[ "padding" + which ] = type; - } - - if ( includeWidth ) { - attrs.opacity = attrs.width = type; - } - - return attrs; -} - -function createTween( value, prop, animation ) { - var tween, - collection = ( Animation.tweeners[ prop ] || [] ).concat( Animation.tweeners[ "*" ] ), - index = 0, - length = collection.length; - for ( ; index < length; index++ ) { - if ( ( tween = collection[ index ].call( animation, prop, value ) ) ) { - - // We're done with this property - return tween; - } - } -} - -function defaultPrefilter( elem, props, opts ) { - var prop, value, toggle, hooks, oldfire, propTween, restoreDisplay, display, - isBox = "width" in props || "height" in props, - anim = this, - orig = {}, - style = elem.style, - hidden = elem.nodeType && isHiddenWithinTree( elem ), - dataShow = dataPriv.get( elem, "fxshow" ); - - // Queue-skipping animations hijack the fx hooks - if ( !opts.queue ) { - hooks = jQuery._queueHooks( elem, "fx" ); - if ( hooks.unqueued == null ) { - hooks.unqueued = 0; - oldfire = hooks.empty.fire; - hooks.empty.fire = function() { - if ( !hooks.unqueued ) { - oldfire(); - } - }; - } - hooks.unqueued++; - - anim.always( function() { - - // Ensure the complete handler is called before this completes - anim.always( function() { - hooks.unqueued--; - if ( !jQuery.queue( elem, "fx" ).length ) { - hooks.empty.fire(); - } - } ); - } ); - } - - // Detect show/hide animations - for ( prop in props ) { - value = props[ prop ]; - if ( rfxtypes.test( value ) ) { - delete props[ prop ]; - toggle = toggle || value === "toggle"; - if ( value === ( hidden ? "hide" : "show" ) ) { - - // Pretend to be hidden if this is a "show" and - // there is still data from a stopped show/hide - if ( value === "show" && dataShow && dataShow[ prop ] !== undefined ) { - hidden = true; - - // Ignore all other no-op show/hide data - } else { - continue; - } - } - orig[ prop ] = dataShow && dataShow[ prop ] || jQuery.style( elem, prop ); - } - } - - // Bail out if this is a no-op like .hide().hide() - propTween = !jQuery.isEmptyObject( props ); - if ( !propTween && jQuery.isEmptyObject( orig ) ) { - return; - } - - // Restrict "overflow" and "display" styles during box animations - if ( isBox && elem.nodeType === 1 ) { - - // Support: IE <=9 - 11, Edge 12 - 15 - // Record all 3 overflow attributes because IE does not infer the shorthand - // from identically-valued overflowX and overflowY and Edge just mirrors - // the overflowX value there. - opts.overflow = [ style.overflow, style.overflowX, style.overflowY ]; - - // Identify a display type, preferring old show/hide data over the CSS cascade - restoreDisplay = dataShow && dataShow.display; - if ( restoreDisplay == null ) { - restoreDisplay = dataPriv.get( elem, "display" ); - } - display = jQuery.css( elem, "display" ); - if ( display === "none" ) { - if ( restoreDisplay ) { - display = restoreDisplay; - } else { - - // Get nonempty value(s) by temporarily forcing visibility - showHide( [ elem ], true ); - restoreDisplay = elem.style.display || restoreDisplay; - display = jQuery.css( elem, "display" ); - showHide( [ elem ] ); - } - } - - // Animate inline elements as inline-block - if ( display === "inline" || display === "inline-block" && restoreDisplay != null ) { - if ( jQuery.css( elem, "float" ) === "none" ) { - - // Restore the original display value at the end of pure show/hide animations - if ( !propTween ) { - anim.done( function() { - style.display = restoreDisplay; - } ); - if ( restoreDisplay == null ) { - display = style.display; - restoreDisplay = display === "none" ? "" : display; - } - } - style.display = "inline-block"; - } - } - } - - if ( opts.overflow ) { - style.overflow = "hidden"; - anim.always( function() { - style.overflow = opts.overflow[ 0 ]; - style.overflowX = opts.overflow[ 1 ]; - style.overflowY = opts.overflow[ 2 ]; - } ); - } - - // Implement show/hide animations - propTween = false; - for ( prop in orig ) { - - // General show/hide setup for this element animation - if ( !propTween ) { - if ( dataShow ) { - if ( "hidden" in dataShow ) { - hidden = dataShow.hidden; - } - } else { - dataShow = dataPriv.access( elem, "fxshow", { display: restoreDisplay } ); - } - - // Store hidden/visible for toggle so `.stop().toggle()` "reverses" - if ( toggle ) { - dataShow.hidden = !hidden; - } - - // Show elements before animating them - if ( hidden ) { - showHide( [ elem ], true ); - } - - /* eslint-disable no-loop-func */ - - anim.done( function() { - - /* eslint-enable no-loop-func */ - - // The final step of a "hide" animation is actually hiding the element - if ( !hidden ) { - showHide( [ elem ] ); - } - dataPriv.remove( elem, "fxshow" ); - for ( prop in orig ) { - jQuery.style( elem, prop, orig[ prop ] ); - } - } ); - } - - // Per-property setup - propTween = createTween( hidden ? dataShow[ prop ] : 0, prop, anim ); - if ( !( prop in dataShow ) ) { - dataShow[ prop ] = propTween.start; - if ( hidden ) { - propTween.end = propTween.start; - propTween.start = 0; - } - } - } -} - -function propFilter( props, specialEasing ) { - var index, name, easing, value, hooks; - - // camelCase, specialEasing and expand cssHook pass - for ( index in props ) { - name = camelCase( index ); - easing = specialEasing[ name ]; - value = props[ index ]; - if ( Array.isArray( value ) ) { - easing = value[ 1 ]; - value = props[ index ] = value[ 0 ]; - } - - if ( index !== name ) { - props[ name ] = value; - delete props[ index ]; - } - - hooks = jQuery.cssHooks[ name ]; - if ( hooks && "expand" in hooks ) { - value = hooks.expand( value ); - delete props[ name ]; - - // Not quite $.extend, this won't overwrite existing keys. - // Reusing 'index' because we have the correct "name" - for ( index in value ) { - if ( !( index in props ) ) { - props[ index ] = value[ index ]; - specialEasing[ index ] = easing; - } - } - } else { - specialEasing[ name ] = easing; - } - } -} - -function Animation( elem, properties, options ) { - var result, - stopped, - index = 0, - length = Animation.prefilters.length, - deferred = jQuery.Deferred().always( function() { - - // Don't match elem in the :animated selector - delete tick.elem; - } ), - tick = function() { - if ( stopped ) { - return false; - } - var currentTime = fxNow || createFxNow(), - remaining = Math.max( 0, animation.startTime + animation.duration - currentTime ), - - // Support: Android 2.3 only - // Archaic crash bug won't allow us to use `1 - ( 0.5 || 0 )` (#12497) - temp = remaining / animation.duration || 0, - percent = 1 - temp, - index = 0, - length = animation.tweens.length; - - for ( ; index < length; index++ ) { - animation.tweens[ index ].run( percent ); - } - - deferred.notifyWith( elem, [ animation, percent, remaining ] ); - - // If there's more to do, yield - if ( percent < 1 && length ) { - return remaining; - } - - // If this was an empty animation, synthesize a final progress notification - if ( !length ) { - deferred.notifyWith( elem, [ animation, 1, 0 ] ); - } - - // Resolve the animation and report its conclusion - deferred.resolveWith( elem, [ animation ] ); - return false; - }, - animation = deferred.promise( { - elem: elem, - props: jQuery.extend( {}, properties ), - opts: jQuery.extend( true, { - specialEasing: {}, - easing: jQuery.easing._default - }, options ), - originalProperties: properties, - originalOptions: options, - startTime: fxNow || createFxNow(), - duration: options.duration, - tweens: [], - createTween: function( prop, end ) { - var tween = jQuery.Tween( elem, animation.opts, prop, end, - animation.opts.specialEasing[ prop ] || animation.opts.easing ); - animation.tweens.push( tween ); - return tween; - }, - stop: function( gotoEnd ) { - var index = 0, - - // If we are going to the end, we want to run all the tweens - // otherwise we skip this part - length = gotoEnd ? animation.tweens.length : 0; - if ( stopped ) { - return this; - } - stopped = true; - for ( ; index < length; index++ ) { - animation.tweens[ index ].run( 1 ); - } - - // Resolve when we played the last frame; otherwise, reject - if ( gotoEnd ) { - deferred.notifyWith( elem, [ animation, 1, 0 ] ); - deferred.resolveWith( elem, [ animation, gotoEnd ] ); - } else { - deferred.rejectWith( elem, [ animation, gotoEnd ] ); - } - return this; - } - } ), - props = animation.props; - - propFilter( props, animation.opts.specialEasing ); - - for ( ; index < length; index++ ) { - result = Animation.prefilters[ index ].call( animation, elem, props, animation.opts ); - if ( result ) { - if ( isFunction( result.stop ) ) { - jQuery._queueHooks( animation.elem, animation.opts.queue ).stop = - result.stop.bind( result ); - } - return result; - } - } - - jQuery.map( props, createTween, animation ); - - if ( isFunction( animation.opts.start ) ) { - animation.opts.start.call( elem, animation ); - } - - // Attach callbacks from options - animation - .progress( animation.opts.progress ) - .done( animation.opts.done, animation.opts.complete ) - .fail( animation.opts.fail ) - .always( animation.opts.always ); - - jQuery.fx.timer( - jQuery.extend( tick, { - elem: elem, - anim: animation, - queue: animation.opts.queue - } ) - ); - - return animation; -} - -jQuery.Animation = jQuery.extend( Animation, { - - tweeners: { - "*": [ function( prop, value ) { - var tween = this.createTween( prop, value ); - adjustCSS( tween.elem, prop, rcssNum.exec( value ), tween ); - return tween; - } ] - }, - - tweener: function( props, callback ) { - if ( isFunction( props ) ) { - callback = props; - props = [ "*" ]; - } else { - props = props.match( rnothtmlwhite ); - } - - var prop, - index = 0, - length = props.length; - - for ( ; index < length; index++ ) { - prop = props[ index ]; - Animation.tweeners[ prop ] = Animation.tweeners[ prop ] || []; - Animation.tweeners[ prop ].unshift( callback ); - } - }, - - prefilters: [ defaultPrefilter ], - - prefilter: function( callback, prepend ) { - if ( prepend ) { - Animation.prefilters.unshift( callback ); - } else { - Animation.prefilters.push( callback ); - } - } -} ); - -jQuery.speed = function( speed, easing, fn ) { - var opt = speed && typeof speed === "object" ? jQuery.extend( {}, speed ) : { - complete: fn || !fn && easing || - isFunction( speed ) && speed, - duration: speed, - easing: fn && easing || easing && !isFunction( easing ) && easing - }; - - // Go to the end state if fx are off - if ( jQuery.fx.off ) { - opt.duration = 0; - - } else { - if ( typeof opt.duration !== "number" ) { - if ( opt.duration in jQuery.fx.speeds ) { - opt.duration = jQuery.fx.speeds[ opt.duration ]; - - } else { - opt.duration = jQuery.fx.speeds._default; - } - } - } - - // Normalize opt.queue - true/undefined/null -> "fx" - if ( opt.queue == null || opt.queue === true ) { - opt.queue = "fx"; - } - - // Queueing - opt.old = opt.complete; - - opt.complete = function() { - if ( isFunction( opt.old ) ) { - opt.old.call( this ); - } - - if ( opt.queue ) { - jQuery.dequeue( this, opt.queue ); - } - }; - - return opt; -}; - -jQuery.fn.extend( { - fadeTo: function( speed, to, easing, callback ) { - - // Show any hidden elements after setting opacity to 0 - return this.filter( isHiddenWithinTree ).css( "opacity", 0 ).show() - - // Animate to the value specified - .end().animate( { opacity: to }, speed, easing, callback ); - }, - animate: function( prop, speed, easing, callback ) { - var empty = jQuery.isEmptyObject( prop ), - optall = jQuery.speed( speed, easing, callback ), - doAnimation = function() { - - // Operate on a copy of prop so per-property easing won't be lost - var anim = Animation( this, jQuery.extend( {}, prop ), optall ); - - // Empty animations, or finishing resolves immediately - if ( empty || dataPriv.get( this, "finish" ) ) { - anim.stop( true ); - } - }; - doAnimation.finish = doAnimation; - - return empty || optall.queue === false ? - this.each( doAnimation ) : - this.queue( optall.queue, doAnimation ); - }, - stop: function( type, clearQueue, gotoEnd ) { - var stopQueue = function( hooks ) { - var stop = hooks.stop; - delete hooks.stop; - stop( gotoEnd ); - }; - - if ( typeof type !== "string" ) { - gotoEnd = clearQueue; - clearQueue = type; - type = undefined; - } - if ( clearQueue ) { - this.queue( type || "fx", [] ); - } - - return this.each( function() { - var dequeue = true, - index = type != null && type + "queueHooks", - timers = jQuery.timers, - data = dataPriv.get( this ); - - if ( index ) { - if ( data[ index ] && data[ index ].stop ) { - stopQueue( data[ index ] ); - } - } else { - for ( index in data ) { - if ( data[ index ] && data[ index ].stop && rrun.test( index ) ) { - stopQueue( data[ index ] ); - } - } - } - - for ( index = timers.length; index--; ) { - if ( timers[ index ].elem === this && - ( type == null || timers[ index ].queue === type ) ) { - - timers[ index ].anim.stop( gotoEnd ); - dequeue = false; - timers.splice( index, 1 ); - } - } - - // Start the next in the queue if the last step wasn't forced. - // Timers currently will call their complete callbacks, which - // will dequeue but only if they were gotoEnd. - if ( dequeue || !gotoEnd ) { - jQuery.dequeue( this, type ); - } - } ); - }, - finish: function( type ) { - if ( type !== false ) { - type = type || "fx"; - } - return this.each( function() { - var index, - data = dataPriv.get( this ), - queue = data[ type + "queue" ], - hooks = data[ type + "queueHooks" ], - timers = jQuery.timers, - length = queue ? queue.length : 0; - - // Enable finishing flag on private data - data.finish = true; - - // Empty the queue first - jQuery.queue( this, type, [] ); - - if ( hooks && hooks.stop ) { - hooks.stop.call( this, true ); - } - - // Look for any active animations, and finish them - for ( index = timers.length; index--; ) { - if ( timers[ index ].elem === this && timers[ index ].queue === type ) { - timers[ index ].anim.stop( true ); - timers.splice( index, 1 ); - } - } - - // Look for any animations in the old queue and finish them - for ( index = 0; index < length; index++ ) { - if ( queue[ index ] && queue[ index ].finish ) { - queue[ index ].finish.call( this ); - } - } - - // Turn off finishing flag - delete data.finish; - } ); - } -} ); - -jQuery.each( [ "toggle", "show", "hide" ], function( _i, name ) { - var cssFn = jQuery.fn[ name ]; - jQuery.fn[ name ] = function( speed, easing, callback ) { - return speed == null || typeof speed === "boolean" ? - cssFn.apply( this, arguments ) : - this.animate( genFx( name, true ), speed, easing, callback ); - }; -} ); - -// Generate shortcuts for custom animations -jQuery.each( { - slideDown: genFx( "show" ), - slideUp: genFx( "hide" ), - slideToggle: genFx( "toggle" ), - fadeIn: { opacity: "show" }, - fadeOut: { opacity: "hide" }, - fadeToggle: { opacity: "toggle" } -}, function( name, props ) { - jQuery.fn[ name ] = function( speed, easing, callback ) { - return this.animate( props, speed, easing, callback ); - }; -} ); - -jQuery.timers = []; -jQuery.fx.tick = function() { - var timer, - i = 0, - timers = jQuery.timers; - - fxNow = Date.now(); - - for ( ; i < timers.length; i++ ) { - timer = timers[ i ]; - - // Run the timer and safely remove it when done (allowing for external removal) - if ( !timer() && timers[ i ] === timer ) { - timers.splice( i--, 1 ); - } - } - - if ( !timers.length ) { - jQuery.fx.stop(); - } - fxNow = undefined; -}; - -jQuery.fx.timer = function( timer ) { - jQuery.timers.push( timer ); - jQuery.fx.start(); -}; - -jQuery.fx.interval = 13; -jQuery.fx.start = function() { - if ( inProgress ) { - return; - } - - inProgress = true; - schedule(); -}; - -jQuery.fx.stop = function() { - inProgress = null; -}; - -jQuery.fx.speeds = { - slow: 600, - fast: 200, - - // Default speed - _default: 400 -}; - - -// Based off of the plugin by Clint Helfers, with permission. -// https://web.archive.org/web/20100324014747/http://blindsignals.com/index.php/2009/07/jquery-delay/ -jQuery.fn.delay = function( time, type ) { - time = jQuery.fx ? jQuery.fx.speeds[ time ] || time : time; - type = type || "fx"; - - return this.queue( type, function( next, hooks ) { - var timeout = window.setTimeout( next, time ); - hooks.stop = function() { - window.clearTimeout( timeout ); - }; - } ); -}; - - -( function() { - var input = document.createElement( "input" ), - select = document.createElement( "select" ), - opt = select.appendChild( document.createElement( "option" ) ); - - input.type = "checkbox"; - - // Support: Android <=4.3 only - // Default value for a checkbox should be "on" - support.checkOn = input.value !== ""; - - // Support: IE <=11 only - // Must access selectedIndex to make default options select - support.optSelected = opt.selected; - - // Support: IE <=11 only - // An input loses its value after becoming a radio - input = document.createElement( "input" ); - input.value = "t"; - input.type = "radio"; - support.radioValue = input.value === "t"; -} )(); - - -var boolHook, - attrHandle = jQuery.expr.attrHandle; - -jQuery.fn.extend( { - attr: function( name, value ) { - return access( this, jQuery.attr, name, value, arguments.length > 1 ); - }, - - removeAttr: function( name ) { - return this.each( function() { - jQuery.removeAttr( this, name ); - } ); - } -} ); - -jQuery.extend( { - attr: function( elem, name, value ) { - var ret, hooks, - nType = elem.nodeType; - - // Don't get/set attributes on text, comment and attribute nodes - if ( nType === 3 || nType === 8 || nType === 2 ) { - return; - } - - // Fallback to prop when attributes are not supported - if ( typeof elem.getAttribute === "undefined" ) { - return jQuery.prop( elem, name, value ); - } - - // Attribute hooks are determined by the lowercase version - // Grab necessary hook if one is defined - if ( nType !== 1 || !jQuery.isXMLDoc( elem ) ) { - hooks = jQuery.attrHooks[ name.toLowerCase() ] || - ( jQuery.expr.match.bool.test( name ) ? boolHook : undefined ); - } - - if ( value !== undefined ) { - if ( value === null ) { - jQuery.removeAttr( elem, name ); - return; - } - - if ( hooks && "set" in hooks && - ( ret = hooks.set( elem, value, name ) ) !== undefined ) { - return ret; - } - - elem.setAttribute( name, value + "" ); - return value; - } - - if ( hooks && "get" in hooks && ( ret = hooks.get( elem, name ) ) !== null ) { - return ret; - } - - ret = jQuery.find.attr( elem, name ); - - // Non-existent attributes return null, we normalize to undefined - return ret == null ? undefined : ret; - }, - - attrHooks: { - type: { - set: function( elem, value ) { - if ( !support.radioValue && value === "radio" && - nodeName( elem, "input" ) ) { - var val = elem.value; - elem.setAttribute( "type", value ); - if ( val ) { - elem.value = val; - } - return value; - } - } - } - }, - - removeAttr: function( elem, value ) { - var name, - i = 0, - - // Attribute names can contain non-HTML whitespace characters - // https://html.spec.whatwg.org/multipage/syntax.html#attributes-2 - attrNames = value && value.match( rnothtmlwhite ); - - if ( attrNames && elem.nodeType === 1 ) { - while ( ( name = attrNames[ i++ ] ) ) { - elem.removeAttribute( name ); - } - } - } -} ); - -// Hooks for boolean attributes -boolHook = { - set: function( elem, value, name ) { - if ( value === false ) { - - // Remove boolean attributes when set to false - jQuery.removeAttr( elem, name ); - } else { - elem.setAttribute( name, name ); - } - return name; - } -}; - -jQuery.each( jQuery.expr.match.bool.source.match( /\w+/g ), function( _i, name ) { - var getter = attrHandle[ name ] || jQuery.find.attr; - - attrHandle[ name ] = function( elem, name, isXML ) { - var ret, handle, - lowercaseName = name.toLowerCase(); - - if ( !isXML ) { - - // Avoid an infinite loop by temporarily removing this function from the getter - handle = attrHandle[ lowercaseName ]; - attrHandle[ lowercaseName ] = ret; - ret = getter( elem, name, isXML ) != null ? - lowercaseName : - null; - attrHandle[ lowercaseName ] = handle; - } - return ret; - }; -} ); - - - - -var rfocusable = /^(?:input|select|textarea|button)$/i, - rclickable = /^(?:a|area)$/i; - -jQuery.fn.extend( { - prop: function( name, value ) { - return access( this, jQuery.prop, name, value, arguments.length > 1 ); - }, - - removeProp: function( name ) { - return this.each( function() { - delete this[ jQuery.propFix[ name ] || name ]; - } ); - } -} ); - -jQuery.extend( { - prop: function( elem, name, value ) { - var ret, hooks, - nType = elem.nodeType; - - // Don't get/set properties on text, comment and attribute nodes - if ( nType === 3 || nType === 8 || nType === 2 ) { - return; - } - - if ( nType !== 1 || !jQuery.isXMLDoc( elem ) ) { - - // Fix name and attach hooks - name = jQuery.propFix[ name ] || name; - hooks = jQuery.propHooks[ name ]; - } - - if ( value !== undefined ) { - if ( hooks && "set" in hooks && - ( ret = hooks.set( elem, value, name ) ) !== undefined ) { - return ret; - } - - return ( elem[ name ] = value ); - } - - if ( hooks && "get" in hooks && ( ret = hooks.get( elem, name ) ) !== null ) { - return ret; - } - - return elem[ name ]; - }, - - propHooks: { - tabIndex: { - get: function( elem ) { - - // Support: IE <=9 - 11 only - // elem.tabIndex doesn't always return the - // correct value when it hasn't been explicitly set - // https://web.archive.org/web/20141116233347/http://fluidproject.org/blog/2008/01/09/getting-setting-and-removing-tabindex-values-with-javascript/ - // Use proper attribute retrieval(#12072) - var tabindex = jQuery.find.attr( elem, "tabindex" ); - - if ( tabindex ) { - return parseInt( tabindex, 10 ); - } - - if ( - rfocusable.test( elem.nodeName ) || - rclickable.test( elem.nodeName ) && - elem.href - ) { - return 0; - } - - return -1; - } - } - }, - - propFix: { - "for": "htmlFor", - "class": "className" - } -} ); - -// Support: IE <=11 only -// Accessing the selectedIndex property -// forces the browser to respect setting selected -// on the option -// The getter ensures a default option is selected -// when in an optgroup -// eslint rule "no-unused-expressions" is disabled for this code -// since it considers such accessions noop -if ( !support.optSelected ) { - jQuery.propHooks.selected = { - get: function( elem ) { - - /* eslint no-unused-expressions: "off" */ - - var parent = elem.parentNode; - if ( parent && parent.parentNode ) { - parent.parentNode.selectedIndex; - } - return null; - }, - set: function( elem ) { - - /* eslint no-unused-expressions: "off" */ - - var parent = elem.parentNode; - if ( parent ) { - parent.selectedIndex; - - if ( parent.parentNode ) { - parent.parentNode.selectedIndex; - } - } - } - }; -} - -jQuery.each( [ - "tabIndex", - "readOnly", - "maxLength", - "cellSpacing", - "cellPadding", - "rowSpan", - "colSpan", - "useMap", - "frameBorder", - "contentEditable" -], function() { - jQuery.propFix[ this.toLowerCase() ] = this; -} ); - - - - - // Strip and collapse whitespace according to HTML spec - // https://infra.spec.whatwg.org/#strip-and-collapse-ascii-whitespace - function stripAndCollapse( value ) { - var tokens = value.match( rnothtmlwhite ) || []; - return tokens.join( " " ); - } - - -function getClass( elem ) { - return elem.getAttribute && elem.getAttribute( "class" ) || ""; -} - -function classesToArray( value ) { - if ( Array.isArray( value ) ) { - return value; - } - if ( typeof value === "string" ) { - return value.match( rnothtmlwhite ) || []; - } - return []; -} - -jQuery.fn.extend( { - addClass: function( value ) { - var classes, elem, cur, curValue, clazz, j, finalValue, - i = 0; - - if ( isFunction( value ) ) { - return this.each( function( j ) { - jQuery( this ).addClass( value.call( this, j, getClass( this ) ) ); - } ); - } - - classes = classesToArray( value ); - - if ( classes.length ) { - while ( ( elem = this[ i++ ] ) ) { - curValue = getClass( elem ); - cur = elem.nodeType === 1 && ( " " + stripAndCollapse( curValue ) + " " ); - - if ( cur ) { - j = 0; - while ( ( clazz = classes[ j++ ] ) ) { - if ( cur.indexOf( " " + clazz + " " ) < 0 ) { - cur += clazz + " "; - } - } - - // Only assign if different to avoid unneeded rendering. - finalValue = stripAndCollapse( cur ); - if ( curValue !== finalValue ) { - elem.setAttribute( "class", finalValue ); - } - } - } - } - - return this; - }, - - removeClass: function( value ) { - var classes, elem, cur, curValue, clazz, j, finalValue, - i = 0; - - if ( isFunction( value ) ) { - return this.each( function( j ) { - jQuery( this ).removeClass( value.call( this, j, getClass( this ) ) ); - } ); - } - - if ( !arguments.length ) { - return this.attr( "class", "" ); - } - - classes = classesToArray( value ); - - if ( classes.length ) { - while ( ( elem = this[ i++ ] ) ) { - curValue = getClass( elem ); - - // This expression is here for better compressibility (see addClass) - cur = elem.nodeType === 1 && ( " " + stripAndCollapse( curValue ) + " " ); - - if ( cur ) { - j = 0; - while ( ( clazz = classes[ j++ ] ) ) { - - // Remove *all* instances - while ( cur.indexOf( " " + clazz + " " ) > -1 ) { - cur = cur.replace( " " + clazz + " ", " " ); - } - } - - // Only assign if different to avoid unneeded rendering. - finalValue = stripAndCollapse( cur ); - if ( curValue !== finalValue ) { - elem.setAttribute( "class", finalValue ); - } - } - } - } - - return this; - }, - - toggleClass: function( value, stateVal ) { - var type = typeof value, - isValidValue = type === "string" || Array.isArray( value ); - - if ( typeof stateVal === "boolean" && isValidValue ) { - return stateVal ? this.addClass( value ) : this.removeClass( value ); - } - - if ( isFunction( value ) ) { - return this.each( function( i ) { - jQuery( this ).toggleClass( - value.call( this, i, getClass( this ), stateVal ), - stateVal - ); - } ); - } - - return this.each( function() { - var className, i, self, classNames; - - if ( isValidValue ) { - - // Toggle individual class names - i = 0; - self = jQuery( this ); - classNames = classesToArray( value ); - - while ( ( className = classNames[ i++ ] ) ) { - - // Check each className given, space separated list - if ( self.hasClass( className ) ) { - self.removeClass( className ); - } else { - self.addClass( className ); - } - } - - // Toggle whole class name - } else if ( value === undefined || type === "boolean" ) { - className = getClass( this ); - if ( className ) { - - // Store className if set - dataPriv.set( this, "__className__", className ); - } - - // If the element has a class name or if we're passed `false`, - // then remove the whole classname (if there was one, the above saved it). - // Otherwise bring back whatever was previously saved (if anything), - // falling back to the empty string if nothing was stored. - if ( this.setAttribute ) { - this.setAttribute( "class", - className || value === false ? - "" : - dataPriv.get( this, "__className__" ) || "" - ); - } - } - } ); - }, - - hasClass: function( selector ) { - var className, elem, - i = 0; - - className = " " + selector + " "; - while ( ( elem = this[ i++ ] ) ) { - if ( elem.nodeType === 1 && - ( " " + stripAndCollapse( getClass( elem ) ) + " " ).indexOf( className ) > -1 ) { - return true; - } - } - - return false; - } -} ); - - - - -var rreturn = /\r/g; - -jQuery.fn.extend( { - val: function( value ) { - var hooks, ret, valueIsFunction, - elem = this[ 0 ]; - - if ( !arguments.length ) { - if ( elem ) { - hooks = jQuery.valHooks[ elem.type ] || - jQuery.valHooks[ elem.nodeName.toLowerCase() ]; - - if ( hooks && - "get" in hooks && - ( ret = hooks.get( elem, "value" ) ) !== undefined - ) { - return ret; - } - - ret = elem.value; - - // Handle most common string cases - if ( typeof ret === "string" ) { - return ret.replace( rreturn, "" ); - } - - // Handle cases where value is null/undef or number - return ret == null ? "" : ret; - } - - return; - } - - valueIsFunction = isFunction( value ); - - return this.each( function( i ) { - var val; - - if ( this.nodeType !== 1 ) { - return; - } - - if ( valueIsFunction ) { - val = value.call( this, i, jQuery( this ).val() ); - } else { - val = value; - } - - // Treat null/undefined as ""; convert numbers to string - if ( val == null ) { - val = ""; - - } else if ( typeof val === "number" ) { - val += ""; - - } else if ( Array.isArray( val ) ) { - val = jQuery.map( val, function( value ) { - return value == null ? "" : value + ""; - } ); - } - - hooks = jQuery.valHooks[ this.type ] || jQuery.valHooks[ this.nodeName.toLowerCase() ]; - - // If set returns undefined, fall back to normal setting - if ( !hooks || !( "set" in hooks ) || hooks.set( this, val, "value" ) === undefined ) { - this.value = val; - } - } ); - } -} ); - -jQuery.extend( { - valHooks: { - option: { - get: function( elem ) { - - var val = jQuery.find.attr( elem, "value" ); - return val != null ? - val : - - // Support: IE <=10 - 11 only - // option.text throws exceptions (#14686, #14858) - // Strip and collapse whitespace - // https://html.spec.whatwg.org/#strip-and-collapse-whitespace - stripAndCollapse( jQuery.text( elem ) ); - } - }, - select: { - get: function( elem ) { - var value, option, i, - options = elem.options, - index = elem.selectedIndex, - one = elem.type === "select-one", - values = one ? null : [], - max = one ? index + 1 : options.length; - - if ( index < 0 ) { - i = max; - - } else { - i = one ? index : 0; - } - - // Loop through all the selected options - for ( ; i < max; i++ ) { - option = options[ i ]; - - // Support: IE <=9 only - // IE8-9 doesn't update selected after form reset (#2551) - if ( ( option.selected || i === index ) && - - // Don't return options that are disabled or in a disabled optgroup - !option.disabled && - ( !option.parentNode.disabled || - !nodeName( option.parentNode, "optgroup" ) ) ) { - - // Get the specific value for the option - value = jQuery( option ).val(); - - // We don't need an array for one selects - if ( one ) { - return value; - } - - // Multi-Selects return an array - values.push( value ); - } - } - - return values; - }, - - set: function( elem, value ) { - var optionSet, option, - options = elem.options, - values = jQuery.makeArray( value ), - i = options.length; - - while ( i-- ) { - option = options[ i ]; - - /* eslint-disable no-cond-assign */ - - if ( option.selected = - jQuery.inArray( jQuery.valHooks.option.get( option ), values ) > -1 - ) { - optionSet = true; - } - - /* eslint-enable no-cond-assign */ - } - - // Force browsers to behave consistently when non-matching value is set - if ( !optionSet ) { - elem.selectedIndex = -1; - } - return values; - } - } - } -} ); - -// Radios and checkboxes getter/setter -jQuery.each( [ "radio", "checkbox" ], function() { - jQuery.valHooks[ this ] = { - set: function( elem, value ) { - if ( Array.isArray( value ) ) { - return ( elem.checked = jQuery.inArray( jQuery( elem ).val(), value ) > -1 ); - } - } - }; - if ( !support.checkOn ) { - jQuery.valHooks[ this ].get = function( elem ) { - return elem.getAttribute( "value" ) === null ? "on" : elem.value; - }; - } -} ); - - - - -// Return jQuery for attributes-only inclusion - - -support.focusin = "onfocusin" in window; - - -var rfocusMorph = /^(?:focusinfocus|focusoutblur)$/, - stopPropagationCallback = function( e ) { - e.stopPropagation(); - }; - -jQuery.extend( jQuery.event, { - - trigger: function( event, data, elem, onlyHandlers ) { - - var i, cur, tmp, bubbleType, ontype, handle, special, lastElement, - eventPath = [ elem || document ], - type = hasOwn.call( event, "type" ) ? event.type : event, - namespaces = hasOwn.call( event, "namespace" ) ? event.namespace.split( "." ) : []; - - cur = lastElement = tmp = elem = elem || document; - - // Don't do events on text and comment nodes - if ( elem.nodeType === 3 || elem.nodeType === 8 ) { - return; - } - - // focus/blur morphs to focusin/out; ensure we're not firing them right now - if ( rfocusMorph.test( type + jQuery.event.triggered ) ) { - return; - } - - if ( type.indexOf( "." ) > -1 ) { - - // Namespaced trigger; create a regexp to match event type in handle() - namespaces = type.split( "." ); - type = namespaces.shift(); - namespaces.sort(); - } - ontype = type.indexOf( ":" ) < 0 && "on" + type; - - // Caller can pass in a jQuery.Event object, Object, or just an event type string - event = event[ jQuery.expando ] ? - event : - new jQuery.Event( type, typeof event === "object" && event ); - - // Trigger bitmask: & 1 for native handlers; & 2 for jQuery (always true) - event.isTrigger = onlyHandlers ? 2 : 3; - event.namespace = namespaces.join( "." ); - event.rnamespace = event.namespace ? - new RegExp( "(^|\\.)" + namespaces.join( "\\.(?:.*\\.|)" ) + "(\\.|$)" ) : - null; - - // Clean up the event in case it is being reused - event.result = undefined; - if ( !event.target ) { - event.target = elem; - } - - // Clone any incoming data and prepend the event, creating the handler arg list - data = data == null ? - [ event ] : - jQuery.makeArray( data, [ event ] ); - - // Allow special events to draw outside the lines - special = jQuery.event.special[ type ] || {}; - if ( !onlyHandlers && special.trigger && special.trigger.apply( elem, data ) === false ) { - return; - } - - // Determine event propagation path in advance, per W3C events spec (#9951) - // Bubble up to document, then to window; watch for a global ownerDocument var (#9724) - if ( !onlyHandlers && !special.noBubble && !isWindow( elem ) ) { - - bubbleType = special.delegateType || type; - if ( !rfocusMorph.test( bubbleType + type ) ) { - cur = cur.parentNode; - } - for ( ; cur; cur = cur.parentNode ) { - eventPath.push( cur ); - tmp = cur; - } - - // Only add window if we got to document (e.g., not plain obj or detached DOM) - if ( tmp === ( elem.ownerDocument || document ) ) { - eventPath.push( tmp.defaultView || tmp.parentWindow || window ); - } - } - - // Fire handlers on the event path - i = 0; - while ( ( cur = eventPath[ i++ ] ) && !event.isPropagationStopped() ) { - lastElement = cur; - event.type = i > 1 ? - bubbleType : - special.bindType || type; - - // jQuery handler - handle = ( - dataPriv.get( cur, "events" ) || Object.create( null ) - )[ event.type ] && - dataPriv.get( cur, "handle" ); - if ( handle ) { - handle.apply( cur, data ); - } - - // Native handler - handle = ontype && cur[ ontype ]; - if ( handle && handle.apply && acceptData( cur ) ) { - event.result = handle.apply( cur, data ); - if ( event.result === false ) { - event.preventDefault(); - } - } - } - event.type = type; - - // If nobody prevented the default action, do it now - if ( !onlyHandlers && !event.isDefaultPrevented() ) { - - if ( ( !special._default || - special._default.apply( eventPath.pop(), data ) === false ) && - acceptData( elem ) ) { - - // Call a native DOM method on the target with the same name as the event. - // Don't do default actions on window, that's where global variables be (#6170) - if ( ontype && isFunction( elem[ type ] ) && !isWindow( elem ) ) { - - // Don't re-trigger an onFOO event when we call its FOO() method - tmp = elem[ ontype ]; - - if ( tmp ) { - elem[ ontype ] = null; - } - - // Prevent re-triggering of the same event, since we already bubbled it above - jQuery.event.triggered = type; - - if ( event.isPropagationStopped() ) { - lastElement.addEventListener( type, stopPropagationCallback ); - } - - elem[ type ](); - - if ( event.isPropagationStopped() ) { - lastElement.removeEventListener( type, stopPropagationCallback ); - } - - jQuery.event.triggered = undefined; - - if ( tmp ) { - elem[ ontype ] = tmp; - } - } - } - } - - return event.result; - }, - - // Piggyback on a donor event to simulate a different one - // Used only for `focus(in | out)` events - simulate: function( type, elem, event ) { - var e = jQuery.extend( - new jQuery.Event(), - event, - { - type: type, - isSimulated: true - } - ); - - jQuery.event.trigger( e, null, elem ); - } - -} ); - -jQuery.fn.extend( { - - trigger: function( type, data ) { - return this.each( function() { - jQuery.event.trigger( type, data, this ); - } ); - }, - triggerHandler: function( type, data ) { - var elem = this[ 0 ]; - if ( elem ) { - return jQuery.event.trigger( type, data, elem, true ); - } - } -} ); - - -// Support: Firefox <=44 -// Firefox doesn't have focus(in | out) events -// Related ticket - https://bugzilla.mozilla.org/show_bug.cgi?id=687787 -// -// Support: Chrome <=48 - 49, Safari <=9.0 - 9.1 -// focus(in | out) events fire after focus & blur events, -// which is spec violation - http://www.w3.org/TR/DOM-Level-3-Events/#events-focusevent-event-order -// Related ticket - https://bugs.chromium.org/p/chromium/issues/detail?id=449857 -if ( !support.focusin ) { - jQuery.each( { focus: "focusin", blur: "focusout" }, function( orig, fix ) { - - // Attach a single capturing handler on the document while someone wants focusin/focusout - var handler = function( event ) { - jQuery.event.simulate( fix, event.target, jQuery.event.fix( event ) ); - }; - - jQuery.event.special[ fix ] = { - setup: function() { - - // Handle: regular nodes (via `this.ownerDocument`), window - // (via `this.document`) & document (via `this`). - var doc = this.ownerDocument || this.document || this, - attaches = dataPriv.access( doc, fix ); - - if ( !attaches ) { - doc.addEventListener( orig, handler, true ); - } - dataPriv.access( doc, fix, ( attaches || 0 ) + 1 ); - }, - teardown: function() { - var doc = this.ownerDocument || this.document || this, - attaches = dataPriv.access( doc, fix ) - 1; - - if ( !attaches ) { - doc.removeEventListener( orig, handler, true ); - dataPriv.remove( doc, fix ); - - } else { - dataPriv.access( doc, fix, attaches ); - } - } - }; - } ); -} -var location = window.location; - -var nonce = { guid: Date.now() }; - -var rquery = ( /\?/ ); - - - -// Cross-browser xml parsing -jQuery.parseXML = function( data ) { - var xml; - if ( !data || typeof data !== "string" ) { - return null; - } - - // Support: IE 9 - 11 only - // IE throws on parseFromString with invalid input. - try { - xml = ( new window.DOMParser() ).parseFromString( data, "text/xml" ); - } catch ( e ) { - xml = undefined; - } - - if ( !xml || xml.getElementsByTagName( "parsererror" ).length ) { - jQuery.error( "Invalid XML: " + data ); - } - return xml; -}; - - -var - rbracket = /\[\]$/, - rCRLF = /\r?\n/g, - rsubmitterTypes = /^(?:submit|button|image|reset|file)$/i, - rsubmittable = /^(?:input|select|textarea|keygen)/i; - -function buildParams( prefix, obj, traditional, add ) { - var name; - - if ( Array.isArray( obj ) ) { - - // Serialize array item. - jQuery.each( obj, function( i, v ) { - if ( traditional || rbracket.test( prefix ) ) { - - // Treat each array item as a scalar. - add( prefix, v ); - - } else { - - // Item is non-scalar (array or object), encode its numeric index. - buildParams( - prefix + "[" + ( typeof v === "object" && v != null ? i : "" ) + "]", - v, - traditional, - add - ); - } - } ); - - } else if ( !traditional && toType( obj ) === "object" ) { - - // Serialize object item. - for ( name in obj ) { - buildParams( prefix + "[" + name + "]", obj[ name ], traditional, add ); - } - - } else { - - // Serialize scalar item. - add( prefix, obj ); - } -} - -// Serialize an array of form elements or a set of -// key/values into a query string -jQuery.param = function( a, traditional ) { - var prefix, - s = [], - add = function( key, valueOrFunction ) { - - // If value is a function, invoke it and use its return value - var value = isFunction( valueOrFunction ) ? - valueOrFunction() : - valueOrFunction; - - s[ s.length ] = encodeURIComponent( key ) + "=" + - encodeURIComponent( value == null ? "" : value ); - }; - - if ( a == null ) { - return ""; - } - - // If an array was passed in, assume that it is an array of form elements. - if ( Array.isArray( a ) || ( a.jquery && !jQuery.isPlainObject( a ) ) ) { - - // Serialize the form elements - jQuery.each( a, function() { - add( this.name, this.value ); - } ); - - } else { - - // If traditional, encode the "old" way (the way 1.3.2 or older - // did it), otherwise encode params recursively. - for ( prefix in a ) { - buildParams( prefix, a[ prefix ], traditional, add ); - } - } - - // Return the resulting serialization - return s.join( "&" ); -}; - -jQuery.fn.extend( { - serialize: function() { - return jQuery.param( this.serializeArray() ); - }, - serializeArray: function() { - return this.map( function() { - - // Can add propHook for "elements" to filter or add form elements - var elements = jQuery.prop( this, "elements" ); - return elements ? jQuery.makeArray( elements ) : this; - } ) - .filter( function() { - var type = this.type; - - // Use .is( ":disabled" ) so that fieldset[disabled] works - return this.name && !jQuery( this ).is( ":disabled" ) && - rsubmittable.test( this.nodeName ) && !rsubmitterTypes.test( type ) && - ( this.checked || !rcheckableType.test( type ) ); - } ) - .map( function( _i, elem ) { - var val = jQuery( this ).val(); - - if ( val == null ) { - return null; - } - - if ( Array.isArray( val ) ) { - return jQuery.map( val, function( val ) { - return { name: elem.name, value: val.replace( rCRLF, "\r\n" ) }; - } ); - } - - return { name: elem.name, value: val.replace( rCRLF, "\r\n" ) }; - } ).get(); - } -} ); - - -var - r20 = /%20/g, - rhash = /#.*$/, - rantiCache = /([?&])_=[^&]*/, - rheaders = /^(.*?):[ \t]*([^\r\n]*)$/mg, - - // #7653, #8125, #8152: local protocol detection - rlocalProtocol = /^(?:about|app|app-storage|.+-extension|file|res|widget):$/, - rnoContent = /^(?:GET|HEAD)$/, - rprotocol = /^\/\//, - - /* Prefilters - * 1) They are useful to introduce custom dataTypes (see ajax/jsonp.js for an example) - * 2) These are called: - * - BEFORE asking for a transport - * - AFTER param serialization (s.data is a string if s.processData is true) - * 3) key is the dataType - * 4) the catchall symbol "*" can be used - * 5) execution will start with transport dataType and THEN continue down to "*" if needed - */ - prefilters = {}, - - /* Transports bindings - * 1) key is the dataType - * 2) the catchall symbol "*" can be used - * 3) selection will start with transport dataType and THEN go to "*" if needed - */ - transports = {}, - - // Avoid comment-prolog char sequence (#10098); must appease lint and evade compression - allTypes = "*/".concat( "*" ), - - // Anchor tag for parsing the document origin - originAnchor = document.createElement( "a" ); - originAnchor.href = location.href; - -// Base "constructor" for jQuery.ajaxPrefilter and jQuery.ajaxTransport -function addToPrefiltersOrTransports( structure ) { - - // dataTypeExpression is optional and defaults to "*" - return function( dataTypeExpression, func ) { - - if ( typeof dataTypeExpression !== "string" ) { - func = dataTypeExpression; - dataTypeExpression = "*"; - } - - var dataType, - i = 0, - dataTypes = dataTypeExpression.toLowerCase().match( rnothtmlwhite ) || []; - - if ( isFunction( func ) ) { - - // For each dataType in the dataTypeExpression - while ( ( dataType = dataTypes[ i++ ] ) ) { - - // Prepend if requested - if ( dataType[ 0 ] === "+" ) { - dataType = dataType.slice( 1 ) || "*"; - ( structure[ dataType ] = structure[ dataType ] || [] ).unshift( func ); - - // Otherwise append - } else { - ( structure[ dataType ] = structure[ dataType ] || [] ).push( func ); - } - } - } - }; -} - -// Base inspection function for prefilters and transports -function inspectPrefiltersOrTransports( structure, options, originalOptions, jqXHR ) { - - var inspected = {}, - seekingTransport = ( structure === transports ); - - function inspect( dataType ) { - var selected; - inspected[ dataType ] = true; - jQuery.each( structure[ dataType ] || [], function( _, prefilterOrFactory ) { - var dataTypeOrTransport = prefilterOrFactory( options, originalOptions, jqXHR ); - if ( typeof dataTypeOrTransport === "string" && - !seekingTransport && !inspected[ dataTypeOrTransport ] ) { - - options.dataTypes.unshift( dataTypeOrTransport ); - inspect( dataTypeOrTransport ); - return false; - } else if ( seekingTransport ) { - return !( selected = dataTypeOrTransport ); - } - } ); - return selected; - } - - return inspect( options.dataTypes[ 0 ] ) || !inspected[ "*" ] && inspect( "*" ); -} - -// A special extend for ajax options -// that takes "flat" options (not to be deep extended) -// Fixes #9887 -function ajaxExtend( target, src ) { - var key, deep, - flatOptions = jQuery.ajaxSettings.flatOptions || {}; - - for ( key in src ) { - if ( src[ key ] !== undefined ) { - ( flatOptions[ key ] ? target : ( deep || ( deep = {} ) ) )[ key ] = src[ key ]; - } - } - if ( deep ) { - jQuery.extend( true, target, deep ); - } - - return target; -} - -/* Handles responses to an ajax request: - * - finds the right dataType (mediates between content-type and expected dataType) - * - returns the corresponding response - */ -function ajaxHandleResponses( s, jqXHR, responses ) { - - var ct, type, finalDataType, firstDataType, - contents = s.contents, - dataTypes = s.dataTypes; - - // Remove auto dataType and get content-type in the process - while ( dataTypes[ 0 ] === "*" ) { - dataTypes.shift(); - if ( ct === undefined ) { - ct = s.mimeType || jqXHR.getResponseHeader( "Content-Type" ); - } - } - - // Check if we're dealing with a known content-type - if ( ct ) { - for ( type in contents ) { - if ( contents[ type ] && contents[ type ].test( ct ) ) { - dataTypes.unshift( type ); - break; - } - } - } - - // Check to see if we have a response for the expected dataType - if ( dataTypes[ 0 ] in responses ) { - finalDataType = dataTypes[ 0 ]; - } else { - - // Try convertible dataTypes - for ( type in responses ) { - if ( !dataTypes[ 0 ] || s.converters[ type + " " + dataTypes[ 0 ] ] ) { - finalDataType = type; - break; - } - if ( !firstDataType ) { - firstDataType = type; - } - } - - // Or just use first one - finalDataType = finalDataType || firstDataType; - } - - // If we found a dataType - // We add the dataType to the list if needed - // and return the corresponding response - if ( finalDataType ) { - if ( finalDataType !== dataTypes[ 0 ] ) { - dataTypes.unshift( finalDataType ); - } - return responses[ finalDataType ]; - } -} - -/* Chain conversions given the request and the original response - * Also sets the responseXXX fields on the jqXHR instance - */ -function ajaxConvert( s, response, jqXHR, isSuccess ) { - var conv2, current, conv, tmp, prev, - converters = {}, - - // Work with a copy of dataTypes in case we need to modify it for conversion - dataTypes = s.dataTypes.slice(); - - // Create converters map with lowercased keys - if ( dataTypes[ 1 ] ) { - for ( conv in s.converters ) { - converters[ conv.toLowerCase() ] = s.converters[ conv ]; - } - } - - current = dataTypes.shift(); - - // Convert to each sequential dataType - while ( current ) { - - if ( s.responseFields[ current ] ) { - jqXHR[ s.responseFields[ current ] ] = response; - } - - // Apply the dataFilter if provided - if ( !prev && isSuccess && s.dataFilter ) { - response = s.dataFilter( response, s.dataType ); - } - - prev = current; - current = dataTypes.shift(); - - if ( current ) { - - // There's only work to do if current dataType is non-auto - if ( current === "*" ) { - - current = prev; - - // Convert response if prev dataType is non-auto and differs from current - } else if ( prev !== "*" && prev !== current ) { - - // Seek a direct converter - conv = converters[ prev + " " + current ] || converters[ "* " + current ]; - - // If none found, seek a pair - if ( !conv ) { - for ( conv2 in converters ) { - - // If conv2 outputs current - tmp = conv2.split( " " ); - if ( tmp[ 1 ] === current ) { - - // If prev can be converted to accepted input - conv = converters[ prev + " " + tmp[ 0 ] ] || - converters[ "* " + tmp[ 0 ] ]; - if ( conv ) { - - // Condense equivalence converters - if ( conv === true ) { - conv = converters[ conv2 ]; - - // Otherwise, insert the intermediate dataType - } else if ( converters[ conv2 ] !== true ) { - current = tmp[ 0 ]; - dataTypes.unshift( tmp[ 1 ] ); - } - break; - } - } - } - } - - // Apply converter (if not an equivalence) - if ( conv !== true ) { - - // Unless errors are allowed to bubble, catch and return them - if ( conv && s.throws ) { - response = conv( response ); - } else { - try { - response = conv( response ); - } catch ( e ) { - return { - state: "parsererror", - error: conv ? e : "No conversion from " + prev + " to " + current - }; - } - } - } - } - } - } - - return { state: "success", data: response }; -} - -jQuery.extend( { - - // Counter for holding the number of active queries - active: 0, - - // Last-Modified header cache for next request - lastModified: {}, - etag: {}, - - ajaxSettings: { - url: location.href, - type: "GET", - isLocal: rlocalProtocol.test( location.protocol ), - global: true, - processData: true, - async: true, - contentType: "application/x-www-form-urlencoded; charset=UTF-8", - - /* - timeout: 0, - data: null, - dataType: null, - username: null, - password: null, - cache: null, - throws: false, - traditional: false, - headers: {}, - */ - - accepts: { - "*": allTypes, - text: "text/plain", - html: "text/html", - xml: "application/xml, text/xml", - json: "application/json, text/javascript" - }, - - contents: { - xml: /\bxml\b/, - html: /\bhtml/, - json: /\bjson\b/ - }, - - responseFields: { - xml: "responseXML", - text: "responseText", - json: "responseJSON" - }, - - // Data converters - // Keys separate source (or catchall "*") and destination types with a single space - converters: { - - // Convert anything to text - "* text": String, - - // Text to html (true = no transformation) - "text html": true, - - // Evaluate text as a json expression - "text json": JSON.parse, - - // Parse text as xml - "text xml": jQuery.parseXML - }, - - // For options that shouldn't be deep extended: - // you can add your own custom options here if - // and when you create one that shouldn't be - // deep extended (see ajaxExtend) - flatOptions: { - url: true, - context: true - } - }, - - // Creates a full fledged settings object into target - // with both ajaxSettings and settings fields. - // If target is omitted, writes into ajaxSettings. - ajaxSetup: function( target, settings ) { - return settings ? - - // Building a settings object - ajaxExtend( ajaxExtend( target, jQuery.ajaxSettings ), settings ) : - - // Extending ajaxSettings - ajaxExtend( jQuery.ajaxSettings, target ); - }, - - ajaxPrefilter: addToPrefiltersOrTransports( prefilters ), - ajaxTransport: addToPrefiltersOrTransports( transports ), - - // Main method - ajax: function( url, options ) { - - // If url is an object, simulate pre-1.5 signature - if ( typeof url === "object" ) { - options = url; - url = undefined; - } - - // Force options to be an object - options = options || {}; - - var transport, - - // URL without anti-cache param - cacheURL, - - // Response headers - responseHeadersString, - responseHeaders, - - // timeout handle - timeoutTimer, - - // Url cleanup var - urlAnchor, - - // Request state (becomes false upon send and true upon completion) - completed, - - // To know if global events are to be dispatched - fireGlobals, - - // Loop variable - i, - - // uncached part of the url - uncached, - - // Create the final options object - s = jQuery.ajaxSetup( {}, options ), - - // Callbacks context - callbackContext = s.context || s, - - // Context for global events is callbackContext if it is a DOM node or jQuery collection - globalEventContext = s.context && - ( callbackContext.nodeType || callbackContext.jquery ) ? - jQuery( callbackContext ) : - jQuery.event, - - // Deferreds - deferred = jQuery.Deferred(), - completeDeferred = jQuery.Callbacks( "once memory" ), - - // Status-dependent callbacks - statusCode = s.statusCode || {}, - - // Headers (they are sent all at once) - requestHeaders = {}, - requestHeadersNames = {}, - - // Default abort message - strAbort = "canceled", - - // Fake xhr - jqXHR = { - readyState: 0, - - // Builds headers hashtable if needed - getResponseHeader: function( key ) { - var match; - if ( completed ) { - if ( !responseHeaders ) { - responseHeaders = {}; - while ( ( match = rheaders.exec( responseHeadersString ) ) ) { - responseHeaders[ match[ 1 ].toLowerCase() + " " ] = - ( responseHeaders[ match[ 1 ].toLowerCase() + " " ] || [] ) - .concat( match[ 2 ] ); - } - } - match = responseHeaders[ key.toLowerCase() + " " ]; - } - return match == null ? null : match.join( ", " ); - }, - - // Raw string - getAllResponseHeaders: function() { - return completed ? responseHeadersString : null; - }, - - // Caches the header - setRequestHeader: function( name, value ) { - if ( completed == null ) { - name = requestHeadersNames[ name.toLowerCase() ] = - requestHeadersNames[ name.toLowerCase() ] || name; - requestHeaders[ name ] = value; - } - return this; - }, - - // Overrides response content-type header - overrideMimeType: function( type ) { - if ( completed == null ) { - s.mimeType = type; - } - return this; - }, - - // Status-dependent callbacks - statusCode: function( map ) { - var code; - if ( map ) { - if ( completed ) { - - // Execute the appropriate callbacks - jqXHR.always( map[ jqXHR.status ] ); - } else { - - // Lazy-add the new callbacks in a way that preserves old ones - for ( code in map ) { - statusCode[ code ] = [ statusCode[ code ], map[ code ] ]; - } - } - } - return this; - }, - - // Cancel the request - abort: function( statusText ) { - var finalText = statusText || strAbort; - if ( transport ) { - transport.abort( finalText ); - } - done( 0, finalText ); - return this; - } - }; - - // Attach deferreds - deferred.promise( jqXHR ); - - // Add protocol if not provided (prefilters might expect it) - // Handle falsy url in the settings object (#10093: consistency with old signature) - // We also use the url parameter if available - s.url = ( ( url || s.url || location.href ) + "" ) - .replace( rprotocol, location.protocol + "//" ); - - // Alias method option to type as per ticket #12004 - s.type = options.method || options.type || s.method || s.type; - - // Extract dataTypes list - s.dataTypes = ( s.dataType || "*" ).toLowerCase().match( rnothtmlwhite ) || [ "" ]; - - // A cross-domain request is in order when the origin doesn't match the current origin. - if ( s.crossDomain == null ) { - urlAnchor = document.createElement( "a" ); - - // Support: IE <=8 - 11, Edge 12 - 15 - // IE throws exception on accessing the href property if url is malformed, - // e.g. http://example.com:80x/ - try { - urlAnchor.href = s.url; - - // Support: IE <=8 - 11 only - // Anchor's host property isn't correctly set when s.url is relative - urlAnchor.href = urlAnchor.href; - s.crossDomain = originAnchor.protocol + "//" + originAnchor.host !== - urlAnchor.protocol + "//" + urlAnchor.host; - } catch ( e ) { - - // If there is an error parsing the URL, assume it is crossDomain, - // it can be rejected by the transport if it is invalid - s.crossDomain = true; - } - } - - // Convert data if not already a string - if ( s.data && s.processData && typeof s.data !== "string" ) { - s.data = jQuery.param( s.data, s.traditional ); - } - - // Apply prefilters - inspectPrefiltersOrTransports( prefilters, s, options, jqXHR ); - - // If request was aborted inside a prefilter, stop there - if ( completed ) { - return jqXHR; - } - - // We can fire global events as of now if asked to - // Don't fire events if jQuery.event is undefined in an AMD-usage scenario (#15118) - fireGlobals = jQuery.event && s.global; - - // Watch for a new set of requests - if ( fireGlobals && jQuery.active++ === 0 ) { - jQuery.event.trigger( "ajaxStart" ); - } - - // Uppercase the type - s.type = s.type.toUpperCase(); - - // Determine if request has content - s.hasContent = !rnoContent.test( s.type ); - - // Save the URL in case we're toying with the If-Modified-Since - // and/or If-None-Match header later on - // Remove hash to simplify url manipulation - cacheURL = s.url.replace( rhash, "" ); - - // More options handling for requests with no content - if ( !s.hasContent ) { - - // Remember the hash so we can put it back - uncached = s.url.slice( cacheURL.length ); - - // If data is available and should be processed, append data to url - if ( s.data && ( s.processData || typeof s.data === "string" ) ) { - cacheURL += ( rquery.test( cacheURL ) ? "&" : "?" ) + s.data; - - // #9682: remove data so that it's not used in an eventual retry - delete s.data; - } - - // Add or update anti-cache param if needed - if ( s.cache === false ) { - cacheURL = cacheURL.replace( rantiCache, "$1" ); - uncached = ( rquery.test( cacheURL ) ? "&" : "?" ) + "_=" + ( nonce.guid++ ) + - uncached; - } - - // Put hash and anti-cache on the URL that will be requested (gh-1732) - s.url = cacheURL + uncached; - - // Change '%20' to '+' if this is encoded form body content (gh-2658) - } else if ( s.data && s.processData && - ( s.contentType || "" ).indexOf( "application/x-www-form-urlencoded" ) === 0 ) { - s.data = s.data.replace( r20, "+" ); - } - - // Set the If-Modified-Since and/or If-None-Match header, if in ifModified mode. - if ( s.ifModified ) { - if ( jQuery.lastModified[ cacheURL ] ) { - jqXHR.setRequestHeader( "If-Modified-Since", jQuery.lastModified[ cacheURL ] ); - } - if ( jQuery.etag[ cacheURL ] ) { - jqXHR.setRequestHeader( "If-None-Match", jQuery.etag[ cacheURL ] ); - } - } - - // Set the correct header, if data is being sent - if ( s.data && s.hasContent && s.contentType !== false || options.contentType ) { - jqXHR.setRequestHeader( "Content-Type", s.contentType ); - } - - // Set the Accepts header for the server, depending on the dataType - jqXHR.setRequestHeader( - "Accept", - s.dataTypes[ 0 ] && s.accepts[ s.dataTypes[ 0 ] ] ? - s.accepts[ s.dataTypes[ 0 ] ] + - ( s.dataTypes[ 0 ] !== "*" ? ", " + allTypes + "; q=0.01" : "" ) : - s.accepts[ "*" ] - ); - - // Check for headers option - for ( i in s.headers ) { - jqXHR.setRequestHeader( i, s.headers[ i ] ); - } - - // Allow custom headers/mimetypes and early abort - if ( s.beforeSend && - ( s.beforeSend.call( callbackContext, jqXHR, s ) === false || completed ) ) { - - // Abort if not done already and return - return jqXHR.abort(); - } - - // Aborting is no longer a cancellation - strAbort = "abort"; - - // Install callbacks on deferreds - completeDeferred.add( s.complete ); - jqXHR.done( s.success ); - jqXHR.fail( s.error ); - - // Get transport - transport = inspectPrefiltersOrTransports( transports, s, options, jqXHR ); - - // If no transport, we auto-abort - if ( !transport ) { - done( -1, "No Transport" ); - } else { - jqXHR.readyState = 1; - - // Send global event - if ( fireGlobals ) { - globalEventContext.trigger( "ajaxSend", [ jqXHR, s ] ); - } - - // If request was aborted inside ajaxSend, stop there - if ( completed ) { - return jqXHR; - } - - // Timeout - if ( s.async && s.timeout > 0 ) { - timeoutTimer = window.setTimeout( function() { - jqXHR.abort( "timeout" ); - }, s.timeout ); - } - - try { - completed = false; - transport.send( requestHeaders, done ); - } catch ( e ) { - - // Rethrow post-completion exceptions - if ( completed ) { - throw e; - } - - // Propagate others as results - done( -1, e ); - } - } - - // Callback for when everything is done - function done( status, nativeStatusText, responses, headers ) { - var isSuccess, success, error, response, modified, - statusText = nativeStatusText; - - // Ignore repeat invocations - if ( completed ) { - return; - } - - completed = true; - - // Clear timeout if it exists - if ( timeoutTimer ) { - window.clearTimeout( timeoutTimer ); - } - - // Dereference transport for early garbage collection - // (no matter how long the jqXHR object will be used) - transport = undefined; - - // Cache response headers - responseHeadersString = headers || ""; - - // Set readyState - jqXHR.readyState = status > 0 ? 4 : 0; - - // Determine if successful - isSuccess = status >= 200 && status < 300 || status === 304; - - // Get response data - if ( responses ) { - response = ajaxHandleResponses( s, jqXHR, responses ); - } - - // Use a noop converter for missing script - if ( !isSuccess && jQuery.inArray( "script", s.dataTypes ) > -1 ) { - s.converters[ "text script" ] = function() {}; - } - - // Convert no matter what (that way responseXXX fields are always set) - response = ajaxConvert( s, response, jqXHR, isSuccess ); - - // If successful, handle type chaining - if ( isSuccess ) { - - // Set the If-Modified-Since and/or If-None-Match header, if in ifModified mode. - if ( s.ifModified ) { - modified = jqXHR.getResponseHeader( "Last-Modified" ); - if ( modified ) { - jQuery.lastModified[ cacheURL ] = modified; - } - modified = jqXHR.getResponseHeader( "etag" ); - if ( modified ) { - jQuery.etag[ cacheURL ] = modified; - } - } - - // if no content - if ( status === 204 || s.type === "HEAD" ) { - statusText = "nocontent"; - - // if not modified - } else if ( status === 304 ) { - statusText = "notmodified"; - - // If we have data, let's convert it - } else { - statusText = response.state; - success = response.data; - error = response.error; - isSuccess = !error; - } - } else { - - // Extract error from statusText and normalize for non-aborts - error = statusText; - if ( status || !statusText ) { - statusText = "error"; - if ( status < 0 ) { - status = 0; - } - } - } - - // Set data for the fake xhr object - jqXHR.status = status; - jqXHR.statusText = ( nativeStatusText || statusText ) + ""; - - // Success/Error - if ( isSuccess ) { - deferred.resolveWith( callbackContext, [ success, statusText, jqXHR ] ); - } else { - deferred.rejectWith( callbackContext, [ jqXHR, statusText, error ] ); - } - - // Status-dependent callbacks - jqXHR.statusCode( statusCode ); - statusCode = undefined; - - if ( fireGlobals ) { - globalEventContext.trigger( isSuccess ? "ajaxSuccess" : "ajaxError", - [ jqXHR, s, isSuccess ? success : error ] ); - } - - // Complete - completeDeferred.fireWith( callbackContext, [ jqXHR, statusText ] ); - - if ( fireGlobals ) { - globalEventContext.trigger( "ajaxComplete", [ jqXHR, s ] ); - - // Handle the global AJAX counter - if ( !( --jQuery.active ) ) { - jQuery.event.trigger( "ajaxStop" ); - } - } - } - - return jqXHR; - }, - - getJSON: function( url, data, callback ) { - return jQuery.get( url, data, callback, "json" ); - }, - - getScript: function( url, callback ) { - return jQuery.get( url, undefined, callback, "script" ); - } -} ); - -jQuery.each( [ "get", "post" ], function( _i, method ) { - jQuery[ method ] = function( url, data, callback, type ) { - - // Shift arguments if data argument was omitted - if ( isFunction( data ) ) { - type = type || callback; - callback = data; - data = undefined; - } - - // The url can be an options object (which then must have .url) - return jQuery.ajax( jQuery.extend( { - url: url, - type: method, - dataType: type, - data: data, - success: callback - }, jQuery.isPlainObject( url ) && url ) ); - }; -} ); - -jQuery.ajaxPrefilter( function( s ) { - var i; - for ( i in s.headers ) { - if ( i.toLowerCase() === "content-type" ) { - s.contentType = s.headers[ i ] || ""; - } - } -} ); - - -jQuery._evalUrl = function( url, options, doc ) { - return jQuery.ajax( { - url: url, - - // Make this explicit, since user can override this through ajaxSetup (#11264) - type: "GET", - dataType: "script", - cache: true, - async: false, - global: false, - - // Only evaluate the response if it is successful (gh-4126) - // dataFilter is not invoked for failure responses, so using it instead - // of the default converter is kludgy but it works. - converters: { - "text script": function() {} - }, - dataFilter: function( response ) { - jQuery.globalEval( response, options, doc ); - } - } ); -}; - - -jQuery.fn.extend( { - wrapAll: function( html ) { - var wrap; - - if ( this[ 0 ] ) { - if ( isFunction( html ) ) { - html = html.call( this[ 0 ] ); - } - - // The elements to wrap the target around - wrap = jQuery( html, this[ 0 ].ownerDocument ).eq( 0 ).clone( true ); - - if ( this[ 0 ].parentNode ) { - wrap.insertBefore( this[ 0 ] ); - } - - wrap.map( function() { - var elem = this; - - while ( elem.firstElementChild ) { - elem = elem.firstElementChild; - } - - return elem; - } ).append( this ); - } - - return this; - }, - - wrapInner: function( html ) { - if ( isFunction( html ) ) { - return this.each( function( i ) { - jQuery( this ).wrapInner( html.call( this, i ) ); - } ); - } - - return this.each( function() { - var self = jQuery( this ), - contents = self.contents(); - - if ( contents.length ) { - contents.wrapAll( html ); - - } else { - self.append( html ); - } - } ); - }, - - wrap: function( html ) { - var htmlIsFunction = isFunction( html ); - - return this.each( function( i ) { - jQuery( this ).wrapAll( htmlIsFunction ? html.call( this, i ) : html ); - } ); - }, - - unwrap: function( selector ) { - this.parent( selector ).not( "body" ).each( function() { - jQuery( this ).replaceWith( this.childNodes ); - } ); - return this; - } -} ); - - -jQuery.expr.pseudos.hidden = function( elem ) { - return !jQuery.expr.pseudos.visible( elem ); -}; -jQuery.expr.pseudos.visible = function( elem ) { - return !!( elem.offsetWidth || elem.offsetHeight || elem.getClientRects().length ); -}; - - - - -jQuery.ajaxSettings.xhr = function() { - try { - return new window.XMLHttpRequest(); - } catch ( e ) {} -}; - -var xhrSuccessStatus = { - - // File protocol always yields status code 0, assume 200 - 0: 200, - - // Support: IE <=9 only - // #1450: sometimes IE returns 1223 when it should be 204 - 1223: 204 - }, - xhrSupported = jQuery.ajaxSettings.xhr(); - -support.cors = !!xhrSupported && ( "withCredentials" in xhrSupported ); -support.ajax = xhrSupported = !!xhrSupported; - -jQuery.ajaxTransport( function( options ) { - var callback, errorCallback; - - // Cross domain only allowed if supported through XMLHttpRequest - if ( support.cors || xhrSupported && !options.crossDomain ) { - return { - send: function( headers, complete ) { - var i, - xhr = options.xhr(); - - xhr.open( - options.type, - options.url, - options.async, - options.username, - options.password - ); - - // Apply custom fields if provided - if ( options.xhrFields ) { - for ( i in options.xhrFields ) { - xhr[ i ] = options.xhrFields[ i ]; - } - } - - // Override mime type if needed - if ( options.mimeType && xhr.overrideMimeType ) { - xhr.overrideMimeType( options.mimeType ); - } - - // X-Requested-With header - // For cross-domain requests, seeing as conditions for a preflight are - // akin to a jigsaw puzzle, we simply never set it to be sure. - // (it can always be set on a per-request basis or even using ajaxSetup) - // For same-domain requests, won't change header if already provided. - if ( !options.crossDomain && !headers[ "X-Requested-With" ] ) { - headers[ "X-Requested-With" ] = "XMLHttpRequest"; - } - - // Set headers - for ( i in headers ) { - xhr.setRequestHeader( i, headers[ i ] ); - } - - // Callback - callback = function( type ) { - return function() { - if ( callback ) { - callback = errorCallback = xhr.onload = - xhr.onerror = xhr.onabort = xhr.ontimeout = - xhr.onreadystatechange = null; - - if ( type === "abort" ) { - xhr.abort(); - } else if ( type === "error" ) { - - // Support: IE <=9 only - // On a manual native abort, IE9 throws - // errors on any property access that is not readyState - if ( typeof xhr.status !== "number" ) { - complete( 0, "error" ); - } else { - complete( - - // File: protocol always yields status 0; see #8605, #14207 - xhr.status, - xhr.statusText - ); - } - } else { - complete( - xhrSuccessStatus[ xhr.status ] || xhr.status, - xhr.statusText, - - // Support: IE <=9 only - // IE9 has no XHR2 but throws on binary (trac-11426) - // For XHR2 non-text, let the caller handle it (gh-2498) - ( xhr.responseType || "text" ) !== "text" || - typeof xhr.responseText !== "string" ? - { binary: xhr.response } : - { text: xhr.responseText }, - xhr.getAllResponseHeaders() - ); - } - } - }; - }; - - // Listen to events - xhr.onload = callback(); - errorCallback = xhr.onerror = xhr.ontimeout = callback( "error" ); - - // Support: IE 9 only - // Use onreadystatechange to replace onabort - // to handle uncaught aborts - if ( xhr.onabort !== undefined ) { - xhr.onabort = errorCallback; - } else { - xhr.onreadystatechange = function() { - - // Check readyState before timeout as it changes - if ( xhr.readyState === 4 ) { - - // Allow onerror to be called first, - // but that will not handle a native abort - // Also, save errorCallback to a variable - // as xhr.onerror cannot be accessed - window.setTimeout( function() { - if ( callback ) { - errorCallback(); - } - } ); - } - }; - } - - // Create the abort callback - callback = callback( "abort" ); - - try { - - // Do send the request (this may raise an exception) - xhr.send( options.hasContent && options.data || null ); - } catch ( e ) { - - // #14683: Only rethrow if this hasn't been notified as an error yet - if ( callback ) { - throw e; - } - } - }, - - abort: function() { - if ( callback ) { - callback(); - } - } - }; - } -} ); - - - - -// Prevent auto-execution of scripts when no explicit dataType was provided (See gh-2432) -jQuery.ajaxPrefilter( function( s ) { - if ( s.crossDomain ) { - s.contents.script = false; - } -} ); - -// Install script dataType -jQuery.ajaxSetup( { - accepts: { - script: "text/javascript, application/javascript, " + - "application/ecmascript, application/x-ecmascript" - }, - contents: { - script: /\b(?:java|ecma)script\b/ - }, - converters: { - "text script": function( text ) { - jQuery.globalEval( text ); - return text; - } - } -} ); - -// Handle cache's special case and crossDomain -jQuery.ajaxPrefilter( "script", function( s ) { - if ( s.cache === undefined ) { - s.cache = false; - } - if ( s.crossDomain ) { - s.type = "GET"; - } -} ); - -// Bind script tag hack transport -jQuery.ajaxTransport( "script", function( s ) { - - // This transport only deals with cross domain or forced-by-attrs requests - if ( s.crossDomain || s.scriptAttrs ) { - var script, callback; - return { - send: function( _, complete ) { - script = jQuery( "\r\n"; - -// inject VBScript -document.write(IEBinaryToArray_ByteStr_Script); - -global.JSZipUtils._getBinaryFromXHR = function (xhr) { - var binary = xhr.responseBody; - var byteMapping = {}; - for ( var i = 0; i < 256; i++ ) { - for ( var j = 0; j < 256; j++ ) { - byteMapping[ String.fromCharCode( i + (j << 8) ) ] = - String.fromCharCode(i) + String.fromCharCode(j); - } - } - var rawBytes = IEBinaryToArray_ByteStr(binary); - var lastChr = IEBinaryToArray_ByteStr_Last(binary); - return rawBytes.replace(/[\s\S]/g, function( match ) { - return byteMapping[match]; - }) + lastChr; -}; - -// enforcing Stuk's coding style -// vim: set shiftwidth=4 softtabstop=4: - -},{}]},{},[1]) -; diff --git a/website/static/javadoc/0.8-SNAPSHOT/jquery/jszip-utils/dist/jszip-utils-ie.min.js b/website/static/javadoc/0.8-SNAPSHOT/jquery/jszip-utils/dist/jszip-utils-ie.min.js deleted file mode 100644 index 93d8bc8ef2..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/jquery/jszip-utils/dist/jszip-utils-ie.min.js +++ /dev/null @@ -1,10 +0,0 @@ -/*! - -JSZipUtils - A collection of cross-browser utilities to go along with JSZip. - - -(c) 2014 Stuart Knightley, David Duponchel -Dual licenced under the MIT license or GPLv3. See https://raw.github.com/Stuk/jszip-utils/master/LICENSE.markdown. - -*/ -!function a(b,c,d){function e(g,h){if(!c[g]){if(!b[g]){var i="function"==typeof require&&require;if(!h&&i)return i(g,!0);if(f)return f(g,!0);throw new Error("Cannot find module '"+g+"'")}var j=c[g]={exports:{}};b[g][0].call(j.exports,function(a){var c=b[g][1][a];return e(c?c:a)},j,j.exports,a,b,c,d)}return c[g].exports}for(var f="function"==typeof require&&require,g=0;g\r\n";document.write(b),a.JSZipUtils._getBinaryFromXHR=function(a){for(var b=a.responseBody,c={},d=0;256>d;d++)for(var e=0;256>e;e++)c[String.fromCharCode(d+(e<<8))]=String.fromCharCode(d)+String.fromCharCode(e);var f=IEBinaryToArray_ByteStr(b),g=IEBinaryToArray_ByteStr_Last(b);return f.replace(/[\s\S]/g,function(a){return c[a]})+g}},{}]},{},[1]); diff --git a/website/static/javadoc/0.8-SNAPSHOT/jquery/jszip-utils/dist/jszip-utils.js b/website/static/javadoc/0.8-SNAPSHOT/jquery/jszip-utils/dist/jszip-utils.js deleted file mode 100644 index 775895ec92..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/jquery/jszip-utils/dist/jszip-utils.js +++ /dev/null @@ -1,118 +0,0 @@ -/*! - -JSZipUtils - A collection of cross-browser utilities to go along with JSZip. - - -(c) 2014 Stuart Knightley, David Duponchel -Dual licenced under the MIT license or GPLv3. See https://raw.github.com/Stuk/jszip-utils/master/LICENSE.markdown. - -*/ -!function(e){"object"==typeof exports?module.exports=e():"function"==typeof define&&define.amd?define(e):"undefined"!=typeof window?window.JSZipUtils=e():"undefined"!=typeof global?global.JSZipUtils=e():"undefined"!=typeof self&&(self.JSZipUtils=e())}(function(){var define,module,exports;return (function e(t,n,r){function s(o,u){if(!n[o]){if(!t[o]){var a=typeof require=="function"&&require;if(!u&&a)return a(o,!0);if(i)return i(o,!0);throw new Error("Cannot find module '"+o+"'")}var f=n[o]={exports:{}};t[o][0].call(f.exports,function(e){var n=t[o][1][e];return s(n?n:e)},f,f.exports,e,t,n,r)}return n[o].exports}var i=typeof require=="function"&&require;for(var o=0;o - -(c) 2014 Stuart Knightley, David Duponchel -Dual licenced under the MIT license or GPLv3. See https://raw.github.com/Stuk/jszip-utils/master/LICENSE.markdown. - -*/ -!function(a){"object"==typeof exports?module.exports=a():"function"==typeof define&&define.amd?define(a):"undefined"!=typeof window?window.JSZipUtils=a():"undefined"!=typeof global?global.JSZipUtils=a():"undefined"!=typeof self&&(self.JSZipUtils=a())}(function(){return function a(b,c,d){function e(g,h){if(!c[g]){if(!b[g]){var i="function"==typeof require&&require;if(!h&&i)return i(g,!0);if(f)return f(g,!0);throw new Error("Cannot find module '"+g+"'")}var j=c[g]={exports:{}};b[g][0].call(j.exports,function(a){var c=b[g][1][a];return e(c?c:a)},j,j.exports,a,b,c,d)}return c[g].exports}for(var f="function"==typeof require&&require,g=0;g - -(c) 2009-2016 Stuart Knightley -Dual licenced under the MIT license or GPLv3. See https://raw.github.com/Stuk/jszip/master/LICENSE.markdown. - -JSZip uses the library pako released under the MIT license : -https://github.com/nodeca/pako/blob/master/LICENSE -*/ - -(function(f){if(typeof exports==="object"&&typeof module!=="undefined"){module.exports=f()}else if(typeof define==="function"&&define.amd){define([],f)}else{var g;if(typeof window!=="undefined"){g=window}else if(typeof global!=="undefined"){g=global}else if(typeof self!=="undefined"){g=self}else{g=this}g.JSZip = f()}})(function(){var define,module,exports;return (function e(t,n,r){function s(o,u){if(!n[o]){if(!t[o]){var a=typeof require=="function"&&require;if(!u&&a)return a(o,!0);if(i)return i(o,!0);var f=new Error("Cannot find module '"+o+"'");throw f.code="MODULE_NOT_FOUND",f}var l=n[o]={exports:{}};t[o][0].call(l.exports,function(e){var n=t[o][1][e];return s(n?n:e)},l,l.exports,e,t,n,r)}return n[o].exports}var i=typeof require=="function"&&require;for(var o=0;o> 2; - enc2 = ((chr1 & 3) << 4) | (chr2 >> 4); - enc3 = remainingBytes > 1 ? (((chr2 & 15) << 2) | (chr3 >> 6)) : 64; - enc4 = remainingBytes > 2 ? (chr3 & 63) : 64; - - output.push(_keyStr.charAt(enc1) + _keyStr.charAt(enc2) + _keyStr.charAt(enc3) + _keyStr.charAt(enc4)); - - } - - return output.join(""); -}; - -// public method for decoding -exports.decode = function(input) { - var chr1, chr2, chr3; - var enc1, enc2, enc3, enc4; - var i = 0, resultIndex = 0; - - var dataUrlPrefix = "data:"; - - if (input.substr(0, dataUrlPrefix.length) === dataUrlPrefix) { - // This is a common error: people give a data url - // (data:image/png;base64,iVBOR...) with a {base64: true} and - // wonders why things don't work. - // We can detect that the string input looks like a data url but we - // *can't* be sure it is one: removing everything up to the comma would - // be too dangerous. - throw new Error("Invalid base64 input, it looks like a data url."); - } - - input = input.replace(/[^A-Za-z0-9\+\/\=]/g, ""); - - var totalLength = input.length * 3 / 4; - if(input.charAt(input.length - 1) === _keyStr.charAt(64)) { - totalLength--; - } - if(input.charAt(input.length - 2) === _keyStr.charAt(64)) { - totalLength--; - } - if (totalLength % 1 !== 0) { - // totalLength is not an integer, the length does not match a valid - // base64 content. That can happen if: - // - the input is not a base64 content - // - the input is *almost* a base64 content, with a extra chars at the - // beginning or at the end - // - the input uses a base64 variant (base64url for example) - throw new Error("Invalid base64 input, bad content length."); - } - var output; - if (support.uint8array) { - output = new Uint8Array(totalLength|0); - } else { - output = new Array(totalLength|0); - } - - while (i < input.length) { - - enc1 = _keyStr.indexOf(input.charAt(i++)); - enc2 = _keyStr.indexOf(input.charAt(i++)); - enc3 = _keyStr.indexOf(input.charAt(i++)); - enc4 = _keyStr.indexOf(input.charAt(i++)); - - chr1 = (enc1 << 2) | (enc2 >> 4); - chr2 = ((enc2 & 15) << 4) | (enc3 >> 2); - chr3 = ((enc3 & 3) << 6) | enc4; - - output[resultIndex++] = chr1; - - if (enc3 !== 64) { - output[resultIndex++] = chr2; - } - if (enc4 !== 64) { - output[resultIndex++] = chr3; - } - - } - - return output; -}; - -},{"./support":30,"./utils":32}],2:[function(require,module,exports){ -'use strict'; - -var external = require("./external"); -var DataWorker = require('./stream/DataWorker'); -var Crc32Probe = require('./stream/Crc32Probe'); -var DataLengthProbe = require('./stream/DataLengthProbe'); - -/** - * Represent a compressed object, with everything needed to decompress it. - * @constructor - * @param {number} compressedSize the size of the data compressed. - * @param {number} uncompressedSize the size of the data after decompression. - * @param {number} crc32 the crc32 of the decompressed file. - * @param {object} compression the type of compression, see lib/compressions.js. - * @param {String|ArrayBuffer|Uint8Array|Buffer} data the compressed data. - */ -function CompressedObject(compressedSize, uncompressedSize, crc32, compression, data) { - this.compressedSize = compressedSize; - this.uncompressedSize = uncompressedSize; - this.crc32 = crc32; - this.compression = compression; - this.compressedContent = data; -} - -CompressedObject.prototype = { - /** - * Create a worker to get the uncompressed content. - * @return {GenericWorker} the worker. - */ - getContentWorker: function () { - var worker = new DataWorker(external.Promise.resolve(this.compressedContent)) - .pipe(this.compression.uncompressWorker()) - .pipe(new DataLengthProbe("data_length")); - - var that = this; - worker.on("end", function () { - if (this.streamInfo['data_length'] !== that.uncompressedSize) { - throw new Error("Bug : uncompressed data size mismatch"); - } - }); - return worker; - }, - /** - * Create a worker to get the compressed content. - * @return {GenericWorker} the worker. - */ - getCompressedWorker: function () { - return new DataWorker(external.Promise.resolve(this.compressedContent)) - .withStreamInfo("compressedSize", this.compressedSize) - .withStreamInfo("uncompressedSize", this.uncompressedSize) - .withStreamInfo("crc32", this.crc32) - .withStreamInfo("compression", this.compression) - ; - } -}; - -/** - * Chain the given worker with other workers to compress the content with the - * given compression. - * @param {GenericWorker} uncompressedWorker the worker to pipe. - * @param {Object} compression the compression object. - * @param {Object} compressionOptions the options to use when compressing. - * @return {GenericWorker} the new worker compressing the content. - */ -CompressedObject.createWorkerFrom = function (uncompressedWorker, compression, compressionOptions) { - return uncompressedWorker - .pipe(new Crc32Probe()) - .pipe(new DataLengthProbe("uncompressedSize")) - .pipe(compression.compressWorker(compressionOptions)) - .pipe(new DataLengthProbe("compressedSize")) - .withStreamInfo("compression", compression); -}; - -module.exports = CompressedObject; - -},{"./external":6,"./stream/Crc32Probe":25,"./stream/DataLengthProbe":26,"./stream/DataWorker":27}],3:[function(require,module,exports){ -'use strict'; - -var GenericWorker = require("./stream/GenericWorker"); - -exports.STORE = { - magic: "\x00\x00", - compressWorker : function (compressionOptions) { - return new GenericWorker("STORE compression"); - }, - uncompressWorker : function () { - return new GenericWorker("STORE decompression"); - } -}; -exports.DEFLATE = require('./flate'); - -},{"./flate":7,"./stream/GenericWorker":28}],4:[function(require,module,exports){ -'use strict'; - -var utils = require('./utils'); - -/** - * The following functions come from pako, from pako/lib/zlib/crc32.js - * released under the MIT license, see pako https://github.com/nodeca/pako/ - */ - -// Use ordinary array, since untyped makes no boost here -function makeTable() { - var c, table = []; - - for(var n =0; n < 256; n++){ - c = n; - for(var k =0; k < 8; k++){ - c = ((c&1) ? (0xEDB88320 ^ (c >>> 1)) : (c >>> 1)); - } - table[n] = c; - } - - return table; -} - -// Create table on load. Just 255 signed longs. Not a problem. -var crcTable = makeTable(); - - -function crc32(crc, buf, len, pos) { - var t = crcTable, end = pos + len; - - crc = crc ^ (-1); - - for (var i = pos; i < end; i++ ) { - crc = (crc >>> 8) ^ t[(crc ^ buf[i]) & 0xFF]; - } - - return (crc ^ (-1)); // >>> 0; -} - -// That's all for the pako functions. - -/** - * Compute the crc32 of a string. - * This is almost the same as the function crc32, but for strings. Using the - * same function for the two use cases leads to horrible performances. - * @param {Number} crc the starting value of the crc. - * @param {String} str the string to use. - * @param {Number} len the length of the string. - * @param {Number} pos the starting position for the crc32 computation. - * @return {Number} the computed crc32. - */ -function crc32str(crc, str, len, pos) { - var t = crcTable, end = pos + len; - - crc = crc ^ (-1); - - for (var i = pos; i < end; i++ ) { - crc = (crc >>> 8) ^ t[(crc ^ str.charCodeAt(i)) & 0xFF]; - } - - return (crc ^ (-1)); // >>> 0; -} - -module.exports = function crc32wrapper(input, crc) { - if (typeof input === "undefined" || !input.length) { - return 0; - } - - var isArray = utils.getTypeOf(input) !== "string"; - - if(isArray) { - return crc32(crc|0, input, input.length, 0); - } else { - return crc32str(crc|0, input, input.length, 0); - } -}; - -},{"./utils":32}],5:[function(require,module,exports){ -'use strict'; -exports.base64 = false; -exports.binary = false; -exports.dir = false; -exports.createFolders = true; -exports.date = null; -exports.compression = null; -exports.compressionOptions = null; -exports.comment = null; -exports.unixPermissions = null; -exports.dosPermissions = null; - -},{}],6:[function(require,module,exports){ -/* global Promise */ -'use strict'; - -// load the global object first: -// - it should be better integrated in the system (unhandledRejection in node) -// - the environment may have a custom Promise implementation (see zone.js) -var ES6Promise = null; -if (typeof Promise !== "undefined") { - ES6Promise = Promise; -} else { - ES6Promise = require("lie"); -} - -/** - * Let the user use/change some implementations. - */ -module.exports = { - Promise: ES6Promise -}; - -},{"lie":37}],7:[function(require,module,exports){ -'use strict'; -var USE_TYPEDARRAY = (typeof Uint8Array !== 'undefined') && (typeof Uint16Array !== 'undefined') && (typeof Uint32Array !== 'undefined'); - -var pako = require("pako"); -var utils = require("./utils"); -var GenericWorker = require("./stream/GenericWorker"); - -var ARRAY_TYPE = USE_TYPEDARRAY ? "uint8array" : "array"; - -exports.magic = "\x08\x00"; - -/** - * Create a worker that uses pako to inflate/deflate. - * @constructor - * @param {String} action the name of the pako function to call : either "Deflate" or "Inflate". - * @param {Object} options the options to use when (de)compressing. - */ -function FlateWorker(action, options) { - GenericWorker.call(this, "FlateWorker/" + action); - - this._pako = null; - this._pakoAction = action; - this._pakoOptions = options; - // the `meta` object from the last chunk received - // this allow this worker to pass around metadata - this.meta = {}; -} - -utils.inherits(FlateWorker, GenericWorker); - -/** - * @see GenericWorker.processChunk - */ -FlateWorker.prototype.processChunk = function (chunk) { - this.meta = chunk.meta; - if (this._pako === null) { - this._createPako(); - } - this._pako.push(utils.transformTo(ARRAY_TYPE, chunk.data), false); -}; - -/** - * @see GenericWorker.flush - */ -FlateWorker.prototype.flush = function () { - GenericWorker.prototype.flush.call(this); - if (this._pako === null) { - this._createPako(); - } - this._pako.push([], true); -}; -/** - * @see GenericWorker.cleanUp - */ -FlateWorker.prototype.cleanUp = function () { - GenericWorker.prototype.cleanUp.call(this); - this._pako = null; -}; - -/** - * Create the _pako object. - * TODO: lazy-loading this object isn't the best solution but it's the - * quickest. The best solution is to lazy-load the worker list. See also the - * issue #446. - */ -FlateWorker.prototype._createPako = function () { - this._pako = new pako[this._pakoAction]({ - raw: true, - level: this._pakoOptions.level || -1 // default compression - }); - var self = this; - this._pako.onData = function(data) { - self.push({ - data : data, - meta : self.meta - }); - }; -}; - -exports.compressWorker = function (compressionOptions) { - return new FlateWorker("Deflate", compressionOptions); -}; -exports.uncompressWorker = function () { - return new FlateWorker("Inflate", {}); -}; - -},{"./stream/GenericWorker":28,"./utils":32,"pako":38}],8:[function(require,module,exports){ -'use strict'; - -var utils = require('../utils'); -var GenericWorker = require('../stream/GenericWorker'); -var utf8 = require('../utf8'); -var crc32 = require('../crc32'); -var signature = require('../signature'); - -/** - * Transform an integer into a string in hexadecimal. - * @private - * @param {number} dec the number to convert. - * @param {number} bytes the number of bytes to generate. - * @returns {string} the result. - */ -var decToHex = function(dec, bytes) { - var hex = "", i; - for (i = 0; i < bytes; i++) { - hex += String.fromCharCode(dec & 0xff); - dec = dec >>> 8; - } - return hex; -}; - -/** - * Generate the UNIX part of the external file attributes. - * @param {Object} unixPermissions the unix permissions or null. - * @param {Boolean} isDir true if the entry is a directory, false otherwise. - * @return {Number} a 32 bit integer. - * - * adapted from http://unix.stackexchange.com/questions/14705/the-zip-formats-external-file-attribute : - * - * TTTTsstrwxrwxrwx0000000000ADVSHR - * ^^^^____________________________ file type, see zipinfo.c (UNX_*) - * ^^^_________________________ setuid, setgid, sticky - * ^^^^^^^^^________________ permissions - * ^^^^^^^^^^______ not used ? - * ^^^^^^ DOS attribute bits : Archive, Directory, Volume label, System file, Hidden, Read only - */ -var generateUnixExternalFileAttr = function (unixPermissions, isDir) { - - var result = unixPermissions; - if (!unixPermissions) { - // I can't use octal values in strict mode, hence the hexa. - // 040775 => 0x41fd - // 0100664 => 0x81b4 - result = isDir ? 0x41fd : 0x81b4; - } - return (result & 0xFFFF) << 16; -}; - -/** - * Generate the DOS part of the external file attributes. - * @param {Object} dosPermissions the dos permissions or null. - * @param {Boolean} isDir true if the entry is a directory, false otherwise. - * @return {Number} a 32 bit integer. - * - * Bit 0 Read-Only - * Bit 1 Hidden - * Bit 2 System - * Bit 3 Volume Label - * Bit 4 Directory - * Bit 5 Archive - */ -var generateDosExternalFileAttr = function (dosPermissions, isDir) { - - // the dir flag is already set for compatibility - return (dosPermissions || 0) & 0x3F; -}; - -/** - * Generate the various parts used in the construction of the final zip file. - * @param {Object} streamInfo the hash with information about the compressed file. - * @param {Boolean} streamedContent is the content streamed ? - * @param {Boolean} streamingEnded is the stream finished ? - * @param {number} offset the current offset from the start of the zip file. - * @param {String} platform let's pretend we are this platform (change platform dependents fields) - * @param {Function} encodeFileName the function to encode the file name / comment. - * @return {Object} the zip parts. - */ -var generateZipParts = function(streamInfo, streamedContent, streamingEnded, offset, platform, encodeFileName) { - var file = streamInfo['file'], - compression = streamInfo['compression'], - useCustomEncoding = encodeFileName !== utf8.utf8encode, - encodedFileName = utils.transformTo("string", encodeFileName(file.name)), - utfEncodedFileName = utils.transformTo("string", utf8.utf8encode(file.name)), - comment = file.comment, - encodedComment = utils.transformTo("string", encodeFileName(comment)), - utfEncodedComment = utils.transformTo("string", utf8.utf8encode(comment)), - useUTF8ForFileName = utfEncodedFileName.length !== file.name.length, - useUTF8ForComment = utfEncodedComment.length !== comment.length, - dosTime, - dosDate, - extraFields = "", - unicodePathExtraField = "", - unicodeCommentExtraField = "", - dir = file.dir, - date = file.date; - - - var dataInfo = { - crc32 : 0, - compressedSize : 0, - uncompressedSize : 0 - }; - - // if the content is streamed, the sizes/crc32 are only available AFTER - // the end of the stream. - if (!streamedContent || streamingEnded) { - dataInfo.crc32 = streamInfo['crc32']; - dataInfo.compressedSize = streamInfo['compressedSize']; - dataInfo.uncompressedSize = streamInfo['uncompressedSize']; - } - - var bitflag = 0; - if (streamedContent) { - // Bit 3: the sizes/crc32 are set to zero in the local header. - // The correct values are put in the data descriptor immediately - // following the compressed data. - bitflag |= 0x0008; - } - if (!useCustomEncoding && (useUTF8ForFileName || useUTF8ForComment)) { - // Bit 11: Language encoding flag (EFS). - bitflag |= 0x0800; - } - - - var extFileAttr = 0; - var versionMadeBy = 0; - if (dir) { - // dos or unix, we set the dos dir flag - extFileAttr |= 0x00010; - } - if(platform === "UNIX") { - versionMadeBy = 0x031E; // UNIX, version 3.0 - extFileAttr |= generateUnixExternalFileAttr(file.unixPermissions, dir); - } else { // DOS or other, fallback to DOS - versionMadeBy = 0x0014; // DOS, version 2.0 - extFileAttr |= generateDosExternalFileAttr(file.dosPermissions, dir); - } - - // date - // @see http://www.delorie.com/djgpp/doc/rbinter/it/52/13.html - // @see http://www.delorie.com/djgpp/doc/rbinter/it/65/16.html - // @see http://www.delorie.com/djgpp/doc/rbinter/it/66/16.html - - dosTime = date.getUTCHours(); - dosTime = dosTime << 6; - dosTime = dosTime | date.getUTCMinutes(); - dosTime = dosTime << 5; - dosTime = dosTime | date.getUTCSeconds() / 2; - - dosDate = date.getUTCFullYear() - 1980; - dosDate = dosDate << 4; - dosDate = dosDate | (date.getUTCMonth() + 1); - dosDate = dosDate << 5; - dosDate = dosDate | date.getUTCDate(); - - if (useUTF8ForFileName) { - // set the unicode path extra field. unzip needs at least one extra - // field to correctly handle unicode path, so using the path is as good - // as any other information. This could improve the situation with - // other archive managers too. - // This field is usually used without the utf8 flag, with a non - // unicode path in the header (winrar, winzip). This helps (a bit) - // with the messy Windows' default compressed folders feature but - // breaks on p7zip which doesn't seek the unicode path extra field. - // So for now, UTF-8 everywhere ! - unicodePathExtraField = - // Version - decToHex(1, 1) + - // NameCRC32 - decToHex(crc32(encodedFileName), 4) + - // UnicodeName - utfEncodedFileName; - - extraFields += - // Info-ZIP Unicode Path Extra Field - "\x75\x70" + - // size - decToHex(unicodePathExtraField.length, 2) + - // content - unicodePathExtraField; - } - - if(useUTF8ForComment) { - - unicodeCommentExtraField = - // Version - decToHex(1, 1) + - // CommentCRC32 - decToHex(crc32(encodedComment), 4) + - // UnicodeName - utfEncodedComment; - - extraFields += - // Info-ZIP Unicode Path Extra Field - "\x75\x63" + - // size - decToHex(unicodeCommentExtraField.length, 2) + - // content - unicodeCommentExtraField; - } - - var header = ""; - - // version needed to extract - header += "\x0A\x00"; - // general purpose bit flag - header += decToHex(bitflag, 2); - // compression method - header += compression.magic; - // last mod file time - header += decToHex(dosTime, 2); - // last mod file date - header += decToHex(dosDate, 2); - // crc-32 - header += decToHex(dataInfo.crc32, 4); - // compressed size - header += decToHex(dataInfo.compressedSize, 4); - // uncompressed size - header += decToHex(dataInfo.uncompressedSize, 4); - // file name length - header += decToHex(encodedFileName.length, 2); - // extra field length - header += decToHex(extraFields.length, 2); - - - var fileRecord = signature.LOCAL_FILE_HEADER + header + encodedFileName + extraFields; - - var dirRecord = signature.CENTRAL_FILE_HEADER + - // version made by (00: DOS) - decToHex(versionMadeBy, 2) + - // file header (common to file and central directory) - header + - // file comment length - decToHex(encodedComment.length, 2) + - // disk number start - "\x00\x00" + - // internal file attributes TODO - "\x00\x00" + - // external file attributes - decToHex(extFileAttr, 4) + - // relative offset of local header - decToHex(offset, 4) + - // file name - encodedFileName + - // extra field - extraFields + - // file comment - encodedComment; - - return { - fileRecord: fileRecord, - dirRecord: dirRecord - }; -}; - -/** - * Generate the EOCD record. - * @param {Number} entriesCount the number of entries in the zip file. - * @param {Number} centralDirLength the length (in bytes) of the central dir. - * @param {Number} localDirLength the length (in bytes) of the local dir. - * @param {String} comment the zip file comment as a binary string. - * @param {Function} encodeFileName the function to encode the comment. - * @return {String} the EOCD record. - */ -var generateCentralDirectoryEnd = function (entriesCount, centralDirLength, localDirLength, comment, encodeFileName) { - var dirEnd = ""; - var encodedComment = utils.transformTo("string", encodeFileName(comment)); - - // end of central dir signature - dirEnd = signature.CENTRAL_DIRECTORY_END + - // number of this disk - "\x00\x00" + - // number of the disk with the start of the central directory - "\x00\x00" + - // total number of entries in the central directory on this disk - decToHex(entriesCount, 2) + - // total number of entries in the central directory - decToHex(entriesCount, 2) + - // size of the central directory 4 bytes - decToHex(centralDirLength, 4) + - // offset of start of central directory with respect to the starting disk number - decToHex(localDirLength, 4) + - // .ZIP file comment length - decToHex(encodedComment.length, 2) + - // .ZIP file comment - encodedComment; - - return dirEnd; -}; - -/** - * Generate data descriptors for a file entry. - * @param {Object} streamInfo the hash generated by a worker, containing information - * on the file entry. - * @return {String} the data descriptors. - */ -var generateDataDescriptors = function (streamInfo) { - var descriptor = ""; - descriptor = signature.DATA_DESCRIPTOR + - // crc-32 4 bytes - decToHex(streamInfo['crc32'], 4) + - // compressed size 4 bytes - decToHex(streamInfo['compressedSize'], 4) + - // uncompressed size 4 bytes - decToHex(streamInfo['uncompressedSize'], 4); - - return descriptor; -}; - - -/** - * A worker to concatenate other workers to create a zip file. - * @param {Boolean} streamFiles `true` to stream the content of the files, - * `false` to accumulate it. - * @param {String} comment the comment to use. - * @param {String} platform the platform to use, "UNIX" or "DOS". - * @param {Function} encodeFileName the function to encode file names and comments. - */ -function ZipFileWorker(streamFiles, comment, platform, encodeFileName) { - GenericWorker.call(this, "ZipFileWorker"); - // The number of bytes written so far. This doesn't count accumulated chunks. - this.bytesWritten = 0; - // The comment of the zip file - this.zipComment = comment; - // The platform "generating" the zip file. - this.zipPlatform = platform; - // the function to encode file names and comments. - this.encodeFileName = encodeFileName; - // Should we stream the content of the files ? - this.streamFiles = streamFiles; - // If `streamFiles` is false, we will need to accumulate the content of the - // files to calculate sizes / crc32 (and write them *before* the content). - // This boolean indicates if we are accumulating chunks (it will change a lot - // during the lifetime of this worker). - this.accumulate = false; - // The buffer receiving chunks when accumulating content. - this.contentBuffer = []; - // The list of generated directory records. - this.dirRecords = []; - // The offset (in bytes) from the beginning of the zip file for the current source. - this.currentSourceOffset = 0; - // The total number of entries in this zip file. - this.entriesCount = 0; - // the name of the file currently being added, null when handling the end of the zip file. - // Used for the emitted metadata. - this.currentFile = null; - - - - this._sources = []; -} -utils.inherits(ZipFileWorker, GenericWorker); - -/** - * @see GenericWorker.push - */ -ZipFileWorker.prototype.push = function (chunk) { - - var currentFilePercent = chunk.meta.percent || 0; - var entriesCount = this.entriesCount; - var remainingFiles = this._sources.length; - - if(this.accumulate) { - this.contentBuffer.push(chunk); - } else { - this.bytesWritten += chunk.data.length; - - GenericWorker.prototype.push.call(this, { - data : chunk.data, - meta : { - currentFile : this.currentFile, - percent : entriesCount ? (currentFilePercent + 100 * (entriesCount - remainingFiles - 1)) / entriesCount : 100 - } - }); - } -}; - -/** - * The worker started a new source (an other worker). - * @param {Object} streamInfo the streamInfo object from the new source. - */ -ZipFileWorker.prototype.openedSource = function (streamInfo) { - this.currentSourceOffset = this.bytesWritten; - this.currentFile = streamInfo['file'].name; - - var streamedContent = this.streamFiles && !streamInfo['file'].dir; - - // don't stream folders (because they don't have any content) - if(streamedContent) { - var record = generateZipParts(streamInfo, streamedContent, false, this.currentSourceOffset, this.zipPlatform, this.encodeFileName); - this.push({ - data : record.fileRecord, - meta : {percent:0} - }); - } else { - // we need to wait for the whole file before pushing anything - this.accumulate = true; - } -}; - -/** - * The worker finished a source (an other worker). - * @param {Object} streamInfo the streamInfo object from the finished source. - */ -ZipFileWorker.prototype.closedSource = function (streamInfo) { - this.accumulate = false; - var streamedContent = this.streamFiles && !streamInfo['file'].dir; - var record = generateZipParts(streamInfo, streamedContent, true, this.currentSourceOffset, this.zipPlatform, this.encodeFileName); - - this.dirRecords.push(record.dirRecord); - if(streamedContent) { - // after the streamed file, we put data descriptors - this.push({ - data : generateDataDescriptors(streamInfo), - meta : {percent:100} - }); - } else { - // the content wasn't streamed, we need to push everything now - // first the file record, then the content - this.push({ - data : record.fileRecord, - meta : {percent:0} - }); - while(this.contentBuffer.length) { - this.push(this.contentBuffer.shift()); - } - } - this.currentFile = null; -}; - -/** - * @see GenericWorker.flush - */ -ZipFileWorker.prototype.flush = function () { - - var localDirLength = this.bytesWritten; - for(var i = 0; i < this.dirRecords.length; i++) { - this.push({ - data : this.dirRecords[i], - meta : {percent:100} - }); - } - var centralDirLength = this.bytesWritten - localDirLength; - - var dirEnd = generateCentralDirectoryEnd(this.dirRecords.length, centralDirLength, localDirLength, this.zipComment, this.encodeFileName); - - this.push({ - data : dirEnd, - meta : {percent:100} - }); -}; - -/** - * Prepare the next source to be read. - */ -ZipFileWorker.prototype.prepareNextSource = function () { - this.previous = this._sources.shift(); - this.openedSource(this.previous.streamInfo); - if (this.isPaused) { - this.previous.pause(); - } else { - this.previous.resume(); - } -}; - -/** - * @see GenericWorker.registerPrevious - */ -ZipFileWorker.prototype.registerPrevious = function (previous) { - this._sources.push(previous); - var self = this; - - previous.on('data', function (chunk) { - self.processChunk(chunk); - }); - previous.on('end', function () { - self.closedSource(self.previous.streamInfo); - if(self._sources.length) { - self.prepareNextSource(); - } else { - self.end(); - } - }); - previous.on('error', function (e) { - self.error(e); - }); - return this; -}; - -/** - * @see GenericWorker.resume - */ -ZipFileWorker.prototype.resume = function () { - if(!GenericWorker.prototype.resume.call(this)) { - return false; - } - - if (!this.previous && this._sources.length) { - this.prepareNextSource(); - return true; - } - if (!this.previous && !this._sources.length && !this.generatedError) { - this.end(); - return true; - } -}; - -/** - * @see GenericWorker.error - */ -ZipFileWorker.prototype.error = function (e) { - var sources = this._sources; - if(!GenericWorker.prototype.error.call(this, e)) { - return false; - } - for(var i = 0; i < sources.length; i++) { - try { - sources[i].error(e); - } catch(e) { - // the `error` exploded, nothing to do - } - } - return true; -}; - -/** - * @see GenericWorker.lock - */ -ZipFileWorker.prototype.lock = function () { - GenericWorker.prototype.lock.call(this); - var sources = this._sources; - for(var i = 0; i < sources.length; i++) { - sources[i].lock(); - } -}; - -module.exports = ZipFileWorker; - -},{"../crc32":4,"../signature":23,"../stream/GenericWorker":28,"../utf8":31,"../utils":32}],9:[function(require,module,exports){ -'use strict'; - -var compressions = require('../compressions'); -var ZipFileWorker = require('./ZipFileWorker'); - -/** - * Find the compression to use. - * @param {String} fileCompression the compression defined at the file level, if any. - * @param {String} zipCompression the compression defined at the load() level. - * @return {Object} the compression object to use. - */ -var getCompression = function (fileCompression, zipCompression) { - - var compressionName = fileCompression || zipCompression; - var compression = compressions[compressionName]; - if (!compression) { - throw new Error(compressionName + " is not a valid compression method !"); - } - return compression; -}; - -/** - * Create a worker to generate a zip file. - * @param {JSZip} zip the JSZip instance at the right root level. - * @param {Object} options to generate the zip file. - * @param {String} comment the comment to use. - */ -exports.generateWorker = function (zip, options, comment) { - - var zipFileWorker = new ZipFileWorker(options.streamFiles, comment, options.platform, options.encodeFileName); - var entriesCount = 0; - try { - - zip.forEach(function (relativePath, file) { - entriesCount++; - var compression = getCompression(file.options.compression, options.compression); - var compressionOptions = file.options.compressionOptions || options.compressionOptions || {}; - var dir = file.dir, date = file.date; - - file._compressWorker(compression, compressionOptions) - .withStreamInfo("file", { - name : relativePath, - dir : dir, - date : date, - comment : file.comment || "", - unixPermissions : file.unixPermissions, - dosPermissions : file.dosPermissions - }) - .pipe(zipFileWorker); - }); - zipFileWorker.entriesCount = entriesCount; - } catch (e) { - zipFileWorker.error(e); - } - - return zipFileWorker; -}; - -},{"../compressions":3,"./ZipFileWorker":8}],10:[function(require,module,exports){ -'use strict'; - -/** - * Representation a of zip file in js - * @constructor - */ -function JSZip() { - // if this constructor is used without `new`, it adds `new` before itself: - if(!(this instanceof JSZip)) { - return new JSZip(); - } - - if(arguments.length) { - throw new Error("The constructor with parameters has been removed in JSZip 3.0, please check the upgrade guide."); - } - - // object containing the files : - // { - // "folder/" : {...}, - // "folder/data.txt" : {...} - // } - // NOTE: we use a null prototype because we do not - // want filenames like "toString" coming from a zip file - // to overwrite methods and attributes in a normal Object. - this.files = Object.create(null); - - this.comment = null; - - // Where we are in the hierarchy - this.root = ""; - this.clone = function() { - var newObj = new JSZip(); - for (var i in this) { - if (typeof this[i] !== "function") { - newObj[i] = this[i]; - } - } - return newObj; - }; -} -JSZip.prototype = require('./object'); -JSZip.prototype.loadAsync = require('./load'); -JSZip.support = require('./support'); -JSZip.defaults = require('./defaults'); - -// TODO find a better way to handle this version, -// a require('package.json').version doesn't work with webpack, see #327 -JSZip.version = "3.7.1"; - -JSZip.loadAsync = function (content, options) { - return new JSZip().loadAsync(content, options); -}; - -JSZip.external = require("./external"); -module.exports = JSZip; - -},{"./defaults":5,"./external":6,"./load":11,"./object":15,"./support":30}],11:[function(require,module,exports){ -'use strict'; -var utils = require('./utils'); -var external = require("./external"); -var utf8 = require('./utf8'); -var ZipEntries = require('./zipEntries'); -var Crc32Probe = require('./stream/Crc32Probe'); -var nodejsUtils = require("./nodejsUtils"); - -/** - * Check the CRC32 of an entry. - * @param {ZipEntry} zipEntry the zip entry to check. - * @return {Promise} the result. - */ -function checkEntryCRC32(zipEntry) { - return new external.Promise(function (resolve, reject) { - var worker = zipEntry.decompressed.getContentWorker().pipe(new Crc32Probe()); - worker.on("error", function (e) { - reject(e); - }) - .on("end", function () { - if (worker.streamInfo.crc32 !== zipEntry.decompressed.crc32) { - reject(new Error("Corrupted zip : CRC32 mismatch")); - } else { - resolve(); - } - }) - .resume(); - }); -} - -module.exports = function (data, options) { - var zip = this; - options = utils.extend(options || {}, { - base64: false, - checkCRC32: false, - optimizedBinaryString: false, - createFolders: false, - decodeFileName: utf8.utf8decode - }); - - if (nodejsUtils.isNode && nodejsUtils.isStream(data)) { - return external.Promise.reject(new Error("JSZip can't accept a stream when loading a zip file.")); - } - - return utils.prepareContent("the loaded zip file", data, true, options.optimizedBinaryString, options.base64) - .then(function (data) { - var zipEntries = new ZipEntries(options); - zipEntries.load(data); - return zipEntries; - }).then(function checkCRC32(zipEntries) { - var promises = [external.Promise.resolve(zipEntries)]; - var files = zipEntries.files; - if (options.checkCRC32) { - for (var i = 0; i < files.length; i++) { - promises.push(checkEntryCRC32(files[i])); - } - } - return external.Promise.all(promises); - }).then(function addFiles(results) { - var zipEntries = results.shift(); - var files = zipEntries.files; - for (var i = 0; i < files.length; i++) { - var input = files[i]; - zip.file(input.fileNameStr, input.decompressed, { - binary: true, - optimizedBinaryString: true, - date: input.date, - dir: input.dir, - comment: input.fileCommentStr.length ? input.fileCommentStr : null, - unixPermissions: input.unixPermissions, - dosPermissions: input.dosPermissions, - createFolders: options.createFolders - }); - } - if (zipEntries.zipComment.length) { - zip.comment = zipEntries.zipComment; - } - - return zip; - }); -}; - -},{"./external":6,"./nodejsUtils":14,"./stream/Crc32Probe":25,"./utf8":31,"./utils":32,"./zipEntries":33}],12:[function(require,module,exports){ -"use strict"; - -var utils = require('../utils'); -var GenericWorker = require('../stream/GenericWorker'); - -/** - * A worker that use a nodejs stream as source. - * @constructor - * @param {String} filename the name of the file entry for this stream. - * @param {Readable} stream the nodejs stream. - */ -function NodejsStreamInputAdapter(filename, stream) { - GenericWorker.call(this, "Nodejs stream input adapter for " + filename); - this._upstreamEnded = false; - this._bindStream(stream); -} - -utils.inherits(NodejsStreamInputAdapter, GenericWorker); - -/** - * Prepare the stream and bind the callbacks on it. - * Do this ASAP on node 0.10 ! A lazy binding doesn't always work. - * @param {Stream} stream the nodejs stream to use. - */ -NodejsStreamInputAdapter.prototype._bindStream = function (stream) { - var self = this; - this._stream = stream; - stream.pause(); - stream - .on("data", function (chunk) { - self.push({ - data: chunk, - meta : { - percent : 0 - } - }); - }) - .on("error", function (e) { - if(self.isPaused) { - this.generatedError = e; - } else { - self.error(e); - } - }) - .on("end", function () { - if(self.isPaused) { - self._upstreamEnded = true; - } else { - self.end(); - } - }); -}; -NodejsStreamInputAdapter.prototype.pause = function () { - if(!GenericWorker.prototype.pause.call(this)) { - return false; - } - this._stream.pause(); - return true; -}; -NodejsStreamInputAdapter.prototype.resume = function () { - if(!GenericWorker.prototype.resume.call(this)) { - return false; - } - - if(this._upstreamEnded) { - this.end(); - } else { - this._stream.resume(); - } - - return true; -}; - -module.exports = NodejsStreamInputAdapter; - -},{"../stream/GenericWorker":28,"../utils":32}],13:[function(require,module,exports){ -'use strict'; - -var Readable = require('readable-stream').Readable; - -var utils = require('../utils'); -utils.inherits(NodejsStreamOutputAdapter, Readable); - -/** -* A nodejs stream using a worker as source. -* @see the SourceWrapper in http://nodejs.org/api/stream.html -* @constructor -* @param {StreamHelper} helper the helper wrapping the worker -* @param {Object} options the nodejs stream options -* @param {Function} updateCb the update callback. -*/ -function NodejsStreamOutputAdapter(helper, options, updateCb) { - Readable.call(this, options); - this._helper = helper; - - var self = this; - helper.on("data", function (data, meta) { - if (!self.push(data)) { - self._helper.pause(); - } - if(updateCb) { - updateCb(meta); - } - }) - .on("error", function(e) { - self.emit('error', e); - }) - .on("end", function () { - self.push(null); - }); -} - - -NodejsStreamOutputAdapter.prototype._read = function() { - this._helper.resume(); -}; - -module.exports = NodejsStreamOutputAdapter; - -},{"../utils":32,"readable-stream":16}],14:[function(require,module,exports){ -'use strict'; - -module.exports = { - /** - * True if this is running in Nodejs, will be undefined in a browser. - * In a browser, browserify won't include this file and the whole module - * will be resolved an empty object. - */ - isNode : typeof Buffer !== "undefined", - /** - * Create a new nodejs Buffer from an existing content. - * @param {Object} data the data to pass to the constructor. - * @param {String} encoding the encoding to use. - * @return {Buffer} a new Buffer. - */ - newBufferFrom: function(data, encoding) { - if (Buffer.from && Buffer.from !== Uint8Array.from) { - return Buffer.from(data, encoding); - } else { - if (typeof data === "number") { - // Safeguard for old Node.js versions. On newer versions, - // Buffer.from(number) / Buffer(number, encoding) already throw. - throw new Error("The \"data\" argument must not be a number"); - } - return new Buffer(data, encoding); - } - }, - /** - * Create a new nodejs Buffer with the specified size. - * @param {Integer} size the size of the buffer. - * @return {Buffer} a new Buffer. - */ - allocBuffer: function (size) { - if (Buffer.alloc) { - return Buffer.alloc(size); - } else { - var buf = new Buffer(size); - buf.fill(0); - return buf; - } - }, - /** - * Find out if an object is a Buffer. - * @param {Object} b the object to test. - * @return {Boolean} true if the object is a Buffer, false otherwise. - */ - isBuffer : function(b){ - return Buffer.isBuffer(b); - }, - - isStream : function (obj) { - return obj && - typeof obj.on === "function" && - typeof obj.pause === "function" && - typeof obj.resume === "function"; - } -}; - -},{}],15:[function(require,module,exports){ -'use strict'; -var utf8 = require('./utf8'); -var utils = require('./utils'); -var GenericWorker = require('./stream/GenericWorker'); -var StreamHelper = require('./stream/StreamHelper'); -var defaults = require('./defaults'); -var CompressedObject = require('./compressedObject'); -var ZipObject = require('./zipObject'); -var generate = require("./generate"); -var nodejsUtils = require("./nodejsUtils"); -var NodejsStreamInputAdapter = require("./nodejs/NodejsStreamInputAdapter"); - - -/** - * Add a file in the current folder. - * @private - * @param {string} name the name of the file - * @param {String|ArrayBuffer|Uint8Array|Buffer} data the data of the file - * @param {Object} originalOptions the options of the file - * @return {Object} the new file. - */ -var fileAdd = function(name, data, originalOptions) { - // be sure sub folders exist - var dataType = utils.getTypeOf(data), - parent; - - - /* - * Correct options. - */ - - var o = utils.extend(originalOptions || {}, defaults); - o.date = o.date || new Date(); - if (o.compression !== null) { - o.compression = o.compression.toUpperCase(); - } - - if (typeof o.unixPermissions === "string") { - o.unixPermissions = parseInt(o.unixPermissions, 8); - } - - // UNX_IFDIR 0040000 see zipinfo.c - if (o.unixPermissions && (o.unixPermissions & 0x4000)) { - o.dir = true; - } - // Bit 4 Directory - if (o.dosPermissions && (o.dosPermissions & 0x0010)) { - o.dir = true; - } - - if (o.dir) { - name = forceTrailingSlash(name); - } - if (o.createFolders && (parent = parentFolder(name))) { - folderAdd.call(this, parent, true); - } - - var isUnicodeString = dataType === "string" && o.binary === false && o.base64 === false; - if (!originalOptions || typeof originalOptions.binary === "undefined") { - o.binary = !isUnicodeString; - } - - - var isCompressedEmpty = (data instanceof CompressedObject) && data.uncompressedSize === 0; - - if (isCompressedEmpty || o.dir || !data || data.length === 0) { - o.base64 = false; - o.binary = true; - data = ""; - o.compression = "STORE"; - dataType = "string"; - } - - /* - * Convert content to fit. - */ - - var zipObjectContent = null; - if (data instanceof CompressedObject || data instanceof GenericWorker) { - zipObjectContent = data; - } else if (nodejsUtils.isNode && nodejsUtils.isStream(data)) { - zipObjectContent = new NodejsStreamInputAdapter(name, data); - } else { - zipObjectContent = utils.prepareContent(name, data, o.binary, o.optimizedBinaryString, o.base64); - } - - var object = new ZipObject(name, zipObjectContent, o); - this.files[name] = object; - /* - TODO: we can't throw an exception because we have async promises - (we can have a promise of a Date() for example) but returning a - promise is useless because file(name, data) returns the JSZip - object for chaining. Should we break that to allow the user - to catch the error ? - - return external.Promise.resolve(zipObjectContent) - .then(function () { - return object; - }); - */ -}; - -/** - * Find the parent folder of the path. - * @private - * @param {string} path the path to use - * @return {string} the parent folder, or "" - */ -var parentFolder = function (path) { - if (path.slice(-1) === '/') { - path = path.substring(0, path.length - 1); - } - var lastSlash = path.lastIndexOf('/'); - return (lastSlash > 0) ? path.substring(0, lastSlash) : ""; -}; - -/** - * Returns the path with a slash at the end. - * @private - * @param {String} path the path to check. - * @return {String} the path with a trailing slash. - */ -var forceTrailingSlash = function(path) { - // Check the name ends with a / - if (path.slice(-1) !== "/") { - path += "/"; // IE doesn't like substr(-1) - } - return path; -}; - -/** - * Add a (sub) folder in the current folder. - * @private - * @param {string} name the folder's name - * @param {boolean=} [createFolders] If true, automatically create sub - * folders. Defaults to false. - * @return {Object} the new folder. - */ -var folderAdd = function(name, createFolders) { - createFolders = (typeof createFolders !== 'undefined') ? createFolders : defaults.createFolders; - - name = forceTrailingSlash(name); - - // Does this folder already exist? - if (!this.files[name]) { - fileAdd.call(this, name, null, { - dir: true, - createFolders: createFolders - }); - } - return this.files[name]; -}; - -/** -* Cross-window, cross-Node-context regular expression detection -* @param {Object} object Anything -* @return {Boolean} true if the object is a regular expression, -* false otherwise -*/ -function isRegExp(object) { - return Object.prototype.toString.call(object) === "[object RegExp]"; -} - -// return the actual prototype of JSZip -var out = { - /** - * @see loadAsync - */ - load: function() { - throw new Error("This method has been removed in JSZip 3.0, please check the upgrade guide."); - }, - - - /** - * Call a callback function for each entry at this folder level. - * @param {Function} cb the callback function: - * function (relativePath, file) {...} - * It takes 2 arguments : the relative path and the file. - */ - forEach: function(cb) { - var filename, relativePath, file; - /* jshint ignore:start */ - // ignore warning about unwanted properties because this.files is a null prototype object - for (filename in this.files) { - file = this.files[filename]; - relativePath = filename.slice(this.root.length, filename.length); - if (relativePath && filename.slice(0, this.root.length) === this.root) { // the file is in the current root - cb(relativePath, file); // TODO reverse the parameters ? need to be clean AND consistent with the filter search fn... - } - } - /* jshint ignore:end */ - }, - - /** - * Filter nested files/folders with the specified function. - * @param {Function} search the predicate to use : - * function (relativePath, file) {...} - * It takes 2 arguments : the relative path and the file. - * @return {Array} An array of matching elements. - */ - filter: function(search) { - var result = []; - this.forEach(function (relativePath, entry) { - if (search(relativePath, entry)) { // the file matches the function - result.push(entry); - } - - }); - return result; - }, - - /** - * Add a file to the zip file, or search a file. - * @param {string|RegExp} name The name of the file to add (if data is defined), - * the name of the file to find (if no data) or a regex to match files. - * @param {String|ArrayBuffer|Uint8Array|Buffer} data The file data, either raw or base64 encoded - * @param {Object} o File options - * @return {JSZip|Object|Array} this JSZip object (when adding a file), - * a file (when searching by string) or an array of files (when searching by regex). - */ - file: function(name, data, o) { - if (arguments.length === 1) { - if (isRegExp(name)) { - var regexp = name; - return this.filter(function(relativePath, file) { - return !file.dir && regexp.test(relativePath); - }); - } - else { // text - var obj = this.files[this.root + name]; - if (obj && !obj.dir) { - return obj; - } else { - return null; - } - } - } - else { // more than one argument : we have data ! - name = this.root + name; - fileAdd.call(this, name, data, o); - } - return this; - }, - - /** - * Add a directory to the zip file, or search. - * @param {String|RegExp} arg The name of the directory to add, or a regex to search folders. - * @return {JSZip} an object with the new directory as the root, or an array containing matching folders. - */ - folder: function(arg) { - if (!arg) { - return this; - } - - if (isRegExp(arg)) { - return this.filter(function(relativePath, file) { - return file.dir && arg.test(relativePath); - }); - } - - // else, name is a new folder - var name = this.root + arg; - var newFolder = folderAdd.call(this, name); - - // Allow chaining by returning a new object with this folder as the root - var ret = this.clone(); - ret.root = newFolder.name; - return ret; - }, - - /** - * Delete a file, or a directory and all sub-files, from the zip - * @param {string} name the name of the file to delete - * @return {JSZip} this JSZip object - */ - remove: function(name) { - name = this.root + name; - var file = this.files[name]; - if (!file) { - // Look for any folders - if (name.slice(-1) !== "/") { - name += "/"; - } - file = this.files[name]; - } - - if (file && !file.dir) { - // file - delete this.files[name]; - } else { - // maybe a folder, delete recursively - var kids = this.filter(function(relativePath, file) { - return file.name.slice(0, name.length) === name; - }); - for (var i = 0; i < kids.length; i++) { - delete this.files[kids[i].name]; - } - } - - return this; - }, - - /** - * Generate the complete zip file - * @param {Object} options the options to generate the zip file : - * - compression, "STORE" by default. - * - type, "base64" by default. Values are : string, base64, uint8array, arraybuffer, blob. - * @return {String|Uint8Array|ArrayBuffer|Buffer|Blob} the zip file - */ - generate: function(options) { - throw new Error("This method has been removed in JSZip 3.0, please check the upgrade guide."); - }, - - /** - * Generate the complete zip file as an internal stream. - * @param {Object} options the options to generate the zip file : - * - compression, "STORE" by default. - * - type, "base64" by default. Values are : string, base64, uint8array, arraybuffer, blob. - * @return {StreamHelper} the streamed zip file. - */ - generateInternalStream: function(options) { - var worker, opts = {}; - try { - opts = utils.extend(options || {}, { - streamFiles: false, - compression: "STORE", - compressionOptions : null, - type: "", - platform: "DOS", - comment: null, - mimeType: 'application/zip', - encodeFileName: utf8.utf8encode - }); - - opts.type = opts.type.toLowerCase(); - opts.compression = opts.compression.toUpperCase(); - - // "binarystring" is preferred but the internals use "string". - if(opts.type === "binarystring") { - opts.type = "string"; - } - - if (!opts.type) { - throw new Error("No output type specified."); - } - - utils.checkSupport(opts.type); - - // accept nodejs `process.platform` - if( - opts.platform === 'darwin' || - opts.platform === 'freebsd' || - opts.platform === 'linux' || - opts.platform === 'sunos' - ) { - opts.platform = "UNIX"; - } - if (opts.platform === 'win32') { - opts.platform = "DOS"; - } - - var comment = opts.comment || this.comment || ""; - worker = generate.generateWorker(this, opts, comment); - } catch (e) { - worker = new GenericWorker("error"); - worker.error(e); - } - return new StreamHelper(worker, opts.type || "string", opts.mimeType); - }, - /** - * Generate the complete zip file asynchronously. - * @see generateInternalStream - */ - generateAsync: function(options, onUpdate) { - return this.generateInternalStream(options).accumulate(onUpdate); - }, - /** - * Generate the complete zip file asynchronously. - * @see generateInternalStream - */ - generateNodeStream: function(options, onUpdate) { - options = options || {}; - if (!options.type) { - options.type = "nodebuffer"; - } - return this.generateInternalStream(options).toNodejsStream(onUpdate); - } -}; -module.exports = out; - -},{"./compressedObject":2,"./defaults":5,"./generate":9,"./nodejs/NodejsStreamInputAdapter":12,"./nodejsUtils":14,"./stream/GenericWorker":28,"./stream/StreamHelper":29,"./utf8":31,"./utils":32,"./zipObject":35}],16:[function(require,module,exports){ -/* - * This file is used by module bundlers (browserify/webpack/etc) when - * including a stream implementation. We use "readable-stream" to get a - * consistent behavior between nodejs versions but bundlers often have a shim - * for "stream". Using this shim greatly improve the compatibility and greatly - * reduce the final size of the bundle (only one stream implementation, not - * two). - */ -module.exports = require("stream"); - -},{"stream":undefined}],17:[function(require,module,exports){ -'use strict'; -var DataReader = require('./DataReader'); -var utils = require('../utils'); - -function ArrayReader(data) { - DataReader.call(this, data); - for(var i = 0; i < this.data.length; i++) { - data[i] = data[i] & 0xFF; - } -} -utils.inherits(ArrayReader, DataReader); -/** - * @see DataReader.byteAt - */ -ArrayReader.prototype.byteAt = function(i) { - return this.data[this.zero + i]; -}; -/** - * @see DataReader.lastIndexOfSignature - */ -ArrayReader.prototype.lastIndexOfSignature = function(sig) { - var sig0 = sig.charCodeAt(0), - sig1 = sig.charCodeAt(1), - sig2 = sig.charCodeAt(2), - sig3 = sig.charCodeAt(3); - for (var i = this.length - 4; i >= 0; --i) { - if (this.data[i] === sig0 && this.data[i + 1] === sig1 && this.data[i + 2] === sig2 && this.data[i + 3] === sig3) { - return i - this.zero; - } - } - - return -1; -}; -/** - * @see DataReader.readAndCheckSignature - */ -ArrayReader.prototype.readAndCheckSignature = function (sig) { - var sig0 = sig.charCodeAt(0), - sig1 = sig.charCodeAt(1), - sig2 = sig.charCodeAt(2), - sig3 = sig.charCodeAt(3), - data = this.readData(4); - return sig0 === data[0] && sig1 === data[1] && sig2 === data[2] && sig3 === data[3]; -}; -/** - * @see DataReader.readData - */ -ArrayReader.prototype.readData = function(size) { - this.checkOffset(size); - if(size === 0) { - return []; - } - var result = this.data.slice(this.zero + this.index, this.zero + this.index + size); - this.index += size; - return result; -}; -module.exports = ArrayReader; - -},{"../utils":32,"./DataReader":18}],18:[function(require,module,exports){ -'use strict'; -var utils = require('../utils'); - -function DataReader(data) { - this.data = data; // type : see implementation - this.length = data.length; - this.index = 0; - this.zero = 0; -} -DataReader.prototype = { - /** - * Check that the offset will not go too far. - * @param {string} offset the additional offset to check. - * @throws {Error} an Error if the offset is out of bounds. - */ - checkOffset: function(offset) { - this.checkIndex(this.index + offset); - }, - /** - * Check that the specified index will not be too far. - * @param {string} newIndex the index to check. - * @throws {Error} an Error if the index is out of bounds. - */ - checkIndex: function(newIndex) { - if (this.length < this.zero + newIndex || newIndex < 0) { - throw new Error("End of data reached (data length = " + this.length + ", asked index = " + (newIndex) + "). Corrupted zip ?"); - } - }, - /** - * Change the index. - * @param {number} newIndex The new index. - * @throws {Error} if the new index is out of the data. - */ - setIndex: function(newIndex) { - this.checkIndex(newIndex); - this.index = newIndex; - }, - /** - * Skip the next n bytes. - * @param {number} n the number of bytes to skip. - * @throws {Error} if the new index is out of the data. - */ - skip: function(n) { - this.setIndex(this.index + n); - }, - /** - * Get the byte at the specified index. - * @param {number} i the index to use. - * @return {number} a byte. - */ - byteAt: function(i) { - // see implementations - }, - /** - * Get the next number with a given byte size. - * @param {number} size the number of bytes to read. - * @return {number} the corresponding number. - */ - readInt: function(size) { - var result = 0, - i; - this.checkOffset(size); - for (i = this.index + size - 1; i >= this.index; i--) { - result = (result << 8) + this.byteAt(i); - } - this.index += size; - return result; - }, - /** - * Get the next string with a given byte size. - * @param {number} size the number of bytes to read. - * @return {string} the corresponding string. - */ - readString: function(size) { - return utils.transformTo("string", this.readData(size)); - }, - /** - * Get raw data without conversion, bytes. - * @param {number} size the number of bytes to read. - * @return {Object} the raw data, implementation specific. - */ - readData: function(size) { - // see implementations - }, - /** - * Find the last occurrence of a zip signature (4 bytes). - * @param {string} sig the signature to find. - * @return {number} the index of the last occurrence, -1 if not found. - */ - lastIndexOfSignature: function(sig) { - // see implementations - }, - /** - * Read the signature (4 bytes) at the current position and compare it with sig. - * @param {string} sig the expected signature - * @return {boolean} true if the signature matches, false otherwise. - */ - readAndCheckSignature: function(sig) { - // see implementations - }, - /** - * Get the next date. - * @return {Date} the date. - */ - readDate: function() { - var dostime = this.readInt(4); - return new Date(Date.UTC( - ((dostime >> 25) & 0x7f) + 1980, // year - ((dostime >> 21) & 0x0f) - 1, // month - (dostime >> 16) & 0x1f, // day - (dostime >> 11) & 0x1f, // hour - (dostime >> 5) & 0x3f, // minute - (dostime & 0x1f) << 1)); // second - } -}; -module.exports = DataReader; - -},{"../utils":32}],19:[function(require,module,exports){ -'use strict'; -var Uint8ArrayReader = require('./Uint8ArrayReader'); -var utils = require('../utils'); - -function NodeBufferReader(data) { - Uint8ArrayReader.call(this, data); -} -utils.inherits(NodeBufferReader, Uint8ArrayReader); - -/** - * @see DataReader.readData - */ -NodeBufferReader.prototype.readData = function(size) { - this.checkOffset(size); - var result = this.data.slice(this.zero + this.index, this.zero + this.index + size); - this.index += size; - return result; -}; -module.exports = NodeBufferReader; - -},{"../utils":32,"./Uint8ArrayReader":21}],20:[function(require,module,exports){ -'use strict'; -var DataReader = require('./DataReader'); -var utils = require('../utils'); - -function StringReader(data) { - DataReader.call(this, data); -} -utils.inherits(StringReader, DataReader); -/** - * @see DataReader.byteAt - */ -StringReader.prototype.byteAt = function(i) { - return this.data.charCodeAt(this.zero + i); -}; -/** - * @see DataReader.lastIndexOfSignature - */ -StringReader.prototype.lastIndexOfSignature = function(sig) { - return this.data.lastIndexOf(sig) - this.zero; -}; -/** - * @see DataReader.readAndCheckSignature - */ -StringReader.prototype.readAndCheckSignature = function (sig) { - var data = this.readData(4); - return sig === data; -}; -/** - * @see DataReader.readData - */ -StringReader.prototype.readData = function(size) { - this.checkOffset(size); - // this will work because the constructor applied the "& 0xff" mask. - var result = this.data.slice(this.zero + this.index, this.zero + this.index + size); - this.index += size; - return result; -}; -module.exports = StringReader; - -},{"../utils":32,"./DataReader":18}],21:[function(require,module,exports){ -'use strict'; -var ArrayReader = require('./ArrayReader'); -var utils = require('../utils'); - -function Uint8ArrayReader(data) { - ArrayReader.call(this, data); -} -utils.inherits(Uint8ArrayReader, ArrayReader); -/** - * @see DataReader.readData - */ -Uint8ArrayReader.prototype.readData = function(size) { - this.checkOffset(size); - if(size === 0) { - // in IE10, when using subarray(idx, idx), we get the array [0x00] instead of []. - return new Uint8Array(0); - } - var result = this.data.subarray(this.zero + this.index, this.zero + this.index + size); - this.index += size; - return result; -}; -module.exports = Uint8ArrayReader; - -},{"../utils":32,"./ArrayReader":17}],22:[function(require,module,exports){ -'use strict'; - -var utils = require('../utils'); -var support = require('../support'); -var ArrayReader = require('./ArrayReader'); -var StringReader = require('./StringReader'); -var NodeBufferReader = require('./NodeBufferReader'); -var Uint8ArrayReader = require('./Uint8ArrayReader'); - -/** - * Create a reader adapted to the data. - * @param {String|ArrayBuffer|Uint8Array|Buffer} data the data to read. - * @return {DataReader} the data reader. - */ -module.exports = function (data) { - var type = utils.getTypeOf(data); - utils.checkSupport(type); - if (type === "string" && !support.uint8array) { - return new StringReader(data); - } - if (type === "nodebuffer") { - return new NodeBufferReader(data); - } - if (support.uint8array) { - return new Uint8ArrayReader(utils.transformTo("uint8array", data)); - } - return new ArrayReader(utils.transformTo("array", data)); -}; - -},{"../support":30,"../utils":32,"./ArrayReader":17,"./NodeBufferReader":19,"./StringReader":20,"./Uint8ArrayReader":21}],23:[function(require,module,exports){ -'use strict'; -exports.LOCAL_FILE_HEADER = "PK\x03\x04"; -exports.CENTRAL_FILE_HEADER = "PK\x01\x02"; -exports.CENTRAL_DIRECTORY_END = "PK\x05\x06"; -exports.ZIP64_CENTRAL_DIRECTORY_LOCATOR = "PK\x06\x07"; -exports.ZIP64_CENTRAL_DIRECTORY_END = "PK\x06\x06"; -exports.DATA_DESCRIPTOR = "PK\x07\x08"; - -},{}],24:[function(require,module,exports){ -'use strict'; - -var GenericWorker = require('./GenericWorker'); -var utils = require('../utils'); - -/** - * A worker which convert chunks to a specified type. - * @constructor - * @param {String} destType the destination type. - */ -function ConvertWorker(destType) { - GenericWorker.call(this, "ConvertWorker to " + destType); - this.destType = destType; -} -utils.inherits(ConvertWorker, GenericWorker); - -/** - * @see GenericWorker.processChunk - */ -ConvertWorker.prototype.processChunk = function (chunk) { - this.push({ - data : utils.transformTo(this.destType, chunk.data), - meta : chunk.meta - }); -}; -module.exports = ConvertWorker; - -},{"../utils":32,"./GenericWorker":28}],25:[function(require,module,exports){ -'use strict'; - -var GenericWorker = require('./GenericWorker'); -var crc32 = require('../crc32'); -var utils = require('../utils'); - -/** - * A worker which calculate the crc32 of the data flowing through. - * @constructor - */ -function Crc32Probe() { - GenericWorker.call(this, "Crc32Probe"); - this.withStreamInfo("crc32", 0); -} -utils.inherits(Crc32Probe, GenericWorker); - -/** - * @see GenericWorker.processChunk - */ -Crc32Probe.prototype.processChunk = function (chunk) { - this.streamInfo.crc32 = crc32(chunk.data, this.streamInfo.crc32 || 0); - this.push(chunk); -}; -module.exports = Crc32Probe; - -},{"../crc32":4,"../utils":32,"./GenericWorker":28}],26:[function(require,module,exports){ -'use strict'; - -var utils = require('../utils'); -var GenericWorker = require('./GenericWorker'); - -/** - * A worker which calculate the total length of the data flowing through. - * @constructor - * @param {String} propName the name used to expose the length - */ -function DataLengthProbe(propName) { - GenericWorker.call(this, "DataLengthProbe for " + propName); - this.propName = propName; - this.withStreamInfo(propName, 0); -} -utils.inherits(DataLengthProbe, GenericWorker); - -/** - * @see GenericWorker.processChunk - */ -DataLengthProbe.prototype.processChunk = function (chunk) { - if(chunk) { - var length = this.streamInfo[this.propName] || 0; - this.streamInfo[this.propName] = length + chunk.data.length; - } - GenericWorker.prototype.processChunk.call(this, chunk); -}; -module.exports = DataLengthProbe; - - -},{"../utils":32,"./GenericWorker":28}],27:[function(require,module,exports){ -'use strict'; - -var utils = require('../utils'); -var GenericWorker = require('./GenericWorker'); - -// the size of the generated chunks -// TODO expose this as a public variable -var DEFAULT_BLOCK_SIZE = 16 * 1024; - -/** - * A worker that reads a content and emits chunks. - * @constructor - * @param {Promise} dataP the promise of the data to split - */ -function DataWorker(dataP) { - GenericWorker.call(this, "DataWorker"); - var self = this; - this.dataIsReady = false; - this.index = 0; - this.max = 0; - this.data = null; - this.type = ""; - - this._tickScheduled = false; - - dataP.then(function (data) { - self.dataIsReady = true; - self.data = data; - self.max = data && data.length || 0; - self.type = utils.getTypeOf(data); - if(!self.isPaused) { - self._tickAndRepeat(); - } - }, function (e) { - self.error(e); - }); -} - -utils.inherits(DataWorker, GenericWorker); - -/** - * @see GenericWorker.cleanUp - */ -DataWorker.prototype.cleanUp = function () { - GenericWorker.prototype.cleanUp.call(this); - this.data = null; -}; - -/** - * @see GenericWorker.resume - */ -DataWorker.prototype.resume = function () { - if(!GenericWorker.prototype.resume.call(this)) { - return false; - } - - if (!this._tickScheduled && this.dataIsReady) { - this._tickScheduled = true; - utils.delay(this._tickAndRepeat, [], this); - } - return true; -}; - -/** - * Trigger a tick a schedule an other call to this function. - */ -DataWorker.prototype._tickAndRepeat = function() { - this._tickScheduled = false; - if(this.isPaused || this.isFinished) { - return; - } - this._tick(); - if(!this.isFinished) { - utils.delay(this._tickAndRepeat, [], this); - this._tickScheduled = true; - } -}; - -/** - * Read and push a chunk. - */ -DataWorker.prototype._tick = function() { - - if(this.isPaused || this.isFinished) { - return false; - } - - var size = DEFAULT_BLOCK_SIZE; - var data = null, nextIndex = Math.min(this.max, this.index + size); - if (this.index >= this.max) { - // EOF - return this.end(); - } else { - switch(this.type) { - case "string": - data = this.data.substring(this.index, nextIndex); - break; - case "uint8array": - data = this.data.subarray(this.index, nextIndex); - break; - case "array": - case "nodebuffer": - data = this.data.slice(this.index, nextIndex); - break; - } - this.index = nextIndex; - return this.push({ - data : data, - meta : { - percent : this.max ? this.index / this.max * 100 : 0 - } - }); - } -}; - -module.exports = DataWorker; - -},{"../utils":32,"./GenericWorker":28}],28:[function(require,module,exports){ -'use strict'; - -/** - * A worker that does nothing but passing chunks to the next one. This is like - * a nodejs stream but with some differences. On the good side : - * - it works on IE 6-9 without any issue / polyfill - * - it weights less than the full dependencies bundled with browserify - * - it forwards errors (no need to declare an error handler EVERYWHERE) - * - * A chunk is an object with 2 attributes : `meta` and `data`. The former is an - * object containing anything (`percent` for example), see each worker for more - * details. The latter is the real data (String, Uint8Array, etc). - * - * @constructor - * @param {String} name the name of the stream (mainly used for debugging purposes) - */ -function GenericWorker(name) { - // the name of the worker - this.name = name || "default"; - // an object containing metadata about the workers chain - this.streamInfo = {}; - // an error which happened when the worker was paused - this.generatedError = null; - // an object containing metadata to be merged by this worker into the general metadata - this.extraStreamInfo = {}; - // true if the stream is paused (and should not do anything), false otherwise - this.isPaused = true; - // true if the stream is finished (and should not do anything), false otherwise - this.isFinished = false; - // true if the stream is locked to prevent further structure updates (pipe), false otherwise - this.isLocked = false; - // the event listeners - this._listeners = { - 'data':[], - 'end':[], - 'error':[] - }; - // the previous worker, if any - this.previous = null; -} - -GenericWorker.prototype = { - /** - * Push a chunk to the next workers. - * @param {Object} chunk the chunk to push - */ - push : function (chunk) { - this.emit("data", chunk); - }, - /** - * End the stream. - * @return {Boolean} true if this call ended the worker, false otherwise. - */ - end : function () { - if (this.isFinished) { - return false; - } - - this.flush(); - try { - this.emit("end"); - this.cleanUp(); - this.isFinished = true; - } catch (e) { - this.emit("error", e); - } - return true; - }, - /** - * End the stream with an error. - * @param {Error} e the error which caused the premature end. - * @return {Boolean} true if this call ended the worker with an error, false otherwise. - */ - error : function (e) { - if (this.isFinished) { - return false; - } - - if(this.isPaused) { - this.generatedError = e; - } else { - this.isFinished = true; - - this.emit("error", e); - - // in the workers chain exploded in the middle of the chain, - // the error event will go downward but we also need to notify - // workers upward that there has been an error. - if(this.previous) { - this.previous.error(e); - } - - this.cleanUp(); - } - return true; - }, - /** - * Add a callback on an event. - * @param {String} name the name of the event (data, end, error) - * @param {Function} listener the function to call when the event is triggered - * @return {GenericWorker} the current object for chainability - */ - on : function (name, listener) { - this._listeners[name].push(listener); - return this; - }, - /** - * Clean any references when a worker is ending. - */ - cleanUp : function () { - this.streamInfo = this.generatedError = this.extraStreamInfo = null; - this._listeners = []; - }, - /** - * Trigger an event. This will call registered callback with the provided arg. - * @param {String} name the name of the event (data, end, error) - * @param {Object} arg the argument to call the callback with. - */ - emit : function (name, arg) { - if (this._listeners[name]) { - for(var i = 0; i < this._listeners[name].length; i++) { - this._listeners[name][i].call(this, arg); - } - } - }, - /** - * Chain a worker with an other. - * @param {Worker} next the worker receiving events from the current one. - * @return {worker} the next worker for chainability - */ - pipe : function (next) { - return next.registerPrevious(this); - }, - /** - * Same as `pipe` in the other direction. - * Using an API with `pipe(next)` is very easy. - * Implementing the API with the point of view of the next one registering - * a source is easier, see the ZipFileWorker. - * @param {Worker} previous the previous worker, sending events to this one - * @return {Worker} the current worker for chainability - */ - registerPrevious : function (previous) { - if (this.isLocked) { - throw new Error("The stream '" + this + "' has already been used."); - } - - // sharing the streamInfo... - this.streamInfo = previous.streamInfo; - // ... and adding our own bits - this.mergeStreamInfo(); - this.previous = previous; - var self = this; - previous.on('data', function (chunk) { - self.processChunk(chunk); - }); - previous.on('end', function () { - self.end(); - }); - previous.on('error', function (e) { - self.error(e); - }); - return this; - }, - /** - * Pause the stream so it doesn't send events anymore. - * @return {Boolean} true if this call paused the worker, false otherwise. - */ - pause : function () { - if(this.isPaused || this.isFinished) { - return false; - } - this.isPaused = true; - - if(this.previous) { - this.previous.pause(); - } - return true; - }, - /** - * Resume a paused stream. - * @return {Boolean} true if this call resumed the worker, false otherwise. - */ - resume : function () { - if(!this.isPaused || this.isFinished) { - return false; - } - this.isPaused = false; - - // if true, the worker tried to resume but failed - var withError = false; - if(this.generatedError) { - this.error(this.generatedError); - withError = true; - } - if(this.previous) { - this.previous.resume(); - } - - return !withError; - }, - /** - * Flush any remaining bytes as the stream is ending. - */ - flush : function () {}, - /** - * Process a chunk. This is usually the method overridden. - * @param {Object} chunk the chunk to process. - */ - processChunk : function(chunk) { - this.push(chunk); - }, - /** - * Add a key/value to be added in the workers chain streamInfo once activated. - * @param {String} key the key to use - * @param {Object} value the associated value - * @return {Worker} the current worker for chainability - */ - withStreamInfo : function (key, value) { - this.extraStreamInfo[key] = value; - this.mergeStreamInfo(); - return this; - }, - /** - * Merge this worker's streamInfo into the chain's streamInfo. - */ - mergeStreamInfo : function () { - for(var key in this.extraStreamInfo) { - if (!this.extraStreamInfo.hasOwnProperty(key)) { - continue; - } - this.streamInfo[key] = this.extraStreamInfo[key]; - } - }, - - /** - * Lock the stream to prevent further updates on the workers chain. - * After calling this method, all calls to pipe will fail. - */ - lock: function () { - if (this.isLocked) { - throw new Error("The stream '" + this + "' has already been used."); - } - this.isLocked = true; - if (this.previous) { - this.previous.lock(); - } - }, - - /** - * - * Pretty print the workers chain. - */ - toString : function () { - var me = "Worker " + this.name; - if (this.previous) { - return this.previous + " -> " + me; - } else { - return me; - } - } -}; - -module.exports = GenericWorker; - -},{}],29:[function(require,module,exports){ -'use strict'; - -var utils = require('../utils'); -var ConvertWorker = require('./ConvertWorker'); -var GenericWorker = require('./GenericWorker'); -var base64 = require('../base64'); -var support = require("../support"); -var external = require("../external"); - -var NodejsStreamOutputAdapter = null; -if (support.nodestream) { - try { - NodejsStreamOutputAdapter = require('../nodejs/NodejsStreamOutputAdapter'); - } catch(e) {} -} - -/** - * Apply the final transformation of the data. If the user wants a Blob for - * example, it's easier to work with an U8intArray and finally do the - * ArrayBuffer/Blob conversion. - * @param {String} type the name of the final type - * @param {String|Uint8Array|Buffer} content the content to transform - * @param {String} mimeType the mime type of the content, if applicable. - * @return {String|Uint8Array|ArrayBuffer|Buffer|Blob} the content in the right format. - */ -function transformZipOutput(type, content, mimeType) { - switch(type) { - case "blob" : - return utils.newBlob(utils.transformTo("arraybuffer", content), mimeType); - case "base64" : - return base64.encode(content); - default : - return utils.transformTo(type, content); - } -} - -/** - * Concatenate an array of data of the given type. - * @param {String} type the type of the data in the given array. - * @param {Array} dataArray the array containing the data chunks to concatenate - * @return {String|Uint8Array|Buffer} the concatenated data - * @throws Error if the asked type is unsupported - */ -function concat (type, dataArray) { - var i, index = 0, res = null, totalLength = 0; - for(i = 0; i < dataArray.length; i++) { - totalLength += dataArray[i].length; - } - switch(type) { - case "string": - return dataArray.join(""); - case "array": - return Array.prototype.concat.apply([], dataArray); - case "uint8array": - res = new Uint8Array(totalLength); - for(i = 0; i < dataArray.length; i++) { - res.set(dataArray[i], index); - index += dataArray[i].length; - } - return res; - case "nodebuffer": - return Buffer.concat(dataArray); - default: - throw new Error("concat : unsupported type '" + type + "'"); - } -} - -/** - * Listen a StreamHelper, accumulate its content and concatenate it into a - * complete block. - * @param {StreamHelper} helper the helper to use. - * @param {Function} updateCallback a callback called on each update. Called - * with one arg : - * - the metadata linked to the update received. - * @return Promise the promise for the accumulation. - */ -function accumulate(helper, updateCallback) { - return new external.Promise(function (resolve, reject){ - var dataArray = []; - var chunkType = helper._internalType, - resultType = helper._outputType, - mimeType = helper._mimeType; - helper - .on('data', function (data, meta) { - dataArray.push(data); - if(updateCallback) { - updateCallback(meta); - } - }) - .on('error', function(err) { - dataArray = []; - reject(err); - }) - .on('end', function (){ - try { - var result = transformZipOutput(resultType, concat(chunkType, dataArray), mimeType); - resolve(result); - } catch (e) { - reject(e); - } - dataArray = []; - }) - .resume(); - }); -} - -/** - * An helper to easily use workers outside of JSZip. - * @constructor - * @param {Worker} worker the worker to wrap - * @param {String} outputType the type of data expected by the use - * @param {String} mimeType the mime type of the content, if applicable. - */ -function StreamHelper(worker, outputType, mimeType) { - var internalType = outputType; - switch(outputType) { - case "blob": - case "arraybuffer": - internalType = "uint8array"; - break; - case "base64": - internalType = "string"; - break; - } - - try { - // the type used internally - this._internalType = internalType; - // the type used to output results - this._outputType = outputType; - // the mime type - this._mimeType = mimeType; - utils.checkSupport(internalType); - this._worker = worker.pipe(new ConvertWorker(internalType)); - // the last workers can be rewired without issues but we need to - // prevent any updates on previous workers. - worker.lock(); - } catch(e) { - this._worker = new GenericWorker("error"); - this._worker.error(e); - } -} - -StreamHelper.prototype = { - /** - * Listen a StreamHelper, accumulate its content and concatenate it into a - * complete block. - * @param {Function} updateCb the update callback. - * @return Promise the promise for the accumulation. - */ - accumulate : function (updateCb) { - return accumulate(this, updateCb); - }, - /** - * Add a listener on an event triggered on a stream. - * @param {String} evt the name of the event - * @param {Function} fn the listener - * @return {StreamHelper} the current helper. - */ - on : function (evt, fn) { - var self = this; - - if(evt === "data") { - this._worker.on(evt, function (chunk) { - fn.call(self, chunk.data, chunk.meta); - }); - } else { - this._worker.on(evt, function () { - utils.delay(fn, arguments, self); - }); - } - return this; - }, - /** - * Resume the flow of chunks. - * @return {StreamHelper} the current helper. - */ - resume : function () { - utils.delay(this._worker.resume, [], this._worker); - return this; - }, - /** - * Pause the flow of chunks. - * @return {StreamHelper} the current helper. - */ - pause : function () { - this._worker.pause(); - return this; - }, - /** - * Return a nodejs stream for this helper. - * @param {Function} updateCb the update callback. - * @return {NodejsStreamOutputAdapter} the nodejs stream. - */ - toNodejsStream : function (updateCb) { - utils.checkSupport("nodestream"); - if (this._outputType !== "nodebuffer") { - // an object stream containing blob/arraybuffer/uint8array/string - // is strange and I don't know if it would be useful. - // I you find this comment and have a good usecase, please open a - // bug report ! - throw new Error(this._outputType + " is not supported by this method"); - } - - return new NodejsStreamOutputAdapter(this, { - objectMode : this._outputType !== "nodebuffer" - }, updateCb); - } -}; - - -module.exports = StreamHelper; - -},{"../base64":1,"../external":6,"../nodejs/NodejsStreamOutputAdapter":13,"../support":30,"../utils":32,"./ConvertWorker":24,"./GenericWorker":28}],30:[function(require,module,exports){ -'use strict'; - -exports.base64 = true; -exports.array = true; -exports.string = true; -exports.arraybuffer = typeof ArrayBuffer !== "undefined" && typeof Uint8Array !== "undefined"; -exports.nodebuffer = typeof Buffer !== "undefined"; -// contains true if JSZip can read/generate Uint8Array, false otherwise. -exports.uint8array = typeof Uint8Array !== "undefined"; - -if (typeof ArrayBuffer === "undefined") { - exports.blob = false; -} -else { - var buffer = new ArrayBuffer(0); - try { - exports.blob = new Blob([buffer], { - type: "application/zip" - }).size === 0; - } - catch (e) { - try { - var Builder = self.BlobBuilder || self.WebKitBlobBuilder || self.MozBlobBuilder || self.MSBlobBuilder; - var builder = new Builder(); - builder.append(buffer); - exports.blob = builder.getBlob('application/zip').size === 0; - } - catch (e) { - exports.blob = false; - } - } -} - -try { - exports.nodestream = !!require('readable-stream').Readable; -} catch(e) { - exports.nodestream = false; -} - -},{"readable-stream":16}],31:[function(require,module,exports){ -'use strict'; - -var utils = require('./utils'); -var support = require('./support'); -var nodejsUtils = require('./nodejsUtils'); -var GenericWorker = require('./stream/GenericWorker'); - -/** - * The following functions come from pako, from pako/lib/utils/strings - * released under the MIT license, see pako https://github.com/nodeca/pako/ - */ - -// Table with utf8 lengths (calculated by first byte of sequence) -// Note, that 5 & 6-byte values and some 4-byte values can not be represented in JS, -// because max possible codepoint is 0x10ffff -var _utf8len = new Array(256); -for (var i=0; i<256; i++) { - _utf8len[i] = (i >= 252 ? 6 : i >= 248 ? 5 : i >= 240 ? 4 : i >= 224 ? 3 : i >= 192 ? 2 : 1); -} -_utf8len[254]=_utf8len[254]=1; // Invalid sequence start - -// convert string to array (typed, when possible) -var string2buf = function (str) { - var buf, c, c2, m_pos, i, str_len = str.length, buf_len = 0; - - // count binary size - for (m_pos = 0; m_pos < str_len; m_pos++) { - c = str.charCodeAt(m_pos); - if ((c & 0xfc00) === 0xd800 && (m_pos+1 < str_len)) { - c2 = str.charCodeAt(m_pos+1); - if ((c2 & 0xfc00) === 0xdc00) { - c = 0x10000 + ((c - 0xd800) << 10) + (c2 - 0xdc00); - m_pos++; - } - } - buf_len += c < 0x80 ? 1 : c < 0x800 ? 2 : c < 0x10000 ? 3 : 4; - } - - // allocate buffer - if (support.uint8array) { - buf = new Uint8Array(buf_len); - } else { - buf = new Array(buf_len); - } - - // convert - for (i=0, m_pos = 0; i < buf_len; m_pos++) { - c = str.charCodeAt(m_pos); - if ((c & 0xfc00) === 0xd800 && (m_pos+1 < str_len)) { - c2 = str.charCodeAt(m_pos+1); - if ((c2 & 0xfc00) === 0xdc00) { - c = 0x10000 + ((c - 0xd800) << 10) + (c2 - 0xdc00); - m_pos++; - } - } - if (c < 0x80) { - /* one byte */ - buf[i++] = c; - } else if (c < 0x800) { - /* two bytes */ - buf[i++] = 0xC0 | (c >>> 6); - buf[i++] = 0x80 | (c & 0x3f); - } else if (c < 0x10000) { - /* three bytes */ - buf[i++] = 0xE0 | (c >>> 12); - buf[i++] = 0x80 | (c >>> 6 & 0x3f); - buf[i++] = 0x80 | (c & 0x3f); - } else { - /* four bytes */ - buf[i++] = 0xf0 | (c >>> 18); - buf[i++] = 0x80 | (c >>> 12 & 0x3f); - buf[i++] = 0x80 | (c >>> 6 & 0x3f); - buf[i++] = 0x80 | (c & 0x3f); - } - } - - return buf; -}; - -// Calculate max possible position in utf8 buffer, -// that will not break sequence. If that's not possible -// - (very small limits) return max size as is. -// -// buf[] - utf8 bytes array -// max - length limit (mandatory); -var utf8border = function(buf, max) { - var pos; - - max = max || buf.length; - if (max > buf.length) { max = buf.length; } - - // go back from last position, until start of sequence found - pos = max-1; - while (pos >= 0 && (buf[pos] & 0xC0) === 0x80) { pos--; } - - // Fuckup - very small and broken sequence, - // return max, because we should return something anyway. - if (pos < 0) { return max; } - - // If we came to start of buffer - that means vuffer is too small, - // return max too. - if (pos === 0) { return max; } - - return (pos + _utf8len[buf[pos]] > max) ? pos : max; -}; - -// convert array to string -var buf2string = function (buf) { - var str, i, out, c, c_len; - var len = buf.length; - - // Reserve max possible length (2 words per char) - // NB: by unknown reasons, Array is significantly faster for - // String.fromCharCode.apply than Uint16Array. - var utf16buf = new Array(len*2); - - for (out=0, i=0; i 4) { utf16buf[out++] = 0xfffd; i += c_len-1; continue; } - - // apply mask on first byte - c &= c_len === 2 ? 0x1f : c_len === 3 ? 0x0f : 0x07; - // join the rest - while (c_len > 1 && i < len) { - c = (c << 6) | (buf[i++] & 0x3f); - c_len--; - } - - // terminated by end of string? - if (c_len > 1) { utf16buf[out++] = 0xfffd; continue; } - - if (c < 0x10000) { - utf16buf[out++] = c; - } else { - c -= 0x10000; - utf16buf[out++] = 0xd800 | ((c >> 10) & 0x3ff); - utf16buf[out++] = 0xdc00 | (c & 0x3ff); - } - } - - // shrinkBuf(utf16buf, out) - if (utf16buf.length !== out) { - if(utf16buf.subarray) { - utf16buf = utf16buf.subarray(0, out); - } else { - utf16buf.length = out; - } - } - - // return String.fromCharCode.apply(null, utf16buf); - return utils.applyFromCharCode(utf16buf); -}; - - -// That's all for the pako functions. - - -/** - * Transform a javascript string into an array (typed if possible) of bytes, - * UTF-8 encoded. - * @param {String} str the string to encode - * @return {Array|Uint8Array|Buffer} the UTF-8 encoded string. - */ -exports.utf8encode = function utf8encode(str) { - if (support.nodebuffer) { - return nodejsUtils.newBufferFrom(str, "utf-8"); - } - - return string2buf(str); -}; - - -/** - * Transform a bytes array (or a representation) representing an UTF-8 encoded - * string into a javascript string. - * @param {Array|Uint8Array|Buffer} buf the data de decode - * @return {String} the decoded string. - */ -exports.utf8decode = function utf8decode(buf) { - if (support.nodebuffer) { - return utils.transformTo("nodebuffer", buf).toString("utf-8"); - } - - buf = utils.transformTo(support.uint8array ? "uint8array" : "array", buf); - - return buf2string(buf); -}; - -/** - * A worker to decode utf8 encoded binary chunks into string chunks. - * @constructor - */ -function Utf8DecodeWorker() { - GenericWorker.call(this, "utf-8 decode"); - // the last bytes if a chunk didn't end with a complete codepoint. - this.leftOver = null; -} -utils.inherits(Utf8DecodeWorker, GenericWorker); - -/** - * @see GenericWorker.processChunk - */ -Utf8DecodeWorker.prototype.processChunk = function (chunk) { - - var data = utils.transformTo(support.uint8array ? "uint8array" : "array", chunk.data); - - // 1st step, re-use what's left of the previous chunk - if (this.leftOver && this.leftOver.length) { - if(support.uint8array) { - var previousData = data; - data = new Uint8Array(previousData.length + this.leftOver.length); - data.set(this.leftOver, 0); - data.set(previousData, this.leftOver.length); - } else { - data = this.leftOver.concat(data); - } - this.leftOver = null; - } - - var nextBoundary = utf8border(data); - var usableData = data; - if (nextBoundary !== data.length) { - if (support.uint8array) { - usableData = data.subarray(0, nextBoundary); - this.leftOver = data.subarray(nextBoundary, data.length); - } else { - usableData = data.slice(0, nextBoundary); - this.leftOver = data.slice(nextBoundary, data.length); - } - } - - this.push({ - data : exports.utf8decode(usableData), - meta : chunk.meta - }); -}; - -/** - * @see GenericWorker.flush - */ -Utf8DecodeWorker.prototype.flush = function () { - if(this.leftOver && this.leftOver.length) { - this.push({ - data : exports.utf8decode(this.leftOver), - meta : {} - }); - this.leftOver = null; - } -}; -exports.Utf8DecodeWorker = Utf8DecodeWorker; - -/** - * A worker to endcode string chunks into utf8 encoded binary chunks. - * @constructor - */ -function Utf8EncodeWorker() { - GenericWorker.call(this, "utf-8 encode"); -} -utils.inherits(Utf8EncodeWorker, GenericWorker); - -/** - * @see GenericWorker.processChunk - */ -Utf8EncodeWorker.prototype.processChunk = function (chunk) { - this.push({ - data : exports.utf8encode(chunk.data), - meta : chunk.meta - }); -}; -exports.Utf8EncodeWorker = Utf8EncodeWorker; - -},{"./nodejsUtils":14,"./stream/GenericWorker":28,"./support":30,"./utils":32}],32:[function(require,module,exports){ -'use strict'; - -var support = require('./support'); -var base64 = require('./base64'); -var nodejsUtils = require('./nodejsUtils'); -var setImmediate = require('set-immediate-shim'); -var external = require("./external"); - - -/** - * Convert a string that pass as a "binary string": it should represent a byte - * array but may have > 255 char codes. Be sure to take only the first byte - * and returns the byte array. - * @param {String} str the string to transform. - * @return {Array|Uint8Array} the string in a binary format. - */ -function string2binary(str) { - var result = null; - if (support.uint8array) { - result = new Uint8Array(str.length); - } else { - result = new Array(str.length); - } - return stringToArrayLike(str, result); -} - -/** - * Create a new blob with the given content and the given type. - * @param {String|ArrayBuffer} part the content to put in the blob. DO NOT use - * an Uint8Array because the stock browser of android 4 won't accept it (it - * will be silently converted to a string, "[object Uint8Array]"). - * - * Use only ONE part to build the blob to avoid a memory leak in IE11 / Edge: - * when a large amount of Array is used to create the Blob, the amount of - * memory consumed is nearly 100 times the original data amount. - * - * @param {String} type the mime type of the blob. - * @return {Blob} the created blob. - */ -exports.newBlob = function(part, type) { - exports.checkSupport("blob"); - - try { - // Blob constructor - return new Blob([part], { - type: type - }); - } - catch (e) { - - try { - // deprecated, browser only, old way - var Builder = self.BlobBuilder || self.WebKitBlobBuilder || self.MozBlobBuilder || self.MSBlobBuilder; - var builder = new Builder(); - builder.append(part); - return builder.getBlob(type); - } - catch (e) { - - // well, fuck ?! - throw new Error("Bug : can't construct the Blob."); - } - } - - -}; -/** - * The identity function. - * @param {Object} input the input. - * @return {Object} the same input. - */ -function identity(input) { - return input; -} - -/** - * Fill in an array with a string. - * @param {String} str the string to use. - * @param {Array|ArrayBuffer|Uint8Array|Buffer} array the array to fill in (will be mutated). - * @return {Array|ArrayBuffer|Uint8Array|Buffer} the updated array. - */ -function stringToArrayLike(str, array) { - for (var i = 0; i < str.length; ++i) { - array[i] = str.charCodeAt(i) & 0xFF; - } - return array; -} - -/** - * An helper for the function arrayLikeToString. - * This contains static information and functions that - * can be optimized by the browser JIT compiler. - */ -var arrayToStringHelper = { - /** - * Transform an array of int into a string, chunk by chunk. - * See the performances notes on arrayLikeToString. - * @param {Array|ArrayBuffer|Uint8Array|Buffer} array the array to transform. - * @param {String} type the type of the array. - * @param {Integer} chunk the chunk size. - * @return {String} the resulting string. - * @throws Error if the chunk is too big for the stack. - */ - stringifyByChunk: function(array, type, chunk) { - var result = [], k = 0, len = array.length; - // shortcut - if (len <= chunk) { - return String.fromCharCode.apply(null, array); - } - while (k < len) { - if (type === "array" || type === "nodebuffer") { - result.push(String.fromCharCode.apply(null, array.slice(k, Math.min(k + chunk, len)))); - } - else { - result.push(String.fromCharCode.apply(null, array.subarray(k, Math.min(k + chunk, len)))); - } - k += chunk; - } - return result.join(""); - }, - /** - * Call String.fromCharCode on every item in the array. - * This is the naive implementation, which generate A LOT of intermediate string. - * This should be used when everything else fail. - * @param {Array|ArrayBuffer|Uint8Array|Buffer} array the array to transform. - * @return {String} the result. - */ - stringifyByChar: function(array){ - var resultStr = ""; - for(var i = 0; i < array.length; i++) { - resultStr += String.fromCharCode(array[i]); - } - return resultStr; - }, - applyCanBeUsed : { - /** - * true if the browser accepts to use String.fromCharCode on Uint8Array - */ - uint8array : (function () { - try { - return support.uint8array && String.fromCharCode.apply(null, new Uint8Array(1)).length === 1; - } catch (e) { - return false; - } - })(), - /** - * true if the browser accepts to use String.fromCharCode on nodejs Buffer. - */ - nodebuffer : (function () { - try { - return support.nodebuffer && String.fromCharCode.apply(null, nodejsUtils.allocBuffer(1)).length === 1; - } catch (e) { - return false; - } - })() - } -}; - -/** - * Transform an array-like object to a string. - * @param {Array|ArrayBuffer|Uint8Array|Buffer} array the array to transform. - * @return {String} the result. - */ -function arrayLikeToString(array) { - // Performances notes : - // -------------------- - // String.fromCharCode.apply(null, array) is the fastest, see - // see http://jsperf.com/converting-a-uint8array-to-a-string/2 - // but the stack is limited (and we can get huge arrays !). - // - // result += String.fromCharCode(array[i]); generate too many strings ! - // - // This code is inspired by http://jsperf.com/arraybuffer-to-string-apply-performance/2 - // TODO : we now have workers that split the work. Do we still need that ? - var chunk = 65536, - type = exports.getTypeOf(array), - canUseApply = true; - if (type === "uint8array") { - canUseApply = arrayToStringHelper.applyCanBeUsed.uint8array; - } else if (type === "nodebuffer") { - canUseApply = arrayToStringHelper.applyCanBeUsed.nodebuffer; - } - - if (canUseApply) { - while (chunk > 1) { - try { - return arrayToStringHelper.stringifyByChunk(array, type, chunk); - } catch (e) { - chunk = Math.floor(chunk / 2); - } - } - } - - // no apply or chunk error : slow and painful algorithm - // default browser on android 4.* - return arrayToStringHelper.stringifyByChar(array); -} - -exports.applyFromCharCode = arrayLikeToString; - - -/** - * Copy the data from an array-like to an other array-like. - * @param {Array|ArrayBuffer|Uint8Array|Buffer} arrayFrom the origin array. - * @param {Array|ArrayBuffer|Uint8Array|Buffer} arrayTo the destination array which will be mutated. - * @return {Array|ArrayBuffer|Uint8Array|Buffer} the updated destination array. - */ -function arrayLikeToArrayLike(arrayFrom, arrayTo) { - for (var i = 0; i < arrayFrom.length; i++) { - arrayTo[i] = arrayFrom[i]; - } - return arrayTo; -} - -// a matrix containing functions to transform everything into everything. -var transform = {}; - -// string to ? -transform["string"] = { - "string": identity, - "array": function(input) { - return stringToArrayLike(input, new Array(input.length)); - }, - "arraybuffer": function(input) { - return transform["string"]["uint8array"](input).buffer; - }, - "uint8array": function(input) { - return stringToArrayLike(input, new Uint8Array(input.length)); - }, - "nodebuffer": function(input) { - return stringToArrayLike(input, nodejsUtils.allocBuffer(input.length)); - } -}; - -// array to ? -transform["array"] = { - "string": arrayLikeToString, - "array": identity, - "arraybuffer": function(input) { - return (new Uint8Array(input)).buffer; - }, - "uint8array": function(input) { - return new Uint8Array(input); - }, - "nodebuffer": function(input) { - return nodejsUtils.newBufferFrom(input); - } -}; - -// arraybuffer to ? -transform["arraybuffer"] = { - "string": function(input) { - return arrayLikeToString(new Uint8Array(input)); - }, - "array": function(input) { - return arrayLikeToArrayLike(new Uint8Array(input), new Array(input.byteLength)); - }, - "arraybuffer": identity, - "uint8array": function(input) { - return new Uint8Array(input); - }, - "nodebuffer": function(input) { - return nodejsUtils.newBufferFrom(new Uint8Array(input)); - } -}; - -// uint8array to ? -transform["uint8array"] = { - "string": arrayLikeToString, - "array": function(input) { - return arrayLikeToArrayLike(input, new Array(input.length)); - }, - "arraybuffer": function(input) { - return input.buffer; - }, - "uint8array": identity, - "nodebuffer": function(input) { - return nodejsUtils.newBufferFrom(input); - } -}; - -// nodebuffer to ? -transform["nodebuffer"] = { - "string": arrayLikeToString, - "array": function(input) { - return arrayLikeToArrayLike(input, new Array(input.length)); - }, - "arraybuffer": function(input) { - return transform["nodebuffer"]["uint8array"](input).buffer; - }, - "uint8array": function(input) { - return arrayLikeToArrayLike(input, new Uint8Array(input.length)); - }, - "nodebuffer": identity -}; - -/** - * Transform an input into any type. - * The supported output type are : string, array, uint8array, arraybuffer, nodebuffer. - * If no output type is specified, the unmodified input will be returned. - * @param {String} outputType the output type. - * @param {String|Array|ArrayBuffer|Uint8Array|Buffer} input the input to convert. - * @throws {Error} an Error if the browser doesn't support the requested output type. - */ -exports.transformTo = function(outputType, input) { - if (!input) { - // undefined, null, etc - // an empty string won't harm. - input = ""; - } - if (!outputType) { - return input; - } - exports.checkSupport(outputType); - var inputType = exports.getTypeOf(input); - var result = transform[inputType][outputType](input); - return result; -}; - -/** - * Return the type of the input. - * The type will be in a format valid for JSZip.utils.transformTo : string, array, uint8array, arraybuffer. - * @param {Object} input the input to identify. - * @return {String} the (lowercase) type of the input. - */ -exports.getTypeOf = function(input) { - if (typeof input === "string") { - return "string"; - } - if (Object.prototype.toString.call(input) === "[object Array]") { - return "array"; - } - if (support.nodebuffer && nodejsUtils.isBuffer(input)) { - return "nodebuffer"; - } - if (support.uint8array && input instanceof Uint8Array) { - return "uint8array"; - } - if (support.arraybuffer && input instanceof ArrayBuffer) { - return "arraybuffer"; - } -}; - -/** - * Throw an exception if the type is not supported. - * @param {String} type the type to check. - * @throws {Error} an Error if the browser doesn't support the requested type. - */ -exports.checkSupport = function(type) { - var supported = support[type.toLowerCase()]; - if (!supported) { - throw new Error(type + " is not supported by this platform"); - } -}; - -exports.MAX_VALUE_16BITS = 65535; -exports.MAX_VALUE_32BITS = -1; // well, "\xFF\xFF\xFF\xFF\xFF\xFF\xFF\xFF" is parsed as -1 - -/** - * Prettify a string read as binary. - * @param {string} str the string to prettify. - * @return {string} a pretty string. - */ -exports.pretty = function(str) { - var res = '', - code, i; - for (i = 0; i < (str || "").length; i++) { - code = str.charCodeAt(i); - res += '\\x' + (code < 16 ? "0" : "") + code.toString(16).toUpperCase(); - } - return res; -}; - -/** - * Defer the call of a function. - * @param {Function} callback the function to call asynchronously. - * @param {Array} args the arguments to give to the callback. - */ -exports.delay = function(callback, args, self) { - setImmediate(function () { - callback.apply(self || null, args || []); - }); -}; - -/** - * Extends a prototype with an other, without calling a constructor with - * side effects. Inspired by nodejs' `utils.inherits` - * @param {Function} ctor the constructor to augment - * @param {Function} superCtor the parent constructor to use - */ -exports.inherits = function (ctor, superCtor) { - var Obj = function() {}; - Obj.prototype = superCtor.prototype; - ctor.prototype = new Obj(); -}; - -/** - * Merge the objects passed as parameters into a new one. - * @private - * @param {...Object} var_args All objects to merge. - * @return {Object} a new object with the data of the others. - */ -exports.extend = function() { - var result = {}, i, attr; - for (i = 0; i < arguments.length; i++) { // arguments is not enumerable in some browsers - for (attr in arguments[i]) { - if (arguments[i].hasOwnProperty(attr) && typeof result[attr] === "undefined") { - result[attr] = arguments[i][attr]; - } - } - } - return result; -}; - -/** - * Transform arbitrary content into a Promise. - * @param {String} name a name for the content being processed. - * @param {Object} inputData the content to process. - * @param {Boolean} isBinary true if the content is not an unicode string - * @param {Boolean} isOptimizedBinaryString true if the string content only has one byte per character. - * @param {Boolean} isBase64 true if the string content is encoded with base64. - * @return {Promise} a promise in a format usable by JSZip. - */ -exports.prepareContent = function(name, inputData, isBinary, isOptimizedBinaryString, isBase64) { - - // if inputData is already a promise, this flatten it. - var promise = external.Promise.resolve(inputData).then(function(data) { - - - var isBlob = support.blob && (data instanceof Blob || ['[object File]', '[object Blob]'].indexOf(Object.prototype.toString.call(data)) !== -1); - - if (isBlob && typeof FileReader !== "undefined") { - return new external.Promise(function (resolve, reject) { - var reader = new FileReader(); - - reader.onload = function(e) { - resolve(e.target.result); - }; - reader.onerror = function(e) { - reject(e.target.error); - }; - reader.readAsArrayBuffer(data); - }); - } else { - return data; - } - }); - - return promise.then(function(data) { - var dataType = exports.getTypeOf(data); - - if (!dataType) { - return external.Promise.reject( - new Error("Can't read the data of '" + name + "'. Is it " + - "in a supported JavaScript type (String, Blob, ArrayBuffer, etc) ?") - ); - } - // special case : it's way easier to work with Uint8Array than with ArrayBuffer - if (dataType === "arraybuffer") { - data = exports.transformTo("uint8array", data); - } else if (dataType === "string") { - if (isBase64) { - data = base64.decode(data); - } - else if (isBinary) { - // optimizedBinaryString === true means that the file has already been filtered with a 0xFF mask - if (isOptimizedBinaryString !== true) { - // this is a string, not in a base64 format. - // Be sure that this is a correct "binary string" - data = string2binary(data); - } - } - } - return data; - }); -}; - -},{"./base64":1,"./external":6,"./nodejsUtils":14,"./support":30,"set-immediate-shim":54}],33:[function(require,module,exports){ -'use strict'; -var readerFor = require('./reader/readerFor'); -var utils = require('./utils'); -var sig = require('./signature'); -var ZipEntry = require('./zipEntry'); -var utf8 = require('./utf8'); -var support = require('./support'); -// class ZipEntries {{{ -/** - * All the entries in the zip file. - * @constructor - * @param {Object} loadOptions Options for loading the stream. - */ -function ZipEntries(loadOptions) { - this.files = []; - this.loadOptions = loadOptions; -} -ZipEntries.prototype = { - /** - * Check that the reader is on the specified signature. - * @param {string} expectedSignature the expected signature. - * @throws {Error} if it is an other signature. - */ - checkSignature: function(expectedSignature) { - if (!this.reader.readAndCheckSignature(expectedSignature)) { - this.reader.index -= 4; - var signature = this.reader.readString(4); - throw new Error("Corrupted zip or bug: unexpected signature " + "(" + utils.pretty(signature) + ", expected " + utils.pretty(expectedSignature) + ")"); - } - }, - /** - * Check if the given signature is at the given index. - * @param {number} askedIndex the index to check. - * @param {string} expectedSignature the signature to expect. - * @return {boolean} true if the signature is here, false otherwise. - */ - isSignature: function(askedIndex, expectedSignature) { - var currentIndex = this.reader.index; - this.reader.setIndex(askedIndex); - var signature = this.reader.readString(4); - var result = signature === expectedSignature; - this.reader.setIndex(currentIndex); - return result; - }, - /** - * Read the end of the central directory. - */ - readBlockEndOfCentral: function() { - this.diskNumber = this.reader.readInt(2); - this.diskWithCentralDirStart = this.reader.readInt(2); - this.centralDirRecordsOnThisDisk = this.reader.readInt(2); - this.centralDirRecords = this.reader.readInt(2); - this.centralDirSize = this.reader.readInt(4); - this.centralDirOffset = this.reader.readInt(4); - - this.zipCommentLength = this.reader.readInt(2); - // warning : the encoding depends of the system locale - // On a linux machine with LANG=en_US.utf8, this field is utf8 encoded. - // On a windows machine, this field is encoded with the localized windows code page. - var zipComment = this.reader.readData(this.zipCommentLength); - var decodeParamType = support.uint8array ? "uint8array" : "array"; - // To get consistent behavior with the generation part, we will assume that - // this is utf8 encoded unless specified otherwise. - var decodeContent = utils.transformTo(decodeParamType, zipComment); - this.zipComment = this.loadOptions.decodeFileName(decodeContent); - }, - /** - * Read the end of the Zip 64 central directory. - * Not merged with the method readEndOfCentral : - * The end of central can coexist with its Zip64 brother, - * I don't want to read the wrong number of bytes ! - */ - readBlockZip64EndOfCentral: function() { - this.zip64EndOfCentralSize = this.reader.readInt(8); - this.reader.skip(4); - // this.versionMadeBy = this.reader.readString(2); - // this.versionNeeded = this.reader.readInt(2); - this.diskNumber = this.reader.readInt(4); - this.diskWithCentralDirStart = this.reader.readInt(4); - this.centralDirRecordsOnThisDisk = this.reader.readInt(8); - this.centralDirRecords = this.reader.readInt(8); - this.centralDirSize = this.reader.readInt(8); - this.centralDirOffset = this.reader.readInt(8); - - this.zip64ExtensibleData = {}; - var extraDataSize = this.zip64EndOfCentralSize - 44, - index = 0, - extraFieldId, - extraFieldLength, - extraFieldValue; - while (index < extraDataSize) { - extraFieldId = this.reader.readInt(2); - extraFieldLength = this.reader.readInt(4); - extraFieldValue = this.reader.readData(extraFieldLength); - this.zip64ExtensibleData[extraFieldId] = { - id: extraFieldId, - length: extraFieldLength, - value: extraFieldValue - }; - } - }, - /** - * Read the end of the Zip 64 central directory locator. - */ - readBlockZip64EndOfCentralLocator: function() { - this.diskWithZip64CentralDirStart = this.reader.readInt(4); - this.relativeOffsetEndOfZip64CentralDir = this.reader.readInt(8); - this.disksCount = this.reader.readInt(4); - if (this.disksCount > 1) { - throw new Error("Multi-volumes zip are not supported"); - } - }, - /** - * Read the local files, based on the offset read in the central part. - */ - readLocalFiles: function() { - var i, file; - for (i = 0; i < this.files.length; i++) { - file = this.files[i]; - this.reader.setIndex(file.localHeaderOffset); - this.checkSignature(sig.LOCAL_FILE_HEADER); - file.readLocalPart(this.reader); - file.handleUTF8(); - file.processAttributes(); - } - }, - /** - * Read the central directory. - */ - readCentralDir: function() { - var file; - - this.reader.setIndex(this.centralDirOffset); - while (this.reader.readAndCheckSignature(sig.CENTRAL_FILE_HEADER)) { - file = new ZipEntry({ - zip64: this.zip64 - }, this.loadOptions); - file.readCentralPart(this.reader); - this.files.push(file); - } - - if (this.centralDirRecords !== this.files.length) { - if (this.centralDirRecords !== 0 && this.files.length === 0) { - // We expected some records but couldn't find ANY. - // This is really suspicious, as if something went wrong. - throw new Error("Corrupted zip or bug: expected " + this.centralDirRecords + " records in central dir, got " + this.files.length); - } else { - // We found some records but not all. - // Something is wrong but we got something for the user: no error here. - // console.warn("expected", this.centralDirRecords, "records in central dir, got", this.files.length); - } - } - }, - /** - * Read the end of central directory. - */ - readEndOfCentral: function() { - var offset = this.reader.lastIndexOfSignature(sig.CENTRAL_DIRECTORY_END); - if (offset < 0) { - // Check if the content is a truncated zip or complete garbage. - // A "LOCAL_FILE_HEADER" is not required at the beginning (auto - // extractible zip for example) but it can give a good hint. - // If an ajax request was used without responseType, we will also - // get unreadable data. - var isGarbage = !this.isSignature(0, sig.LOCAL_FILE_HEADER); - - if (isGarbage) { - throw new Error("Can't find end of central directory : is this a zip file ? " + - "If it is, see https://stuk.github.io/jszip/documentation/howto/read_zip.html"); - } else { - throw new Error("Corrupted zip: can't find end of central directory"); - } - - } - this.reader.setIndex(offset); - var endOfCentralDirOffset = offset; - this.checkSignature(sig.CENTRAL_DIRECTORY_END); - this.readBlockEndOfCentral(); - - - /* extract from the zip spec : - 4) If one of the fields in the end of central directory - record is too small to hold required data, the field - should be set to -1 (0xFFFF or 0xFFFFFFFF) and the - ZIP64 format record should be created. - 5) The end of central directory record and the - Zip64 end of central directory locator record must - reside on the same disk when splitting or spanning - an archive. - */ - if (this.diskNumber === utils.MAX_VALUE_16BITS || this.diskWithCentralDirStart === utils.MAX_VALUE_16BITS || this.centralDirRecordsOnThisDisk === utils.MAX_VALUE_16BITS || this.centralDirRecords === utils.MAX_VALUE_16BITS || this.centralDirSize === utils.MAX_VALUE_32BITS || this.centralDirOffset === utils.MAX_VALUE_32BITS) { - this.zip64 = true; - - /* - Warning : the zip64 extension is supported, but ONLY if the 64bits integer read from - the zip file can fit into a 32bits integer. This cannot be solved : JavaScript represents - all numbers as 64-bit double precision IEEE 754 floating point numbers. - So, we have 53bits for integers and bitwise operations treat everything as 32bits. - see https://developer.mozilla.org/en-US/docs/JavaScript/Reference/Operators/Bitwise_Operators - and http://www.ecma-international.org/publications/files/ECMA-ST/ECMA-262.pdf section 8.5 - */ - - // should look for a zip64 EOCD locator - offset = this.reader.lastIndexOfSignature(sig.ZIP64_CENTRAL_DIRECTORY_LOCATOR); - if (offset < 0) { - throw new Error("Corrupted zip: can't find the ZIP64 end of central directory locator"); - } - this.reader.setIndex(offset); - this.checkSignature(sig.ZIP64_CENTRAL_DIRECTORY_LOCATOR); - this.readBlockZip64EndOfCentralLocator(); - - // now the zip64 EOCD record - if (!this.isSignature(this.relativeOffsetEndOfZip64CentralDir, sig.ZIP64_CENTRAL_DIRECTORY_END)) { - // console.warn("ZIP64 end of central directory not where expected."); - this.relativeOffsetEndOfZip64CentralDir = this.reader.lastIndexOfSignature(sig.ZIP64_CENTRAL_DIRECTORY_END); - if (this.relativeOffsetEndOfZip64CentralDir < 0) { - throw new Error("Corrupted zip: can't find the ZIP64 end of central directory"); - } - } - this.reader.setIndex(this.relativeOffsetEndOfZip64CentralDir); - this.checkSignature(sig.ZIP64_CENTRAL_DIRECTORY_END); - this.readBlockZip64EndOfCentral(); - } - - var expectedEndOfCentralDirOffset = this.centralDirOffset + this.centralDirSize; - if (this.zip64) { - expectedEndOfCentralDirOffset += 20; // end of central dir 64 locator - expectedEndOfCentralDirOffset += 12 /* should not include the leading 12 bytes */ + this.zip64EndOfCentralSize; - } - - var extraBytes = endOfCentralDirOffset - expectedEndOfCentralDirOffset; - - if (extraBytes > 0) { - // console.warn(extraBytes, "extra bytes at beginning or within zipfile"); - if (this.isSignature(endOfCentralDirOffset, sig.CENTRAL_FILE_HEADER)) { - // The offsets seem wrong, but we have something at the specified offset. - // So… we keep it. - } else { - // the offset is wrong, update the "zero" of the reader - // this happens if data has been prepended (crx files for example) - this.reader.zero = extraBytes; - } - } else if (extraBytes < 0) { - throw new Error("Corrupted zip: missing " + Math.abs(extraBytes) + " bytes."); - } - }, - prepareReader: function(data) { - this.reader = readerFor(data); - }, - /** - * Read a zip file and create ZipEntries. - * @param {String|ArrayBuffer|Uint8Array|Buffer} data the binary string representing a zip file. - */ - load: function(data) { - this.prepareReader(data); - this.readEndOfCentral(); - this.readCentralDir(); - this.readLocalFiles(); - } -}; -// }}} end of ZipEntries -module.exports = ZipEntries; - -},{"./reader/readerFor":22,"./signature":23,"./support":30,"./utf8":31,"./utils":32,"./zipEntry":34}],34:[function(require,module,exports){ -'use strict'; -var readerFor = require('./reader/readerFor'); -var utils = require('./utils'); -var CompressedObject = require('./compressedObject'); -var crc32fn = require('./crc32'); -var utf8 = require('./utf8'); -var compressions = require('./compressions'); -var support = require('./support'); - -var MADE_BY_DOS = 0x00; -var MADE_BY_UNIX = 0x03; - -/** - * Find a compression registered in JSZip. - * @param {string} compressionMethod the method magic to find. - * @return {Object|null} the JSZip compression object, null if none found. - */ -var findCompression = function(compressionMethod) { - for (var method in compressions) { - if (!compressions.hasOwnProperty(method)) { - continue; - } - if (compressions[method].magic === compressionMethod) { - return compressions[method]; - } - } - return null; -}; - -// class ZipEntry {{{ -/** - * An entry in the zip file. - * @constructor - * @param {Object} options Options of the current file. - * @param {Object} loadOptions Options for loading the stream. - */ -function ZipEntry(options, loadOptions) { - this.options = options; - this.loadOptions = loadOptions; -} -ZipEntry.prototype = { - /** - * say if the file is encrypted. - * @return {boolean} true if the file is encrypted, false otherwise. - */ - isEncrypted: function() { - // bit 1 is set - return (this.bitFlag & 0x0001) === 0x0001; - }, - /** - * say if the file has utf-8 filename/comment. - * @return {boolean} true if the filename/comment is in utf-8, false otherwise. - */ - useUTF8: function() { - // bit 11 is set - return (this.bitFlag & 0x0800) === 0x0800; - }, - /** - * Read the local part of a zip file and add the info in this object. - * @param {DataReader} reader the reader to use. - */ - readLocalPart: function(reader) { - var compression, localExtraFieldsLength; - - // we already know everything from the central dir ! - // If the central dir data are false, we are doomed. - // On the bright side, the local part is scary : zip64, data descriptors, both, etc. - // The less data we get here, the more reliable this should be. - // Let's skip the whole header and dash to the data ! - reader.skip(22); - // in some zip created on windows, the filename stored in the central dir contains \ instead of /. - // Strangely, the filename here is OK. - // I would love to treat these zip files as corrupted (see http://www.info-zip.org/FAQ.html#backslashes - // or APPNOTE#4.4.17.1, "All slashes MUST be forward slashes '/'") but there are a lot of bad zip generators... - // Search "unzip mismatching "local" filename continuing with "central" filename version" on - // the internet. - // - // I think I see the logic here : the central directory is used to display - // content and the local directory is used to extract the files. Mixing / and \ - // may be used to display \ to windows users and use / when extracting the files. - // Unfortunately, this lead also to some issues : http://seclists.org/fulldisclosure/2009/Sep/394 - this.fileNameLength = reader.readInt(2); - localExtraFieldsLength = reader.readInt(2); // can't be sure this will be the same as the central dir - // the fileName is stored as binary data, the handleUTF8 method will take care of the encoding. - this.fileName = reader.readData(this.fileNameLength); - reader.skip(localExtraFieldsLength); - - if (this.compressedSize === -1 || this.uncompressedSize === -1) { - throw new Error("Bug or corrupted zip : didn't get enough information from the central directory " + "(compressedSize === -1 || uncompressedSize === -1)"); - } - - compression = findCompression(this.compressionMethod); - if (compression === null) { // no compression found - throw new Error("Corrupted zip : compression " + utils.pretty(this.compressionMethod) + " unknown (inner file : " + utils.transformTo("string", this.fileName) + ")"); - } - this.decompressed = new CompressedObject(this.compressedSize, this.uncompressedSize, this.crc32, compression, reader.readData(this.compressedSize)); - }, - - /** - * Read the central part of a zip file and add the info in this object. - * @param {DataReader} reader the reader to use. - */ - readCentralPart: function(reader) { - this.versionMadeBy = reader.readInt(2); - reader.skip(2); - // this.versionNeeded = reader.readInt(2); - this.bitFlag = reader.readInt(2); - this.compressionMethod = reader.readString(2); - this.date = reader.readDate(); - this.crc32 = reader.readInt(4); - this.compressedSize = reader.readInt(4); - this.uncompressedSize = reader.readInt(4); - var fileNameLength = reader.readInt(2); - this.extraFieldsLength = reader.readInt(2); - this.fileCommentLength = reader.readInt(2); - this.diskNumberStart = reader.readInt(2); - this.internalFileAttributes = reader.readInt(2); - this.externalFileAttributes = reader.readInt(4); - this.localHeaderOffset = reader.readInt(4); - - if (this.isEncrypted()) { - throw new Error("Encrypted zip are not supported"); - } - - // will be read in the local part, see the comments there - reader.skip(fileNameLength); - this.readExtraFields(reader); - this.parseZIP64ExtraField(reader); - this.fileComment = reader.readData(this.fileCommentLength); - }, - - /** - * Parse the external file attributes and get the unix/dos permissions. - */ - processAttributes: function () { - this.unixPermissions = null; - this.dosPermissions = null; - var madeBy = this.versionMadeBy >> 8; - - // Check if we have the DOS directory flag set. - // We look for it in the DOS and UNIX permissions - // but some unknown platform could set it as a compatibility flag. - this.dir = this.externalFileAttributes & 0x0010 ? true : false; - - if(madeBy === MADE_BY_DOS) { - // first 6 bits (0 to 5) - this.dosPermissions = this.externalFileAttributes & 0x3F; - } - - if(madeBy === MADE_BY_UNIX) { - this.unixPermissions = (this.externalFileAttributes >> 16) & 0xFFFF; - // the octal permissions are in (this.unixPermissions & 0x01FF).toString(8); - } - - // fail safe : if the name ends with a / it probably means a folder - if (!this.dir && this.fileNameStr.slice(-1) === '/') { - this.dir = true; - } - }, - - /** - * Parse the ZIP64 extra field and merge the info in the current ZipEntry. - * @param {DataReader} reader the reader to use. - */ - parseZIP64ExtraField: function(reader) { - - if (!this.extraFields[0x0001]) { - return; - } - - // should be something, preparing the extra reader - var extraReader = readerFor(this.extraFields[0x0001].value); - - // I really hope that these 64bits integer can fit in 32 bits integer, because js - // won't let us have more. - if (this.uncompressedSize === utils.MAX_VALUE_32BITS) { - this.uncompressedSize = extraReader.readInt(8); - } - if (this.compressedSize === utils.MAX_VALUE_32BITS) { - this.compressedSize = extraReader.readInt(8); - } - if (this.localHeaderOffset === utils.MAX_VALUE_32BITS) { - this.localHeaderOffset = extraReader.readInt(8); - } - if (this.diskNumberStart === utils.MAX_VALUE_32BITS) { - this.diskNumberStart = extraReader.readInt(4); - } - }, - /** - * Read the central part of a zip file and add the info in this object. - * @param {DataReader} reader the reader to use. - */ - readExtraFields: function(reader) { - var end = reader.index + this.extraFieldsLength, - extraFieldId, - extraFieldLength, - extraFieldValue; - - if (!this.extraFields) { - this.extraFields = {}; - } - - while (reader.index + 4 < end) { - extraFieldId = reader.readInt(2); - extraFieldLength = reader.readInt(2); - extraFieldValue = reader.readData(extraFieldLength); - - this.extraFields[extraFieldId] = { - id: extraFieldId, - length: extraFieldLength, - value: extraFieldValue - }; - } - - reader.setIndex(end); - }, - /** - * Apply an UTF8 transformation if needed. - */ - handleUTF8: function() { - var decodeParamType = support.uint8array ? "uint8array" : "array"; - if (this.useUTF8()) { - this.fileNameStr = utf8.utf8decode(this.fileName); - this.fileCommentStr = utf8.utf8decode(this.fileComment); - } else { - var upath = this.findExtraFieldUnicodePath(); - if (upath !== null) { - this.fileNameStr = upath; - } else { - // ASCII text or unsupported code page - var fileNameByteArray = utils.transformTo(decodeParamType, this.fileName); - this.fileNameStr = this.loadOptions.decodeFileName(fileNameByteArray); - } - - var ucomment = this.findExtraFieldUnicodeComment(); - if (ucomment !== null) { - this.fileCommentStr = ucomment; - } else { - // ASCII text or unsupported code page - var commentByteArray = utils.transformTo(decodeParamType, this.fileComment); - this.fileCommentStr = this.loadOptions.decodeFileName(commentByteArray); - } - } - }, - - /** - * Find the unicode path declared in the extra field, if any. - * @return {String} the unicode path, null otherwise. - */ - findExtraFieldUnicodePath: function() { - var upathField = this.extraFields[0x7075]; - if (upathField) { - var extraReader = readerFor(upathField.value); - - // wrong version - if (extraReader.readInt(1) !== 1) { - return null; - } - - // the crc of the filename changed, this field is out of date. - if (crc32fn(this.fileName) !== extraReader.readInt(4)) { - return null; - } - - return utf8.utf8decode(extraReader.readData(upathField.length - 5)); - } - return null; - }, - - /** - * Find the unicode comment declared in the extra field, if any. - * @return {String} the unicode comment, null otherwise. - */ - findExtraFieldUnicodeComment: function() { - var ucommentField = this.extraFields[0x6375]; - if (ucommentField) { - var extraReader = readerFor(ucommentField.value); - - // wrong version - if (extraReader.readInt(1) !== 1) { - return null; - } - - // the crc of the comment changed, this field is out of date. - if (crc32fn(this.fileComment) !== extraReader.readInt(4)) { - return null; - } - - return utf8.utf8decode(extraReader.readData(ucommentField.length - 5)); - } - return null; - } -}; -module.exports = ZipEntry; - -},{"./compressedObject":2,"./compressions":3,"./crc32":4,"./reader/readerFor":22,"./support":30,"./utf8":31,"./utils":32}],35:[function(require,module,exports){ -'use strict'; - -var StreamHelper = require('./stream/StreamHelper'); -var DataWorker = require('./stream/DataWorker'); -var utf8 = require('./utf8'); -var CompressedObject = require('./compressedObject'); -var GenericWorker = require('./stream/GenericWorker'); - -/** - * A simple object representing a file in the zip file. - * @constructor - * @param {string} name the name of the file - * @param {String|ArrayBuffer|Uint8Array|Buffer} data the data - * @param {Object} options the options of the file - */ -var ZipObject = function(name, data, options) { - this.name = name; - this.dir = options.dir; - this.date = options.date; - this.comment = options.comment; - this.unixPermissions = options.unixPermissions; - this.dosPermissions = options.dosPermissions; - - this._data = data; - this._dataBinary = options.binary; - // keep only the compression - this.options = { - compression : options.compression, - compressionOptions : options.compressionOptions - }; -}; - -ZipObject.prototype = { - /** - * Create an internal stream for the content of this object. - * @param {String} type the type of each chunk. - * @return StreamHelper the stream. - */ - internalStream: function (type) { - var result = null, outputType = "string"; - try { - if (!type) { - throw new Error("No output type specified."); - } - outputType = type.toLowerCase(); - var askUnicodeString = outputType === "string" || outputType === "text"; - if (outputType === "binarystring" || outputType === "text") { - outputType = "string"; - } - result = this._decompressWorker(); - - var isUnicodeString = !this._dataBinary; - - if (isUnicodeString && !askUnicodeString) { - result = result.pipe(new utf8.Utf8EncodeWorker()); - } - if (!isUnicodeString && askUnicodeString) { - result = result.pipe(new utf8.Utf8DecodeWorker()); - } - } catch (e) { - result = new GenericWorker("error"); - result.error(e); - } - - return new StreamHelper(result, outputType, ""); - }, - - /** - * Prepare the content in the asked type. - * @param {String} type the type of the result. - * @param {Function} onUpdate a function to call on each internal update. - * @return Promise the promise of the result. - */ - async: function (type, onUpdate) { - return this.internalStream(type).accumulate(onUpdate); - }, - - /** - * Prepare the content as a nodejs stream. - * @param {String} type the type of each chunk. - * @param {Function} onUpdate a function to call on each internal update. - * @return Stream the stream. - */ - nodeStream: function (type, onUpdate) { - return this.internalStream(type || "nodebuffer").toNodejsStream(onUpdate); - }, - - /** - * Return a worker for the compressed content. - * @private - * @param {Object} compression the compression object to use. - * @param {Object} compressionOptions the options to use when compressing. - * @return Worker the worker. - */ - _compressWorker: function (compression, compressionOptions) { - if ( - this._data instanceof CompressedObject && - this._data.compression.magic === compression.magic - ) { - return this._data.getCompressedWorker(); - } else { - var result = this._decompressWorker(); - if(!this._dataBinary) { - result = result.pipe(new utf8.Utf8EncodeWorker()); - } - return CompressedObject.createWorkerFrom(result, compression, compressionOptions); - } - }, - /** - * Return a worker for the decompressed content. - * @private - * @return Worker the worker. - */ - _decompressWorker : function () { - if (this._data instanceof CompressedObject) { - return this._data.getContentWorker(); - } else if (this._data instanceof GenericWorker) { - return this._data; - } else { - return new DataWorker(this._data); - } - } -}; - -var removedMethods = ["asText", "asBinary", "asNodeBuffer", "asUint8Array", "asArrayBuffer"]; -var removedFn = function () { - throw new Error("This method has been removed in JSZip 3.0, please check the upgrade guide."); -}; - -for(var i = 0; i < removedMethods.length; i++) { - ZipObject.prototype[removedMethods[i]] = removedFn; -} -module.exports = ZipObject; - -},{"./compressedObject":2,"./stream/DataWorker":27,"./stream/GenericWorker":28,"./stream/StreamHelper":29,"./utf8":31}],36:[function(require,module,exports){ -(function (global){ -'use strict'; -var Mutation = global.MutationObserver || global.WebKitMutationObserver; - -var scheduleDrain; - -{ - if (Mutation) { - var called = 0; - var observer = new Mutation(nextTick); - var element = global.document.createTextNode(''); - observer.observe(element, { - characterData: true - }); - scheduleDrain = function () { - element.data = (called = ++called % 2); - }; - } else if (!global.setImmediate && typeof global.MessageChannel !== 'undefined') { - var channel = new global.MessageChannel(); - channel.port1.onmessage = nextTick; - scheduleDrain = function () { - channel.port2.postMessage(0); - }; - } else if ('document' in global && 'onreadystatechange' in global.document.createElement('script')) { - scheduleDrain = function () { - - // Create a - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface Connection

    -
    -
    -
    -
      -
    • -
      -
      All Superinterfaces:
      -
      AutoCloseable
      -
      -
      -
      All Known Implementing Classes:
      -
      FlussConnection
      -
      -
      -
      @PublicEvolving
      -@ThreadSafe
      -public interface Connection
      -extends AutoCloseable
      -
      A cluster connection encapsulating lower level individual connections to actual Fluss servers. - Connections are instantiated through the ConnectionFactory class. The lifecycle of the - connection is managed by the caller, who has to close() the connection to release the - resources. The connection object contains logic to find the Coordinator, locate table buckets out - on the cluster, keeps a cache of locations and then knows how to re-calibrate after they move. - The individual connections to servers, meta cache, etc. are all shared by the Table and - Admin instances obtained from this connection. - -

      Connection creation is a heavy-weight operation. Connection implementations are thread-safe, - so that the client can create a connection once, and share it with different threads. - Table and Admin instances, on the other hand, are light-weight and are not thread-safe. - Typically, a single connection per client application is instantiated and every thread will - obtain its own Table instance. Caching or pooling of Table and Admin is - not recommended.

      -
      -
      Since:
      -
      0.1
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Summary

        - - - - - - - - - - - - - - - - - - - - - - - - - - - -
        All Methods Instance Methods Abstract Methods 
        Modifier and TypeMethodDescription
        voidclose() -
        Close the connection and release all resources.
        -
        AdmingetAdmin() -
        Retrieve a new Admin client to administer a Fluss cluster.
        -
        org.apache.fluss.config.ConfigurationgetConfiguration() -
        Retrieve the configuration used to create this connection.
        -
        TablegetTable​(org.apache.fluss.metadata.TablePath tablePath) -
        Retrieve a new Table client to operate data in table.
        -
        -
      • -
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getConfiguration

          -
          org.apache.fluss.config.Configuration getConfiguration()
          -
          Retrieve the configuration used to create this connection.
          -
        • -
        - - - -
          -
        • -

          getAdmin

          -
          Admin getAdmin()
          -
          Retrieve a new Admin client to administer a Fluss cluster.
          -
        • -
        - - - -
          -
        • -

          getTable

          -
          Table getTable​(org.apache.fluss.metadata.TablePath tablePath)
          -
          Retrieve a new Table client to operate data in table.
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/ConnectionFactory.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/ConnectionFactory.html deleted file mode 100644 index 5848e8c490..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/ConnectionFactory.html +++ /dev/null @@ -1,314 +0,0 @@ - - - - - -ConnectionFactory (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class ConnectionFactory

    -
    -
    - -
    -
      -
    • -
      -
      @PublicEvolving
      -public class ConnectionFactory
      -extends Object
      -
      A non-instantiable class that manages creation of Connections. Managing the lifecycle of - the Connections to the cluster is the responsibility of the caller. From a Connection, Admin implementations are retrieved with Connection.getAdmin().
      -
      -
      Since:
      -
      0.1
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          createConnection

          -
          public static Connection createConnection​(org.apache.fluss.config.Configuration conf)
          -
          Creates a new Connection to the Fluss cluster. The given configuration at least needs - to contain "bootstrap.servers" to discover the Fluss cluster. Here is a simple example: - -
          
          - Configuration conf = new Configuration();
          - conf.setString("bootstrap.servers", "localhost:9092");
          - Connection connection = ConnectionFactory.createConnection(conf);
          - Admin admin = connection.getAdmin();
          - try {
          -    // Use the admin as needed, for a single operation and a single thread
          -  } finally {
          -    admin.close();
          -    connection.close();
          -  }
          - 
          -
        • -
        - - - -
          -
        • -

          createConnection

          -
          public static Connection createConnection​(org.apache.fluss.config.Configuration conf,
          -                                          org.apache.fluss.metrics.registry.MetricRegistry metricRegistry)
          -
          Create a new Connection to the Fluss cluster with registering metrics to the given - metricRegistry. It's mainly used for client to register metrics to external metrics - system. - -

          See more comments in method createConnection(Configuration)

          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/FlussConnection.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/FlussConnection.html deleted file mode 100644 index 62fa30c299..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/FlussConnection.html +++ /dev/null @@ -1,435 +0,0 @@ - - - - - -FlussConnection (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlussConnection

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      AutoCloseable, Connection
      -
      -
      -
      public final class FlussConnection
      -extends Object
      -implements Connection
      -
      A connection to Fluss cluster, and holds the client session resources.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getConfiguration

          -
          public org.apache.fluss.config.Configuration getConfiguration()
          -
          Description copied from interface: Connection
          -
          Retrieve the configuration used to create this connection.
          -
          -
          Specified by:
          -
          getConfiguration in interface Connection
          -
          -
        • -
        - - - -
          -
        • -

          getAdmin

          -
          public Admin getAdmin()
          -
          Description copied from interface: Connection
          -
          Retrieve a new Admin client to administer a Fluss cluster.
          -
          -
          Specified by:
          -
          getAdmin in interface Connection
          -
          -
        • -
        - - - -
          -
        • -

          getTable

          -
          public Table getTable​(org.apache.fluss.metadata.TablePath tablePath)
          -
          Description copied from interface: Connection
          -
          Retrieve a new Table client to operate data in table.
          -
          -
          Specified by:
          -
          getTable in interface Connection
          -
          -
        • -
        - - - -
          -
        • -

          getRpcClient

          -
          public org.apache.fluss.rpc.RpcClient getRpcClient()
          -
        • -
        - - - - - - - -
          -
        • -

          getClientMetricGroup

          -
          public org.apache.fluss.rpc.metrics.ClientMetricGroup getClientMetricGroup()
          -
        • -
        - - - -
          -
        • -

          getOrCreateWriterClient

          -
          public WriterClient getOrCreateWriterClient()
          -
        • -
        - - - -
          -
        • -

          getOrCreateLookupClient

          -
          public LookupClient getOrCreateLookupClient()
          -
        • -
        - - - -
          -
        • -

          getOrCreateRemoteFileDownloader

          -
          public RemoteFileDownloader getOrCreateRemoteFileDownloader()
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/Admin.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/Admin.html deleted file mode 100644 index 8c9115ccd0..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/Admin.html +++ /dev/null @@ -1,1040 +0,0 @@ - - - - - -Admin (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface Admin

    -
    -
    -
    -
      -
    • -
      -
      All Superinterfaces:
      -
      AutoCloseable
      -
      -
      -
      All Known Implementing Classes:
      -
      FlussAdmin
      -
      -
      -
      @PublicEvolving
      -public interface Admin
      -extends AutoCloseable
      -
      The administrative client for Fluss, which supports managing and inspecting tables, servers, - configurations and ACLs.
      -
      -
      Since:
      -
      0.1
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getServerNodes

          -
          CompletableFuture<List<org.apache.fluss.cluster.ServerNode>> getServerNodes()
          -
          Get the current server node information. asynchronously.
          -
        • -
        - - - -
          -
        • -

          getTableSchema

          -
          CompletableFuture<org.apache.fluss.metadata.SchemaInfo> getTableSchema​(org.apache.fluss.metadata.TablePath tablePath)
          -
          Get the latest table schema of the given table asynchronously. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • TableNotExistException if the table does not exist. -
          -
          -
          Parameters:
          -
          tablePath - the table path of the table.
          -
          -
        • -
        - - - -
          -
        • -

          getTableSchema

          -
          CompletableFuture<org.apache.fluss.metadata.SchemaInfo> getTableSchema​(org.apache.fluss.metadata.TablePath tablePath,
          -                                                                       int schemaId)
          -
          Get the specific table schema of the given table by schema id asynchronously. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • TableNotExistException if the table does not exist. -
          • SchemaNotExistException if the schema does not exist. -
          -
          -
          Parameters:
          -
          tablePath - the table path of the table.
          -
          -
        • -
        - - - -
          -
        • -

          createDatabase

          -
          CompletableFuture<Void> createDatabase​(String databaseName,
          -                                       org.apache.fluss.metadata.DatabaseDescriptor databaseDescriptor,
          -                                       boolean ignoreIfExists)
          -
          Create a new database asynchronously. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • DatabaseAlreadyExistException if the database already exists and - ignoreIfExists is false. -
          -
          -
          Parameters:
          -
          databaseName - The name of the database to create.
          -
          databaseDescriptor - The descriptor of the database to create.
          -
          ignoreIfExists - Flag to specify behavior when a database with the given name already - exists: if set to false, throw a DatabaseAlreadyExistException, if set to true, do - nothing.
          -
          Throws:
          -
          org.apache.fluss.exception.InvalidDatabaseException - if the database name is invalid, e.g., contains illegal - characters, or exceeds the maximum length.
          -
          -
        • -
        - - - -
          -
        • -

          getDatabaseInfo

          -
          CompletableFuture<org.apache.fluss.metadata.DatabaseInfo> getDatabaseInfo​(String databaseName)
          -
          Get the database with the given database name asynchronously. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • DatabaseNotExistException if the database does not exist. -
          -
          -
          Parameters:
          -
          databaseName - The database name of the database.
          -
          -
        • -
        - - - -
          -
        • -

          dropDatabase

          -
          CompletableFuture<Void> dropDatabase​(String databaseName,
          -                                     boolean ignoreIfNotExists,
          -                                     boolean cascade)
          -
          Drop the database with the given name asynchronously. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • DatabaseNotExistException if the database does not exist and - ignoreIfNotExists is false. -
          • DatabaseNotEmptyException if the database is not empty and cascade is - false. -
          -
          -
          Parameters:
          -
          databaseName - The name of the database to delete.
          -
          ignoreIfNotExists - Flag to specify behavior when a database with the given name does - not exist: if set to false, throw a DatabaseNotExistException, if set to true, do - nothing.
          -
          cascade - Flag to specify whether to delete all tables in the database.
          -
          -
        • -
        - - - -
          -
        • -

          databaseExists

          -
          CompletableFuture<Boolean> databaseExists​(String databaseName)
          -
          Get whether database exists asynchronously.
          -
          -
          Parameters:
          -
          databaseName - The name of the database to check.
          -
          -
        • -
        - - - - - - - -
          -
        • -

          createTable

          -
          CompletableFuture<Void> createTable​(org.apache.fluss.metadata.TablePath tablePath,
          -                                    org.apache.fluss.metadata.TableDescriptor tableDescriptor,
          -                                    boolean ignoreIfExists)
          -                             throws org.apache.fluss.exception.InvalidTableException,
          -                                    org.apache.fluss.exception.InvalidDatabaseException
          -
          Create a new table asynchronously. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • DatabaseNotExistException if the database in the table path does not exist. -
          • TableAlreadyExistException if the table already exists and - ignoreIfExists is false. -
          • InvalidReplicationFactorException if the table's replication factor is larger - than the number of available tablet servers. -
          -
          -
          Parameters:
          -
          tablePath - The tablePath of the table.
          -
          tableDescriptor - The table to create.
          -
          Throws:
          -
          org.apache.fluss.exception.InvalidTableException - if the table name is invalid, e.g., contains illegal - characters, or exceeds the maximum length.
          -
          org.apache.fluss.exception.InvalidDatabaseException - if the database name is invalid, e.g., contains illegal - characters, or exceeds the maximum length.
          -
          -
        • -
        - - - -
          -
        • -

          getTableInfo

          -
          CompletableFuture<org.apache.fluss.metadata.TableInfo> getTableInfo​(org.apache.fluss.metadata.TablePath tablePath)
          -
          Get the table with the given table path asynchronously. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • TableNotExistException if the table does not exist. -
          -
          -
          Parameters:
          -
          tablePath - The table path of the table.
          -
          -
        • -
        - - - -
          -
        • -

          dropTable

          -
          CompletableFuture<Void> dropTable​(org.apache.fluss.metadata.TablePath tablePath,
          -                                  boolean ignoreIfNotExists)
          -
          Drop the table with the given table path asynchronously. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • TableNotExistException if the table does not exist and - ignoreIfNotExists is false. -
          -
          -
          Parameters:
          -
          tablePath - The table path of the table.
          -
          ignoreIfNotExists - Flag to specify behavior when a table with the given name does not - exist: if set to false, throw a TableNotExistException, if set to true, do nothing.
          -
          -
        • -
        - - - -
          -
        • -

          tableExists

          -
          CompletableFuture<Boolean> tableExists​(org.apache.fluss.metadata.TablePath tablePath)
          -
          Get whether table exists asynchronously.
          -
          -
          Parameters:
          -
          tablePath - The table path of the table.
          -
          -
        • -
        - - - -
          -
        • -

          listTables

          -
          CompletableFuture<List<String>> listTables​(String databaseName)
          -
          List all tables in the given database in fluss cluster asynchronously. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • DatabaseNotExistException if the database does not exist. -
          -
          -
          Parameters:
          -
          databaseName - The name of the database.
          -
          -
        • -
        - - - -
          -
        • -

          listPartitionInfos

          -
          CompletableFuture<List<org.apache.fluss.metadata.PartitionInfo>> listPartitionInfos​(org.apache.fluss.metadata.TablePath tablePath)
          -
          List all partitions in the given table in fluss cluster asynchronously. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • TableNotExistException if the table does not exist. -
          • TableNotPartitionedException if the table is not partitioned. -
          -
          -
          Parameters:
          -
          tablePath - The path of the table.
          -
          -
        • -
        - - - -
          -
        • -

          listPartitionInfos

          -
          CompletableFuture<List<org.apache.fluss.metadata.PartitionInfo>> listPartitionInfos​(org.apache.fluss.metadata.TablePath tablePath,
          -                                                                                    org.apache.fluss.metadata.PartitionSpec partialPartitionSpec)
          -
          List all partitions in fluss cluster that are under the given table and the given partial - PartitionSpec asynchronously. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • TableNotExistException if the table does not exist. -
          • TableNotPartitionedException if the table is not partitioned. -
          • InvalidPartitionException if the input partition spec is invalid. -
          -
          -
          Parameters:
          -
          tablePath - The path of the table.
          -
          partialPartitionSpec - Part of table partition spec
          -
          -
        • -
        - - - -
          -
        • -

          createPartition

          -
          CompletableFuture<Void> createPartition​(org.apache.fluss.metadata.TablePath tablePath,
          -                                        org.apache.fluss.metadata.PartitionSpec partitionSpec,
          -                                        boolean ignoreIfExists)
          -
          Create a new partition for a partitioned table. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • TableNotExistException if the table does not exist. -
          • TableNotPartitionedException if the table is not partitioned. -
          • PartitionAlreadyExistsException if the partition already exists and - ignoreIfExists is false. -
          • InvalidPartitionException if the input partition spec is invalid. -
          • TooManyPartitionsException if the number of partitions is larger than the - maximum number of partitions of one table, see ConfigOptions.MAX_PARTITION_NUM. -
          • TooManyBucketsException if the number of buckets is larger than the maximum - number of buckets of one table, see ConfigOptions.MAX_BUCKET_NUM. -
          -
          -
          Parameters:
          -
          tablePath - The table path of the table.
          -
          partitionSpec - The partition spec to add.
          -
          ignoreIfExists - Flag to specify behavior when a partition with the given name already - exists: if set to false, throw a PartitionAlreadyExistsException, if set to true, do - nothing.
          -
          -
        • -
        - - - -
          -
        • -

          dropPartition

          -
          CompletableFuture<Void> dropPartition​(org.apache.fluss.metadata.TablePath tablePath,
          -                                      org.apache.fluss.metadata.PartitionSpec partitionSpec,
          -                                      boolean ignoreIfNotExists)
          -
          Drop a partition from a partitioned table. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • TableNotExistException if the table does not exist. -
          • TableNotPartitionedException if the table is not partitioned. -
          • PartitionNotExistException if the partition not exists and - ignoreIfExists is false. -
          • InvalidPartitionException if the input partition spec is invalid. -
          -
          -
          Parameters:
          -
          tablePath - The table path of the table.
          -
          partitionSpec - The partition spec to drop.
          -
          ignoreIfNotExists - Flag to specify behavior when a partition with the given name does - not exist: if set to false, throw a PartitionNotExistException, if set to true, do - nothing.
          -
          -
        • -
        - - - -
          -
        • -

          getLatestKvSnapshots

          -
          CompletableFuture<KvSnapshots> getLatestKvSnapshots​(org.apache.fluss.metadata.TablePath tablePath)
          -
          Get the latest kv snapshots of the given table asynchronously. A kv snapshot is a snapshot of - a bucket of a primary key table at a certain point in time. Therefore, there are at-most - N snapshots for a primary key table, N is the number of buckets. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • TableNotExistException if the table does not exist. -
          • NonPrimaryKeyTableException if the table is not a primary key table. -
          • PartitionNotExistException if the table is partitioned, use getLatestKvSnapshots(TablePath, String) instead to get the latest kv snapshot of a - partition of a partitioned table. -
          • -
          -
          -
          Parameters:
          -
          tablePath - the table path of the table.
          -
          -
        • -
        - - - -
          -
        • -

          getLatestKvSnapshots

          -
          CompletableFuture<KvSnapshots> getLatestKvSnapshots​(org.apache.fluss.metadata.TablePath tablePath,
          -                                                    String partitionName)
          -
          Get the latest kv snapshots of the given table partition asynchronously. A kv snapshot is a - snapshot of a bucket of a primary key table at a certain point in time. Therefore, there are - at-most N snapshots for a partition of a primary key table, N is the number - of buckets. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • TableNotExistException if the table does not exist. -
          • NonPrimaryKeyTableException if the table is not a primary key table. -
          • PartitionNotExistException if the partition does not exist -
          • TableNotPartitionedException if the table is not partitioned, use getLatestKvSnapshots(TablePath) instead to get the latest kv snapshots for a - non-partitioned table. -
          -
          -
          Parameters:
          -
          tablePath - the table path of the table.
          -
          partitionName - the partition name, see ResolvedPartitionSpec.getPartitionName() - for the format of the partition name.
          -
          -
        • -
        - - - -
          -
        • -

          getKvSnapshotMetadata

          -
          CompletableFuture<KvSnapshotMetadata> getKvSnapshotMetadata​(org.apache.fluss.metadata.TableBucket bucket,
          -                                                            long snapshotId)
          -
          Get the kv snapshot metadata of the given kv snapshot asynchronously. The kv snapshot - metadata including the snapshot files for the kv tablet and the log offset for the changelog - at the snapshot time. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • KvSnapshotNotExistException if the snapshot does not exist. -
          -
          -
          Parameters:
          -
          bucket - the table bucket of the kv snapshot.
          -
          snapshotId - the snapshot id.
          -
          -
        • -
        - - - -
          -
        • -

          getLatestLakeSnapshot

          -
          CompletableFuture<LakeSnapshot> getLatestLakeSnapshot​(org.apache.fluss.metadata.TablePath tablePath)
          -
          Get table lake snapshot info of the given table asynchronously. - -

          It'll get the latest snapshot for all the buckets of the table. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • TableNotExistException if the table does not exist. -
          • LakeTableSnapshotNotExistException if no any lake snapshot exist. -
          -
          -
          Parameters:
          -
          tablePath - the table path of the table.
          -
          -
        • -
        - - - -
          -
        • -

          listOffsets

          -
          ListOffsetsResult listOffsets​(org.apache.fluss.metadata.TablePath tablePath,
          -                              Collection<Integer> buckets,
          -                              OffsetSpec offsetSpec)
          -
          List offset for the specified buckets. This operation enables to find the beginning offset, - end offset as well as the offset matching a timestamp in buckets.
          -
          -
          Parameters:
          -
          tablePath - the table path of the table.
          -
          buckets - the buckets to fetch offset.
          -
          offsetSpec - the offset spec to fetch.
          -
          -
        • -
        - - - -
          -
        • -

          listOffsets

          -
          ListOffsetsResult listOffsets​(org.apache.fluss.metadata.TablePath tablePath,
          -                              String partitionName,
          -                              Collection<Integer> buckets,
          -                              OffsetSpec offsetSpec)
          -
          List offset for the specified buckets. This operation enables to find the beginning offset, - end offset as well as the offset matching a timestamp in buckets.
          -
          -
          Parameters:
          -
          tablePath - the table path of the table.
          -
          partitionName - the partition name of the partition,see ResolvedPartitionSpec.getPartitionName() * for the format of the partition name.
          -
          buckets - the buckets to fetch offset.
          -
          offsetSpec - the offset spec to fetch.
          -
          -
        • -
        - - - -
          -
        • -

          listAcls

          -
          CompletableFuture<Collection<org.apache.fluss.security.acl.AclBinding>> listAcls​(org.apache.fluss.security.acl.AclBindingFilter aclBindingFilter)
          -
          Retrieves ACL entries filtered by principal for the specified resource. - -

          1. Validates the user has 'describe' permission on the resource. 2. Returns entries - matching the principal if permitted; throws an exception otherwise.

          -
          -
          Returns:
          -
          A CompletableFuture containing the filtered ACL entries.
          -
          -
        • -
        - - - -
          -
        • -

          createAcls

          -
          CreateAclsResult createAcls​(Collection<org.apache.fluss.security.acl.AclBinding> aclBindings)
          -
          Creates multiple ACL entries in a single atomic operation. - -

          1. Validates the user has 'alter' permission on the resource. 2. Creates the ACL entries - if valid and permitted. - -

          Each entry in aclBindings must have a valid principal, operation and permission.

          -
          -
          Parameters:
          -
          aclBindings - List of ACL entries to create.
          -
          Returns:
          -
          A CompletableFuture indicating completion of the operation.
          -
          -
        • -
        - - - -
          -
        • -

          dropAcls

          -
          DropAclsResult dropAcls​(Collection<org.apache.fluss.security.acl.AclBindingFilter> filters)
          -
          Removes multiple ACL entries in a single atomic operation. - -

          1. Validates the user has 'alter' permission on the resource. 2. Removes entries only if - they exactly match the provided entries (principal, operation, permission). 3. Does not - remove entries if any of the ACL entries do not exist.

          -
          -
          Parameters:
          -
          filters - List of ACL entries to remove.
          -
          Returns:
          -
          A CompletableFuture indicating completion of the operation.
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/CreateAclsResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/CreateAclsResult.html deleted file mode 100644 index ed5c10e6c1..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/CreateAclsResult.html +++ /dev/null @@ -1,385 +0,0 @@ - - - - - -CreateAclsResult (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class CreateAclsResult

    -
    -
    - -
    -
      -
    • -
      -
      public class CreateAclsResult
      -extends Object
      -
      Represents the result of a batch ACL operation, managing asynchronous completion of individual - ACL operations. - -

      This class tracks the execution status of multiple ACL operations (e.g., create/drop) by - associating each AclBinding with its corresponding CompletableFuture. It - processes RPC responses to complete or fail individual futures based on server-side results.

      -
      -
      Since:
      -
      0.6
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          CreateAclsResult

          -
          public CreateAclsResult​(Collection<org.apache.fluss.security.acl.AclBinding> aclBindings)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - - - - - -
          -
        • -

          getFutures

          -
          public Map<org.apache.fluss.security.acl.AclBinding,​CompletableFuture<Void>> getFutures()
          -
          Gets the map of ACL bindings to their associated futures.
          -
          -
          Returns:
          -
          The map of ACL bindings to futures.
          -
          -
        • -
        - - - -
          -
        • -

          complete

          -
          public void complete​(List<org.apache.fluss.rpc.messages.PbCreateAclRespInfo> pbAclRespInfos)
          -
          Completes individual futures based on RPC response information. - -

          For each PbCreateAclRespInfo in the collection, Completes the future with success - or failure based on the response's error code.

          -
          -
          Parameters:
          -
          pbAclRespInfos - Collection of protobuf response messages containing ACL operation - results.
          -
          -
        • -
        - - - -
          -
        • -

          completeExceptionally

          -
          public void completeExceptionally​(Throwable t)
          -
          Marks all futures as exceptionally completed with the provided throwable. - -

          This method propagates a common exception (e.g., network error) to all tracked futures.

          -
          -
          Parameters:
          -
          t - The throwable to propagate to all futures
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.FilterResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.FilterResult.html deleted file mode 100644 index 515fe92f51..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.FilterResult.html +++ /dev/null @@ -1,292 +0,0 @@ - - - - - -DropAclsResult.FilterResult (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class DropAclsResult.FilterResult

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.admin.DropAclsResult.FilterResult
      • -
      -
    • -
    -
    -
      -
    • -
      -
      Enclosing class:
      -
      DropAclsResult
      -
      -
      -
      public static class DropAclsResult.FilterResult
      -extends Object
      -
      A class containing either the deleted ACL binding or an exception if the delete failed.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          binding

          -
          public org.apache.fluss.security.acl.AclBinding binding()
          -
          Return the deleted ACL binding or null if there was an error.
          -
        • -
        - - - -
          -
        • -

          exception

          -
          @Nullable
          -public org.apache.fluss.exception.ApiException exception()
          -
          Return an exception if the ACL delete was not successful or null if it was.
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.FilterResults.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.FilterResults.html deleted file mode 100644 index cea8bc22d2..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.FilterResults.html +++ /dev/null @@ -1,274 +0,0 @@ - - - - - -DropAclsResult.FilterResults (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class DropAclsResult.FilterResults

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.admin.DropAclsResult.FilterResults
      • -
      -
    • -
    -
    -
      -
    • -
      -
      Enclosing class:
      -
      DropAclsResult
      -
      -
      -
      public static class DropAclsResult.FilterResults
      -extends Object
      -
      A class containing the results of the delete ACLs operation.
      -
    • -
    -
    -
    - -
    -
    - -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.html deleted file mode 100644 index 56535cd81a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/DropAclsResult.html +++ /dev/null @@ -1,352 +0,0 @@ - - - - - -DropAclsResult (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class DropAclsResult

    -
    -
    - -
    -
      -
    • -
      -
      public class DropAclsResult
      -extends Object
      -
      Represents the result of a drop ACLs operation.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          all

          -
          public CompletableFuture<Collection<org.apache.fluss.security.acl.AclBinding>> all()
          -
          Return a future which succeeds only if all the ACLs deletions succeed, and which contains all - the deleted ACLs. Note that it if the filters don't match any ACLs, this is not considered an - error.
          -
        • -
        - - - -
          -
        • -

          completeExceptionally

          -
          public void completeExceptionally​(Throwable t)
          -
        • -
        - - - -
          -
        • -

          complete

          -
          public void complete​(List<org.apache.fluss.rpc.messages.PbDropAclsFilterResult> results)
          -
        • -
        - - - -
          -
        • -

          values

          -
          public Map<org.apache.fluss.security.acl.AclBindingFilter,​CompletableFuture<DropAclsResult.FilterResults>> values()
          -
          Return a map from acl filters to futures which can be used to check the status of the - deletions by each filter.
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/FlussAdmin.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/FlussAdmin.html deleted file mode 100644 index 14af9dc59b..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/FlussAdmin.html +++ /dev/null @@ -1,1173 +0,0 @@ - - - - - -FlussAdmin (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlussAdmin

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      AutoCloseable, Admin
      -
      -
      -
      public class FlussAdmin
      -extends Object
      -implements Admin
      -
      The default implementation of Admin. - -

      This class is thread-safe. The API of this class is evolving, see Admin for details.

      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlussAdmin

          -
          public FlussAdmin​(org.apache.fluss.rpc.RpcClient client,
          -                  MetadataUpdater metadataUpdater)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getServerNodes

          -
          public CompletableFuture<List<org.apache.fluss.cluster.ServerNode>> getServerNodes()
          -
          Description copied from interface: Admin
          -
          Get the current server node information. asynchronously.
          -
          -
          Specified by:
          -
          getServerNodes in interface Admin
          -
          -
        • -
        - - - -
          -
        • -

          getTableSchema

          -
          public CompletableFuture<org.apache.fluss.metadata.SchemaInfo> getTableSchema​(org.apache.fluss.metadata.TablePath tablePath)
          -
          Description copied from interface: Admin
          -
          Get the latest table schema of the given table asynchronously. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • TableNotExistException if the table does not exist. -
          -
          -
          Specified by:
          -
          getTableSchema in interface Admin
          -
          Parameters:
          -
          tablePath - the table path of the table.
          -
          -
        • -
        - - - -
          -
        • -

          getTableSchema

          -
          public CompletableFuture<org.apache.fluss.metadata.SchemaInfo> getTableSchema​(org.apache.fluss.metadata.TablePath tablePath,
          -                                                                              int schemaId)
          -
          Description copied from interface: Admin
          -
          Get the specific table schema of the given table by schema id asynchronously. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • TableNotExistException if the table does not exist. -
          • SchemaNotExistException if the schema does not exist. -
          -
          -
          Specified by:
          -
          getTableSchema in interface Admin
          -
          Parameters:
          -
          tablePath - the table path of the table.
          -
          -
        • -
        - - - -
          -
        • -

          createDatabase

          -
          public CompletableFuture<Void> createDatabase​(String databaseName,
          -                                              org.apache.fluss.metadata.DatabaseDescriptor databaseDescriptor,
          -                                              boolean ignoreIfExists)
          -
          Description copied from interface: Admin
          -
          Create a new database asynchronously. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • DatabaseAlreadyExistException if the database already exists and - ignoreIfExists is false. -
          -
          -
          Specified by:
          -
          createDatabase in interface Admin
          -
          Parameters:
          -
          databaseName - The name of the database to create.
          -
          databaseDescriptor - The descriptor of the database to create.
          -
          ignoreIfExists - Flag to specify behavior when a database with the given name already - exists: if set to false, throw a DatabaseAlreadyExistException, if set to true, do - nothing.
          -
          -
        • -
        - - - -
          -
        • -

          getDatabaseInfo

          -
          public CompletableFuture<org.apache.fluss.metadata.DatabaseInfo> getDatabaseInfo​(String databaseName)
          -
          Description copied from interface: Admin
          -
          Get the database with the given database name asynchronously. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • DatabaseNotExistException if the database does not exist. -
          -
          -
          Specified by:
          -
          getDatabaseInfo in interface Admin
          -
          Parameters:
          -
          databaseName - The database name of the database.
          -
          -
        • -
        - - - -
          -
        • -

          dropDatabase

          -
          public CompletableFuture<Void> dropDatabase​(String databaseName,
          -                                            boolean ignoreIfNotExists,
          -                                            boolean cascade)
          -
          Description copied from interface: Admin
          -
          Drop the database with the given name asynchronously. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • DatabaseNotExistException if the database does not exist and - ignoreIfNotExists is false. -
          • DatabaseNotEmptyException if the database is not empty and cascade is - false. -
          -
          -
          Specified by:
          -
          dropDatabase in interface Admin
          -
          Parameters:
          -
          databaseName - The name of the database to delete.
          -
          ignoreIfNotExists - Flag to specify behavior when a database with the given name does - not exist: if set to false, throw a DatabaseNotExistException, if set to true, do - nothing.
          -
          cascade - Flag to specify whether to delete all tables in the database.
          -
          -
        • -
        - - - -
          -
        • -

          databaseExists

          -
          public CompletableFuture<Boolean> databaseExists​(String databaseName)
          -
          Description copied from interface: Admin
          -
          Get whether database exists asynchronously.
          -
          -
          Specified by:
          -
          databaseExists in interface Admin
          -
          Parameters:
          -
          databaseName - The name of the database to check.
          -
          -
        • -
        - - - - - - - -
          -
        • -

          createTable

          -
          public CompletableFuture<Void> createTable​(org.apache.fluss.metadata.TablePath tablePath,
          -                                           org.apache.fluss.metadata.TableDescriptor tableDescriptor,
          -                                           boolean ignoreIfExists)
          -
          Description copied from interface: Admin
          -
          Create a new table asynchronously. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • DatabaseNotExistException if the database in the table path does not exist. -
          • TableAlreadyExistException if the table already exists and - ignoreIfExists is false. -
          • InvalidReplicationFactorException if the table's replication factor is larger - than the number of available tablet servers. -
          -
          -
          Specified by:
          -
          createTable in interface Admin
          -
          Parameters:
          -
          tablePath - The tablePath of the table.
          -
          tableDescriptor - The table to create.
          -
          -
        • -
        - - - -
          -
        • -

          getTableInfo

          -
          public CompletableFuture<org.apache.fluss.metadata.TableInfo> getTableInfo​(org.apache.fluss.metadata.TablePath tablePath)
          -
          Description copied from interface: Admin
          -
          Get the table with the given table path asynchronously. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • TableNotExistException if the table does not exist. -
          -
          -
          Specified by:
          -
          getTableInfo in interface Admin
          -
          Parameters:
          -
          tablePath - The table path of the table.
          -
          -
        • -
        - - - -
          -
        • -

          dropTable

          -
          public CompletableFuture<Void> dropTable​(org.apache.fluss.metadata.TablePath tablePath,
          -                                         boolean ignoreIfNotExists)
          -
          Description copied from interface: Admin
          -
          Drop the table with the given table path asynchronously. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • TableNotExistException if the table does not exist and - ignoreIfNotExists is false. -
          -
          -
          Specified by:
          -
          dropTable in interface Admin
          -
          Parameters:
          -
          tablePath - The table path of the table.
          -
          ignoreIfNotExists - Flag to specify behavior when a table with the given name does not - exist: if set to false, throw a TableNotExistException, if set to true, do nothing.
          -
          -
        • -
        - - - -
          -
        • -

          tableExists

          -
          public CompletableFuture<Boolean> tableExists​(org.apache.fluss.metadata.TablePath tablePath)
          -
          Description copied from interface: Admin
          -
          Get whether table exists asynchronously.
          -
          -
          Specified by:
          -
          tableExists in interface Admin
          -
          Parameters:
          -
          tablePath - The table path of the table.
          -
          -
        • -
        - - - -
          -
        • -

          listTables

          -
          public CompletableFuture<List<String>> listTables​(String databaseName)
          -
          Description copied from interface: Admin
          -
          List all tables in the given database in fluss cluster asynchronously. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • DatabaseNotExistException if the database does not exist. -
          -
          -
          Specified by:
          -
          listTables in interface Admin
          -
          Parameters:
          -
          databaseName - The name of the database.
          -
          -
        • -
        - - - -
          -
        • -

          listPartitionInfos

          -
          public CompletableFuture<List<org.apache.fluss.metadata.PartitionInfo>> listPartitionInfos​(org.apache.fluss.metadata.TablePath tablePath)
          -
          Description copied from interface: Admin
          -
          List all partitions in the given table in fluss cluster asynchronously. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • TableNotExistException if the table does not exist. -
          • TableNotPartitionedException if the table is not partitioned. -
          -
          -
          Specified by:
          -
          listPartitionInfos in interface Admin
          -
          Parameters:
          -
          tablePath - The path of the table.
          -
          -
        • -
        - - - -
          -
        • -

          listPartitionInfos

          -
          public CompletableFuture<List<org.apache.fluss.metadata.PartitionInfo>> listPartitionInfos​(org.apache.fluss.metadata.TablePath tablePath,
          -                                                                                           org.apache.fluss.metadata.PartitionSpec partitionSpec)
          -
          Description copied from interface: Admin
          -
          List all partitions in fluss cluster that are under the given table and the given partial - PartitionSpec asynchronously. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • TableNotExistException if the table does not exist. -
          • TableNotPartitionedException if the table is not partitioned. -
          • InvalidPartitionException if the input partition spec is invalid. -
          -
          -
          Specified by:
          -
          listPartitionInfos in interface Admin
          -
          Parameters:
          -
          tablePath - The path of the table.
          -
          partitionSpec - Part of table partition spec
          -
          -
        • -
        - - - -
          -
        • -

          createPartition

          -
          public CompletableFuture<Void> createPartition​(org.apache.fluss.metadata.TablePath tablePath,
          -                                               org.apache.fluss.metadata.PartitionSpec partitionSpec,
          -                                               boolean ignoreIfExists)
          -
          Description copied from interface: Admin
          -
          Create a new partition for a partitioned table. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • TableNotExistException if the table does not exist. -
          • TableNotPartitionedException if the table is not partitioned. -
          • PartitionAlreadyExistsException if the partition already exists and - ignoreIfExists is false. -
          • InvalidPartitionException if the input partition spec is invalid. -
          • TooManyPartitionsException if the number of partitions is larger than the - maximum number of partitions of one table, see ConfigOptions.MAX_PARTITION_NUM. -
          • TooManyBucketsException if the number of buckets is larger than the maximum - number of buckets of one table, see ConfigOptions.MAX_BUCKET_NUM. -
          -
          -
          Specified by:
          -
          createPartition in interface Admin
          -
          Parameters:
          -
          tablePath - The table path of the table.
          -
          partitionSpec - The partition spec to add.
          -
          ignoreIfExists - Flag to specify behavior when a partition with the given name already - exists: if set to false, throw a PartitionAlreadyExistsException, if set to true, do - nothing.
          -
          -
        • -
        - - - -
          -
        • -

          dropPartition

          -
          public CompletableFuture<Void> dropPartition​(org.apache.fluss.metadata.TablePath tablePath,
          -                                             org.apache.fluss.metadata.PartitionSpec partitionSpec,
          -                                             boolean ignoreIfNotExists)
          -
          Description copied from interface: Admin
          -
          Drop a partition from a partitioned table. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • TableNotExistException if the table does not exist. -
          • TableNotPartitionedException if the table is not partitioned. -
          • PartitionNotExistException if the partition not exists and - ignoreIfExists is false. -
          • InvalidPartitionException if the input partition spec is invalid. -
          -
          -
          Specified by:
          -
          dropPartition in interface Admin
          -
          Parameters:
          -
          tablePath - The table path of the table.
          -
          partitionSpec - The partition spec to drop.
          -
          ignoreIfNotExists - Flag to specify behavior when a partition with the given name does - not exist: if set to false, throw a PartitionNotExistException, if set to true, do - nothing.
          -
          -
        • -
        - - - -
          -
        • -

          getLatestKvSnapshots

          -
          public CompletableFuture<KvSnapshots> getLatestKvSnapshots​(org.apache.fluss.metadata.TablePath tablePath)
          -
          Description copied from interface: Admin
          -
          Get the latest kv snapshots of the given table asynchronously. A kv snapshot is a snapshot of - a bucket of a primary key table at a certain point in time. Therefore, there are at-most - N snapshots for a primary key table, N is the number of buckets. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • TableNotExistException if the table does not exist. -
          • NonPrimaryKeyTableException if the table is not a primary key table. -
          • PartitionNotExistException if the table is partitioned, use Admin.getLatestKvSnapshots(TablePath, String) instead to get the latest kv snapshot of a - partition of a partitioned table. -
          • -
          -
          -
          Specified by:
          -
          getLatestKvSnapshots in interface Admin
          -
          Parameters:
          -
          tablePath - the table path of the table.
          -
          -
        • -
        - - - -
          -
        • -

          getLatestKvSnapshots

          -
          public CompletableFuture<KvSnapshots> getLatestKvSnapshots​(org.apache.fluss.metadata.TablePath tablePath,
          -                                                           String partitionName)
          -
          Description copied from interface: Admin
          -
          Get the latest kv snapshots of the given table partition asynchronously. A kv snapshot is a - snapshot of a bucket of a primary key table at a certain point in time. Therefore, there are - at-most N snapshots for a partition of a primary key table, N is the number - of buckets. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • TableNotExistException if the table does not exist. -
          • NonPrimaryKeyTableException if the table is not a primary key table. -
          • PartitionNotExistException if the partition does not exist -
          • TableNotPartitionedException if the table is not partitioned, use Admin.getLatestKvSnapshots(TablePath) instead to get the latest kv snapshots for a - non-partitioned table. -
          -
          -
          Specified by:
          -
          getLatestKvSnapshots in interface Admin
          -
          Parameters:
          -
          tablePath - the table path of the table.
          -
          partitionName - the partition name, see ResolvedPartitionSpec.getPartitionName() - for the format of the partition name.
          -
          -
        • -
        - - - -
          -
        • -

          getKvSnapshotMetadata

          -
          public CompletableFuture<KvSnapshotMetadata> getKvSnapshotMetadata​(org.apache.fluss.metadata.TableBucket bucket,
          -                                                                   long snapshotId)
          -
          Description copied from interface: Admin
          -
          Get the kv snapshot metadata of the given kv snapshot asynchronously. The kv snapshot - metadata including the snapshot files for the kv tablet and the log offset for the changelog - at the snapshot time. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • KvSnapshotNotExistException if the snapshot does not exist. -
          -
          -
          Specified by:
          -
          getKvSnapshotMetadata in interface Admin
          -
          Parameters:
          -
          bucket - the table bucket of the kv snapshot.
          -
          snapshotId - the snapshot id.
          -
          -
        • -
        - - - -
          -
        • -

          getLatestLakeSnapshot

          -
          public CompletableFuture<LakeSnapshot> getLatestLakeSnapshot​(org.apache.fluss.metadata.TablePath tablePath)
          -
          Description copied from interface: Admin
          -
          Get table lake snapshot info of the given table asynchronously. - -

          It'll get the latest snapshot for all the buckets of the table. - -

          The following exceptions can be anticipated when calling get() on returned future. - -

            -
          • TableNotExistException if the table does not exist. -
          • LakeTableSnapshotNotExistException if no any lake snapshot exist. -
          -
          -
          Specified by:
          -
          getLatestLakeSnapshot in interface Admin
          -
          Parameters:
          -
          tablePath - the table path of the table.
          -
          -
        • -
        - - - -
          -
        • -

          listOffsets

          -
          public ListOffsetsResult listOffsets​(org.apache.fluss.metadata.TablePath tablePath,
          -                                     Collection<Integer> buckets,
          -                                     OffsetSpec offsetSpec)
          -
          Description copied from interface: Admin
          -
          List offset for the specified buckets. This operation enables to find the beginning offset, - end offset as well as the offset matching a timestamp in buckets.
          -
          -
          Specified by:
          -
          listOffsets in interface Admin
          -
          Parameters:
          -
          tablePath - the table path of the table.
          -
          buckets - the buckets to fetch offset.
          -
          offsetSpec - the offset spec to fetch.
          -
          -
        • -
        - - - -
          -
        • -

          listOffsets

          -
          public ListOffsetsResult listOffsets​(org.apache.fluss.metadata.TablePath tablePath,
          -                                     String partitionName,
          -                                     Collection<Integer> buckets,
          -                                     OffsetSpec offsetSpec)
          -
          Description copied from interface: Admin
          -
          List offset for the specified buckets. This operation enables to find the beginning offset, - end offset as well as the offset matching a timestamp in buckets.
          -
          -
          Specified by:
          -
          listOffsets in interface Admin
          -
          Parameters:
          -
          tablePath - the table path of the table.
          -
          partitionName - the partition name of the partition,see ResolvedPartitionSpec.getPartitionName() * for the format of the partition name.
          -
          buckets - the buckets to fetch offset.
          -
          offsetSpec - the offset spec to fetch.
          -
          -
        • -
        - - - -
          -
        • -

          listAcls

          -
          public CompletableFuture<Collection<org.apache.fluss.security.acl.AclBinding>> listAcls​(org.apache.fluss.security.acl.AclBindingFilter aclBindingFilter)
          -
          Description copied from interface: Admin
          -
          Retrieves ACL entries filtered by principal for the specified resource. - -

          1. Validates the user has 'describe' permission on the resource. 2. Returns entries - matching the principal if permitted; throws an exception otherwise.

          -
          -
          Specified by:
          -
          listAcls in interface Admin
          -
          Returns:
          -
          A CompletableFuture containing the filtered ACL entries.
          -
          -
        • -
        - - - -
          -
        • -

          createAcls

          -
          public CreateAclsResult createAcls​(Collection<org.apache.fluss.security.acl.AclBinding> aclBindings)
          -
          Description copied from interface: Admin
          -
          Creates multiple ACL entries in a single atomic operation. - -

          1. Validates the user has 'alter' permission on the resource. 2. Creates the ACL entries - if valid and permitted. - -

          Each entry in aclBindings must have a valid principal, operation and permission.

          -
          -
          Specified by:
          -
          createAcls in interface Admin
          -
          Parameters:
          -
          aclBindings - List of ACL entries to create.
          -
          Returns:
          -
          A CompletableFuture indicating completion of the operation.
          -
          -
        • -
        - - - -
          -
        • -

          dropAcls

          -
          public DropAclsResult dropAcls​(Collection<org.apache.fluss.security.acl.AclBindingFilter> filters)
          -
          Description copied from interface: Admin
          -
          Removes multiple ACL entries in a single atomic operation. - -

          1. Validates the user has 'alter' permission on the resource. 2. Removes entries only if - they exactly match the provided entries (principal, operation, permission). 3. Does not - remove entries if any of the ACL entries do not exist.

          -
          -
          Specified by:
          -
          dropAcls in interface Admin
          -
          Parameters:
          -
          filters - List of ACL entries to remove.
          -
          Returns:
          -
          A CompletableFuture indicating completion of the operation.
          -
          -
        • -
        - - - -
          -
        • -

          close

          -
          public void close()
          -
          -
          Specified by:
          -
          close in interface AutoCloseable
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/ListOffsetsResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/ListOffsetsResult.html deleted file mode 100644 index d1cd4527fd..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/ListOffsetsResult.html +++ /dev/null @@ -1,326 +0,0 @@ - - - - - -ListOffsetsResult (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class ListOffsetsResult

    -
    -
    - -
    -
      -
    • -
      -
      @PublicEvolving
      -public class ListOffsetsResult
      -extends Object
      -
      Result of list offsets request.
      -
      -
      Since:
      -
      0.2
      -
      -
    • -
    -
    -
    - -
    -
    - -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.EarliestSpec.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.EarliestSpec.html deleted file mode 100644 index 4de7b8e4f5..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.EarliestSpec.html +++ /dev/null @@ -1,312 +0,0 @@ - - - - - -OffsetSpec.EarliestSpec (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class OffsetSpec.EarliestSpec

    -
    -
    - -
    -
      -
    • -
      -
      Enclosing class:
      -
      OffsetSpec
      -
      -
      -
      public static class OffsetSpec.EarliestSpec
      -extends OffsetSpec
      -
      Earliest offset spec.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          EarliestSpec

          -
          public EarliestSpec()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.LatestSpec.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.LatestSpec.html deleted file mode 100644 index 2c28048181..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.LatestSpec.html +++ /dev/null @@ -1,312 +0,0 @@ - - - - - -OffsetSpec.LatestSpec (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class OffsetSpec.LatestSpec

    -
    -
    - -
    -
      -
    • -
      -
      Enclosing class:
      -
      OffsetSpec
      -
      -
      -
      public static class OffsetSpec.LatestSpec
      -extends OffsetSpec
      -
      latest offset spec.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LatestSpec

          -
          public LatestSpec()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.TimestampSpec.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.TimestampSpec.html deleted file mode 100644 index ce8f828fe7..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.TimestampSpec.html +++ /dev/null @@ -1,350 +0,0 @@ - - - - - -OffsetSpec.TimestampSpec (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class OffsetSpec.TimestampSpec

    -
    -
    - -
    -
      -
    • -
      -
      Enclosing class:
      -
      OffsetSpec
      -
      -
      -
      public static class OffsetSpec.TimestampSpec
      -extends OffsetSpec
      -
      timestamp offset spec.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          TimestampSpec

          -
          public TimestampSpec​(long timestamp)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getTimestamp

          -
          public long getTimestamp()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.html deleted file mode 100644 index 209c725dad..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/OffsetSpec.html +++ /dev/null @@ -1,399 +0,0 @@ - - - - - -OffsetSpec (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class OffsetSpec

    -
    -
    - -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - -
          -
        • -

          LIST_EARLIEST_OFFSET

          -
          public static final int LIST_EARLIEST_OFFSET
          -
          -
          See Also:
          -
          Constant Field Values
          -
          -
        • -
        - - - -
          -
        • -

          LIST_LATEST_OFFSET

          -
          public static final int LIST_LATEST_OFFSET
          -
          -
          See Also:
          -
          Constant Field Values
          -
          -
        • -
        - - - -
          -
        • -

          LIST_OFFSET_FROM_TIMESTAMP

          -
          public static final int LIST_OFFSET_FROM_TIMESTAMP
          -
          -
          See Also:
          -
          Constant Field Values
          -
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          OffsetSpec

          -
          public OffsetSpec()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/Admin.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/Admin.html deleted file mode 100644 index b28444501a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/Admin.html +++ /dev/null @@ -1,413 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.client.admin.Admin (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.client.admin.Admin

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/CreateAclsResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/CreateAclsResult.html deleted file mode 100644 index 86a49a89a5..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/CreateAclsResult.html +++ /dev/null @@ -1,199 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.admin.CreateAclsResult (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.admin.CreateAclsResult

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.FilterResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.FilterResult.html deleted file mode 100644 index fa3bfcc6d6..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.FilterResult.html +++ /dev/null @@ -1,194 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.admin.DropAclsResult.FilterResult (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.admin.DropAclsResult.FilterResult

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.FilterResults.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.FilterResults.html deleted file mode 100644 index 209d0e8a17..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.FilterResults.html +++ /dev/null @@ -1,195 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.admin.DropAclsResult.FilterResults (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.admin.DropAclsResult.FilterResults

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.html deleted file mode 100644 index 69f1a8dd38..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/DropAclsResult.html +++ /dev/null @@ -1,199 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.admin.DropAclsResult (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.admin.DropAclsResult

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/FlussAdmin.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/FlussAdmin.html deleted file mode 100644 index 9cf453a2cf..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/FlussAdmin.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.admin.FlussAdmin (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.admin.FlussAdmin

    -
    -
    No usage of org.apache.fluss.client.admin.FlussAdmin
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/ListOffsetsResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/ListOffsetsResult.html deleted file mode 100644 index a91a047c1d..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/ListOffsetsResult.html +++ /dev/null @@ -1,221 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.admin.ListOffsetsResult (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.admin.ListOffsetsResult

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.EarliestSpec.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.EarliestSpec.html deleted file mode 100644 index aaa11f8ee2..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.EarliestSpec.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.admin.OffsetSpec.EarliestSpec (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.admin.OffsetSpec.EarliestSpec

    -
    -
    No usage of org.apache.fluss.client.admin.OffsetSpec.EarliestSpec
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.LatestSpec.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.LatestSpec.html deleted file mode 100644 index dc1bb528dd..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.LatestSpec.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.admin.OffsetSpec.LatestSpec (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.admin.OffsetSpec.LatestSpec

    -
    -
    No usage of org.apache.fluss.client.admin.OffsetSpec.LatestSpec
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.TimestampSpec.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.TimestampSpec.html deleted file mode 100644 index 73c30490d0..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.TimestampSpec.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.admin.OffsetSpec.TimestampSpec (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.admin.OffsetSpec.TimestampSpec

    -
    -
    No usage of org.apache.fluss.client.admin.OffsetSpec.TimestampSpec
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.html deleted file mode 100644 index 158176b94c..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/class-use/OffsetSpec.html +++ /dev/null @@ -1,281 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.admin.OffsetSpec (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.admin.OffsetSpec

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-summary.html deleted file mode 100644 index 1e2f9fbf88..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-summary.html +++ /dev/null @@ -1,241 +0,0 @@ - - - - - -org.apache.fluss.client.admin (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.client.admin

    -
    -
    -
      -
    • - - - - - - - - - - - - -
      Interface Summary 
      InterfaceDescription
      Admin -
      The administrative client for Fluss, which supports managing and inspecting tables, servers, - configurations and ACLs.
      -
      -
    • -
    • - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
      Class Summary 
      ClassDescription
      CreateAclsResult -
      Represents the result of a batch ACL operation, managing asynchronous completion of individual - ACL operations.
      -
      DropAclsResult -
      Represents the result of a drop ACLs operation.
      -
      DropAclsResult.FilterResult -
      A class containing either the deleted ACL binding or an exception if the delete failed.
      -
      DropAclsResult.FilterResults -
      A class containing the results of the delete ACLs operation.
      -
      FlussAdmin -
      The default implementation of Admin.
      -
      ListOffsetsResult -
      Result of list offsets request.
      -
      OffsetSpec -
      Offset spec.
      -
      OffsetSpec.EarliestSpec -
      Earliest offset spec.
      -
      OffsetSpec.LatestSpec -
      latest offset spec.
      -
      OffsetSpec.TimestampSpec -
      timestamp offset spec.
      -
      -
    • -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-tree.html deleted file mode 100644 index 25cf96694a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-tree.html +++ /dev/null @@ -1,185 +0,0 @@ - - - - - -org.apache.fluss.client.admin Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.client.admin

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -

    Interface Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-use.html deleted file mode 100644 index 039ff4fe58..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/admin/package-use.html +++ /dev/null @@ -1,414 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.client.admin (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.client.admin

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/Connection.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/Connection.html deleted file mode 100644 index 9885f40391..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/Connection.html +++ /dev/null @@ -1,246 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.client.Connection (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.client.Connection

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/ConnectionFactory.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/ConnectionFactory.html deleted file mode 100644 index be7483606e..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/ConnectionFactory.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.ConnectionFactory (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.ConnectionFactory

    -
    -
    No usage of org.apache.fluss.client.ConnectionFactory
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/FlussConnection.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/FlussConnection.html deleted file mode 100644 index 103483b660..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/class-use/FlussConnection.html +++ /dev/null @@ -1,217 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.FlussConnection (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.FlussConnection

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/AbstractLookupBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/AbstractLookupBatch.html deleted file mode 100644 index 1b5c6204ad..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/AbstractLookupBatch.html +++ /dev/null @@ -1,412 +0,0 @@ - - - - - -AbstractLookupBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class AbstractLookupBatch<T>

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.lookup.AbstractLookupBatch<T>
      • -
      -
    • -
    -
    -
      -
    • -
      -
      @Internal
      -public abstract class AbstractLookupBatch<T>
      -extends Object
      -
      An abstract lookup batch.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      - -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          AbstractLookupBatch

          -
          public AbstractLookupBatch​(org.apache.fluss.metadata.TableBucket tableBucket)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          complete

          -
          public abstract void complete​(List<T> values)
          -
          Complete the lookup operations using given values .
          -
        • -
        - - - - - - - - - - - -
          -
        • -

          tableBucket

          -
          public org.apache.fluss.metadata.TableBucket tableBucket()
          -
        • -
        - - - -
          -
        • -

          completeExceptionally

          -
          public void completeExceptionally​(Exception exception)
          -
          Complete the get operations with given exception.
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/AbstractLookupQuery.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/AbstractLookupQuery.html deleted file mode 100644 index cdb368599e..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/AbstractLookupQuery.html +++ /dev/null @@ -1,356 +0,0 @@ - - - - - -AbstractLookupQuery (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class AbstractLookupQuery<T>

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.lookup.AbstractLookupQuery<T>
      • -
      -
    • -
    -
    -
      -
    • -
      -
      Direct Known Subclasses:
      -
      LookupQuery, PrefixLookupQuery
      -
      -
      -
      @Internal
      -public abstract class AbstractLookupQuery<T>
      -extends Object
      -
      Abstract Class to represent a lookup operation.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          AbstractLookupQuery

          -
          public AbstractLookupQuery​(org.apache.fluss.metadata.TableBucket tableBucket,
          -                           byte[] key)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          key

          -
          public byte[] key()
          -
        • -
        - - - -
          -
        • -

          tableBucket

          -
          public org.apache.fluss.metadata.TableBucket tableBucket()
          -
        • -
        - - - -
          -
        • -

          lookupType

          -
          public abstract LookupType lookupType()
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/Lookup.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/Lookup.html deleted file mode 100644 index d8a9c3bf51..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/Lookup.html +++ /dev/null @@ -1,354 +0,0 @@ - - - - - -Lookup (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface Lookup

    -
    -
    -
    -
      -
    • -
      -
      All Known Implementing Classes:
      -
      TableLookup
      -
      -
      -
      @PublicEvolving
      -public interface Lookup
      -
      Used to configure and create a Lookuper to lookup rows of a primary key table. The built - Lookuper can be a primary key lookuper that lookups by the primary key, or a prefix key lookup - that lookups by the prefix key of the primary key. - -

      Lookup objects are immutable and can be shared between threads. Refinement methods, - like lookupBy(java.util.List<java.lang.String>), create new Lookup instances. - -

      Example1: Create a Primary Key Lookuper. Given a table with primary key column [k STRING]. - -

      
      - Lookuper lookuper = table.newLookup().createLookuper();
      - CompletableFuture<LookupResult> resultFuture = lookuper.lookup(GenericRow.of("key1"));
      - resultFuture.get().getRows().forEach(row -> {
      -    System.out.println(row);
      - });
      - 
      - -

      Example2: Create a Prefix Key Lookuper. Given a table with primary key column [a INT, b - STRING, c BIGINT] and bucket key [a, b]. - -

      
      - Lookuper lookuper = table.newLookup().lookupBy("a", "b").createLookuper();
      - CompletableFuture<LookupResult> resultFuture = lookuper.lookup(GenericRow.of(1, "b1"));
      - resultFuture.get().getRows().forEach(row -> {
      -   System.out.println(row);
      - });
      - 
      -
      -
      Since:
      -
      0.6
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          lookupBy

          -
          Lookup lookupBy​(List<String> lookupColumnNames)
          -
          Returns a new Lookup instance with the given lookup columns. By default, the lookup columns - are the primary key columns of the table. The specified columns must be a prefix subset of - the physical primary key columns (i.e., the primary key columns exclude partition columns). - -

          Note: Currently, if you want to lookup with specified lookup columns (i.e., prefix key - lookup), the table you created must both define the primary key and the bucket key, in - addition, the lookup columns needs to be equals with bucket key, and to be a part of the - primary key and must be a prefix of the primary key. For example, if a table has fields - [a,b,c,d], and the primary key is set to [a, b, c], with the bucket key set to [a, b], then - the schema of the lookup columns would also be [a, b]. This pattern can create a Prefix Key - Lookuper lookup rows by prefix of the primary key. - -

          TODO: currently, the interface only support bucket key as the prefix key to lookup. - Generalize the prefix key lookup to support any prefix keys. - -

          We also support prefix key lookup for partitioned tables. The schema of the lookup columns - should contain partition fields and bucket key. In addition, the schema of the lookup columns - excluded partition fields should be a prefix of primary key excluded partition fields.

          -
          -
          Parameters:
          -
          lookupColumnNames - the specified columns to lookup by
          -
          -
        • -
        - - - -
          -
        • -

          lookupBy

          -
          default Lookup lookupBy​(String... lookupColumnNames)
          -
          -
          Parameters:
          -
          lookupColumnNames - the specified columns to lookup by
          -
          See Also:
          -
          for more details.
          -
          -
        • -
        - - - -
          -
        • -

          createLookuper

          -
          Lookuper createLookuper()
          -
          Creates a Lookuper instance to lookup rows of a primary key table by the specified - lookup columns. By default, the lookup columns are the primary key columns, but can be - changed with (lookupBy(List)) method.
          -
          -
          Returns:
          -
          the lookuper
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupBatch.html deleted file mode 100644 index 06b041d1ee..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupBatch.html +++ /dev/null @@ -1,370 +0,0 @@ - - - - - -LookupBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LookupBatch

    -
    -
    - -
    -
      -
    • -
      -
      @Internal
      -public class LookupBatch
      -extends Object
      -
      A batch that contains the lookup operations that send to same tablet bucket together.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LookupBatch

          -
          public LookupBatch​(org.apache.fluss.metadata.TableBucket tableBucket)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          addLookup

          -
          public void addLookup​(LookupQuery lookup)
          -
        • -
        - - - - - - - -
          -
        • -

          tableBucket

          -
          public org.apache.fluss.metadata.TableBucket tableBucket()
          -
        • -
        - - - -
          -
        • -

          complete

          -
          public void complete​(List<byte[]> values)
          -
          Complete the lookup operations using given values .
          -
        • -
        - - - -
          -
        • -

          completeExceptionally

          -
          public void completeExceptionally​(Exception exception)
          -
          Complete the lookup operations with given exception.
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupClient.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupClient.html deleted file mode 100644 index ff1eef4b16..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupClient.html +++ /dev/null @@ -1,397 +0,0 @@ - - - - - -LookupClient (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LookupClient

    -
    -
    - -
    -
      -
    • -
      -
      @ThreadSafe
      -@Internal
      -public class LookupClient
      -extends Object
      -
      A client that lookups value of keys from server. - -

      The lookup client contains of a queue of pending lookup operations and background I/O threads - that is responsible for turning these lookup operations into network requests and transmitting - them to the cluster. - -

      The lookup(TableBucket, byte[]) method is asynchronous, when called, it adds the - lookup operation to a queue of pending lookup operations and immediately returns. This allows the - lookup operations to batch together individual lookup operations for efficiency.

      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      - -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LookupClient

          -
          public LookupClient​(org.apache.fluss.config.Configuration conf,
          -                    MetadataUpdater metadataUpdater)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          lookup

          -
          public CompletableFuture<byte[]> lookup​(org.apache.fluss.metadata.TableBucket tableBucket,
          -                                        byte[] keyBytes)
          -
        • -
        - - - -
          -
        • -

          prefixLookup

          -
          public CompletableFuture<List<byte[]>> prefixLookup​(org.apache.fluss.metadata.TableBucket tableBucket,
          -                                                    byte[] keyBytes)
          -
        • -
        - - - -
          -
        • -

          close

          -
          public void close​(Duration timeout)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupQuery.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupQuery.html deleted file mode 100644 index 4fc6064837..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupQuery.html +++ /dev/null @@ -1,303 +0,0 @@ - - - - - -LookupQuery (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LookupQuery

    -
    -
    - -
    -
      -
    • -
      -
      @Internal
      -public class LookupQuery
      -extends AbstractLookupQuery<byte[]>
      -
      Class to represent a Lookup operation, it contains the table bucket that the key should lookup - from, the bytes of the key, and a future for the lookup operation.
      -
    • -
    -
    -
    - -
    -
    - -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupResult.html deleted file mode 100644 index 489cc5438f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupResult.html +++ /dev/null @@ -1,395 +0,0 @@ - - - - - -LookupResult (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LookupResult

    -
    -
    - -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LookupResult

          -
          public LookupResult​(@Nullable
          -                    org.apache.fluss.row.InternalRow row)
          -
        • -
        - - - -
          -
        • -

          LookupResult

          -
          public LookupResult​(List<org.apache.fluss.row.InternalRow> rowList)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getRowList

          -
          public List<org.apache.fluss.row.InternalRow> getRowList()
          -
        • -
        - - - -
          -
        • -

          getSingletonRow

          -
          @Nullable
          -public org.apache.fluss.row.InternalRow getSingletonRow()
          -
        • -
        - - - -
          -
        • -

          equals

          -
          public boolean equals​(Object o)
          -
          -
          Overrides:
          -
          equals in class Object
          -
          -
        • -
        - - - -
          -
        • -

          hashCode

          -
          public int hashCode()
          -
          -
          Overrides:
          -
          hashCode in class Object
          -
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupType.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupType.html deleted file mode 100644 index 618eee9430..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/LookupType.html +++ /dev/null @@ -1,380 +0,0 @@ - - - - - -LookupType (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Enum LookupType

    -
    -
    - -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Enum Constant Detail

        - - - -
          -
        • -

          LOOKUP

          -
          public static final LookupType LOOKUP
          -
        • -
        - - - -
          -
        • -

          PREFIX_LOOKUP

          -
          public static final LookupType PREFIX_LOOKUP
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          values

          -
          public static LookupType[] values()
          -
          Returns an array containing the constants of this enum type, in -the order they are declared. This method may be used to iterate -over the constants as follows: -
          -for (LookupType c : LookupType.values())
          -    System.out.println(c);
          -
          -
          -
          Returns:
          -
          an array containing the constants of this enum type, in the order they are declared
          -
          -
        • -
        - - - -
          -
        • -

          valueOf

          -
          public static LookupType valueOf​(String name)
          -
          Returns the enum constant of this type with the specified name. -The string must match exactly an identifier used to declare an -enum constant in this type. (Extraneous whitespace characters are -not permitted.)
          -
          -
          Parameters:
          -
          name - the name of the enum constant to be returned.
          -
          Returns:
          -
          the enum constant with the specified name
          -
          Throws:
          -
          IllegalArgumentException - if this enum type has no constant with the specified name
          -
          NullPointerException - if the argument is null
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/Lookuper.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/Lookuper.html deleted file mode 100644 index e1055c8b17..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/Lookuper.html +++ /dev/null @@ -1,269 +0,0 @@ - - - - - -Lookuper (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface Lookuper

    -
    -
    -
    -
      -
    • -
      -
      @PublicEvolving
      -public interface Lookuper
      -
      The lookup-er is used to lookup row of a primary key table by primary key or prefix key.
      -
      -
      Since:
      -
      0.6
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          lookup

          -
          CompletableFuture<LookupResult> lookup​(org.apache.fluss.row.InternalRow lookupKey)
          -
          Lookups certain row from the given lookup key. - -

          The lookup key must be a primary key if the lookuper is a Primary Key Lookuper (created by - table.newLookup().createLookuper()), or be the prefix key if the lookuper is a Prefix - Key Lookuper (created by table.newLookup().lookupBy(prefixKeys).createLookuper()).

          -
          -
          Parameters:
          -
          lookupKey - the lookup key.
          -
          Returns:
          -
          the result of lookup.
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookup.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookup.html deleted file mode 100644 index 1446848f55..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookup.html +++ /dev/null @@ -1,312 +0,0 @@ - - - - - -PrefixLookup (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class PrefixLookup

    -
    -
    - -
    -
      -
    • -
      -
      @PublicEvolving
      -public class PrefixLookup
      -extends Object
      -
      Used to describe the operation to prefix lookup by PrefixLookuper to a primary key table.
      -
      -
      Since:
      -
      0.6
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          PrefixLookup

          -
          public PrefixLookup​(List<String> lookupColumnNames)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getLookupColumnNames

          -
          public List<String> getLookupColumnNames()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookupBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookupBatch.html deleted file mode 100644 index 6bcda800e8..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookupBatch.html +++ /dev/null @@ -1,368 +0,0 @@ - - - - - -PrefixLookupBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class PrefixLookupBatch

    -
    -
    - -
    -
      -
    • -
      -
      @Internal
      -public class PrefixLookupBatch
      -extends Object
      -
      A batch that contains the prefix lookup operations that send to same destination and some table - together.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          PrefixLookupBatch

          -
          public PrefixLookupBatch​(org.apache.fluss.metadata.TableBucket tableBucket)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - - - - - - - - - -
          -
        • -

          tableBucket

          -
          public org.apache.fluss.metadata.TableBucket tableBucket()
          -
        • -
        - - - -
          -
        • -

          complete

          -
          public void complete​(List<List<byte[]>> values)
          -
        • -
        - - - -
          -
        • -

          completeExceptionally

          -
          public void completeExceptionally​(Exception exception)
          -
          Complete the get operations with given exception.
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookupQuery.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookupQuery.html deleted file mode 100644 index 67db91fbca..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/PrefixLookupQuery.html +++ /dev/null @@ -1,303 +0,0 @@ - - - - - -PrefixLookupQuery (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class PrefixLookupQuery

    -
    -
    - -
    -
      -
    • -
      -
      @Internal
      -public class PrefixLookupQuery
      -extends AbstractLookupQuery<List<byte[]>>
      -
      Class to represent a prefix lookup operation, it contains the table id, bucketNums and related - CompletableFuture.
      -
    • -
    -
    -
    - -
    -
    - -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/TableLookup.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/TableLookup.html deleted file mode 100644 index 8b0c7a3a23..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/TableLookup.html +++ /dev/null @@ -1,377 +0,0 @@ - - - - - -TableLookup (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TableLookup

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      Lookup
      -
      -
      -
      public class TableLookup
      -extends Object
      -implements Lookup
      -
      API for configuring and creating Lookuper.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          TableLookup

          -
          public TableLookup​(org.apache.fluss.metadata.TableInfo tableInfo,
          -                   MetadataUpdater metadataUpdater,
          -                   LookupClient lookupClient)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          lookupBy

          -
          public Lookup lookupBy​(List<String> lookupColumnNames)
          -
          Description copied from interface: Lookup
          -
          Returns a new Lookup instance with the given lookup columns. By default, the lookup columns - are the primary key columns of the table. The specified columns must be a prefix subset of - the physical primary key columns (i.e., the primary key columns exclude partition columns). - -

          Note: Currently, if you want to lookup with specified lookup columns (i.e., prefix key - lookup), the table you created must both define the primary key and the bucket key, in - addition, the lookup columns needs to be equals with bucket key, and to be a part of the - primary key and must be a prefix of the primary key. For example, if a table has fields - [a,b,c,d], and the primary key is set to [a, b, c], with the bucket key set to [a, b], then - the schema of the lookup columns would also be [a, b]. This pattern can create a Prefix Key - Lookuper lookup rows by prefix of the primary key. - -

          TODO: currently, the interface only support bucket key as the prefix key to lookup. - Generalize the prefix key lookup to support any prefix keys. - -

          We also support prefix key lookup for partitioned tables. The schema of the lookup columns - should contain partition fields and bucket key. In addition, the schema of the lookup columns - excluded partition fields should be a prefix of primary key excluded partition fields.

          -
          -
          Specified by:
          -
          lookupBy in interface Lookup
          -
          Parameters:
          -
          lookupColumnNames - the specified columns to lookup by
          -
          -
        • -
        - - - -
          -
        • -

          createLookuper

          -
          public Lookuper createLookuper()
          -
          Description copied from interface: Lookup
          -
          Creates a Lookuper instance to lookup rows of a primary key table by the specified - lookup columns. By default, the lookup columns are the primary key columns, but can be - changed with (Lookup.lookupBy(List)) method.
          -
          -
          Specified by:
          -
          createLookuper in interface Lookup
          -
          Returns:
          -
          the lookuper
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/AbstractLookupBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/AbstractLookupBatch.html deleted file mode 100644 index 30dbc32a46..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/AbstractLookupBatch.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.lookup.AbstractLookupBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.lookup.AbstractLookupBatch

    -
    -
    No usage of org.apache.fluss.client.lookup.AbstractLookupBatch
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/AbstractLookupQuery.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/AbstractLookupQuery.html deleted file mode 100644 index 400c0c9cd2..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/AbstractLookupQuery.html +++ /dev/null @@ -1,248 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.lookup.AbstractLookupQuery (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.lookup.AbstractLookupQuery

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/Lookup.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/Lookup.html deleted file mode 100644 index dfb691b6a3..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/Lookup.html +++ /dev/null @@ -1,255 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.client.lookup.Lookup (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.client.lookup.Lookup

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupBatch.html deleted file mode 100644 index 948d2ce7a1..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupBatch.html +++ /dev/null @@ -1,193 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.lookup.LookupBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.lookup.LookupBatch

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupClient.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupClient.html deleted file mode 100644 index 4b508f55ba..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupClient.html +++ /dev/null @@ -1,218 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.lookup.LookupClient (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.lookup.LookupClient

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupQuery.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupQuery.html deleted file mode 100644 index 20b01d6de4..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupQuery.html +++ /dev/null @@ -1,207 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.lookup.LookupQuery (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.lookup.LookupQuery

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupResult.html deleted file mode 100644 index dd71ef8ac5..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupResult.html +++ /dev/null @@ -1,194 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.lookup.LookupResult (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.lookup.LookupResult

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupType.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupType.html deleted file mode 100644 index 75301e71b9..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/LookupType.html +++ /dev/null @@ -1,246 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.lookup.LookupType (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.lookup.LookupType

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/Lookuper.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/Lookuper.html deleted file mode 100644 index a70953e7f2..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/Lookuper.html +++ /dev/null @@ -1,200 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.client.lookup.Lookuper (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.client.lookup.Lookuper

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookup.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookup.html deleted file mode 100644 index e57351cf06..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookup.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.lookup.PrefixLookup (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.lookup.PrefixLookup

    -
    -
    No usage of org.apache.fluss.client.lookup.PrefixLookup
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookupBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookupBatch.html deleted file mode 100644 index dee71159a8..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookupBatch.html +++ /dev/null @@ -1,193 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.lookup.PrefixLookupBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.lookup.PrefixLookupBatch

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookupQuery.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookupQuery.html deleted file mode 100644 index 821ae6a97b..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/PrefixLookupQuery.html +++ /dev/null @@ -1,207 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.lookup.PrefixLookupQuery (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.lookup.PrefixLookupQuery

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/TableLookup.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/TableLookup.html deleted file mode 100644 index 64aa9eb0ec..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/class-use/TableLookup.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.lookup.TableLookup (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.lookup.TableLookup

    -
    -
    No usage of org.apache.fluss.client.lookup.TableLookup
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-summary.html deleted file mode 100644 index 3f3e4d96dd..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-summary.html +++ /dev/null @@ -1,265 +0,0 @@ - - - - - -org.apache.fluss.client.lookup (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.client.lookup

    -
    -
    -
      -
    • - - - - - - - - - - - - - - - - -
      Interface Summary 
      InterfaceDescription
      Lookup -
      Used to configure and create a Lookuper to lookup rows of a primary key table.
      -
      Lookuper -
      The lookup-er is used to lookup row of a primary key table by primary key or prefix key.
      -
      -
    • -
    • - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
      Class Summary 
      ClassDescription
      AbstractLookupBatch<T> -
      An abstract lookup batch.
      -
      AbstractLookupQuery<T> -
      Abstract Class to represent a lookup operation.
      -
      LookupBatch -
      A batch that contains the lookup operations that send to same tablet bucket together.
      -
      LookupClient -
      A client that lookups value of keys from server.
      -
      LookupQuery -
      Class to represent a Lookup operation, it contains the table bucket that the key should lookup - from, the bytes of the key, and a future for the lookup operation.
      -
      LookupResult - -
      PrefixLookup -
      Used to describe the operation to prefix lookup by PrefixLookuper to a primary key table.
      -
      PrefixLookupBatch -
      A batch that contains the prefix lookup operations that send to same destination and some table - together.
      -
      PrefixLookupQuery -
      Class to represent a prefix lookup operation, it contains the table id, bucketNums and related - CompletableFuture.
      -
      TableLookup -
      API for configuring and creating Lookuper.
      -
      -
    • -
    • - - - - - - - - - - - - -
      Enum Summary 
      EnumDescription
      LookupType -
      Enum to represent the type of lookup operation.
      -
      -
    • -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-tree.html deleted file mode 100644 index 6ebb77a7af..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-tree.html +++ /dev/null @@ -1,196 +0,0 @@ - - - - - -org.apache.fluss.client.lookup Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.client.lookup

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -

    Interface Hierarchy

    -
      -
    • org.apache.fluss.client.lookup.Lookup
    • -
    • org.apache.fluss.client.lookup.Lookuper
    • -
    -
    -
    -

    Enum Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-use.html deleted file mode 100644 index e8b7161486..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/lookup/package-use.html +++ /dev/null @@ -1,328 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.client.lookup (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.client.lookup

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/KvSnapshotMetadata.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/KvSnapshotMetadata.html deleted file mode 100644 index efc9ff38d7..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/KvSnapshotMetadata.html +++ /dev/null @@ -1,354 +0,0 @@ - - - - - -KvSnapshotMetadata (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class KvSnapshotMetadata

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.metadata.KvSnapshotMetadata
      • -
      -
    • -
    -
    -
      -
    • -
      -
      @PublicEvolving
      -public class KvSnapshotMetadata
      -extends Object
      -
      A class to represent the snapshot info of a kv tablet (bucket). It contains: - -
        -
      • The snapshot files of the bucket. -
      • The log offset corresponding to the snapshot. -
      - -

      To read the full data of a bucket, it requires to read the snapshot and the log from the - corresponding log offset.

      -
      -
      Since:
      -
      0.6
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          KvSnapshotMetadata

          -
          public KvSnapshotMetadata​(List<org.apache.fluss.fs.FsPathAndFileName> snapshotFiles,
          -                          long logOffset)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getSnapshotFiles

          -
          public List<org.apache.fluss.fs.FsPathAndFileName> getSnapshotFiles()
          -
        • -
        - - - -
          -
        • -

          getLogOffset

          -
          public long getLogOffset()
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/KvSnapshots.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/KvSnapshots.html deleted file mode 100644 index 96a8da7bcc..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/KvSnapshots.html +++ /dev/null @@ -1,385 +0,0 @@ - - - - - -KvSnapshots (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class KvSnapshots

    -
    -
    - -
    -
      -
    • -
      -
      @PublicEvolving
      -public class KvSnapshots
      -extends Object
      -
      A class representing the kv snapshots of a table or a partition. It contains multiple snapshots - for each kv tablet (bucket).
      -
      -
      Since:
      -
      0.6
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          KvSnapshots

          -
          public KvSnapshots​(long tableId,
          -                   @Nullable
          -                   Long partitionId,
          -                   Map<Integer,​Long> snapshotIds,
          -                   Map<Integer,​Long> logOffsets)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getTableId

          -
          public long getTableId()
          -
        • -
        - - - -
          -
        • -

          getPartitionId

          -
          @Nullable
          -public Long getPartitionId()
          -
        • -
        - - - -
          -
        • -

          getBucketIds

          -
          public Set<Integer> getBucketIds()
          -
        • -
        - - - -
          -
        • -

          getSnapshotId

          -
          public OptionalLong getSnapshotId​(int bucketId)
          -
          Get the latest snapshot id for this kv tablet (bucket), or empty if there are no snapshots.
          -
        • -
        - - - -
          -
        • -

          getLogOffset

          -
          public OptionalLong getLogOffset​(int bucketId)
          -
          Get the log offset to read after the snapshot for this kv tablet (bucket), or empty if there - are no snapshots.
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/LakeSnapshot.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/LakeSnapshot.html deleted file mode 100644 index 7779ded118..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/LakeSnapshot.html +++ /dev/null @@ -1,363 +0,0 @@ - - - - - -LakeSnapshot (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LakeSnapshot

    -
    -
    - -
    -
      -
    • -
      -
      @PublicEvolving
      -public class LakeSnapshot
      -extends Object
      -
      A class representing the lake snapshot information of a table. It contains: -
    • The snapshot id and the log offset for each bucket.
    -
    -
    Since:
    -
    0.3
    -
    - - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LakeSnapshot

          -
          public LakeSnapshot​(long snapshotId,
          -                    Map<org.apache.fluss.metadata.TableBucket,​Long> tableBucketsOffset,
          -                    Map<Long,​String> partitionNameById)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getSnapshotId

          -
          public long getSnapshotId()
          -
        • -
        - - - -
          -
        • -

          getTableBucketsOffset

          -
          public Map<org.apache.fluss.metadata.TableBucket,​Long> getTableBucketsOffset()
          -
        • -
        - - - -
          -
        • -

          getPartitionNameById

          -
          public Map<Long,​String> getPartitionNameById()
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    - -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/MetadataUpdater.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/MetadataUpdater.html deleted file mode 100644 index d62e985e38..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/MetadataUpdater.html +++ /dev/null @@ -1,721 +0,0 @@ - - - - - -MetadataUpdater (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class MetadataUpdater

    -
    -
    - -
    -
      -
    • -
      -
      public class MetadataUpdater
      -extends Object
      -
      The updater to initialize and update client metadata.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - -
          -
        • -

          cluster

          -
          protected volatile org.apache.fluss.cluster.Cluster cluster
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          MetadataUpdater

          -
          public MetadataUpdater​(org.apache.fluss.config.Configuration configuration,
          -                       org.apache.fluss.rpc.RpcClient rpcClient)
          -
        • -
        - - - -
          -
        • -

          MetadataUpdater

          -
          @VisibleForTesting
          -public MetadataUpdater​(org.apache.fluss.rpc.RpcClient rpcClient,
          -                       org.apache.fluss.cluster.Cluster cluster)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getCluster

          -
          public org.apache.fluss.cluster.Cluster getCluster()
          -
        • -
        - - - -
          -
        • -

          getCoordinatorServer

          -
          @Nullable
          -public org.apache.fluss.cluster.ServerNode getCoordinatorServer()
          -
        • -
        - - - -
          -
        • -

          getTableId

          -
          public long getTableId​(org.apache.fluss.metadata.TablePath tablePath)
          -
        • -
        - - - -
          -
        • -

          getPartitionId

          -
          public Optional<Long> getPartitionId​(org.apache.fluss.metadata.PhysicalTablePath physicalTablePath)
          -
        • -
        - - - -
          -
        • -

          getPartitionIdOrElseThrow

          -
          public Long getPartitionIdOrElseThrow​(org.apache.fluss.metadata.PhysicalTablePath physicalTablePath)
          -
        • -
        - - - -
          -
        • -

          getTableInfoOrElseThrow

          -
          public org.apache.fluss.metadata.TableInfo getTableInfoOrElseThrow​(org.apache.fluss.metadata.TablePath tablePath)
          -
        • -
        - - - -
          -
        • -

          getBucketLocation

          -
          public Optional<org.apache.fluss.cluster.BucketLocation> getBucketLocation​(org.apache.fluss.metadata.TableBucket tableBucket)
          -
        • -
        - - - -
          -
        • -

          getTableInfoOrElseThrow

          -
          public org.apache.fluss.metadata.TableInfo getTableInfoOrElseThrow​(long tableId)
          -
        • -
        - - - -
          -
        • -

          leaderFor

          -
          public int leaderFor​(org.apache.fluss.metadata.TableBucket tableBucket)
          -
        • -
        - - - -
          -
        • -

          getRandomTabletServer

          -
          @Nullable
          -public org.apache.fluss.cluster.ServerNode getRandomTabletServer()
          -
        • -
        - - - -
          -
        • -

          newCoordinatorServerClient

          -
          public org.apache.fluss.rpc.gateway.CoordinatorGateway newCoordinatorServerClient()
          -
        • -
        - - - -
          -
        • -

          newRandomTabletServerClient

          -
          public org.apache.fluss.rpc.gateway.TabletServerGateway newRandomTabletServerClient()
          -
        • -
        - - - -
          -
        • -

          newTabletServerClientForNode

          -
          @Nullable
          -public org.apache.fluss.rpc.gateway.TabletServerGateway newTabletServerClientForNode​(int serverId)
          -
        • -
        - - - -
          -
        • -

          checkAndUpdateTableMetadata

          -
          public void checkAndUpdateTableMetadata​(Set<org.apache.fluss.metadata.TablePath> tablePaths)
          -
        • -
        - - - -
          -
        • -

          checkAndUpdatePartitionMetadata

          -
          public boolean checkAndUpdatePartitionMetadata​(org.apache.fluss.metadata.PhysicalTablePath physicalTablePath)
          -                                        throws org.apache.fluss.exception.PartitionNotExistException
          -
          Check the partition exists in metadata cache, if not, try to update the metadata cache, if - not exist yet, throw exception. - -

          and update partition metadata .

          -
          -
          Throws:
          -
          org.apache.fluss.exception.PartitionNotExistException
          -
          -
        • -
        - - - -
          -
        • -

          checkAndUpdateMetadata

          -
          public void checkAndUpdateMetadata​(org.apache.fluss.metadata.TablePath tablePath,
          -                                   org.apache.fluss.metadata.TableBucket tableBucket)
          -
          Check the table/partition info for the given table bucket exist in metadata cache, if not, - try to update the metadata cache.
          -
        • -
        - - - -
          -
        • -

          checkAndUpdatePartitionMetadata

          -
          public void checkAndUpdatePartitionMetadata​(org.apache.fluss.metadata.TablePath tablePath,
          -                                            Collection<Long> partitionIds)
          -
          Check the partitions info for the given partition ids exist in metadata cache, if not, try to - update the metadata cache. - -

          Note: it'll assume the partition ids belong to the given tablePath

          -
        • -
        - - - -
          -
        • -

          updateTableOrPartitionMetadata

          -
          public void updateTableOrPartitionMetadata​(org.apache.fluss.metadata.TablePath tablePath,
          -                                           @Nullable
          -                                           Long partitionId)
          -
        • -
        - - - -
          -
        • -

          updatePhysicalTableMetadata

          -
          public void updatePhysicalTableMetadata​(Set<org.apache.fluss.metadata.PhysicalTablePath> physicalTablePaths)
          -
          Update the table or partition metadata info.
          -
        • -
        - - - -
          -
        • -

          updateMetadata

          -
          @VisibleForTesting
          -protected void updateMetadata​(@Nullable
          -                              Set<org.apache.fluss.metadata.TablePath> tablePaths,
          -                              @Nullable
          -                              Collection<org.apache.fluss.metadata.PhysicalTablePath> tablePartitionNames,
          -                              @Nullable
          -                              Collection<Long> tablePartitionIds)
          -                       throws org.apache.fluss.exception.PartitionNotExistException
          -
          -
          Throws:
          -
          org.apache.fluss.exception.PartitionNotExistException
          -
          -
        • -
        - - - -
          -
        • -

          invalidPhysicalTableBucketMeta

          -
          public void invalidPhysicalTableBucketMeta​(Set<org.apache.fluss.metadata.PhysicalTablePath> physicalTablesToInvalid)
          -
          Invalid the bucket metadata for the given physical table paths.
          -
        • -
        - - - -
          -
        • -

          getPhysicalTablePathByIds

          -
          public Set<org.apache.fluss.metadata.PhysicalTablePath> getPhysicalTablePathByIds​(@Nullable
          -                                                                                  Collection<Long> tableId,
          -                                                                                  @Nullable
          -                                                                                  Collection<org.apache.fluss.metadata.TablePartition> tablePartitions)
          -
          Get the table physical paths by table ids and partition ids.
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/KvSnapshotMetadata.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/KvSnapshotMetadata.html deleted file mode 100644 index 3371e6022f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/KvSnapshotMetadata.html +++ /dev/null @@ -1,227 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.metadata.KvSnapshotMetadata (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.metadata.KvSnapshotMetadata

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/KvSnapshots.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/KvSnapshots.html deleted file mode 100644 index 79d0606216..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/KvSnapshots.html +++ /dev/null @@ -1,239 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.metadata.KvSnapshots (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.metadata.KvSnapshots

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/LakeSnapshot.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/LakeSnapshot.html deleted file mode 100644 index 7cbf96d092..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/LakeSnapshot.html +++ /dev/null @@ -1,225 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.metadata.LakeSnapshot (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.metadata.LakeSnapshot

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/MetadataUpdater.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/MetadataUpdater.html deleted file mode 100644 index 521c06b7dd..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/class-use/MetadataUpdater.html +++ /dev/null @@ -1,398 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.metadata.MetadataUpdater (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.metadata.MetadataUpdater

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-summary.html deleted file mode 100644 index 2912039719..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-summary.html +++ /dev/null @@ -1,186 +0,0 @@ - - - - - -org.apache.fluss.client.metadata (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.client.metadata

    -
    -
    -
      -
    • - - - - - - - - - - - - - - - - - - - - - - - - -
      Class Summary 
      ClassDescription
      KvSnapshotMetadata -
      A class to represent the snapshot info of a kv tablet (bucket).
      -
      KvSnapshots -
      A class representing the kv snapshots of a table or a partition.
      -
      LakeSnapshot -
      A class representing the lake snapshot information of a table.
      -
      MetadataUpdater -
      The updater to initialize and update client metadata.
      -
      -
    • -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-tree.html deleted file mode 100644 index e3f50d5f39..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-tree.html +++ /dev/null @@ -1,166 +0,0 @@ - - - - - -org.apache.fluss.client.metadata Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.client.metadata

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-use.html deleted file mode 100644 index 691feed838..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metadata/package-use.html +++ /dev/null @@ -1,359 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.client.metadata (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.client.metadata

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/ScannerMetricGroup.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/ScannerMetricGroup.html deleted file mode 100644 index 56df1dee93..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/ScannerMetricGroup.html +++ /dev/null @@ -1,494 +0,0 @@ - - - - - -ScannerMetricGroup (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class ScannerMetricGroup

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.metrics.groups.AbstractMetricGroup
      • -
      • -
          -
        • org.apache.fluss.client.metrics.ScannerMetricGroup
        • -
        -
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.fluss.metrics.groups.MetricGroup
      -
      -
      -
      @Internal
      -public class ScannerMetricGroup
      -extends org.apache.fluss.metrics.groups.AbstractMetricGroup
      -
      The metric group for scanner, including LogScanner and BatchScanner.
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Nested Class Summary

        -
          -
        • - - -

          Nested classes/interfaces inherited from class org.apache.fluss.metrics.groups.AbstractMetricGroup

          -org.apache.fluss.metrics.groups.AbstractMetricGroup.ChildType
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Field Summary

        -
          -
        • - - -

          Fields inherited from class org.apache.fluss.metrics.groups.AbstractMetricGroup

          -LOG, parent, registry, variables
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Summary

        - - - - - - - - - - -
        Constructors 
        ConstructorDescription
        ScannerMetricGroup​(org.apache.fluss.rpc.metrics.ClientMetricGroup parent, - org.apache.fluss.metadata.TablePath tablePath) 
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          ScannerMetricGroup

          -
          public ScannerMetricGroup​(org.apache.fluss.rpc.metrics.ClientMetricGroup parent,
          -                          org.apache.fluss.metadata.TablePath tablePath)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          fetchRequestCount

          -
          public org.apache.fluss.metrics.Counter fetchRequestCount()
          -
        • -
        - - - -
          -
        • -

          bytesPerRequest

          -
          public org.apache.fluss.metrics.Histogram bytesPerRequest()
          -
        • -
        - - - -
          -
        • -

          remoteFetchBytes

          -
          public org.apache.fluss.metrics.Counter remoteFetchBytes()
          -
        • -
        - - - -
          -
        • -

          remoteFetchRequestCount

          -
          public org.apache.fluss.metrics.Counter remoteFetchRequestCount()
          -
        • -
        - - - -
          -
        • -

          remoteFetchErrorCount

          -
          public org.apache.fluss.metrics.Counter remoteFetchErrorCount()
          -
        • -
        - - - -
          -
        • -

          recordPollStart

          -
          public void recordPollStart​(long pollStartMs)
          -
        • -
        - - - -
          -
        • -

          recordPollEnd

          -
          public void recordPollEnd​(long pollEndMs)
          -
        • -
        - - - -
          -
        • -

          updateFetchLatency

          -
          public void updateFetchLatency​(long latencyInMs)
          -
        • -
        - - - -
          -
        • -

          getGroupName

          -
          protected String getGroupName​(org.apache.fluss.metrics.CharacterFilter filter)
          -
          -
          Specified by:
          -
          getGroupName in class org.apache.fluss.metrics.groups.AbstractMetricGroup
          -
          -
        • -
        - - - -
          -
        • -

          putVariables

          -
          protected final void putVariables​(Map<String,​String> variables)
          -
          -
          Overrides:
          -
          putVariables in class org.apache.fluss.metrics.groups.AbstractMetricGroup
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/WriterMetricGroup.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/WriterMetricGroup.html deleted file mode 100644 index c852040807..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/WriterMetricGroup.html +++ /dev/null @@ -1,460 +0,0 @@ - - - - - -WriterMetricGroup (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class WriterMetricGroup

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.metrics.groups.AbstractMetricGroup
      • -
      • -
          -
        • org.apache.fluss.client.metrics.WriterMetricGroup
        • -
        -
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.fluss.metrics.groups.MetricGroup
      -
      -
      -
      @Internal
      -public class WriterMetricGroup
      -extends org.apache.fluss.metrics.groups.AbstractMetricGroup
      -
      Metrics for WriterClient.
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Nested Class Summary

        -
          -
        • - - -

          Nested classes/interfaces inherited from class org.apache.fluss.metrics.groups.AbstractMetricGroup

          -org.apache.fluss.metrics.groups.AbstractMetricGroup.ChildType
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Field Summary

        -
          -
        • - - -

          Fields inherited from class org.apache.fluss.metrics.groups.AbstractMetricGroup

          -LOG, parent, registry, variables
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Summary

        - - - - - - - - - - -
        Constructors 
        ConstructorDescription
        WriterMetricGroup​(org.apache.fluss.rpc.metrics.ClientMetricGroup parent) 
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          WriterMetricGroup

          -
          public WriterMetricGroup​(org.apache.fluss.rpc.metrics.ClientMetricGroup parent)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          setBatchQueueTimeMs

          -
          public void setBatchQueueTimeMs​(long batchQueueTimeMs)
          -
        • -
        - - - -
          -
        • -

          setSendLatencyInMs

          -
          public void setSendLatencyInMs​(long sendLatencyInMs)
          -
        • -
        - - - -
          -
        • -

          recordsRetryTotal

          -
          public org.apache.fluss.metrics.Counter recordsRetryTotal()
          -
        • -
        - - - -
          -
        • -

          recordsSendTotal

          -
          public org.apache.fluss.metrics.Counter recordsSendTotal()
          -
        • -
        - - - -
          -
        • -

          bytesPerBatch

          -
          public org.apache.fluss.metrics.Histogram bytesPerBatch()
          -
        • -
        - - - -
          -
        • -

          bytesSendTotal

          -
          public org.apache.fluss.metrics.Counter bytesSendTotal()
          -
        • -
        - - - -
          -
        • -

          recordPerBatch

          -
          public org.apache.fluss.metrics.Histogram recordPerBatch()
          -
        • -
        - - - -
          -
        • -

          getGroupName

          -
          protected String getGroupName​(org.apache.fluss.metrics.CharacterFilter filter)
          -
          -
          Specified by:
          -
          getGroupName in class org.apache.fluss.metrics.groups.AbstractMetricGroup
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/class-use/ScannerMetricGroup.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/class-use/ScannerMetricGroup.html deleted file mode 100644 index b55b238b64..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/class-use/ScannerMetricGroup.html +++ /dev/null @@ -1,203 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.metrics.ScannerMetricGroup (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.metrics.ScannerMetricGroup

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/class-use/WriterMetricGroup.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/class-use/WriterMetricGroup.html deleted file mode 100644 index d393a72f6a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/class-use/WriterMetricGroup.html +++ /dev/null @@ -1,197 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.metrics.WriterMetricGroup (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.metrics.WriterMetricGroup

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-summary.html deleted file mode 100644 index bf07f802d3..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-summary.html +++ /dev/null @@ -1,174 +0,0 @@ - - - - - -org.apache.fluss.client.metrics (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.client.metrics

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-tree.html deleted file mode 100644 index ccebcb4fb1..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-tree.html +++ /dev/null @@ -1,168 +0,0 @@ - - - - - -org.apache.fluss.client.metrics Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.client.metrics

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    -
      -
    • java.lang.Object -
        -
      • org.apache.fluss.metrics.groups.AbstractMetricGroup (implements org.apache.fluss.metrics.groups.MetricGroup) - -
      • -
      -
    • -
    -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-use.html deleted file mode 100644 index ee24d19b1a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/metrics/package-use.html +++ /dev/null @@ -1,208 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.client.metrics (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.client.metrics

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-summary.html deleted file mode 100644 index 444abc2cbf..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-summary.html +++ /dev/null @@ -1,191 +0,0 @@ - - - - - -org.apache.fluss.client (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.client

    -
    -
    -
      -
    • - - - - - - - - - - - - -
      Interface Summary 
      InterfaceDescription
      Connection -
      A cluster connection encapsulating lower level individual connections to actual Fluss servers.
      -
      -
    • -
    • - - - - - - - - - - - - - - - - -
      Class Summary 
      ClassDescription
      ConnectionFactory -
      A non-instantiable class that manages creation of Connections.
      -
      FlussConnection -
      A connection to Fluss cluster, and holds the client session resources.
      -
      -
    • -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-tree.html deleted file mode 100644 index d44d4c30ce..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-tree.html +++ /dev/null @@ -1,174 +0,0 @@ - - - - - -org.apache.fluss.client Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.client

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -

    Interface Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-use.html deleted file mode 100644 index df0ee0dd15..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/package-use.html +++ /dev/null @@ -1,254 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.client (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.client

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/FlussTable.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/FlussTable.html deleted file mode 100644 index 50c06a5414..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/FlussTable.html +++ /dev/null @@ -1,451 +0,0 @@ - - - - - -FlussTable (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlussTable

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      AutoCloseable, Table
      -
      -
      -
      @PublicEvolving
      -public class FlussTable
      -extends Object
      -implements Table
      -
      The base impl of Table.
      -
      -
      Since:
      -
      0.1
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlussTable

          -
          public FlussTable​(FlussConnection conn,
          -                  org.apache.fluss.metadata.TablePath tablePath,
          -                  org.apache.fluss.metadata.TableInfo tableInfo)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getTableInfo

          -
          public org.apache.fluss.metadata.TableInfo getTableInfo()
          -
          Description copied from interface: Table
          -
          Get the TableInfo for this table. - -

          Note: the table info of this Table is set during the creation of this Table and will not be updated after that, even if the table info of the table has been - changed. Therefore, if there are any changes to the table info, it may be necessary to - reconstruct the Table.

          -
          -
          Specified by:
          -
          getTableInfo in interface Table
          -
          -
        • -
        - - - -
          -
        • -

          newScan

          -
          public Scan newScan()
          -
          Description copied from interface: Table
          -
          Creates a new Scan for this table to configure and create a scanner to scan data for - this table. The scanner can be a log scanner to continuously read streaming log data or a - batch scanner to read batch data.
          -
          -
          Specified by:
          -
          newScan in interface Table
          -
          -
        • -
        - - - -
          -
        • -

          newLookup

          -
          public Lookup newLookup()
          -
          Description copied from interface: Table
          -
          Creates a new Lookup for this table to configure and create a Lookuper to - lookup data for this table by primary key or a prefix of primary key.
          -
          -
          Specified by:
          -
          newLookup in interface Table
          -
          -
        • -
        - - - -
          -
        • -

          newAppend

          -
          public Append newAppend()
          -
          Description copied from interface: Table
          -
          Creates a new Append to build a AppendWriter to append data to this table - (requires to be a Log Table).
          -
          -
          Specified by:
          -
          newAppend in interface Table
          -
          -
        • -
        - - - -
          -
        • -

          newUpsert

          -
          public Upsert newUpsert()
          -
          Description copied from interface: Table
          -
          Creates a new Upsert to build a UpsertWriter to upsert and delete data to - this table (requires to be a Primary Key Table).
          -
          -
          Specified by:
          -
          newUpsert in interface Table
          -
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/Table.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/Table.html deleted file mode 100644 index 03552fa610..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/Table.html +++ /dev/null @@ -1,358 +0,0 @@ - - - - - -Table (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface Table

    -
    -
    -
    -
      -
    • -
      -
      All Superinterfaces:
      -
      AutoCloseable
      -
      -
      -
      All Known Implementing Classes:
      -
      FlussTable
      -
      -
      -
      @PublicEvolving
      -public interface Table
      -extends AutoCloseable
      -
      Used to communicate with a single Fluss table. Obtain an instance from a Connection. - -

      Table can be used to get, put, delete or scan data from a fluss table.

      -
      -
      Since:
      -
      0.1
      -
      -
    • -
    -
    -
    -
      -
    • - -
      - -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getTableInfo

          -
          org.apache.fluss.metadata.TableInfo getTableInfo()
          -
          Get the TableInfo for this table. - -

          Note: the table info of this Table is set during the creation of this Table and will not be updated after that, even if the table info of the table has been - changed. Therefore, if there are any changes to the table info, it may be necessary to - reconstruct the Table.

          -
        • -
        - - - -
          -
        • -

          newScan

          -
          Scan newScan()
          -
          Creates a new Scan for this table to configure and create a scanner to scan data for - this table. The scanner can be a log scanner to continuously read streaming log data or a - batch scanner to read batch data.
          -
        • -
        - - - -
          -
        • -

          newLookup

          -
          Lookup newLookup()
          -
          Creates a new Lookup for this table to configure and create a Lookuper to - lookup data for this table by primary key or a prefix of primary key.
          -
        • -
        - - - -
          -
        • -

          newAppend

          -
          Append newAppend()
          -
          Creates a new Append to build a AppendWriter to append data to this table - (requires to be a Log Table).
          -
        • -
        - - - -
          -
        • -

          newUpsert

          -
          Upsert newUpsert()
          -
          Creates a new Upsert to build a UpsertWriter to upsert and delete data to - this table (requires to be a Primary Key Table).
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/class-use/FlussTable.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/class-use/FlussTable.html deleted file mode 100644 index 3a9ebf5a3e..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/class-use/FlussTable.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.FlussTable (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.FlussTable

    -
    -
    No usage of org.apache.fluss.client.table.FlussTable
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/class-use/Table.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/class-use/Table.html deleted file mode 100644 index 51bae671c7..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/class-use/Table.html +++ /dev/null @@ -1,306 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.client.table.Table (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.client.table.Table

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/PartitionGetter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/PartitionGetter.html deleted file mode 100644 index 5c0474d514..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/PartitionGetter.html +++ /dev/null @@ -1,309 +0,0 @@ - - - - - -PartitionGetter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class PartitionGetter

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.table.getter.PartitionGetter
      • -
      -
    • -
    -
    -
      -
    • -
      -
      public class PartitionGetter
      -extends Object
      -
      A getter to get partition name from a row.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          PartitionGetter

          -
          public PartitionGetter​(org.apache.fluss.types.RowType rowType,
          -                       List<String> partitionKeys)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getPartition

          -
          public String getPartition​(org.apache.fluss.row.InternalRow row)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/class-use/PartitionGetter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/class-use/PartitionGetter.html deleted file mode 100644 index 237574e2a2..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/class-use/PartitionGetter.html +++ /dev/null @@ -1,197 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.getter.PartitionGetter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.getter.PartitionGetter

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-summary.html deleted file mode 100644 index bddc3fd654..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-summary.html +++ /dev/null @@ -1,168 +0,0 @@ - - - - - -org.apache.fluss.client.table.getter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.client.table.getter

    -
    -
    -
      -
    • - - - - - - - - - - - - -
      Class Summary 
      ClassDescription
      PartitionGetter -
      A getter to get partition name from a row.
      -
      -
    • -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-tree.html deleted file mode 100644 index 27b4de7e79..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-tree.html +++ /dev/null @@ -1,163 +0,0 @@ - - - - - -org.apache.fluss.client.table.getter Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.client.table.getter

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-use.html deleted file mode 100644 index c86e5968cb..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/getter/package-use.html +++ /dev/null @@ -1,185 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.client.table.getter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.client.table.getter

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-summary.html deleted file mode 100644 index 4c6b88ad7f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-summary.html +++ /dev/null @@ -1,185 +0,0 @@ - - - - - -org.apache.fluss.client.table (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.client.table

    -
    -
    -
      -
    • - - - - - - - - - - - - -
      Interface Summary 
      InterfaceDescription
      Table -
      Used to communicate with a single Fluss table.
      -
      -
    • -
    • - - - - - - - - - - - - -
      Class Summary 
      ClassDescription
      FlussTable -
      The base impl of Table.
      -
      -
    • -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-tree.html deleted file mode 100644 index 0c3b04ded2..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-tree.html +++ /dev/null @@ -1,173 +0,0 @@ - - - - - -org.apache.fluss.client.table Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.client.table

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    -
      -
    • java.lang.Object -
        -
      • org.apache.fluss.client.table.FlussTable (implements org.apache.fluss.client.table.Table)
      • -
      -
    • -
    -
    -
    -

    Interface Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-use.html deleted file mode 100644 index f8fd961b49..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/package-use.html +++ /dev/null @@ -1,277 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.client.table (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.client.table

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/RemoteFileDownloader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/RemoteFileDownloader.html deleted file mode 100644 index 97dd7905de..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/RemoteFileDownloader.html +++ /dev/null @@ -1,435 +0,0 @@ - - - - - -RemoteFileDownloader (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class RemoteFileDownloader

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.table.scanner.RemoteFileDownloader
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      Closeable, AutoCloseable
      -
      -
      -
      public class RemoteFileDownloader
      -extends Object
      -implements Closeable
      -
      The downloader that has a IO thread pool to download the remote files (like kv snapshots files, - log segment files).
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      - -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          RemoteFileDownloader

          -
          public RemoteFileDownloader​(int threadNum)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          downloadFileAsync

          -
          public CompletableFuture<Long> downloadFileAsync​(org.apache.fluss.fs.FsPathAndFileName fsPathAndFileName,
          -                                                 Path targetDirectory)
          -
          Downloads the file from the given remote file path to the target directory asynchronously, - returns a Future object of the number of downloaded bytes. The Future will fail if the - download fails after retrying for RETRY_COUNT times.
          -
        • -
        - - - -
          -
        • -

          downloadFile

          -
          protected long downloadFile​(Path targetFilePath,
          -                            org.apache.fluss.fs.FsPath remoteFilePath)
          -                     throws IOException
          -
          Copies the file from a remote file path to the given target file path, returns the number of - downloaded bytes.
          -
          -
          Throws:
          -
          IOException
          -
          -
        • -
        - - - - - - - -
          -
        • -

          transferAllToDirectory

          -
          public void transferAllToDirectory​(List<org.apache.fluss.fs.FsPathAndFileName> fsPathAndFileNames,
          -                                   Path targetDirectory,
          -                                   org.apache.fluss.utils.CloseableRegistry closeableRegistry)
          -                            throws IOException
          -
          -
          Throws:
          -
          IOException
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/Scan.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/Scan.html deleted file mode 100644 index 1f1ab5c3b2..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/Scan.html +++ /dev/null @@ -1,374 +0,0 @@ - - - - - -Scan (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface Scan

    -
    -
    -
    -
      -
    • -
      -
      All Known Implementing Classes:
      -
      TableScan
      -
      -
      -
      @PublicEvolving
      -public interface Scan
      -
      Used to configure and create a scanner to scan data for a table. - -

      Scan objects are immutable and can be shared between threads. Refinement methods, like - project(int[]) and limit(int), create new Scan instances.

      -
      -
      Since:
      -
      0.6
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Summary

        - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
        All Methods Instance Methods Abstract Methods 
        Modifier and TypeMethodDescription
        BatchScannercreateBatchScanner​(org.apache.fluss.metadata.TableBucket tableBucket) -
        Creates a BatchScanner to read current data in the given table bucket for this scan.
        -
        BatchScannercreateBatchScanner​(org.apache.fluss.metadata.TableBucket tableBucket, - long snapshotId) -
        Creates a BatchScanner to read given snapshot data in the given table bucket for this - scan.
        -
        LogScannercreateLogScanner() -
        Creates a LogScanner to continuously read log data for this scan.
        -
        Scanlimit​(int rowNumber) -
        Returns a new scan from this that will read the given limited row number.
        -
        Scanproject​(int[] projectedColumns) -
        Returns a new scan from this that will read the given data columns.
        -
        Scanproject​(List<String> projectedColumnNames) -
        Returns a new scan from this that will read the given data columns.
        -
        -
      • -
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          project

          -
          Scan project​(@Nullable
          -             int[] projectedColumns)
          -
          Returns a new scan from this that will read the given data columns.
          -
          -
          Parameters:
          -
          projectedColumns - the selected column indexes
          -
          -
        • -
        - - - -
          -
        • -

          project

          -
          Scan project​(List<String> projectedColumnNames)
          -
          Returns a new scan from this that will read the given data columns.
          -
          -
          Parameters:
          -
          projectedColumnNames - the selected column names
          -
          -
        • -
        - - - -
          -
        • -

          limit

          -
          Scan limit​(int rowNumber)
          -
          Returns a new scan from this that will read the given limited row number.
          -
          -
          Parameters:
          -
          rowNumber - the limited row number to read
          -
          -
        • -
        - - - -
          -
        • -

          createLogScanner

          -
          LogScanner createLogScanner()
          -
          Creates a LogScanner to continuously read log data for this scan. - -

          Note: this API doesn't support pre-configured with limit(int).

          -
        • -
        - - - -
          -
        • -

          createBatchScanner

          -
          BatchScanner createBatchScanner​(org.apache.fluss.metadata.TableBucket tableBucket)
          -
          Creates a BatchScanner to read current data in the given table bucket for this scan. - -

          Note: this API doesn't support pre-configured with project(int[]).

          -
        • -
        - - - -
          -
        • -

          createBatchScanner

          -
          BatchScanner createBatchScanner​(org.apache.fluss.metadata.TableBucket tableBucket,
          -                                long snapshotId)
          -
          Creates a BatchScanner to read given snapshot data in the given table bucket for this - scan. - -

          Note: this API doesn't support pre-configured with project(int[]) and limit(int) and only support for Primary Key Tables.

          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/ScanRecord.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/ScanRecord.html deleted file mode 100644 index 8c7501644f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/ScanRecord.html +++ /dev/null @@ -1,447 +0,0 @@ - - - - - -ScanRecord (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class ScanRecord

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.fluss.record.LogRecord
      -
      -
      -
      @Internal
      -public class ScanRecord
      -extends Object
      -implements org.apache.fluss.record.LogRecord
      -
      one scan record.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          ScanRecord

          -
          public ScanRecord​(org.apache.fluss.row.InternalRow row)
          -
        • -
        - - - -
          -
        • -

          ScanRecord

          -
          public ScanRecord​(long offset,
          -                  long timestamp,
          -                  org.apache.fluss.record.ChangeType changeType,
          -                  org.apache.fluss.row.InternalRow row)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          logOffset

          -
          public long logOffset()
          -
          The position of this record in the corresponding fluss table bucket.
          -
          -
          Specified by:
          -
          logOffset in interface org.apache.fluss.record.LogRecord
          -
          -
        • -
        - - - -
          -
        • -

          timestamp

          -
          public long timestamp()
          -
          -
          Specified by:
          -
          timestamp in interface org.apache.fluss.record.LogRecord
          -
          -
        • -
        - - - -
          -
        • -

          getChangeType

          -
          public org.apache.fluss.record.ChangeType getChangeType()
          -
          -
          Specified by:
          -
          getChangeType in interface org.apache.fluss.record.LogRecord
          -
          -
        • -
        - - - -
          -
        • -

          getRow

          -
          public org.apache.fluss.row.InternalRow getRow()
          -
          -
          Specified by:
          -
          getRow in interface org.apache.fluss.record.LogRecord
          -
          -
        • -
        - - - -
          -
        • -

          equals

          -
          public boolean equals​(Object o)
          -
          -
          Overrides:
          -
          equals in class Object
          -
          -
        • -
        - - - -
          -
        • -

          hashCode

          -
          public int hashCode()
          -
          -
          Overrides:
          -
          hashCode in class Object
          -
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/TableScan.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/TableScan.html deleted file mode 100644 index 057f3828ff..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/TableScan.html +++ /dev/null @@ -1,449 +0,0 @@ - - - - - -TableScan (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TableScan

    -
    -
    - -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          TableScan

          -
          public TableScan​(FlussConnection conn,
          -                 org.apache.fluss.metadata.TableInfo tableInfo)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          project

          -
          public Scan project​(@Nullable
          -                    int[] projectedColumns)
          -
          Description copied from interface: Scan
          -
          Returns a new scan from this that will read the given data columns.
          -
          -
          Specified by:
          -
          project in interface Scan
          -
          Parameters:
          -
          projectedColumns - the selected column indexes
          -
          -
        • -
        - - - -
          -
        • -

          project

          -
          public Scan project​(List<String> projectedColumnNames)
          -
          Description copied from interface: Scan
          -
          Returns a new scan from this that will read the given data columns.
          -
          -
          Specified by:
          -
          project in interface Scan
          -
          Parameters:
          -
          projectedColumnNames - the selected column names
          -
          -
        • -
        - - - -
          -
        • -

          limit

          -
          public Scan limit​(int rowNumber)
          -
          Description copied from interface: Scan
          -
          Returns a new scan from this that will read the given limited row number.
          -
          -
          Specified by:
          -
          limit in interface Scan
          -
          Parameters:
          -
          rowNumber - the limited row number to read
          -
          -
        • -
        - - - -
          -
        • -

          createLogScanner

          -
          public LogScanner createLogScanner()
          -
          Description copied from interface: Scan
          -
          Creates a LogScanner to continuously read log data for this scan. - -

          Note: this API doesn't support pre-configured with Scan.limit(int).

          -
          -
          Specified by:
          -
          createLogScanner in interface Scan
          -
          -
        • -
        - - - -
          -
        • -

          createBatchScanner

          -
          public BatchScanner createBatchScanner​(org.apache.fluss.metadata.TableBucket tableBucket)
          -
          Description copied from interface: Scan
          -
          Creates a BatchScanner to read current data in the given table bucket for this scan. - -

          Note: this API doesn't support pre-configured with Scan.project(int[]).

          -
          -
          Specified by:
          -
          createBatchScanner in interface Scan
          -
          -
        • -
        - - - -
          -
        • -

          createBatchScanner

          -
          public BatchScanner createBatchScanner​(org.apache.fluss.metadata.TableBucket tableBucket,
          -                                       long snapshotId)
          -
          Description copied from interface: Scan
          -
          Creates a BatchScanner to read given snapshot data in the given table bucket for this - scan. - -

          Note: this API doesn't support pre-configured with Scan.project(int[]) and Scan.limit(int) and only support for Primary Key Tables.

          -
          -
          Specified by:
          -
          createBatchScanner in interface Scan
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/BatchScanUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/BatchScanUtils.html deleted file mode 100644 index 8d98ac7d97..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/BatchScanUtils.html +++ /dev/null @@ -1,367 +0,0 @@ - - - - - -BatchScanUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class BatchScanUtils

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.table.scanner.batch.BatchScanUtils
      • -
      -
    • -
    -
    -
      -
    • -
      -
      public class BatchScanUtils
      -extends Object
      -
      Utility class for batch scan.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          BatchScanUtils

          -
          public BatchScanUtils()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          collectRows

          -
          public static List<org.apache.fluss.row.InternalRow> collectRows​(BatchScanner scanner)
          -
          Collect all rows from the scanner.
          -
        • -
        - - - -
          -
        • -

          collectAllRows

          -
          public static List<org.apache.fluss.row.InternalRow> collectAllRows​(List<BatchScanner> scanners)
          -
          Collect all rows from the scanners.
          -
        • -
        - - - -
          -
        • -

          collectLimitedRows

          -
          public static List<org.apache.fluss.row.InternalRow> collectLimitedRows​(List<BatchScanner> scanners,
          -                                                                        int limit)
          -
          Collect rows from the scanners until the number of rows reaches the limit.
          -
        • -
        - - - -
          -
        • -

          toList

          -
          public static <T> List<T> toList​(org.apache.fluss.utils.CloseableIterator<T> iterator)
          -
          Collect all rows from the iterator and close it.
          -
          -
          Returns:
          -
          return a list of rows
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/BatchScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/BatchScanner.html deleted file mode 100644 index d97399b796..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/BatchScanner.html +++ /dev/null @@ -1,305 +0,0 @@ - - - - - -BatchScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface BatchScanner

    -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Summary

        - - - - - - - - - - - - - - - - - -
        All Methods Instance Methods Abstract Methods 
        Modifier and TypeMethodDescription
        voidclose() -
        Closes the scanner and should release all resources.
        -
        org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow>pollBatch​(Duration timeout) -
        Poll one batch records.
        -
        -
      • -
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          pollBatch

          -
          @Nullable
          -org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow> pollBatch​(Duration timeout)
          -                                                                              throws IOException
          -
          Poll one batch records. The method should return null when reaching the end of the input.
          -
          -
          Parameters:
          -
          timeout - The maximum time to block (must not be greater than Long.MAX_VALUE - milliseconds)
          -
          Throws:
          -
          IOException
          -
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/KvSnapshotBatchScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/KvSnapshotBatchScanner.html deleted file mode 100644 index bf91387213..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/KvSnapshotBatchScanner.html +++ /dev/null @@ -1,423 +0,0 @@ - - - - - -KvSnapshotBatchScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class KvSnapshotBatchScanner

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.table.scanner.batch.KvSnapshotBatchScanner
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      Closeable, AutoCloseable, BatchScanner
      -
      -
      -
      @Internal
      -public class KvSnapshotBatchScanner
      -extends Object
      -implements BatchScanner
      -
      A scanner to scan the snapshot data of a kv bucket. - -

      When the scanner is created, It will download the given snapshot files from the remote storage - and create a reader to read the files into records asynchronously. - -

      In the BatchScanner.pollBatch(Duration) method: - -

        -
      • if the reader is not ready in given time, return an empty iterator -
      • If the reader is ready, always return the reader if there remains any data in the reader. - Otherwise, return null -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Summary

        - - - - - - - - - - - - -
        Fields 
        Modifier and TypeFieldDescription
        static org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow>NO_DATA_AVAILABLE 
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Summary

        - - - - - - - - - - -
        Constructors 
        ConstructorDescription
        KvSnapshotBatchScanner​(org.apache.fluss.types.RowType tableRowType, - org.apache.fluss.metadata.TableBucket tableBucket, - List<org.apache.fluss.fs.FsPathAndFileName> fsPathAndFileNames, - int[] projectedFields, - String scannerTmpDir, - org.apache.fluss.metadata.KvFormat kvFormat, - RemoteFileDownloader remoteFileDownloader) 
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - -
          -
        • -

          NO_DATA_AVAILABLE

          -
          public static final org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow> NO_DATA_AVAILABLE
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          KvSnapshotBatchScanner

          -
          public KvSnapshotBatchScanner​(org.apache.fluss.types.RowType tableRowType,
          -                              org.apache.fluss.metadata.TableBucket tableBucket,
          -                              List<org.apache.fluss.fs.FsPathAndFileName> fsPathAndFileNames,
          -                              @Nullable
          -                              int[] projectedFields,
          -                              String scannerTmpDir,
          -                              org.apache.fluss.metadata.KvFormat kvFormat,
          -                              RemoteFileDownloader remoteFileDownloader)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          pollBatch

          -
          @Nullable
          -public org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow> pollBatch​(Duration timeout)
          -                                                                                     throws IOException
          -
          Fetch data from snapshot. - -

          If the snapshot file reader is not ready in given maximum block time, return an empty - iterator. If the reader is ready, always return the reader if there remains any record in the - reader, otherwise, return null.

          -
          -
          Specified by:
          -
          pollBatch in interface BatchScanner
          -
          Parameters:
          -
          timeout - The maximum time to block (must not be greater than Long.MAX_VALUE - milliseconds)
          -
          Throws:
          -
          IOException
          -
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/LimitBatchScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/LimitBatchScanner.html deleted file mode 100644 index f4c95f0785..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/LimitBatchScanner.html +++ /dev/null @@ -1,365 +0,0 @@ - - - - - -LimitBatchScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LimitBatchScanner

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.table.scanner.batch.LimitBatchScanner
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LimitBatchScanner

          -
          public LimitBatchScanner​(org.apache.fluss.metadata.TableInfo tableInfo,
          -                         org.apache.fluss.metadata.TableBucket tableBucket,
          -                         MetadataUpdater metadataUpdater,
          -                         @Nullable
          -                         int[] projectedFields,
          -                         int limit)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          pollBatch

          -
          @Nullable
          -public org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow> pollBatch​(Duration timeout)
          -                                                                                     throws IOException
          -
          Description copied from interface: BatchScanner
          -
          Poll one batch records. The method should return null when reaching the end of the input.
          -
          -
          Specified by:
          -
          pollBatch in interface BatchScanner
          -
          Parameters:
          -
          timeout - The maximum time to block (must not be greater than Long.MAX_VALUE - milliseconds)
          -
          Throws:
          -
          IOException
          -
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/BatchScanUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/BatchScanUtils.html deleted file mode 100644 index 75777dd2d8..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/BatchScanUtils.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.scanner.batch.BatchScanUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.scanner.batch.BatchScanUtils

    -
    -
    No usage of org.apache.fluss.client.table.scanner.batch.BatchScanUtils
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/BatchScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/BatchScanner.html deleted file mode 100644 index 6be498059a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/BatchScanner.html +++ /dev/null @@ -1,358 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.client.table.scanner.batch.BatchScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.client.table.scanner.batch.BatchScanner

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/KvSnapshotBatchScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/KvSnapshotBatchScanner.html deleted file mode 100644 index 75269c070d..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/KvSnapshotBatchScanner.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.scanner.batch.KvSnapshotBatchScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.scanner.batch.KvSnapshotBatchScanner

    -
    -
    No usage of org.apache.fluss.client.table.scanner.batch.KvSnapshotBatchScanner
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/LimitBatchScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/LimitBatchScanner.html deleted file mode 100644 index 7499305f36..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/class-use/LimitBatchScanner.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.scanner.batch.LimitBatchScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.scanner.batch.LimitBatchScanner

    -
    -
    No usage of org.apache.fluss.client.table.scanner.batch.LimitBatchScanner
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-summary.html deleted file mode 100644 index 579e32d9b7..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-summary.html +++ /dev/null @@ -1,197 +0,0 @@ - - - - - -org.apache.fluss.client.table.scanner.batch (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.client.table.scanner.batch

    -
    -
    -
      -
    • - - - - - - - - - - - - -
      Interface Summary 
      InterfaceDescription
      BatchScanner -
      The scanner that reads records form a table in a batch fashion.
      -
      -
    • -
    • - - - - - - - - - - - - - - - - - - - - -
      Class Summary 
      ClassDescription
      BatchScanUtils -
      Utility class for batch scan.
      -
      KvSnapshotBatchScanner -
      A scanner to scan the snapshot data of a kv bucket.
      -
      LimitBatchScanner -
      A BatchScanner implementation that scans a limited number of records from a table.
      -
      -
    • -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-tree.html deleted file mode 100644 index c3eaf1e8a8..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-tree.html +++ /dev/null @@ -1,179 +0,0 @@ - - - - - -org.apache.fluss.client.table.scanner.batch Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.client.table.scanner.batch

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -

    Interface Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-use.html deleted file mode 100644 index 0277e6da45..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/batch/package-use.html +++ /dev/null @@ -1,254 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.client.table.scanner.batch (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.client.table.scanner.batch

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/RemoteFileDownloader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/RemoteFileDownloader.html deleted file mode 100644 index cafc462740..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/RemoteFileDownloader.html +++ /dev/null @@ -1,268 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.scanner.RemoteFileDownloader (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.scanner.RemoteFileDownloader

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/Scan.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/Scan.html deleted file mode 100644 index 3a07484684..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/Scan.html +++ /dev/null @@ -1,274 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.client.table.scanner.Scan (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.client.table.scanner.Scan

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/ScanRecord.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/ScanRecord.html deleted file mode 100644 index b99c0ef1f1..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/ScanRecord.html +++ /dev/null @@ -1,351 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.scanner.ScanRecord (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.scanner.ScanRecord

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/TableScan.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/TableScan.html deleted file mode 100644 index c6ab351270..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/class-use/TableScan.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.scanner.TableScan (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.scanner.TableScan

    -
    -
    No usage of org.apache.fluss.client.table.scanner.TableScan
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetchBuffer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetchBuffer.html deleted file mode 100644 index 6bf6232f39..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetchBuffer.html +++ /dev/null @@ -1,340 +0,0 @@ - - - - - -LogFetchBuffer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LogFetchBuffer

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.table.scanner.log.LogFetchBuffer
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      AutoCloseable
      -
      -
      -
      @ThreadSafe
      -@Internal
      -public class LogFetchBuffer
      -extends Object
      -implements AutoCloseable
      -
      LogFetchBuffer buffers up the results from the tablet server - responses as they are received. It's essentially a wrapper around a Queue of - CompletedFetch. There is at most one LogFetchBuffer per bucket in the queue. - -

      Note: this class is thread-safe with the intention that the data will - be created by a background thread and consumed by the application thread.

      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LogFetchBuffer

          -
          public LogFetchBuffer()
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetchCollector.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetchCollector.html deleted file mode 100644 index 7b6d0349e4..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetchCollector.html +++ /dev/null @@ -1,329 +0,0 @@ - - - - - -LogFetchCollector (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LogFetchCollector

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.table.scanner.log.LogFetchCollector
      • -
      -
    • -
    -
    -
      -
    • -
      -
      @ThreadSafe
      -@Internal
      -public class LogFetchCollector
      -extends Object
      -
      LogFetchCollector operates at the LogRecordBatch level, as that is what is stored - in the LogFetchBuffer. Each LogRecord in the LogRecordBatch is converted - to a ScanRecord and added to the returned LogFetcher.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LogFetchCollector

          -
          public LogFetchCollector​(org.apache.fluss.metadata.TablePath tablePath,
          -                         LogScannerStatus logScannerStatus,
          -                         org.apache.fluss.config.Configuration conf,
          -                         MetadataUpdater metadataUpdater)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          collectFetch

          -
          public Map<org.apache.fluss.metadata.TableBucket,​List<ScanRecord>> collectFetch​(LogFetchBuffer logFetchBuffer)
          -
          Return the fetched log records, empty the record buffer and update the consumed position. - -

          NOTE: returning empty records guarantees the consumed position are NOT updated.

          -
          -
          Returns:
          -
          The fetched records per partition
          -
          Throws:
          -
          org.apache.fluss.exception.LogOffsetOutOfRangeException - If there is OffsetOutOfRange error in fetchResponse and - the defaultResetPolicy is NONE
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetcher.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetcher.html deleted file mode 100644 index b52d9827ac..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogFetcher.html +++ /dev/null @@ -1,423 +0,0 @@ - - - - - -LogFetcher (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LogFetcher

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.table.scanner.log.LogFetcher
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LogFetcher

          -
          public LogFetcher​(org.apache.fluss.metadata.TableInfo tableInfo,
          -                  @Nullable
          -                  org.apache.fluss.utils.Projection projection,
          -                  LogScannerStatus logScannerStatus,
          -                  org.apache.fluss.config.Configuration conf,
          -                  MetadataUpdater metadataUpdater,
          -                  ScannerMetricGroup scannerMetricGroup,
          -                  RemoteFileDownloader remoteFileDownloader)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          hasAvailableFetches

          -
          public boolean hasAvailableFetches()
          -
          Return whether we have any completed fetches that are fetch-able. This method is thread-safe.
          -
          -
          Returns:
          -
          true if there are completed fetches that can be returned, false otherwise
          -
          -
        • -
        - - - -
          -
        • -

          collectFetch

          -
          public Map<org.apache.fluss.metadata.TableBucket,​List<ScanRecord>> collectFetch()
          -
        • -
        - - - -
          -
        • -

          sendFetches

          -
          public void sendFetches()
          -
          Set up a fetch request for any node that we have assigned buckets for which doesn't already - have an in-flight fetch or pending fetch data.
          -
        • -
        - - - -
          -
        • -

          awaitNotEmpty

          -
          public boolean awaitNotEmpty​(long deadlineNanos)
          -
          -
          Parameters:
          -
          deadlineNanos - the deadline time to wait until
          -
          Returns:
          -
          false if the waiting time detectably elapsed before return from the method, else true
          -
          -
        • -
        - - - -
          -
        • -

          wakeup

          -
          public void wakeup()
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScan.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScan.html deleted file mode 100644 index 5e6e5c3a6c..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScan.html +++ /dev/null @@ -1,334 +0,0 @@ - - - - - -LogScan (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LogScan

    -
    -
    - -
    -
      -
    • -
      -
      @PublicEvolving
      -public class LogScan
      -extends Object
      -
      Used to describe the operation to scan log data by LogScanner to a table.
      -
      -
      Since:
      -
      0.1
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LogScan

          -
          public LogScan()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          withProjectedFields

          -
          public LogScan withProjectedFields​(int[] projectedFields)
          -
          Returns a new instance of LogScan description with column projection.
          -
          -
          Parameters:
          -
          projectedFields - the projection fields
          -
          -
        • -
        - - - -
          -
        • -

          getProjectedFields

          -
          @Nullable
          -public int[] getProjectedFields()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScanner.html deleted file mode 100644 index 89c4845c60..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScanner.html +++ /dev/null @@ -1,503 +0,0 @@ - - - - - -LogScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface LogScanner

    -
    -
    -
    -
      -
    • -
      -
      All Superinterfaces:
      -
      AutoCloseable
      -
      -
      -
      All Known Implementing Classes:
      -
      LogScannerImpl
      -
      -
      -
      @PublicEvolving
      -public interface LogScanner
      -extends AutoCloseable
      -
      The scanner is used to scan log data of specify table from Fluss.
      -
      -
      Since:
      -
      0.1
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Summary

        - - - - - - - - - - - - -
        Fields 
        Modifier and TypeFieldDescription
        static longEARLIEST_OFFSET -
        The earliest offset to fetch from.
        -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Summary

        - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
        All Methods Instance Methods Abstract Methods Default Methods 
        Modifier and TypeMethodDescription
        ScanRecordspoll​(Duration timeout) -
        Poll log data from tablet server.
        -
        voidsubscribe​(int bucket, - long offset) -
        Subscribe to the given table bucket in given offset dynamically.
        -
        voidsubscribe​(long partitionId, - int bucket, - long offset) -
        Subscribe to the given partitioned table bucket in given offset dynamically.
        -
        default voidsubscribeFromBeginning​(int bucket) -
        Subscribe to the given table buckets from beginning dynamically.
        -
        default voidsubscribeFromBeginning​(long partitionId, - int bucket) -
        Subscribe to the given partitioned table bucket from beginning dynamically.
        -
        voidunsubscribe​(long partitionId, - int bucket) -
        Unsubscribe from the given bucket of given partition dynamically.
        -
        voidwakeup() -
        Wake up the log scanner in case the fetcher thread in log scanner is blocking in poll(Duration timeout).
        -
        - -
      • -
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - -
          -
        • -

          EARLIEST_OFFSET

          -
          static final long EARLIEST_OFFSET
          -
          The earliest offset to fetch from. Fluss uses "-2" to indicate fetching from log start - offset.
          -
          -
          See Also:
          -
          Constant Field Values
          -
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          poll

          -
          ScanRecords poll​(Duration timeout)
          -
          Poll log data from tablet server. - -

          On each poll, scanner will try to use the last scanned offset as the starting offset and - fetch sequentially. The latest offset will be cached in memory, and if there is no offset - cached in memory, scanner will start either from the earliest offset or from the latest - offset for each bucket according to the policy.

          -
          -
          Parameters:
          -
          timeout - the timeout to poll.
          -
          Returns:
          -
          the result of poll.
          -
          Throws:
          -
          IllegalStateException - if the scanner is not subscribed to any buckets to - read from.
          -
          -
        • -
        - - - -
          -
        • -

          subscribe

          -
          void subscribe​(int bucket,
          -               long offset)
          -
          Subscribe to the given table bucket in given offset dynamically. If the table bucket is - already subscribed, the offset will be updated. - -

          Please use subscribe(long, int, long) to subscribe a partitioned table.

          -
          -
          Parameters:
          -
          bucket - the table bucket to subscribe.
          -
          offset - the offset to start from.
          -
          Throws:
          -
          IllegalStateException - if the table is a partitioned table.
          -
          -
        • -
        - - - -
          -
        • -

          subscribeFromBeginning

          -
          default void subscribeFromBeginning​(int bucket)
          -
          Subscribe to the given table buckets from beginning dynamically. If the table bucket is - already subscribed, the start offset will be updated. - -

          It equals to call subscribe(int, long) with offset EARLIEST_OFFSET for - the subscribed bucket. - -

          Please use subscribeFromBeginning(long, int) to subscribe a partitioned table.

          -
          -
          Parameters:
          -
          bucket - the table bucket to subscribe.
          -
          Throws:
          -
          IllegalStateException - if the table is a partitioned table.
          -
          -
        • -
        - - - -
          -
        • -

          subscribe

          -
          void subscribe​(long partitionId,
          -               int bucket,
          -               long offset)
          -
          Subscribe to the given partitioned table bucket in given offset dynamically. If the table - bucket is already subscribed, the offset will be updated. - -

          Please use subscribe(int, long) to subscribe a non-partitioned table.

          -
          -
          Parameters:
          -
          partitionId - the partition id of the table partition to subscribe.
          -
          bucket - the table bucket to subscribe.
          -
          offset - the offset to start from.
          -
          Throws:
          -
          IllegalStateException - if the table is a non-partitioned table.
          -
          -
        • -
        - - - -
          -
        • -

          unsubscribe

          -
          void unsubscribe​(long partitionId,
          -                 int bucket)
          -
          Unsubscribe from the given bucket of given partition dynamically.
          -
          -
          Parameters:
          -
          partitionId - the partition id of the table partition to unsubscribe.
          -
          bucket - the table bucket to unsubscribe.
          -
          Throws:
          -
          IllegalStateException - if the table is a non-partitioned table.
          -
          -
        • -
        - - - -
          -
        • -

          subscribeFromBeginning

          -
          default void subscribeFromBeginning​(long partitionId,
          -                                    int bucket)
          -
          Subscribe to the given partitioned table bucket from beginning dynamically. If the table - bucket is already subscribed, the start offset will be updated. - -

          It equals to call subscribe(long, int, long) with offset EARLIEST_OFFSET - for the subscribed bucket. - -

          Please use subscribeFromBeginning(int) to subscribe a non-partitioned table.

          -
          -
          Parameters:
          -
          bucket - the table bucket to subscribe.
          -
          Throws:
          -
          IllegalStateException - if the table is a non-partitioned table.
          -
          -
        • -
        - - - -
          -
        • -

          wakeup

          -
          void wakeup()
          -
          Wake up the log scanner in case the fetcher thread in log scanner is blocking in poll(Duration timeout).
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScannerImpl.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScannerImpl.html deleted file mode 100644 index f63ef7841d..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScannerImpl.html +++ /dev/null @@ -1,502 +0,0 @@ - - - - - -LogScannerImpl (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LogScannerImpl

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.table.scanner.log.LogScannerImpl
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LogScannerImpl

          -
          public LogScannerImpl​(org.apache.fluss.config.Configuration conf,
          -                      org.apache.fluss.metadata.TableInfo tableInfo,
          -                      MetadataUpdater metadataUpdater,
          -                      org.apache.fluss.rpc.metrics.ClientMetricGroup clientMetricGroup,
          -                      RemoteFileDownloader remoteFileDownloader,
          -                      @Nullable
          -                      int[] projectedFields)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          poll

          -
          public ScanRecords poll​(Duration timeout)
          -
          Description copied from interface: LogScanner
          -
          Poll log data from tablet server. - -

          On each poll, scanner will try to use the last scanned offset as the starting offset and - fetch sequentially. The latest offset will be cached in memory, and if there is no offset - cached in memory, scanner will start either from the earliest offset or from the latest - offset for each bucket according to the policy.

          -
          -
          Specified by:
          -
          poll in interface LogScanner
          -
          Parameters:
          -
          timeout - the timeout to poll.
          -
          Returns:
          -
          the result of poll.
          -
          -
        • -
        - - - -
          -
        • -

          subscribe

          -
          public void subscribe​(int bucket,
          -                      long offset)
          -
          Description copied from interface: LogScanner
          -
          Subscribe to the given table bucket in given offset dynamically. If the table bucket is - already subscribed, the offset will be updated. - -

          Please use LogScanner.subscribe(long, int, long) to subscribe a partitioned table.

          -
          -
          Specified by:
          -
          subscribe in interface LogScanner
          -
          Parameters:
          -
          bucket - the table bucket to subscribe.
          -
          offset - the offset to start from.
          -
          -
        • -
        - - - -
          -
        • -

          subscribe

          -
          public void subscribe​(long partitionId,
          -                      int bucket,
          -                      long offset)
          -
          Description copied from interface: LogScanner
          -
          Subscribe to the given partitioned table bucket in given offset dynamically. If the table - bucket is already subscribed, the offset will be updated. - -

          Please use LogScanner.subscribe(int, long) to subscribe a non-partitioned table.

          -
          -
          Specified by:
          -
          subscribe in interface LogScanner
          -
          Parameters:
          -
          partitionId - the partition id of the table partition to subscribe.
          -
          bucket - the table bucket to subscribe.
          -
          offset - the offset to start from.
          -
          -
        • -
        - - - -
          -
        • -

          unsubscribe

          -
          public void unsubscribe​(long partitionId,
          -                        int bucket)
          -
          Description copied from interface: LogScanner
          -
          Unsubscribe from the given bucket of given partition dynamically.
          -
          -
          Specified by:
          -
          unsubscribe in interface LogScanner
          -
          Parameters:
          -
          partitionId - the partition id of the table partition to unsubscribe.
          -
          bucket - the table bucket to unsubscribe.
          -
          -
        • -
        - - - - - - - -
          -
        • -

          close

          -
          public void close()
          -
          -
          Specified by:
          -
          close in interface AutoCloseable
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScannerStatus.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScannerStatus.html deleted file mode 100644 index 7e944852be..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/LogScannerStatus.html +++ /dev/null @@ -1,271 +0,0 @@ - - - - - -LogScannerStatus (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LogScannerStatus

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.table.scanner.log.LogScannerStatus
      • -
      -
    • -
    -
    -
      -
    • -
      -
      @ThreadSafe
      -@Internal
      -public class LogScannerStatus
      -extends Object
      -
      The status of a LogScanner.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LogScannerStatus

          -
          public LogScannerStatus()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/RemoteLogDownloadFuture.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/RemoteLogDownloadFuture.html deleted file mode 100644 index 887caf6a81..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/RemoteLogDownloadFuture.html +++ /dev/null @@ -1,351 +0,0 @@ - - - - - -RemoteLogDownloadFuture (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class RemoteLogDownloadFuture

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.table.scanner.log.RemoteLogDownloadFuture
      • -
      -
    • -
    -
    -
      -
    • -
      -
      public class RemoteLogDownloadFuture
      -extends Object
      -
      Represents the future of a remote log download request.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      - -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          isDone

          -
          public boolean isDone()
          -
        • -
        - - - -
          -
        • -

          getFileLogRecords

          -
          public org.apache.fluss.record.FileLogRecords getFileLogRecords​(int startPosition)
          -
        • -
        - - - -
          -
        • -

          getRecycleCallback

          -
          public Runnable getRecycleCallback()
          -
        • -
        - - - -
          -
        • -

          onComplete

          -
          public void onComplete​(Runnable callback)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/RemoteLogDownloader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/RemoteLogDownloader.html deleted file mode 100644 index 3ef589fba8..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/RemoteLogDownloader.html +++ /dev/null @@ -1,378 +0,0 @@ - - - - - -RemoteLogDownloader (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class RemoteLogDownloader

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.table.scanner.log.RemoteLogDownloader
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      Closeable, AutoCloseable
      -
      -
      -
      @ThreadSafe
      -@Internal
      -public class RemoteLogDownloader
      -extends Object
      -implements Closeable
      -
      Downloader to read remote log files to local disk.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          RemoteLogDownloader

          -
          public RemoteLogDownloader​(org.apache.fluss.metadata.TablePath tablePath,
          -                           org.apache.fluss.config.Configuration conf,
          -                           RemoteFileDownloader remoteFileDownloader,
          -                           ScannerMetricGroup scannerMetricGroup)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          start

          -
          public void start()
          -
        • -
        - - - -
          -
        • -

          requestRemoteLog

          -
          public RemoteLogDownloadFuture requestRemoteLog​(org.apache.fluss.fs.FsPath logTabletDir,
          -                                                org.apache.fluss.remote.RemoteLogSegment segment)
          -
          Request to fetch remote log segment to local. This method is non-blocking.
          -
        • -
        - - - - - - - -
          -
        • -

          getFsPathAndFileName

          -
          protected static org.apache.fluss.fs.FsPathAndFileName getFsPathAndFileName​(org.apache.fluss.fs.FsPath remoteLogTabletDir,
          -                                                                            org.apache.fluss.remote.RemoteLogSegment segment)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/ScanRecords.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/ScanRecords.html deleted file mode 100644 index ea78b81fd8..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/ScanRecords.html +++ /dev/null @@ -1,445 +0,0 @@ - - - - - -ScanRecords (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class ScanRecords

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.table.scanner.log.ScanRecords
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      - -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          ScanRecords

          -
          public ScanRecords​(Map<org.apache.fluss.metadata.TableBucket,​List<ScanRecord>> records)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          records

          -
          public List<ScanRecord> records​(org.apache.fluss.metadata.TableBucket scanBucket)
          -
          Get just the records for the given bucketId.
          -
          -
          Parameters:
          -
          scanBucket - The bucket to get records for
          -
          -
        • -
        - - - -
          -
        • -

          buckets

          -
          public Set<org.apache.fluss.metadata.TableBucket> buckets()
          -
          Get the bucket ids which have records contained in this record set.
          -
          -
          Returns:
          -
          the set of partitions with data in this record set (maybe empty if no data was - returned)
          -
          -
        • -
        - - - -
          -
        • -

          count

          -
          public int count()
          -
          The number of records for all buckets.
          -
        • -
        - - - -
          -
        • -

          isEmpty

          -
          public boolean isEmpty()
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetchBuffer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetchBuffer.html deleted file mode 100644 index f112afd140..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetchBuffer.html +++ /dev/null @@ -1,194 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.scanner.log.LogFetchBuffer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.scanner.log.LogFetchBuffer

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetchCollector.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetchCollector.html deleted file mode 100644 index 1ff1177f3d..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetchCollector.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.scanner.log.LogFetchCollector (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.scanner.log.LogFetchCollector

    -
    -
    No usage of org.apache.fluss.client.table.scanner.log.LogFetchCollector
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetcher.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetcher.html deleted file mode 100644 index 386a2a300a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogFetcher.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.scanner.log.LogFetcher (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.scanner.log.LogFetcher

    -
    -
    No usage of org.apache.fluss.client.table.scanner.log.LogFetcher
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScan.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScan.html deleted file mode 100644 index 05a1f661fe..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScan.html +++ /dev/null @@ -1,194 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.scanner.log.LogScan (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.scanner.log.LogScan

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScanner.html deleted file mode 100644 index 687fc1c7a3..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScanner.html +++ /dev/null @@ -1,227 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.client.table.scanner.log.LogScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.client.table.scanner.log.LogScanner

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScannerImpl.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScannerImpl.html deleted file mode 100644 index 4d28490e78..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScannerImpl.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.scanner.log.LogScannerImpl (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.scanner.log.LogScannerImpl

    -
    -
    No usage of org.apache.fluss.client.table.scanner.log.LogScannerImpl
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScannerStatus.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScannerStatus.html deleted file mode 100644 index 7bc84832b0..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/LogScannerStatus.html +++ /dev/null @@ -1,203 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.scanner.log.LogScannerStatus (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.scanner.log.LogScannerStatus

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/RemoteLogDownloadFuture.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/RemoteLogDownloadFuture.html deleted file mode 100644 index 0fb9637dbc..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/RemoteLogDownloadFuture.html +++ /dev/null @@ -1,195 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.scanner.log.RemoteLogDownloadFuture (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.scanner.log.RemoteLogDownloadFuture

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/RemoteLogDownloader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/RemoteLogDownloader.html deleted file mode 100644 index 7edfade352..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/RemoteLogDownloader.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.scanner.log.RemoteLogDownloader (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.scanner.log.RemoteLogDownloader

    -
    -
    No usage of org.apache.fluss.client.table.scanner.log.RemoteLogDownloader
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/ScanRecords.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/ScanRecords.html deleted file mode 100644 index 1df5840ed6..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/class-use/ScanRecords.html +++ /dev/null @@ -1,214 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.scanner.log.ScanRecords (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.scanner.log.ScanRecords

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-summary.html deleted file mode 100644 index ae4e7c6a9b..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-summary.html +++ /dev/null @@ -1,235 +0,0 @@ - - - - - -org.apache.fluss.client.table.scanner.log (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.client.table.scanner.log

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-tree.html deleted file mode 100644 index ca0798d3f8..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-tree.html +++ /dev/null @@ -1,181 +0,0 @@ - - - - - -org.apache.fluss.client.table.scanner.log Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.client.table.scanner.log

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -

    Interface Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-use.html deleted file mode 100644 index b00018f3f1..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/log/package-use.html +++ /dev/null @@ -1,239 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.client.table.scanner.log (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.client.table.scanner.log

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-summary.html deleted file mode 100644 index 14fee667ce..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-summary.html +++ /dev/null @@ -1,198 +0,0 @@ - - - - - -org.apache.fluss.client.table.scanner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.client.table.scanner

    -
    -
    -
      -
    • - - - - - - - - - - - - -
      Interface Summary 
      InterfaceDescription
      Scan -
      Used to configure and create a scanner to scan data for a table.
      -
      -
    • -
    • - - - - - - - - - - - - - - - - - - - - -
      Class Summary 
      ClassDescription
      RemoteFileDownloader -
      The downloader that has a IO thread pool to download the remote files (like kv snapshots files, - log segment files).
      -
      ScanRecord -
      one scan record.
      -
      TableScan -
      API for configuring and creating LogScanner and BatchScanner.
      -
      -
    • -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-tree.html deleted file mode 100644 index 950aedd9f5..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-tree.html +++ /dev/null @@ -1,171 +0,0 @@ - - - - - -org.apache.fluss.client.table.scanner Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.client.table.scanner

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    -
      -
    • java.lang.Object -
        -
      • org.apache.fluss.client.table.scanner.RemoteFileDownloader (implements java.io.Closeable)
      • -
      • org.apache.fluss.client.table.scanner.ScanRecord (implements org.apache.fluss.record.LogRecord)
      • -
      • org.apache.fluss.client.table.scanner.TableScan (implements org.apache.fluss.client.table.scanner.Scan)
      • -
      -
    • -
    -
    -
    -

    Interface Hierarchy

    -
      -
    • org.apache.fluss.client.table.scanner.Scan
    • -
    -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-use.html deleted file mode 100644 index 208e6182c9..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/scanner/package-use.html +++ /dev/null @@ -1,332 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.client.table.scanner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.client.table.scanner

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AbstractTableWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AbstractTableWriter.html deleted file mode 100644 index ca23cb24b7..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AbstractTableWriter.html +++ /dev/null @@ -1,443 +0,0 @@ - - - - - -AbstractTableWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class AbstractTableWriter

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.table.writer.AbstractTableWriter
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - -
          -
        • -

          tablePath

          -
          protected final org.apache.fluss.metadata.TablePath tablePath
          -
        • -
        - - - -
          -
        • -

          writerClient

          -
          protected final WriterClient writerClient
          -
        • -
        - - - -
          -
        • -

          fieldCount

          -
          protected final int fieldCount
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          AbstractTableWriter

          -
          protected AbstractTableWriter​(org.apache.fluss.metadata.TablePath tablePath,
          -                              org.apache.fluss.metadata.TableInfo tableInfo,
          -                              WriterClient writerClient)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          flush

          -
          public void flush()
          -
          Flush data written that have not yet been sent to the server, forcing the client to send the - requests to server and blocks on the completion of the requests associated with these - records. A request is considered completed when it is successfully acknowledged according to - the ConfigOptions.CLIENT_WRITER_ACKS configuration you have specified or else it - results in an error.
          -
          -
          Specified by:
          -
          flush in interface TableWriter
          -
          -
        • -
        - - - - - - - -
          -
        • -

          getPhysicalPath

          -
          protected org.apache.fluss.metadata.PhysicalTablePath getPhysicalPath​(org.apache.fluss.row.InternalRow row)
          -
        • -
        - - - -
          -
        • -

          checkFieldCount

          -
          protected void checkFieldCount​(org.apache.fluss.row.InternalRow row)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/Append.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/Append.html deleted file mode 100644 index 06f795ca1e..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/Append.html +++ /dev/null @@ -1,265 +0,0 @@ - - - - - -Append (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface Append

    -
    -
    -
    -
      -
    • -
      -
      All Known Implementing Classes:
      -
      TableAppend
      -
      -
      -
      @PublicEvolving
      -public interface Append
      -
      Used to configure and create a AppendWriter to write data to a Log Table. - -

      Append objects are immutable and can be shared between threads.

      -
      -
      Since:
      -
      0.6
      -
      -
    • -
    -
    -
    - -
    -
    - -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AppendResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AppendResult.html deleted file mode 100644 index 79f99a359e..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AppendResult.html +++ /dev/null @@ -1,274 +0,0 @@ - - - - - -AppendResult (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class AppendResult

    -
    -
    - -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          AppendResult

          -
          public AppendResult()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AppendWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AppendWriter.html deleted file mode 100644 index 269dcbba63..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/AppendWriter.html +++ /dev/null @@ -1,277 +0,0 @@ - - - - - -AppendWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface AppendWriter

    -
    -
    -
    -
      -
    • -
      -
      All Superinterfaces:
      -
      TableWriter
      -
      -
      -
      @PublicEvolving
      -public interface AppendWriter
      -extends TableWriter
      -
      The writer to write data to the log table.
      -
      -
      Since:
      -
      0.2
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          append

          -
          CompletableFuture<AppendResult> append​(org.apache.fluss.row.InternalRow row)
          -
          Append row into a Log Table.
          -
          -
          Parameters:
          -
          row - the row to append.
          -
          Returns:
          -
          A CompletableFuture that always returns append result when complete normally.
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/DeleteResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/DeleteResult.html deleted file mode 100644 index 5279c4e11d..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/DeleteResult.html +++ /dev/null @@ -1,274 +0,0 @@ - - - - - -DeleteResult (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class DeleteResult

    -
    -
    - -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          DeleteResult

          -
          public DeleteResult()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableAppend.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableAppend.html deleted file mode 100644 index 0eb32d06da..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableAppend.html +++ /dev/null @@ -1,324 +0,0 @@ - - - - - -TableAppend (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TableAppend

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      Append
      -
      -
      -
      public class TableAppend
      -extends Object
      -implements Append
      -
      API for configuring and creating AppendWriter.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          TableAppend

          -
          public TableAppend​(org.apache.fluss.metadata.TablePath tablePath,
          -                   org.apache.fluss.metadata.TableInfo tableInfo,
          -                   WriterClient writerClient)
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableUpsert.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableUpsert.html deleted file mode 100644 index 5e424bdbd4..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableUpsert.html +++ /dev/null @@ -1,384 +0,0 @@ - - - - - -TableUpsert (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TableUpsert

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      Upsert
      -
      -
      -
      public class TableUpsert
      -extends Object
      -implements Upsert
      -
      API for configuring and creating UpsertWriter.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          TableUpsert

          -
          public TableUpsert​(org.apache.fluss.metadata.TablePath tablePath,
          -                   org.apache.fluss.metadata.TableInfo tableInfo,
          -                   WriterClient writerClient)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          partialUpdate

          -
          public Upsert partialUpdate​(@Nullable
          -                            int[] targetColumns)
          -
          Description copied from interface: Upsert
          -
          Apply partial update columns and returns a new Upsert instance. - -

          For UpsertWriter.upsert(InternalRow) operation, only the specified columns will be - updated and other columns will remain unchanged if the row exists or set to null if the row - doesn't exist. - -

          For UpsertWriter.delete(InternalRow) operation, the entire row will not be - removed, but only the specified columns except primary key will be set to null. The entire - row will be removed when all columns except primary key are null after a UpsertWriter.delete(InternalRow) operation. - -

          Note: The specified columns must be a contains all columns of primary key, and all columns - except primary key should be nullable.

          -
          -
          Specified by:
          -
          partialUpdate in interface Upsert
          -
          Parameters:
          -
          targetColumns - the column indexes to partial update
          -
          -
        • -
        - - - -
          -
        • -

          partialUpdate

          -
          public Upsert partialUpdate​(String... targetColumnNames)
          -
          -
          Specified by:
          -
          partialUpdate in interface Upsert
          -
          Parameters:
          -
          targetColumnNames - the column names to partial update
          -
          See Also:
          -
          for more details.
          -
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableWriter.html deleted file mode 100644 index 6bd9d9f062..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/TableWriter.html +++ /dev/null @@ -1,275 +0,0 @@ - - - - - -TableWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface TableWriter

    -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Summary

        - - - - - - - - - - - - -
        All Methods Instance Methods Abstract Methods 
        Modifier and TypeMethodDescription
        voidflush() -
        Flush data written that have not yet been sent to the server, forcing the client to send the - requests to server and blocks on the completion of the requests associated with these - records.
        -
        -
      • -
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          flush

          -
          void flush()
          -
          Flush data written that have not yet been sent to the server, forcing the client to send the - requests to server and blocks on the completion of the requests associated with these - records. A request is considered completed when it is successfully acknowledged according to - the ConfigOptions.CLIENT_WRITER_ACKS configuration you have specified or else it - results in an error.
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/Upsert.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/Upsert.html deleted file mode 100644 index 4b46328362..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/Upsert.html +++ /dev/null @@ -1,322 +0,0 @@ - - - - - -Upsert (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface Upsert

    -
    -
    -
    -
      -
    • -
      -
      All Known Implementing Classes:
      -
      TableUpsert
      -
      -
      -
      @PublicEvolving
      -public interface Upsert
      -
      Used to configure and create UpsertWriter to upsert and delete data to a Primary Key - Table. - -

      Upsert objects are immutable and can be shared between threads. Refinement methods, - like partialUpdate(int[]), create new Upsert instances.

      -
      -
      Since:
      -
      0.6
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          partialUpdate

          -
          Upsert partialUpdate​(@Nullable
          -                     int[] targetColumns)
          -
          Apply partial update columns and returns a new Upsert instance. - -

          For UpsertWriter.upsert(InternalRow) operation, only the specified columns will be - updated and other columns will remain unchanged if the row exists or set to null if the row - doesn't exist. - -

          For UpsertWriter.delete(InternalRow) operation, the entire row will not be - removed, but only the specified columns except primary key will be set to null. The entire - row will be removed when all columns except primary key are null after a UpsertWriter.delete(InternalRow) operation. - -

          Note: The specified columns must be a contains all columns of primary key, and all columns - except primary key should be nullable.

          -
          -
          Parameters:
          -
          targetColumns - the column indexes to partial update
          -
          -
        • -
        - - - -
          -
        • -

          partialUpdate

          -
          Upsert partialUpdate​(String... targetColumnNames)
          -
          -
          Parameters:
          -
          targetColumnNames - the column names to partial update
          -
          See Also:
          -
          for more details.
          -
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertResult.html deleted file mode 100644 index 4666920ba2..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertResult.html +++ /dev/null @@ -1,274 +0,0 @@ - - - - - -UpsertResult (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class UpsertResult

    -
    -
    - -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          UpsertResult

          -
          public UpsertResult()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertWrite.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertWrite.html deleted file mode 100644 index 26f95ac8f9..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertWrite.html +++ /dev/null @@ -1,352 +0,0 @@ - - - - - -UpsertWrite (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class UpsertWrite

    -
    -
    - -
    -
      -
    • -
      -
      @PublicEvolving
      -public class UpsertWrite
      -extends Object
      -
      Used to describe the operation to write data by UpsertWriter to a table.
      -
      -
      Since:
      -
      0.2
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          UpsertWrite

          -
          public UpsertWrite()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          withPartialUpdate

          -
          public UpsertWrite withPartialUpdate​(int[] targetColumns)
          -
          Returns a new instance of UpsertWrite description with partial update the specified columns. - -

          For UpsertWriter.upsert(InternalRow) operation, only the specified columns will be - updated and other columns will remain unchanged if the row exists or set to null if the row - doesn't exist. - -

          For UpsertWriter.delete(InternalRow) operation, the entire row will not be - removed, but only the specified columns except primary key will be set to null. The entire - row will be removed when all columns except primary key are null after a UpsertWriter.delete(InternalRow) operation. - -

          Note: The specified columns must be a contains all columns of primary key, and all columns - except primary key should be nullable.

          -
          -
          Parameters:
          -
          targetColumns - the columns to partial update,
          -
          -
        • -
        - - - -
          -
        • -

          getPartialUpdateColumns

          -
          @Nullable
          -public int[] getPartialUpdateColumns()
          -
          Returns the columns to partial update. Returns null if update all columns.
          -
          -
          Returns:
          -
          the columns to partial update.
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertWriter.html deleted file mode 100644 index 572b271623..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/UpsertWriter.html +++ /dev/null @@ -1,302 +0,0 @@ - - - - - -UpsertWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface UpsertWriter

    -
    -
    -
    -
      -
    • -
      -
      All Superinterfaces:
      -
      TableWriter
      -
      -
      -
      @PublicEvolving
      -public interface UpsertWriter
      -extends TableWriter
      -
      The writer to write data to the primary key table.
      -
      -
      Since:
      -
      0.2
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          upsert

          -
          CompletableFuture<UpsertResult> upsert​(org.apache.fluss.row.InternalRow row)
          -
          Inserts row into Fluss table if they do not already exist, or updates them if they do exist.
          -
          -
          Parameters:
          -
          row - the row to upsert.
          -
          Returns:
          -
          A CompletableFuture that always returns upsert result when complete normally.
          -
          -
        • -
        - - - -
          -
        • -

          delete

          -
          CompletableFuture<DeleteResult> delete​(org.apache.fluss.row.InternalRow row)
          -
          Delete certain row by the input row in Fluss table, the input row must contain the primary - key.
          -
          -
          Parameters:
          -
          row - the row to delete.
          -
          Returns:
          -
          A CompletableFuture that always delete result when complete normally.
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AbstractTableWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AbstractTableWriter.html deleted file mode 100644 index 73ea9aa792..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AbstractTableWriter.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.writer.AbstractTableWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.writer.AbstractTableWriter

    -
    -
    No usage of org.apache.fluss.client.table.writer.AbstractTableWriter
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/Append.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/Append.html deleted file mode 100644 index 085bff6a69..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/Append.html +++ /dev/null @@ -1,228 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.client.table.writer.Append (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.client.table.writer.Append

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AppendResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AppendResult.html deleted file mode 100644 index 723de1052b..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AppendResult.html +++ /dev/null @@ -1,194 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.writer.AppendResult (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.writer.AppendResult

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AppendWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AppendWriter.html deleted file mode 100644 index 06c8fbae39..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/AppendWriter.html +++ /dev/null @@ -1,199 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.client.table.writer.AppendWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.client.table.writer.AppendWriter

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/DeleteResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/DeleteResult.html deleted file mode 100644 index b4caa93de4..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/DeleteResult.html +++ /dev/null @@ -1,195 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.writer.DeleteResult (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.writer.DeleteResult

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableAppend.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableAppend.html deleted file mode 100644 index 444502f386..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableAppend.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.writer.TableAppend (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.writer.TableAppend

    -
    -
    No usage of org.apache.fluss.client.table.writer.TableAppend
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableUpsert.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableUpsert.html deleted file mode 100644 index 105586ccef..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableUpsert.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.writer.TableUpsert (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.writer.TableUpsert

    -
    -
    No usage of org.apache.fluss.client.table.writer.TableUpsert
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableWriter.html deleted file mode 100644 index b359c9b2dc..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/TableWriter.html +++ /dev/null @@ -1,218 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.client.table.writer.TableWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.client.table.writer.TableWriter

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/Upsert.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/Upsert.html deleted file mode 100644 index bb4715456e..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/Upsert.html +++ /dev/null @@ -1,260 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.client.table.writer.Upsert (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.client.table.writer.Upsert

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertResult.html deleted file mode 100644 index 0a4ccc2776..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertResult.html +++ /dev/null @@ -1,194 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.writer.UpsertResult (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.writer.UpsertResult

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertWrite.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertWrite.html deleted file mode 100644 index 77198995ee..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertWrite.html +++ /dev/null @@ -1,194 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.table.writer.UpsertWrite (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.table.writer.UpsertWrite

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertWriter.html deleted file mode 100644 index dd17066ba6..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/class-use/UpsertWriter.html +++ /dev/null @@ -1,200 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.client.table.writer.UpsertWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.client.table.writer.UpsertWriter

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-summary.html deleted file mode 100644 index ffa217be8d..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-summary.html +++ /dev/null @@ -1,246 +0,0 @@ - - - - - -org.apache.fluss.client.table.writer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.client.table.writer

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-tree.html deleted file mode 100644 index e1bb348e82..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-tree.html +++ /dev/null @@ -1,182 +0,0 @@ - - - - - -org.apache.fluss.client.table.writer Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.client.table.writer

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -

    Interface Hierarchy

    -
      -
    • org.apache.fluss.client.table.writer.Append
    • -
    • org.apache.fluss.client.table.writer.TableWriter - -
    • -
    • org.apache.fluss.client.table.writer.Upsert
    • -
    -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-use.html deleted file mode 100644 index 8cc4618d20..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/table/writer/package-use.html +++ /dev/null @@ -1,264 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.client.table.writer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.client.table.writer

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/DefaultSecurityTokenManager.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/DefaultSecurityTokenManager.html deleted file mode 100644 index d7b3a2302f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/DefaultSecurityTokenManager.html +++ /dev/null @@ -1,360 +0,0 @@ - - - - - -DefaultSecurityTokenManager (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class DefaultSecurityTokenManager

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.token.DefaultSecurityTokenManager
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          DefaultSecurityTokenManager

          -
          public DefaultSecurityTokenManager​(org.apache.fluss.config.Configuration configuration,
          -                                   SecurityTokenProvider securityTokenProvider)
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/DefaultSecurityTokenProvider.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/DefaultSecurityTokenProvider.html deleted file mode 100644 index d8d96b7f18..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/DefaultSecurityTokenProvider.html +++ /dev/null @@ -1,325 +0,0 @@ - - - - - -DefaultSecurityTokenProvider (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class DefaultSecurityTokenProvider

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.token.DefaultSecurityTokenProvider
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          DefaultSecurityTokenProvider

          -
          public DefaultSecurityTokenProvider​(org.apache.fluss.rpc.gateway.AdminReadOnlyGateway adminReadOnlyGateway)
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/SecurityTokenManager.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/SecurityTokenManager.html deleted file mode 100644 index cf24d051cf..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/SecurityTokenManager.html +++ /dev/null @@ -1,280 +0,0 @@ - - - - - -SecurityTokenManager (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface SecurityTokenManager

    -
    -
    -
    -
      -
    • -
      -
      All Known Implementing Classes:
      -
      DefaultSecurityTokenManager
      -
      -
      -
      public interface SecurityTokenManager
      -
      API for security token manager.
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Summary

        - - - - - - - - - - - - - - - - - -
        All Methods Instance Methods Abstract Methods 
        Modifier and TypeMethodDescription
        voidstart() -
        Creates a re-occurring task which obtains new tokens.
        -
        voidstop() -
        Stops re-occurring token obtain task.
        -
        -
      • -
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          start

          -
          void start()
          -    throws Exception
          -
          Creates a re-occurring task which obtains new tokens.
          -
          -
          Throws:
          -
          Exception
          -
          -
        • -
        - - - -
          -
        • -

          stop

          -
          void stop()
          -
          Stops re-occurring token obtain task.
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/SecurityTokenProvider.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/SecurityTokenProvider.html deleted file mode 100644 index 949327705f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/SecurityTokenProvider.html +++ /dev/null @@ -1,265 +0,0 @@ - - - - - -SecurityTokenProvider (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface SecurityTokenProvider

    -
    -
    -
    -
      -
    • -
      -
      All Known Implementing Classes:
      -
      DefaultSecurityTokenProvider
      -
      -
      -
      public interface SecurityTokenProvider
      -
      Security token provider API.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          obtainSecurityToken

          -
          org.apache.fluss.fs.token.ObtainedSecurityToken obtainSecurityToken()
          -                                                             throws Exception
          -
          Obtain security token.
          -
          -
          Returns:
          -
          the obtained security token.
          -
          Throws:
          -
          Exception
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/DefaultSecurityTokenManager.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/DefaultSecurityTokenManager.html deleted file mode 100644 index 5fee791735..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/DefaultSecurityTokenManager.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.token.DefaultSecurityTokenManager (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.token.DefaultSecurityTokenManager

    -
    -
    No usage of org.apache.fluss.client.token.DefaultSecurityTokenManager
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/DefaultSecurityTokenProvider.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/DefaultSecurityTokenProvider.html deleted file mode 100644 index 616cb1a45a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/DefaultSecurityTokenProvider.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.token.DefaultSecurityTokenProvider (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.token.DefaultSecurityTokenProvider

    -
    -
    No usage of org.apache.fluss.client.token.DefaultSecurityTokenProvider
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/SecurityTokenManager.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/SecurityTokenManager.html deleted file mode 100644 index 43e5a197aa..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/SecurityTokenManager.html +++ /dev/null @@ -1,194 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.client.token.SecurityTokenManager (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.client.token.SecurityTokenManager

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/SecurityTokenProvider.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/SecurityTokenProvider.html deleted file mode 100644 index f44ca74b92..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/class-use/SecurityTokenProvider.html +++ /dev/null @@ -1,208 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.client.token.SecurityTokenProvider (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.client.token.SecurityTokenProvider

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-summary.html deleted file mode 100644 index 1595497445..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-summary.html +++ /dev/null @@ -1,197 +0,0 @@ - - - - - -org.apache.fluss.client.token (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.client.token

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-tree.html deleted file mode 100644 index 44327a8226..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-tree.html +++ /dev/null @@ -1,171 +0,0 @@ - - - - - -org.apache.fluss.client.token Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.client.token

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -

    Interface Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-use.html deleted file mode 100644 index a5c0a11515..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/token/package-use.html +++ /dev/null @@ -1,191 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.client.token (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.client.token

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/ClientRpcMessageUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/ClientRpcMessageUtils.html deleted file mode 100644 index 3198c1c8ad..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/ClientRpcMessageUtils.html +++ /dev/null @@ -1,556 +0,0 @@ - - - - - -ClientRpcMessageUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class ClientRpcMessageUtils

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.utils.ClientRpcMessageUtils
      • -
      -
    • -
    -
    -
      -
    • -
      -
      public class ClientRpcMessageUtils
      -extends Object
      -
      Utils for making rpc request/response from inner object or convert inner class to rpc - request/response for client.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          ClientRpcMessageUtils

          -
          public ClientRpcMessageUtils()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          makeProduceLogRequest

          -
          public static org.apache.fluss.rpc.messages.ProduceLogRequest makeProduceLogRequest​(long tableId,
          -                                                                                    int acks,
          -                                                                                    int maxRequestTimeoutMs,
          -                                                                                    List<ReadyWriteBatch> readyBatches)
          -
        • -
        - - - -
          -
        • -

          makePutKvRequest

          -
          public static org.apache.fluss.rpc.messages.PutKvRequest makePutKvRequest​(long tableId,
          -                                                                          int acks,
          -                                                                          int maxRequestTimeoutMs,
          -                                                                          List<ReadyWriteBatch> readyWriteBatches)
          -
        • -
        - - - -
          -
        • -

          makeLookupRequest

          -
          public static org.apache.fluss.rpc.messages.LookupRequest makeLookupRequest​(long tableId,
          -                                                                            Collection<LookupBatch> lookupBatches)
          -
        • -
        - - - -
          -
        • -

          makePrefixLookupRequest

          -
          public static org.apache.fluss.rpc.messages.PrefixLookupRequest makePrefixLookupRequest​(long tableId,
          -                                                                                        Collection<PrefixLookupBatch> lookupBatches)
          -
        • -
        - - - -
          -
        • -

          toKvSnapshots

          -
          public static KvSnapshots toKvSnapshots​(org.apache.fluss.rpc.messages.GetLatestKvSnapshotsResponse response)
          -
        • -
        - - - -
          -
        • -

          toKvSnapshotMetadata

          -
          public static KvSnapshotMetadata toKvSnapshotMetadata​(org.apache.fluss.rpc.messages.GetKvSnapshotMetadataResponse response)
          -
        • -
        - - - -
          -
        • -

          toLakeTableSnapshotInfo

          -
          public static LakeSnapshot toLakeTableSnapshotInfo​(org.apache.fluss.rpc.messages.GetLatestLakeSnapshotResponse response)
          -
        • -
        - - - -
          -
        • -

          toFsPathAndFileName

          -
          public static List<org.apache.fluss.fs.FsPathAndFileName> toFsPathAndFileName​(List<org.apache.fluss.rpc.messages.PbRemotePathAndLocalFile> pbFileHandles)
          -
        • -
        - - - -
          -
        • -

          toSecurityToken

          -
          public static org.apache.fluss.fs.token.ObtainedSecurityToken toSecurityToken​(org.apache.fluss.rpc.messages.GetFileSystemSecurityTokenResponse response)
          -
        • -
        - - - -
          -
        • -

          makeMetadataRequest

          -
          public static org.apache.fluss.rpc.messages.MetadataRequest makeMetadataRequest​(@Nullable
          -                                                                                Set<org.apache.fluss.metadata.TablePath> tablePaths,
          -                                                                                @Nullable
          -                                                                                Collection<org.apache.fluss.metadata.PhysicalTablePath> tablePathPartitionNames,
          -                                                                                @Nullable
          -                                                                                Collection<Long> tablePathPartitionIds)
          -
        • -
        - - - -
          -
        • -

          makeListOffsetsRequest

          -
          public static org.apache.fluss.rpc.messages.ListOffsetsRequest makeListOffsetsRequest​(long tableId,
          -                                                                                      @Nullable
          -                                                                                      Long partitionId,
          -                                                                                      List<Integer> bucketIdList,
          -                                                                                      OffsetSpec offsetSpec)
          -
        • -
        - - - -
          -
        • -

          makeCreatePartitionRequest

          -
          public static org.apache.fluss.rpc.messages.CreatePartitionRequest makeCreatePartitionRequest​(org.apache.fluss.metadata.TablePath tablePath,
          -                                                                                              org.apache.fluss.metadata.PartitionSpec partitionSpec,
          -                                                                                              boolean ignoreIfNotExists)
          -
        • -
        - - - -
          -
        • -

          makeDropPartitionRequest

          -
          public static org.apache.fluss.rpc.messages.DropPartitionRequest makeDropPartitionRequest​(org.apache.fluss.metadata.TablePath tablePath,
          -                                                                                          org.apache.fluss.metadata.PartitionSpec partitionSpec,
          -                                                                                          boolean ignoreIfNotExists)
          -
        • -
        - - - -
          -
        • -

          toPartitionInfos

          -
          public static List<org.apache.fluss.metadata.PartitionInfo> toPartitionInfos​(org.apache.fluss.rpc.messages.ListPartitionInfosResponse response)
          -
        • -
        - - - -
          -
        • -

          toKeyValueMap

          -
          public static Map<String,​String> toKeyValueMap​(List<org.apache.fluss.rpc.messages.PbKeyValue> pbKeyValues)
          -
        • -
        - - - -
          -
        • -

          makePbPartitionSpec

          -
          public static org.apache.fluss.rpc.messages.PbPartitionSpec makePbPartitionSpec​(org.apache.fluss.metadata.PartitionSpec partitionSpec)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/ClientUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/ClientUtils.html deleted file mode 100644 index ed680cdd9e..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/ClientUtils.html +++ /dev/null @@ -1,343 +0,0 @@ - - - - - -ClientUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class ClientUtils

    -
    -
    - -
    -
      -
    • -
      -
      public final class ClientUtils
      -extends Object
      -
      Utils for Fluss Client.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - - - - - -
          -
        • -

          getHost

          -
          public static String getHost​(String address)
          -
          Extracts the hostname from a "host:port" address string.
          -
          -
          Parameters:
          -
          address - address string to parse
          -
          Returns:
          -
          hostname or null if the given address is incorrect
          -
          -
        • -
        - - - -
          -
        • -

          getPort

          -
          public static Integer getPort​(String address)
          -
          Extracts the port number from a "host:port" address string.
          -
          -
          Parameters:
          -
          address - address string to parse
          -
          Returns:
          -
          port number or null if the given address is incorrect
          -
          -
        • -
        - - - -
          -
        • -

          getPartitionId

          -
          public static Long getPartitionId​(org.apache.fluss.row.InternalRow row,
          -                                  PartitionGetter partitionGetter,
          -                                  org.apache.fluss.metadata.TablePath tablePath,
          -                                  MetadataUpdater metadataUpdater)
          -                           throws org.apache.fluss.exception.PartitionNotExistException
          -
          Return the id of the partition the row belongs to. It'll try to update the metadata if the - partition doesn't exist. If the partition doesn't exist yet after update metadata, it'll - throw PartitionNotExistException.
          -
          -
          Throws:
          -
          org.apache.fluss.exception.PartitionNotExistException
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/MetadataUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/MetadataUtils.html deleted file mode 100644 index 66feafba5f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/MetadataUtils.html +++ /dev/null @@ -1,420 +0,0 @@ - - - - - -MetadataUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class MetadataUtils

    -
    -
    - -
    -
      -
    • -
      -
      public class MetadataUtils
      -extends Object
      -
      Utils for metadata for client.
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Summary

        - - - - - - - - - - -
        Constructors 
        ConstructorDescription
        MetadataUtils() 
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Summary

        - - - - - - - - - - - - - - - - - - - - - - - - - - - -
        All Methods Static Methods Concrete Methods 
        Modifier and TypeMethodDescription
        static org.apache.fluss.cluster.ServerNodegetOneAvailableTabletServerNode​(org.apache.fluss.cluster.Cluster cluster) 
        static org.apache.fluss.cluster.ClustersendMetadataRequestAndRebuildCluster​(org.apache.fluss.cluster.Cluster cluster, - org.apache.fluss.rpc.RpcClient client, - Set<org.apache.fluss.metadata.TablePath> tablePaths, - Collection<org.apache.fluss.metadata.PhysicalTablePath> tablePartitionNames, - Collection<Long> tablePartitionIds) -
        Partial update cluster, means we will rebuild the cluster by sending metadata request to - request the input tables/partitions in physicalTablePaths, after that add those - tables/partitions into cluster.
        -
        static org.apache.fluss.cluster.ClustersendMetadataRequestAndRebuildCluster​(org.apache.fluss.rpc.gateway.AdminReadOnlyGateway gateway, - boolean partialUpdate, - org.apache.fluss.cluster.Cluster originCluster, - Set<org.apache.fluss.metadata.TablePath> tablePaths, - Collection<org.apache.fluss.metadata.PhysicalTablePath> tablePartitions, - Collection<Long> tablePartitionIds) -
        maybe partial update cluster.
        -
        static org.apache.fluss.cluster.ClustersendMetadataRequestAndRebuildCluster​(org.apache.fluss.rpc.gateway.AdminReadOnlyGateway gateway, - Set<org.apache.fluss.metadata.TablePath> tablePaths) -
        full update cluster, means we will rebuild the cluster by clearing all cached table in - cluster, and then send metadata request to request the input tables in tablePaths, after that - add those table into cluster.
        -
        - -
      • -
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          MetadataUtils

          -
          public MetadataUtils()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          sendMetadataRequestAndRebuildCluster

          -
          public static org.apache.fluss.cluster.Cluster sendMetadataRequestAndRebuildCluster​(org.apache.fluss.rpc.gateway.AdminReadOnlyGateway gateway,
          -                                                                                    Set<org.apache.fluss.metadata.TablePath> tablePaths)
          -                                                                             throws ExecutionException,
          -                                                                                    InterruptedException,
          -                                                                                    TimeoutException
          -
          full update cluster, means we will rebuild the cluster by clearing all cached table in - cluster, and then send metadata request to request the input tables in tablePaths, after that - add those table into cluster.
          -
          -
          Throws:
          -
          ExecutionException
          -
          InterruptedException
          -
          TimeoutException
          -
          -
        • -
        - - - -
          -
        • -

          sendMetadataRequestAndRebuildCluster

          -
          public static org.apache.fluss.cluster.Cluster sendMetadataRequestAndRebuildCluster​(org.apache.fluss.cluster.Cluster cluster,
          -                                                                                    org.apache.fluss.rpc.RpcClient client,
          -                                                                                    @Nullable
          -                                                                                    Set<org.apache.fluss.metadata.TablePath> tablePaths,
          -                                                                                    @Nullable
          -                                                                                    Collection<org.apache.fluss.metadata.PhysicalTablePath> tablePartitionNames,
          -                                                                                    @Nullable
          -                                                                                    Collection<Long> tablePartitionIds)
          -                                                                             throws ExecutionException,
          -                                                                                    InterruptedException,
          -                                                                                    TimeoutException
          -
          Partial update cluster, means we will rebuild the cluster by sending metadata request to - request the input tables/partitions in physicalTablePaths, after that add those - tables/partitions into cluster. The origin tables/partitions in cluster will not be cleared, - but will be updated.
          -
          -
          Throws:
          -
          ExecutionException
          -
          InterruptedException
          -
          TimeoutException
          -
          -
        • -
        - - - - - - - -
          -
        • -

          getOneAvailableTabletServerNode

          -
          public static org.apache.fluss.cluster.ServerNode getOneAvailableTabletServerNode​(org.apache.fluss.cluster.Cluster cluster)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/ClientRpcMessageUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/ClientRpcMessageUtils.html deleted file mode 100644 index bb1d01f261..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/ClientRpcMessageUtils.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.utils.ClientRpcMessageUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.utils.ClientRpcMessageUtils

    -
    -
    No usage of org.apache.fluss.client.utils.ClientRpcMessageUtils
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/ClientUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/ClientUtils.html deleted file mode 100644 index f13ca48e77..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/ClientUtils.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.utils.ClientUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.utils.ClientUtils

    -
    -
    No usage of org.apache.fluss.client.utils.ClientUtils
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/MetadataUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/MetadataUtils.html deleted file mode 100644 index 4351d0c34b..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/class-use/MetadataUtils.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.utils.MetadataUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.utils.MetadataUtils

    -
    -
    No usage of org.apache.fluss.client.utils.MetadataUtils
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-summary.html deleted file mode 100644 index 263757778c..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-summary.html +++ /dev/null @@ -1,181 +0,0 @@ - - - - - -org.apache.fluss.client.utils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.client.utils

    -
    -
    -
      -
    • - - - - - - - - - - - - - - - - - - - - -
      Class Summary 
      ClassDescription
      ClientRpcMessageUtils -
      Utils for making rpc request/response from inner object or convert inner class to rpc - request/response for client.
      -
      ClientUtils -
      Utils for Fluss Client.
      -
      MetadataUtils -
      Utils for metadata for client.
      -
      -
    • -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-tree.html deleted file mode 100644 index 3065645f24..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-tree.html +++ /dev/null @@ -1,165 +0,0 @@ - - - - - -org.apache.fluss.client.utils Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.client.utils

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-use.html deleted file mode 100644 index e62e4f11c1..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/utils/package-use.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.client.utils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.client.utils

    -
    -
    No usage of org.apache.fluss.client.utils
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/ArrowLogWriteBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/ArrowLogWriteBatch.html deleted file mode 100644 index b977cc6057..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/ArrowLogWriteBatch.html +++ /dev/null @@ -1,597 +0,0 @@ - - - - - -ArrowLogWriteBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class ArrowLogWriteBatch

    -
    -
    - -
    -
      -
    • -
      -
      @NotThreadSafe
      -@Internal
      -public class ArrowLogWriteBatch
      -extends WriteBatch
      -
      A batch of log records managed in ARROW format that is or will be sent to server by ProduceLogRequest. - -

      This class is not thread safe and external synchronization must be used when modifying it.

      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          ArrowLogWriteBatch

          -
          public ArrowLogWriteBatch​(int bucketId,
          -                          org.apache.fluss.metadata.PhysicalTablePath physicalTablePath,
          -                          int schemaId,
          -                          org.apache.fluss.row.arrow.ArrowWriter arrowWriter,
          -                          org.apache.fluss.memory.AbstractPagedOutputView outputView,
          -                          long createdMs)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          tryAppend

          -
          public boolean tryAppend​(WriteRecord writeRecord,
          -                         WriteCallback callback)
          -                  throws Exception
          -
          Description copied from class: WriteBatch
          -
          try to append one write record to the record batch.
          -
          -
          Specified by:
          -
          tryAppend in class WriteBatch
          -
          Parameters:
          -
          writeRecord - the record to write
          -
          callback - the callback to send back to writer
          -
          Returns:
          -
          true if append success, false if the batch is full.
          -
          Throws:
          -
          Exception
          -
          -
        • -
        - - - -
          -
        • -

          build

          -
          public org.apache.fluss.record.bytesview.BytesView build()
          -
          Description copied from class: WriteBatch
          -
          Gets the memory segment bytes view of the batch. This includes the latest updated WriteBatch.setWriterState(long, int) in the bytes view.
          -
          -
          Specified by:
          -
          build in class WriteBatch
          -
          -
        • -
        - - - - - - - -
          -
        • -

          isClosed

          -
          public boolean isClosed()
          -
          Description copied from class: WriteBatch
          -
          check if the batch is closed.
          -
          -
          Specified by:
          -
          isClosed in class WriteBatch
          -
          Returns:
          -
          true if closed, false otherwise
          -
          -
        • -
        - - - -
          -
        • -

          estimatedSizeInBytes

          -
          public int estimatedSizeInBytes()
          -
          Description copied from class: WriteBatch
          -
          Get an estimate of the number of bytes written to the underlying buffer. The returned value - is exactly correct if the record set is not compressed or if the batch has been WriteBatch.build().
          -
          -
          Specified by:
          -
          estimatedSizeInBytes in class WriteBatch
          -
          -
        • -
        - - - -
          -
        • -

          pooledMemorySegments

          -
          public List<org.apache.fluss.memory.MemorySegment> pooledMemorySegments()
          -
          Description copied from class: WriteBatch
          -
          get pooled memory segments to de-allocate. After produceLog/PutKv acks, the WriteBatch need to de-allocate the allocated pooled MemorySegments back to MemorySegmentPool for reusing.
          -
          -
          Specified by:
          -
          pooledMemorySegments in class WriteBatch
          -
          Returns:
          -
          the pooled memory segment this batch allocated
          -
          -
        • -
        - - - -
          -
        • -

          setWriterState

          -
          public void setWriterState​(long writerId,
          -                           int batchSequence)
          -
          -
          Specified by:
          -
          setWriterState in class WriteBatch
          -
          -
        • -
        - - - -
          -
        • -

          resetWriterState

          -
          public void resetWriterState​(long writerId,
          -                             int batchSequence)
          -
          -
          Overrides:
          -
          resetWriterState in class WriteBatch
          -
          -
        • -
        - - - -
          -
        • -

          writerId

          -
          public long writerId()
          -
          -
          Specified by:
          -
          writerId in class WriteBatch
          -
          -
        • -
        - - - - - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/DynamicPartitionCreator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/DynamicPartitionCreator.html deleted file mode 100644 index 6527436850..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/DynamicPartitionCreator.html +++ /dev/null @@ -1,314 +0,0 @@ - - - - - -DynamicPartitionCreator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class DynamicPartitionCreator

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.write.DynamicPartitionCreator
      • -
      -
    • -
    -
    -
      -
    • -
      -
      @ThreadSafe
      -public class DynamicPartitionCreator
      -extends Object
      -
      A creator to create partition when dynamic partition create enable for table.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          DynamicPartitionCreator

          -
          public DynamicPartitionCreator​(MetadataUpdater metadataUpdater,
          -                               Admin admin,
          -                               boolean dynamicPartitionEnabled,
          -                               Consumer<Throwable> fatalErrorHandler)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          checkAndCreatePartitionAsync

          -
          public void checkAndCreatePartitionAsync​(org.apache.fluss.metadata.PhysicalTablePath physicalTablePath)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/DynamicWriteBatchSizeEstimator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/DynamicWriteBatchSizeEstimator.html deleted file mode 100644 index aa6048863b..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/DynamicWriteBatchSizeEstimator.html +++ /dev/null @@ -1,329 +0,0 @@ - - - - - -DynamicWriteBatchSizeEstimator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class DynamicWriteBatchSizeEstimator

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.write.DynamicWriteBatchSizeEstimator
      • -
      -
    • -
    -
    -
      -
    • -
      -
      @Internal
      -@ThreadSafe
      -public class DynamicWriteBatchSizeEstimator
      -extends Object
      -
      An estimator to estimate the buffer usage of a writeBatch.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          DynamicWriteBatchSizeEstimator

          -
          public DynamicWriteBatchSizeEstimator​(boolean dynamicBatchSizeEnabled,
          -                                      int maxBatchSize,
          -                                      int pageSize)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          updateEstimation

          -
          public void updateEstimation​(org.apache.fluss.metadata.PhysicalTablePath physicalTablePath,
          -                             int observedBatchSize)
          -
        • -
        - - - -
          -
        • -

          getEstimatedBatchSize

          -
          public int getEstimatedBatchSize​(org.apache.fluss.metadata.PhysicalTablePath physicalTablePath)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/HashBucketAssigner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/HashBucketAssigner.html deleted file mode 100644 index bf7a4e19fd..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/HashBucketAssigner.html +++ /dev/null @@ -1,370 +0,0 @@ - - - - - -HashBucketAssigner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class HashBucketAssigner

    -
    -
    - -
    -
      -
    • -
      -
      @Internal
      -public class HashBucketAssigner
      -extends Object
      -
      Hash bucket assigner.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          HashBucketAssigner

          -
          public HashBucketAssigner​(int numBuckets)
          -
        • -
        - - - -
          -
        • -

          HashBucketAssigner

          -
          public HashBucketAssigner​(int numBuckets,
          -                          org.apache.fluss.bucketing.BucketingFunction function)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          assignBucket

          -
          public int assignBucket​(byte[] bucketKeys)
          -
        • -
        - - - -
          -
        • -

          assignBucket

          -
          public int assignBucket​(@Nullable
          -                        byte[] bucketKey,
          -                        org.apache.fluss.cluster.Cluster cluster)
          -
        • -
        - - - -
          -
        • -

          abortIfBatchFull

          -
          public boolean abortIfBatchFull()
          -
        • -
        - - - -
          -
        • -

          onNewBatch

          -
          public void onNewBatch​(org.apache.fluss.cluster.Cluster cluster,
          -                       int prevBucketId)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceBucketEntry.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceBucketEntry.html deleted file mode 100644 index 9b9641e3dd..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceBucketEntry.html +++ /dev/null @@ -1,224 +0,0 @@ - - - - - -IdempotenceBucketEntry (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class IdempotenceBucketEntry

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.write.IdempotenceBucketEntry
      • -
      -
    • -
    -
    -
      -
    • -
      -
      @Internal
      -public class IdempotenceBucketEntry
      -extends Object
      -
      Entry to store the idempotence information of each table-bucket.
      -
    • -
    -
    -
    - -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceBucketMap.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceBucketMap.html deleted file mode 100644 index e95b2cbf6b..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceBucketMap.html +++ /dev/null @@ -1,270 +0,0 @@ - - - - - -IdempotenceBucketMap (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class IdempotenceBucketMap

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.write.IdempotenceBucketMap
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          IdempotenceBucketMap

          -
          public IdempotenceBucketMap()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceManager.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceManager.html deleted file mode 100644 index c6fdd26670..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IdempotenceManager.html +++ /dev/null @@ -1,276 +0,0 @@ - - - - - -IdempotenceManager (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class IdempotenceManager

    -
    -
    - -
    -
      -
    • -
      -
      @Internal
      -@ThreadSafe
      -public class IdempotenceManager
      -extends Object
      -
      A class which manages the idempotence in writer keeps the state necessary to ensure idempotent of - writer.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          IdempotenceManager

          -
          public IdempotenceManager​(boolean idempotenceEnabled,
          -                          int maxInflightRequestsPerBucket,
          -                          org.apache.fluss.rpc.gateway.TabletServerGateway tabletServerGateway)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IndexedLogWriteBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IndexedLogWriteBatch.html deleted file mode 100644 index 7a44311887..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/IndexedLogWriteBatch.html +++ /dev/null @@ -1,597 +0,0 @@ - - - - - -IndexedLogWriteBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class IndexedLogWriteBatch

    -
    -
    - -
    -
      -
    • -
      -
      @NotThreadSafe
      -@Internal
      -public final class IndexedLogWriteBatch
      -extends WriteBatch
      -
      A batch of log records managed in INDEXED format that is or will be sent to server by ProduceLogRequest. - -

      This class is not thread safe and external synchronization must be used when modifying it.

      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          IndexedLogWriteBatch

          -
          public IndexedLogWriteBatch​(int bucketId,
          -                            org.apache.fluss.metadata.PhysicalTablePath physicalTablePath,
          -                            int schemaId,
          -                            int writeLimit,
          -                            org.apache.fluss.memory.AbstractPagedOutputView outputView,
          -                            long createdMs)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          tryAppend

          -
          public boolean tryAppend​(WriteRecord writeRecord,
          -                         WriteCallback callback)
          -                  throws Exception
          -
          Description copied from class: WriteBatch
          -
          try to append one write record to the record batch.
          -
          -
          Specified by:
          -
          tryAppend in class WriteBatch
          -
          Parameters:
          -
          writeRecord - the record to write
          -
          callback - the callback to send back to writer
          -
          Returns:
          -
          true if append success, false if the batch is full.
          -
          Throws:
          -
          Exception
          -
          -
        • -
        - - - -
          -
        • -

          build

          -
          public org.apache.fluss.record.bytesview.BytesView build()
          -
          Description copied from class: WriteBatch
          -
          Gets the memory segment bytes view of the batch. This includes the latest updated WriteBatch.setWriterState(long, int) in the bytes view.
          -
          -
          Specified by:
          -
          build in class WriteBatch
          -
          -
        • -
        - - - -
          -
        • -

          isClosed

          -
          public boolean isClosed()
          -
          Description copied from class: WriteBatch
          -
          check if the batch is closed.
          -
          -
          Specified by:
          -
          isClosed in class WriteBatch
          -
          Returns:
          -
          true if closed, false otherwise
          -
          -
        • -
        - - - - - - - -
          -
        • -

          pooledMemorySegments

          -
          public List<org.apache.fluss.memory.MemorySegment> pooledMemorySegments()
          -
          Description copied from class: WriteBatch
          -
          get pooled memory segments to de-allocate. After produceLog/PutKv acks, the WriteBatch need to de-allocate the allocated pooled MemorySegments back to MemorySegmentPool for reusing.
          -
          -
          Specified by:
          -
          pooledMemorySegments in class WriteBatch
          -
          Returns:
          -
          the pooled memory segment this batch allocated
          -
          -
        • -
        - - - -
          -
        • -

          setWriterState

          -
          public void setWriterState​(long writerId,
          -                           int batchSequence)
          -
          -
          Specified by:
          -
          setWriterState in class WriteBatch
          -
          -
        • -
        - - - -
          -
        • -

          writerId

          -
          public long writerId()
          -
          -
          Specified by:
          -
          writerId in class WriteBatch
          -
          -
        • -
        - - - - - - - - - - - -
          -
        • -

          resetWriterState

          -
          public void resetWriterState​(long writerId,
          -                             int batchSequence)
          -
          -
          Overrides:
          -
          resetWriterState in class WriteBatch
          -
          -
        • -
        - - - -
          -
        • -

          estimatedSizeInBytes

          -
          public int estimatedSizeInBytes()
          -
          Description copied from class: WriteBatch
          -
          Get an estimate of the number of bytes written to the underlying buffer. The returned value - is exactly correct if the record set is not compressed or if the batch has been WriteBatch.build().
          -
          -
          Specified by:
          -
          estimatedSizeInBytes in class WriteBatch
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/KvWriteBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/KvWriteBatch.html deleted file mode 100644 index ce2a15f4ec..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/KvWriteBatch.html +++ /dev/null @@ -1,617 +0,0 @@ - - - - - -KvWriteBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class KvWriteBatch

    -
    -
    - -
    -
      -
    • -
      -
      @NotThreadSafe
      -@Internal
      -public class KvWriteBatch
      -extends WriteBatch
      -
      A batch of kv records that is or will be sent to server by PutKvRequest. - -

      This class is not thread safe and external synchronization must be used when modifying it.

      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          KvWriteBatch

          -
          public KvWriteBatch​(int bucketId,
          -                    org.apache.fluss.metadata.PhysicalTablePath physicalTablePath,
          -                    int schemaId,
          -                    org.apache.fluss.metadata.KvFormat kvFormat,
          -                    int writeLimit,
          -                    org.apache.fluss.memory.AbstractPagedOutputView outputView,
          -                    @Nullable
          -                    int[] targetColumns,
          -                    long createdMs)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          tryAppend

          -
          public boolean tryAppend​(WriteRecord writeRecord,
          -                         WriteCallback callback)
          -                  throws Exception
          -
          Description copied from class: WriteBatch
          -
          try to append one write record to the record batch.
          -
          -
          Specified by:
          -
          tryAppend in class WriteBatch
          -
          Parameters:
          -
          writeRecord - the record to write
          -
          callback - the callback to send back to writer
          -
          Returns:
          -
          true if append success, false if the batch is full.
          -
          Throws:
          -
          Exception
          -
          -
        • -
        - - - -
          -
        • -

          getTargetColumns

          -
          @Nullable
          -public int[] getTargetColumns()
          -
        • -
        - - - -
          -
        • -

          build

          -
          public org.apache.fluss.record.bytesview.BytesView build()
          -
          Description copied from class: WriteBatch
          -
          Gets the memory segment bytes view of the batch. This includes the latest updated WriteBatch.setWriterState(long, int) in the bytes view.
          -
          -
          Specified by:
          -
          build in class WriteBatch
          -
          -
        • -
        - - - - - - - -
          -
        • -

          isClosed

          -
          public boolean isClosed()
          -
          Description copied from class: WriteBatch
          -
          check if the batch is closed.
          -
          -
          Specified by:
          -
          isClosed in class WriteBatch
          -
          Returns:
          -
          true if closed, false otherwise
          -
          -
        • -
        - - - -
          -
        • -

          estimatedSizeInBytes

          -
          public int estimatedSizeInBytes()
          -
          Description copied from class: WriteBatch
          -
          Get an estimate of the number of bytes written to the underlying buffer. The returned value - is exactly correct if the record set is not compressed or if the batch has been WriteBatch.build().
          -
          -
          Specified by:
          -
          estimatedSizeInBytes in class WriteBatch
          -
          -
        • -
        - - - -
          -
        • -

          pooledMemorySegments

          -
          public List<org.apache.fluss.memory.MemorySegment> pooledMemorySegments()
          -
          Description copied from class: WriteBatch
          -
          get pooled memory segments to de-allocate. After produceLog/PutKv acks, the WriteBatch need to de-allocate the allocated pooled MemorySegments back to MemorySegmentPool for reusing.
          -
          -
          Specified by:
          -
          pooledMemorySegments in class WriteBatch
          -
          Returns:
          -
          the pooled memory segment this batch allocated
          -
          -
        • -
        - - - -
          -
        • -

          setWriterState

          -
          public void setWriterState​(long writerId,
          -                           int batchSequence)
          -
          -
          Specified by:
          -
          setWriterState in class WriteBatch
          -
          -
        • -
        - - - -
          -
        • -

          writerId

          -
          public long writerId()
          -
          -
          Specified by:
          -
          writerId in class WriteBatch
          -
          -
        • -
        - - - - - - - - - - - -
          -
        • -

          resetWriterState

          -
          public void resetWriterState​(long writerId,
          -                             int batchSequence)
          -
          -
          Overrides:
          -
          resetWriterState in class WriteBatch
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/ReadyWriteBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/ReadyWriteBatch.html deleted file mode 100644 index a2ea9f7a02..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/ReadyWriteBatch.html +++ /dev/null @@ -1,325 +0,0 @@ - - - - - -ReadyWriteBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class ReadyWriteBatch

    -
    -
    - -
    -
      -
    • -
      -
      public class ReadyWriteBatch
      -extends Object
      -
      The WriteBatch already ready in sender. The difference with WriteBatch is that - the partitionId of the tableBucket in ReadyWriteBatch has already been determined in the dynamic - partition create scenario
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          ReadyWriteBatch

          -
          public ReadyWriteBatch​(org.apache.fluss.metadata.TableBucket tableBucket,
          -                       WriteBatch writeBatch)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          tableBucket

          -
          public org.apache.fluss.metadata.TableBucket tableBucket()
          -
        • -
        - - - -
          -
        • -

          writeBatch

          -
          public WriteBatch writeBatch()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.ReadyCheckResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.ReadyCheckResult.html deleted file mode 100644 index 6d51bc30f6..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.ReadyCheckResult.html +++ /dev/null @@ -1,347 +0,0 @@ - - - - - -RecordAccumulator.ReadyCheckResult (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class RecordAccumulator.ReadyCheckResult

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.write.RecordAccumulator.ReadyCheckResult
      • -
      -
    • -
    -
    -
      -
    • -
      -
      Enclosing class:
      -
      RecordAccumulator
      -
      -
      -
      public static final class RecordAccumulator.ReadyCheckResult
      -extends Object
      -
      The set of nodes that have at leader one complete record batch in the accumulator.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - -
          -
        • -

          readyNodes

          -
          public final Set<Integer> readyNodes
          -
        • -
        - - - -
          -
        • -

          nextReadyCheckDelayMs

          -
          public final long nextReadyCheckDelayMs
          -
        • -
        - - - -
          -
        • -

          unknownLeaderTables

          -
          public final Set<org.apache.fluss.metadata.PhysicalTablePath> unknownLeaderTables
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          ReadyCheckResult

          -
          public ReadyCheckResult​(Set<Integer> readyNodes,
          -                        long nextReadyCheckDelayMs,
          -                        Set<org.apache.fluss.metadata.PhysicalTablePath> unknownLeaderTables)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.RecordAppendResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.RecordAppendResult.html deleted file mode 100644 index d8351ff7e1..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.RecordAppendResult.html +++ /dev/null @@ -1,350 +0,0 @@ - - - - - -RecordAccumulator.RecordAppendResult (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class RecordAccumulator.RecordAppendResult

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.write.RecordAccumulator.RecordAppendResult
      • -
      -
    • -
    -
    -
      -
    • -
      -
      Enclosing class:
      -
      RecordAccumulator
      -
      -
      -
      public static final class RecordAccumulator.RecordAppendResult
      -extends Object
      -
      Metadata about a record just appended to the record accumulator.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - -
          -
        • -

          batchIsFull

          -
          public final boolean batchIsFull
          -
        • -
        - - - -
          -
        • -

          newBatchCreated

          -
          public final boolean newBatchCreated
          -
        • -
        - - - -
          -
        • -

          abortRecordForNewBatch

          -
          public final boolean abortRecordForNewBatch
          -
          Whether this record was abort because the new batch created in record accumulator.
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          RecordAppendResult

          -
          public RecordAppendResult​(boolean batchIsFull,
          -                          boolean newBatchCreated,
          -                          boolean abortRecordForNewBatch)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.html deleted file mode 100644 index 2c51fab987..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RecordAccumulator.html +++ /dev/null @@ -1,541 +0,0 @@ - - - - - -RecordAccumulator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class RecordAccumulator

    -
    -
    - -
    -
      -
    • -
      -
      @Internal
      -public final class RecordAccumulator
      -extends Object
      -
      This class act as a queue that accumulates records into WriteBatch instances to be sent - to tablet servers.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          append

          -
          public RecordAccumulator.RecordAppendResult append​(WriteRecord writeRecord,
          -                                                   WriteCallback callback,
          -                                                   org.apache.fluss.cluster.Cluster cluster,
          -                                                   int bucketId,
          -                                                   boolean abortIfBatchFull)
          -                                            throws Exception
          -
          Add a record to the accumulator, return to append result. - -

          The append result will contain the future metadata, and flag for whether the appended - batch is full or a new batch is created.

          -
          -
          Throws:
          -
          Exception
          -
          -
        • -
        - - - -
          -
        • -

          ready

          -
          public RecordAccumulator.ReadyCheckResult ready​(org.apache.fluss.cluster.Cluster cluster)
          -
          Get a list of nodes whose buckets are ready to be sent. - -

          Also return the flag for whether there are any unknown leaders for the accumulated bucket - batches. - -

          A destination node is ready to send data if: - -

          -     1.There is at least one bucket that is not backing off its send.
          -     2.The record set is full
          -     3.The record set has sat in the accumulator for at least lingerMs milliseconds
          -     4.The accumulator is out of memory and threads are blocking waiting for data (in
          -     this case all buckets are immediately considered ready).
          -     5.The accumulator has been closed
          - 
          -
        • -
        - - - -
          -
        • -

          drain

          -
          public Map<Integer,​List<ReadyWriteBatch>> drain​(org.apache.fluss.cluster.Cluster cluster,
          -                                                      Set<Integer> nodes,
          -                                                      int maxSize)
          -                                               throws Exception
          -
          Drain all the data for the given nodes and collate them into a list of batches that will fit - within the specified size on a per-node basis. This method attempts to avoid choosing the - same table-node over and over.
          -
          -
          Parameters:
          -
          cluster - The current cluster metadata
          -
          nodes - The list of node to drain
          -
          maxSize - The maximum number of bytes to drain
          -
          Returns:
          -
          A list of ReadyWriteBatch for each node specified with total size less than - the requested maxSize.
          -
          Throws:
          -
          Exception
          -
          -
        • -
        - - - -
          -
        • -

          reEnqueue

          -
          public void reEnqueue​(ReadyWriteBatch readyWriteBatch)
          -
        • -
        - - - -
          -
        • -

          abortBatches

          -
          public void abortBatches​(Exception reason)
          -
          Abort all incomplete batches (whether they have been sent or not).
          -
        • -
        - - - -
          -
        • -

          hasUnDrained

          -
          public boolean hasUnDrained()
          -
          Check whether there are any batches which haven't been drained.
          -
        • -
        - - - -
          -
        • -

          hasIncomplete

          -
          public boolean hasIncomplete()
          -
          Check whether there are any pending batches (whether sent or unsent).
          -
        • -
        - - - -
          -
        • -

          beginFlush

          -
          public void beginFlush()
          -
          Initiate the flushing of data from the accumulator...this makes all requests immediately - ready.
          -
        • -
        - - - -
          -
        • -

          awaitFlushCompletion

          -
          public void awaitFlushCompletion()
          -                          throws InterruptedException
          -
          Mark all buckets as ready to send and block until to send is complete.
          -
          -
          Throws:
          -
          InterruptedException
          -
          -
        • -
        - - - -
          -
        • -

          deallocate

          -
          public void deallocate​(WriteBatch batch)
          -
          Deallocate the record batch.
          -
        • -
        - - - -
          -
        • -

          getPhysicalTablePathsInBatches

          -
          public Set<org.apache.fluss.metadata.PhysicalTablePath> getPhysicalTablePathsInBatches()
          -
        • -
        - - - -
          -
        • -

          close

          -
          public void close()
          -
          Close this accumulator and force all the record buffers to be drained.
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RoundRobinBucketAssigner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RoundRobinBucketAssigner.html deleted file mode 100644 index 19cb757a81..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/RoundRobinBucketAssigner.html +++ /dev/null @@ -1,355 +0,0 @@ - - - - - -RoundRobinBucketAssigner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class RoundRobinBucketAssigner

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.write.RoundRobinBucketAssigner
      • -
      -
    • -
    -
    -
      -
    • -
      -
      @Internal
      -public class RoundRobinBucketAssigner
      -extends Object
      -
      The bucket assigner use round-robin strategy.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          RoundRobinBucketAssigner

          -
          public RoundRobinBucketAssigner​(org.apache.fluss.metadata.PhysicalTablePath physicalTablePath)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          assignBucket

          -
          public int assignBucket​(org.apache.fluss.cluster.Cluster cluster)
          -
        • -
        - - - -
          -
        • -

          abortIfBatchFull

          -
          public boolean abortIfBatchFull()
          -
        • -
        - - - -
          -
        • -

          onNewBatch

          -
          public void onNewBatch​(org.apache.fluss.cluster.Cluster cluster,
          -                       int prevBucketId)
          -
        • -
        - - - -
          -
        • -

          assignBucket

          -
          public int assignBucket​(@Nullable
          -                        byte[] bucketKey,
          -                        org.apache.fluss.cluster.Cluster cluster)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/Sender.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/Sender.html deleted file mode 100644 index fb5e8ffc4e..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/Sender.html +++ /dev/null @@ -1,402 +0,0 @@ - - - - - -Sender (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class Sender

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      Runnable
      -
      -
      -
      public class Sender
      -extends Object
      -implements Runnable
      -
      This background thread handles the sending of produce requests to the tablet server. This thread - makes metadata requests to renew its view of the cluster and then sends produce requests to the - appropriate nodes.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      - -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          run

          -
          public void run()
          -
          -
          Specified by:
          -
          run in interface Runnable
          -
          -
        • -
        - - - -
          -
        • -

          runOnce

          -
          public void runOnce()
          -             throws Exception
          -
          Run a single iteration of sending.
          -
          -
          Throws:
          -
          Exception
          -
          -
        • -
        - - - -
          -
        • -

          isRunning

          -
          public boolean isRunning()
          -
        • -
        - - - -
          -
        • -

          forceClose

          -
          public void forceClose()
          -
          Closes the sender without sending out any pending messages.
          -
        • -
        - - - -
          -
        • -

          initiateClose

          -
          public void initiateClose()
          -
          Start closing the sender (won't actually complete until all data is sent out).
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/StickyBucketAssigner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/StickyBucketAssigner.html deleted file mode 100644 index fa79f19320..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/StickyBucketAssigner.html +++ /dev/null @@ -1,356 +0,0 @@ - - - - - -StickyBucketAssigner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class StickyBucketAssigner

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.write.StickyBucketAssigner
      • -
      -
    • -
    -
    -
      -
    • -
      -
      @Internal
      -public class StickyBucketAssigner
      -extends Object
      -
      The bucket assigner with sticky strategy. The assigned bucket id maybe changed only if one new - batch created in record accumulator. Otherwise, we will always return the same bucket id.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          StickyBucketAssigner

          -
          public StickyBucketAssigner​(org.apache.fluss.metadata.PhysicalTablePath physicalTablePath)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          assignBucket

          -
          public int assignBucket​(org.apache.fluss.cluster.Cluster cluster)
          -
        • -
        - - - -
          -
        • -

          abortIfBatchFull

          -
          public boolean abortIfBatchFull()
          -
        • -
        - - - -
          -
        • -

          onNewBatch

          -
          public void onNewBatch​(org.apache.fluss.cluster.Cluster cluster,
          -                       int prevBucketId)
          -
        • -
        - - - -
          -
        • -

          assignBucket

          -
          public int assignBucket​(@Nullable
          -                        byte[] bucketKey,
          -                        org.apache.fluss.cluster.Cluster cluster)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteBatch.RequestFuture.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteBatch.RequestFuture.html deleted file mode 100644 index bfb3e02837..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteBatch.RequestFuture.html +++ /dev/null @@ -1,336 +0,0 @@ - - - - - -WriteBatch.RequestFuture (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class WriteBatch.RequestFuture

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.client.write.WriteBatch.RequestFuture
      • -
      -
    • -
    -
    -
      -
    • -
      -
      Enclosing class:
      -
      WriteBatch
      -
      -
      -
      public static class WriteBatch.RequestFuture
      -extends Object
      -
      The future for this batch.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          RequestFuture

          -
          public RequestFuture()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          done

          -
          public void done()
          -
          Mark this request as complete and unblock any threads waiting on its completion.
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteBatch.html deleted file mode 100644 index 74610acf96..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteBatch.html +++ /dev/null @@ -1,763 +0,0 @@ - - - - - -WriteBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class WriteBatch

    -
    -
    - -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - - - - - -
          -
        • -

          reopened

          -
          protected boolean reopened
          -
        • -
        - - - -
          -
        • -

          recordCount

          -
          protected int recordCount
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          WriteBatch

          -
          public WriteBatch​(int bucketId,
          -                  org.apache.fluss.metadata.PhysicalTablePath physicalTablePath,
          -                  long createdMs)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          tryAppend

          -
          public abstract boolean tryAppend​(WriteRecord writeRecord,
          -                                  WriteCallback callback)
          -                           throws Exception
          -
          try to append one write record to the record batch.
          -
          -
          Parameters:
          -
          writeRecord - the record to write
          -
          callback - the callback to send back to writer
          -
          Returns:
          -
          true if append success, false if the batch is full.
          -
          Throws:
          -
          Exception
          -
          -
        • -
        - - - -
          -
        • -

          build

          -
          public abstract org.apache.fluss.record.bytesview.BytesView build()
          -
          Gets the memory segment bytes view of the batch. This includes the latest updated setWriterState(long, int) in the bytes view.
          -
        • -
        - - - -
          -
        • -

          close

          -
          public abstract void close()
          -                    throws Exception
          -
          close the batch to not append new records.
          -
          -
          Throws:
          -
          Exception
          -
          -
        • -
        - - - -
          -
        • -

          isClosed

          -
          public abstract boolean isClosed()
          -
          check if the batch is closed.
          -
          -
          Returns:
          -
          true if closed, false otherwise
          -
          -
        • -
        - - - -
          -
        • -

          estimatedSizeInBytes

          -
          public abstract int estimatedSizeInBytes()
          -
          Get an estimate of the number of bytes written to the underlying buffer. The returned value - is exactly correct if the record set is not compressed or if the batch has been build().
          -
        • -
        - - - -
          -
        • -

          pooledMemorySegments

          -
          public abstract List<org.apache.fluss.memory.MemorySegment> pooledMemorySegments()
          -
          get pooled memory segments to de-allocate. After produceLog/PutKv acks, the WriteBatch need to de-allocate the allocated pooled MemorySegments back to MemorySegmentPool for reusing.
          -
          -
          Returns:
          -
          the pooled memory segment this batch allocated
          -
          -
        • -
        - - - -
          -
        • -

          setWriterState

          -
          public abstract void setWriterState​(long writerId,
          -                                    int batchSequence)
          -
        • -
        - - - -
          -
        • -

          writerId

          -
          public abstract long writerId()
          -
        • -
        - - - -
          -
        • -

          batchSequence

          -
          public abstract int batchSequence()
          -
        • -
        - - - -
          -
        • -

          abortRecordAppends

          -
          public abstract void abortRecordAppends()
          -
        • -
        - - - -
          -
        • -

          hasBatchSequence

          -
          public boolean hasBatchSequence()
          -
        • -
        - - - -
          -
        • -

          resetWriterState

          -
          public void resetWriterState​(long writerId,
          -                             int batchSequence)
          -
        • -
        - - - -
          -
        • -

          abort

          -
          public void abort​(Exception exception)
          -
          Abort the batch and complete the future and callbacks.
          -
        • -
        - - - -
          -
        • -

          sequenceHasBeenReset

          -
          public boolean sequenceHasBeenReset()
          -
        • -
        - - - -
          -
        • -

          bucketId

          -
          public int bucketId()
          -
        • -
        - - - -
          -
        • -

          physicalTablePath

          -
          public org.apache.fluss.metadata.PhysicalTablePath physicalTablePath()
          -
        • -
        - - - - - - - -
          -
        • -

          waitedTimeMs

          -
          public long waitedTimeMs​(long nowMs)
          -
        • -
        - - - -
          -
        • -

          getRecordCount

          -
          public int getRecordCount()
          -
        • -
        - - - -
          -
        • -

          getQueueTimeMs

          -
          public long getQueueTimeMs()
          -
        • -
        - - - -
          -
        • -

          complete

          -
          public boolean complete()
          -
          Complete the batch successfully.
          -
        • -
        - - - -
          -
        • -

          completeExceptionally

          -
          public boolean completeExceptionally​(Exception exception)
          -
          Complete the batch exceptionally. The provided exception will be used for each record future - contained in the batch.
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteCallback.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteCallback.html deleted file mode 100644 index 8ec2431bac..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteCallback.html +++ /dev/null @@ -1,255 +0,0 @@ - - - - - -WriteCallback (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface WriteCallback

    -
    -
    -
    -
      -
    • -
      -
      @Internal
      -public interface WriteCallback
      -
      A callback interface that the user can implement to allow code to execute when the write request - is complete. This callback will generally execute in the background I/O thread, so it should be - fast.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          onCompletion

          -
          void onCompletion​(@Nullable
          -                  Exception exception)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteFormat.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteFormat.html deleted file mode 100644 index 8438b335c9..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteFormat.html +++ /dev/null @@ -1,393 +0,0 @@ - - - - - -WriteFormat (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Enum WriteFormat

    -
    -
    - -
    - -
    -
    - -
    -
    -
      -
    • - -
      - -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          values

          -
          public static WriteFormat[] values()
          -
          Returns an array containing the constants of this enum type, in -the order they are declared. This method may be used to iterate -over the constants as follows: -
          -for (WriteFormat c : WriteFormat.values())
          -    System.out.println(c);
          -
          -
          -
          Returns:
          -
          an array containing the constants of this enum type, in the order they are declared
          -
          -
        • -
        - - - -
          -
        • -

          valueOf

          -
          public static WriteFormat valueOf​(String name)
          -
          Returns the enum constant of this type with the specified name. -The string must match exactly an identifier used to declare an -enum constant in this type. (Extraneous whitespace characters are -not permitted.)
          -
          -
          Parameters:
          -
          name - the name of the enum constant to be returned.
          -
          Returns:
          -
          the enum constant with the specified name
          -
          Throws:
          -
          IllegalArgumentException - if this enum type has no constant with the specified name
          -
          NullPointerException - if the argument is null
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteRecord.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteRecord.html deleted file mode 100644 index a897784627..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriteRecord.html +++ /dev/null @@ -1,461 +0,0 @@ - - - - - -WriteRecord (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class WriteRecord

    -
    -
    - -
    -
      -
    • -
      -
      @Internal
      -public final class WriteRecord
      -extends Object
      -
      A record to write to a table. It can represent an upsert operation, a delete operation, or an - append operation.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          forUpsert

          -
          public static WriteRecord forUpsert​(org.apache.fluss.metadata.PhysicalTablePath tablePath,
          -                                    org.apache.fluss.row.BinaryRow row,
          -                                    byte[] key,
          -                                    byte[] bucketKey,
          -                                    @Nullable
          -                                    int[] targetColumns)
          -
          Create a write record for upsert operation and partial-upsert operation.
          -
        • -
        - - - -
          -
        • -

          forDelete

          -
          public static WriteRecord forDelete​(org.apache.fluss.metadata.PhysicalTablePath tablePath,
          -                                    byte[] key,
          -                                    byte[] bucketKey,
          -                                    @Nullable
          -                                    int[] targetColumns)
          -
          Create a write record for delete operation and partial-delete update.
          -
        • -
        - - - -
          -
        • -

          forIndexedAppend

          -
          public static WriteRecord forIndexedAppend​(org.apache.fluss.metadata.PhysicalTablePath tablePath,
          -                                           org.apache.fluss.row.indexed.IndexedRow row,
          -                                           @Nullable
          -                                           byte[] bucketKey)
          -
          Create a write record for append operation for indexed format.
          -
        • -
        - - - -
          -
        • -

          forArrowAppend

          -
          public static WriteRecord forArrowAppend​(org.apache.fluss.metadata.PhysicalTablePath tablePath,
          -                                         org.apache.fluss.row.InternalRow row,
          -                                         @Nullable
          -                                         byte[] bucketKey)
          -
          Creates a write record for append operation for Arrow format.
          -
        • -
        - - - -
          -
        • -

          getPhysicalTablePath

          -
          public org.apache.fluss.metadata.PhysicalTablePath getPhysicalTablePath()
          -
        • -
        - - - -
          -
        • -

          getKey

          -
          @Nullable
          -public byte[] getKey()
          -
        • -
        - - - -
          -
        • -

          getBucketKey

          -
          @Nullable
          -public byte[] getBucketKey()
          -
        • -
        - - - -
          -
        • -

          getRow

          -
          @Nullable
          -public org.apache.fluss.row.InternalRow getRow()
          -
        • -
        - - - -
          -
        • -

          getTargetColumns

          -
          @Nullable
          -public int[] getTargetColumns()
          -
        • -
        - - - -
          -
        • -

          getWriteFormat

          -
          public WriteFormat getWriteFormat()
          -
        • -
        - - - -
          -
        • -

          getEstimatedSizeInBytes

          -
          public int getEstimatedSizeInBytes()
          -
          Get the estimated size in bytes of the record with batch header.
          -
          -
          Returns:
          -
          the estimated size in bytes of the record with batch header
          -
          Throws:
          -
          IllegalStateException - if the estimated size in bytes is not supported for the write - format
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriterClient.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriterClient.html deleted file mode 100644 index faee5f1fa1..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/WriterClient.html +++ /dev/null @@ -1,419 +0,0 @@ - - - - - -WriterClient (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class WriterClient

    -
    -
    - -
    -
      -
    • -
      -
      @ThreadSafe
      -@Internal
      -public class WriterClient
      -extends Object
      -
      A client that write records to server. - -

      The writer consists of a pool of buffer space that holds records that haven't yet been - transmitted to the tablet server as well as a background I/O thread that is responsible for - turning these records into requests and transmitting them to the cluster. Failure to close the - WriterClient after use will leak these resources. - -

      The send method is asynchronous. When called, it adds the log record to a buffer of pending - record sends and immediately returns. This allows the wrote record to batch together individual - records for efficiency.

      -
    • -
    -
    -
    -
      -
    • - -
      - -
      - -
      -
        -
      • - - -

        Constructor Summary

        - - - - - - - - - - -
        Constructors 
        ConstructorDescription
        WriterClient​(org.apache.fluss.config.Configuration conf, - MetadataUpdater metadataUpdater, - org.apache.fluss.rpc.metrics.ClientMetricGroup clientMetricGroup, - Admin admin) 
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
      -
    • - -
      - -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          WriterClient

          -
          public WriterClient​(org.apache.fluss.config.Configuration conf,
          -                    MetadataUpdater metadataUpdater,
          -                    org.apache.fluss.rpc.metrics.ClientMetricGroup clientMetricGroup,
          -                    Admin admin)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          send

          -
          public void send​(WriteRecord record,
          -                 WriteCallback callback)
          -
          Asynchronously send a record to a table and invoke the provided callback when to send has - been acknowledged.
          -
        • -
        - - - -
          -
        • -

          flush

          -
          public void flush()
          -
          Invoking this method makes all buffered records immediately available to send (even if - linger.ms is greater than 0) and blocks on the completion of the requests associated - with these records. The post-condition of flush() is that any previously sent - record will have completed (e.g. Future.isDone() == true). A request is - considered completed when it is successfully acknowledged according to the acks - configuration you have specified or else it results in an error. - -

          Other threads can continue sending records while one thread is blocked waiting for a flush - call to complete, however no guarantee is made about the completion of records sent after the - flush call begins.

          -
        • -
        - - - -
          -
        • -

          close

          -
          public void close​(Duration timeout)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/ArrowLogWriteBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/ArrowLogWriteBatch.html deleted file mode 100644 index 6402d55796..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/ArrowLogWriteBatch.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.write.ArrowLogWriteBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.write.ArrowLogWriteBatch

    -
    -
    No usage of org.apache.fluss.client.write.ArrowLogWriteBatch
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/DynamicPartitionCreator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/DynamicPartitionCreator.html deleted file mode 100644 index 45e3d22c05..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/DynamicPartitionCreator.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.write.DynamicPartitionCreator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.write.DynamicPartitionCreator

    -
    -
    No usage of org.apache.fluss.client.write.DynamicPartitionCreator
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/DynamicWriteBatchSizeEstimator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/DynamicWriteBatchSizeEstimator.html deleted file mode 100644 index 170647c8b5..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/DynamicWriteBatchSizeEstimator.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.write.DynamicWriteBatchSizeEstimator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.write.DynamicWriteBatchSizeEstimator

    -
    -
    No usage of org.apache.fluss.client.write.DynamicWriteBatchSizeEstimator
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/HashBucketAssigner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/HashBucketAssigner.html deleted file mode 100644 index c54e8bc87a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/HashBucketAssigner.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.write.HashBucketAssigner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.write.HashBucketAssigner

    -
    -
    No usage of org.apache.fluss.client.write.HashBucketAssigner
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceBucketEntry.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceBucketEntry.html deleted file mode 100644 index 4f01b83fda..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceBucketEntry.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.write.IdempotenceBucketEntry (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.write.IdempotenceBucketEntry

    -
    -
    No usage of org.apache.fluss.client.write.IdempotenceBucketEntry
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceBucketMap.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceBucketMap.html deleted file mode 100644 index 1638841664..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceBucketMap.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.write.IdempotenceBucketMap (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.write.IdempotenceBucketMap

    -
    -
    No usage of org.apache.fluss.client.write.IdempotenceBucketMap
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceManager.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceManager.html deleted file mode 100644 index c54387dcc6..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IdempotenceManager.html +++ /dev/null @@ -1,197 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.write.IdempotenceManager (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.write.IdempotenceManager

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IndexedLogWriteBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IndexedLogWriteBatch.html deleted file mode 100644 index d7d1d13ae5..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/IndexedLogWriteBatch.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.write.IndexedLogWriteBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.write.IndexedLogWriteBatch

    -
    -
    No usage of org.apache.fluss.client.write.IndexedLogWriteBatch
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/KvWriteBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/KvWriteBatch.html deleted file mode 100644 index 1ff1e9eeec..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/KvWriteBatch.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.write.KvWriteBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.write.KvWriteBatch

    -
    -
    No usage of org.apache.fluss.client.write.KvWriteBatch
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/ReadyWriteBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/ReadyWriteBatch.html deleted file mode 100644 index b6bb435e15..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/ReadyWriteBatch.html +++ /dev/null @@ -1,249 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.write.ReadyWriteBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.write.ReadyWriteBatch

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.ReadyCheckResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.ReadyCheckResult.html deleted file mode 100644 index 46d4a8fb32..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.ReadyCheckResult.html +++ /dev/null @@ -1,194 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.write.RecordAccumulator.ReadyCheckResult (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.write.RecordAccumulator.ReadyCheckResult

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.RecordAppendResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.RecordAppendResult.html deleted file mode 100644 index d65d4073a0..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.RecordAppendResult.html +++ /dev/null @@ -1,198 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.write.RecordAccumulator.RecordAppendResult (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.write.RecordAccumulator.RecordAppendResult

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.html deleted file mode 100644 index 37e8d34a73..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RecordAccumulator.html +++ /dev/null @@ -1,197 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.write.RecordAccumulator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.write.RecordAccumulator

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RoundRobinBucketAssigner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RoundRobinBucketAssigner.html deleted file mode 100644 index d0e8c090f4..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/RoundRobinBucketAssigner.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.write.RoundRobinBucketAssigner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.write.RoundRobinBucketAssigner

    -
    -
    No usage of org.apache.fluss.client.write.RoundRobinBucketAssigner
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/Sender.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/Sender.html deleted file mode 100644 index f04b48f1dd..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/Sender.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.write.Sender (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.write.Sender

    -
    -
    No usage of org.apache.fluss.client.write.Sender
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/StickyBucketAssigner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/StickyBucketAssigner.html deleted file mode 100644 index 27f4aa71fa..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/StickyBucketAssigner.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.write.StickyBucketAssigner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.write.StickyBucketAssigner

    -
    -
    No usage of org.apache.fluss.client.write.StickyBucketAssigner
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteBatch.RequestFuture.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteBatch.RequestFuture.html deleted file mode 100644 index ccb00ed1d7..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteBatch.RequestFuture.html +++ /dev/null @@ -1,192 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.write.WriteBatch.RequestFuture (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.write.WriteBatch.RequestFuture

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteBatch.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteBatch.html deleted file mode 100644 index 804445429f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteBatch.html +++ /dev/null @@ -1,254 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.write.WriteBatch (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.write.WriteBatch

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteCallback.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteCallback.html deleted file mode 100644 index d2f4d2195c..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteCallback.html +++ /dev/null @@ -1,248 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.client.write.WriteCallback (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.client.write.WriteCallback

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteFormat.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteFormat.html deleted file mode 100644 index bb433ec798..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteFormat.html +++ /dev/null @@ -1,207 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.write.WriteFormat (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.write.WriteFormat

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteRecord.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteRecord.html deleted file mode 100644 index ede1e65e35..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriteRecord.html +++ /dev/null @@ -1,308 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.write.WriteRecord (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.write.WriteRecord

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriterClient.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriterClient.html deleted file mode 100644 index cbf1d1a245..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/class-use/WriterClient.html +++ /dev/null @@ -1,245 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.client.write.WriterClient (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.client.write.WriterClient

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-summary.html deleted file mode 100644 index 5ed026a837..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-summary.html +++ /dev/null @@ -1,319 +0,0 @@ - - - - - -org.apache.fluss.client.write (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.client.write

    -
    -
    -
      -
    • - - - - - - - - - - - - -
      Interface Summary 
      InterfaceDescription
      WriteCallback -
      A callback interface that the user can implement to allow code to execute when the write request - is complete.
      -
      -
    • -
    • - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
      Class Summary 
      ClassDescription
      ArrowLogWriteBatch -
      A batch of log records managed in ARROW format that is or will be sent to server by ProduceLogRequest.
      -
      DynamicPartitionCreator -
      A creator to create partition when dynamic partition create enable for table.
      -
      DynamicWriteBatchSizeEstimator -
      An estimator to estimate the buffer usage of a writeBatch.
      -
      HashBucketAssigner -
      Hash bucket assigner.
      -
      IdempotenceBucketEntry -
      Entry to store the idempotence information of each table-bucket.
      -
      IdempotenceBucketMap -
      Map to manage IdempotenceBucketEntry of all table-bucket.
      -
      IdempotenceManager -
      A class which manages the idempotence in writer keeps the state necessary to ensure idempotent of - writer.
      -
      IndexedLogWriteBatch -
      A batch of log records managed in INDEXED format that is or will be sent to server by ProduceLogRequest.
      -
      KvWriteBatch -
      A batch of kv records that is or will be sent to server by PutKvRequest.
      -
      ReadyWriteBatch -
      The WriteBatch already ready in sender.
      -
      RecordAccumulator -
      This class act as a queue that accumulates records into WriteBatch instances to be sent - to tablet servers.
      -
      RecordAccumulator.ReadyCheckResult -
      The set of nodes that have at leader one complete record batch in the accumulator.
      -
      RecordAccumulator.RecordAppendResult -
      Metadata about a record just appended to the record accumulator.
      -
      RoundRobinBucketAssigner -
      The bucket assigner use round-robin strategy.
      -
      Sender -
      This background thread handles the sending of produce requests to the tablet server.
      -
      StickyBucketAssigner -
      The bucket assigner with sticky strategy.
      -
      WriteBatch -
      The abstract write batch contains write callback object to wait write request feedback.
      -
      WriteBatch.RequestFuture -
      The future for this batch.
      -
      WriterClient -
      A client that write records to server.
      -
      WriteRecord -
      A record to write to a table.
      -
      -
    • -
    • - - - - - - - - - - - - -
      Enum Summary 
      EnumDescription
      WriteFormat -
      The format of the write record.
      -
      -
    • -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-tree.html deleted file mode 100644 index cbb0ac96a7..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-tree.html +++ /dev/null @@ -1,205 +0,0 @@ - - - - - -org.apache.fluss.client.write Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.client.write

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -

    Interface Hierarchy

    - -
    -
    -

    Enum Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-use.html deleted file mode 100644 index da9cf93515..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/client/write/package-use.html +++ /dev/null @@ -1,317 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.client.write (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.client.write

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/FlinkConnectorOptions.ScanStartupMode.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/FlinkConnectorOptions.ScanStartupMode.html deleted file mode 100644 index 59f8fce695..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/FlinkConnectorOptions.ScanStartupMode.html +++ /dev/null @@ -1,446 +0,0 @@ - - - - - -FlinkConnectorOptions.ScanStartupMode (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Enum FlinkConnectorOptions.ScanStartupMode

    -
    -
    - -
    - -
    -
    - -
    -
    -
      -
    • - -
      - -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          values

          -
          public static FlinkConnectorOptions.ScanStartupMode[] values()
          -
          Returns an array containing the constants of this enum type, in -the order they are declared. This method may be used to iterate -over the constants as follows: -
          -for (FlinkConnectorOptions.ScanStartupMode c : FlinkConnectorOptions.ScanStartupMode.values())
          -    System.out.println(c);
          -
          -
          -
          Returns:
          -
          an array containing the constants of this enum type, in the order they are declared
          -
          -
        • -
        - - - -
          -
        • -

          valueOf

          -
          public static FlinkConnectorOptions.ScanStartupMode valueOf​(String name)
          -
          Returns the enum constant of this type with the specified name. -The string must match exactly an identifier used to declare an -enum constant in this type. (Extraneous whitespace characters are -not permitted.)
          -
          -
          Parameters:
          -
          name - the name of the enum constant to be returned.
          -
          Returns:
          -
          the enum constant with the specified name
          -
          Throws:
          -
          IllegalArgumentException - if this enum type has no constant with the specified name
          -
          NullPointerException - if the argument is null
          -
          -
        • -
        - - - - - - - -
          -
        • -

          getDescription

          -
          public org.apache.flink.configuration.description.InlineElement getDescription()
          -
          -
          Specified by:
          -
          getDescription in interface org.apache.flink.configuration.DescribedEnum
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/FlinkConnectorOptions.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/FlinkConnectorOptions.html deleted file mode 100644 index 898bc12894..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/FlinkConnectorOptions.html +++ /dev/null @@ -1,476 +0,0 @@ - - - - - -FlinkConnectorOptions (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkConnectorOptions

    -
    -
    - -
    -
      -
    • -
      -
      public class FlinkConnectorOptions
      -extends Object
      -
      Options for flink connector.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - -
          -
        • -

          BUCKET_NUMBER

          -
          public static final org.apache.flink.configuration.ConfigOption<Integer> BUCKET_NUMBER
          -
        • -
        - - - -
          -
        • -

          BUCKET_KEY

          -
          public static final org.apache.flink.configuration.ConfigOption<String> BUCKET_KEY
          -
        • -
        - - - -
          -
        • -

          BOOTSTRAP_SERVERS

          -
          public static final org.apache.flink.configuration.ConfigOption<String> BOOTSTRAP_SERVERS
          -
        • -
        - - - -
          -
        • -

          LOOKUP_ASYNC

          -
          public static final org.apache.flink.configuration.ConfigOption<Boolean> LOOKUP_ASYNC
          -
        • -
        - - - - - - - -
          -
        • -

          SCAN_STARTUP_TIMESTAMP

          -
          public static final org.apache.flink.configuration.ConfigOption<String> SCAN_STARTUP_TIMESTAMP
          -
        • -
        - - - -
          -
        • -

          SCAN_PARTITION_DISCOVERY_INTERVAL

          -
          public static final org.apache.flink.configuration.ConfigOption<Duration> SCAN_PARTITION_DISCOVERY_INTERVAL
          -
        • -
        - - - -
          -
        • -

          SINK_IGNORE_DELETE

          -
          public static final org.apache.flink.configuration.ConfigOption<Boolean> SINK_IGNORE_DELETE
          -
        • -
        - - - -
          -
        • -

          SINK_BUCKET_SHUFFLE

          -
          public static final org.apache.flink.configuration.ConfigOption<Boolean> SINK_BUCKET_SHUFFLE
          -
        • -
        - - - -
          -
        • -

          TABLE_OPTIONS

          -
          public static final List<org.apache.flink.configuration.ConfigOption<?>> TABLE_OPTIONS
          -
        • -
        - - - -
          -
        • -

          CLIENT_OPTIONS

          -
          public static final List<org.apache.flink.configuration.ConfigOption<?>> CLIENT_OPTIONS
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlinkConnectorOptions

          -
          public FlinkConnectorOptions()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/CatalogTableAdapter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/CatalogTableAdapter.html deleted file mode 100644 index f906e27d30..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/CatalogTableAdapter.html +++ /dev/null @@ -1,315 +0,0 @@ - - - - - -CatalogTableAdapter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class CatalogTableAdapter

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.adapter.CatalogTableAdapter
      • -
      -
    • -
    -
    -
      -
    • -
      -
      public class CatalogTableAdapter
      -extends Object
      -
      A adapter for CatalogTable constructor. TODO: remove this class when no longer support - flink 1.18 and 1.19.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          CatalogTableAdapter

          -
          public CatalogTableAdapter()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          toCatalogTable

          -
          public static org.apache.flink.table.catalog.CatalogTable toCatalogTable​(org.apache.flink.table.api.Schema schema,
          -                                                                         @Nullable
          -                                                                         String comment,
          -                                                                         List<String> partitionKeys,
          -                                                                         Map<String,​String> options)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/SingleThreadFetcherManagerAdapter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/SingleThreadFetcherManagerAdapter.html deleted file mode 100644 index cfcca073db..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/SingleThreadFetcherManagerAdapter.html +++ /dev/null @@ -1,320 +0,0 @@ - - - - - -SingleThreadFetcherManagerAdapter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class SingleThreadFetcherManagerAdapter<E,​SplitT extends org.apache.flink.api.connector.source.SourceSplit>

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager<E,​SplitT>
      • -
      • -
          -
        • org.apache.flink.connector.base.source.reader.fetcher.SingleThreadFetcherManager<E,​SplitT>
        • -
        • -
            -
          • org.apache.fluss.flink.adapter.SingleThreadFetcherManagerAdapter<E,​SplitT>
          • -
          -
        • -
        -
      • -
      -
    • -
    -
    -
      -
    • -
      -
      Direct Known Subclasses:
      -
      FlinkSourceFetcherManager
      -
      -
      -
      public class SingleThreadFetcherManagerAdapter<E,​SplitT extends org.apache.flink.api.connector.source.SourceSplit>
      -extends org.apache.flink.connector.base.source.reader.fetcher.SingleThreadFetcherManager<E,​SplitT>
      -
      Adapter for SingleThreadFetcherManager.TODO: remove it until not supported in flink 1.18.
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Summary

        -
          -
        • - - -

          Fields inherited from class org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager

          -fetchers
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Summary

        - - - - - - - - - - -
        Constructors 
        ConstructorDescription
        SingleThreadFetcherManagerAdapter​(org.apache.flink.connector.base.source.reader.synchronization.FutureCompletingBlockingQueue<org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<E>> elementsQueue, - Supplier<org.apache.flink.connector.base.source.reader.splitreader.SplitReader<E,​SplitT>> splitReaderSupplier, - org.apache.flink.configuration.Configuration configuration, - Consumer<Collection<String>> splitFinishedHook) 
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Summary

        -
          -
        • - - -

          Methods inherited from class org.apache.flink.connector.base.source.reader.fetcher.SingleThreadFetcherManager

          -addSplits, getRunningFetcher, removeSplits
        • -
        -
          -
        • - - -

          Methods inherited from class org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager

          -checkErrors, close, createSplitFetcher, getNumAliveFetchers, getQueue, maybeShutdownFinishedFetchers, pauseOrResumeSplits, startFetcher
        • -
        - -
      • -
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          SingleThreadFetcherManagerAdapter

          -
          public SingleThreadFetcherManagerAdapter​(org.apache.flink.connector.base.source.reader.synchronization.FutureCompletingBlockingQueue<org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<E>> elementsQueue,
          -                                         Supplier<org.apache.flink.connector.base.source.reader.splitreader.SplitReader<E,​SplitT>> splitReaderSupplier,
          -                                         org.apache.flink.configuration.Configuration configuration,
          -                                         Consumer<Collection<String>> splitFinishedHook)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/SingleThreadMultiplexSourceReaderBaseAdapter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/SingleThreadMultiplexSourceReaderBaseAdapter.html deleted file mode 100644 index 57b0023240..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/SingleThreadMultiplexSourceReaderBaseAdapter.html +++ /dev/null @@ -1,334 +0,0 @@ - - - - - -SingleThreadMultiplexSourceReaderBaseAdapter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class SingleThreadMultiplexSourceReaderBaseAdapter<E,​T,​SplitT extends org.apache.flink.api.connector.source.SourceSplit,​SplitStateT>

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.flink.connector.base.source.reader.SourceReaderBase<E,​T,​SplitT,​SplitStateT>
      • -
      • -
          -
        • org.apache.flink.connector.base.source.reader.SingleThreadMultiplexSourceReaderBase<E,​T,​SplitT,​SplitStateT>
        • -
        • -
            -
          • org.apache.fluss.flink.adapter.SingleThreadMultiplexSourceReaderBaseAdapter<E,​T,​SplitT,​SplitStateT>
          • -
          -
        • -
        -
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      AutoCloseable, org.apache.flink.api.common.state.CheckpointListener, org.apache.flink.api.connector.source.SourceReader<T,​SplitT>
      -
      -
      -
      Direct Known Subclasses:
      -
      FlinkSourceReader
      -
      -
      -
      public abstract class SingleThreadMultiplexSourceReaderBaseAdapter<E,​T,​SplitT extends org.apache.flink.api.connector.source.SourceSplit,​SplitStateT>
      -extends org.apache.flink.connector.base.source.reader.SingleThreadMultiplexSourceReaderBase<E,​T,​SplitT,​SplitStateT>
      -
      Adapter for SingleThreadMultiplexSourceReaderBase.TODO: remove it until not supported in - flink 1.18.
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Summary

        -
          -
        • - - -

          Fields inherited from class org.apache.flink.connector.base.source.reader.SourceReaderBase

          -config, context, eofRecordEvaluator, options, recordEmitter, splitFetcherManager
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Summary

        - - - - - - - - - - -
        Constructors 
        ConstructorDescription
        SingleThreadMultiplexSourceReaderBaseAdapter​(org.apache.flink.connector.base.source.reader.synchronization.FutureCompletingBlockingQueue<org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<E>> elementsQueue, - org.apache.flink.connector.base.source.reader.fetcher.SingleThreadFetcherManager<E,​SplitT> splitFetcherManager, - org.apache.flink.connector.base.source.reader.RecordEmitter<E,​T,​SplitStateT> recordEmitter, - org.apache.flink.configuration.Configuration config, - org.apache.flink.api.connector.source.SourceReaderContext context) 
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Summary

        -
          -
        • - - -

          Methods inherited from class org.apache.flink.connector.base.source.reader.SourceReaderBase

          -addSplits, close, getNumberOfCurrentlyAssignedSplits, handleSourceEvents, initializedState, isAvailable, notifyNoMoreSplits, onSplitFinished, pauseOrResumeSplits, pollNext, snapshotState, start, toSplitType
        • -
        - -
          -
        • - - -

          Methods inherited from interface org.apache.flink.api.common.state.CheckpointListener

          -notifyCheckpointAborted
        • -
        -
          -
        • - - -

          Methods inherited from interface org.apache.flink.api.connector.source.SourceReader

          -notifyCheckpointComplete
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          SingleThreadMultiplexSourceReaderBaseAdapter

          -
          public SingleThreadMultiplexSourceReaderBaseAdapter​(org.apache.flink.connector.base.source.reader.synchronization.FutureCompletingBlockingQueue<org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<E>> elementsQueue,
          -                                                    org.apache.flink.connector.base.source.reader.fetcher.SingleThreadFetcherManager<E,​SplitT> splitFetcherManager,
          -                                                    org.apache.flink.connector.base.source.reader.RecordEmitter<E,​T,​SplitStateT> recordEmitter,
          -                                                    org.apache.flink.configuration.Configuration config,
          -                                                    org.apache.flink.api.connector.source.SourceReaderContext context)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/CatalogTableAdapter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/CatalogTableAdapter.html deleted file mode 100644 index 208b884445..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/CatalogTableAdapter.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.adapter.CatalogTableAdapter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.adapter.CatalogTableAdapter

    -
    -
    No usage of org.apache.fluss.flink.adapter.CatalogTableAdapter
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/SingleThreadFetcherManagerAdapter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/SingleThreadFetcherManagerAdapter.html deleted file mode 100644 index b3236fc615..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/SingleThreadFetcherManagerAdapter.html +++ /dev/null @@ -1,194 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.adapter.SingleThreadFetcherManagerAdapter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.adapter.SingleThreadFetcherManagerAdapter

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/SingleThreadMultiplexSourceReaderBaseAdapter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/SingleThreadMultiplexSourceReaderBaseAdapter.html deleted file mode 100644 index ff25e2207c..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/class-use/SingleThreadMultiplexSourceReaderBaseAdapter.html +++ /dev/null @@ -1,194 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.adapter.SingleThreadMultiplexSourceReaderBaseAdapter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.adapter.SingleThreadMultiplexSourceReaderBaseAdapter

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-summary.html deleted file mode 100644 index d3e6c47b31..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-summary.html +++ /dev/null @@ -1,181 +0,0 @@ - - - - - -org.apache.fluss.flink.adapter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.adapter

    -
    -
    -
      -
    • - - - - - - - - - - - - - - - - - - - - -
      Class Summary 
      ClassDescription
      CatalogTableAdapter -
      A adapter for CatalogTable constructor.
      -
      SingleThreadFetcherManagerAdapter<E,​SplitT extends org.apache.flink.api.connector.source.SourceSplit> -
      Adapter for SingleThreadFetcherManager.TODO: remove it until not supported in flink 1.18.
      -
      SingleThreadMultiplexSourceReaderBaseAdapter<E,​T,​SplitT extends org.apache.flink.api.connector.source.SourceSplit,​SplitStateT> -
      Adapter for SingleThreadMultiplexSourceReaderBase.TODO: remove it until not supported in - flink 1.18.
      -
      -
    • -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-tree.html deleted file mode 100644 index b4633f7b55..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-tree.html +++ /dev/null @@ -1,181 +0,0 @@ - - - - - -org.apache.fluss.flink.adapter Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.adapter

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    -
      -
    • java.lang.Object -
        -
      • org.apache.fluss.flink.adapter.CatalogTableAdapter
      • -
      • org.apache.flink.connector.base.source.reader.SourceReaderBase<E,​T,​SplitT,​SplitStateT> (implements org.apache.flink.api.connector.source.SourceReader<T,​SplitT>) -
          -
        • org.apache.flink.connector.base.source.reader.SingleThreadMultiplexSourceReaderBase<E,​T,​SplitT,​SplitStateT> - -
        • -
        -
      • -
      • org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager<E,​SplitT> -
          -
        • org.apache.flink.connector.base.source.reader.fetcher.SingleThreadFetcherManager<E,​SplitT> - -
        • -
        -
      • -
      -
    • -
    -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-use.html deleted file mode 100644 index 4513815309..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/adapter/package-use.html +++ /dev/null @@ -1,209 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.adapter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.adapter

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalog.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalog.html deleted file mode 100644 index 2306ae203e..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalog.html +++ /dev/null @@ -1,1463 +0,0 @@ - - - - - -FlinkCatalog (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkCatalog

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.flink.table.catalog.AbstractCatalog
      • -
      • -
          -
        • org.apache.fluss.flink.catalog.FlinkCatalog
        • -
        -
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.table.catalog.Catalog
      -
      -
      -
      public class FlinkCatalog
      -extends org.apache.flink.table.catalog.AbstractCatalog
      -
      A Flink Catalog for fluss. - -

      Currently, this class must extend the internal Flink class AbstractCatalog because an - incompatibility bug ( FLINK-38030) in flink 2.0.0. - -

      TODO: Once this issue is resolved in a future version of Flink (likely 2.1+), refactor this - class to implement the public interface Catalog instead of - extending the internal class AbstractCatalog.

      -
    • -
    -
    -
    -
      -
    • - -
      - -
      - -
      - -
      - -
      -
        -
      • - - -

        Method Summary

        - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
        All Methods Instance Methods Concrete Methods 
        Modifier and TypeMethodDescription
        voidalterDatabase​(String databaseName, - org.apache.flink.table.catalog.CatalogDatabase catalogDatabase, - boolean b) 
        voidalterFunction​(org.apache.flink.table.catalog.ObjectPath objectPath, - org.apache.flink.table.catalog.CatalogFunction catalogFunction, - boolean b) 
        voidalterPartition​(org.apache.flink.table.catalog.ObjectPath objectPath, - org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec, - org.apache.flink.table.catalog.CatalogPartition catalogPartition, - boolean b) 
        voidalterPartitionColumnStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath, - org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec, - org.apache.flink.table.catalog.stats.CatalogColumnStatistics catalogColumnStatistics, - boolean b) 
        voidalterPartitionStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath, - org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec, - org.apache.flink.table.catalog.stats.CatalogTableStatistics catalogTableStatistics, - boolean b) 
        voidalterTable​(org.apache.flink.table.catalog.ObjectPath objectPath, - org.apache.flink.table.catalog.CatalogBaseTable catalogBaseTable, - boolean b) 
        voidalterTableColumnStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath, - org.apache.flink.table.catalog.stats.CatalogColumnStatistics catalogColumnStatistics, - boolean b) 
        voidalterTableStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath, - org.apache.flink.table.catalog.stats.CatalogTableStatistics catalogTableStatistics, - boolean b) 
        voidclose() 
        voidcreateDatabase​(String databaseName, - org.apache.flink.table.catalog.CatalogDatabase database, - boolean ignoreIfExists) 
        voidcreateFunction​(org.apache.flink.table.catalog.ObjectPath objectPath, - org.apache.flink.table.catalog.CatalogFunction catalogFunction, - boolean b) 
        voidcreatePartition​(org.apache.flink.table.catalog.ObjectPath objectPath, - org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec, - org.apache.flink.table.catalog.CatalogPartition catalogPartition, - boolean b) 
        voidcreateTable​(org.apache.flink.table.catalog.ObjectPath objectPath, - org.apache.flink.table.catalog.CatalogBaseTable table, - boolean ignoreIfExist) 
        booleandatabaseExists​(String databaseName) 
        voiddropDatabase​(String databaseName, - boolean ignoreIfNotExists, - boolean cascade) 
        voiddropFunction​(org.apache.flink.table.catalog.ObjectPath objectPath, - boolean b) 
        voiddropPartition​(org.apache.flink.table.catalog.ObjectPath objectPath, - org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec, - boolean b) 
        voiddropTable​(org.apache.flink.table.catalog.ObjectPath objectPath, - boolean ignoreIfNotExists) 
        booleanfunctionExists​(org.apache.flink.table.catalog.ObjectPath objectPath) 
        org.apache.flink.table.catalog.CatalogDatabasegetDatabase​(String databaseName) 
        Optional<org.apache.flink.table.factories.Factory>getFactory() 
        org.apache.flink.table.catalog.CatalogFunctiongetFunction​(org.apache.flink.table.catalog.ObjectPath functionPath) 
        protected org.apache.flink.table.catalog.CatalogBaseTablegetLakeTable​(String databaseName, - String tableName, - org.apache.fluss.config.Configuration properties) 
        org.apache.flink.table.catalog.CatalogPartitiongetPartition​(org.apache.flink.table.catalog.ObjectPath objectPath, - org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec) 
        org.apache.flink.table.catalog.stats.CatalogColumnStatisticsgetPartitionColumnStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath, - org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec) 
        org.apache.flink.table.catalog.stats.CatalogTableStatisticsgetPartitionStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath, - org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec) 
        org.apache.flink.table.procedures.ProceduregetProcedure​(org.apache.flink.table.catalog.ObjectPath procedurePath) 
        Map<String,​String>getSecurityConfigs() 
        org.apache.flink.table.catalog.CatalogBaseTablegetTable​(org.apache.flink.table.catalog.ObjectPath objectPath) 
        org.apache.flink.table.catalog.stats.CatalogColumnStatisticsgetTableColumnStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath) 
        org.apache.flink.table.catalog.stats.CatalogTableStatisticsgetTableStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath) 
        List<String>listDatabases() 
        List<String>listFunctions​(String s) 
        List<org.apache.flink.table.catalog.CatalogPartitionSpec>listPartitions​(org.apache.flink.table.catalog.ObjectPath objectPath) 
        List<org.apache.flink.table.catalog.CatalogPartitionSpec>listPartitions​(org.apache.flink.table.catalog.ObjectPath objectPath, - org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec) 
        List<org.apache.flink.table.catalog.CatalogPartitionSpec>listPartitionsByFilter​(org.apache.flink.table.catalog.ObjectPath objectPath, - List<org.apache.flink.table.expressions.Expression> list) 
        List<String>listProcedures​(String dbName) 
        List<String>listTables​(String databaseName) 
        List<String>listViews​(String s) 
        voidopen() 
        booleanpartitionExists​(org.apache.flink.table.catalog.ObjectPath objectPath, - org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec) 
        voidrenameTable​(org.apache.flink.table.catalog.ObjectPath objectPath, - String s, - boolean b) 
        booleantableExists​(org.apache.flink.table.catalog.ObjectPath objectPath) 
        protected org.apache.fluss.metadata.TablePathtoTablePath​(org.apache.flink.table.catalog.ObjectPath objectPath) 
        -
          -
        • - - -

          Methods inherited from class org.apache.flink.table.catalog.AbstractCatalog

          -getDefaultDatabase, getName
        • -
        - -
          -
        • - - -

          Methods inherited from interface org.apache.flink.table.catalog.Catalog

          -alterTable, bulkGetPartitionColumnStatistics, bulkGetPartitionStatistics, dropDatabase, getFunctionDefinitionFactory, getTable, getTableFactory, supportsManagedTable
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - - - - - -
          -
        • -

          classLoader

          -
          protected final ClassLoader classLoader
          -
        • -
        - - - -
          -
        • -

          catalogName

          -
          protected final String catalogName
          -
        • -
        - - - -
          -
        • -

          defaultDatabase

          -
          protected final String defaultDatabase
          -
        • -
        - - - -
          -
        • -

          bootstrapServers

          -
          protected final String bootstrapServers
          -
        • -
        - - - -
          -
        • -

          connection

          -
          protected Connection connection
          -
        • -
        - - - -
          -
        • -

          admin

          -
          protected Admin admin
          -
        • -
        -
      • -
      -
      - -
      - -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getFactory

          -
          public Optional<org.apache.flink.table.factories.Factory> getFactory()
          -
        • -
        - - - -
          -
        • -

          open

          -
          public void open()
          -          throws org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          close

          -
          public void close()
          -           throws org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          listDatabases

          -
          public List<String> listDatabases()
          -                           throws org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          getDatabase

          -
          public org.apache.flink.table.catalog.CatalogDatabase getDatabase​(String databaseName)
          -                                                           throws org.apache.flink.table.catalog.exceptions.DatabaseNotExistException,
          -                                                                  org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.DatabaseNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          databaseExists

          -
          public boolean databaseExists​(String databaseName)
          -                       throws org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          createDatabase

          -
          public void createDatabase​(String databaseName,
          -                           org.apache.flink.table.catalog.CatalogDatabase database,
          -                           boolean ignoreIfExists)
          -                    throws org.apache.flink.table.catalog.exceptions.DatabaseAlreadyExistException,
          -                           org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.DatabaseAlreadyExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          dropDatabase

          -
          public void dropDatabase​(String databaseName,
          -                         boolean ignoreIfNotExists,
          -                         boolean cascade)
          -                  throws org.apache.flink.table.catalog.exceptions.DatabaseNotExistException,
          -                         org.apache.flink.table.catalog.exceptions.DatabaseNotEmptyException,
          -                         org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.DatabaseNotExistException
          -
          org.apache.flink.table.catalog.exceptions.DatabaseNotEmptyException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          alterDatabase

          -
          public void alterDatabase​(String databaseName,
          -                          org.apache.flink.table.catalog.CatalogDatabase catalogDatabase,
          -                          boolean b)
          -                   throws org.apache.flink.table.catalog.exceptions.DatabaseNotExistException,
          -                          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.DatabaseNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          listTables

          -
          public List<String> listTables​(String databaseName)
          -                        throws org.apache.flink.table.catalog.exceptions.DatabaseNotExistException,
          -                               org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.DatabaseNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          listViews

          -
          public List<String> listViews​(String s)
          -                       throws org.apache.flink.table.catalog.exceptions.DatabaseNotExistException,
          -                              org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.DatabaseNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          getTable

          -
          public org.apache.flink.table.catalog.CatalogBaseTable getTable​(org.apache.flink.table.catalog.ObjectPath objectPath)
          -                                                         throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
          -                                                                org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.TableNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          getLakeTable

          -
          protected org.apache.flink.table.catalog.CatalogBaseTable getLakeTable​(String databaseName,
          -                                                                       String tableName,
          -                                                                       org.apache.fluss.config.Configuration properties)
          -                                                                throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
          -                                                                       org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.TableNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          tableExists

          -
          public boolean tableExists​(org.apache.flink.table.catalog.ObjectPath objectPath)
          -                    throws org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          dropTable

          -
          public void dropTable​(org.apache.flink.table.catalog.ObjectPath objectPath,
          -                      boolean ignoreIfNotExists)
          -               throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
          -                      org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.TableNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          renameTable

          -
          public void renameTable​(org.apache.flink.table.catalog.ObjectPath objectPath,
          -                        String s,
          -                        boolean b)
          -                 throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
          -                        org.apache.flink.table.catalog.exceptions.TableAlreadyExistException,
          -                        org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.TableNotExistException
          -
          org.apache.flink.table.catalog.exceptions.TableAlreadyExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          createTable

          -
          public void createTable​(org.apache.flink.table.catalog.ObjectPath objectPath,
          -                        org.apache.flink.table.catalog.CatalogBaseTable table,
          -                        boolean ignoreIfExist)
          -                 throws org.apache.flink.table.catalog.exceptions.TableAlreadyExistException,
          -                        org.apache.flink.table.catalog.exceptions.DatabaseNotExistException,
          -                        org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.TableAlreadyExistException
          -
          org.apache.flink.table.catalog.exceptions.DatabaseNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          alterTable

          -
          public void alterTable​(org.apache.flink.table.catalog.ObjectPath objectPath,
          -                       org.apache.flink.table.catalog.CatalogBaseTable catalogBaseTable,
          -                       boolean b)
          -                throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
          -                       org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.TableNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          listPartitions

          -
          public List<org.apache.flink.table.catalog.CatalogPartitionSpec> listPartitions​(org.apache.flink.table.catalog.ObjectPath objectPath)
          -                                                                         throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
          -                                                                                org.apache.flink.table.catalog.exceptions.TableNotPartitionedException,
          -                                                                                org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.TableNotExistException
          -
          org.apache.flink.table.catalog.exceptions.TableNotPartitionedException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          listPartitions

          -
          public List<org.apache.flink.table.catalog.CatalogPartitionSpec> listPartitions​(org.apache.flink.table.catalog.ObjectPath objectPath,
          -                                                                                org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec)
          -                                                                         throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
          -                                                                                org.apache.flink.table.catalog.exceptions.TableNotPartitionedException,
          -                                                                                org.apache.flink.table.catalog.exceptions.PartitionSpecInvalidException,
          -                                                                                org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.TableNotExistException
          -
          org.apache.flink.table.catalog.exceptions.TableNotPartitionedException
          -
          org.apache.flink.table.catalog.exceptions.PartitionSpecInvalidException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          listPartitionsByFilter

          -
          public List<org.apache.flink.table.catalog.CatalogPartitionSpec> listPartitionsByFilter​(org.apache.flink.table.catalog.ObjectPath objectPath,
          -                                                                                        List<org.apache.flink.table.expressions.Expression> list)
          -                                                                                 throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
          -                                                                                        org.apache.flink.table.catalog.exceptions.TableNotPartitionedException,
          -                                                                                        org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.TableNotExistException
          -
          org.apache.flink.table.catalog.exceptions.TableNotPartitionedException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          getPartition

          -
          public org.apache.flink.table.catalog.CatalogPartition getPartition​(org.apache.flink.table.catalog.ObjectPath objectPath,
          -                                                                    org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec)
          -                                                             throws org.apache.flink.table.catalog.exceptions.PartitionNotExistException,
          -                                                                    org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.PartitionNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          partitionExists

          -
          public boolean partitionExists​(org.apache.flink.table.catalog.ObjectPath objectPath,
          -                               org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec)
          -                        throws org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          createPartition

          -
          public void createPartition​(org.apache.flink.table.catalog.ObjectPath objectPath,
          -                            org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec,
          -                            org.apache.flink.table.catalog.CatalogPartition catalogPartition,
          -                            boolean b)
          -                     throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
          -                            org.apache.flink.table.catalog.exceptions.TableNotPartitionedException,
          -                            org.apache.flink.table.catalog.exceptions.PartitionSpecInvalidException,
          -                            org.apache.flink.table.catalog.exceptions.PartitionAlreadyExistsException,
          -                            org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.TableNotExistException
          -
          org.apache.flink.table.catalog.exceptions.TableNotPartitionedException
          -
          org.apache.flink.table.catalog.exceptions.PartitionSpecInvalidException
          -
          org.apache.flink.table.catalog.exceptions.PartitionAlreadyExistsException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          dropPartition

          -
          public void dropPartition​(org.apache.flink.table.catalog.ObjectPath objectPath,
          -                          org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec,
          -                          boolean b)
          -                   throws org.apache.flink.table.catalog.exceptions.PartitionNotExistException,
          -                          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.PartitionNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          alterPartition

          -
          public void alterPartition​(org.apache.flink.table.catalog.ObjectPath objectPath,
          -                           org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec,
          -                           org.apache.flink.table.catalog.CatalogPartition catalogPartition,
          -                           boolean b)
          -                    throws org.apache.flink.table.catalog.exceptions.PartitionNotExistException,
          -                           org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.PartitionNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          listFunctions

          -
          public List<String> listFunctions​(String s)
          -                           throws org.apache.flink.table.catalog.exceptions.DatabaseNotExistException,
          -                                  org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.DatabaseNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          getFunction

          -
          public org.apache.flink.table.catalog.CatalogFunction getFunction​(org.apache.flink.table.catalog.ObjectPath functionPath)
          -                                                           throws org.apache.flink.table.catalog.exceptions.FunctionNotExistException,
          -                                                                  org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.FunctionNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          functionExists

          -
          public boolean functionExists​(org.apache.flink.table.catalog.ObjectPath objectPath)
          -                       throws org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          createFunction

          -
          public void createFunction​(org.apache.flink.table.catalog.ObjectPath objectPath,
          -                           org.apache.flink.table.catalog.CatalogFunction catalogFunction,
          -                           boolean b)
          -                    throws org.apache.flink.table.catalog.exceptions.FunctionAlreadyExistException,
          -                           org.apache.flink.table.catalog.exceptions.DatabaseNotExistException,
          -                           org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.FunctionAlreadyExistException
          -
          org.apache.flink.table.catalog.exceptions.DatabaseNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          alterFunction

          -
          public void alterFunction​(org.apache.flink.table.catalog.ObjectPath objectPath,
          -                          org.apache.flink.table.catalog.CatalogFunction catalogFunction,
          -                          boolean b)
          -                   throws org.apache.flink.table.catalog.exceptions.FunctionNotExistException,
          -                          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.FunctionNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          dropFunction

          -
          public void dropFunction​(org.apache.flink.table.catalog.ObjectPath objectPath,
          -                         boolean b)
          -                  throws org.apache.flink.table.catalog.exceptions.FunctionNotExistException,
          -                         org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.FunctionNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          getTableStatistics

          -
          public org.apache.flink.table.catalog.stats.CatalogTableStatistics getTableStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath)
          -                                                                               throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
          -                                                                                      org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.TableNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          getTableColumnStatistics

          -
          public org.apache.flink.table.catalog.stats.CatalogColumnStatistics getTableColumnStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath)
          -                                                                                      throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
          -                                                                                             org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.TableNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          getPartitionStatistics

          -
          public org.apache.flink.table.catalog.stats.CatalogTableStatistics getPartitionStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath,
          -                                                                                          org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec)
          -                                                                                   throws org.apache.flink.table.catalog.exceptions.PartitionNotExistException,
          -                                                                                          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.PartitionNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          getPartitionColumnStatistics

          -
          public org.apache.flink.table.catalog.stats.CatalogColumnStatistics getPartitionColumnStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath,
          -                                                                                                 org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec)
          -                                                                                          throws org.apache.flink.table.catalog.exceptions.PartitionNotExistException,
          -                                                                                                 org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.PartitionNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          alterTableStatistics

          -
          public void alterTableStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath,
          -                                 org.apache.flink.table.catalog.stats.CatalogTableStatistics catalogTableStatistics,
          -                                 boolean b)
          -                          throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
          -                                 org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.TableNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          alterTableColumnStatistics

          -
          public void alterTableColumnStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath,
          -                                       org.apache.flink.table.catalog.stats.CatalogColumnStatistics catalogColumnStatistics,
          -                                       boolean b)
          -                                throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
          -                                       org.apache.flink.table.catalog.exceptions.CatalogException,
          -                                       org.apache.flink.table.catalog.exceptions.TablePartitionedException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.TableNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          org.apache.flink.table.catalog.exceptions.TablePartitionedException
          -
          -
        • -
        - - - -
          -
        • -

          alterPartitionStatistics

          -
          public void alterPartitionStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath,
          -                                     org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec,
          -                                     org.apache.flink.table.catalog.stats.CatalogTableStatistics catalogTableStatistics,
          -                                     boolean b)
          -                              throws org.apache.flink.table.catalog.exceptions.PartitionNotExistException,
          -                                     org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.PartitionNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          alterPartitionColumnStatistics

          -
          public void alterPartitionColumnStatistics​(org.apache.flink.table.catalog.ObjectPath objectPath,
          -                                           org.apache.flink.table.catalog.CatalogPartitionSpec catalogPartitionSpec,
          -                                           org.apache.flink.table.catalog.stats.CatalogColumnStatistics catalogColumnStatistics,
          -                                           boolean b)
          -                                    throws org.apache.flink.table.catalog.exceptions.PartitionNotExistException,
          -                                           org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.PartitionNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          toTablePath

          -
          protected org.apache.fluss.metadata.TablePath toTablePath​(org.apache.flink.table.catalog.ObjectPath objectPath)
          -
        • -
        - - - -
          -
        • -

          listProcedures

          -
          public List<String> listProcedures​(String dbName)
          -                            throws org.apache.flink.table.catalog.exceptions.DatabaseNotExistException,
          -                                   org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.DatabaseNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          getProcedure

          -
          public org.apache.flink.table.procedures.Procedure getProcedure​(org.apache.flink.table.catalog.ObjectPath procedurePath)
          -                                                         throws org.apache.flink.table.catalog.exceptions.ProcedureNotExistException,
          -                                                                org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.ProcedureNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        - - - -
          -
        • -

          getSecurityConfigs

          -
          @VisibleForTesting
          -public Map<String,​String> getSecurityConfigs()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalogFactory.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalogFactory.html deleted file mode 100644 index 10fa409d9d..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalogFactory.html +++ /dev/null @@ -1,446 +0,0 @@ - - - - - -FlinkCatalogFactory (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkCatalogFactory

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.catalog.FlinkCatalogFactory
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.table.factories.CatalogFactory, org.apache.flink.table.factories.Factory, org.apache.flink.table.factories.TableFactory
      -
      -
      -
      public class FlinkCatalogFactory
      -extends Object
      -implements org.apache.flink.table.factories.CatalogFactory
      -
      Factory for FlinkCatalog.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      - -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlinkCatalogFactory

          -
          public FlinkCatalogFactory()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          factoryIdentifier

          -
          public String factoryIdentifier()
          -
          -
          Specified by:
          -
          factoryIdentifier in interface org.apache.flink.table.factories.CatalogFactory
          -
          Specified by:
          -
          factoryIdentifier in interface org.apache.flink.table.factories.Factory
          -
          -
        • -
        - - - -
          -
        • -

          requiredOptions

          -
          public Set<org.apache.flink.configuration.ConfigOption<?>> requiredOptions()
          -
          -
          Specified by:
          -
          requiredOptions in interface org.apache.flink.table.factories.CatalogFactory
          -
          Specified by:
          -
          requiredOptions in interface org.apache.flink.table.factories.Factory
          -
          -
        • -
        - - - -
          -
        • -

          optionalOptions

          -
          public Set<org.apache.flink.configuration.ConfigOption<?>> optionalOptions()
          -
          -
          Specified by:
          -
          optionalOptions in interface org.apache.flink.table.factories.CatalogFactory
          -
          Specified by:
          -
          optionalOptions in interface org.apache.flink.table.factories.Factory
          -
          -
        • -
        - - - -
          -
        • -

          createCatalog

          -
          public FlinkCatalog createCatalog​(org.apache.flink.table.factories.CatalogFactory.Context context)
          -
          -
          Specified by:
          -
          createCatalog in interface org.apache.flink.table.factories.CatalogFactory
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalogOptions.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalogOptions.html deleted file mode 100644 index 43470494bc..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkCatalogOptions.html +++ /dev/null @@ -1,271 +0,0 @@ - - - - - -FlinkCatalogOptions (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkCatalogOptions

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.catalog.FlinkCatalogOptions
      • -
      -
    • -
    -
    -
      -
    • -
      -
      public class FlinkCatalogOptions
      -extends Object
      -
      Options for flink catalog.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - -
          -
        • -

          DEFAULT_DATABASE

          -
          public static final org.apache.flink.configuration.ConfigOption<String> DEFAULT_DATABASE
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkTableFactory.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkTableFactory.html deleted file mode 100644 index 3acc063ab3..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/FlinkTableFactory.html +++ /dev/null @@ -1,412 +0,0 @@ - - - - - -FlinkTableFactory (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkTableFactory

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.table.factories.DynamicTableFactory, org.apache.flink.table.factories.DynamicTableSinkFactory, org.apache.flink.table.factories.DynamicTableSourceFactory, org.apache.flink.table.factories.Factory
      -
      -
      -
      public class FlinkTableFactory
      -extends Object
      -implements org.apache.flink.table.factories.DynamicTableSourceFactory, org.apache.flink.table.factories.DynamicTableSinkFactory
      -
      Factory to create table source and table sink for Fluss.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlinkTableFactory

          -
          public FlinkTableFactory()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          createDynamicTableSource

          -
          public org.apache.flink.table.connector.source.DynamicTableSource createDynamicTableSource​(org.apache.flink.table.factories.DynamicTableFactory.Context context)
          -
          -
          Specified by:
          -
          createDynamicTableSource in interface org.apache.flink.table.factories.DynamicTableSourceFactory
          -
          -
        • -
        - - - -
          -
        • -

          createDynamicTableSink

          -
          public org.apache.flink.table.connector.sink.DynamicTableSink createDynamicTableSink​(org.apache.flink.table.factories.DynamicTableFactory.Context context)
          -
          -
          Specified by:
          -
          createDynamicTableSink in interface org.apache.flink.table.factories.DynamicTableSinkFactory
          -
          -
        • -
        - - - -
          -
        • -

          factoryIdentifier

          -
          public String factoryIdentifier()
          -
          -
          Specified by:
          -
          factoryIdentifier in interface org.apache.flink.table.factories.Factory
          -
          -
        • -
        - - - -
          -
        • -

          requiredOptions

          -
          public Set<org.apache.flink.configuration.ConfigOption<?>> requiredOptions()
          -
          -
          Specified by:
          -
          requiredOptions in interface org.apache.flink.table.factories.Factory
          -
          -
        • -
        - - - -
          -
        • -

          optionalOptions

          -
          public Set<org.apache.flink.configuration.ConfigOption<?>> optionalOptions()
          -
          -
          Specified by:
          -
          optionalOptions in interface org.apache.flink.table.factories.Factory
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalog.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalog.html deleted file mode 100644 index 030fa1be56..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalog.html +++ /dev/null @@ -1,192 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.catalog.FlinkCatalog (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.catalog.FlinkCatalog

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalogFactory.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalogFactory.html deleted file mode 100644 index 2fcddfb73f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalogFactory.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.catalog.FlinkCatalogFactory (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.catalog.FlinkCatalogFactory

    -
    -
    No usage of org.apache.fluss.flink.catalog.FlinkCatalogFactory
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalogOptions.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalogOptions.html deleted file mode 100644 index 5c770b2e96..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkCatalogOptions.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.catalog.FlinkCatalogOptions (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.catalog.FlinkCatalogOptions

    -
    -
    No usage of org.apache.fluss.flink.catalog.FlinkCatalogOptions
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkTableFactory.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkTableFactory.html deleted file mode 100644 index 49ec529ce6..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/class-use/FlinkTableFactory.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.catalog.FlinkTableFactory (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.catalog.FlinkTableFactory

    -
    -
    No usage of org.apache.fluss.flink.catalog.FlinkTableFactory
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-summary.html deleted file mode 100644 index ce294770f5..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-summary.html +++ /dev/null @@ -1,186 +0,0 @@ - - - - - -org.apache.fluss.flink.catalog (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.catalog

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-tree.html deleted file mode 100644 index 06a58001b7..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-tree.html +++ /dev/null @@ -1,170 +0,0 @@ - - - - - -org.apache.fluss.flink.catalog Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.catalog

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    -
      -
    • java.lang.Object -
        -
      • org.apache.flink.table.catalog.AbstractCatalog (implements org.apache.flink.table.catalog.Catalog) - -
      • -
      • org.apache.fluss.flink.catalog.FlinkCatalogFactory (implements org.apache.flink.table.factories.CatalogFactory)
      • -
      • org.apache.fluss.flink.catalog.FlinkCatalogOptions
      • -
      • org.apache.fluss.flink.catalog.FlinkTableFactory (implements org.apache.flink.table.factories.DynamicTableSinkFactory, org.apache.flink.table.factories.DynamicTableSourceFactory)
      • -
      -
    • -
    -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-use.html deleted file mode 100644 index d252449fe8..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/catalog/package-use.html +++ /dev/null @@ -1,185 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.catalog (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.catalog

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/class-use/FlinkConnectorOptions.ScanStartupMode.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/class-use/FlinkConnectorOptions.ScanStartupMode.html deleted file mode 100644 index 713da7e1ec..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/class-use/FlinkConnectorOptions.ScanStartupMode.html +++ /dev/null @@ -1,243 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.FlinkConnectorOptions.ScanStartupMode (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.FlinkConnectorOptions.ScanStartupMode

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/class-use/FlinkConnectorOptions.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/class-use/FlinkConnectorOptions.html deleted file mode 100644 index e67feffe7a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/class-use/FlinkConnectorOptions.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.FlinkConnectorOptions (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.FlinkConnectorOptions

    -
    -
    No usage of org.apache.fluss.flink.FlinkConnectorOptions
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeCatalog.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeCatalog.html deleted file mode 100644 index 1bb2510369..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeCatalog.html +++ /dev/null @@ -1,318 +0,0 @@ - - - - - -LakeCatalog (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LakeCatalog

    -
    -
    - -
    -
      -
    • -
      -
      public class LakeCatalog
      -extends Object
      -
      A lake catalog to delegate the operations on lake table.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      - -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getTable

          -
          public org.apache.flink.table.catalog.CatalogBaseTable getTable​(org.apache.flink.table.catalog.ObjectPath objectPath)
          -                                                         throws org.apache.flink.table.catalog.exceptions.TableNotExistException,
          -                                                                org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
          Throws:
          -
          org.apache.flink.table.catalog.exceptions.TableNotExistException
          -
          org.apache.flink.table.catalog.exceptions.CatalogException
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeRecordRecordEmitter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeRecordRecordEmitter.html deleted file mode 100644 index c1ff70840f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeRecordRecordEmitter.html +++ /dev/null @@ -1,311 +0,0 @@ - - - - - -LakeRecordRecordEmitter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LakeRecordRecordEmitter<OUT>

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.lake.LakeRecordRecordEmitter<OUT>
      • -
      -
    • -
    -
    -
      -
    • -
      -
      public class LakeRecordRecordEmitter<OUT>
      -extends Object
      -
      The emitter to emit record from lake split.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LakeRecordRecordEmitter

          -
          public LakeRecordRecordEmitter​(BiConsumer<ScanRecord,​org.apache.flink.api.connector.source.SourceOutput<OUT>> sourceOutputFunc)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          emitRecord

          -
          public void emitRecord​(SourceSplitState splitState,
          -                       org.apache.flink.api.connector.source.SourceOutput<OUT> sourceOutput,
          -                       RecordAndPos recordAndPos)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitGenerator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitGenerator.html deleted file mode 100644 index ab2438b53d..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitGenerator.html +++ /dev/null @@ -1,330 +0,0 @@ - - - - - -LakeSplitGenerator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LakeSplitGenerator

    -
    -
    - -
    -
      -
    • -
      -
      public class LakeSplitGenerator
      -extends Object
      -
      A generator for lake splits.
      -
    • -
    -
    -
    - -
    -
    - -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitReaderGenerator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitReaderGenerator.html deleted file mode 100644 index 217d601269..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitReaderGenerator.html +++ /dev/null @@ -1,329 +0,0 @@ - - - - - -LakeSplitReaderGenerator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LakeSplitReaderGenerator

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.lake.LakeSplitReaderGenerator
      • -
      -
    • -
    -
    -
      -
    • -
      -
      public class LakeSplitReaderGenerator
      -extends Object
      -
      A generator to generate reader for lake split.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LakeSplitReaderGenerator

          -
          public LakeSplitReaderGenerator​(Table table,
          -                                @Nullable
          -                                int[] projectedFields,
          -                                @Nullable
          -                                org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource)
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitSerializer.html deleted file mode 100644 index 39cfbbbbc1..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitSerializer.html +++ /dev/null @@ -1,340 +0,0 @@ - - - - - -LakeSplitSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LakeSplitSerializer

    -
    -
    - -
    -
      -
    • -
      -
      public class LakeSplitSerializer
      -extends Object
      -
      A serializer for lake split.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LakeSplitSerializer

          -
          public LakeSplitSerializer​(org.apache.fluss.lake.serializer.SimpleVersionedSerializer<org.apache.fluss.lake.source.LakeSplit> sourceSplitSerializer)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - - - - - -
          -
        • -

          deserialize

          -
          public SourceSplitBase deserialize​(byte splitKind,
          -                                   org.apache.fluss.metadata.TableBucket tableBucket,
          -                                   @Nullable
          -                                   String partition,
          -                                   org.apache.flink.core.memory.DataInputDeserializer input)
          -                            throws IOException
          -
          -
          Throws:
          -
          IOException
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitStateInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitStateInitializer.html deleted file mode 100644 index 2df0c2c4b2..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeSplitStateInitializer.html +++ /dev/null @@ -1,307 +0,0 @@ - - - - - -LakeSplitStateInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LakeSplitStateInitializer

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.lake.LakeSplitStateInitializer
      • -
      -
    • -
    -
    -
      -
    • -
      -
      public class LakeSplitStateInitializer
      -extends Object
      -
      The state initializer for lake split.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LakeSplitStateInitializer

          -
          public LakeSplitStateInitializer()
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeTableFactory.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeTableFactory.html deleted file mode 100644 index ab55e3d90f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/LakeTableFactory.html +++ /dev/null @@ -1,309 +0,0 @@ - - - - - -LakeTableFactory (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LakeTableFactory

    -
    -
    - -
    -
      -
    • -
      -
      public class LakeTableFactory
      -extends Object
      -
      A factory to create DynamicTableSource for lake table.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LakeTableFactory

          -
          public LakeTableFactory()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          createDynamicTableSource

          -
          public org.apache.flink.table.connector.source.DynamicTableSource createDynamicTableSource​(org.apache.flink.table.factories.DynamicTableFactory.Context context,
          -                                                                                           String tableName)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeCatalog.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeCatalog.html deleted file mode 100644 index f852d878f8..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeCatalog.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.lake.LakeCatalog (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.lake.LakeCatalog

    -
    -
    No usage of org.apache.fluss.flink.lake.LakeCatalog
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeRecordRecordEmitter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeRecordRecordEmitter.html deleted file mode 100644 index 08283101e1..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeRecordRecordEmitter.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.lake.LakeRecordRecordEmitter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.lake.LakeRecordRecordEmitter

    -
    -
    No usage of org.apache.fluss.flink.lake.LakeRecordRecordEmitter
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitGenerator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitGenerator.html deleted file mode 100644 index ea625b4ad0..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitGenerator.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.lake.LakeSplitGenerator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.lake.LakeSplitGenerator

    -
    -
    No usage of org.apache.fluss.flink.lake.LakeSplitGenerator
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitReaderGenerator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitReaderGenerator.html deleted file mode 100644 index 39072450d6..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitReaderGenerator.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.lake.LakeSplitReaderGenerator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.lake.LakeSplitReaderGenerator

    -
    -
    No usage of org.apache.fluss.flink.lake.LakeSplitReaderGenerator
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitSerializer.html deleted file mode 100644 index 89ca9d3711..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitSerializer.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.lake.LakeSplitSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.lake.LakeSplitSerializer

    -
    -
    No usage of org.apache.fluss.flink.lake.LakeSplitSerializer
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitStateInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitStateInitializer.html deleted file mode 100644 index cf9dd9dd1a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeSplitStateInitializer.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.lake.LakeSplitStateInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.lake.LakeSplitStateInitializer

    -
    -
    No usage of org.apache.fluss.flink.lake.LakeSplitStateInitializer
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeTableFactory.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeTableFactory.html deleted file mode 100644 index a0feff2eab..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/class-use/LakeTableFactory.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.lake.LakeTableFactory (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.lake.LakeTableFactory

    -
    -
    No usage of org.apache.fluss.flink.lake.LakeTableFactory
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-summary.html deleted file mode 100644 index 1629a25457..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-summary.html +++ /dev/null @@ -1,204 +0,0 @@ - - - - - -org.apache.fluss.flink.lake (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.lake

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-tree.html deleted file mode 100644 index 433c85bd89..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-tree.html +++ /dev/null @@ -1,169 +0,0 @@ - - - - - -org.apache.fluss.flink.lake Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.lake

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-use.html deleted file mode 100644 index c48882c657..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/package-use.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.lake (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.lake

    -
    -
    No usage of org.apache.fluss.flink.lake
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/IndexedLakeSplitRecordIterator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/IndexedLakeSplitRecordIterator.html deleted file mode 100644 index d726d01215..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/IndexedLakeSplitRecordIterator.html +++ /dev/null @@ -1,393 +0,0 @@ - - - - - -IndexedLakeSplitRecordIterator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class IndexedLakeSplitRecordIterator

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.lake.reader.IndexedLakeSplitRecordIterator
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      Closeable, AutoCloseable, Iterator<org.apache.fluss.row.InternalRow>, org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow>
      -
      -
      -
      public class IndexedLakeSplitRecordIterator
      -extends Object
      -implements org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow>
      -
      An iterator wrapper that converts LogRecord objects to InternalRow objects while tracking the - current LakeSplit index being processed. - -

      This class serves as an adapter between the underlying LogRecord iterator and the InternalRow - interface expected by consumers. It maintains reference to the specific LakeSplit index that is - currently being iterated. - -

      Primary responsibilities: - -

        -
      • Wraps a LogRecord iterator and exposes InternalRow objects -
      • Preserves the index of the LakeSplit being processed -
      • Provides clean resource management through Closeable interface -
      • Maintains iterator semantics for sequential data access -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          IndexedLakeSplitRecordIterator

          -
          public IndexedLakeSplitRecordIterator​(org.apache.fluss.utils.CloseableIterator<org.apache.fluss.record.LogRecord> logRecordIterators,
          -                                      int currentLakeSplitIndex)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getCurrentLakeSplitIndex

          -
          public int getCurrentLakeSplitIndex()
          -
        • -
        - - - -
          -
        • -

          close

          -
          public void close()
          -
          -
          Specified by:
          -
          close in interface AutoCloseable
          -
          Specified by:
          -
          close in interface Closeable
          -
          Specified by:
          -
          close in interface org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow>
          -
          -
        • -
        - - - -
          -
        • -

          hasNext

          -
          public boolean hasNext()
          -
          -
          Specified by:
          -
          hasNext in interface Iterator<org.apache.fluss.row.InternalRow>
          -
          -
        • -
        - - - -
          -
        • -

          next

          -
          public org.apache.fluss.row.InternalRow next()
          -
          -
          Specified by:
          -
          next in interface Iterator<org.apache.fluss.row.InternalRow>
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/KeyValueRow.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/KeyValueRow.html deleted file mode 100644 index 08278b13e8..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/KeyValueRow.html +++ /dev/null @@ -1,339 +0,0 @@ - - - - - -KeyValueRow (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class KeyValueRow

    -
    -
    - -
    -
      -
    • -
      -
      public class KeyValueRow
      -extends Object
      -
      An InternalRow with the key part.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          KeyValueRow

          -
          public KeyValueRow​(int[] keyIndexes,
          -                   org.apache.fluss.row.InternalRow valueRow,
          -                   boolean isDelete)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          isDelete

          -
          public boolean isDelete()
          -
        • -
        - - - -
          -
        • -

          keyRow

          -
          public org.apache.fluss.row.InternalRow keyRow()
          -
        • -
        - - - -
          -
        • -

          valueRow

          -
          public org.apache.fluss.row.InternalRow valueRow()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/LakeSnapshotAndLogSplitScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/LakeSnapshotAndLogSplitScanner.html deleted file mode 100644 index 1d0daa3306..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/LakeSnapshotAndLogSplitScanner.html +++ /dev/null @@ -1,363 +0,0 @@ - - - - - -LakeSnapshotAndLogSplitScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LakeSnapshotAndLogSplitScanner

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.lake.reader.LakeSnapshotAndLogSplitScanner
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LakeSnapshotAndLogSplitScanner

          -
          public LakeSnapshotAndLogSplitScanner​(Table table,
          -                                      org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource,
          -                                      LakeSnapshotAndFlussLogSplit lakeSnapshotAndFlussLogSplit,
          -                                      @Nullable
          -                                      int[] projectedFields)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          pollBatch

          -
          @Nullable
          -public org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow> pollBatch​(Duration timeout)
          -                                                                                     throws IOException
          -
          Description copied from interface: BatchScanner
          -
          Poll one batch records. The method should return null when reaching the end of the input.
          -
          -
          Specified by:
          -
          pollBatch in interface BatchScanner
          -
          Parameters:
          -
          timeout - The maximum time to block (must not be greater than Long.MAX_VALUE - milliseconds)
          -
          Throws:
          -
          IOException
          -
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/LakeSnapshotScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/LakeSnapshotScanner.html deleted file mode 100644 index 7beb53751c..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/LakeSnapshotScanner.html +++ /dev/null @@ -1,358 +0,0 @@ - - - - - -LakeSnapshotScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LakeSnapshotScanner

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.lake.reader.LakeSnapshotScanner
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LakeSnapshotScanner

          -
          public LakeSnapshotScanner​(org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource,
          -                           LakeSnapshotSplit lakeSnapshotSplit)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          pollBatch

          -
          @Nullable
          -public org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow> pollBatch​(Duration timeout)
          -                                                                                     throws IOException
          -
          Description copied from interface: BatchScanner
          -
          Poll one batch records. The method should return null when reaching the end of the input.
          -
          -
          Specified by:
          -
          pollBatch in interface BatchScanner
          -
          Parameters:
          -
          timeout - The maximum time to block (must not be greater than Long.MAX_VALUE - milliseconds)
          -
          Throws:
          -
          IOException
          -
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/SeekableLakeSnapshotSplitScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/SeekableLakeSnapshotSplitScanner.html deleted file mode 100644 index d46ff11ce9..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/SeekableLakeSnapshotSplitScanner.html +++ /dev/null @@ -1,373 +0,0 @@ - - - - - -SeekableLakeSnapshotSplitScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class SeekableLakeSnapshotSplitScanner

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.lake.reader.SeekableLakeSnapshotSplitScanner
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      Closeable, AutoCloseable, BatchScanner
      -
      -
      -
      public class SeekableLakeSnapshotSplitScanner
      -extends Object
      -implements BatchScanner
      -
      A scanner that supports seeking to a specific LakeSplit and reading from that point. - -

      This scanner enables direct positioning to any LakeSplit in the list using its index, and then - sequentially reads data starting from that specific split. It provides fine-grained control over - where to begin the scanning process. - -

      Key capabilities: - -

        -
      • Direct seeking to any LakeSplit by index -
      • Sequential reading starting from the sought split -
      • Precise positioning within the split collection -
      • Resumable scanning from arbitrary positions -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          SeekableLakeSnapshotSplitScanner

          -
          public SeekableLakeSnapshotSplitScanner​(org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource,
          -                                        List<org.apache.fluss.lake.source.LakeSplit> lakeSplits,
          -                                        int currentLakeSplitIndex)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          pollBatch

          -
          @Nullable
          -public org.apache.fluss.utils.CloseableIterator<org.apache.fluss.row.InternalRow> pollBatch​(Duration timeout)
          -                                                                                     throws IOException
          -
          Description copied from interface: BatchScanner
          -
          Poll one batch records. The method should return null when reaching the end of the input.
          -
          -
          Specified by:
          -
          pollBatch in interface BatchScanner
          -
          Parameters:
          -
          timeout - The maximum time to block (must not be greater than Long.MAX_VALUE - milliseconds)
          -
          Throws:
          -
          IOException
          -
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/IndexedLakeSplitRecordIterator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/IndexedLakeSplitRecordIterator.html deleted file mode 100644 index 4516c7f2b3..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/IndexedLakeSplitRecordIterator.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.lake.reader.IndexedLakeSplitRecordIterator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.lake.reader.IndexedLakeSplitRecordIterator

    -
    -
    No usage of org.apache.fluss.flink.lake.reader.IndexedLakeSplitRecordIterator
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/KeyValueRow.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/KeyValueRow.html deleted file mode 100644 index a9c2f25fb0..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/KeyValueRow.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.lake.reader.KeyValueRow (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.lake.reader.KeyValueRow

    -
    -
    No usage of org.apache.fluss.flink.lake.reader.KeyValueRow
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/LakeSnapshotAndLogSplitScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/LakeSnapshotAndLogSplitScanner.html deleted file mode 100644 index 34e03e956e..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/LakeSnapshotAndLogSplitScanner.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.lake.reader.LakeSnapshotAndLogSplitScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.lake.reader.LakeSnapshotAndLogSplitScanner

    -
    -
    No usage of org.apache.fluss.flink.lake.reader.LakeSnapshotAndLogSplitScanner
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/LakeSnapshotScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/LakeSnapshotScanner.html deleted file mode 100644 index 5e0c53a699..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/LakeSnapshotScanner.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.lake.reader.LakeSnapshotScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.lake.reader.LakeSnapshotScanner

    -
    -
    No usage of org.apache.fluss.flink.lake.reader.LakeSnapshotScanner
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/SeekableLakeSnapshotSplitScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/SeekableLakeSnapshotSplitScanner.html deleted file mode 100644 index b7a6d1173f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/class-use/SeekableLakeSnapshotSplitScanner.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.lake.reader.SeekableLakeSnapshotSplitScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.lake.reader.SeekableLakeSnapshotSplitScanner

    -
    -
    No usage of org.apache.fluss.flink.lake.reader.SeekableLakeSnapshotSplitScanner
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-summary.html deleted file mode 100644 index 76c18f9c2b..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-summary.html +++ /dev/null @@ -1,193 +0,0 @@ - - - - - -org.apache.fluss.flink.lake.reader (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.lake.reader

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-tree.html deleted file mode 100644 index 62e38a9240..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-tree.html +++ /dev/null @@ -1,167 +0,0 @@ - - - - - -org.apache.fluss.flink.lake.reader Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.lake.reader

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-use.html deleted file mode 100644 index cb01e963c5..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/reader/package-use.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.lake.reader (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.lake.reader

    -
    -
    No usage of org.apache.fluss.flink.lake.reader
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/LakeSnapshotAndFlussLogSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/LakeSnapshotAndFlussLogSplit.html deleted file mode 100644 index 645fdc89da..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/LakeSnapshotAndFlussLogSplit.html +++ /dev/null @@ -1,624 +0,0 @@ - - - - - -LakeSnapshotAndFlussLogSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LakeSnapshotAndFlussLogSplit

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.api.connector.source.SourceSplit
      -
      -
      -
      public class LakeSnapshotAndFlussLogSplit
      -extends SourceSplitBase
      -
      A split mixing Lake snapshot and Fluss log.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - -
          -
        • -

          LAKE_SNAPSHOT_FLUSS_LOG_SPLIT_KIND

          -
          public static final byte LAKE_SNAPSHOT_FLUSS_LOG_SPLIT_KIND
          -
          -
          See Also:
          -
          Constant Field Values
          -
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LakeSnapshotAndFlussLogSplit

          -
          public LakeSnapshotAndFlussLogSplit​(org.apache.fluss.metadata.TableBucket tableBucket,
          -                                    @Nullable
          -                                    String partitionName,
          -                                    @Nullable
          -                                    List<org.apache.fluss.lake.source.LakeSplit> snapshotSplits,
          -                                    long startingOffset,
          -                                    long stoppingOffset)
          -
        • -
        - - - -
          -
        • -

          LakeSnapshotAndFlussLogSplit

          -
          public LakeSnapshotAndFlussLogSplit​(org.apache.fluss.metadata.TableBucket tableBucket,
          -                                    @Nullable
          -                                    String partitionName,
          -                                    @Nullable
          -                                    List<org.apache.fluss.lake.source.LakeSplit> snapshotSplits,
          -                                    long startingOffset,
          -                                    long stoppingOffset,
          -                                    long recordsToSkip,
          -                                    int currentLakeSplitIndex,
          -                                    boolean isLakeSplitFinished)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - - - - - -
          -
        • -

          updateWithCurrentLakeSplitIndex

          -
          public LakeSnapshotAndFlussLogSplit updateWithCurrentLakeSplitIndex​(int currentLakeSplitIndex)
          -
        • -
        - - - - - - - - - - - -
          -
        • -

          getRecordsToSkip

          -
          public long getRecordsToSkip()
          -
        • -
        - - - -
          -
        • -

          getStartingOffset

          -
          public long getStartingOffset()
          -
        • -
        - - - -
          -
        • -

          isLakeSplitFinished

          -
          public boolean isLakeSplitFinished()
          -
        • -
        - - - -
          -
        • -

          getStoppingOffset

          -
          public Optional<Long> getStoppingOffset()
          -
        • -
        - - - - - - - -
          -
        • -

          isStreaming

          -
          public boolean isStreaming()
          -
        • -
        - - - - - - - -
          -
        • -

          splitId

          -
          public String splitId()
          -
        • -
        - - - -
          -
        • -

          getLakeSplits

          -
          @Nullable
          -public List<org.apache.fluss.lake.source.LakeSplit> getLakeSplits()
          -
        • -
        - - - -
          -
        • -

          getCurrentLakeSplitIndex

          -
          public int getCurrentLakeSplitIndex()
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/LakeSnapshotSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/LakeSnapshotSplit.html deleted file mode 100644 index af027ee0df..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/LakeSnapshotSplit.html +++ /dev/null @@ -1,501 +0,0 @@ - - - - - -LakeSnapshotSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LakeSnapshotSplit

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.api.connector.source.SourceSplit
      -
      -
      -
      public class LakeSnapshotSplit
      -extends SourceSplitBase
      -
      A split for reading a snapshot of lake.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - -
          -
        • -

          LAKE_SNAPSHOT_SPLIT_KIND

          -
          public static final byte LAKE_SNAPSHOT_SPLIT_KIND
          -
          -
          See Also:
          -
          Constant Field Values
          -
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LakeSnapshotSplit

          -
          public LakeSnapshotSplit​(org.apache.fluss.metadata.TableBucket tableBucket,
          -                         @Nullable
          -                         String partitionName,
          -                         org.apache.fluss.lake.source.LakeSplit lakeSplit,
          -                         int splitIndex)
          -
        • -
        - - - -
          -
        • -

          LakeSnapshotSplit

          -
          public LakeSnapshotSplit​(org.apache.fluss.metadata.TableBucket tableBucket,
          -                         @Nullable
          -                         String partitionName,
          -                         org.apache.fluss.lake.source.LakeSplit lakeSplit,
          -                         int splitIndex,
          -                         long recordsToSkip)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getLakeSplit

          -
          public org.apache.fluss.lake.source.LakeSplit getLakeSplit()
          -
        • -
        - - - -
          -
        • -

          getRecordsToSkip

          -
          public long getRecordsToSkip()
          -
        • -
        - - - -
          -
        • -

          getSplitIndex

          -
          public int getSplitIndex()
          -
        • -
        - - - -
          -
        • -

          splitId

          -
          public String splitId()
          -
        • -
        - - - - - - - - - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/class-use/LakeSnapshotAndFlussLogSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/class-use/LakeSnapshotAndFlussLogSplit.html deleted file mode 100644 index 71eedad85c..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/class-use/LakeSnapshotAndFlussLogSplit.html +++ /dev/null @@ -1,258 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.lake.split.LakeSnapshotAndFlussLogSplit

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/class-use/LakeSnapshotSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/class-use/LakeSnapshotSplit.html deleted file mode 100644 index 90ef942ece..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/class-use/LakeSnapshotSplit.html +++ /dev/null @@ -1,215 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.lake.split.LakeSnapshotSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.lake.split.LakeSnapshotSplit

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-summary.html deleted file mode 100644 index c3cb71f828..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-summary.html +++ /dev/null @@ -1,174 +0,0 @@ - - - - - -org.apache.fluss.flink.lake.split (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.lake.split

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-tree.html deleted file mode 100644 index 8cc69f148f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-tree.html +++ /dev/null @@ -1,168 +0,0 @@ - - - - - -org.apache.fluss.flink.lake.split Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.lake.split

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-use.html deleted file mode 100644 index cbd3b7b60c..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/split/package-use.html +++ /dev/null @@ -1,243 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.lake.split (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.lake.split

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/LakeSnapshotAndFlussLogSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/LakeSnapshotAndFlussLogSplitState.html deleted file mode 100644 index a6559716cc..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/LakeSnapshotAndFlussLogSplitState.html +++ /dev/null @@ -1,365 +0,0 @@ - - - - - -LakeSnapshotAndFlussLogSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LakeSnapshotAndFlussLogSplitState

    -
    -
    - -
    - -
    -
    - -
    -
    -
      -
    • - -
      - -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          setRecordsToSkip

          -
          public void setRecordsToSkip​(long recordsToSkip)
          -
        • -
        - - - -
          -
        • -

          setCurrentLakeSplitIndex

          -
          public void setCurrentLakeSplitIndex​(int currentLakeSplitIndex)
          -
        • -
        - - - -
          -
        • -

          setNextLogOffset

          -
          public void setNextLogOffset​(long nextOffset)
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/LakeSnapshotSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/LakeSnapshotSplitState.html deleted file mode 100644 index e8af1ac39e..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/LakeSnapshotSplitState.html +++ /dev/null @@ -1,337 +0,0 @@ - - - - - -LakeSnapshotSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LakeSnapshotSplitState

    -
    -
    - -
    - -
    -
    - -
    -
    - -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/class-use/LakeSnapshotAndFlussLogSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/class-use/LakeSnapshotAndFlussLogSplitState.html deleted file mode 100644 index ed62ff9265..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/class-use/LakeSnapshotAndFlussLogSplitState.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.lake.state.LakeSnapshotAndFlussLogSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.lake.state.LakeSnapshotAndFlussLogSplitState

    -
    -
    No usage of org.apache.fluss.flink.lake.state.LakeSnapshotAndFlussLogSplitState
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/class-use/LakeSnapshotSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/class-use/LakeSnapshotSplitState.html deleted file mode 100644 index cc43984171..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/class-use/LakeSnapshotSplitState.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.lake.state.LakeSnapshotSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.lake.state.LakeSnapshotSplitState

    -
    -
    No usage of org.apache.fluss.flink.lake.state.LakeSnapshotSplitState
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-summary.html deleted file mode 100644 index f8d1c51552..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-summary.html +++ /dev/null @@ -1,174 +0,0 @@ - - - - - -org.apache.fluss.flink.lake.state (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.lake.state

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-tree.html deleted file mode 100644 index a24cfa7be7..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-tree.html +++ /dev/null @@ -1,168 +0,0 @@ - - - - - -org.apache.fluss.flink.lake.state Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.lake.state

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-use.html deleted file mode 100644 index b746e9fbd3..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/lake/state/package-use.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.lake.state (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.lake.state

    -
    -
    No usage of org.apache.fluss.flink.lake.state
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkCounter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkCounter.html deleted file mode 100644 index 9138323bba..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkCounter.html +++ /dev/null @@ -1,395 +0,0 @@ - - - - - -FlinkCounter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkCounter

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.metrics.Counter, org.apache.flink.metrics.Metric
      -
      -
      -
      public class FlinkCounter
      -extends Object
      -implements org.apache.flink.metrics.Counter
      -
      An implementation of Flink's Counter which wraps Fluss's Counter.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlinkCounter

          -
          public FlinkCounter​(org.apache.fluss.metrics.Counter wrapped)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          inc

          -
          public void inc()
          -
          -
          Specified by:
          -
          inc in interface org.apache.flink.metrics.Counter
          -
          -
        • -
        - - - -
          -
        • -

          inc

          -
          public void inc​(long n)
          -
          -
          Specified by:
          -
          inc in interface org.apache.flink.metrics.Counter
          -
          -
        • -
        - - - -
          -
        • -

          dec

          -
          public void dec()
          -
          -
          Specified by:
          -
          dec in interface org.apache.flink.metrics.Counter
          -
          -
        • -
        - - - -
          -
        • -

          dec

          -
          public void dec​(long n)
          -
          -
          Specified by:
          -
          dec in interface org.apache.flink.metrics.Counter
          -
          -
        • -
        - - - -
          -
        • -

          getCount

          -
          public long getCount()
          -
          -
          Specified by:
          -
          getCount in interface org.apache.flink.metrics.Counter
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkGauge.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkGauge.html deleted file mode 100644 index 6375c809e5..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkGauge.html +++ /dev/null @@ -1,323 +0,0 @@ - - - - - -FlinkGauge (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkGauge<T>

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.metrics.Gauge<T>, org.apache.flink.metrics.Metric
      -
      -
      -
      public class FlinkGauge<T>
      -extends Object
      -implements org.apache.flink.metrics.Gauge<T>
      -
      An implementation of Flink's Gauge which wraps Fluss's Gauge.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlinkGauge

          -
          public FlinkGauge​(org.apache.fluss.metrics.Gauge<T> wrapped)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getValue

          -
          public T getValue()
          -
          -
          Specified by:
          -
          getValue in interface org.apache.flink.metrics.Gauge<T>
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkHistogram.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkHistogram.html deleted file mode 100644 index fbbf179c37..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkHistogram.html +++ /dev/null @@ -1,359 +0,0 @@ - - - - - -FlinkHistogram (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkHistogram

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.metrics.Histogram, org.apache.flink.metrics.Metric
      -
      -
      -
      public class FlinkHistogram
      -extends Object
      -implements org.apache.flink.metrics.Histogram
      -
      An implementation of Flink's Histogram which wraps Fluss's Histogram.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlinkHistogram

          -
          public FlinkHistogram​(org.apache.fluss.metrics.Histogram wrapped)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          update

          -
          public void update​(long n)
          -
          -
          Specified by:
          -
          update in interface org.apache.flink.metrics.Histogram
          -
          -
        • -
        - - - -
          -
        • -

          getCount

          -
          public long getCount()
          -
          -
          Specified by:
          -
          getCount in interface org.apache.flink.metrics.Histogram
          -
          -
        • -
        - - - -
          -
        • -

          getStatistics

          -
          public org.apache.flink.metrics.HistogramStatistics getStatistics()
          -
          -
          Specified by:
          -
          getStatistics in interface org.apache.flink.metrics.Histogram
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkMeter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkMeter.html deleted file mode 100644 index 840eba1721..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkMeter.html +++ /dev/null @@ -1,377 +0,0 @@ - - - - - -FlinkMeter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkMeter

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.metrics.Meter, org.apache.flink.metrics.Metric
      -
      -
      -
      public class FlinkMeter
      -extends Object
      -implements org.apache.flink.metrics.Meter
      -
      An implementation of Flink's Meter which wraps Fluss's Meter.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlinkMeter

          -
          public FlinkMeter​(org.apache.fluss.metrics.Meter wrapped)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          markEvent

          -
          public void markEvent()
          -
          -
          Specified by:
          -
          markEvent in interface org.apache.flink.metrics.Meter
          -
          -
        • -
        - - - -
          -
        • -

          markEvent

          -
          public void markEvent​(long n)
          -
          -
          Specified by:
          -
          markEvent in interface org.apache.flink.metrics.Meter
          -
          -
        • -
        - - - -
          -
        • -

          getRate

          -
          public double getRate()
          -
          -
          Specified by:
          -
          getRate in interface org.apache.flink.metrics.Meter
          -
          -
        • -
        - - - -
          -
        • -

          getCount

          -
          public long getCount()
          -
          -
          Specified by:
          -
          getCount in interface org.apache.flink.metrics.Meter
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkMetricRegistry.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkMetricRegistry.html deleted file mode 100644 index a011ff7b87..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/FlinkMetricRegistry.html +++ /dev/null @@ -1,477 +0,0 @@ - - - - - -FlinkMetricRegistry (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkMetricRegistry

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.metrics.FlinkMetricRegistry
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      AutoCloseable, org.apache.fluss.metrics.registry.MetricRegistry, org.apache.fluss.utils.AutoCloseableAsync
      -
      -
      -
      public class FlinkMetricRegistry
      -extends Object
      -implements org.apache.fluss.metrics.registry.MetricRegistry
      -
      An implementation of MetricRegistry which registers all metrics into Flink's metric - system. It's mainly used for Fluss client to register its metrics to Flink's metric system. - -

      All metrics of the Fluss client are registered under group "fluss", which is a child group of - OperatorMetricGroup. - -

      For example, the metrics sendLatencyMs will be reported in metric: - "{some_parent_groups}.operator.fluss.client_writer.client_id.sendLatencyMs"

      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      - -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlinkMetricRegistry

          -
          public FlinkMetricRegistry​(org.apache.flink.metrics.MetricGroup flinkOperatorMetricGroup)
          -
        • -
        - - - -
          -
        • -

          FlinkMetricRegistry

          -
          public FlinkMetricRegistry​(org.apache.flink.metrics.MetricGroup flinkOperatorMetricGroup,
          -                           Set<String> exposedMetricNames)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getNumberReporters

          -
          public int getNumberReporters()
          -
          -
          Specified by:
          -
          getNumberReporters in interface org.apache.fluss.metrics.registry.MetricRegistry
          -
          -
        • -
        - - - -
          -
        • -

          register

          -
          public void register​(org.apache.fluss.metrics.Metric metric,
          -                     String metricName,
          -                     org.apache.fluss.metrics.groups.AbstractMetricGroup group)
          -
          -
          Specified by:
          -
          register in interface org.apache.fluss.metrics.registry.MetricRegistry
          -
          -
        • -
        - - - -
          -
        • -

          getFlussMetric

          -
          public org.apache.fluss.metrics.Metric getFlussMetric​(String metricName)
          -
          Exposes the metrics of Fluss metics group for flink.
          -
        • -
        - - - -
          -
        • -

          unregister

          -
          public void unregister​(org.apache.fluss.metrics.Metric metric,
          -                       String metricName,
          -                       org.apache.fluss.metrics.groups.AbstractMetricGroup group)
          -
          -
          Specified by:
          -
          unregister in interface org.apache.fluss.metrics.registry.MetricRegistry
          -
          -
        • -
        - - - -
          -
        • -

          closeAsync

          -
          public CompletableFuture<Void> closeAsync()
          -
          -
          Specified by:
          -
          closeAsync in interface org.apache.fluss.utils.AutoCloseableAsync
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkCounter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkCounter.html deleted file mode 100644 index 4da2cba43b..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkCounter.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.metrics.FlinkCounter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.metrics.FlinkCounter

    -
    -
    No usage of org.apache.fluss.flink.metrics.FlinkCounter
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkGauge.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkGauge.html deleted file mode 100644 index 18ae20f5c8..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkGauge.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.metrics.FlinkGauge (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.metrics.FlinkGauge

    -
    -
    No usage of org.apache.fluss.flink.metrics.FlinkGauge
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkHistogram.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkHistogram.html deleted file mode 100644 index e655003d0d..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkHistogram.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.metrics.FlinkHistogram (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.metrics.FlinkHistogram

    -
    -
    No usage of org.apache.fluss.flink.metrics.FlinkHistogram
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkMeter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkMeter.html deleted file mode 100644 index 2363d21a3e..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkMeter.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.metrics.FlinkMeter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.metrics.FlinkMeter

    -
    -
    No usage of org.apache.fluss.flink.metrics.FlinkMeter
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkMetricRegistry.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkMetricRegistry.html deleted file mode 100644 index 90af203b33..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/class-use/FlinkMetricRegistry.html +++ /dev/null @@ -1,192 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.metrics.FlinkMetricRegistry (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.metrics.FlinkMetricRegistry

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-summary.html deleted file mode 100644 index f56c2ad488..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-summary.html +++ /dev/null @@ -1,193 +0,0 @@ - - - - - -org.apache.fluss.flink.metrics (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.metrics

    -
    -
    -
      -
    • - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
      Class Summary 
      ClassDescription
      FlinkCounter -
      An implementation of Flink's Counter which wraps Fluss's Counter.
      -
      FlinkGauge<T> -
      An implementation of Flink's Gauge which wraps Fluss's Gauge.
      -
      FlinkHistogram -
      An implementation of Flink's Histogram which wraps Fluss's Histogram.
      -
      FlinkMeter -
      An implementation of Flink's Meter which wraps Fluss's Meter.
      -
      FlinkMetricRegistry -
      An implementation of MetricRegistry which registers all metrics into Flink's metric - system.
      -
      -
    • -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-tree.html deleted file mode 100644 index 244a85acd2..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-tree.html +++ /dev/null @@ -1,167 +0,0 @@ - - - - - -org.apache.fluss.flink.metrics Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.metrics

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    -
      -
    • java.lang.Object -
        -
      • org.apache.fluss.flink.metrics.FlinkCounter (implements org.apache.flink.metrics.Counter)
      • -
      • org.apache.fluss.flink.metrics.FlinkGauge<T> (implements org.apache.flink.metrics.Gauge<T>)
      • -
      • org.apache.fluss.flink.metrics.FlinkHistogram (implements org.apache.flink.metrics.Histogram)
      • -
      • org.apache.fluss.flink.metrics.FlinkMeter (implements org.apache.flink.metrics.Meter)
      • -
      • org.apache.fluss.flink.metrics.FlinkMetricRegistry (implements org.apache.fluss.metrics.registry.MetricRegistry)
      • -
      -
    • -
    -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-use.html deleted file mode 100644 index c6b70f1d48..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/metrics/package-use.html +++ /dev/null @@ -1,186 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.metrics (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.metrics

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-summary.html deleted file mode 100644 index acab3a57ef..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-summary.html +++ /dev/null @@ -1,185 +0,0 @@ - - - - - -org.apache.fluss.flink (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-tree.html deleted file mode 100644 index 72302d1b1f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-tree.html +++ /dev/null @@ -1,177 +0,0 @@ - - - - - -org.apache.fluss.flink Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -

    Enum Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-use.html deleted file mode 100644 index 78c86f7bdc..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/package-use.html +++ /dev/null @@ -1,208 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/AbstractAclProcedure.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/AbstractAclProcedure.html deleted file mode 100644 index 7a6158f929..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/AbstractAclProcedure.html +++ /dev/null @@ -1,419 +0,0 @@ - - - - - -AbstractAclProcedure (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class AbstractAclProcedure

    -
    -
    - -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          AbstractAclProcedure

          -
          public AbstractAclProcedure()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          internalCall

          -
          protected String[] internalCall​(@Nullable
          -                                String resource,
          -                                @Nullable
          -                                String permission,
          -                                @Nullable
          -                                String principal,
          -                                @Nullable
          -                                String operation,
          -                                @Nullable
          -                                String host)
          -                         throws Exception
          -
          Invokes the ACL operation as a Flink table procedure. - -

          This method serves as the entry point for executing ACL operations (ADD, DROP, LIST) - through the Flink SQL procedure interface. It delegates execution to the internalCall method.

          -
          -
          Parameters:
          -
          resource - Resource on which the ACL operation applies. The format must be one of: -
            -
          • cluster - cluster level -
          • cluster.db_name - database level -
          • cluster.db_name.table_name - table level -
          -
          permission - Permission type to grant or revoke. Common values include in PermissionType.
          -
          principal - Principal (user or role) to apply the ACL to. Accepts: -
            -
          • ANY -
          • ALL -
          • PrincipalType:PrincipalName, e.g., User:alice -
          -
          operation - Operation type applied on the resource. Common values include in OperationType.
          -
          Returns:
          -
          An array of strings representing the result of the operation: -
            -
          • ["success"] for ADD/DROP operations upon success. -
          • For LIST operations, returns a list of formatted ACL entries as strings. -
          -
          Throws:
          -
          ExecutionException - if an error occurs during the execution of the ACL operation.
          -
          InterruptedException - if the current thread is interrupted while waiting for the - operation to complete.
          -
          Exception
          -
          -
        • -
        - - - -
          -
        • -

          aclOperation

          -
          protected abstract String[] aclOperation​(org.apache.fluss.security.acl.Resource resource,
          -                                         org.apache.fluss.security.acl.PermissionType permission,
          -                                         org.apache.fluss.security.acl.FlussPrincipal flussPrincipal,
          -                                         org.apache.fluss.security.acl.OperationType operationType,
          -                                         String host)
          -                                  throws Exception
          -
          -
          Throws:
          -
          Exception
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/AddAclProcedure.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/AddAclProcedure.html deleted file mode 100644 index 0d1b896e61..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/AddAclProcedure.html +++ /dev/null @@ -1,397 +0,0 @@ - - - - - -AddAclProcedure (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class AddAclProcedure

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.table.procedures.Procedure
      -
      -
      -
      public class AddAclProcedure
      -extends AbstractAclProcedure
      -
      Procedure to add acl.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          AddAclProcedure

          -
          public AddAclProcedure()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - - - - - -
          -
        • -

          aclOperation

          -
          protected String[] aclOperation​(org.apache.fluss.security.acl.Resource resource,
          -                                org.apache.fluss.security.acl.PermissionType permission,
          -                                org.apache.fluss.security.acl.FlussPrincipal flussPrincipal,
          -                                org.apache.fluss.security.acl.OperationType operationType,
          -                                String host)
          -                         throws Exception
          -
          -
          Specified by:
          -
          aclOperation in class AbstractAclProcedure
          -
          Throws:
          -
          Exception
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/DropAclProcedure.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/DropAclProcedure.html deleted file mode 100644 index 3aabc9c04a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/DropAclProcedure.html +++ /dev/null @@ -1,401 +0,0 @@ - - - - - -DropAclProcedure (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class DropAclProcedure

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.table.procedures.Procedure
      -
      -
      -
      public class DropAclProcedure
      -extends AbstractAclProcedure
      -
      Procedure to drop acl.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          DropAclProcedure

          -
          public DropAclProcedure()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          call

          -
          public String[] call​(org.apache.flink.table.procedure.ProcedureContext context,
          -                     @Nullable
          -                     String resource,
          -                     @Nullable
          -                     String permission,
          -                     @Nullable
          -                     String principal,
          -                     @Nullable
          -                     String operation,
          -                     @Nullable
          -                     String host)
          -              throws Exception
          -
          -
          Throws:
          -
          Exception
          -
          -
        • -
        - - - -
          -
        • -

          aclOperation

          -
          protected String[] aclOperation​(org.apache.fluss.security.acl.Resource resource,
          -                                org.apache.fluss.security.acl.PermissionType permission,
          -                                org.apache.fluss.security.acl.FlussPrincipal flussPrincipal,
          -                                org.apache.fluss.security.acl.OperationType operationType,
          -                                String host)
          -                         throws Exception
          -
          -
          Specified by:
          -
          aclOperation in class AbstractAclProcedure
          -
          Throws:
          -
          Exception
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ListAclProcedure.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ListAclProcedure.html deleted file mode 100644 index 38ef0eee0a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ListAclProcedure.html +++ /dev/null @@ -1,401 +0,0 @@ - - - - - -ListAclProcedure (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class ListAclProcedure

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.table.procedures.Procedure
      -
      -
      -
      public class ListAclProcedure
      -extends AbstractAclProcedure
      -
      Procedure to list acl.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          ListAclProcedure

          -
          public ListAclProcedure()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          call

          -
          public String[] call​(org.apache.flink.table.procedure.ProcedureContext context,
          -                     @Nullable
          -                     String resource,
          -                     @Nullable
          -                     String permission,
          -                     @Nullable
          -                     String principal,
          -                     @Nullable
          -                     String operation,
          -                     @Nullable
          -                     String host)
          -              throws Exception
          -
          -
          Throws:
          -
          Exception
          -
          -
        • -
        - - - -
          -
        • -

          aclOperation

          -
          protected String[] aclOperation​(org.apache.fluss.security.acl.Resource resource,
          -                                org.apache.fluss.security.acl.PermissionType permission,
          -                                org.apache.fluss.security.acl.FlussPrincipal flussPrincipal,
          -                                org.apache.fluss.security.acl.OperationType operationType,
          -                                String host)
          -                         throws Exception
          -
          -
          Specified by:
          -
          aclOperation in class AbstractAclProcedure
          -
          Throws:
          -
          Exception
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ProcedureBase.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ProcedureBase.html deleted file mode 100644 index aaa755d9cc..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ProcedureBase.html +++ /dev/null @@ -1,358 +0,0 @@ - - - - - -ProcedureBase (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class ProcedureBase

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.table.procedures.Procedure
      -
      -
      -
      Direct Known Subclasses:
      -
      AbstractAclProcedure
      -
      -
      -
      public abstract class ProcedureBase
      -extends Object
      -implements org.apache.flink.table.procedures.Procedure
      -
      Base implementation for flink Procedure.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - -
          -
        • -

          admin

          -
          protected Admin admin
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          ProcedureBase

          -
          public ProcedureBase()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          withAdmin

          -
          public void withAdmin​(Admin admin)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ProcedureManager.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ProcedureManager.html deleted file mode 100644 index 2d9fec3eaf..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/ProcedureManager.html +++ /dev/null @@ -1,323 +0,0 @@ - - - - - -ProcedureManager (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class ProcedureManager

    -
    -
    - -
    -
      -
    • -
      -
      public class ProcedureManager
      -extends Object
      -
      ProcedureUtil to load procedure.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          ProcedureManager

          -
          public ProcedureManager()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          listProcedures

          -
          public static List<String> listProcedures()
          -
        • -
        - - - -
          -
        • -

          getProcedure

          -
          public static Optional<org.apache.flink.table.procedures.Procedure> getProcedure​(Admin admin,
          -                                                                                 org.apache.flink.table.catalog.ObjectPath procedurePath)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/AbstractAclProcedure.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/AbstractAclProcedure.html deleted file mode 100644 index afab9c8d15..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/AbstractAclProcedure.html +++ /dev/null @@ -1,208 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.procedure.AbstractAclProcedure (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.procedure.AbstractAclProcedure

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/AddAclProcedure.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/AddAclProcedure.html deleted file mode 100644 index 1231d74227..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/AddAclProcedure.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.procedure.AddAclProcedure (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.procedure.AddAclProcedure

    -
    -
    No usage of org.apache.fluss.flink.procedure.AddAclProcedure
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/DropAclProcedure.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/DropAclProcedure.html deleted file mode 100644 index 09eee4cd64..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/DropAclProcedure.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.procedure.DropAclProcedure (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.procedure.DropAclProcedure

    -
    -
    No usage of org.apache.fluss.flink.procedure.DropAclProcedure
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ListAclProcedure.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ListAclProcedure.html deleted file mode 100644 index 4aa65cc577..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ListAclProcedure.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.procedure.ListAclProcedure (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.procedure.ListAclProcedure

    -
    -
    No usage of org.apache.fluss.flink.procedure.ListAclProcedure
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ProcedureBase.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ProcedureBase.html deleted file mode 100644 index 9a10f1c89d..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ProcedureBase.html +++ /dev/null @@ -1,215 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.procedure.ProcedureBase (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.procedure.ProcedureBase

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ProcedureManager.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ProcedureManager.html deleted file mode 100644 index ca411c3606..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/class-use/ProcedureManager.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.procedure.ProcedureManager (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.procedure.ProcedureManager

    -
    -
    No usage of org.apache.fluss.flink.procedure.ProcedureManager
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-summary.html deleted file mode 100644 index 2bec0feff4..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-summary.html +++ /dev/null @@ -1,198 +0,0 @@ - - - - - -org.apache.fluss.flink.procedure (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.procedure

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-tree.html deleted file mode 100644 index 7606c3a71f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-tree.html +++ /dev/null @@ -1,174 +0,0 @@ - - - - - -org.apache.fluss.flink.procedure Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.procedure

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-use.html deleted file mode 100644 index dfecb1472b..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/procedure/package-use.html +++ /dev/null @@ -1,191 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.procedure (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.procedure

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/FlinkAsFlussRow.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/FlinkAsFlussRow.html deleted file mode 100644 index a24da9fd54..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/FlinkAsFlussRow.html +++ /dev/null @@ -1,643 +0,0 @@ - - - - - -FlinkAsFlussRow (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkAsFlussRow

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.fluss.row.InternalRow
      -
      -
      -
      public class FlinkAsFlussRow
      -extends Object
      -implements org.apache.fluss.row.InternalRow
      -
      Wraps a Flink RowData as a Fluss InternalRow.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlinkAsFlussRow

          -
          public FlinkAsFlussRow()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          replace

          -
          public FlinkAsFlussRow replace​(org.apache.flink.table.data.RowData flinkRow)
          -
        • -
        - - - -
          -
        • -

          getFieldCount

          -
          public int getFieldCount()
          -
          -
          Specified by:
          -
          getFieldCount in interface org.apache.fluss.row.InternalRow
          -
          -
        • -
        - - - -
          -
        • -

          isNullAt

          -
          public boolean isNullAt​(int pos)
          -
          -
          Specified by:
          -
          isNullAt in interface org.apache.fluss.row.InternalRow
          -
          -
        • -
        - - - -
          -
        • -

          getBoolean

          -
          public boolean getBoolean​(int pos)
          -
          -
          Specified by:
          -
          getBoolean in interface org.apache.fluss.row.InternalRow
          -
          -
        • -
        - - - -
          -
        • -

          getByte

          -
          public byte getByte​(int pos)
          -
          -
          Specified by:
          -
          getByte in interface org.apache.fluss.row.InternalRow
          -
          -
        • -
        - - - -
          -
        • -

          getShort

          -
          public short getShort​(int pos)
          -
          -
          Specified by:
          -
          getShort in interface org.apache.fluss.row.InternalRow
          -
          -
        • -
        - - - -
          -
        • -

          getInt

          -
          public int getInt​(int pos)
          -
          -
          Specified by:
          -
          getInt in interface org.apache.fluss.row.InternalRow
          -
          -
        • -
        - - - -
          -
        • -

          getLong

          -
          public long getLong​(int pos)
          -
          -
          Specified by:
          -
          getLong in interface org.apache.fluss.row.InternalRow
          -
          -
        • -
        - - - -
          -
        • -

          getFloat

          -
          public float getFloat​(int pos)
          -
          -
          Specified by:
          -
          getFloat in interface org.apache.fluss.row.InternalRow
          -
          -
        • -
        - - - -
          -
        • -

          getDouble

          -
          public double getDouble​(int pos)
          -
          -
          Specified by:
          -
          getDouble in interface org.apache.fluss.row.InternalRow
          -
          -
        • -
        - - - -
          -
        • -

          getChar

          -
          public org.apache.fluss.row.BinaryString getChar​(int pos,
          -                                                 int length)
          -
          -
          Specified by:
          -
          getChar in interface org.apache.fluss.row.InternalRow
          -
          -
        • -
        - - - -
          -
        • -

          getString

          -
          public org.apache.fluss.row.BinaryString getString​(int pos)
          -
          -
          Specified by:
          -
          getString in interface org.apache.fluss.row.InternalRow
          -
          -
        • -
        - - - -
          -
        • -

          getDecimal

          -
          public org.apache.fluss.row.Decimal getDecimal​(int pos,
          -                                               int precision,
          -                                               int scale)
          -
          -
          Specified by:
          -
          getDecimal in interface org.apache.fluss.row.InternalRow
          -
          -
        • -
        - - - -
          -
        • -

          fromFlinkDecimal

          -
          public static org.apache.fluss.row.Decimal fromFlinkDecimal​(org.apache.flink.table.data.DecimalData decimal)
          -
        • -
        - - - -
          -
        • -

          getTimestampNtz

          -
          public org.apache.fluss.row.TimestampNtz getTimestampNtz​(int pos,
          -                                                         int precision)
          -
          -
          Specified by:
          -
          getTimestampNtz in interface org.apache.fluss.row.InternalRow
          -
          -
        • -
        - - - -
          -
        • -

          getTimestampLtz

          -
          public org.apache.fluss.row.TimestampLtz getTimestampLtz​(int pos,
          -                                                         int precision)
          -
          -
          Specified by:
          -
          getTimestampLtz in interface org.apache.fluss.row.InternalRow
          -
          -
        • -
        - - - -
          -
        • -

          getBinary

          -
          public byte[] getBinary​(int pos,
          -                        int length)
          -
          -
          Specified by:
          -
          getBinary in interface org.apache.fluss.row.InternalRow
          -
          -
        • -
        - - - -
          -
        • -

          getBytes

          -
          public byte[] getBytes​(int pos)
          -
          -
          Specified by:
          -
          getBytes in interface org.apache.fluss.row.InternalRow
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/OperationType.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/OperationType.html deleted file mode 100644 index 49bb4d2285..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/OperationType.html +++ /dev/null @@ -1,432 +0,0 @@ - - - - - -OperationType (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Enum OperationType

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      Serializable, Comparable<OperationType>
      -
      -
      -
      public enum OperationType
      -extends Enum<OperationType>
      -
      Enumeration of row operation types used in Fluss-Flink data processing. - -

      This enum represents the type of operation associated with a row, such as an append (insert), - upsert (update or insert), delete, or ignore. It is used to indicate how a row should be - interpreted or processed in downstream systems. - -

        -
      • APPEND - Represents an append-only (insert) operation. -
      • UPSERT - Represents an upsert operation (update or insert). -
      • DELETE - Represents a delete operation. -
      • IGNORE - Represents an operation that should be ignored. -
      -
      -
      See Also:
      -
      RowWithOp
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Enum Constant Detail

        - - - -
          -
        • -

          APPEND

          -
          public static final OperationType APPEND
          -
          Represents an append-only (insert) operation.
          -
        • -
        - - - -
          -
        • -

          UPSERT

          -
          public static final OperationType UPSERT
          -
          Represents an upsert operation (update or insert).
          -
        • -
        - - - -
          -
        • -

          DELETE

          -
          public static final OperationType DELETE
          -
          Represents a delete operation.
          -
        • -
        - - - -
          -
        • -

          IGNORE

          -
          public static final OperationType IGNORE
          -
          Represents an operation that should be ignored.
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          values

          -
          public static OperationType[] values()
          -
          Returns an array containing the constants of this enum type, in -the order they are declared. This method may be used to iterate -over the constants as follows: -
          -for (OperationType c : OperationType.values())
          -    System.out.println(c);
          -
          -
          -
          Returns:
          -
          an array containing the constants of this enum type, in the order they are declared
          -
          -
        • -
        - - - -
          -
        • -

          valueOf

          -
          public static OperationType valueOf​(String name)
          -
          Returns the enum constant of this type with the specified name. -The string must match exactly an identifier used to declare an -enum constant in this type. (Extraneous whitespace characters are -not permitted.)
          -
          -
          Parameters:
          -
          name - the name of the enum constant to be returned.
          -
          Returns:
          -
          the enum constant with the specified name
          -
          Throws:
          -
          IllegalArgumentException - if this enum type has no constant with the specified name
          -
          NullPointerException - if the argument is null
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/RowWithOp.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/RowWithOp.html deleted file mode 100644 index 97e4da77b0..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/RowWithOp.html +++ /dev/null @@ -1,406 +0,0 @@ - - - - - -RowWithOp (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class RowWithOp

    -
    -
    - -
    -
      -
    • -
      -
      public class RowWithOp
      -extends Object
      -
      A wrapper class that associates an InternalRow with an OperationType for use in - Fluss-Flink data processing. - -

      This class is used to represent a row of data along with its corresponding operation type, - such as APPEND, UPSERT, or DELETE, as defined by OperationType.

      -
      -
      See Also:
      -
      InternalRow, -OperationType
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          RowWithOp

          -
          public RowWithOp​(org.apache.fluss.row.InternalRow row,
          -                 @Nullable
          -                 OperationType opType)
          -
          Constructs a RowWithOp with the specified internal row and operation type.
          -
          -
          Parameters:
          -
          row - the internal row data (must not be null)
          -
          opType - the operation type (must not be null)
          -
          Throws:
          -
          NullPointerException - if row or opType is null
          -
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getRow

          -
          public org.apache.fluss.row.InternalRow getRow()
          -
          Returns the internal row data.
          -
          -
          Returns:
          -
          the internal row
          -
          -
        • -
        - - - -
          -
        • -

          getOperationType

          -
          public OperationType getOperationType()
          -
          Returns the operation type associated with this row.
          -
          -
          Returns:
          -
          the operation type
          -
          -
        • -
        - - - -
          -
        • -

          equals

          -
          public boolean equals​(Object o)
          -
          Indicates whether some other object is "equal to" this one. Two RowWithOp objects are - considered equal if their internal rows and operation types are equal.
          -
          -
          Overrides:
          -
          equals in class Object
          -
          Parameters:
          -
          o - the reference object with which to compare
          -
          Returns:
          -
          true if this object is the same as the obj argument; false otherwise
          -
          -
        • -
        - - - -
          -
        • -

          hashCode

          -
          public int hashCode()
          -
          Returns a hash code value for the object.
          -
          -
          Overrides:
          -
          hashCode in class Object
          -
          Returns:
          -
          a hash code value for this object
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/FlinkAsFlussRow.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/FlinkAsFlussRow.html deleted file mode 100644 index 9720e91a3f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/FlinkAsFlussRow.html +++ /dev/null @@ -1,192 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.row.FlinkAsFlussRow (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.row.FlinkAsFlussRow

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/OperationType.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/OperationType.html deleted file mode 100644 index 689f0fb241..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/OperationType.html +++ /dev/null @@ -1,225 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.row.OperationType (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.row.OperationType

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/RowWithOp.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/RowWithOp.html deleted file mode 100644 index f1f9d30a53..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/class-use/RowWithOp.html +++ /dev/null @@ -1,201 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.row.RowWithOp (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.row.RowWithOp

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-summary.html deleted file mode 100644 index 8c2ec0c37b..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-summary.html +++ /dev/null @@ -1,192 +0,0 @@ - - - - - -org.apache.fluss.flink.row (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.row

    -
    -
    -
      -
    • - - - - - - - - - - - - - - - - -
      Class Summary 
      ClassDescription
      FlinkAsFlussRow -
      Wraps a Flink RowData as a Fluss InternalRow.
      -
      RowWithOp -
      A wrapper class that associates an InternalRow with an OperationType for use in - Fluss-Flink data processing.
      -
      -
    • -
    • - - - - - - - - - - - - -
      Enum Summary 
      EnumDescription
      OperationType -
      Enumeration of row operation types used in Fluss-Flink data processing.
      -
      -
    • -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-tree.html deleted file mode 100644 index d4423f8e93..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-tree.html +++ /dev/null @@ -1,178 +0,0 @@ - - - - - -org.apache.fluss.flink.row Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.row

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    -
      -
    • java.lang.Object -
        -
      • org.apache.fluss.flink.row.FlinkAsFlussRow (implements org.apache.fluss.row.InternalRow)
      • -
      • org.apache.fluss.flink.row.RowWithOp
      • -
      -
    • -
    -
    -
    -

    Enum Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-use.html deleted file mode 100644 index 5117a575e5..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/row/package-use.html +++ /dev/null @@ -1,215 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.row (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.row

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/ChannelComputer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/ChannelComputer.html deleted file mode 100644 index 9dd9070b73..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/ChannelComputer.html +++ /dev/null @@ -1,315 +0,0 @@ - - - - - -ChannelComputer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface ChannelComputer<T>

    -
    -
    -
    -
      -
    • -
      -
      Type Parameters:
      -
      T - type of record
      -
      -
      -
      All Superinterfaces:
      -
      Serializable
      -
      -
      -
      All Known Implementing Classes:
      -
      FlinkRowDataChannelComputer
      -
      -
      -
      public interface ChannelComputer<T>
      -extends Serializable
      -
      A utility class to compute which downstream channel a given record should be sent to before flink - sink.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          setup

          -
          void setup​(int numChannels)
          -
        • -
        - - - - - -
          -
        • -

          channel

          -
          int channel​(T record)
          -
        • -
        - - - -
          -
        • -

          select

          -
          static int select​(String partitionName,
          -                  int bucket,
          -                  int numChannels)
          -
        • -
        - - - -
          -
        • -

          select

          -
          static int select​(int bucket,
          -                  int numChannels)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkRowDataChannelComputer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkRowDataChannelComputer.html deleted file mode 100644 index 71bf412e89..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkRowDataChannelComputer.html +++ /dev/null @@ -1,369 +0,0 @@ - - - - - -FlinkRowDataChannelComputer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkRowDataChannelComputer<InputT>

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.sink.FlinkRowDataChannelComputer<InputT>
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    - -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkStreamPartitioner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkStreamPartitioner.html deleted file mode 100644 index 10bddbc51b..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkStreamPartitioner.html +++ /dev/null @@ -1,454 +0,0 @@ - - - - - -FlinkStreamPartitioner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkStreamPartitioner<T>

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.flink.streaming.runtime.partitioner.StreamPartitioner<T>
      • -
      • -
          -
        • org.apache.fluss.flink.sink.FlinkStreamPartitioner<T>
        • -
        -
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      Serializable, org.apache.flink.runtime.io.network.api.writer.ChannelSelector<org.apache.flink.runtime.plugable.SerializationDelegate<org.apache.flink.streaming.runtime.streamrecord.StreamRecord<T>>>
      -
      -
      -
      public class FlinkStreamPartitioner<T>
      -extends org.apache.flink.streaming.runtime.partitioner.StreamPartitioner<T>
      -
      A StreamPartitioner which wraps a ChannelComputer.
      -
      -
      See Also:
      -
      Serialized Form
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Summary

        -
          -
        • - - -

          Fields inherited from class org.apache.flink.streaming.runtime.partitioner.StreamPartitioner

          -numberOfChannels
        • -
        -
      • -
      -
      - -
      - -
      - -
      -
        -
      • - - -

        Method Summary

        - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
        All Methods Static Methods Instance Methods Concrete Methods 
        Modifier and TypeMethodDescription
        org.apache.flink.streaming.runtime.partitioner.StreamPartitioner<T>copy() 
        org.apache.flink.runtime.io.network.api.writer.SubtaskStateMappergetDownstreamSubtaskStateMapper() 
        booleanisPointwise() 
        static <T> org.apache.flink.streaming.api.datastream.DataStream<T>partition​(org.apache.flink.streaming.api.datastream.DataStream<T> input, - ChannelComputer<T> channelComputer, - Integer parallelism) 
        intselectChannel​(org.apache.flink.runtime.plugable.SerializationDelegate<org.apache.flink.streaming.runtime.streamrecord.StreamRecord<T>> record) 
        voidsetup​(int numberOfChannels) 
        StringtoString() 
        -
          -
        • - - -

          Methods inherited from class org.apache.flink.streaming.runtime.partitioner.StreamPartitioner

          -disableUnalignedCheckpoints, equals, getUpstreamSubtaskStateMapper, hashCode, isBroadcast, isSupportsUnalignedCheckpoint
        • -
        - -
      • -
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlinkStreamPartitioner

          -
          public FlinkStreamPartitioner​(ChannelComputer<T> channelComputer)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          setup

          -
          public void setup​(int numberOfChannels)
          -
          -
          Specified by:
          -
          setup in interface org.apache.flink.runtime.io.network.api.writer.ChannelSelector<T>
          -
          Overrides:
          -
          setup in class org.apache.flink.streaming.runtime.partitioner.StreamPartitioner<T>
          -
          -
        • -
        - - - -
          -
        • -

          copy

          -
          public org.apache.flink.streaming.runtime.partitioner.StreamPartitioner<T> copy()
          -
          -
          Specified by:
          -
          copy in class org.apache.flink.streaming.runtime.partitioner.StreamPartitioner<T>
          -
          -
        • -
        - - - -
          -
        • -

          getDownstreamSubtaskStateMapper

          -
          public org.apache.flink.runtime.io.network.api.writer.SubtaskStateMapper getDownstreamSubtaskStateMapper()
          -
          -
          Specified by:
          -
          getDownstreamSubtaskStateMapper in class org.apache.flink.streaming.runtime.partitioner.StreamPartitioner<T>
          -
          -
        • -
        - - - -
          -
        • -

          isPointwise

          -
          public boolean isPointwise()
          -
          -
          Specified by:
          -
          isPointwise in class org.apache.flink.streaming.runtime.partitioner.StreamPartitioner<T>
          -
          -
        • -
        - - - - - - - -
          -
        • -

          selectChannel

          -
          public int selectChannel​(org.apache.flink.runtime.plugable.SerializationDelegate<org.apache.flink.streaming.runtime.streamrecord.StreamRecord<T>> record)
          -
        • -
        - - - -
          -
        • -

          partition

          -
          public static <T> org.apache.flink.streaming.api.datastream.DataStream<T> partition​(org.apache.flink.streaming.api.datastream.DataStream<T> input,
          -                                                                                    ChannelComputer<T> channelComputer,
          -                                                                                    Integer parallelism)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkTableSink.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkTableSink.html deleted file mode 100644 index e4aec6207b..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlinkTableSink.html +++ /dev/null @@ -1,541 +0,0 @@ - - - - - -FlinkTableSink (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkTableSink

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.table.connector.sink.abilities.SupportsDeletePushDown, org.apache.flink.table.connector.sink.abilities.SupportsPartitioning, org.apache.flink.table.connector.sink.abilities.SupportsRowLevelDelete, org.apache.flink.table.connector.sink.abilities.SupportsRowLevelUpdate, org.apache.flink.table.connector.sink.DynamicTableSink
      -
      -
      -
      public class FlinkTableSink
      -extends Object
      -implements org.apache.flink.table.connector.sink.DynamicTableSink, org.apache.flink.table.connector.sink.abilities.SupportsPartitioning, org.apache.flink.table.connector.sink.abilities.SupportsDeletePushDown, org.apache.flink.table.connector.sink.abilities.SupportsRowLevelDelete, org.apache.flink.table.connector.sink.abilities.SupportsRowLevelUpdate
      -
      A Flink DynamicTableSink.
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Nested Class Summary

        -
          -
        • - - -

          Nested classes/interfaces inherited from interface org.apache.flink.table.connector.sink.DynamicTableSink

          -org.apache.flink.table.connector.sink.DynamicTableSink.Context, org.apache.flink.table.connector.sink.DynamicTableSink.DataStructureConverter, org.apache.flink.table.connector.sink.DynamicTableSink.SinkRuntimeProvider
        • -
        -
          -
        • - - -

          Nested classes/interfaces inherited from interface org.apache.flink.table.connector.sink.abilities.SupportsRowLevelDelete

          -org.apache.flink.table.connector.sink.abilities.SupportsRowLevelDelete.RowLevelDeleteInfo, org.apache.flink.table.connector.sink.abilities.SupportsRowLevelDelete.RowLevelDeleteMode
        • -
        -
          -
        • - - -

          Nested classes/interfaces inherited from interface org.apache.flink.table.connector.sink.abilities.SupportsRowLevelUpdate

          -org.apache.flink.table.connector.sink.abilities.SupportsRowLevelUpdate.RowLevelUpdateInfo, org.apache.flink.table.connector.sink.abilities.SupportsRowLevelUpdate.RowLevelUpdateMode
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Summary

        - - - - - - - - - - -
        Constructors 
        ConstructorDescription
        FlinkTableSink​(org.apache.fluss.metadata.TablePath tablePath, - org.apache.fluss.config.Configuration flussConfig, - org.apache.flink.table.types.logical.RowType tableRowType, - int[] primaryKeyIndexes, - List<String> partitionKeys, - boolean streaming, - org.apache.fluss.metadata.MergeEngineType mergeEngineType, - org.apache.fluss.metadata.DataLakeFormat lakeFormat, - boolean ignoreDelete, - int numBucket, - List<String> bucketKeys, - boolean shuffleByBucketId) 
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlinkTableSink

          -
          public FlinkTableSink​(org.apache.fluss.metadata.TablePath tablePath,
          -                      org.apache.fluss.config.Configuration flussConfig,
          -                      org.apache.flink.table.types.logical.RowType tableRowType,
          -                      int[] primaryKeyIndexes,
          -                      List<String> partitionKeys,
          -                      boolean streaming,
          -                      @Nullable
          -                      org.apache.fluss.metadata.MergeEngineType mergeEngineType,
          -                      @Nullable
          -                      org.apache.fluss.metadata.DataLakeFormat lakeFormat,
          -                      boolean ignoreDelete,
          -                      int numBucket,
          -                      List<String> bucketKeys,
          -                      boolean shuffleByBucketId)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getChangelogMode

          -
          public org.apache.flink.table.connector.ChangelogMode getChangelogMode​(org.apache.flink.table.connector.ChangelogMode requestedMode)
          -
          -
          Specified by:
          -
          getChangelogMode in interface org.apache.flink.table.connector.sink.DynamicTableSink
          -
          -
        • -
        - - - -
          -
        • -

          getSinkRuntimeProvider

          -
          public org.apache.flink.table.connector.sink.DynamicTableSink.SinkRuntimeProvider getSinkRuntimeProvider​(org.apache.flink.table.connector.sink.DynamicTableSink.Context context)
          -
          -
          Specified by:
          -
          getSinkRuntimeProvider in interface org.apache.flink.table.connector.sink.DynamicTableSink
          -
          -
        • -
        - - - -
          -
        • -

          copy

          -
          public org.apache.flink.table.connector.sink.DynamicTableSink copy()
          -
          -
          Specified by:
          -
          copy in interface org.apache.flink.table.connector.sink.DynamicTableSink
          -
          -
        • -
        - - - -
          -
        • -

          asSummaryString

          -
          public String asSummaryString()
          -
          -
          Specified by:
          -
          asSummaryString in interface org.apache.flink.table.connector.sink.DynamicTableSink
          -
          -
        • -
        - - - -
          -
        • -

          applyStaticPartition

          -
          public void applyStaticPartition​(Map<String,​String> partition)
          -
          -
          Specified by:
          -
          applyStaticPartition in interface org.apache.flink.table.connector.sink.abilities.SupportsPartitioning
          -
          -
        • -
        - - - -
          -
        • -

          applyDeleteFilters

          -
          public boolean applyDeleteFilters​(List<org.apache.flink.table.expressions.ResolvedExpression> filters)
          -
          -
          Specified by:
          -
          applyDeleteFilters in interface org.apache.flink.table.connector.sink.abilities.SupportsDeletePushDown
          -
          -
        • -
        - - - -
          -
        • -

          executeDeletion

          -
          public Optional<Long> executeDeletion()
          -
          -
          Specified by:
          -
          executeDeletion in interface org.apache.flink.table.connector.sink.abilities.SupportsDeletePushDown
          -
          -
        • -
        - - - -
          -
        • -

          applyRowLevelDelete

          -
          public org.apache.flink.table.connector.sink.abilities.SupportsRowLevelDelete.RowLevelDeleteInfo applyRowLevelDelete​(@Nullable
          -                                                                                                                     org.apache.flink.table.connector.RowLevelModificationScanContext rowLevelModificationScanContext)
          -
          -
          Specified by:
          -
          applyRowLevelDelete in interface org.apache.flink.table.connector.sink.abilities.SupportsRowLevelDelete
          -
          -
        • -
        - - - -
          -
        • -

          applyRowLevelUpdate

          -
          public org.apache.flink.table.connector.sink.abilities.SupportsRowLevelUpdate.RowLevelUpdateInfo applyRowLevelUpdate​(List<org.apache.flink.table.catalog.Column> updatedColumns,
          -                                                                                                                     @Nullable
          -                                                                                                                     org.apache.flink.table.connector.RowLevelModificationScanContext rowLevelModificationScanContext)
          -
          -
          Specified by:
          -
          applyRowLevelUpdate in interface org.apache.flink.table.connector.sink.abilities.SupportsRowLevelUpdate
          -
          -
        • -
        - - - -
          -
        • -

          getBucketKeys

          -
          @VisibleForTesting
          -public List<String> getBucketKeys()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlussSink.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlussSink.html deleted file mode 100644 index fc87dbd901..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlussSink.html +++ /dev/null @@ -1,376 +0,0 @@ - - - - - -FlussSink (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlussSink<InputT>

    -
    -
    - -
    -
      -
    • -
      -
      Type Parameters:
      -
      InputT - the type of input elements accepted by the sink
      -
      -
      -
      All Implemented Interfaces:
      -
      Serializable, org.apache.flink.api.connector.sink2.Sink<InputT>, org.apache.flink.streaming.api.connector.sink2.SupportsPreWriteTopology<InputT>
      -
      -
      -
      @PublicEvolving
      -public class FlussSink<InputT>
      -extends Object
      -
      FlussSink is a specialized Flink sink for writing data to Fluss. - -

      This class extends FlinkSink and provides a builder for constructing Fluss sink - instances with custom configurations. It is intended to be used as the main entry point for - integrating Fluss as a sink in Flink data pipelines.

      -
      -
      Since:
      -
      0.7
      -
      See Also:
      -
      Serialized Form
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Nested Class Summary

        -
          -
        • - - -

          Nested classes/interfaces inherited from interface org.apache.flink.api.connector.sink2.Sink

          -org.apache.flink.api.connector.sink2.Sink.InitContext, org.apache.flink.api.connector.sink2.Sink.InitContextWrapper
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          builder

          -
          public static <T> FlussSinkBuilder<T> builder()
          -
          Creates a new FlussSinkBuilder instance for building a FlussSink.
          -
          -
          Type Parameters:
          -
          T - the type of input elements
          -
          Returns:
          -
          a new FlussSinkBuilder instance
          -
          -
        • -
        - - - -
          -
        • -

          createWriter

          -
          @Deprecated
          -public org.apache.flink.api.connector.sink2.SinkWriter<InputT> createWriter​(org.apache.flink.api.connector.sink2.Sink.InitContext context)
          -                                                                     throws IOException
          -
          Deprecated.
          -
          -
          Specified by:
          -
          createWriter in interface org.apache.flink.api.connector.sink2.Sink<InputT>
          -
          Throws:
          -
          IOException
          -
          -
        • -
        - - - -
          -
        • -

          createWriter

          -
          public org.apache.flink.api.connector.sink2.SinkWriter<InputT> createWriter​(org.apache.flink.api.connector.sink2.WriterInitContext context)
          -                                                                     throws IOException
          -
          -
          Specified by:
          -
          createWriter in interface org.apache.flink.api.connector.sink2.Sink<InputT>
          -
          Throws:
          -
          IOException
          -
          -
        • -
        - - - -
          -
        • -

          addPreWriteTopology

          -
          public org.apache.flink.streaming.api.datastream.DataStream<InputT> addPreWriteTopology​(org.apache.flink.streaming.api.datastream.DataStream<InputT> input)
          -
          -
          Specified by:
          -
          addPreWriteTopology in interface org.apache.flink.streaming.api.connector.sink2.SupportsPreWriteTopology<InputT>
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlussSinkBuilder.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlussSinkBuilder.html deleted file mode 100644 index d282dbba89..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/FlussSinkBuilder.html +++ /dev/null @@ -1,455 +0,0 @@ - - - - - -FlussSinkBuilder (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlussSinkBuilder<InputT>

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.sink.FlussSinkBuilder<InputT>
      • -
      -
    • -
    -
    -
      -
    • -
      -
      Type Parameters:
      -
      InputT - > The input type of records to be written to Fluss
      -
      -
      -
      @PublicEvolving
      -public class FlussSinkBuilder<InputT>
      -extends Object
      -
      Builder for creating and configuring Fluss sink connectors for Apache Flink. - -

      The builder supports automatic schema inference from POJO classes using reflection and - provides options for customizing data conversion logic through custom converters. - -

      Example usage: - -

      
      - FlinkSink<Order> sink = new FlussSinkBuilder<Order>()
      -          .setBootstrapServers(bootstrapServers)
      -          .setTable(tableName)
      -          .setDatabase(databaseName)
      -          .setRowType(orderRowType)
      -          .setSerializationSchema(new OrderSerializationSchema())
      -          .build())
      - 
      -
      -
      Since:
      -
      0.7
      -
      -
    • -
    -
    -
    - -
    -
    - -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/ChannelComputer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/ChannelComputer.html deleted file mode 100644 index ed7448d8d5..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/ChannelComputer.html +++ /dev/null @@ -1,224 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.flink.sink.ChannelComputer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.flink.sink.ChannelComputer

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkRowDataChannelComputer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkRowDataChannelComputer.html deleted file mode 100644 index b900c2955e..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkRowDataChannelComputer.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.sink.FlinkRowDataChannelComputer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.sink.FlinkRowDataChannelComputer

    -
    -
    No usage of org.apache.fluss.flink.sink.FlinkRowDataChannelComputer
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkStreamPartitioner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkStreamPartitioner.html deleted file mode 100644 index dc77de29c7..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkStreamPartitioner.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.sink.FlinkStreamPartitioner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.sink.FlinkStreamPartitioner

    -
    -
    No usage of org.apache.fluss.flink.sink.FlinkStreamPartitioner
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkTableSink.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkTableSink.html deleted file mode 100644 index 604643d4b5..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlinkTableSink.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.sink.FlinkTableSink (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.sink.FlinkTableSink

    -
    -
    No usage of org.apache.fluss.flink.sink.FlinkTableSink
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlussSink.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlussSink.html deleted file mode 100644 index 7f59abc3c6..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlussSink.html +++ /dev/null @@ -1,194 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.sink.FlussSink (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.sink.FlussSink

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlussSinkBuilder.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlussSinkBuilder.html deleted file mode 100644 index e00ce2f85d..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/class-use/FlussSinkBuilder.html +++ /dev/null @@ -1,244 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.sink.FlussSinkBuilder (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.sink.FlussSinkBuilder

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-summary.html deleted file mode 100644 index 8b207fc299..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-summary.html +++ /dev/null @@ -1,210 +0,0 @@ - - - - - -org.apache.fluss.flink.sink (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.sink

    -
    -
    -
      -
    • - - - - - - - - - - - - -
      Interface Summary 
      InterfaceDescription
      ChannelComputer<T> -
      A utility class to compute which downstream channel a given record should be sent to before flink - sink.
      -
      -
    • -
    • - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
      Class Summary 
      ClassDescription
      FlinkRowDataChannelComputer<InputT> -
      ChannelComputer for flink RowData.
      -
      FlinkStreamPartitioner<T> -
      A StreamPartitioner which wraps a ChannelComputer.
      -
      FlinkTableSink -
      A Flink DynamicTableSink.
      -
      FlussSink<InputT> -
      FlussSink is a specialized Flink sink for writing data to Fluss.
      -
      FlussSinkBuilder<InputT> -
      Builder for creating and configuring Fluss sink connectors for Apache Flink.
      -
      -
    • -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-tree.html deleted file mode 100644 index 2902cda3a2..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-tree.html +++ /dev/null @@ -1,181 +0,0 @@ - - - - - -org.apache.fluss.flink.sink Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.sink

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    -
      -
    • java.lang.Object -
        -
      • org.apache.fluss.flink.sink.FlinkRowDataChannelComputer<InputT> (implements org.apache.fluss.flink.sink.ChannelComputer<T>)
      • -
      • org.apache.fluss.flink.sink.FlinkTableSink (implements org.apache.flink.table.connector.sink.DynamicTableSink, org.apache.flink.table.connector.sink.abilities.SupportsDeletePushDown, org.apache.flink.table.connector.sink.abilities.SupportsPartitioning, org.apache.flink.table.connector.sink.abilities.SupportsRowLevelDelete, org.apache.flink.table.connector.sink.abilities.SupportsRowLevelUpdate)
      • -
      • org.apache.fluss.flink.sink.FlussSink<InputT>
      • -
      • org.apache.fluss.flink.sink.FlussSinkBuilder<InputT>
      • -
      • org.apache.flink.streaming.runtime.partitioner.StreamPartitioner<T> (implements org.apache.flink.runtime.io.network.api.writer.ChannelSelector<T>, java.io.Serializable) - -
      • -
      -
    • -
    -
    -
    -

    Interface Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-use.html deleted file mode 100644 index a2d6f78228..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/package-use.html +++ /dev/null @@ -1,198 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.sink (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.sink

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/FlussSerializationSchema.InitializationContext.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/FlussSerializationSchema.InitializationContext.html deleted file mode 100644 index baa1d71267..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/FlussSerializationSchema.InitializationContext.html +++ /dev/null @@ -1,272 +0,0 @@ - - - - - -FlussSerializationSchema.InitializationContext (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface FlussSerializationSchema.InitializationContext

    -
    -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getRowSchema

          -
          org.apache.fluss.types.RowType getRowSchema()
          -
          Returns the target row schema.
          -
          -
          Returns:
          -
          The schema of the target row.
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/FlussSerializationSchema.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/FlussSerializationSchema.html deleted file mode 100644 index 3521e0ddc4..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/FlussSerializationSchema.html +++ /dev/null @@ -1,331 +0,0 @@ - - - - - -FlussSerializationSchema (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface FlussSerializationSchema<T>

    -
    -
    -
    -
      -
    • -
      -
      Type Parameters:
      -
      T - The type to be serialized.
      -
      -
      -
      All Superinterfaces:
      -
      Serializable
      -
      -
      -
      All Known Implementing Classes:
      -
      RowDataSerializationSchema
      -
      -
      -
      @PublicEvolving
      -public interface FlussSerializationSchema<T>
      -extends Serializable
      -
      A serialization schema for Fluss.
      -
    • -
    -
    -
    - -
    -
    - -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/RowDataSerializationSchema.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/RowDataSerializationSchema.html deleted file mode 100644 index 536429d799..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/RowDataSerializationSchema.html +++ /dev/null @@ -1,384 +0,0 @@ - - - - - -RowDataSerializationSchema (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class RowDataSerializationSchema

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.sink.serializer.RowDataSerializationSchema
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          RowDataSerializationSchema

          -
          public RowDataSerializationSchema​(boolean isAppendOnly,
          -                                  boolean ignoreDelete)
          -
          Constructs a new RowSerializationSchema.
          -
          -
          Parameters:
          -
          isAppendOnly - whether the schema is append-only (only INSERTs allowed)
          -
          ignoreDelete - whether to ignore DELETE and UPDATE_BEFORE operations
          -
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/SerializerInitContextImpl.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/SerializerInitContextImpl.html deleted file mode 100644 index 423042074a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/SerializerInitContextImpl.html +++ /dev/null @@ -1,322 +0,0 @@ - - - - - -SerializerInitContextImpl (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class SerializerInitContextImpl

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.sink.serializer.SerializerInitContextImpl
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    - -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/FlussSerializationSchema.InitializationContext.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/FlussSerializationSchema.InitializationContext.html deleted file mode 100644 index bd2e20a02c..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/FlussSerializationSchema.InitializationContext.html +++ /dev/null @@ -1,218 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.flink.sink.serializer.FlussSerializationSchema.InitializationContext (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.flink.sink.serializer.FlussSerializationSchema.InitializationContext

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/FlussSerializationSchema.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/FlussSerializationSchema.html deleted file mode 100644 index 8c2adb196f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/FlussSerializationSchema.html +++ /dev/null @@ -1,294 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.flink.sink.serializer.FlussSerializationSchema (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.flink.sink.serializer.FlussSerializationSchema

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/RowDataSerializationSchema.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/RowDataSerializationSchema.html deleted file mode 100644 index 860f84473a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/RowDataSerializationSchema.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.sink.serializer.RowDataSerializationSchema (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.sink.serializer.RowDataSerializationSchema

    -
    -
    No usage of org.apache.fluss.flink.sink.serializer.RowDataSerializationSchema
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/SerializerInitContextImpl.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/SerializerInitContextImpl.html deleted file mode 100644 index 669248f2bc..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/class-use/SerializerInitContextImpl.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.sink.serializer.SerializerInitContextImpl (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.sink.serializer.SerializerInitContextImpl

    -
    -
    No usage of org.apache.fluss.flink.sink.serializer.SerializerInitContextImpl
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-summary.html deleted file mode 100644 index 4c86309875..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-summary.html +++ /dev/null @@ -1,197 +0,0 @@ - - - - - -org.apache.fluss.flink.sink.serializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.sink.serializer

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-tree.html deleted file mode 100644 index cc4e06974d..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-tree.html +++ /dev/null @@ -1,175 +0,0 @@ - - - - - -org.apache.fluss.flink.sink.serializer Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.sink.serializer

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -

    Interface Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-use.html deleted file mode 100644 index 3d62d01fa2..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/serializer/package-use.html +++ /dev/null @@ -1,237 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.sink.serializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.sink.serializer

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/AppendSinkWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/AppendSinkWriter.html deleted file mode 100644 index f731236f7d..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/AppendSinkWriter.html +++ /dev/null @@ -1,399 +0,0 @@ - - - - - -AppendSinkWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class AppendSinkWriter<InputT>

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      AutoCloseable, org.apache.flink.api.connector.sink2.SinkWriter<InputT>
      -
      -
      -
      public class AppendSinkWriter<InputT>
      -extends FlinkSinkWriter<InputT>
      -
      An append only sink writer for fluss log table.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          AppendSinkWriter

          -
          public AppendSinkWriter​(org.apache.fluss.metadata.TablePath tablePath,
          -                        org.apache.fluss.config.Configuration flussConfig,
          -                        org.apache.flink.table.types.logical.RowType tableRowType,
          -                        org.apache.flink.api.common.operators.MailboxExecutor mailboxExecutor,
          -                        FlussSerializationSchema<InputT> serializationSchema)
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/FlinkSinkWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/FlinkSinkWriter.html deleted file mode 100644 index b41947e562..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/FlinkSinkWriter.html +++ /dev/null @@ -1,593 +0,0 @@ - - - - - -FlinkSinkWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkSinkWriter<InputT>

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.sink.writer.FlinkSinkWriter<InputT>
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      AutoCloseable, org.apache.flink.api.connector.sink2.SinkWriter<InputT>
      -
      -
      -
      Direct Known Subclasses:
      -
      AppendSinkWriter, UpsertSinkWriter
      -
      -
      -
      public abstract class FlinkSinkWriter<InputT>
      -extends Object
      -implements org.apache.flink.api.connector.sink2.SinkWriter<InputT>
      -
      Base class for Flink SinkWriter implementations in Fluss.
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Nested Class Summary

        -
          -
        • - - -

          Nested classes/interfaces inherited from interface org.apache.flink.api.connector.sink2.SinkWriter

          -org.apache.flink.api.connector.sink2.SinkWriter.Context
        • -
        -
      • -
      -
      - -
      - -
      - -
      -
        -
      • - - -

        Constructor Summary

        - - - - - - - - - - - - - - -
        Constructors 
        ConstructorDescription
        FlinkSinkWriter​(org.apache.fluss.metadata.TablePath tablePath, - org.apache.fluss.config.Configuration flussConfig, - org.apache.flink.table.types.logical.RowType tableRowType, - int[] targetColumns, - org.apache.flink.api.common.operators.MailboxExecutor mailboxExecutor, - FlussSerializationSchema<InputT> serializationSchema) 
        FlinkSinkWriter​(org.apache.fluss.metadata.TablePath tablePath, - org.apache.fluss.config.Configuration flussConfig, - org.apache.flink.table.types.logical.RowType tableRowType, - org.apache.flink.api.common.operators.MailboxExecutor mailboxExecutor, - FlussSerializationSchema<InputT> serializationSchema) 
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - -
          -
        • -

          LOG

          -
          protected static final org.slf4j.Logger LOG
          -
        • -
        - - - -
          -
        • -

          tableRowType

          -
          protected final org.apache.flink.table.types.logical.RowType tableRowType
          -
        • -
        - - - -
          -
        • -

          targetColumnIndexes

          -
          @Nullable
          -protected final int[] targetColumnIndexes
          -
        • -
        - - - -
          -
        • -

          table

          -
          protected transient Table table
          -
        • -
        - - - - - - - -
          -
        • -

          metricGroup

          -
          protected transient org.apache.flink.metrics.groups.SinkWriterMetricGroup metricGroup
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlinkSinkWriter

          -
          public FlinkSinkWriter​(org.apache.fluss.metadata.TablePath tablePath,
          -                       org.apache.fluss.config.Configuration flussConfig,
          -                       org.apache.flink.table.types.logical.RowType tableRowType,
          -                       org.apache.flink.api.common.operators.MailboxExecutor mailboxExecutor,
          -                       FlussSerializationSchema<InputT> serializationSchema)
          -
        • -
        - - - -
          -
        • -

          FlinkSinkWriter

          -
          public FlinkSinkWriter​(org.apache.fluss.metadata.TablePath tablePath,
          -                       org.apache.fluss.config.Configuration flussConfig,
          -                       org.apache.flink.table.types.logical.RowType tableRowType,
          -                       @Nullable
          -                       int[] targetColumns,
          -                       org.apache.flink.api.common.operators.MailboxExecutor mailboxExecutor,
          -                       FlussSerializationSchema<InputT> serializationSchema)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          initialize

          -
          public void initialize​(org.apache.flink.metrics.groups.SinkWriterMetricGroup metricGroup)
          -
        • -
        - - - -
          -
        • -

          initMetrics

          -
          protected void initMetrics()
          -
        • -
        - - - - - - - - - - - - - - - - - -
          -
        • -

          checkAsyncException

          -
          protected void checkAsyncException()
          -                            throws IOException
          -
          This method should only be invoked in the mailbox thread since the counter is not volatile. - Logic needs to be invoked by write AND flush since we support various semantics.
          -
          -
          Throws:
          -
          IOException
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/UpsertSinkWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/UpsertSinkWriter.html deleted file mode 100644 index c543826c1e..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/UpsertSinkWriter.html +++ /dev/null @@ -1,404 +0,0 @@ - - - - - -UpsertSinkWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class UpsertSinkWriter<InputT>

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      AutoCloseable, org.apache.flink.api.connector.sink2.SinkWriter<InputT>
      -
      -
      -
      public class UpsertSinkWriter<InputT>
      -extends FlinkSinkWriter<InputT>
      -
      An upsert sink writer or fluss primary key table.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          UpsertSinkWriter

          -
          public UpsertSinkWriter​(org.apache.fluss.metadata.TablePath tablePath,
          -                        org.apache.fluss.config.Configuration flussConfig,
          -                        org.apache.flink.table.types.logical.RowType tableRowType,
          -                        @Nullable
          -                        int[] targetColumnIndexes,
          -                        org.apache.flink.api.common.operators.MailboxExecutor mailboxExecutor,
          -                        FlussSerializationSchema<InputT> flussSerializationSchema)
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/AppendSinkWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/AppendSinkWriter.html deleted file mode 100644 index aa1be3168e..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/AppendSinkWriter.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.sink.writer.AppendSinkWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.sink.writer.AppendSinkWriter

    -
    -
    No usage of org.apache.fluss.flink.sink.writer.AppendSinkWriter
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/FlinkSinkWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/FlinkSinkWriter.html deleted file mode 100644 index bb7e8cb4d2..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/FlinkSinkWriter.html +++ /dev/null @@ -1,201 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.sink.writer.FlinkSinkWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.sink.writer.FlinkSinkWriter

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/UpsertSinkWriter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/UpsertSinkWriter.html deleted file mode 100644 index eaa562a56d..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/class-use/UpsertSinkWriter.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.sink.writer.UpsertSinkWriter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.sink.writer.UpsertSinkWriter

    -
    -
    No usage of org.apache.fluss.flink.sink.writer.UpsertSinkWriter
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-summary.html deleted file mode 100644 index dd95f239c9..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-summary.html +++ /dev/null @@ -1,180 +0,0 @@ - - - - - -org.apache.fluss.flink.sink.writer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.sink.writer

    -
    -
    -
      -
    • - - - - - - - - - - - - - - - - - - - - -
      Class Summary 
      ClassDescription
      AppendSinkWriter<InputT> -
      An append only sink writer for fluss log table.
      -
      FlinkSinkWriter<InputT> -
      Base class for Flink SinkWriter implementations in Fluss.
      -
      UpsertSinkWriter<InputT> -
      An upsert sink writer or fluss primary key table.
      -
      -
    • -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-tree.html deleted file mode 100644 index a50e0445c0..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-tree.html +++ /dev/null @@ -1,168 +0,0 @@ - - - - - -org.apache.fluss.flink.sink.writer Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.sink.writer

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    -
      -
    • java.lang.Object -
        -
      • org.apache.fluss.flink.sink.writer.FlinkSinkWriter<InputT> (implements org.apache.flink.api.connector.sink2.SinkWriter<InputT>) - -
      • -
      -
    • -
    -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-use.html deleted file mode 100644 index 25d1f56f15..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/sink/writer/package-use.html +++ /dev/null @@ -1,185 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.sink.writer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.sink.writer

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlinkSource.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlinkSource.html deleted file mode 100644 index 4aef171c5e..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlinkSource.html +++ /dev/null @@ -1,550 +0,0 @@ - - - - - -FlinkSource (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkSource<OUT>

    -
    -
    - -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - - - - - -
          -
        • -

          scanPartitionDiscoveryIntervalMs

          -
          protected final long scanPartitionDiscoveryIntervalMs
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlinkSource

          -
          public FlinkSource​(org.apache.fluss.config.Configuration flussConf,
          -                   org.apache.fluss.metadata.TablePath tablePath,
          -                   boolean hasPrimaryKey,
          -                   boolean isPartitioned,
          -                   org.apache.fluss.types.RowType sourceOutputType,
          -                   @Nullable
          -                   int[] projectedFields,
          -                   OffsetsInitializer offsetsInitializer,
          -                   long scanPartitionDiscoveryIntervalMs,
          -                   FlussDeserializationSchema<OUT> deserializationSchema,
          -                   boolean streaming,
          -                   List<PushdownUtils.FieldEqual> partitionFilters)
          -
        • -
        - - - -
          -
        • -

          FlinkSource

          -
          public FlinkSource​(org.apache.fluss.config.Configuration flussConf,
          -                   org.apache.fluss.metadata.TablePath tablePath,
          -                   boolean hasPrimaryKey,
          -                   boolean isPartitioned,
          -                   org.apache.fluss.types.RowType sourceOutputType,
          -                   @Nullable
          -                   int[] projectedFields,
          -                   OffsetsInitializer offsetsInitializer,
          -                   long scanPartitionDiscoveryIntervalMs,
          -                   FlussDeserializationSchema<OUT> deserializationSchema,
          -                   boolean streaming,
          -                   List<PushdownUtils.FieldEqual> partitionFilters,
          -                   org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getBoundedness

          -
          public org.apache.flink.api.connector.source.Boundedness getBoundedness()
          -
          -
          Specified by:
          -
          getBoundedness in interface org.apache.flink.api.connector.source.Source<OUT,​SourceSplitBase,​SourceEnumeratorState>
          -
          -
        • -
        - - - - - - - - - - - -
          -
        • -

          getSplitSerializer

          -
          public org.apache.flink.core.io.SimpleVersionedSerializer<SourceSplitBase> getSplitSerializer()
          -
          -
          Specified by:
          -
          getSplitSerializer in interface org.apache.flink.api.connector.source.Source<OUT,​SourceSplitBase,​SourceEnumeratorState>
          -
          -
        • -
        - - - -
          -
        • -

          getEnumeratorCheckpointSerializer

          -
          public org.apache.flink.core.io.SimpleVersionedSerializer<SourceEnumeratorState> getEnumeratorCheckpointSerializer()
          -
          -
          Specified by:
          -
          getEnumeratorCheckpointSerializer in interface org.apache.flink.api.connector.source.Source<OUT,​SourceSplitBase,​SourceEnumeratorState>
          -
          -
        • -
        - - - -
          -
        • -

          createReader

          -
          public org.apache.flink.api.connector.source.SourceReader<OUT,​SourceSplitBase> createReader​(org.apache.flink.api.connector.source.SourceReaderContext context)
          -                                                                                           throws Exception
          -
          -
          Specified by:
          -
          createReader in interface org.apache.flink.api.connector.source.SourceReaderFactory<OUT,​SourceSplitBase>
          -
          Throws:
          -
          Exception
          -
          -
        • -
        - - - -
          -
        • -

          getProducedType

          -
          public org.apache.flink.api.common.typeinfo.TypeInformation<OUT> getProducedType()
          -
          -
          Specified by:
          -
          getProducedType in interface org.apache.flink.api.java.typeutils.ResultTypeQueryable<OUT>
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlinkTableSource.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlinkTableSource.html deleted file mode 100644 index fd3bd18574..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlinkTableSource.html +++ /dev/null @@ -1,704 +0,0 @@ - - - - - -FlinkTableSource (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkTableSource

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.table.connector.source.abilities.SupportsAggregatePushDown, org.apache.flink.table.connector.source.abilities.SupportsFilterPushDown, org.apache.flink.table.connector.source.abilities.SupportsLimitPushDown, org.apache.flink.table.connector.source.abilities.SupportsProjectionPushDown, org.apache.flink.table.connector.source.abilities.SupportsRowLevelModificationScan, org.apache.flink.table.connector.source.DynamicTableSource, org.apache.flink.table.connector.source.LookupTableSource, org.apache.flink.table.connector.source.ScanTableSource
      -
      -
      -
      public class FlinkTableSource
      -extends Object
      -implements org.apache.flink.table.connector.source.ScanTableSource, org.apache.flink.table.connector.source.abilities.SupportsProjectionPushDown, org.apache.flink.table.connector.source.abilities.SupportsFilterPushDown, org.apache.flink.table.connector.source.LookupTableSource, org.apache.flink.table.connector.source.abilities.SupportsRowLevelModificationScan, org.apache.flink.table.connector.source.abilities.SupportsLimitPushDown, org.apache.flink.table.connector.source.abilities.SupportsAggregatePushDown
      -
      Flink table source to scan Fluss data.
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Nested Class Summary

        -
          -
        • - - -

          Nested classes/interfaces inherited from interface org.apache.flink.table.connector.source.DynamicTableSource

          -org.apache.flink.table.connector.source.DynamicTableSource.Context, org.apache.flink.table.connector.source.DynamicTableSource.DataStructureConverter
        • -
        -
          -
        • - - -

          Nested classes/interfaces inherited from interface org.apache.flink.table.connector.source.LookupTableSource

          -org.apache.flink.table.connector.source.LookupTableSource.LookupContext, org.apache.flink.table.connector.source.LookupTableSource.LookupRuntimeProvider
        • -
        -
          -
        • - - -

          Nested classes/interfaces inherited from interface org.apache.flink.table.connector.source.ScanTableSource

          -org.apache.flink.table.connector.source.ScanTableSource.ScanContext, org.apache.flink.table.connector.source.ScanTableSource.ScanRuntimeProvider
        • -
        -
          -
        • - - -

          Nested classes/interfaces inherited from interface org.apache.flink.table.connector.source.abilities.SupportsFilterPushDown

          -org.apache.flink.table.connector.source.abilities.SupportsFilterPushDown.Result
        • -
        -
          -
        • - - -

          Nested classes/interfaces inherited from interface org.apache.flink.table.connector.source.abilities.SupportsRowLevelModificationScan

          -org.apache.flink.table.connector.source.abilities.SupportsRowLevelModificationScan.RowLevelModificationType
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Field Summary

        - - - - - - - - - - - - - - - - - -
        Fields 
        Modifier and TypeFieldDescription
        static org.slf4j.LoggerLOG 
        protected booleanselectRowCount 
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Summary

        - - - - - - - - - - -
        Constructors 
        ConstructorDescription
        FlinkTableSource​(org.apache.fluss.metadata.TablePath tablePath, - org.apache.fluss.config.Configuration flussConfig, - org.apache.flink.table.types.logical.RowType tableOutputType, - int[] primaryKeyIndexes, - int[] bucketKeyIndexes, - int[] partitionKeyIndexes, - boolean streaming, - FlinkConnectorOptionsUtils.StartupOptions startupOptions, - int lookupMaxRetryTimes, - boolean lookupAsync, - org.apache.flink.table.connector.source.lookup.cache.LookupCache cache, - long scanPartitionDiscoveryIntervalMs, - boolean isDataLakeEnabled, - org.apache.fluss.metadata.MergeEngineType mergeEngineType, - Map<String,​String> tableOptions) 
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Summary

        - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
        All Methods Instance Methods Concrete Methods 
        Modifier and TypeMethodDescription
        booleanapplyAggregates​(List<int[]> groupingSets, - List<org.apache.flink.table.expressions.AggregateExpression> aggregateExpressions, - org.apache.flink.table.types.DataType dataType) 
        org.apache.flink.table.connector.source.abilities.SupportsFilterPushDown.ResultapplyFilters​(List<org.apache.flink.table.expressions.ResolvedExpression> filters) 
        voidapplyLimit​(long limit) 
        voidapplyProjection​(int[][] projectedFields, - org.apache.flink.table.types.DataType producedDataType) 
        org.apache.flink.table.connector.RowLevelModificationScanContextapplyRowLevelModificationScan​(org.apache.flink.table.connector.source.abilities.SupportsRowLevelModificationScan.RowLevelModificationType rowLevelModificationType, - org.apache.flink.table.connector.RowLevelModificationScanContext rowLevelModificationScanContext) 
        StringasSummaryString() 
        org.apache.flink.table.connector.source.DynamicTableSourcecopy() 
        int[]getBucketKeyIndexes() 
        org.apache.flink.table.connector.source.lookup.cache.LookupCachegetCache() 
        org.apache.flink.table.connector.ChangelogModegetChangelogMode() 
        org.apache.flink.table.connector.source.LookupTableSource.LookupRuntimeProvidergetLookupRuntimeProvider​(org.apache.flink.table.connector.source.LookupTableSource.LookupContext context) 
        int[]getPartitionKeyIndexes() 
        int[]getPrimaryKeyIndexes() 
        org.apache.flink.table.connector.source.ScanTableSource.ScanRuntimeProvidergetScanRuntimeProvider​(org.apache.flink.table.connector.source.ScanTableSource.ScanContext scanContext) 
        booleansupportsNestedProjection() 
        - -
          -
        • - - -

          Methods inherited from interface org.apache.flink.table.connector.source.abilities.SupportsProjectionPushDown

          -applyProjection
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - -
          -
        • -

          LOG

          -
          public static final org.slf4j.Logger LOG
          -
        • -
        - - - -
          -
        • -

          selectRowCount

          -
          protected boolean selectRowCount
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlinkTableSource

          -
          public FlinkTableSource​(org.apache.fluss.metadata.TablePath tablePath,
          -                        org.apache.fluss.config.Configuration flussConfig,
          -                        org.apache.flink.table.types.logical.RowType tableOutputType,
          -                        int[] primaryKeyIndexes,
          -                        int[] bucketKeyIndexes,
          -                        int[] partitionKeyIndexes,
          -                        boolean streaming,
          -                        FlinkConnectorOptionsUtils.StartupOptions startupOptions,
          -                        int lookupMaxRetryTimes,
          -                        boolean lookupAsync,
          -                        @Nullable
          -                        org.apache.flink.table.connector.source.lookup.cache.LookupCache cache,
          -                        long scanPartitionDiscoveryIntervalMs,
          -                        boolean isDataLakeEnabled,
          -                        @Nullable
          -                        org.apache.fluss.metadata.MergeEngineType mergeEngineType,
          -                        Map<String,​String> tableOptions)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getChangelogMode

          -
          public org.apache.flink.table.connector.ChangelogMode getChangelogMode()
          -
          -
          Specified by:
          -
          getChangelogMode in interface org.apache.flink.table.connector.source.ScanTableSource
          -
          -
        • -
        - - - -
          -
        • -

          getScanRuntimeProvider

          -
          public org.apache.flink.table.connector.source.ScanTableSource.ScanRuntimeProvider getScanRuntimeProvider​(org.apache.flink.table.connector.source.ScanTableSource.ScanContext scanContext)
          -
          -
          Specified by:
          -
          getScanRuntimeProvider in interface org.apache.flink.table.connector.source.ScanTableSource
          -
          -
        • -
        - - - -
          -
        • -

          getLookupRuntimeProvider

          -
          public org.apache.flink.table.connector.source.LookupTableSource.LookupRuntimeProvider getLookupRuntimeProvider​(org.apache.flink.table.connector.source.LookupTableSource.LookupContext context)
          -
          -
          Specified by:
          -
          getLookupRuntimeProvider in interface org.apache.flink.table.connector.source.LookupTableSource
          -
          -
        • -
        - - - -
          -
        • -

          copy

          -
          public org.apache.flink.table.connector.source.DynamicTableSource copy()
          -
          -
          Specified by:
          -
          copy in interface org.apache.flink.table.connector.source.DynamicTableSource
          -
          -
        • -
        - - - -
          -
        • -

          asSummaryString

          -
          public String asSummaryString()
          -
          -
          Specified by:
          -
          asSummaryString in interface org.apache.flink.table.connector.source.DynamicTableSource
          -
          -
        • -
        - - - -
          -
        • -

          supportsNestedProjection

          -
          public boolean supportsNestedProjection()
          -
          -
          Specified by:
          -
          supportsNestedProjection in interface org.apache.flink.table.connector.source.abilities.SupportsProjectionPushDown
          -
          -
        • -
        - - - -
          -
        • -

          applyProjection

          -
          public void applyProjection​(int[][] projectedFields,
          -                            org.apache.flink.table.types.DataType producedDataType)
          -
          -
          Specified by:
          -
          applyProjection in interface org.apache.flink.table.connector.source.abilities.SupportsProjectionPushDown
          -
          -
        • -
        - - - -
          -
        • -

          applyFilters

          -
          public org.apache.flink.table.connector.source.abilities.SupportsFilterPushDown.Result applyFilters​(List<org.apache.flink.table.expressions.ResolvedExpression> filters)
          -
          -
          Specified by:
          -
          applyFilters in interface org.apache.flink.table.connector.source.abilities.SupportsFilterPushDown
          -
          -
        • -
        - - - -
          -
        • -

          applyRowLevelModificationScan

          -
          public org.apache.flink.table.connector.RowLevelModificationScanContext applyRowLevelModificationScan​(org.apache.flink.table.connector.source.abilities.SupportsRowLevelModificationScan.RowLevelModificationType rowLevelModificationType,
          -                                                                                                      @Nullable
          -                                                                                                      org.apache.flink.table.connector.RowLevelModificationScanContext rowLevelModificationScanContext)
          -
          -
          Specified by:
          -
          applyRowLevelModificationScan in interface org.apache.flink.table.connector.source.abilities.SupportsRowLevelModificationScan
          -
          -
        • -
        - - - -
          -
        • -

          applyLimit

          -
          public void applyLimit​(long limit)
          -
          -
          Specified by:
          -
          applyLimit in interface org.apache.flink.table.connector.source.abilities.SupportsLimitPushDown
          -
          -
        • -
        - - - -
          -
        • -

          applyAggregates

          -
          public boolean applyAggregates​(List<int[]> groupingSets,
          -                               List<org.apache.flink.table.expressions.AggregateExpression> aggregateExpressions,
          -                               org.apache.flink.table.types.DataType dataType)
          -
          -
          Specified by:
          -
          applyAggregates in interface org.apache.flink.table.connector.source.abilities.SupportsAggregatePushDown
          -
          -
        • -
        - - - -
          -
        • -

          getCache

          -
          @VisibleForTesting
          -@Nullable
          -public org.apache.flink.table.connector.source.lookup.cache.LookupCache getCache()
          -
        • -
        - - - -
          -
        • -

          getPrimaryKeyIndexes

          -
          @VisibleForTesting
          -public int[] getPrimaryKeyIndexes()
          -
        • -
        - - - -
          -
        • -

          getBucketKeyIndexes

          -
          @VisibleForTesting
          -public int[] getBucketKeyIndexes()
          -
        • -
        - - - -
          -
        • -

          getPartitionKeyIndexes

          -
          @VisibleForTesting
          -public int[] getPartitionKeyIndexes()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlussSource.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlussSource.html deleted file mode 100644 index 7dcf087164..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlussSource.html +++ /dev/null @@ -1,338 +0,0 @@ - - - - - -FlussSource (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlussSource<OUT>

    -
    -
    - -
    -
      -
    • -
      -
      Type Parameters:
      -
      OUT - The type of records produced by this source
      -
      -
      -
      All Implemented Interfaces:
      -
      Serializable, org.apache.flink.api.connector.source.Source<OUT,​SourceSplitBase,​SourceEnumeratorState>, org.apache.flink.api.connector.source.SourceReaderFactory<OUT,​SourceSplitBase>, org.apache.flink.api.java.typeutils.ResultTypeQueryable
      -
      -
      -
      public class FlussSource<OUT>
      -extends FlinkSource<OUT>
      -
      A Flink DataStream source implementation for reading data from Fluss tables. - -

      This class extends the FlinkSource base class and implements - ResultTypeQueryable to provide type information for Flink's type system. - -

      Sample usage: - -

      
      - FlussSource<Order> flussSource = FlussSource.<Order>builder()
      -     .setBootstrapServers("localhost:9092")
      -     .setDatabase("mydb")
      -     .setTable("orders")
      -     .setProjectedFields("orderId", "amount")
      -     .setStartingOffsets(OffsetsInitializer.earliest())
      -     .setScanPartitionDiscoveryIntervalMs(1000L)
      -     .setDeserializationSchema(new OrderDeserializationSchema())
      -     .build();
      -
      - DataStreamSource<Order> stream = env.fromSource(
      -     flussSource,
      -     WatermarkStrategy.noWatermarks(),
      -     "Fluss Source"
      - );
      - 
      -
      -
      See Also:
      -
      Serialized Form
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          builder

          -
          public static <T> FlussSourceBuilder<T> builder()
          -
          Get a FlussSourceBuilder to build a FlussSource.
          -
          -
          Returns:
          -
          a Fluss source builder.
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlussSourceBuilder.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlussSourceBuilder.html deleted file mode 100644 index e3e9e95db6..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/FlussSourceBuilder.html +++ /dev/null @@ -1,547 +0,0 @@ - - - - - -FlussSourceBuilder (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlussSourceBuilder<OUT>

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.FlussSourceBuilder<OUT>
      • -
      -
    • -
    -
    -
      -
    • -
      -
      Type Parameters:
      -
      OUT - The type of records produced by the source being built
      -
      -
      -
      public class FlussSourceBuilder<OUT>
      -extends Object
      -
      Builder class for creating FlussSource instances. - -

      The builder allows for step-by-step configuration of a Fluss source connector. It handles the - setup of connection parameters, table metadata retrieval, and source configuration. - -

      Sample usage: - -

      
      - FlussSource<Order> source = FlussSource.<Order>builder()
      -     .setBootstrapServers("localhost:9092")
      -     .setDatabase("mydb")
      -     .setTable("orders")
      -     .setProjectedFields("orderId", "amount")
      -     .setScanPartitionDiscoveryIntervalMs(1000L)
      -     .setStartingOffsets(OffsetsInitializer.earliest())
      -     .setDeserializationSchema(new OrderDeserializationSchema())
      -     .build();
      - 
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlussSourceBuilder

          -
          public FlussSourceBuilder()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          setBootstrapServers

          -
          public FlussSourceBuilder<OUT> setBootstrapServers​(String bootstrapServers)
          -
          Sets the bootstrap servers for the Fluss source connection. - -

          This is a required parameter.

          -
          -
          Parameters:
          -
          bootstrapServers - bootstrap server addresses
          -
          Returns:
          -
          this builder
          -
          -
        • -
        - - - -
          -
        • -

          setDatabase

          -
          public FlussSourceBuilder<OUT> setDatabase​(String database)
          -
          Sets the database name for the Fluss source. - -

          This is a required parameter.

          -
          -
          Parameters:
          -
          database - name of the database
          -
          Returns:
          -
          this builder
          -
          -
        • -
        - - - -
          -
        • -

          setTable

          -
          public FlussSourceBuilder<OUT> setTable​(String table)
          -
          Sets the table name for the Fluss source. - -

          This is a required parameter.

          -
          -
          Parameters:
          -
          table - name of the table
          -
          Returns:
          -
          this builder
          -
          -
        • -
        - - - -
          -
        • -

          setScanPartitionDiscoveryIntervalMs

          -
          public FlussSourceBuilder<OUT> setScanPartitionDiscoveryIntervalMs​(long scanPartitionDiscoveryIntervalMs)
          -
          Sets the scan partition discovery interval in milliseconds. - -

          If not specified, the default value from FlinkConnectorOptions.SCAN_PARTITION_DISCOVERY_INTERVAL is used.

          -
          -
          Parameters:
          -
          scanPartitionDiscoveryIntervalMs - interval in milliseconds
          -
          Returns:
          -
          this builder
          -
          -
        • -
        - - - -
          -
        • -

          setStartingOffsets

          -
          public FlussSourceBuilder<OUT> setStartingOffsets​(OffsetsInitializer offsetsInitializer)
          -
          Sets the starting offsets strategy for the Fluss source. - -

          If not specified, OffsetsInitializer.full() is used by default.

          -
          -
          Parameters:
          -
          offsetsInitializer - the strategy for determining starting offsets
          -
          Returns:
          -
          this builder
          -
          -
        • -
        - - - -
          -
        • -

          setDeserializationSchema

          -
          public FlussSourceBuilder<OUT> setDeserializationSchema​(FlussDeserializationSchema<OUT> deserializationSchema)
          -
          Sets the deserialization schema for converting Fluss records to output records. - -

          This is a required parameter.

          -
          -
          Parameters:
          -
          deserializationSchema - the deserialization schema to use
          -
          Returns:
          -
          this builder
          -
          -
        • -
        - - - -
          -
        • -

          setProjectedFields

          -
          public FlussSourceBuilder<OUT> setProjectedFields​(String... projectedFieldNames)
          -
          Sets the projected fields for this source using field names. - -

          Projection allows selecting a subset of fields from the table. Without projection, all - fields from the table are included.

          -
          -
          Parameters:
          -
          projectedFieldNames - names of the fields to project
          -
          Returns:
          -
          this builder
          -
          Throws:
          -
          NullPointerException - if projectedFieldNames is null
          -
          -
        • -
        - - - -
          -
        • -

          setFlussConfig

          -
          public FlussSourceBuilder<OUT> setFlussConfig​(org.apache.fluss.config.Configuration flussConf)
          -
          Sets custom Fluss configuration properties for the source connector. - -

          If not specified, an empty configuration will be created and populated with required - properties. Any configuration set through this method will be merged with table-specific - properties retrieved from the Fluss system.

          -
          -
          Parameters:
          -
          flussConf - the configuration to use
          -
          Returns:
          -
          this builder
          -
          -
        • -
        - - - -
          -
        • -

          build

          -
          public FlussSource<OUT> build()
          -
          Builds and returns a new FlussSource instance with the configured properties. - -

          This method validates all required parameters, connects to the Fluss system to retrieve - table metadata, and constructs a configured source.

          -
          -
          Returns:
          -
          a new FlussSource instance
          -
          Throws:
          -
          NullPointerException - if any required parameter is missing
          -
          IllegalArgumentException - if any parameter is invalid
          -
          RuntimeException - if connection to Fluss fails or the table cannot be found
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlinkSource.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlinkSource.html deleted file mode 100644 index f08faf14b9..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlinkSource.html +++ /dev/null @@ -1,194 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.FlinkSource (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.FlinkSource

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlinkTableSource.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlinkTableSource.html deleted file mode 100644 index c670250364..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlinkTableSource.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.FlinkTableSource (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.FlinkTableSource

    -
    -
    No usage of org.apache.fluss.flink.source.FlinkTableSource
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlussSource.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlussSource.html deleted file mode 100644 index d8a927b34d..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlussSource.html +++ /dev/null @@ -1,194 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.FlussSource (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.FlussSource

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlussSourceBuilder.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlussSourceBuilder.html deleted file mode 100644 index 1262dd4965..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/class-use/FlussSourceBuilder.html +++ /dev/null @@ -1,250 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.FlussSourceBuilder (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.FlussSourceBuilder

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/DeserializerInitContextImpl.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/DeserializerInitContextImpl.html deleted file mode 100644 index af6e9287d7..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/DeserializerInitContextImpl.html +++ /dev/null @@ -1,380 +0,0 @@ - - - - - -DeserializerInitContextImpl (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class DeserializerInitContextImpl

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.deserializer.DeserializerInitContextImpl
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    - -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/FlussDeserializationSchema.InitializationContext.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/FlussDeserializationSchema.InitializationContext.html deleted file mode 100644 index 24038ddf98..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/FlussDeserializationSchema.InitializationContext.html +++ /dev/null @@ -1,322 +0,0 @@ - - - - - -FlussDeserializationSchema.InitializationContext (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface FlussDeserializationSchema.InitializationContext

    -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Summary

        - - - - - - - - - - - - - - - - - - - - - - -
        All Methods Instance Methods Abstract Methods 
        Modifier and TypeMethodDescription
        org.apache.flink.metrics.MetricGroupgetMetricGroup() -
        Returns the metric group for the parallel subtask of the source that runs this FlussDeserializationSchema.
        -
        org.apache.fluss.types.RowTypegetRowSchema() -
        Returns the schema of the LogRecord.getRow().
        -
        org.apache.flink.util.UserCodeClassLoadergetUserCodeClassLoader() -
        Gets the UserCodeClassLoader to load classes that are not in system's classpath, - but are part of the jar file of a user job.
        -
        -
      • -
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getMetricGroup

          -
          org.apache.flink.metrics.MetricGroup getMetricGroup()
          -
          Returns the metric group for the parallel subtask of the source that runs this FlussDeserializationSchema. - -

          Instances of this class can be used to register new metrics with Flink and to create a - nested hierarchy based on the group names. See MetricGroup for more information - for the metrics system.

          -
          -
          See Also:
          -
          MetricGroup
          -
          -
        • -
        - - - -
          -
        • -

          getUserCodeClassLoader

          -
          org.apache.flink.util.UserCodeClassLoader getUserCodeClassLoader()
          -
          Gets the UserCodeClassLoader to load classes that are not in system's classpath, - but are part of the jar file of a user job.
          -
          -
          See Also:
          -
          UserCodeClassLoader
          -
          -
        • -
        - - - -
          -
        • -

          getRowSchema

          -
          org.apache.fluss.types.RowType getRowSchema()
          -
          Returns the schema of the LogRecord.getRow().
          -
          -
          Returns:
          -
          The schema of the LogRecord.getRow().
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/FlussDeserializationSchema.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/FlussDeserializationSchema.html deleted file mode 100644 index 40cd6d4637..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/FlussDeserializationSchema.html +++ /dev/null @@ -1,357 +0,0 @@ - - - - - -FlussDeserializationSchema (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface FlussDeserializationSchema<T>

    -
    -
    -
    -
      -
    • -
      -
      Type Parameters:
      -
      T - The type created by the deserialization schema.
      -
      -
      -
      All Superinterfaces:
      -
      Serializable
      -
      -
      -
      All Known Implementing Classes:
      -
      JsonStringDeserializationSchema, RowDataDeserializationSchema
      -
      -
      -
      @PublicEvolving
      -public interface FlussDeserializationSchema<T>
      -extends Serializable
      -
      Interface for deserialization schema used to deserialize LogRecord objects into specific - data types.
      -
      -
      Since:
      -
      0.7
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - - - - - -
          -
        • -

          deserialize

          -
          T deserialize​(org.apache.fluss.record.LogRecord record)
          -       throws Exception
          -
          Deserializes a LogRecord into an object of type T.
          -
          -
          Parameters:
          -
          record - The Fluss record to deserialize.
          -
          Returns:
          -
          The deserialized object.
          -
          Throws:
          -
          Exception - If the deserialization fails.
          -
          -
        • -
        - - - -
          -
        • -

          getProducedType

          -
          org.apache.flink.api.common.typeinfo.TypeInformation<T> getProducedType​(org.apache.fluss.types.RowType rowSchema)
          -
          Gets the data type (as a TypeInformation) produced by this deserializer.
          -
          -
          Parameters:
          -
          rowSchema - The schema of the LogRecord.getRow().
          -
          Returns:
          -
          The data type produced by this deserializer.
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/JsonStringDeserializationSchema.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/JsonStringDeserializationSchema.html deleted file mode 100644 index 9852c079b0..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/JsonStringDeserializationSchema.html +++ /dev/null @@ -1,435 +0,0 @@ - - - - - -JsonStringDeserializationSchema (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class JsonStringDeserializationSchema

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.deserializer.JsonStringDeserializationSchema
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      Serializable, FlussDeserializationSchema<String>
      -
      -
      -
      @PublicEvolving
      -public class JsonStringDeserializationSchema
      -extends Object
      -implements FlussDeserializationSchema<String>
      -
      A deserialization schema that converts LogRecord objects to JSON strings. - -

      This implementation serializes Fluss records into JSON strings, making it useful for - debugging, logging, or when the downstream processing requires string-based JSON data. The schema - preserves important metadata such as offset, timestamp, and change type along with the actual row - data. - -

      The resulting JSON has the following structure: - -

      
      - {
      -   "offset": <record_offset>,
      -   "timestamp": <record_timestamp>,
      -   "changeType": <APPEND_ONLY|INSERT|UPDATE_BEFORE|UPDATE_AFTER|DELETE>,
      -   "row": <string_representation_of_row>
      - }
      - 
      - -

      Usage example: - -

      
      - FlussSource<String> source = FlussSource.builder()
      -     .setDeserializationSchema(new JsonStringDeserializationSchema())
      -     .build();
      - 
      -
      -
      Since:
      -
      0.7
      -
      See Also:
      -
      Serialized Form
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          JsonStringDeserializationSchema

          -
          public JsonStringDeserializationSchema()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          open

          -
          public void open​(FlussDeserializationSchema.InitializationContext context)
          -          throws Exception
          -
          Initializes the JSON serialization mechanism. - -

          This method creates a new ObjectMapper instance and configures it with: - -

            -
          • JavaTimeModule for proper serialization of date/time objects -
          • Configuration to render dates in ISO-8601 format rather than timestamps -
          -
          -
          Specified by:
          -
          open in interface FlussDeserializationSchema<String>
          -
          Parameters:
          -
          context - Contextual information for initialization (not used in this implementation)
          -
          Throws:
          -
          Exception - if initialization fails
          -
          -
        • -
        - - - -
          -
        • -

          deserialize

          -
          public String deserialize​(org.apache.fluss.record.LogRecord record)
          -                   throws Exception
          -
          Deserializes a LogRecord into a JSON String. - -

          The method extracts key information from the record (offset, timestamp, change type, and - row data) and serializes it as a JSON string.

          -
          -
          Specified by:
          -
          deserialize in interface FlussDeserializationSchema<String>
          -
          Parameters:
          -
          record - The Fluss LogRecord to deserialize
          -
          Returns:
          -
          JSON string representation of the record
          -
          Throws:
          -
          Exception - If JSON serialization fails
          -
          -
        • -
        - - - -
          -
        • -

          getProducedType

          -
          public org.apache.flink.api.common.typeinfo.TypeInformation<String> getProducedType​(org.apache.fluss.types.RowType rowSchema)
          -
          Returns the TypeInformation for the produced String type.
          -
          -
          Specified by:
          -
          getProducedType in interface FlussDeserializationSchema<String>
          -
          Parameters:
          -
          rowSchema - The schema of the LogRecord.getRow().
          -
          Returns:
          -
          TypeInformation for String class
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/RowDataDeserializationSchema.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/RowDataDeserializationSchema.html deleted file mode 100644 index 233aba8656..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/RowDataDeserializationSchema.html +++ /dev/null @@ -1,414 +0,0 @@ - - - - - -RowDataDeserializationSchema (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class RowDataDeserializationSchema

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.deserializer.RowDataDeserializationSchema
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      Serializable, FlussDeserializationSchema<org.apache.flink.table.data.RowData>
      -
      -
      -
      @PublicEvolving
      -public class RowDataDeserializationSchema
      -extends Object
      -implements FlussDeserializationSchema<org.apache.flink.table.data.RowData>
      -
      A deserialization schema that converts LogRecord objects to Flink's RowData - format. - -

      This implementation takes a RowType in its constructor and uses a FlussRowToFlinkRowConverter to transform Fluss records into Flink's internal row representation. - -

      Usage example: - -

      
      - FlussSource<RowData> source = FlussSource.builder()
      -     .setDeserializationSchema(new RowDataDeserializationSchema())
      -     .build();
      - 
      -
      -
      Since:
      -
      0.7
      -
      See Also:
      -
      Serialized Form
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          RowDataDeserializationSchema

          -
          public RowDataDeserializationSchema()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - - - - - -
          -
        • -

          deserialize

          -
          public org.apache.flink.table.data.RowData deserialize​(org.apache.fluss.record.LogRecord record)
          -                                                throws Exception
          -
          Deserializes a LogRecord into a Flink RowData object.
          -
          -
          Specified by:
          -
          deserialize in interface FlussDeserializationSchema<org.apache.flink.table.data.RowData>
          -
          Parameters:
          -
          record - The Fluss LogRecord to deserialize
          -
          Returns:
          -
          The deserialized RowData
          -
          Throws:
          -
          Exception - If deserialization fails or if the record is not a valid ScanRecord
          -
          -
        • -
        - - - -
          -
        • -

          getProducedType

          -
          public org.apache.flink.api.common.typeinfo.TypeInformation<org.apache.flink.table.data.RowData> getProducedType​(org.apache.fluss.types.RowType rowSchema)
          -
          Returns the TypeInformation for the produced RowData type.
          -
          -
          Specified by:
          -
          getProducedType in interface FlussDeserializationSchema<org.apache.flink.table.data.RowData>
          -
          Parameters:
          -
          rowSchema - The schema of the LogRecord.getRow().
          -
          Returns:
          -
          The data type produced by this deserializer.
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/DeserializerInitContextImpl.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/DeserializerInitContextImpl.html deleted file mode 100644 index ded3fa0b21..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/DeserializerInitContextImpl.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.deserializer.DeserializerInitContextImpl (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.deserializer.DeserializerInitContextImpl

    -
    -
    No usage of org.apache.fluss.flink.source.deserializer.DeserializerInitContextImpl
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/FlussDeserializationSchema.InitializationContext.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/FlussDeserializationSchema.InitializationContext.html deleted file mode 100644 index 5a8866e2f6..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/FlussDeserializationSchema.InitializationContext.html +++ /dev/null @@ -1,225 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.flink.source.deserializer.FlussDeserializationSchema.InitializationContext (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.flink.source.deserializer.FlussDeserializationSchema.InitializationContext

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/FlussDeserializationSchema.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/FlussDeserializationSchema.html deleted file mode 100644 index e616a323c9..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/FlussDeserializationSchema.html +++ /dev/null @@ -1,292 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.flink.source.deserializer.FlussDeserializationSchema (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.flink.source.deserializer.FlussDeserializationSchema

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/JsonStringDeserializationSchema.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/JsonStringDeserializationSchema.html deleted file mode 100644 index 4bee001ebc..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/JsonStringDeserializationSchema.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.deserializer.JsonStringDeserializationSchema (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.deserializer.JsonStringDeserializationSchema

    -
    -
    No usage of org.apache.fluss.flink.source.deserializer.JsonStringDeserializationSchema
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/RowDataDeserializationSchema.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/RowDataDeserializationSchema.html deleted file mode 100644 index 7dd401fdc8..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/class-use/RowDataDeserializationSchema.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.deserializer.RowDataDeserializationSchema (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.deserializer.RowDataDeserializationSchema

    -
    -
    No usage of org.apache.fluss.flink.source.deserializer.RowDataDeserializationSchema
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-summary.html deleted file mode 100644 index 38e10cb557..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-summary.html +++ /dev/null @@ -1,205 +0,0 @@ - - - - - -org.apache.fluss.flink.source.deserializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.source.deserializer

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-tree.html deleted file mode 100644 index 178095e152..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-tree.html +++ /dev/null @@ -1,176 +0,0 @@ - - - - - -org.apache.fluss.flink.source.deserializer Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.source.deserializer

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -

    Interface Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-use.html deleted file mode 100644 index 6e12518ca7..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/deserializer/package-use.html +++ /dev/null @@ -1,240 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.source.deserializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.source.deserializer

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/FlinkRecordEmitter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/FlinkRecordEmitter.html deleted file mode 100644 index 054773ae42..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/FlinkRecordEmitter.html +++ /dev/null @@ -1,327 +0,0 @@ - - - - - -FlinkRecordEmitter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkRecordEmitter<OUT>

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.emitter.FlinkRecordEmitter<OUT>
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.connector.base.source.reader.RecordEmitter<RecordAndPos,​OUT,​SourceSplitState>
      -
      -
      -
      public class FlinkRecordEmitter<OUT>
      -extends Object
      -implements org.apache.flink.connector.base.source.reader.RecordEmitter<RecordAndPos,​OUT,​SourceSplitState>
      -
      The RecordEmitter implementation for FlinkSourceReader. - -

      During emitting records: - -

      when the record is from snapshot data, it'll update the records number to skip which helps to - skip the records has been read while restoring in reading snapshot data phase. - -

      when the record is from log data, it'll update the offset

      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      - -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          emitRecord

          -
          public void emitRecord​(RecordAndPos recordAndPosition,
          -                       org.apache.flink.api.connector.source.SourceOutput<OUT> sourceOutput,
          -                       SourceSplitState splitState)
          -
          -
          Specified by:
          -
          emitRecord in interface org.apache.flink.connector.base.source.reader.RecordEmitter<RecordAndPos,​OUT,​SourceSplitState>
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/class-use/FlinkRecordEmitter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/class-use/FlinkRecordEmitter.html deleted file mode 100644 index 5cbd53ddeb..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/class-use/FlinkRecordEmitter.html +++ /dev/null @@ -1,198 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.emitter.FlinkRecordEmitter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.emitter.FlinkRecordEmitter

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-summary.html deleted file mode 100644 index 277114fc8f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-summary.html +++ /dev/null @@ -1,168 +0,0 @@ - - - - - -org.apache.fluss.flink.source.emitter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.source.emitter

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-tree.html deleted file mode 100644 index e2cb4b1aed..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-tree.html +++ /dev/null @@ -1,163 +0,0 @@ - - - - - -org.apache.fluss.flink.source.emitter Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.source.emitter

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    -
      -
    • java.lang.Object -
        -
      • org.apache.fluss.flink.source.emitter.FlinkRecordEmitter<OUT> (implements org.apache.flink.connector.base.source.reader.RecordEmitter<E,​T,​SplitStateT>)
      • -
      -
    • -
    -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-use.html deleted file mode 100644 index 4b74eb88a1..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/emitter/package-use.html +++ /dev/null @@ -1,185 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.source.emitter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.source.emitter

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/FlinkSourceEnumerator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/FlinkSourceEnumerator.html deleted file mode 100644 index dc735a125e..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/FlinkSourceEnumerator.html +++ /dev/null @@ -1,580 +0,0 @@ - - - - - -FlinkSourceEnumerator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkSourceEnumerator

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      AutoCloseable, org.apache.flink.api.common.state.CheckpointListener, org.apache.flink.api.connector.source.SplitEnumerator<SourceSplitBase,​SourceEnumeratorState>
      -
      -
      -
      public class FlinkSourceEnumerator
      -extends Object
      -implements org.apache.flink.api.connector.source.SplitEnumerator<SourceSplitBase,​SourceEnumeratorState>
      -
      An implementation of SplitEnumerator for the data of Fluss. - -

      The enumerator is responsible for: - -

        -
      • Get the all splits(snapshot split + log split) for a table of Fluss to be read. -
      • Assign the splits to readers with the guarantee that the splits belong to the same bucket - will be assigned to same reader. -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Summary

        - - - - - - - - - - - - - - - - - - -
        Constructors 
        ConstructorDescription
        FlinkSourceEnumerator​(org.apache.fluss.metadata.TablePath tablePath, - org.apache.fluss.config.Configuration flussConf, - boolean hasPrimaryKey, - boolean isPartitioned, - org.apache.flink.api.connector.source.SplitEnumeratorContext<SourceSplitBase> context, - Set<org.apache.fluss.metadata.TableBucket> assignedTableBuckets, - Map<Long,​String> assignedPartitions, - List<SourceSplitBase> pendingHybridLakeFlussSplits, - OffsetsInitializer startingOffsetsInitializer, - long scanPartitionDiscoveryIntervalMs, - boolean streaming, - List<PushdownUtils.FieldEqual> partitionFilters, - org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource) 
        FlinkSourceEnumerator​(org.apache.fluss.metadata.TablePath tablePath, - org.apache.fluss.config.Configuration flussConf, - boolean hasPrimaryKey, - boolean isPartitioned, - org.apache.flink.api.connector.source.SplitEnumeratorContext<SourceSplitBase> context, - OffsetsInitializer startingOffsetsInitializer, - long scanPartitionDiscoveryIntervalMs, - boolean streaming, - List<PushdownUtils.FieldEqual> partitionFilters) 
        FlinkSourceEnumerator​(org.apache.fluss.metadata.TablePath tablePath, - org.apache.fluss.config.Configuration flussConf, - boolean hasPrimaryKey, - boolean isPartitioned, - org.apache.flink.api.connector.source.SplitEnumeratorContext<SourceSplitBase> context, - OffsetsInitializer startingOffsetsInitializer, - long scanPartitionDiscoveryIntervalMs, - boolean streaming, - List<PushdownUtils.FieldEqual> partitionFilters, - org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource) 
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlinkSourceEnumerator

          -
          public FlinkSourceEnumerator​(org.apache.fluss.metadata.TablePath tablePath,
          -                             org.apache.fluss.config.Configuration flussConf,
          -                             boolean hasPrimaryKey,
          -                             boolean isPartitioned,
          -                             org.apache.flink.api.connector.source.SplitEnumeratorContext<SourceSplitBase> context,
          -                             OffsetsInitializer startingOffsetsInitializer,
          -                             long scanPartitionDiscoveryIntervalMs,
          -                             boolean streaming,
          -                             List<PushdownUtils.FieldEqual> partitionFilters)
          -
        • -
        - - - -
          -
        • -

          FlinkSourceEnumerator

          -
          public FlinkSourceEnumerator​(org.apache.fluss.metadata.TablePath tablePath,
          -                             org.apache.fluss.config.Configuration flussConf,
          -                             boolean hasPrimaryKey,
          -                             boolean isPartitioned,
          -                             org.apache.flink.api.connector.source.SplitEnumeratorContext<SourceSplitBase> context,
          -                             OffsetsInitializer startingOffsetsInitializer,
          -                             long scanPartitionDiscoveryIntervalMs,
          -                             boolean streaming,
          -                             List<PushdownUtils.FieldEqual> partitionFilters,
          -                             @Nullable
          -                             org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource)
          -
        • -
        - - - -
          -
        • -

          FlinkSourceEnumerator

          -
          public FlinkSourceEnumerator​(org.apache.fluss.metadata.TablePath tablePath,
          -                             org.apache.fluss.config.Configuration flussConf,
          -                             boolean hasPrimaryKey,
          -                             boolean isPartitioned,
          -                             org.apache.flink.api.connector.source.SplitEnumeratorContext<SourceSplitBase> context,
          -                             Set<org.apache.fluss.metadata.TableBucket> assignedTableBuckets,
          -                             Map<Long,​String> assignedPartitions,
          -                             List<SourceSplitBase> pendingHybridLakeFlussSplits,
          -                             OffsetsInitializer startingOffsetsInitializer,
          -                             long scanPartitionDiscoveryIntervalMs,
          -                             boolean streaming,
          -                             List<PushdownUtils.FieldEqual> partitionFilters,
          -                             @Nullable
          -                             org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - - - - - -
          -
        • -

          getSplitOwner

          -
          @VisibleForTesting
          -protected int getSplitOwner​(SourceSplitBase split)
          -
          Returns the index of the target subtask that a specific split should be assigned to. - -

          The resulting distribution of splits of a single table has the following contract: - -

            -
          • 1. Splits in same bucket are assigned to same subtask -
          • 2. Uniformly distributed across subtasks -
          • 3. For partitioned table, the buckets in same partition are round-robin distributed - (strictly clockwise w.r.t. ascending subtask indices) by using the partition id as the - offset from a starting index. The starting index is the index of the subtask which - bucket 0 of the partition will be assigned to, determined using the partition id to - make sure the partitions' buckets of a table are distributed uniformly -
          -
          -
          Parameters:
          -
          split - the split to assign.
          -
          Returns:
          -
          the id of the subtask that owns the split.
          -
          -
        • -
        - - - -
          -
        • -

          handleSplitRequest

          -
          public void handleSplitRequest​(int subtaskId,
          -                               @Nullable
          -                               String requesterHostname)
          -
          -
          Specified by:
          -
          handleSplitRequest in interface org.apache.flink.api.connector.source.SplitEnumerator<SourceSplitBase,​SourceEnumeratorState>
          -
          -
        • -
        - - - -
          -
        • -

          handleSourceEvent

          -
          public void handleSourceEvent​(int subtaskId,
          -                              org.apache.flink.api.connector.source.SourceEvent sourceEvent)
          -
          -
          Specified by:
          -
          handleSourceEvent in interface org.apache.flink.api.connector.source.SplitEnumerator<SourceSplitBase,​SourceEnumeratorState>
          -
          -
        • -
        - - - - - - - -
          -
        • -

          addReader

          -
          public void addReader​(int subtaskId)
          -
          -
          Specified by:
          -
          addReader in interface org.apache.flink.api.connector.source.SplitEnumerator<SourceSplitBase,​SourceEnumeratorState>
          -
          -
        • -
        - - - - - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/class-use/FlinkSourceEnumerator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/class-use/FlinkSourceEnumerator.html deleted file mode 100644 index 2df23534ee..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/class-use/FlinkSourceEnumerator.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator

    -
    -
    No usage of org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/BucketOffsetsRetrieverImpl.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/BucketOffsetsRetrieverImpl.html deleted file mode 100644 index 2b1458677a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/BucketOffsetsRetrieverImpl.html +++ /dev/null @@ -1,365 +0,0 @@ - - - - - -BucketOffsetsRetrieverImpl (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class BucketOffsetsRetrieverImpl

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.enumerator.initializer.BucketOffsetsRetrieverImpl
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    - -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/LatestOffsetsInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/LatestOffsetsInitializer.html deleted file mode 100644 index a2736bfaeb..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/LatestOffsetsInitializer.html +++ /dev/null @@ -1,357 +0,0 @@ - - - - - -LatestOffsetsInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LatestOffsetsInitializer

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.enumerator.initializer.LatestOffsetsInitializer
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LatestOffsetsInitializer

          -
          public LatestOffsetsInitializer()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getBucketOffsets

          -
          public Map<Integer,​Long> getBucketOffsets​(@Nullable
          -                                                String partitionName,
          -                                                Collection<Integer> buckets,
          -                                                OffsetsInitializer.BucketOffsetsRetriever bucketOffsetsRetriever)
          -
          Description copied from interface: OffsetsInitializer
          -
          Get the initial offsets for the given fluss buckets. These offsets will be used as starting - offsets of the fluss buckets.
          -
          -
          Specified by:
          -
          getBucketOffsets in interface OffsetsInitializer
          -
          Parameters:
          -
          partitionName - the partition name of the buckets if they are partitioned. Otherwise, - null.
          -
          buckets - the fluss buckets to get the starting offsets.
          -
          bucketOffsetsRetriever - a helper to retrieve information of the fluss buckets.
          -
          Returns:
          -
          A mapping from fluss bucket to their offsets to start scanning from.
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/NoStoppingOffsetsInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/NoStoppingOffsetsInitializer.html deleted file mode 100644 index c894aa0b06..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/NoStoppingOffsetsInitializer.html +++ /dev/null @@ -1,356 +0,0 @@ - - - - - -NoStoppingOffsetsInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class NoStoppingOffsetsInitializer

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.enumerator.initializer.NoStoppingOffsetsInitializer
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          NoStoppingOffsetsInitializer

          -
          public NoStoppingOffsetsInitializer()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getBucketOffsets

          -
          public Map<Integer,​Long> getBucketOffsets​(@Nullable
          -                                                String partitionName,
          -                                                Collection<Integer> buckets,
          -                                                OffsetsInitializer.BucketOffsetsRetriever bucketOffsetsRetriever)
          -
          Description copied from interface: OffsetsInitializer
          -
          Get the initial offsets for the given fluss buckets. These offsets will be used as starting - offsets of the fluss buckets.
          -
          -
          Specified by:
          -
          getBucketOffsets in interface OffsetsInitializer
          -
          Parameters:
          -
          partitionName - the partition name of the buckets if they are partitioned. Otherwise, - null.
          -
          buckets - the fluss buckets to get the starting offsets.
          -
          bucketOffsetsRetriever - a helper to retrieve information of the fluss buckets.
          -
          Returns:
          -
          A mapping from fluss bucket to their offsets to start scanning from.
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/OffsetsInitializer.BucketOffsetsRetriever.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/OffsetsInitializer.BucketOffsetsRetriever.html deleted file mode 100644 index bb48bcea13..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/OffsetsInitializer.BucketOffsetsRetriever.html +++ /dev/null @@ -1,299 +0,0 @@ - - - - - -OffsetsInitializer.BucketOffsetsRetriever (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface OffsetsInitializer.BucketOffsetsRetriever

    -
    -
    -
    -
      -
    • -
      -
      All Known Implementing Classes:
      -
      BucketOffsetsRetrieverImpl
      -
      -
      -
      Enclosing interface:
      -
      OffsetsInitializer
      -
      -
      -
      public static interface OffsetsInitializer.BucketOffsetsRetriever
      -
      An interface that provides necessary information to the OffsetsInitializer to get the - initial offsets of the fluss buckets.
      -
    • -
    -
    -
    - -
    -
    - -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/OffsetsInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/OffsetsInitializer.html deleted file mode 100644 index d2a8709429..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/OffsetsInitializer.html +++ /dev/null @@ -1,410 +0,0 @@ - - - - - -OffsetsInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface OffsetsInitializer

    -
    -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getBucketOffsets

          -
          Map<Integer,​Long> getBucketOffsets​(@Nullable
          -                                         String partitionName,
          -                                         Collection<Integer> buckets,
          -                                         OffsetsInitializer.BucketOffsetsRetriever bucketOffsetsRetriever)
          -
          Get the initial offsets for the given fluss buckets. These offsets will be used as starting - offsets of the fluss buckets.
          -
          -
          Parameters:
          -
          partitionName - the partition name of the buckets if they are partitioned. Otherwise, - null.
          -
          buckets - the fluss buckets to get the starting offsets.
          -
          bucketOffsetsRetriever - a helper to retrieve information of the fluss buckets.
          -
          Returns:
          -
          A mapping from fluss bucket to their offsets to start scanning from.
          -
          -
        • -
        - - - - - - - - - - - -
          -
        • -

          full

          -
          static OffsetsInitializer full()
          -
          Get an OffsetsInitializer which performs a full snapshot on the table upon first - startup, and continue to read log with the offset to the snapshot. - -

          If the table to read is a log table, the full snapshot means reading from the earliest log - offset which means "full" OffsetsInitializer equal to the earliest() - OffsetsInitializer. If the table to read is a primary key table, the full snapshot means - reading the latest snapshot which materializes all changes on the table.

          -
          -
          Returns:
          -
          an OffsetsInitializer which initializes the offsets to snapshot offsets.
          -
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/SnapshotOffsetsInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/SnapshotOffsetsInitializer.html deleted file mode 100644 index dae9ff6d1f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/SnapshotOffsetsInitializer.html +++ /dev/null @@ -1,357 +0,0 @@ - - - - - -SnapshotOffsetsInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class SnapshotOffsetsInitializer

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.enumerator.initializer.SnapshotOffsetsInitializer
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          SnapshotOffsetsInitializer

          -
          public SnapshotOffsetsInitializer()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getBucketOffsets

          -
          public Map<Integer,​Long> getBucketOffsets​(@Nullable
          -                                                String partitionName,
          -                                                Collection<Integer> buckets,
          -                                                OffsetsInitializer.BucketOffsetsRetriever bucketOffsetsRetriever)
          -
          For table with primary key. This method will be invoked only when the kv snapshot not exists.
          -
          -
          Specified by:
          -
          getBucketOffsets in interface OffsetsInitializer
          -
          Parameters:
          -
          partitionName - the partition name of the buckets if they are partitioned. Otherwise, - null.
          -
          buckets - the fluss buckets to get the starting offsets.
          -
          bucketOffsetsRetriever - a helper to retrieve information of the fluss buckets.
          -
          Returns:
          -
          A mapping from fluss bucket to their offsets to start scanning from.
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/TimestampOffsetsInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/TimestampOffsetsInitializer.html deleted file mode 100644 index dab4adf861..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/TimestampOffsetsInitializer.html +++ /dev/null @@ -1,356 +0,0 @@ - - - - - -TimestampOffsetsInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TimestampOffsetsInitializer

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.enumerator.initializer.TimestampOffsetsInitializer
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          TimestampOffsetsInitializer

          -
          public TimestampOffsetsInitializer​(long timestamp)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getBucketOffsets

          -
          public Map<Integer,​Long> getBucketOffsets​(@Nullable
          -                                                String partitionName,
          -                                                Collection<Integer> buckets,
          -                                                OffsetsInitializer.BucketOffsetsRetriever bucketOffsetsRetriever)
          -
          Description copied from interface: OffsetsInitializer
          -
          Get the initial offsets for the given fluss buckets. These offsets will be used as starting - offsets of the fluss buckets.
          -
          -
          Specified by:
          -
          getBucketOffsets in interface OffsetsInitializer
          -
          Parameters:
          -
          partitionName - the partition name of the buckets if they are partitioned. Otherwise, - null.
          -
          buckets - the fluss buckets to get the starting offsets.
          -
          bucketOffsetsRetriever - a helper to retrieve information of the fluss buckets.
          -
          Returns:
          -
          A mapping from fluss bucket to their offsets to start scanning from.
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/BucketOffsetsRetrieverImpl.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/BucketOffsetsRetrieverImpl.html deleted file mode 100644 index 2215f82d13..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/BucketOffsetsRetrieverImpl.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.enumerator.initializer.BucketOffsetsRetrieverImpl (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.enumerator.initializer.BucketOffsetsRetrieverImpl

    -
    -
    No usage of org.apache.fluss.flink.source.enumerator.initializer.BucketOffsetsRetrieverImpl
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/LatestOffsetsInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/LatestOffsetsInitializer.html deleted file mode 100644 index 642d051545..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/LatestOffsetsInitializer.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.enumerator.initializer.LatestOffsetsInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.enumerator.initializer.LatestOffsetsInitializer

    -
    -
    No usage of org.apache.fluss.flink.source.enumerator.initializer.LatestOffsetsInitializer
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/NoStoppingOffsetsInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/NoStoppingOffsetsInitializer.html deleted file mode 100644 index 3870876bc6..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/NoStoppingOffsetsInitializer.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.enumerator.initializer.NoStoppingOffsetsInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.enumerator.initializer.NoStoppingOffsetsInitializer

    -
    -
    No usage of org.apache.fluss.flink.source.enumerator.initializer.NoStoppingOffsetsInitializer
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/OffsetsInitializer.BucketOffsetsRetriever.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/OffsetsInitializer.BucketOffsetsRetriever.html deleted file mode 100644 index 7dac3b93c6..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/OffsetsInitializer.BucketOffsetsRetriever.html +++ /dev/null @@ -1,273 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer.BucketOffsetsRetriever (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer.BucketOffsetsRetriever

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/OffsetsInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/OffsetsInitializer.html deleted file mode 100644 index ebd9feca7b..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/OffsetsInitializer.html +++ /dev/null @@ -1,431 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.flink.source.enumerator.initializer.OffsetsInitializer

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/SnapshotOffsetsInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/SnapshotOffsetsInitializer.html deleted file mode 100644 index 364841325c..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/SnapshotOffsetsInitializer.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.enumerator.initializer.SnapshotOffsetsInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.enumerator.initializer.SnapshotOffsetsInitializer

    -
    -
    No usage of org.apache.fluss.flink.source.enumerator.initializer.SnapshotOffsetsInitializer
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/TimestampOffsetsInitializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/TimestampOffsetsInitializer.html deleted file mode 100644 index 1522024ec2..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/class-use/TimestampOffsetsInitializer.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.enumerator.initializer.TimestampOffsetsInitializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.enumerator.initializer.TimestampOffsetsInitializer

    -
    -
    No usage of org.apache.fluss.flink.source.enumerator.initializer.TimestampOffsetsInitializer
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-summary.html deleted file mode 100644 index a7ebe193da..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-summary.html +++ /dev/null @@ -1,217 +0,0 @@ - - - - - -org.apache.fluss.flink.source.enumerator.initializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.source.enumerator.initializer

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-tree.html deleted file mode 100644 index e48355cd6f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-tree.html +++ /dev/null @@ -1,178 +0,0 @@ - - - - - -org.apache.fluss.flink.source.enumerator.initializer Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.source.enumerator.initializer

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -

    Interface Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-use.html deleted file mode 100644 index 529cbdaf1a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/initializer/package-use.html +++ /dev/null @@ -1,268 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.source.enumerator.initializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.source.enumerator.initializer

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-summary.html deleted file mode 100644 index 6e9be7be81..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-summary.html +++ /dev/null @@ -1,168 +0,0 @@ - - - - - -org.apache.fluss.flink.source.enumerator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.source.enumerator

    -
    -
    -
      -
    • - - - - - - - - - - - - -
      Class Summary 
      ClassDescription
      FlinkSourceEnumerator -
      An implementation of SplitEnumerator for the data of Fluss.
      -
      -
    • -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-tree.html deleted file mode 100644 index 499e58e6b0..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-tree.html +++ /dev/null @@ -1,163 +0,0 @@ - - - - - -org.apache.fluss.flink.source.enumerator Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.source.enumerator

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    -
      -
    • java.lang.Object -
        -
      • org.apache.fluss.flink.source.enumerator.FlinkSourceEnumerator (implements org.apache.flink.api.connector.source.SplitEnumerator<SplitT,​CheckpointT>)
      • -
      -
    • -
    -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-use.html deleted file mode 100644 index 0356c36bf7..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/enumerator/package-use.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.source.enumerator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.source.enumerator

    -
    -
    No usage of org.apache.fluss.flink.source.enumerator
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/PartitionBucketsUnsubscribedEvent.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/PartitionBucketsUnsubscribedEvent.html deleted file mode 100644 index 63459e3ab6..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/PartitionBucketsUnsubscribedEvent.html +++ /dev/null @@ -1,371 +0,0 @@ - - - - - -PartitionBucketsUnsubscribedEvent (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class PartitionBucketsUnsubscribedEvent

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.event.PartitionBucketsUnsubscribedEvent
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      Serializable, org.apache.flink.api.connector.source.SourceEvent
      -
      -
      -
      public class PartitionBucketsUnsubscribedEvent
      -extends Object
      -implements org.apache.flink.api.connector.source.SourceEvent
      -
      An event send from reader to enumerator to indicate the splits of the partition buckets have been - removed by the reader. It contains the table buckets that have been removed from the reader.
      -
      -
      See Also:
      -
      Serialized Form
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          PartitionBucketsUnsubscribedEvent

          -
          public PartitionBucketsUnsubscribedEvent​(Collection<org.apache.fluss.metadata.TableBucket> removedTableBuckets)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getRemovedTableBuckets

          -
          public Collection<org.apache.fluss.metadata.TableBucket> getRemovedTableBuckets()
          -
        • -
        - - - -
          -
        • -

          equals

          -
          public boolean equals​(Object o)
          -
          -
          Overrides:
          -
          equals in class Object
          -
          -
        • -
        - - - -
          -
        • -

          hashCode

          -
          public int hashCode()
          -
          -
          Overrides:
          -
          hashCode in class Object
          -
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/PartitionsRemovedEvent.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/PartitionsRemovedEvent.html deleted file mode 100644 index 6702436025..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/PartitionsRemovedEvent.html +++ /dev/null @@ -1,337 +0,0 @@ - - - - - -PartitionsRemovedEvent (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class PartitionsRemovedEvent

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.event.PartitionsRemovedEvent
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      Serializable, org.apache.flink.api.connector.source.SourceEvent
      -
      -
      -
      public class PartitionsRemovedEvent
      -extends Object
      -implements org.apache.flink.api.connector.source.SourceEvent
      -
      A source event to represent partitions is removed to send from enumerator to reader. - -

      It contains the partition bucket of the removed partitions that has been assigned to the - reader.

      -
      -
      See Also:
      -
      Serialized Form
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          PartitionsRemovedEvent

          -
          public PartitionsRemovedEvent​(Map<Long,​String> removedPartitions)
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/class-use/PartitionBucketsUnsubscribedEvent.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/class-use/PartitionBucketsUnsubscribedEvent.html deleted file mode 100644 index 938210a694..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/class-use/PartitionBucketsUnsubscribedEvent.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.event.PartitionBucketsUnsubscribedEvent (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.event.PartitionBucketsUnsubscribedEvent

    -
    -
    No usage of org.apache.fluss.flink.source.event.PartitionBucketsUnsubscribedEvent
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/class-use/PartitionsRemovedEvent.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/class-use/PartitionsRemovedEvent.html deleted file mode 100644 index decfc9e303..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/class-use/PartitionsRemovedEvent.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.event.PartitionsRemovedEvent (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.event.PartitionsRemovedEvent

    -
    -
    No usage of org.apache.fluss.flink.source.event.PartitionsRemovedEvent
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-summary.html deleted file mode 100644 index d3901f7075..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-summary.html +++ /dev/null @@ -1,175 +0,0 @@ - - - - - -org.apache.fluss.flink.source.event (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.source.event

    -
    -
    -
      -
    • - - - - - - - - - - - - - - - - -
      Class Summary 
      ClassDescription
      PartitionBucketsUnsubscribedEvent -
      An event send from reader to enumerator to indicate the splits of the partition buckets have been - removed by the reader.
      -
      PartitionsRemovedEvent -
      A source event to represent partitions is removed to send from enumerator to reader.
      -
      -
    • -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-tree.html deleted file mode 100644 index 8dc88334db..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-tree.html +++ /dev/null @@ -1,164 +0,0 @@ - - - - - -org.apache.fluss.flink.source.event Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.source.event

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-use.html deleted file mode 100644 index 950dcb582c..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/event/package-use.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.source.event (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.source.event

    -
    -
    No usage of org.apache.fluss.flink.source.event
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/FlinkAsyncLookupFunction.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/FlinkAsyncLookupFunction.html deleted file mode 100644 index 66248f02ea..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/FlinkAsyncLookupFunction.html +++ /dev/null @@ -1,417 +0,0 @@ - - - - - -FlinkAsyncLookupFunction (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkAsyncLookupFunction

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.flink.table.functions.UserDefinedFunction
      • -
      • -
          -
        • org.apache.flink.table.functions.AsyncTableFunction<org.apache.flink.table.data.RowData>
        • -
        • -
            -
          • org.apache.flink.table.functions.AsyncLookupFunction
          • -
          • -
              -
            • org.apache.fluss.flink.source.lookup.FlinkAsyncLookupFunction
            • -
            -
          • -
          -
        • -
        -
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      Serializable, org.apache.flink.table.functions.FunctionDefinition
      -
      -
      -
      public class FlinkAsyncLookupFunction
      -extends org.apache.flink.table.functions.AsyncLookupFunction
      -
      A flink async lookup function for fluss.
      -
      -
      See Also:
      -
      Serialized Form
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Summary

        - - - - - - - - - - -
        Constructors 
        ConstructorDescription
        FlinkAsyncLookupFunction​(org.apache.fluss.config.Configuration flussConfig, - org.apache.fluss.metadata.TablePath tablePath, - org.apache.flink.table.types.logical.RowType flinkRowType, - int maxRetryTimes, - LookupNormalizer lookupNormalizer, - int[] projection) 
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Summary

        - - - - - - - - - - - - - - - - - - - - - - -
        All Methods Instance Methods Concrete Methods 
        Modifier and TypeMethodDescription
        CompletableFuture<Collection<org.apache.flink.table.data.RowData>>asyncLookup​(org.apache.flink.table.data.RowData keyRow) -
        The invoke entry point of lookup function.
        -
        voidclose() 
        voidopen​(org.apache.flink.table.functions.FunctionContext context) 
        -
          -
        • - - -

          Methods inherited from class org.apache.flink.table.functions.AsyncLookupFunction

          -eval
        • -
        -
          -
        • - - -

          Methods inherited from class org.apache.flink.table.functions.AsyncTableFunction

          -getKind, getTypeInference
        • -
        -
          -
        • - - -

          Methods inherited from class org.apache.flink.table.functions.UserDefinedFunction

          -functionIdentifier, toString
        • -
        - -
          -
        • - - -

          Methods inherited from interface org.apache.flink.table.functions.FunctionDefinition

          -getRequirements, isDeterministic, supportsConstantFolding
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlinkAsyncLookupFunction

          -
          public FlinkAsyncLookupFunction​(org.apache.fluss.config.Configuration flussConfig,
          -                                org.apache.fluss.metadata.TablePath tablePath,
          -                                org.apache.flink.table.types.logical.RowType flinkRowType,
          -                                int maxRetryTimes,
          -                                LookupNormalizer lookupNormalizer,
          -                                @Nullable
          -                                int[] projection)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          open

          -
          public void open​(org.apache.flink.table.functions.FunctionContext context)
          -
          -
          Overrides:
          -
          open in class org.apache.flink.table.functions.UserDefinedFunction
          -
          -
        • -
        - - - -
          -
        • -

          asyncLookup

          -
          public CompletableFuture<Collection<org.apache.flink.table.data.RowData>> asyncLookup​(org.apache.flink.table.data.RowData keyRow)
          -
          The invoke entry point of lookup function.
          -
          -
          Specified by:
          -
          asyncLookup in class org.apache.flink.table.functions.AsyncLookupFunction
          -
          Parameters:
          -
          keyRow - A RowData that wraps lookup keys.
          -
          -
        • -
        - - - -
          -
        • -

          close

          -
          public void close()
          -           throws Exception
          -
          -
          Overrides:
          -
          close in class org.apache.flink.table.functions.UserDefinedFunction
          -
          Throws:
          -
          Exception
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/FlinkLookupFunction.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/FlinkLookupFunction.html deleted file mode 100644 index c51e6bc2f6..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/FlinkLookupFunction.html +++ /dev/null @@ -1,418 +0,0 @@ - - - - - -FlinkLookupFunction (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkLookupFunction

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.flink.table.functions.UserDefinedFunction
      • -
      • -
          -
        • org.apache.flink.table.functions.TableFunction<org.apache.flink.table.data.RowData>
        • -
        • -
            -
          • org.apache.flink.table.functions.LookupFunction
          • -
          • -
              -
            • org.apache.fluss.flink.source.lookup.FlinkLookupFunction
            • -
            -
          • -
          -
        • -
        -
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      Serializable, org.apache.flink.table.functions.FunctionDefinition
      -
      -
      -
      public class FlinkLookupFunction
      -extends org.apache.flink.table.functions.LookupFunction
      -
      A flink lookup function for fluss.
      -
      -
      See Also:
      -
      Serialized Form
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Summary

        - - - - - - - - - - -
        Constructors 
        ConstructorDescription
        FlinkLookupFunction​(org.apache.fluss.config.Configuration flussConfig, - org.apache.fluss.metadata.TablePath tablePath, - org.apache.flink.table.types.logical.RowType flinkRowType, - int maxRetryTimes, - LookupNormalizer lookupNormalizer, - int[] projection) 
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Summary

        - - - - - - - - - - - - - - - - - - - - - - -
        All Methods Instance Methods Concrete Methods 
        Modifier and TypeMethodDescription
        voidclose() 
        Collection<org.apache.flink.table.data.RowData>lookup​(org.apache.flink.table.data.RowData keyRow) -
        The invoke entry point of lookup function.
        -
        voidopen​(org.apache.flink.table.functions.FunctionContext context) 
        -
          -
        • - - -

          Methods inherited from class org.apache.flink.table.functions.LookupFunction

          -eval
        • -
        -
          -
        • - - -

          Methods inherited from class org.apache.flink.table.functions.TableFunction

          -collect, finish, getKind, getParameterTypes, getResultType, getTypeInference, setCollector
        • -
        -
          -
        • - - -

          Methods inherited from class org.apache.flink.table.functions.UserDefinedFunction

          -functionIdentifier, toString
        • -
        - -
          -
        • - - -

          Methods inherited from interface org.apache.flink.table.functions.FunctionDefinition

          -getRequirements, isDeterministic, supportsConstantFolding
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlinkLookupFunction

          -
          public FlinkLookupFunction​(org.apache.fluss.config.Configuration flussConfig,
          -                           org.apache.fluss.metadata.TablePath tablePath,
          -                           org.apache.flink.table.types.logical.RowType flinkRowType,
          -                           int maxRetryTimes,
          -                           LookupNormalizer lookupNormalizer,
          -                           @Nullable
          -                           int[] projection)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          open

          -
          public void open​(org.apache.flink.table.functions.FunctionContext context)
          -
          -
          Overrides:
          -
          open in class org.apache.flink.table.functions.UserDefinedFunction
          -
          -
        • -
        - - - -
          -
        • -

          lookup

          -
          public Collection<org.apache.flink.table.data.RowData> lookup​(org.apache.flink.table.data.RowData keyRow)
          -
          The invoke entry point of lookup function.
          -
          -
          Specified by:
          -
          lookup in class org.apache.flink.table.functions.LookupFunction
          -
          Parameters:
          -
          keyRow - - A RowData that wraps lookup keys. Currently only support single - rowkey.
          -
          -
        • -
        - - - -
          -
        • -

          close

          -
          public void close()
          -           throws Exception
          -
          -
          Overrides:
          -
          close in class org.apache.flink.table.functions.UserDefinedFunction
          -
          Throws:
          -
          Exception
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/LookupNormalizer.RemainingFilter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/LookupNormalizer.RemainingFilter.html deleted file mode 100644 index 0a1ad4e986..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/LookupNormalizer.RemainingFilter.html +++ /dev/null @@ -1,271 +0,0 @@ - - - - - -LookupNormalizer.RemainingFilter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LookupNormalizer.RemainingFilter

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.lookup.LookupNormalizer.RemainingFilter
      • -
      -
    • -
    -
    -
      -
    • -
      -
      Enclosing class:
      -
      LookupNormalizer
      -
      -
      -
      public static class LookupNormalizer.RemainingFilter
      -extends Object
      -
      A filter to check if the lookup result matches the remaining conditions.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          isMatch

          -
          public boolean isMatch​(org.apache.flink.table.data.RowData result)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/LookupNormalizer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/LookupNormalizer.html deleted file mode 100644 index fd9823cec3..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/LookupNormalizer.html +++ /dev/null @@ -1,418 +0,0 @@ - - - - - -LookupNormalizer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LookupNormalizer

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.lookup.LookupNormalizer
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      Serializable
      -
      -
      -
      public class LookupNormalizer
      -extends Object
      -implements Serializable
      -
      A utility class to normalize the lookup key row to match the Fluss key fields order and drop the - lookup result that doesn't match remaining conditions. - -

      For example, if we have a Fluss table with the following schema: - [id: int, name: string, age: int, score: double] with primary key (name, id). And a lookup - condition dim.id = src.id AND dim.name = src.name AND dim.age = 32. The lookup key - row will be [1001, "Alice", 32]. We need to normalize the lookup key row into - ["Alice", 1001], and construct a remaining filter for {age == 32}.

      -
      -
      See Also:
      -
      Serialized Form
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getLookupType

          -
          public LookupType getLookupType()
          -
          Returns the lookup type (primary key lookup, or prefix key lookup) requested from Flink to - Fluss.
          -
        • -
        - - - -
          -
        • -

          getLookupKeyIndexes

          -
          public int[] getLookupKeyIndexes()
          -
          Returns the indexes of the normalized lookup keys.
          -
        • -
        - - - -
          -
        • -

          normalizeLookupKey

          -
          public org.apache.flink.table.data.RowData normalizeLookupKey​(org.apache.flink.table.data.RowData lookupKey)
          -
          Normalize the lookup key row to match the request key and the key fields order.
          -
        • -
        - - - - - - - -
          -
        • -

          createPrimaryKeyLookupNormalizer

          -
          public static LookupNormalizer createPrimaryKeyLookupNormalizer​(int[] primaryKeys,
          -                                                                org.apache.flink.table.types.logical.RowType schema)
          -
          Create a LookupNormalizer for primary key lookup.
          -
        • -
        - - - -
          -
        • -

          validateAndCreateLookupNormalizer

          -
          public static LookupNormalizer validateAndCreateLookupNormalizer​(int[][] lookupKeyIndexes,
          -                                                                 int[] primaryKeys,
          -                                                                 int[] bucketKeys,
          -                                                                 int[] partitionKeys,
          -                                                                 org.apache.flink.table.types.logical.RowType schema,
          -                                                                 @Nullable
          -                                                                 int[] projectedFields)
          -
          Validate the lookup key indexes and primary keys, and create a LookupNormalizer.
          -
          -
          Parameters:
          -
          lookupKeyIndexes - the indexes of the lookup keys in the table row
          -
          primaryKeys - the indexes of the primary keys of the table
          -
          bucketKeys - the indexes of the bucket keys of the table, must be a part of primary keys
          -
          partitionKeys - the indexes of the partition keys of the table, maybe empty if the table - is not partitioned
          -
          schema - the schema of the table
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/FlinkAsyncLookupFunction.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/FlinkAsyncLookupFunction.html deleted file mode 100644 index 967ef4597f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/FlinkAsyncLookupFunction.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.lookup.FlinkAsyncLookupFunction (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.lookup.FlinkAsyncLookupFunction

    -
    -
    No usage of org.apache.fluss.flink.source.lookup.FlinkAsyncLookupFunction
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/FlinkLookupFunction.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/FlinkLookupFunction.html deleted file mode 100644 index ead4608f83..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/FlinkLookupFunction.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.lookup.FlinkLookupFunction (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.lookup.FlinkLookupFunction

    -
    -
    No usage of org.apache.fluss.flink.source.lookup.FlinkLookupFunction
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/LookupNormalizer.RemainingFilter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/LookupNormalizer.RemainingFilter.html deleted file mode 100644 index f30416656f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/LookupNormalizer.RemainingFilter.html +++ /dev/null @@ -1,192 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.lookup.LookupNormalizer.RemainingFilter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.lookup.LookupNormalizer.RemainingFilter

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/LookupNormalizer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/LookupNormalizer.html deleted file mode 100644 index c45a3480c2..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/class-use/LookupNormalizer.html +++ /dev/null @@ -1,234 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.lookup.LookupNormalizer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.lookup.LookupNormalizer

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-summary.html deleted file mode 100644 index ffa8c63d7a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-summary.html +++ /dev/null @@ -1,187 +0,0 @@ - - - - - -org.apache.fluss.flink.source.lookup (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.source.lookup

    -
    -
    -
      -
    • - - - - - - - - - - - - - - - - - - - - - - - - -
      Class Summary 
      ClassDescription
      FlinkAsyncLookupFunction -
      A flink async lookup function for fluss.
      -
      FlinkLookupFunction -
      A flink lookup function for fluss.
      -
      LookupNormalizer -
      A utility class to normalize the lookup key row to match the Fluss key fields order and drop the - lookup result that doesn't match remaining conditions.
      -
      LookupNormalizer.RemainingFilter -
      A filter to check if the lookup result matches the remaining conditions.
      -
      -
    • -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-tree.html deleted file mode 100644 index 1d32c1ef8d..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-tree.html +++ /dev/null @@ -1,186 +0,0 @@ - - - - - -org.apache.fluss.flink.source.lookup Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.source.lookup

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    -
      -
    • java.lang.Object -
        -
      • org.apache.fluss.flink.source.lookup.LookupNormalizer (implements java.io.Serializable)
      • -
      • org.apache.fluss.flink.source.lookup.LookupNormalizer.RemainingFilter
      • -
      • org.apache.flink.table.functions.UserDefinedFunction (implements org.apache.flink.table.functions.FunctionDefinition, java.io.Serializable) -
          -
        • org.apache.flink.table.functions.AsyncTableFunction<T> -
            -
          • org.apache.flink.table.functions.AsyncLookupFunction - -
          • -
          -
        • -
        • org.apache.flink.table.functions.TableFunction<T> -
            -
          • org.apache.flink.table.functions.LookupFunction - -
          • -
          -
        • -
        -
      • -
      -
    • -
    -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-use.html deleted file mode 100644 index 0d1586ab58..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/lookup/package-use.html +++ /dev/null @@ -1,192 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.source.lookup (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.source.lookup

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/FlinkSourceReaderMetrics.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/FlinkSourceReaderMetrics.html deleted file mode 100644 index 2756ea3a52..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/FlinkSourceReaderMetrics.html +++ /dev/null @@ -1,523 +0,0 @@ - - - - - -FlinkSourceReaderMetrics (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkSourceReaderMetrics

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics
      • -
      -
    • -
    -
    -
      -
    • -
      -
      public class FlinkSourceReaderMetrics
      -extends Object
      -
      A collection class for handling metrics in FlinkSourceReader of Fluss. - -

      All metrics of the source reader are registered under group "fluss.reader", which is a child - group of OperatorMetricGroup. Metrics related to a - specific table bucket will be registered in the group: - -

      "fluss.reader.bucket.{bucket_id}" for non-partitioned bucket or - "fluss.reader.partition.{partition_id}.bucket.{bucket_id}" for partitioned bucket. - -

      For example, current consuming offset of table "my-table" and bucket 1 will be reported in - metric: "{some_parent_groups}.operator.fluss.reader.table.my-table.bucket.1.currentOffset"

      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      - -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlinkSourceReaderMetrics

          -
          public FlinkSourceReaderMetrics​(org.apache.flink.metrics.groups.SourceReaderMetricGroup sourceReaderMetricGroup)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          reportRecordEventTime

          -
          public void reportRecordEventTime​(long lag)
          -
        • -
        - - - -
          -
        • -

          registerTableBucket

          -
          public void registerTableBucket​(org.apache.fluss.metadata.TableBucket tableBucket)
          -
        • -
        - - - -
          -
        • -

          recordCurrentOffset

          -
          public void recordCurrentOffset​(org.apache.fluss.metadata.TableBucket tb,
          -                                long offset)
          -
          Update current consuming offset of the given TableBucket.
          -
          -
          Parameters:
          -
          tb - Updating table bucket
          -
          offset - Current consuming offset
          -
          -
        • -
        - - - -
          -
        • -

          getSourceReaderMetricGroup

          -
          public org.apache.flink.metrics.groups.SourceReaderMetricGroup getSourceReaderMetricGroup()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/class-use/FlinkSourceReaderMetrics.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/class-use/FlinkSourceReaderMetrics.html deleted file mode 100644 index ca451b4e20..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/class-use/FlinkSourceReaderMetrics.html +++ /dev/null @@ -1,242 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.metrics.FlinkSourceReaderMetrics

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-summary.html deleted file mode 100644 index b46b2a0a88..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-summary.html +++ /dev/null @@ -1,168 +0,0 @@ - - - - - -org.apache.fluss.flink.source.metrics (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.source.metrics

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-tree.html deleted file mode 100644 index e61cff8b26..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-tree.html +++ /dev/null @@ -1,163 +0,0 @@ - - - - - -org.apache.fluss.flink.source.metrics Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.source.metrics

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-use.html deleted file mode 100644 index e5028d7525..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/metrics/package-use.html +++ /dev/null @@ -1,185 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.source.metrics (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.source.metrics

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-summary.html deleted file mode 100644 index b97639b357..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-summary.html +++ /dev/null @@ -1,186 +0,0 @@ - - - - - -org.apache.fluss.flink.source (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.source

    -
    -
    -
      -
    • - - - - - - - - - - - - - - - - - - - - - - - - -
      Class Summary 
      ClassDescription
      FlinkSource<OUT> -
      Flink source for Fluss.
      -
      FlinkTableSource -
      Flink table source to scan Fluss data.
      -
      FlussSource<OUT> -
      A Flink DataStream source implementation for reading data from Fluss tables.
      -
      FlussSourceBuilder<OUT> -
      Builder class for creating FlussSource instances.
      -
      -
    • -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-tree.html deleted file mode 100644 index 2bd9bc2ffc..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-tree.html +++ /dev/null @@ -1,169 +0,0 @@ - - - - - -org.apache.fluss.flink.source Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.source

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    -
      -
    • java.lang.Object -
        -
      • org.apache.fluss.flink.source.FlinkSource<OUT> (implements org.apache.flink.api.java.typeutils.ResultTypeQueryable<T>, org.apache.flink.api.connector.source.Source<T,​SplitT,​EnumChkT>) - -
      • -
      • org.apache.fluss.flink.source.FlinkTableSource (implements org.apache.flink.table.connector.source.LookupTableSource, org.apache.flink.table.connector.source.ScanTableSource, org.apache.flink.table.connector.source.abilities.SupportsAggregatePushDown, org.apache.flink.table.connector.source.abilities.SupportsFilterPushDown, org.apache.flink.table.connector.source.abilities.SupportsLimitPushDown, org.apache.flink.table.connector.source.abilities.SupportsProjectionPushDown, org.apache.flink.table.connector.source.abilities.SupportsRowLevelModificationScan)
      • -
      • org.apache.fluss.flink.source.FlussSourceBuilder<OUT>
      • -
      -
    • -
    -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-use.html deleted file mode 100644 index 7d941008d1..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/package-use.html +++ /dev/null @@ -1,197 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.source (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.source

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/BoundedSplitReader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/BoundedSplitReader.html deleted file mode 100644 index 4b736cc8a2..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/BoundedSplitReader.html +++ /dev/null @@ -1,350 +0,0 @@ - - - - - -BoundedSplitReader (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class BoundedSplitReader

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.reader.BoundedSplitReader
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          BoundedSplitReader

          -
          public BoundedSplitReader​(BatchScanner splitScanner,
          -                          long toSkip)
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkRecordsWithSplitIds.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkRecordsWithSplitIds.html deleted file mode 100644 index c83920bbd7..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkRecordsWithSplitIds.html +++ /dev/null @@ -1,444 +0,0 @@ - - - - - -FlinkRecordsWithSplitIds (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkRecordsWithSplitIds

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<RecordAndPos>
      -
      -
      -
      public class FlinkRecordsWithSplitIds
      -extends Object
      -implements org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<RecordAndPos>
      -
      An implementation of RecordsWithSplitIds which contains records from multiple splits.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      - -
      - -
      -
        -
      • - - -

        Method Detail

        - - - - - - - -
          -
        • -

          setTableBucketStoppingOffset

          -
          public void setTableBucketStoppingOffset​(org.apache.fluss.metadata.TableBucket tableBucket,
          -                                         long stoppingOffset)
          -
        • -
        - - - -
          -
        • -

          nextSplit

          -
          @Nullable
          -public String nextSplit()
          -
          -
          Specified by:
          -
          nextSplit in interface org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<RecordAndPos>
          -
          -
        • -
        - - - -
          -
        • -

          nextRecordFromSplit

          -
          @Nullable
          -public RecordAndPos nextRecordFromSplit()
          -
          -
          Specified by:
          -
          nextRecordFromSplit in interface org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<RecordAndPos>
          -
          -
        • -
        - - - -
          -
        • -

          recycle

          -
          public void recycle()
          -
          -
          Specified by:
          -
          recycle in interface org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<RecordAndPos>
          -
          -
        • -
        - - - -
          -
        • -

          finishedSplits

          -
          public Set<String> finishedSplits()
          -
          -
          Specified by:
          -
          finishedSplits in interface org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<RecordAndPos>
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkSourceReader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkSourceReader.html deleted file mode 100644 index 54db07bfc5..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkSourceReader.html +++ /dev/null @@ -1,443 +0,0 @@ - - - - - -FlinkSourceReader (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkSourceReader<OUT>

    -
    -
    - -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Summary

        -
          -
        • - - -

          Fields inherited from class org.apache.flink.connector.base.source.reader.SourceReaderBase

          -config, context, eofRecordEvaluator, options, recordEmitter, splitFetcherManager
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Summary

        - - - - - - - - - - -
        Constructors 
        ConstructorDescription
        FlinkSourceReader​(org.apache.flink.connector.base.source.reader.synchronization.FutureCompletingBlockingQueue<org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<RecordAndPos>> elementsQueue, - org.apache.fluss.config.Configuration flussConfig, - org.apache.fluss.metadata.TablePath tablePath, - org.apache.fluss.types.RowType sourceOutputType, - org.apache.flink.api.connector.source.SourceReaderContext context, - int[] projectedFields, - FlinkSourceReaderMetrics flinkSourceReaderMetrics, - FlinkRecordEmitter<OUT> recordEmitter, - org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource) 
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlinkSourceReader

          -
          public FlinkSourceReader​(org.apache.flink.connector.base.source.reader.synchronization.FutureCompletingBlockingQueue<org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<RecordAndPos>> elementsQueue,
          -                         org.apache.fluss.config.Configuration flussConfig,
          -                         org.apache.fluss.metadata.TablePath tablePath,
          -                         org.apache.fluss.types.RowType sourceOutputType,
          -                         org.apache.flink.api.connector.source.SourceReaderContext context,
          -                         @Nullable
          -                         int[] projectedFields,
          -                         FlinkSourceReaderMetrics flinkSourceReaderMetrics,
          -                         FlinkRecordEmitter<OUT> recordEmitter,
          -                         org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource)
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkSourceSplitReader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkSourceSplitReader.html deleted file mode 100644 index 09017063a3..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/FlinkSourceSplitReader.html +++ /dev/null @@ -1,411 +0,0 @@ - - - - - -FlinkSourceSplitReader (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkSourceSplitReader

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.reader.FlinkSourceSplitReader
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlinkSourceSplitReader

          -
          public FlinkSourceSplitReader​(org.apache.fluss.config.Configuration flussConf,
          -                              org.apache.fluss.metadata.TablePath tablePath,
          -                              org.apache.fluss.types.RowType sourceOutputType,
          -                              @Nullable
          -                              int[] projectedFields,
          -                              FlinkSourceReaderMetrics flinkSourceReaderMetrics,
          -                              @Nullable
          -                              org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - - - - - -
          -
        • -

          handleSplitsChanges

          -
          public void handleSplitsChanges​(org.apache.flink.connector.base.source.reader.splitreader.SplitsChange<SourceSplitBase> splitsChanges)
          -
          -
          Specified by:
          -
          handleSplitsChanges in interface org.apache.flink.connector.base.source.reader.splitreader.SplitReader<RecordAndPos,​SourceSplitBase>
          -
          -
        • -
        - - - -
          -
        • -

          removePartitions

          -
          public Set<org.apache.fluss.metadata.TableBucket> removePartitions​(Map<Long,​String> removedPartitions)
          -
        • -
        - - - -
          -
        • -

          wakeUp

          -
          public void wakeUp()
          -
          -
          Specified by:
          -
          wakeUp in interface org.apache.flink.connector.base.source.reader.splitreader.SplitReader<RecordAndPos,​SourceSplitBase>
          -
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/MutableRecordAndPos.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/MutableRecordAndPos.html deleted file mode 100644 index b2faec87e0..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/MutableRecordAndPos.html +++ /dev/null @@ -1,359 +0,0 @@ - - - - - -MutableRecordAndPos (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class MutableRecordAndPos

    -
    -
    - -
    -
      -
    • -
      -
      public class MutableRecordAndPos
      -extends RecordAndPos
      -
      A mutable version of the RecordAndPos. - -

      This mutable object is useful in cases where only once instance of a RecordAndPos is - needed at a time.

      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          MutableRecordAndPos

          -
          public MutableRecordAndPos()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          setRecord

          -
          public void setRecord​(ScanRecord scanRecord,
          -                      long readRecordsCount)
          -
        • -
        - - - -
          -
        • -

          setRecord

          -
          public void setRecord​(ScanRecord scanRecord,
          -                      long readRecordsCount,
          -                      int currentIteratorIndex)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/RecordAndPos.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/RecordAndPos.html deleted file mode 100644 index 160e6cada2..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/RecordAndPos.html +++ /dev/null @@ -1,544 +0,0 @@ - - - - - -RecordAndPos (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class RecordAndPos

    -
    -
    - -
    -
      -
    • -
      -
      Direct Known Subclasses:
      -
      MutableRecordAndPos
      -
      -
      -
      public class RecordAndPos
      -extends Object
      -
      A record wrapping a Fluss ScanRecord, the readRecordsCount when the record is - from reading snapshot, the code currentSplitIndex where the record is from when the Flink - source split read by split reader contains multiple splits(splittable unit), like LakeSnapshotAndFlussLogSplit contains multiple LakeSplits. - -

      When the record is from reading log, readRecordsCount will always be NO_READ_RECORDS_COUNT. - -

      The readRecordsCount defines the point in the snapshot reader AFTER the record. Record - processing and updating checkpointed state happens atomically. The position points to where the - reader should resume after this record is processed. - -

      For example, the very first record in a snapshot split could have a readRecordsCount - of one. - -

      It's produced by FlinkSourceSplitReader and emitted to FlinkRecordEmitter

      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - -
          -
        • -

          NO_READ_RECORDS_COUNT

          -
          public static final long NO_READ_RECORDS_COUNT
          -
          -
          See Also:
          -
          Constant Field Values
          -
          -
        • -
        - - - -
          -
        • -

          DEFAULT_SPLIT_INDEX

          -
          protected static final int DEFAULT_SPLIT_INDEX
          -
          -
          See Also:
          -
          Constant Field Values
          -
          -
        • -
        - - - -
          -
        • -

          scanRecord

          -
          protected ScanRecord scanRecord
          -
        • -
        - - - -
          -
        • -

          readRecordsCount

          -
          protected long readRecordsCount
          -
        • -
        - - - -
          -
        • -

          currentSplitIndex

          -
          protected int currentSplitIndex
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          RecordAndPos

          -
          public RecordAndPos​(ScanRecord scanRecord)
          -
        • -
        - - - -
          -
        • -

          RecordAndPos

          -
          public RecordAndPos​(ScanRecord scanRecord,
          -                    long readRecordsCount)
          -
        • -
        - - - -
          -
        • -

          RecordAndPos

          -
          public RecordAndPos​(ScanRecord scanRecord,
          -                    long readRecordsCount,
          -                    int currentSplitIndex)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          readRecordsCount

          -
          public long readRecordsCount()
          -
        • -
        - - - -
          -
        • -

          getCurrentSplitIndex

          -
          public int getCurrentSplitIndex()
          -
        • -
        - - - - - - - -
          -
        • -

          equals

          -
          public boolean equals​(Object o)
          -
          -
          Overrides:
          -
          equals in class Object
          -
          -
        • -
        - - - -
          -
        • -

          hashCode

          -
          public int hashCode()
          -
          -
          Overrides:
          -
          hashCode in class Object
          -
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/SplitScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/SplitScanner.html deleted file mode 100644 index 0193fa904f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/SplitScanner.html +++ /dev/null @@ -1,269 +0,0 @@ - - - - - -SplitScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface SplitScanner

    -
    -
    -
    -
      -
    • -
      -
      All Superinterfaces:
      -
      AutoCloseable
      -
      -
      -
      public interface SplitScanner
      -extends AutoCloseable
      -
      A scanner to scan split to records .
      -
    • -
    -
    -
    - -
    -
    - -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/BoundedSplitReader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/BoundedSplitReader.html deleted file mode 100644 index cb8dd8605a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/BoundedSplitReader.html +++ /dev/null @@ -1,192 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.reader.BoundedSplitReader (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.reader.BoundedSplitReader

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkRecordsWithSplitIds.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkRecordsWithSplitIds.html deleted file mode 100644 index 4c98b94075..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkRecordsWithSplitIds.html +++ /dev/null @@ -1,192 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkSourceReader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkSourceReader.html deleted file mode 100644 index 834663d646..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkSourceReader.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.reader.FlinkSourceReader (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.reader.FlinkSourceReader

    -
    -
    No usage of org.apache.fluss.flink.source.reader.FlinkSourceReader
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkSourceSplitReader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkSourceSplitReader.html deleted file mode 100644 index a145442292..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/FlinkSourceSplitReader.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.reader.FlinkSourceSplitReader (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.reader.FlinkSourceSplitReader

    -
    -
    No usage of org.apache.fluss.flink.source.reader.FlinkSourceSplitReader
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/MutableRecordAndPos.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/MutableRecordAndPos.html deleted file mode 100644 index e0d8dd985c..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/MutableRecordAndPos.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.reader.MutableRecordAndPos (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.reader.MutableRecordAndPos

    -
    -
    No usage of org.apache.fluss.flink.source.reader.MutableRecordAndPos
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/RecordAndPos.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/RecordAndPos.html deleted file mode 100644 index 9692165f7b..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/RecordAndPos.html +++ /dev/null @@ -1,359 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.reader.RecordAndPos (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.reader.RecordAndPos

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/SplitScanner.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/SplitScanner.html deleted file mode 100644 index b6fdc28a3c..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/class-use/SplitScanner.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.flink.source.reader.SplitScanner (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.flink.source.reader.SplitScanner

    -
    -
    No usage of org.apache.fluss.flink.source.reader.SplitScanner
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/FlinkSourceFetcherManager.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/FlinkSourceFetcherManager.html deleted file mode 100644 index bda5e4e9a5..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/FlinkSourceFetcherManager.html +++ /dev/null @@ -1,373 +0,0 @@ - - - - - -FlinkSourceFetcherManager (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkSourceFetcherManager

    -
    -
    - -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Summary

        -
          -
        • - - -

          Fields inherited from class org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager

          -fetchers
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Summary

        - - - - - - - - - - -
        Constructors 
        ConstructorDescription
        FlinkSourceFetcherManager​(org.apache.flink.connector.base.source.reader.synchronization.FutureCompletingBlockingQueue<org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<RecordAndPos>> elementsQueue, - Supplier<org.apache.flink.connector.base.source.reader.splitreader.SplitReader<RecordAndPos,​SourceSplitBase>> splitReaderSupplier, - Consumer<Collection<String>> splitFinishedHook) -
        Creates a new SplitFetcherManager with a single I/O threads.
        -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Summary

        - - - - - - - - - - - - -
        All Methods Instance Methods Concrete Methods 
        Modifier and TypeMethodDescription
        voidremovePartitions​(Map<Long,​String> removedPartitions, - Consumer<Set<org.apache.fluss.metadata.TableBucket>> unsubscribeTableBucketsCallback) 
        -
          -
        • - - -

          Methods inherited from class org.apache.flink.connector.base.source.reader.fetcher.SingleThreadFetcherManager

          -addSplits, getRunningFetcher, removeSplits
        • -
        -
          -
        • - - -

          Methods inherited from class org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager

          -checkErrors, close, createSplitFetcher, getNumAliveFetchers, getQueue, maybeShutdownFinishedFetchers, pauseOrResumeSplits, startFetcher
        • -
        - -
      • -
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlinkSourceFetcherManager

          -
          public FlinkSourceFetcherManager​(org.apache.flink.connector.base.source.reader.synchronization.FutureCompletingBlockingQueue<org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<RecordAndPos>> elementsQueue,
          -                                 Supplier<org.apache.flink.connector.base.source.reader.splitreader.SplitReader<RecordAndPos,​SourceSplitBase>> splitReaderSupplier,
          -                                 Consumer<Collection<String>> splitFinishedHook)
          -
          Creates a new SplitFetcherManager with a single I/O threads.
          -
          -
          Parameters:
          -
          elementsQueue - The queue that is used to hand over data from the I/O thread (the - fetchers) to the reader (which emits the records and book-keeps the state. This must be - the same queue instance that is also passed to the SourceReaderBase.
          -
          splitReaderSupplier - The factory for the split reader that connects to the source - system.
          -
          splitFinishedHook - Hook for handling finished splits in split fetchers.
          -
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          removePartitions

          -
          public void removePartitions​(Map<Long,​String> removedPartitions,
          -                             Consumer<Set<org.apache.fluss.metadata.TableBucket>> unsubscribeTableBucketsCallback)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/class-use/FlinkSourceFetcherManager.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/class-use/FlinkSourceFetcherManager.html deleted file mode 100644 index 141d952b72..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/class-use/FlinkSourceFetcherManager.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.reader.fetcher.FlinkSourceFetcherManager (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.reader.fetcher.FlinkSourceFetcherManager

    -
    -
    No usage of org.apache.fluss.flink.source.reader.fetcher.FlinkSourceFetcherManager
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-summary.html deleted file mode 100644 index f1cdd0eb29..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-summary.html +++ /dev/null @@ -1,168 +0,0 @@ - - - - - -org.apache.fluss.flink.source.reader.fetcher (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.source.reader.fetcher

    -
    -
    -
      -
    • - - - - - - - - - - - - -
      Class Summary 
      ClassDescription
      FlinkSourceFetcherManager -
      The SplitFetcherManager for Fluss source.
      -
      -
    • -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-tree.html deleted file mode 100644 index cf0935b731..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-tree.html +++ /dev/null @@ -1,175 +0,0 @@ - - - - - -org.apache.fluss.flink.source.reader.fetcher Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.source.reader.fetcher

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    -
      -
    • java.lang.Object -
        -
      • org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager<E,​SplitT> - -
      • -
      -
    • -
    -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-use.html deleted file mode 100644 index 8045c94b9d..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/fetcher/package-use.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.source.reader.fetcher (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.source.reader.fetcher

    -
    -
    No usage of org.apache.fluss.flink.source.reader.fetcher
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-summary.html deleted file mode 100644 index 73fd8a869d..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-summary.html +++ /dev/null @@ -1,217 +0,0 @@ - - - - - -org.apache.fluss.flink.source.reader (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.source.reader

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-tree.html deleted file mode 100644 index 9f51ef1381..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-tree.html +++ /dev/null @@ -1,193 +0,0 @@ - - - - - -org.apache.fluss.flink.source.reader Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.source.reader

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    -
      -
    • java.lang.Object -
        -
      • org.apache.fluss.flink.source.reader.BoundedSplitReader (implements java.lang.AutoCloseable)
      • -
      • org.apache.fluss.flink.source.reader.FlinkRecordsWithSplitIds (implements org.apache.flink.connector.base.source.reader.RecordsWithSplitIds<E>)
      • -
      • org.apache.fluss.flink.source.reader.FlinkSourceSplitReader (implements org.apache.flink.connector.base.source.reader.splitreader.SplitReader<E,​SplitT>)
      • -
      • org.apache.fluss.flink.source.reader.RecordAndPos - -
      • -
      • org.apache.flink.connector.base.source.reader.SourceReaderBase<E,​T,​SplitT,​SplitStateT> (implements org.apache.flink.api.connector.source.SourceReader<T,​SplitT>) - -
      • -
      -
    • -
    -
    -
    -

    Interface Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-use.html deleted file mode 100644 index 3ae499b518..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/reader/package-use.html +++ /dev/null @@ -1,274 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.source.reader (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.source.reader

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/HybridSnapshotLogSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/HybridSnapshotLogSplit.html deleted file mode 100644 index 90a99a24bc..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/HybridSnapshotLogSplit.html +++ /dev/null @@ -1,476 +0,0 @@ - - - - - -HybridSnapshotLogSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class HybridSnapshotLogSplit

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.api.connector.source.SourceSplit
      -
      -
      -
      public class HybridSnapshotLogSplit
      -extends SnapshotSplit
      -
      The hybrid split for first reading the snapshot files and then switch to read the cdc log from a - specified offset. - -

      Only used for primary key table which will be of snapshot phase and incremental phase of - reading.

      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          HybridSnapshotLogSplit

          -
          public HybridSnapshotLogSplit​(org.apache.fluss.metadata.TableBucket tableBucket,
          -                              @Nullable
          -                              String partitionName,
          -                              long snapshotId,
          -                              long logStartingOffset)
          -
        • -
        - - - -
          -
        • -

          HybridSnapshotLogSplit

          -
          public HybridSnapshotLogSplit​(org.apache.fluss.metadata.TableBucket tableBucket,
          -                              @Nullable
          -                              String partitionName,
          -                              long snapshotId,
          -                              long recordsToSkip,
          -                              boolean isSnapshotFinished,
          -                              long logStartingOffset)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getLogStartingOffset

          -
          public long getLogStartingOffset()
          -
        • -
        - - - -
          -
        • -

          isSnapshotFinished

          -
          public boolean isSnapshotFinished()
          -
        • -
        - - - -
          -
        • -

          splitId

          -
          public String splitId()
          -
        • -
        - - - - - - - - - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/HybridSnapshotLogSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/HybridSnapshotLogSplitState.html deleted file mode 100644 index 9a8c5bf285..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/HybridSnapshotLogSplitState.html +++ /dev/null @@ -1,368 +0,0 @@ - - - - - -HybridSnapshotLogSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class HybridSnapshotLogSplitState

    -
    -
    - -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          HybridSnapshotLogSplitState

          -
          public HybridSnapshotLogSplitState​(HybridSnapshotLogSplit hybridSnapshotLogSplit)
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/LogSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/LogSplit.html deleted file mode 100644 index 62f212715b..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/LogSplit.html +++ /dev/null @@ -1,501 +0,0 @@ - - - - - -LogSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LogSplit

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.api.connector.source.SourceSplit
      -
      -
      -
      public class LogSplit
      -extends SourceSplitBase
      -
      The split for log. It's used to describe the log data of a table bucket.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      - -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LogSplit

          -
          public LogSplit​(org.apache.fluss.metadata.TableBucket tableBucket,
          -                @Nullable
          -                String partitionName,
          -                long startingOffset)
          -
        • -
        - - - -
          -
        • -

          LogSplit

          -
          public LogSplit​(org.apache.fluss.metadata.TableBucket tableBucket,
          -                @Nullable
          -                String partitionName,
          -                long startingOffset,
          -                long stoppingOffset)
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/LogSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/LogSplitState.html deleted file mode 100644 index 443bfd3ceb..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/LogSplitState.html +++ /dev/null @@ -1,354 +0,0 @@ - - - - - -LogSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LogSplitState

    -
    -
    - -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LogSplitState

          -
          public LogSplitState​(LogSplit split)
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SnapshotSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SnapshotSplit.html deleted file mode 100644 index 16e1d08c70..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SnapshotSplit.html +++ /dev/null @@ -1,496 +0,0 @@ - - - - - -SnapshotSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class SnapshotSplit

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.api.connector.source.SourceSplit
      -
      -
      -
      Direct Known Subclasses:
      -
      HybridSnapshotLogSplit
      -
      -
      -
      public abstract class SnapshotSplit
      -extends SourceSplitBase
      -
      The split for snapshot. It's used to describe a snapshot of a table bucket.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - -
          -
        • -

          recordsToSkip

          -
          protected final long recordsToSkip
          -
          The records to skip when reading the snapshot.
          -
        • -
        - - - -
          -
        • -

          snapshotId

          -
          protected final long snapshotId
          -
          The snapshot id. It's used to identify the snapshot for a kv bucket.
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          SnapshotSplit

          -
          public SnapshotSplit​(org.apache.fluss.metadata.TableBucket tableBucket,
          -                     @Nullable
          -                     String partitionName,
          -                     long snapshotId,
          -                     long recordsToSkip)
          -
        • -
        - - - -
          -
        • -

          SnapshotSplit

          -
          public SnapshotSplit​(org.apache.fluss.metadata.TableBucket tableBucket,
          -                     @Nullable
          -                     String partitionName,
          -                     long snapshotId)
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitBase.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitBase.html deleted file mode 100644 index 0ebe839fa9..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitBase.html +++ /dev/null @@ -1,582 +0,0 @@ - - - - - -SourceSplitBase (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class SourceSplitBase

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.split.SourceSplitBase
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - -
          -
        • -

          HYBRID_SNAPSHOT_SPLIT_FLAG

          -
          public static final byte HYBRID_SNAPSHOT_SPLIT_FLAG
          -
          -
          See Also:
          -
          Constant Field Values
          -
          -
        • -
        - - - - - - - -
          -
        • -

          tableBucket

          -
          protected final org.apache.fluss.metadata.TableBucket tableBucket
          -
        • -
        - - - -
          -
        • -

          partitionName

          -
          @Nullable
          -protected final String partitionName
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          SourceSplitBase

          -
          public SourceSplitBase​(org.apache.fluss.metadata.TableBucket tableBucket,
          -                       @Nullable
          -                       String partitionName)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          toSplitId

          -
          protected static String toSplitId​(String splitPrefix,
          -                                  org.apache.fluss.metadata.TableBucket tableBucket)
          -
        • -
        - - - -
          -
        • -

          getTableBucket

          -
          public org.apache.fluss.metadata.TableBucket getTableBucket()
          -
        • -
        - - - -
          -
        • -

          getPartitionName

          -
          @Nullable
          -public String getPartitionName()
          -
        • -
        - - - -
          -
        • -

          isLogSplit

          -
          public final boolean isLogSplit()
          -
          Checks whether this split is a log split.
          -
        • -
        - - - -
          -
        • -

          isLakeSplit

          -
          public boolean isLakeSplit()
          -
        • -
        - - - -
          -
        • -

          isHybridSnapshotLogSplit

          -
          public final boolean isHybridSnapshotLogSplit()
          -
          Checks whether this split is a hybrid snapshot and log split.
          -
        • -
        - - - - - - - -
          -
        • -

          asLogSplit

          -
          public final LogSplit asLogSplit()
          -
          Casts this split into a LogSplit.
          -
        • -
        - - - -
          -
        • -

          splitKind

          -
          protected byte splitKind()
          -
        • -
        - - - -
          -
        • -

          equals

          -
          public boolean equals​(Object o)
          -
          -
          Overrides:
          -
          equals in class Object
          -
          -
        • -
        - - - -
          -
        • -

          hashCode

          -
          public int hashCode()
          -
          -
          Overrides:
          -
          hashCode in class Object
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitSerializer.html deleted file mode 100644 index 6d6f0e9736..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitSerializer.html +++ /dev/null @@ -1,362 +0,0 @@ - - - - - -SourceSplitSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class SourceSplitSerializer

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.split.SourceSplitSerializer
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.core.io.SimpleVersionedSerializer<SourceSplitBase>, org.apache.flink.core.io.Versioned
      -
      -
      -
      public class SourceSplitSerializer
      -extends Object
      -implements org.apache.flink.core.io.SimpleVersionedSerializer<SourceSplitBase>
      -
      A serializer for the SourceSplitBase.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          SourceSplitSerializer

          -
          public SourceSplitSerializer​(org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getVersion

          -
          public int getVersion()
          -
          -
          Specified by:
          -
          getVersion in interface org.apache.flink.core.io.SimpleVersionedSerializer<SourceSplitBase>
          -
          Specified by:
          -
          getVersion in interface org.apache.flink.core.io.Versioned
          -
          -
        • -
        - - - - - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitState.html deleted file mode 100644 index 400493ba79..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/SourceSplitState.html +++ /dev/null @@ -1,435 +0,0 @@ - - - - - -SourceSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class SourceSplitState

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.split.SourceSplitState
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      - -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          SourceSplitState

          -
          public SourceSplitState​(SourceSplitBase split)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          isHybridSnapshotLogSplitState

          -
          public final boolean isHybridSnapshotLogSplitState()
          -
          Checks whether this split state is a hybrid snapshot log split state.
          -
        • -
        - - - -
          -
        • -

          isLogSplitState

          -
          public final boolean isLogSplitState()
          -
          Checks whether this split state is a log split state.
          -
        • -
        - - - - - - - - - - - - - - - -
          -
        • -

          isLakeSplit

          -
          public boolean isLakeSplit()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/HybridSnapshotLogSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/HybridSnapshotLogSplit.html deleted file mode 100644 index c080edbace..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/HybridSnapshotLogSplit.html +++ /dev/null @@ -1,212 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.split.HybridSnapshotLogSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.split.HybridSnapshotLogSplit

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/HybridSnapshotLogSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/HybridSnapshotLogSplitState.html deleted file mode 100644 index ea807e9b17..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/HybridSnapshotLogSplitState.html +++ /dev/null @@ -1,194 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.split.HybridSnapshotLogSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.split.HybridSnapshotLogSplitState

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/LogSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/LogSplit.html deleted file mode 100644 index 351bbac144..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/LogSplit.html +++ /dev/null @@ -1,212 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.split.LogSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.split.LogSplit

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/LogSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/LogSplitState.html deleted file mode 100644 index d8da50545d..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/LogSplitState.html +++ /dev/null @@ -1,194 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.split.LogSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.split.LogSplitState

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SnapshotSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SnapshotSplit.html deleted file mode 100644 index af40d11ab9..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SnapshotSplit.html +++ /dev/null @@ -1,195 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.split.SnapshotSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.split.SnapshotSplit

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitBase.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitBase.html deleted file mode 100644 index f8213f5101..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitBase.html +++ /dev/null @@ -1,733 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.split.SourceSplitBase (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.split.SourceSplitBase

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitSerializer.html deleted file mode 100644 index 89a1e988d1..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitSerializer.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.split.SourceSplitSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.split.SourceSplitSerializer

    -
    -
    No usage of org.apache.fluss.flink.source.split.SourceSplitSerializer
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitState.html deleted file mode 100644 index d9656dc70c..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/class-use/SourceSplitState.html +++ /dev/null @@ -1,364 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.split.SourceSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.split.SourceSplitState

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-summary.html deleted file mode 100644 index 3a1363e0c8..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-summary.html +++ /dev/null @@ -1,211 +0,0 @@ - - - - - -org.apache.fluss.flink.source.split (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.source.split

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-tree.html deleted file mode 100644 index b1b4a7559f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-tree.html +++ /dev/null @@ -1,179 +0,0 @@ - - - - - -org.apache.fluss.flink.source.split Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.source.split

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-use.html deleted file mode 100644 index d7054e5819..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/split/package-use.html +++ /dev/null @@ -1,447 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.source.split (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.source.split

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/FlussSourceEnumeratorStateSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/FlussSourceEnumeratorStateSerializer.html deleted file mode 100644 index 59013f3387..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/FlussSourceEnumeratorStateSerializer.html +++ /dev/null @@ -1,362 +0,0 @@ - - - - - -FlussSourceEnumeratorStateSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlussSourceEnumeratorStateSerializer

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.state.FlussSourceEnumeratorStateSerializer
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.core.io.SimpleVersionedSerializer<SourceEnumeratorState>, org.apache.flink.core.io.Versioned
      -
      -
      -
      public class FlussSourceEnumeratorStateSerializer
      -extends Object
      -implements org.apache.flink.core.io.SimpleVersionedSerializer<SourceEnumeratorState>
      -
      A serializer for SourceEnumeratorState.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlussSourceEnumeratorStateSerializer

          -
          public FlussSourceEnumeratorStateSerializer​(org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> lakeSource)
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/SourceEnumeratorState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/SourceEnumeratorState.html deleted file mode 100644 index f95f9ba6cf..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/SourceEnumeratorState.html +++ /dev/null @@ -1,395 +0,0 @@ - - - - - -SourceEnumeratorState (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class SourceEnumeratorState

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.source.state.SourceEnumeratorState
      • -
      -
    • -
    -
    -
      -
    • -
      -
      public class SourceEnumeratorState
      -extends Object
      -
      A checkpoint of the current state of the containing the buckets that is already assigned.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          SourceEnumeratorState

          -
          public SourceEnumeratorState​(Set<org.apache.fluss.metadata.TableBucket> assignedBuckets,
          -                             Map<Long,​String> assignedPartitions,
          -                             @Nullable
          -                             List<SourceSplitBase> remainingHybridLakeFlussSplits)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getAssignedBuckets

          -
          public Set<org.apache.fluss.metadata.TableBucket> getAssignedBuckets()
          -
        • -
        - - - -
          -
        • -

          getAssignedPartitions

          -
          public Map<Long,​String> getAssignedPartitions()
          -
        • -
        - - - -
          -
        • -

          getRemainingHybridLakeFlussSplits

          -
          @Nullable
          -public List<SourceSplitBase> getRemainingHybridLakeFlussSplits()
          -
        • -
        - - - -
          -
        • -

          equals

          -
          public boolean equals​(Object o)
          -
          -
          Overrides:
          -
          equals in class Object
          -
          -
        • -
        - - - -
          -
        • -

          hashCode

          -
          public int hashCode()
          -
          -
          Overrides:
          -
          hashCode in class Object
          -
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/class-use/FlussSourceEnumeratorStateSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/class-use/FlussSourceEnumeratorStateSerializer.html deleted file mode 100644 index b69d118611..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/class-use/FlussSourceEnumeratorStateSerializer.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.state.FlussSourceEnumeratorStateSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.state.FlussSourceEnumeratorStateSerializer

    -
    -
    No usage of org.apache.fluss.flink.source.state.FlussSourceEnumeratorStateSerializer
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/class-use/SourceEnumeratorState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/class-use/SourceEnumeratorState.html deleted file mode 100644 index 81c4c57e16..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/class-use/SourceEnumeratorState.html +++ /dev/null @@ -1,287 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.source.state.SourceEnumeratorState (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.source.state.SourceEnumeratorState

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-summary.html deleted file mode 100644 index a479c5f931..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-summary.html +++ /dev/null @@ -1,174 +0,0 @@ - - - - - -org.apache.fluss.flink.source.state (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.source.state

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-tree.html deleted file mode 100644 index 632b79ba81..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-tree.html +++ /dev/null @@ -1,164 +0,0 @@ - - - - - -org.apache.fluss.flink.source.state Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.source.state

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-use.html deleted file mode 100644 index b5005bce1b..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/source/state/package-use.html +++ /dev/null @@ -1,231 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.source.state (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.source.state

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/LakeTieringJobBuilder.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/LakeTieringJobBuilder.html deleted file mode 100644 index e5c55dbcf5..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/LakeTieringJobBuilder.html +++ /dev/null @@ -1,292 +0,0 @@ - - - - - -LakeTieringJobBuilder (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LakeTieringJobBuilder

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.tiering.LakeTieringJobBuilder
      • -
      -
    • -
    -
    -
      -
    • -
      -
      public class LakeTieringJobBuilder
      -extends Object
      -
      The builder to build Flink lake tiering job.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          newBuilder

          -
          public static LakeTieringJobBuilder newBuilder​(org.apache.flink.streaming.api.environment.StreamExecutionEnvironment env,
          -                                               org.apache.fluss.config.Configuration flussConfig,
          -                                               org.apache.fluss.config.Configuration dataLakeConfig,
          -                                               String dataLakeFormat)
          -
        • -
        - - - -
          -
        • -

          build

          -
          public org.apache.flink.core.execution.JobClient build()
          -                                                throws Exception
          -
          -
          Throws:
          -
          Exception
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/class-use/LakeTieringJobBuilder.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/class-use/LakeTieringJobBuilder.html deleted file mode 100644 index 2ceba7bc0d..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/class-use/LakeTieringJobBuilder.html +++ /dev/null @@ -1,195 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.LakeTieringJobBuilder (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.LakeTieringJobBuilder

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/CommittableMessage.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/CommittableMessage.html deleted file mode 100644 index be474eaf01..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/CommittableMessage.html +++ /dev/null @@ -1,318 +0,0 @@ - - - - - -CommittableMessage (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class CommittableMessage<Committable>

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.tiering.committer.CommittableMessage<Committable>
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - - - -
          -
        • -

          CommittableMessage

          -
          public CommittableMessage​(Committable committable)
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/CommittableMessageTypeInfo.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/CommittableMessageTypeInfo.html deleted file mode 100644 index 3565d859ca..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/CommittableMessageTypeInfo.html +++ /dev/null @@ -1,485 +0,0 @@ - - - - - -CommittableMessageTypeInfo (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class CommittableMessageTypeInfo<Committable>

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.flink.api.common.typeinfo.TypeInformation<CommittableMessage<Committable>>
      • -
      • -
          -
        • org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo<Committable>
        • -
        -
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          of

          -
          public static <Committable> org.apache.flink.api.common.typeinfo.TypeInformation<CommittableMessage<Committable>> of​(org.apache.flink.util.function.SerializableSupplier<org.apache.fluss.lake.serializer.SimpleVersionedSerializer<Committable>> committableSerializerFactory)
          -
        • -
        - - - -
          -
        • -

          isBasicType

          -
          public boolean isBasicType()
          -
          -
          Specified by:
          -
          isBasicType in class org.apache.flink.api.common.typeinfo.TypeInformation<CommittableMessage<Committable>>
          -
          -
        • -
        - - - -
          -
        • -

          isTupleType

          -
          public boolean isTupleType()
          -
          -
          Specified by:
          -
          isTupleType in class org.apache.flink.api.common.typeinfo.TypeInformation<CommittableMessage<Committable>>
          -
          -
        • -
        - - - -
          -
        • -

          getArity

          -
          public int getArity()
          -
          -
          Specified by:
          -
          getArity in class org.apache.flink.api.common.typeinfo.TypeInformation<CommittableMessage<Committable>>
          -
          -
        • -
        - - - -
          -
        • -

          getTotalFields

          -
          public int getTotalFields()
          -
          -
          Specified by:
          -
          getTotalFields in class org.apache.flink.api.common.typeinfo.TypeInformation<CommittableMessage<Committable>>
          -
          -
        • -
        - - - - - - - -
          -
        • -

          isKeyType

          -
          public boolean isKeyType()
          -
          -
          Specified by:
          -
          isKeyType in class org.apache.flink.api.common.typeinfo.TypeInformation<CommittableMessage<Committable>>
          -
          -
        • -
        - - - -
          -
        • -

          createSerializer

          -
          public org.apache.flink.api.common.typeutils.TypeSerializer<CommittableMessage<Committable>> createSerializer​(org.apache.flink.api.common.ExecutionConfig executionConfig)
          -
          -
          Specified by:
          -
          createSerializer in class org.apache.flink.api.common.typeinfo.TypeInformation<CommittableMessage<Committable>>
          -
          -
        • -
        - - - - - - - -
          -
        • -

          equals

          -
          public boolean equals​(Object obj)
          -
          -
          Specified by:
          -
          equals in class org.apache.flink.api.common.typeinfo.TypeInformation<CommittableMessage<Committable>>
          -
          -
        • -
        - - - -
          -
        • -

          hashCode

          -
          public int hashCode()
          -
          -
          Specified by:
          -
          hashCode in class org.apache.flink.api.common.typeinfo.TypeInformation<CommittableMessage<Committable>>
          -
          -
        • -
        - - - -
          -
        • -

          canEqual

          -
          public boolean canEqual​(Object obj)
          -
          -
          Specified by:
          -
          canEqual in class org.apache.flink.api.common.typeinfo.TypeInformation<CommittableMessage<Committable>>
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/FlussTableLakeSnapshotCommitter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/FlussTableLakeSnapshotCommitter.html deleted file mode 100644 index fd5c357406..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/FlussTableLakeSnapshotCommitter.html +++ /dev/null @@ -1,354 +0,0 @@ - - - - - -FlussTableLakeSnapshotCommitter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlussTableLakeSnapshotCommitter

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.tiering.committer.FlussTableLakeSnapshotCommitter
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      AutoCloseable
      -
      -
      -
      public class FlussTableLakeSnapshotCommitter
      -extends Object
      -implements AutoCloseable
      -
      Committer to commit FlussTableLakeSnapshot of lake to Fluss.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlussTableLakeSnapshotCommitter

          -
          public FlussTableLakeSnapshotCommitter​(org.apache.fluss.config.Configuration flussConf)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          open

          -
          public void open()
          -
        • -
        - - - -
          -
        • -

          commit

          -
          public void commit​(long tableId,
          -                   org.apache.fluss.lake.committer.CommittedLakeSnapshot committedLakeSnapshot)
          -            throws IOException
          -
          -
          Throws:
          -
          IOException
          -
          -
        • -
        - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitOperator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitOperator.html deleted file mode 100644 index 955d05fd87..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitOperator.html +++ /dev/null @@ -1,499 +0,0 @@ - - - - - -TieringCommitOperator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TieringCommitOperator<WriteResult,​Committable>

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.flink.streaming.api.operators.AbstractStreamOperator<CommittableMessage<Committable>>
      • -
      • -
          -
        • org.apache.fluss.flink.tiering.committer.TieringCommitOperator<WriteResult,​Committable>
        • -
        -
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      Serializable, org.apache.flink.api.common.state.CheckpointListener, org.apache.flink.streaming.api.operators.Input<TableBucketWriteResult<WriteResult>>, org.apache.flink.streaming.api.operators.KeyContext, org.apache.flink.streaming.api.operators.KeyContextHandler, org.apache.flink.streaming.api.operators.OneInputStreamOperator<TableBucketWriteResult<WriteResult>,​CommittableMessage<Committable>>, org.apache.flink.streaming.api.operators.SetupableStreamOperator<CommittableMessage<Committable>>, org.apache.flink.streaming.api.operators.StreamOperator<CommittableMessage<Committable>>, org.apache.flink.streaming.api.operators.StreamOperatorStateHandler.CheckpointedStreamOperator, org.apache.flink.streaming.api.operators.YieldingOperator<CommittableMessage<Committable>>
      -
      -
      -
      public class TieringCommitOperator<WriteResult,​Committable>
      -extends org.apache.flink.streaming.api.operators.AbstractStreamOperator<CommittableMessage<Committable>>
      -implements org.apache.flink.streaming.api.operators.OneInputStreamOperator<TableBucketWriteResult<WriteResult>,​CommittableMessage<Committable>>
      -
      A Flink operator to aggregate TieringCommitOperators by table to TieringCommitOperator which will - then be committed to lake & Fluss cluster. - -

      It will collect all TableBucketWriteResults which wraps TieringCommitOperator written by - LakeWriter in TieringSource operator. - -

      When it collects all TableBucketWriteResults of a round of tiering for a table, it - will combine all the TieringCommitOperators to TieringCommitOperator via method LakeCommitter.toCommittable(List), and then call method LakeCommitter.commit(Object, - Map) to commit to lake. - -

      Finally, it will also commit the committed lake snapshot to Fluss cluster to make Fluss aware - of the tiering progress.

      -
      -
      See Also:
      -
      Serialized Form
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Summary

        -
          -
        • - - -

          Fields inherited from class org.apache.flink.streaming.api.operators.AbstractStreamOperator

          -chainingStrategy, config, lastRecordAttributes1, lastRecordAttributes2, latencyStats, LOG, metrics, output, processingTimeService, stateHandler, stateKeySelector1, stateKeySelector2, timeServiceManager
        • -
        -
      • -
      -
      - -
      - -
      - -
      -
        -
      • - - -

        Method Summary

        - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
        All Methods Static Methods Instance Methods Concrete Methods 
        Modifier and TypeMethodDescription
        voidclose() 
        voidopen() 
        voidprocessElement​(org.apache.flink.streaming.runtime.streamrecord.StreamRecord<TableBucketWriteResult<WriteResult>> streamRecord) 
        voidsetup​(org.apache.flink.streaming.runtime.tasks.StreamTask<?,​?> containingTask, - org.apache.flink.streaming.api.graph.StreamConfig config, - org.apache.flink.streaming.api.operators.Output<org.apache.flink.streaming.runtime.streamrecord.StreamRecord<CommittableMessage<Committable>>> output) 
        static Map<String,​String>toBucketOffsetsProperty​(Map<org.apache.fluss.metadata.TableBucket,​Long> tableBucketOffsets, - Map<Long,​String> partitionNameById, - List<String> partitionKeys) 
        -
          -
        • - - -

          Methods inherited from class org.apache.flink.streaming.api.operators.AbstractStreamOperator

          -finish, getChainingStrategy, getContainingTask, getCurrentKey, getExecutionConfig, getInternalTimerService, getKeyedStateBackend, getKeyedStateStore, getMetricGroup, getOperatorConfig, getOperatorID, getOperatorName, getOperatorStateBackend, getOrCreateKeyedState, getPartitionedState, getPartitionedState, getProcessingTimeService, getRuntimeContext, getStateKeySelector1, getStateKeySelector2, getTimeServiceManager, getUserCodeClassloader, hasKeyContext1, hasKeyContext2, initializeState, initializeState, isUsingCustomRawKeyedState, notifyCheckpointAborted, notifyCheckpointComplete, prepareSnapshotPreBarrier, processLatencyMarker, processLatencyMarker1, processLatencyMarker2, processRecordAttributes, processRecordAttributes1, processRecordAttributes2, processWatermark, processWatermark1, processWatermark2, processWatermarkStatus, processWatermarkStatus1, processWatermarkStatus2, reportOrForwardLatencyMarker, setChainingStrategy, setCurrentKey, setKeyContextElement1, setKeyContextElement2, setMailboxExecutor, setProcessingTimeService, snapshotState, snapshotState, useSplittableTimers
        • -
        - -
          -
        • - - -

          Methods inherited from interface org.apache.flink.api.common.state.CheckpointListener

          -notifyCheckpointAborted, notifyCheckpointComplete
        • -
        -
          -
        • - - -

          Methods inherited from interface org.apache.flink.streaming.api.operators.Input

          -processLatencyMarker, processRecordAttributes, processWatermark, processWatermarkStatus
        • -
        -
          -
        • - - -

          Methods inherited from interface org.apache.flink.streaming.api.operators.KeyContext

          -getCurrentKey, setCurrentKey
        • -
        -
          -
        • - - -

          Methods inherited from interface org.apache.flink.streaming.api.operators.KeyContextHandler

          -hasKeyContext
        • -
        -
          -
        • - - -

          Methods inherited from interface org.apache.flink.streaming.api.operators.OneInputStreamOperator

          -setKeyContextElement
        • -
        -
          -
        • - - -

          Methods inherited from interface org.apache.flink.streaming.api.operators.StreamOperator

          -finish, getMetricGroup, getOperatorAttributes, getOperatorID, initializeState, prepareSnapshotPreBarrier, setKeyContextElement1, setKeyContextElement2, snapshotState
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          TieringCommitOperator

          -
          public TieringCommitOperator​(org.apache.flink.streaming.api.operators.StreamOperatorParameters<CommittableMessage<Committable>> parameters,
          -                             org.apache.fluss.config.Configuration flussConf,
          -                             org.apache.fluss.lake.writer.LakeTieringFactory<WriteResult,​Committable> lakeTieringFactory)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          setup

          -
          public void setup​(org.apache.flink.streaming.runtime.tasks.StreamTask<?,​?> containingTask,
          -                  org.apache.flink.streaming.api.graph.StreamConfig config,
          -                  org.apache.flink.streaming.api.operators.Output<org.apache.flink.streaming.runtime.streamrecord.StreamRecord<CommittableMessage<Committable>>> output)
          -
          -
          Specified by:
          -
          setup in interface org.apache.flink.streaming.api.operators.SetupableStreamOperator<WriteResult>
          -
          Overrides:
          -
          setup in class org.apache.flink.streaming.api.operators.AbstractStreamOperator<CommittableMessage<Committable>>
          -
          -
        • -
        - - - -
          -
        • -

          open

          -
          public void open()
          -
          -
          Specified by:
          -
          open in interface org.apache.flink.streaming.api.operators.StreamOperator<WriteResult>
          -
          Overrides:
          -
          open in class org.apache.flink.streaming.api.operators.AbstractStreamOperator<CommittableMessage<Committable>>
          -
          -
        • -
        - - - -
          -
        • -

          processElement

          -
          public void processElement​(org.apache.flink.streaming.runtime.streamrecord.StreamRecord<TableBucketWriteResult<WriteResult>> streamRecord)
          -                    throws Exception
          -
          -
          Specified by:
          -
          processElement in interface org.apache.flink.streaming.api.operators.Input<WriteResult>
          -
          Throws:
          -
          Exception
          -
          -
        • -
        - - - - - - - -
          -
        • -

          close

          -
          public void close()
          -           throws Exception
          -
          -
          Specified by:
          -
          close in interface org.apache.flink.streaming.api.operators.StreamOperator<WriteResult>
          -
          Overrides:
          -
          close in class org.apache.flink.streaming.api.operators.AbstractStreamOperator<CommittableMessage<Committable>>
          -
          Throws:
          -
          Exception
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitOperatorFactory.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitOperatorFactory.html deleted file mode 100644 index 9ae414a189..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitOperatorFactory.html +++ /dev/null @@ -1,376 +0,0 @@ - - - - - -TieringCommitOperatorFactory (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TieringCommitOperatorFactory<WriteResult,​Committable>

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.flink.streaming.api.operators.AbstractStreamOperatorFactory<CommittableMessage<Committable>>
      • -
      • -
          -
        • org.apache.fluss.flink.tiering.committer.TieringCommitOperatorFactory<WriteResult,​Committable>
        • -
        -
      • -
      -
    • -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Summary

        -
          -
        • - - -

          Fields inherited from class org.apache.flink.streaming.api.operators.AbstractStreamOperatorFactory

          -chainingStrategy, processingTimeService
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Summary

        - - - - - - - - - - -
        Constructors 
        ConstructorDescription
        TieringCommitOperatorFactory​(org.apache.fluss.config.Configuration flussConfig, - org.apache.fluss.lake.writer.LakeTieringFactory<WriteResult,​Committable> lakeTieringFactory) 
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          TieringCommitOperatorFactory

          -
          public TieringCommitOperatorFactory​(org.apache.fluss.config.Configuration flussConfig,
          -                                    org.apache.fluss.lake.writer.LakeTieringFactory<WriteResult,​Committable> lakeTieringFactory)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          createStreamOperator

          -
          public <T extends org.apache.flink.streaming.api.operators.StreamOperator<CommittableMessage<Committable>>> T createStreamOperator​(org.apache.flink.streaming.api.operators.StreamOperatorParameters<CommittableMessage<Committable>> parameters)
          -
          -
          Specified by:
          -
          createStreamOperator in interface org.apache.flink.streaming.api.operators.StreamOperatorFactory<WriteResult>
          -
          -
        • -
        - - - -
          -
        • -

          getStreamOperatorClass

          -
          public Class<? extends org.apache.flink.streaming.api.operators.StreamOperator> getStreamOperatorClass​(ClassLoader classLoader)
          -
          -
          Specified by:
          -
          getStreamOperatorClass in interface org.apache.flink.streaming.api.operators.StreamOperatorFactory<WriteResult>
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitterInitContext.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitterInitContext.html deleted file mode 100644 index fb7679febb..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/TieringCommitterInitContext.html +++ /dev/null @@ -1,316 +0,0 @@ - - - - - -TieringCommitterInitContext (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TieringCommitterInitContext

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.tiering.committer.TieringCommitterInitContext
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.fluss.lake.committer.CommitterInitContext
      -
      -
      -
      public class TieringCommitterInitContext
      -extends Object
      -implements org.apache.fluss.lake.committer.CommitterInitContext
      -
      The CommitterInitContext implementation for LakeCommitter.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          TieringCommitterInitContext

          -
          public TieringCommitterInitContext​(org.apache.fluss.metadata.TablePath tablePath)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          tablePath

          -
          public org.apache.fluss.metadata.TablePath tablePath()
          -
          -
          Specified by:
          -
          tablePath in interface org.apache.fluss.lake.committer.CommitterInitContext
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/CommittableMessage.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/CommittableMessage.html deleted file mode 100644 index 76ebd86258..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/CommittableMessage.html +++ /dev/null @@ -1,254 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.committer.CommittableMessage (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.committer.CommittableMessage

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/CommittableMessageTypeInfo.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/CommittableMessageTypeInfo.html deleted file mode 100644 index cf00c5e948..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/CommittableMessageTypeInfo.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo

    -
    -
    No usage of org.apache.fluss.flink.tiering.committer.CommittableMessageTypeInfo
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/FlussTableLakeSnapshotCommitter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/FlussTableLakeSnapshotCommitter.html deleted file mode 100644 index 26d52e1e42..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/FlussTableLakeSnapshotCommitter.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.committer.FlussTableLakeSnapshotCommitter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.committer.FlussTableLakeSnapshotCommitter

    -
    -
    No usage of org.apache.fluss.flink.tiering.committer.FlussTableLakeSnapshotCommitter
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitOperator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitOperator.html deleted file mode 100644 index ca3e77cd2a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitOperator.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.committer.TieringCommitOperator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.committer.TieringCommitOperator

    -
    -
    No usage of org.apache.fluss.flink.tiering.committer.TieringCommitOperator
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitOperatorFactory.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitOperatorFactory.html deleted file mode 100644 index 69ea4c4910..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitOperatorFactory.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.committer.TieringCommitOperatorFactory (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.committer.TieringCommitOperatorFactory

    -
    -
    No usage of org.apache.fluss.flink.tiering.committer.TieringCommitOperatorFactory
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitterInitContext.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitterInitContext.html deleted file mode 100644 index f0f3b237e3..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/class-use/TieringCommitterInitContext.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.committer.TieringCommitterInitContext (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.committer.TieringCommitterInitContext

    -
    -
    No usage of org.apache.fluss.flink.tiering.committer.TieringCommitterInitContext
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-summary.html deleted file mode 100644 index 481cf9e146..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-summary.html +++ /dev/null @@ -1,199 +0,0 @@ - - - - - -org.apache.fluss.flink.tiering.committer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.tiering.committer

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-tree.html deleted file mode 100644 index bb0c1c75c4..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-tree.html +++ /dev/null @@ -1,180 +0,0 @@ - - - - - -org.apache.fluss.flink.tiering.committer Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.tiering.committer

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    -
      -
    • java.lang.Object -
        -
      • org.apache.flink.streaming.api.operators.AbstractStreamOperator<OUT> (implements org.apache.flink.streaming.api.operators.KeyContextHandler, java.io.Serializable, org.apache.flink.streaming.api.operators.SetupableStreamOperator<OUT>, org.apache.flink.streaming.api.operators.StreamOperator<OUT>, org.apache.flink.streaming.api.operators.StreamOperatorStateHandler.CheckpointedStreamOperator, org.apache.flink.streaming.api.operators.YieldingOperator<OUT>) -
          -
        • org.apache.fluss.flink.tiering.committer.TieringCommitOperator<WriteResult,​Committable> (implements org.apache.flink.streaming.api.operators.OneInputStreamOperator<IN,​OUT>)
        • -
        -
      • -
      • org.apache.flink.streaming.api.operators.AbstractStreamOperatorFactory<OUT> (implements org.apache.flink.streaming.runtime.tasks.ProcessingTimeServiceAware, org.apache.flink.streaming.api.operators.StreamOperatorFactory<OUT>) -
          -
        • org.apache.fluss.flink.tiering.committer.TieringCommitOperatorFactory<WriteResult,​Committable> (implements org.apache.flink.streaming.api.operators.OneInputStreamOperatorFactory<IN,​OUT>)
        • -
        -
      • -
      • org.apache.fluss.flink.tiering.committer.CommittableMessage<Committable> (implements java.io.Serializable)
      • -
      • org.apache.fluss.flink.tiering.committer.FlussTableLakeSnapshotCommitter (implements java.lang.AutoCloseable)
      • -
      • org.apache.fluss.flink.tiering.committer.TieringCommitterInitContext (implements org.apache.fluss.lake.committer.CommitterInitContext)
      • -
      • org.apache.flink.api.common.typeinfo.TypeInformation<T> (implements java.io.Serializable) - -
      • -
      -
    • -
    -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-use.html deleted file mode 100644 index ec2b9b3423..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/committer/package-use.html +++ /dev/null @@ -1,185 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.tiering.committer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.tiering.committer

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/FailedTieringEvent.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/FailedTieringEvent.html deleted file mode 100644 index e563a36a04..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/FailedTieringEvent.html +++ /dev/null @@ -1,332 +0,0 @@ - - - - - -FailedTieringEvent (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FailedTieringEvent

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.tiering.event.FailedTieringEvent
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      Serializable, org.apache.flink.api.connector.source.SourceEvent
      -
      -
      -
      public class FailedTieringEvent
      -extends Object
      -implements org.apache.flink.api.connector.source.SourceEvent
      -
      SourceEvent used to represent a Fluss table is failed during tiering.
      -
      -
      See Also:
      -
      Serialized Form
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FailedTieringEvent

          -
          public FailedTieringEvent​(long tableId,
          -                          String failReason)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getTableId

          -
          public long getTableId()
          -
        • -
        - - - -
          -
        • -

          failReason

          -
          public String failReason()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/FinishedTieringEvent.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/FinishedTieringEvent.html deleted file mode 100644 index c5e2342b00..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/FinishedTieringEvent.html +++ /dev/null @@ -1,316 +0,0 @@ - - - - - -FinishedTieringEvent (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FinishedTieringEvent

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.tiering.event.FinishedTieringEvent
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      Serializable, org.apache.flink.api.connector.source.SourceEvent
      -
      -
      -
      public class FinishedTieringEvent
      -extends Object
      -implements org.apache.flink.api.connector.source.SourceEvent
      -
      SourceEvent used to represent a Fluss table has been tiered finished.
      -
      -
      See Also:
      -
      Serialized Form
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FinishedTieringEvent

          -
          public FinishedTieringEvent​(long tableId)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getTableId

          -
          public long getTableId()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/TieringRestoreEvent.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/TieringRestoreEvent.html deleted file mode 100644 index 79b77386e6..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/TieringRestoreEvent.html +++ /dev/null @@ -1,278 +0,0 @@ - - - - - -TieringRestoreEvent (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TieringRestoreEvent

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.tiering.event.TieringRestoreEvent
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      Serializable, org.apache.flink.api.connector.source.SourceEvent
      -
      -
      -
      public class TieringRestoreEvent
      -extends Object
      -implements org.apache.flink.api.connector.source.SourceEvent
      -
      SourceEvent used to represent tiering is restoring.
      -
      -
      See Also:
      -
      Serialized Form
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          TieringRestoreEvent

          -
          public TieringRestoreEvent()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/FailedTieringEvent.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/FailedTieringEvent.html deleted file mode 100644 index 63e3497efd..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/FailedTieringEvent.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.event.FailedTieringEvent (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.event.FailedTieringEvent

    -
    -
    No usage of org.apache.fluss.flink.tiering.event.FailedTieringEvent
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/FinishedTieringEvent.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/FinishedTieringEvent.html deleted file mode 100644 index f735faa31a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/FinishedTieringEvent.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.event.FinishedTieringEvent (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.event.FinishedTieringEvent

    -
    -
    No usage of org.apache.fluss.flink.tiering.event.FinishedTieringEvent
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/TieringRestoreEvent.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/TieringRestoreEvent.html deleted file mode 100644 index 05989a26c2..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/class-use/TieringRestoreEvent.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.event.TieringRestoreEvent (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.event.TieringRestoreEvent

    -
    -
    No usage of org.apache.fluss.flink.tiering.event.TieringRestoreEvent
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-summary.html deleted file mode 100644 index dfa22d3083..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-summary.html +++ /dev/null @@ -1,180 +0,0 @@ - - - - - -org.apache.fluss.flink.tiering.event (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.tiering.event

    -
    -
    -
      -
    • - - - - - - - - - - - - - - - - - - - - -
      Class Summary 
      ClassDescription
      FailedTieringEvent -
      SourceEvent used to represent a Fluss table is failed during tiering.
      -
      FinishedTieringEvent -
      SourceEvent used to represent a Fluss table has been tiered finished.
      -
      TieringRestoreEvent -
      SourceEvent used to represent tiering is restoring.
      -
      -
    • -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-tree.html deleted file mode 100644 index 9becc30f22..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-tree.html +++ /dev/null @@ -1,165 +0,0 @@ - - - - - -org.apache.fluss.flink.tiering.event Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.tiering.event

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    -
      -
    • java.lang.Object -
        -
      • org.apache.fluss.flink.tiering.event.FailedTieringEvent (implements org.apache.flink.api.connector.source.SourceEvent)
      • -
      • org.apache.fluss.flink.tiering.event.FinishedTieringEvent (implements org.apache.flink.api.connector.source.SourceEvent)
      • -
      • org.apache.fluss.flink.tiering.event.TieringRestoreEvent (implements org.apache.flink.api.connector.source.SourceEvent)
      • -
      -
    • -
    -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-use.html deleted file mode 100644 index 5a4d6075d0..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/event/package-use.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.tiering.event (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.tiering.event

    -
    -
    No usage of org.apache.fluss.flink.tiering.event
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-summary.html deleted file mode 100644 index 04a9a2fb1f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-summary.html +++ /dev/null @@ -1,168 +0,0 @@ - - - - - -org.apache.fluss.flink.tiering (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.tiering

    -
    -
    -
      -
    • - - - - - - - - - - - - -
      Class Summary 
      ClassDescription
      LakeTieringJobBuilder -
      The builder to build Flink lake tiering job.
      -
      -
    • -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-tree.html deleted file mode 100644 index df81712c57..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-tree.html +++ /dev/null @@ -1,163 +0,0 @@ - - - - - -org.apache.fluss.flink.tiering Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.tiering

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-use.html deleted file mode 100644 index db2b4586b0..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/package-use.html +++ /dev/null @@ -1,185 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.tiering (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.tiering

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResult.html deleted file mode 100644 index cded453a08..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResult.html +++ /dev/null @@ -1,405 +0,0 @@ - - - - - -TableBucketWriteResult (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TableBucketWriteResult<WriteResult>

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.tiering.source.TableBucketWriteResult<WriteResult>
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      Serializable
      -
      -
      -
      public class TableBucketWriteResult<WriteResult>
      -extends Object
      -implements Serializable
      -
      This class contains the TableBucketWriteResult of LakeWriter, the table path and the bucket - that the write result is for, the end log offset of tiering, the total number of write results in - one round of tiering. It'll be passed to downstream committer operator to collect all the write - results of a table and do commit.
      -
      -
      See Also:
      -
      Serialized Form
      -
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - - - -
          -
        • -

          TableBucketWriteResult

          -
          public TableBucketWriteResult​(org.apache.fluss.metadata.TablePath tablePath,
          -                              org.apache.fluss.metadata.TableBucket tableBucket,
          -                              @Nullable
          -                              String partitionName,
          -                              @Nullable
          -                              WriteResult writeResult,
          -                              long logEndOffset,
          -                              int numberOfWriteResults)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          tablePath

          -
          public org.apache.fluss.metadata.TablePath tablePath()
          -
        • -
        - - - -
          -
        • -

          tableBucket

          -
          public org.apache.fluss.metadata.TableBucket tableBucket()
          -
        • -
        - - - -
          -
        • -

          partitionName

          -
          @Nullable
          -public String partitionName()
          -
        • -
        - - - -
          -
        • -

          writeResult

          -
          @Nullable
          -public WriteResult writeResult()
          -
        • -
        - - - -
          -
        • -

          numberOfWriteResults

          -
          public int numberOfWriteResults()
          -
        • -
        - - - -
          -
        • -

          logEndOffset

          -
          public long logEndOffset()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultEmitter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultEmitter.html deleted file mode 100644 index 2eb6827391..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultEmitter.html +++ /dev/null @@ -1,320 +0,0 @@ - - - - - -TableBucketWriteResultEmitter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TableBucketWriteResultEmitter<WriteResult>

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.tiering.source.TableBucketWriteResultEmitter<WriteResult>
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    - -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultSerializer.html deleted file mode 100644 index 5b3ddd4ece..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultSerializer.html +++ /dev/null @@ -1,362 +0,0 @@ - - - - - -TableBucketWriteResultSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TableBucketWriteResultSerializer<WriteResult>

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.tiering.source.TableBucketWriteResultSerializer<WriteResult>
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.core.io.SimpleVersionedSerializer<TableBucketWriteResult<WriteResult>>, org.apache.flink.core.io.Versioned
      -
      -
      -
      public class TableBucketWriteResultSerializer<WriteResult>
      -extends Object
      -implements org.apache.flink.core.io.SimpleVersionedSerializer<TableBucketWriteResult<WriteResult>>
      -
      The serializer for TableBucketWriteResult.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          TableBucketWriteResultSerializer

          -
          public TableBucketWriteResultSerializer​(org.apache.fluss.lake.serializer.SimpleVersionedSerializer<WriteResult> writeResultSerializer)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getVersion

          -
          public int getVersion()
          -
          -
          Specified by:
          -
          getVersion in interface org.apache.flink.core.io.SimpleVersionedSerializer<WriteResult>
          -
          Specified by:
          -
          getVersion in interface org.apache.flink.core.io.Versioned
          -
          -
        • -
        - - - - - - - - -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultTypeInfo.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultTypeInfo.html deleted file mode 100644 index 37c42bdab6..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TableBucketWriteResultTypeInfo.html +++ /dev/null @@ -1,485 +0,0 @@ - - - - - -TableBucketWriteResultTypeInfo (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TableBucketWriteResultTypeInfo<WriteResult>

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.flink.api.common.typeinfo.TypeInformation<TableBucketWriteResult<WriteResult>>
      • -
      • -
          -
        • org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo<WriteResult>
        • -
        -
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    - -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSource.Builder.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSource.Builder.html deleted file mode 100644 index 0fd59d2c86..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSource.Builder.html +++ /dev/null @@ -1,327 +0,0 @@ - - - - - -TieringSource.Builder (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TieringSource.Builder<WriteResult>

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.tiering.source.TieringSource.Builder<WriteResult>
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          Builder

          -
          public Builder​(org.apache.fluss.config.Configuration flussConf,
          -               org.apache.fluss.lake.writer.LakeTieringFactory<WriteResult,​?> lakeTieringFactory)
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSource.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSource.html deleted file mode 100644 index 4b6563c3ed..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSource.html +++ /dev/null @@ -1,514 +0,0 @@ - - - - - -TieringSource (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TieringSource<WriteResult>

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.tiering.source.TieringSource<WriteResult>
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    - -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSourceOptions.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSourceOptions.html deleted file mode 100644 index d3d667a0c4..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSourceOptions.html +++ /dev/null @@ -1,329 +0,0 @@ - - - - - -TieringSourceOptions (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TieringSourceOptions

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.tiering.source.TieringSourceOptions
      • -
      -
    • -
    -
    -
      -
    • -
      -
      public class TieringSourceOptions
      -extends Object
      -
      Configuration options for the TieringSource.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - - - - - -
          -
        • -

          POLL_TIERING_TABLE_INTERVAL

          -
          public static final org.apache.fluss.config.ConfigOption<Duration> POLL_TIERING_TABLE_INTERVAL
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          TieringSourceOptions

          -
          public TieringSourceOptions()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSourceReader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSourceReader.html deleted file mode 100644 index 05d7dd6881..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSourceReader.html +++ /dev/null @@ -1,446 +0,0 @@ - - - - - -TieringSourceReader (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TieringSourceReader<WriteResult>

    -
    -
    - -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Summary

        -
          -
        • - - -

          Fields inherited from class org.apache.flink.connector.base.source.reader.SourceReaderBase

          -config, context, eofRecordEvaluator, options, recordEmitter, splitFetcherManager
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Summary

        - - - - - - - - - - -
        Constructors 
        ConstructorDescription
        TieringSourceReader​(org.apache.flink.api.connector.source.SourceReaderContext context, - org.apache.fluss.config.Configuration flussConf, - org.apache.fluss.lake.writer.LakeTieringFactory<WriteResult,​?> lakeTieringFactory) 
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    - -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSplitReader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSplitReader.html deleted file mode 100644 index b1c1b86349..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringSplitReader.html +++ /dev/null @@ -1,385 +0,0 @@ - - - - - -TieringSplitReader (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TieringSplitReader<WriteResult>

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.tiering.source.TieringSplitReader<WriteResult>
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      AutoCloseable, org.apache.flink.connector.base.source.reader.splitreader.SplitReader<TableBucketWriteResult<WriteResult>,​TieringSplit>
      -
      -
      -
      public class TieringSplitReader<WriteResult>
      -extends Object
      -implements org.apache.flink.connector.base.source.reader.splitreader.SplitReader<TableBucketWriteResult<WriteResult>,​TieringSplit>
      -
      The SplitReader implementation which will read Fluss and write to lake.
      -
    • -
    -
    -
    - -
    -
    - -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringWriterInitContext.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringWriterInitContext.html deleted file mode 100644 index b4faab11cf..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/TieringWriterInitContext.html +++ /dev/null @@ -1,378 +0,0 @@ - - - - - -TieringWriterInitContext (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TieringWriterInitContext

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.tiering.source.TieringWriterInitContext
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.fluss.lake.writer.WriterInitContext
      -
      -
      -
      public class TieringWriterInitContext
      -extends Object
      -implements org.apache.fluss.lake.writer.WriterInitContext
      -
      The implementation of WriterInitContext.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          TieringWriterInitContext

          -
          public TieringWriterInitContext​(org.apache.fluss.metadata.TablePath tablePath,
          -                                org.apache.fluss.metadata.TableBucket tableBucket,
          -                                @Nullable
          -                                String partition,
          -                                org.apache.fluss.metadata.TableInfo tableInfo)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          tablePath

          -
          public org.apache.fluss.metadata.TablePath tablePath()
          -
          -
          Specified by:
          -
          tablePath in interface org.apache.fluss.lake.writer.WriterInitContext
          -
          -
        • -
        - - - -
          -
        • -

          tableBucket

          -
          public org.apache.fluss.metadata.TableBucket tableBucket()
          -
          -
          Specified by:
          -
          tableBucket in interface org.apache.fluss.lake.writer.WriterInitContext
          -
          -
        • -
        - - - -
          -
        • -

          partition

          -
          @Nullable
          -public String partition()
          -
          -
          Specified by:
          -
          partition in interface org.apache.fluss.lake.writer.WriterInitContext
          -
          -
        • -
        - - - -
          -
        • -

          tableInfo

          -
          public org.apache.fluss.metadata.TableInfo tableInfo()
          -
          -
          Specified by:
          -
          tableInfo in interface org.apache.fluss.lake.writer.WriterInitContext
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResult.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResult.html deleted file mode 100644 index 23010e9cb9..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResult.html +++ /dev/null @@ -1,293 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.source.TableBucketWriteResult (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.source.TableBucketWriteResult

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultEmitter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultEmitter.html deleted file mode 100644 index 518b79232f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultEmitter.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.source.TableBucketWriteResultEmitter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.source.TableBucketWriteResultEmitter

    -
    -
    No usage of org.apache.fluss.flink.tiering.source.TableBucketWriteResultEmitter
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultSerializer.html deleted file mode 100644 index bac06c06c1..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultSerializer.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.source.TableBucketWriteResultSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.source.TableBucketWriteResultSerializer

    -
    -
    No usage of org.apache.fluss.flink.tiering.source.TableBucketWriteResultSerializer
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultTypeInfo.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultTypeInfo.html deleted file mode 100644 index 10688df6da..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TableBucketWriteResultTypeInfo.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo

    -
    -
    No usage of org.apache.fluss.flink.tiering.source.TableBucketWriteResultTypeInfo
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSource.Builder.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSource.Builder.html deleted file mode 100644 index 4e5b587baa..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSource.Builder.html +++ /dev/null @@ -1,192 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.source.TieringSource.Builder (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.source.TieringSource.Builder

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSource.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSource.html deleted file mode 100644 index 2396b530af..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSource.html +++ /dev/null @@ -1,192 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.source.TieringSource (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.source.TieringSource

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSourceOptions.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSourceOptions.html deleted file mode 100644 index b9f0769313..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSourceOptions.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.source.TieringSourceOptions (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.source.TieringSourceOptions

    -
    -
    No usage of org.apache.fluss.flink.tiering.source.TieringSourceOptions
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSourceReader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSourceReader.html deleted file mode 100644 index b02e21be8c..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSourceReader.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.source.TieringSourceReader (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.source.TieringSourceReader

    -
    -
    No usage of org.apache.fluss.flink.tiering.source.TieringSourceReader
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSplitReader.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSplitReader.html deleted file mode 100644 index 1443bbc623..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringSplitReader.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.source.TieringSplitReader (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.source.TieringSplitReader

    -
    -
    No usage of org.apache.fluss.flink.tiering.source.TieringSplitReader
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringWriterInitContext.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringWriterInitContext.html deleted file mode 100644 index 374ebb3728..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/class-use/TieringWriterInitContext.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.source.TieringWriterInitContext (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.source.TieringWriterInitContext

    -
    -
    No usage of org.apache.fluss.flink.tiering.source.TieringWriterInitContext
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/TieringSourceEnumerator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/TieringSourceEnumerator.html deleted file mode 100644 index a09c14db92..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/TieringSourceEnumerator.html +++ /dev/null @@ -1,468 +0,0 @@ - - - - - -TieringSourceEnumerator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TieringSourceEnumerator

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      AutoCloseable, org.apache.flink.api.common.state.CheckpointListener, org.apache.flink.api.connector.source.SplitEnumerator<TieringSplit,​TieringSourceEnumeratorState>
      -
      -
      -
      public class TieringSourceEnumerator
      -extends Object
      -implements org.apache.flink.api.connector.source.SplitEnumerator<TieringSplit,​TieringSourceEnumeratorState>
      -
      An implementation of SplitEnumerator used to request TieringSplit from Fluss - Cluster. - -

      The enumerator is responsible for: - -

        -
      • Register the Tiering Service job that the current TieringSourceEnumerator belongs to with - the Fluss Cluster when the Flink Tiering job starts up. -
      • Request Fluss table splits from Fluss Cluster and assigns to SourceReader to tier. -
      • Un-Register the Tiering Service job that the current TieringSourceEnumerator belongs to - with the Fluss Cluster when the Flink Tiering job shutdown as much as possible. -
      -
    • -
    -
    -
    - -
    -
    - -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/class-use/TieringSourceEnumerator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/class-use/TieringSourceEnumerator.html deleted file mode 100644 index 77b0bdc411..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/class-use/TieringSourceEnumerator.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator

    -
    -
    No usage of org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-summary.html deleted file mode 100644 index c79bfe407c..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-summary.html +++ /dev/null @@ -1,169 +0,0 @@ - - - - - -org.apache.fluss.flink.tiering.source.enumerator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.tiering.source.enumerator

    -
    -
    -
      -
    • - - - - - - - - - - - - -
      Class Summary 
      ClassDescription
      TieringSourceEnumerator -
      An implementation of SplitEnumerator used to request TieringSplit from Fluss - Cluster.
      -
      -
    • -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-tree.html deleted file mode 100644 index 295c459f12..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-tree.html +++ /dev/null @@ -1,163 +0,0 @@ - - - - - -org.apache.fluss.flink.tiering.source.enumerator Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.tiering.source.enumerator

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    -
      -
    • java.lang.Object -
        -
      • org.apache.fluss.flink.tiering.source.enumerator.TieringSourceEnumerator (implements org.apache.flink.api.connector.source.SplitEnumerator<SplitT,​CheckpointT>)
      • -
      -
    • -
    -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-use.html deleted file mode 100644 index 107d0e6ccd..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/enumerator/package-use.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.tiering.source.enumerator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.tiering.source.enumerator

    -
    -
    No usage of org.apache.fluss.flink.tiering.source.enumerator
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-summary.html deleted file mode 100644 index 67afe12196..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-summary.html +++ /dev/null @@ -1,224 +0,0 @@ - - - - - -org.apache.fluss.flink.tiering.source (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.tiering.source

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-tree.html deleted file mode 100644 index a21b107201..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-tree.html +++ /dev/null @@ -1,184 +0,0 @@ - - - - - -org.apache.fluss.flink.tiering.source Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.tiering.source

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    -
      -
    • java.lang.Object -
        -
      • org.apache.flink.connector.base.source.reader.SourceReaderBase<E,​T,​SplitT,​SplitStateT> (implements org.apache.flink.api.connector.source.SourceReader<T,​SplitT>) -
          -
        • org.apache.flink.connector.base.source.reader.SingleThreadMultiplexSourceReaderBase<E,​T,​SplitT,​SplitStateT> - -
        • -
        -
      • -
      • org.apache.fluss.flink.tiering.source.TableBucketWriteResult<WriteResult> (implements java.io.Serializable)
      • -
      • org.apache.fluss.flink.tiering.source.TableBucketWriteResultEmitter<WriteResult> (implements org.apache.flink.connector.base.source.reader.RecordEmitter<E,​T,​SplitStateT>)
      • -
      • org.apache.fluss.flink.tiering.source.TableBucketWriteResultSerializer<WriteResult> (implements org.apache.flink.core.io.SimpleVersionedSerializer<E>)
      • -
      • org.apache.fluss.flink.tiering.source.TieringSource<WriteResult> (implements org.apache.flink.api.connector.source.Source<T,​SplitT,​EnumChkT>)
      • -
      • org.apache.fluss.flink.tiering.source.TieringSource.Builder<WriteResult>
      • -
      • org.apache.fluss.flink.tiering.source.TieringSourceOptions
      • -
      • org.apache.fluss.flink.tiering.source.TieringSplitReader<WriteResult> (implements org.apache.flink.connector.base.source.reader.splitreader.SplitReader<E,​SplitT>)
      • -
      • org.apache.fluss.flink.tiering.source.TieringWriterInitContext (implements org.apache.fluss.lake.writer.WriterInitContext)
      • -
      • org.apache.flink.api.common.typeinfo.TypeInformation<T> (implements java.io.Serializable) - -
      • -
      -
    • -
    -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-use.html deleted file mode 100644 index f786fcfc8b..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/package-use.html +++ /dev/null @@ -1,224 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.tiering.source (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.tiering.source

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringLogSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringLogSplit.html deleted file mode 100644 index c0d364ee74..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringLogSplit.html +++ /dev/null @@ -1,473 +0,0 @@ - - - - - -TieringLogSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TieringLogSplit

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.api.connector.source.SourceSplit
      -
      -
      -
      public class TieringLogSplit
      -extends TieringSplit
      -
      The table split for tiering service. It's used to describe the log data of a table bucket.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          TieringLogSplit

          -
          public TieringLogSplit​(org.apache.fluss.metadata.TablePath tablePath,
          -                       org.apache.fluss.metadata.TableBucket tableBucket,
          -                       @Nullable
          -                       String partitionName,
          -                       long startingOffset,
          -                       long stoppingOffset)
          -
        • -
        - - - -
          -
        • -

          TieringLogSplit

          -
          public TieringLogSplit​(org.apache.fluss.metadata.TablePath tablePath,
          -                       org.apache.fluss.metadata.TableBucket tableBucket,
          -                       @Nullable
          -                       String partitionName,
          -                       long startingOffset,
          -                       long stoppingOffset,
          -                       int numberOfSplits)
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSnapshotSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSnapshotSplit.html deleted file mode 100644 index f3a0c3ab4e..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSnapshotSplit.html +++ /dev/null @@ -1,474 +0,0 @@ - - - - - -TieringSnapshotSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TieringSnapshotSplit

    -
    -
    - -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.api.connector.source.SourceSplit
      -
      -
      -
      public class TieringSnapshotSplit
      -extends TieringSplit
      -
      The table split for tiering service. It's used to describe the snapshot data of a primary key - table bucket.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          TieringSnapshotSplit

          -
          public TieringSnapshotSplit​(org.apache.fluss.metadata.TablePath tablePath,
          -                            org.apache.fluss.metadata.TableBucket tableBucket,
          -                            @Nullable
          -                            String partitionName,
          -                            long snapshotId,
          -                            long logOffsetOfSnapshot)
          -
        • -
        - - - -
          -
        • -

          TieringSnapshotSplit

          -
          public TieringSnapshotSplit​(org.apache.fluss.metadata.TablePath tablePath,
          -                            org.apache.fluss.metadata.TableBucket tableBucket,
          -                            @Nullable
          -                            String partitionName,
          -                            long snapshotId,
          -                            long logOffsetOfSnapshot,
          -                            int numberOfSplits)
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplit.html deleted file mode 100644 index 7a1de92bb8..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplit.html +++ /dev/null @@ -1,660 +0,0 @@ - - - - - -TieringSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TieringSplit

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.tiering.source.split.TieringSplit
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.api.connector.source.SourceSplit
      -
      -
      -
      Direct Known Subclasses:
      -
      TieringLogSplit, TieringSnapshotSplit
      -
      -
      -
      public abstract class TieringSplit
      -extends Object
      -implements org.apache.flink.api.connector.source.SourceSplit
      -
      The base table split for tiering service.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - -
          -
        • -

          TIERING_SNAPSHOT_SPLIT_FLAG

          -
          public static final byte TIERING_SNAPSHOT_SPLIT_FLAG
          -
          -
          See Also:
          -
          Constant Field Values
          -
          -
        • -
        - - - -
          -
        • -

          TIERING_LOG_SPLIT_FLAG

          -
          public static final byte TIERING_LOG_SPLIT_FLAG
          -
          -
          See Also:
          -
          Constant Field Values
          -
          -
        • -
        - - - -
          -
        • -

          UNKNOWN_NUMBER_OF_SPLITS

          -
          protected static final int UNKNOWN_NUMBER_OF_SPLITS
          -
          -
          See Also:
          -
          Constant Field Values
          -
          -
        • -
        - - - -
          -
        • -

          tablePath

          -
          protected final org.apache.fluss.metadata.TablePath tablePath
          -
        • -
        - - - -
          -
        • -

          tableBucket

          -
          protected final org.apache.fluss.metadata.TableBucket tableBucket
          -
        • -
        - - - -
          -
        • -

          partitionName

          -
          @Nullable
          -protected final String partitionName
          -
        • -
        - - - -
          -
        • -

          numberOfSplits

          -
          protected final int numberOfSplits
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          TieringSplit

          -
          public TieringSplit​(org.apache.fluss.metadata.TablePath tablePath,
          -                    org.apache.fluss.metadata.TableBucket tableBucket,
          -                    @Nullable
          -                    String partitionName,
          -                    int numberOfSplits)
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          isTieringSnapshotSplit

          -
          public final boolean isTieringSnapshotSplit()
          -
          Checks whether this split is a primary key table split to tier.
          -
        • -
        - - - - - - - -
          -
        • -

          isTieringLogSplit

          -
          public final boolean isTieringLogSplit()
          -
          Checks whether this split is a log split to tier.
          -
        • -
        - - - - - - - -
          -
        • -

          splitKind

          -
          protected byte splitKind()
          -
        • -
        - - - -
          -
        • -

          getNumberOfSplits

          -
          public int getNumberOfSplits()
          -
        • -
        - - - -
          -
        • -

          toSplitId

          -
          protected static String toSplitId​(String splitPrefix,
          -                                  org.apache.fluss.metadata.TableBucket tableBucket)
          -
        • -
        - - - -
          -
        • -

          getTablePath

          -
          public org.apache.fluss.metadata.TablePath getTablePath()
          -
        • -
        - - - -
          -
        • -

          getTableBucket

          -
          public org.apache.fluss.metadata.TableBucket getTableBucket()
          -
        • -
        - - - -
          -
        • -

          getPartitionName

          -
          @Nullable
          -public String getPartitionName()
          -
        • -
        - - - -
          -
        • -

          copy

          -
          public abstract TieringSplit copy​(int numberOfSplits)
          -
        • -
        - - - -
          -
        • -

          equals

          -
          public boolean equals​(Object object)
          -
          -
          Overrides:
          -
          equals in class Object
          -
          -
        • -
        - - - -
          -
        • -

          hashCode

          -
          public int hashCode()
          -
          -
          Overrides:
          -
          hashCode in class Object
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplitGenerator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplitGenerator.html deleted file mode 100644 index 56c432b7e9..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplitGenerator.html +++ /dev/null @@ -1,312 +0,0 @@ - - - - - -TieringSplitGenerator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TieringSplitGenerator

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.tiering.source.split.TieringSplitGenerator
      • -
      -
    • -
    -
    -
      -
    • -
      -
      public class TieringSplitGenerator
      -extends Object
      -
      A generator for lake splits.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          TieringSplitGenerator

          -
          public TieringSplitGenerator​(Admin flussAdmin)
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplitSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplitSerializer.html deleted file mode 100644 index 5da2d5a8a7..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/TieringSplitSerializer.html +++ /dev/null @@ -1,404 +0,0 @@ - - - - - -TieringSplitSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TieringSplitSerializer

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.tiering.source.split.TieringSplitSerializer
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      org.apache.flink.core.io.SimpleVersionedSerializer<TieringSplit>, org.apache.flink.core.io.Versioned
      -
      -
      -
      public class TieringSplitSerializer
      -extends Object
      -implements org.apache.flink.core.io.SimpleVersionedSerializer<TieringSplit>
      -
      A serializer for the TieringSplit.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      - -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          TieringSplitSerializer

          -
          public TieringSplitSerializer()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getVersion

          -
          public int getVersion()
          -
          -
          Specified by:
          -
          getVersion in interface org.apache.flink.core.io.SimpleVersionedSerializer<TieringSplit>
          -
          Specified by:
          -
          getVersion in interface org.apache.flink.core.io.Versioned
          -
          -
        • -
        - - - - - - - -
          -
        • -

          deserialize

          -
          public TieringSplit deserialize​(int version,
          -                                byte[] serialized)
          -                         throws IOException
          -
          -
          Specified by:
          -
          deserialize in interface org.apache.flink.core.io.SimpleVersionedSerializer<TieringSplit>
          -
          Throws:
          -
          IOException
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringLogSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringLogSplit.html deleted file mode 100644 index 2d747eed87..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringLogSplit.html +++ /dev/null @@ -1,199 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.source.split.TieringLogSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.source.split.TieringLogSplit

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSnapshotSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSnapshotSplit.html deleted file mode 100644 index 0f15d3f778..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSnapshotSplit.html +++ /dev/null @@ -1,199 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.source.split.TieringSnapshotSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.source.split.TieringSnapshotSplit

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplit.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplit.html deleted file mode 100644 index 2e7ff6ba79..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplit.html +++ /dev/null @@ -1,452 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.source.split.TieringSplit (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.source.split.TieringSplit

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplitGenerator.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplitGenerator.html deleted file mode 100644 index e761d1dfd3..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplitGenerator.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.source.split.TieringSplitGenerator (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.source.split.TieringSplitGenerator

    -
    -
    No usage of org.apache.fluss.flink.tiering.source.split.TieringSplitGenerator
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplitSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplitSerializer.html deleted file mode 100644 index 6dbd1382dd..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/class-use/TieringSplitSerializer.html +++ /dev/null @@ -1,192 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.source.split.TieringSplitSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.source.split.TieringSplitSerializer

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-summary.html deleted file mode 100644 index 6aa578aedf..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-summary.html +++ /dev/null @@ -1,192 +0,0 @@ - - - - - -org.apache.fluss.flink.tiering.source.split (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.tiering.source.split

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-tree.html deleted file mode 100644 index 6234a6edc8..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-tree.html +++ /dev/null @@ -1,170 +0,0 @@ - - - - - -org.apache.fluss.flink.tiering.source.split Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.tiering.source.split

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    -
      -
    • java.lang.Object -
        -
      • org.apache.fluss.flink.tiering.source.split.TieringSplit (implements org.apache.flink.api.connector.source.SourceSplit) - -
      • -
      • org.apache.fluss.flink.tiering.source.split.TieringSplitGenerator
      • -
      • org.apache.fluss.flink.tiering.source.split.TieringSplitSerializer (implements org.apache.flink.core.io.SimpleVersionedSerializer<E>)
      • -
      -
    • -
    -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-use.html deleted file mode 100644 index f3414c3da6..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/split/package-use.html +++ /dev/null @@ -1,272 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.tiering.source.split (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.tiering.source.split

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSourceEnumeratorState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSourceEnumeratorState.html deleted file mode 100644 index 1b89a480c5..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSourceEnumeratorState.html +++ /dev/null @@ -1,329 +0,0 @@ - - - - - -TieringSourceEnumeratorState (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TieringSourceEnumeratorState

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorState
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          TieringSourceEnumeratorState

          -
          public TieringSourceEnumeratorState()
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSourceEnumeratorStateSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSourceEnumeratorStateSerializer.html deleted file mode 100644 index d5251a8376..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSourceEnumeratorStateSerializer.html +++ /dev/null @@ -1,404 +0,0 @@ - - - - - -TieringSourceEnumeratorStateSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TieringSourceEnumeratorStateSerializer

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorStateSerializer
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    - -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSplitState.html deleted file mode 100644 index 84f796833b..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/TieringSplitState.html +++ /dev/null @@ -1,353 +0,0 @@ - - - - - -TieringSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class TieringSplitState

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.tiering.source.state.TieringSplitState
      • -
      -
    • -
    -
    -
      -
    • -
      -
      public class TieringSplitState
      -extends Object
      -
      The state of a TieringSplit. - -

      Note: The tiering service adopts a stateless design and does not store any progress - information in state during checkpoints. All splits are re-requested from the Fluss cluster in - case of failover.

      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      - -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          TieringSplitState

          -
          public TieringSplitState​(TieringSplit tieringSplit)
          -
        • -
        -
      • -
      -
      - -
      - -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSourceEnumeratorState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSourceEnumeratorState.html deleted file mode 100644 index 090c958c1a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSourceEnumeratorState.html +++ /dev/null @@ -1,287 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorState (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorState

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSourceEnumeratorStateSerializer.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSourceEnumeratorStateSerializer.html deleted file mode 100644 index 1d1957f129..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSourceEnumeratorStateSerializer.html +++ /dev/null @@ -1,192 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorStateSerializer (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.source.state.TieringSourceEnumeratorStateSerializer

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSplitState.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSplitState.html deleted file mode 100644 index d388abca60..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/class-use/TieringSplitState.html +++ /dev/null @@ -1,230 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.tiering.source.state.TieringSplitState (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.tiering.source.state.TieringSplitState

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-summary.html deleted file mode 100644 index bdfe225f86..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-summary.html +++ /dev/null @@ -1,180 +0,0 @@ - - - - - -org.apache.fluss.flink.tiering.source.state (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.tiering.source.state

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-tree.html deleted file mode 100644 index 180acf46ee..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-tree.html +++ /dev/null @@ -1,165 +0,0 @@ - - - - - -org.apache.fluss.flink.tiering.source.state Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.tiering.source.state

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-use.html deleted file mode 100644 index b9fefc7117..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/tiering/source/state/package-use.html +++ /dev/null @@ -1,243 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.tiering.source.state (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.tiering.source.state

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/CatalogExceptionUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/CatalogExceptionUtils.html deleted file mode 100644 index e7beacfa47..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/CatalogExceptionUtils.html +++ /dev/null @@ -1,407 +0,0 @@ - - - - - -CatalogExceptionUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class CatalogExceptionUtils

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.utils.CatalogExceptionUtils
      • -
      -
    • -
    -
    -
      -
    • -
      -
      public class CatalogExceptionUtils
      -extends Object
      -
      Utility class for catalog exceptions.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          isDatabaseNotExist

          -
          public static boolean isDatabaseNotExist​(Throwable throwable)
          -
        • -
        - - - -
          -
        • -

          isDatabaseNotEmpty

          -
          public static boolean isDatabaseNotEmpty​(Throwable throwable)
          -
        • -
        - - - -
          -
        • -

          isDatabaseAlreadyExist

          -
          public static boolean isDatabaseAlreadyExist​(Throwable throwable)
          -
        • -
        - - - -
          -
        • -

          isTableNotExist

          -
          public static boolean isTableNotExist​(Throwable throwable)
          -
        • -
        - - - -
          -
        • -

          isTableAlreadyExist

          -
          public static boolean isTableAlreadyExist​(Throwable throwable)
          -
        • -
        - - - -
          -
        • -

          isLakeTableAlreadyExist

          -
          public static boolean isLakeTableAlreadyExist​(Throwable throwable)
          -
        • -
        - - - -
          -
        • -

          isTableInvalid

          -
          public static boolean isTableInvalid​(Throwable throwable)
          -
        • -
        - - - -
          -
        • -

          isTableNotPartitioned

          -
          public static boolean isTableNotPartitioned​(Throwable throwable)
          -
        • -
        - - - -
          -
        • -

          isPartitionAlreadyExists

          -
          public static boolean isPartitionAlreadyExists​(Throwable throwable)
          -
        • -
        - - - -
          -
        • -

          isPartitionNotExist

          -
          public static boolean isPartitionNotExist​(Throwable throwable)
          -
        • -
        - - - -
          -
        • -

          isPartitionInvalid

          -
          public static boolean isPartitionInvalid​(Throwable throwable)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/CatalogPropertiesUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/CatalogPropertiesUtils.html deleted file mode 100644 index 3a1c3b6ce5..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/CatalogPropertiesUtils.html +++ /dev/null @@ -1,379 +0,0 @@ - - - - - -CatalogPropertiesUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class CatalogPropertiesUtils

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.utils.CatalogPropertiesUtils
      • -
      -
    • -
    -
    -
      -
    • -
      -
      public class CatalogPropertiesUtils
      -extends Object
      -
      Utilities for de/serializing Catalog objects into a map of string properties. - -

      Copied from Flink.

      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - - - - - -
          -
        • -

          deserializeWatermark

          -
          public static void deserializeWatermark​(Map<String,​String> map,
          -                                        org.apache.flink.table.api.Schema.Builder builder)
          -
        • -
        - - - -
          -
        • -

          deserializeComputedColumn

          -
          public static void deserializeComputedColumn​(Map<String,​String> map,
          -                                             int columnIndex,
          -                                             org.apache.flink.table.api.Schema.Builder builder)
          -
        • -
        - - - -
          -
        • -

          serializeComputedColumns

          -
          public static void serializeComputedColumns​(Map<String,​String> map,
          -                                            List<org.apache.flink.table.catalog.Column> columns)
          -
        • -
        - - - -
          -
        • -

          nonPhysicalColumnsCount

          -
          public static int nonPhysicalColumnsCount​(Map<String,​String> tableOptions,
          -                                          List<String> physicalColumns)
          -
        • -
        - - - -
          -
        • -

          serializeComputedColumns

          -
          public static String[] serializeComputedColumns​(List<org.apache.flink.table.catalog.Column> columns)
          -
        • -
        - - - -
          -
        • -

          serializeWatermarkSpecs

          -
          public static void serializeWatermarkSpecs​(Map<String,​String> map,
          -                                           List<org.apache.flink.table.catalog.WatermarkSpec> specs)
          -
        • -
        - - - -
          -
        • -

          columnKey

          -
          public static String columnKey​(int index)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/DataLakeUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/DataLakeUtils.html deleted file mode 100644 index 55a6d08dd7..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/DataLakeUtils.html +++ /dev/null @@ -1,321 +0,0 @@ - - - - - -DataLakeUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class DataLakeUtils

    -
    -
    - -
    -
      -
    • -
      -
      public class DataLakeUtils
      -extends Object
      -
      Utility class for accessing data lake related configurations.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          DataLakeUtils

          -
          public DataLakeUtils()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          extractLakeCatalogProperties

          -
          public static Map<String,​String> extractLakeCatalogProperties​(org.apache.fluss.config.Configuration tableOptions)
          -
        • -
        - - - -
          -
        • -

          getDatalakeFormat

          -
          public static Optional<org.apache.fluss.metadata.DataLakeFormat> getDatalakeFormat​(org.apache.flink.configuration.ReadableConfig tableOptions)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConnectorOptionsUtils.StartupOptions.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConnectorOptionsUtils.StartupOptions.html deleted file mode 100644 index 907e5a103f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConnectorOptionsUtils.StartupOptions.html +++ /dev/null @@ -1,329 +0,0 @@ - - - - - -FlinkConnectorOptionsUtils.StartupOptions (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkConnectorOptionsUtils.StartupOptions

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils.StartupOptions
      • -
      -
    • -
    -
    -
      -
    • -
      -
      Enclosing class:
      -
      FlinkConnectorOptionsUtils
      -
      -
      -
      public static class FlinkConnectorOptionsUtils.StartupOptions
      -extends Object
      -
      Fluss startup options. *
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      - -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          StartupOptions

          -
          public StartupOptions()
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConnectorOptionsUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConnectorOptionsUtils.html deleted file mode 100644 index 013bb8b14a..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConnectorOptionsUtils.html +++ /dev/null @@ -1,424 +0,0 @@ - - - - - -FlinkConnectorOptionsUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkConnectorOptionsUtils

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlinkConnectorOptionsUtils

          -
          public FlinkConnectorOptionsUtils()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          getLocalTimeZone

          -
          public static ZoneId getLocalTimeZone​(String timeZone)
          -
        • -
        - - - -
          -
        • -

          validateTableSourceOptions

          -
          public static void validateTableSourceOptions​(org.apache.flink.configuration.ReadableConfig tableOptions)
          -
        • -
        - - - - - - - -
          -
        • -

          getBucketKeys

          -
          public static List<String> getBucketKeys​(org.apache.flink.configuration.ReadableConfig tableOptions)
          -
        • -
        - - - -
          -
        • -

          getBucketKeyIndexes

          -
          public static int[] getBucketKeyIndexes​(org.apache.flink.configuration.ReadableConfig tableOptions,
          -                                        org.apache.flink.table.types.logical.RowType schema)
          -
        • -
        - - - -
          -
        • -

          parseTimestamp

          -
          public static long parseTimestamp​(String timestampStr,
          -                                  String optionKey,
          -                                  ZoneId timeZone)
          -
          Parses timestamp String to Long. - -

          timestamp String format was given as following: - -

          -     scan.startup.timestamp = 1678883047356
          -     scan.startup.timestamp = 2023-12-09 23:09:12
          - 
          -
          -
          Returns:
          -
          timestamp as long value
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConversions.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConversions.html deleted file mode 100644 index 4113968e04..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkConversions.html +++ /dev/null @@ -1,422 +0,0 @@ - - - - - -FlinkConversions (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkConversions

    -
    -
    - -
    -
      -
    • -
      -
      public class FlinkConversions
      -extends Object
      -
      Utils for conversion between Flink and Fluss.
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Summary

        - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
        All Methods Static Methods Concrete Methods 
        Modifier and TypeMethodDescription
        static <T> org.apache.flink.configuration.ConfigOption<T>toFlinkOption​(org.apache.fluss.config.ConfigOption<T> flussOption) -
        Convert Fluss's ConfigOption to Flink's ConfigOption.
        -
        static List<org.apache.flink.configuration.ConfigOption<?>>toFlinkOptions​(Collection<org.apache.fluss.config.ConfigOption<?>> flussOption) -
        Convert Fluss's ConfigOptions to Flink's ConfigOptions.
        -
        static org.apache.flink.types.RowKindtoFlinkRowKind​(org.apache.fluss.record.ChangeType changeType) 
        static org.apache.flink.table.types.logical.RowTypetoFlinkRowType​(org.apache.fluss.types.RowType flussRowType) -
        Convert Fluss's RowType to Flink's RowType.
        -
        static org.apache.flink.table.catalog.CatalogTabletoFlinkTable​(org.apache.fluss.metadata.TableInfo tableInfo) -
        Convert Fluss's table to Flink's table.
        -
        static org.apache.flink.table.types.DataTypetoFlinkType​(org.apache.fluss.types.DataType flussDataType) -
        Convert Fluss's type to Flink's type.
        -
        static org.apache.fluss.metadata.DatabaseDescriptortoFlussDatabase​(org.apache.flink.table.catalog.CatalogDatabase catalogDatabase) -
        Convert Flink's table to Fluss's database.
        -
        static org.apache.fluss.types.RowTypetoFlussRowType​(org.apache.flink.table.types.logical.RowType flinkRowType) -
        Convert Flink's RowType to Fluss' RowType.
        -
        static org.apache.fluss.metadata.TableDescriptortoFlussTable​(org.apache.flink.table.catalog.ResolvedCatalogTable catalogTable) -
        Convert Flink's table to Fluss's table.
        -
        static org.apache.fluss.types.DataTypetoFlussType​(org.apache.flink.table.types.DataType flinkDataType) -
        Convert Flink's physical type to Fluss' type.
        -
        - -
      • -
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          toFlinkType

          -
          @VisibleForTesting
          -public static org.apache.flink.table.types.DataType toFlinkType​(org.apache.fluss.types.DataType flussDataType)
          -
          Convert Fluss's type to Flink's type.
          -
        • -
        - - - -
          -
        • -

          toFlinkRowType

          -
          public static org.apache.flink.table.types.logical.RowType toFlinkRowType​(org.apache.fluss.types.RowType flussRowType)
          -
          Convert Fluss's RowType to Flink's RowType.
          -
        • -
        - - - -
          -
        • -

          toFlussType

          -
          @VisibleForTesting
          -public static org.apache.fluss.types.DataType toFlussType​(org.apache.flink.table.types.DataType flinkDataType)
          -
          Convert Flink's physical type to Fluss' type.
          -
        • -
        - - - -
          -
        • -

          toFlussRowType

          -
          public static org.apache.fluss.types.RowType toFlussRowType​(org.apache.flink.table.types.logical.RowType flinkRowType)
          -
          Convert Flink's RowType to Fluss' RowType.
          -
        • -
        - - - -
          -
        • -

          toFlinkTable

          -
          public static org.apache.flink.table.catalog.CatalogTable toFlinkTable​(org.apache.fluss.metadata.TableInfo tableInfo)
          -
          Convert Fluss's table to Flink's table.
          -
        • -
        - - - -
          -
        • -

          toFlussTable

          -
          public static org.apache.fluss.metadata.TableDescriptor toFlussTable​(org.apache.flink.table.catalog.ResolvedCatalogTable catalogTable)
          -
          Convert Flink's table to Fluss's table.
          -
        • -
        - - - -
          -
        • -

          toFlussDatabase

          -
          public static org.apache.fluss.metadata.DatabaseDescriptor toFlussDatabase​(org.apache.flink.table.catalog.CatalogDatabase catalogDatabase)
          -
          Convert Flink's table to Fluss's database.
          -
        • -
        - - - -
          -
        • -

          toFlinkOptions

          -
          public static List<org.apache.flink.configuration.ConfigOption<?>> toFlinkOptions​(Collection<org.apache.fluss.config.ConfigOption<?>> flussOption)
          -
          Convert Fluss's ConfigOptions to Flink's ConfigOptions.
          -
        • -
        - - - -
          -
        • -

          toFlinkOption

          -
          public static <T> org.apache.flink.configuration.ConfigOption<T> toFlinkOption​(org.apache.fluss.config.ConfigOption<T> flussOption)
          -
          Convert Fluss's ConfigOption to Flink's ConfigOption.
          -
        • -
        - - - -
          -
        • -

          toFlinkRowKind

          -
          public static org.apache.flink.types.RowKind toFlinkRowKind​(org.apache.fluss.record.ChangeType changeType)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkRowToFlussRowConverter.FlussSerializationConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkRowToFlussRowConverter.FlussSerializationConverter.html deleted file mode 100644 index b9e7393b33..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkRowToFlussRowConverter.FlussSerializationConverter.html +++ /dev/null @@ -1,275 +0,0 @@ - - - - - -FlinkRowToFlussRowConverter.FlussSerializationConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface FlinkRowToFlussRowConverter.FlussSerializationConverter

    -
    -
    -
    -
      -
    • -
      -
      All Superinterfaces:
      -
      Serializable
      -
      -
      -
      Enclosing class:
      -
      FlinkRowToFlussRowConverter
      -
      -
      -
      Functional Interface:
      -
      This is a functional interface and can therefore be used as the assignment target for a lambda expression or method reference.
      -
      -
      -
      @FunctionalInterface
      -public static interface FlinkRowToFlussRowConverter.FlussSerializationConverter
      -extends Serializable
      -
      Runtime converter to convert field in Flink's RowData to Fluss's InternalRow - type object.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          serialize

          -
          Object serialize​(Object flinkField)
          -
          Convert a Flink field object of RowData to the Fluss's internal data structure - object.
          -
          -
          Parameters:
          -
          flinkField - A single field of a RowData
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkRowToFlussRowConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkRowToFlussRowConverter.html deleted file mode 100644 index 156ba82ca7..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkRowToFlussRowConverter.html +++ /dev/null @@ -1,371 +0,0 @@ - - - - - -FlinkRowToFlussRowConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkRowToFlussRowConverter

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.utils.FlinkRowToFlussRowConverter
      • -
      -
    • -
    -
    -
      -
    • -
      -
      All Implemented Interfaces:
      -
      AutoCloseable
      -
      -
      -
      @Deprecated
      -public class FlinkRowToFlussRowConverter
      -extends Object
      -implements AutoCloseable
      -
      Deprecated. -
      This class is deprecated and will be removed in future versions. It is recommended to - use the FlinkAsFlussRow to wraps Flink's RowData as Fluss's InternalRow.
      -
      -
      A converter to convert Flink's RowData to Fluss's InternalRow.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      - -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkUtils.html deleted file mode 100644 index a7bbc1a7d3..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlinkUtils.html +++ /dev/null @@ -1,317 +0,0 @@ - - - - - -FlinkUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlinkUtils

    -
    -
    - -
    -
      -
    • -
      -
      public class FlinkUtils
      -extends Object
      -
      Utils for Flink classes.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FlinkUtils

          -
          public FlinkUtils()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          projectRowType

          -
          public static org.apache.flink.table.types.logical.RowType projectRowType​(org.apache.flink.table.types.logical.RowType rowType,
          -                                                                          int[] projection)
          -
          Returns projected RowType by given projection indexes over original RowType.
          -
          -
          Parameters:
          -
          rowType - the original row type
          -
          projection - the projection indexes
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToFlinkRowConverter.FlussDeserializationConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToFlinkRowConverter.FlussDeserializationConverter.html deleted file mode 100644 index 32b142dd94..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToFlinkRowConverter.FlussDeserializationConverter.html +++ /dev/null @@ -1,275 +0,0 @@ - - - - - -FlussRowToFlinkRowConverter.FlussDeserializationConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface FlussRowToFlinkRowConverter.FlussDeserializationConverter

    -
    -
    -
    -
      -
    • -
      -
      All Superinterfaces:
      -
      Serializable
      -
      -
      -
      Enclosing class:
      -
      FlussRowToFlinkRowConverter
      -
      -
      -
      Functional Interface:
      -
      This is a functional interface and can therefore be used as the assignment target for a lambda expression or method reference.
      -
      -
      -
      @FunctionalInterface
      -public static interface FlussRowToFlinkRowConverter.FlussDeserializationConverter
      -extends Serializable
      -
      Runtime converter to convert field in Fluss's InternalRow to Flink's RowData - type object.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          deserialize

          -
          Object deserialize​(Object flussField)
          -
          Convert a Fluss field object of InternalRow to the Flink's internal data - structure object.
          -
          -
          Parameters:
          -
          flussField - A single field of a InternalRow
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToFlinkRowConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToFlinkRowConverter.html deleted file mode 100644 index d115d3317e..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToFlinkRowConverter.html +++ /dev/null @@ -1,381 +0,0 @@ - - - - - -FlussRowToFlinkRowConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlussRowToFlinkRowConverter

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter
      • -
      -
    • -
    -
    -
      -
    • -
      -
      public class FlussRowToFlinkRowConverter
      -extends Object
      -
      A converter to convert Fluss's InternalRow to Flink's RowData. - -

      Note: fluss-datalake-tiering also contains the same class, we need to keep them in sync if we - modify this class.

      -
    • -
    -
    -
    - -
    -
    - -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToJsonConverters.FlussRowToJsonConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToJsonConverters.FlussRowToJsonConverter.html deleted file mode 100644 index cfc8c61e0c..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToJsonConverters.FlussRowToJsonConverter.html +++ /dev/null @@ -1,269 +0,0 @@ - - - - - -FlussRowToJsonConverters.FlussRowToJsonConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Interface FlussRowToJsonConverters.FlussRowToJsonConverter

    -
    -
    -
    -
      -
    • -
      -
      All Superinterfaces:
      -
      Serializable
      -
      -
      -
      Enclosing class:
      -
      FlussRowToJsonConverters
      -
      -
      -
      Functional Interface:
      -
      This is a functional interface and can therefore be used as the assignment target for a lambda expression or method reference.
      -
      -
      -
      @FunctionalInterface
      -public static interface FlussRowToJsonConverters.FlussRowToJsonConverter
      -extends Serializable
      -
      Runtime converter that converts objects of Fluss data structures to corresponding JsonNodes.
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Summary

        - - - - - - - - - - - - -
        All Methods Instance Methods Abstract Methods 
        Modifier and TypeMethodDescription
        org.apache.fluss.shaded.jackson2.com.fasterxml.jackson.databind.JsonNodeconvert​(org.apache.fluss.shaded.jackson2.com.fasterxml.jackson.databind.ObjectMapper mapper, - org.apache.fluss.shaded.jackson2.com.fasterxml.jackson.databind.JsonNode reuse, - Object value) 
        -
      • -
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          convert

          -
          org.apache.fluss.shaded.jackson2.com.fasterxml.jackson.databind.JsonNode convert​(org.apache.fluss.shaded.jackson2.com.fasterxml.jackson.databind.ObjectMapper mapper,
          -                                                                                 org.apache.fluss.shaded.jackson2.com.fasterxml.jackson.databind.JsonNode reuse,
          -                                                                                 Object value)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToJsonConverters.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToJsonConverters.html deleted file mode 100644 index 9a46068d72..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/FlussRowToJsonConverters.html +++ /dev/null @@ -1,332 +0,0 @@ - - - - - -FlussRowToJsonConverters (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class FlussRowToJsonConverters

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.utils.FlussRowToJsonConverters
      • -
      -
    • -
    -
    -
      -
    • -
      -
      public class FlussRowToJsonConverters
      -extends Object
      -
      A converter to convert Fluss's InternalRow to JsonNode.
      -
    • -
    -
    -
    - -
    -
    - -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/LakeSourceUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/LakeSourceUtils.html deleted file mode 100644 index b4b71f1834..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/LakeSourceUtils.html +++ /dev/null @@ -1,356 +0,0 @@ - - - - - -LakeSourceUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class LakeSourceUtils

    -
    -
    - -
    -
      -
    • -
      -
      public class LakeSourceUtils
      -extends Object
      -
      Utils for create lake source.
      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - -
          -
        • -

          LOG

          -
          public static final org.slf4j.Logger LOG
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          LakeSourceUtils

          -
          public LakeSourceUtils()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          createLakeSource

          -
          @Nullable
          -public static org.apache.fluss.lake.source.LakeSource<org.apache.fluss.lake.source.LakeSplit> createLakeSource​(org.apache.fluss.metadata.TablePath tablePath,
          -                                                                                                               Map<String,​String> properties)
          -
          Return the lake source of the given table. Return null when the lake storage doesn't support - create lake source.
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PojoToRowConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PojoToRowConverter.html deleted file mode 100644 index 70060c64d3..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PojoToRowConverter.html +++ /dev/null @@ -1,349 +0,0 @@ - - - - - -PojoToRowConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class PojoToRowConverter<T>

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.utils.PojoToRowConverter<T>
      • -
      -
    • -
    -
    -
      -
    • -
      -
      Type Parameters:
      -
      T - The POJO type to convert
      -
      -
      -
      public class PojoToRowConverter<T>
      -extends Object
      -
      Utility class for converting Java POJOs to Fluss's InternalRow format. - -

      This utility uses Flink's POJO type information to map fields from POJOs to InternalRow based - on a given schema. - -

      Example usage: - -

      
      - // Create a converter
      - PojoToRowConverter<Order> converter =
      -     new PojoToRowConverter<>(Order.class, rowType);
      -
      - // Convert a POJO to GenericRow
      - Order order = new Order(1001L, 5001L, 10, "123 Mumbai");
      - GenericRow row = converter.convert(order);
      - 
      - -

      Note: Nested POJO fields are not supported in the current implementation.

      -
    • -
    -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          PojoToRowConverter

          -
          public PojoToRowConverter​(Class<T> pojoClass,
          -                          org.apache.fluss.types.RowType rowType)
          -
          Creates a new converter for the specified POJO class and row type.
          -
          -
          Parameters:
          -
          pojoClass - The class of POJOs to convert
          -
          rowType - The row schema to use for conversion
          -
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - - - -
          -
        • -

          convert

          -
          public org.apache.fluss.row.GenericRow convert​(T pojo)
          -
          Converts a POJO to a GenericRow object according to the schema.
          -
          -
          Parameters:
          -
          pojo - The POJO to convert
          -
          Returns:
          -
          The converted GenericRow, or null if the input is null
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.FieldEqual.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.FieldEqual.html deleted file mode 100644 index 2ffa75b64b..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.FieldEqual.html +++ /dev/null @@ -1,340 +0,0 @@ - - - - - -PushdownUtils.FieldEqual (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class PushdownUtils.FieldEqual

    -
    -
    -
      -
    • java.lang.Object
    • -
    • -
        -
      • org.apache.fluss.flink.utils.PushdownUtils.FieldEqual
      • -
      -
    • -
    -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Field Detail

        - - - -
          -
        • -

          fieldIndex

          -
          public final int fieldIndex
          -
        • -
        - - - -
          -
        • -

          equalValue

          -
          public final Object equalValue
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          FieldEqual

          -
          public FieldEqual​(int fieldIndex,
          -                  Object equalValue)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.ValueConversion.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.ValueConversion.html deleted file mode 100644 index 53f6d6aeeb..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.ValueConversion.html +++ /dev/null @@ -1,383 +0,0 @@ - - - - - -PushdownUtils.ValueConversion (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Enum PushdownUtils.ValueConversion

    -
    -
    - -
    - -
    -
    - -
    -
    -
      -
    • - -
      - -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          values

          -
          public static PushdownUtils.ValueConversion[] values()
          -
          Returns an array containing the constants of this enum type, in -the order they are declared. This method may be used to iterate -over the constants as follows: -
          -for (PushdownUtils.ValueConversion c : PushdownUtils.ValueConversion.values())
          -    System.out.println(c);
          -
          -
          -
          Returns:
          -
          an array containing the constants of this enum type, in the order they are declared
          -
          -
        • -
        - - - -
          -
        • -

          valueOf

          -
          public static PushdownUtils.ValueConversion valueOf​(String name)
          -
          Returns the enum constant of this type with the specified name. -The string must match exactly an identifier used to declare an -enum constant in this type. (Extraneous whitespace characters are -not permitted.)
          -
          -
          Parameters:
          -
          name - the name of the enum constant to be returned.
          -
          Returns:
          -
          the enum constant with the specified name
          -
          Throws:
          -
          IllegalArgumentException - if this enum type has no constant with the specified name
          -
          NullPointerException - if the argument is null
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.html deleted file mode 100644 index bb5dcd7710..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/PushdownUtils.html +++ /dev/null @@ -1,434 +0,0 @@ - - - - - -PushdownUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Class PushdownUtils

    -
    -
    - -
    -
      -
    • -
      -
      public class PushdownUtils
      -extends Object
      -
      Utilities for pushdown abilities.
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Nested Class Summary

        - - - - - - - - - - - - - - - - - -
        Nested Classes 
        Modifier and TypeClassDescription
        static class PushdownUtils.FieldEqual -
        A structure represents a source field equal literal expression.
        -
        static class PushdownUtils.ValueConversion -
        The value conversion type between Flink internal value and Fluss internal value.
        -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Constructor Summary

        - - - - - - - - - - -
        Constructors 
        ConstructorDescription
        PushdownUtils() 
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Summary

        - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
        All Methods Static Methods Concrete Methods 
        Modifier and TypeMethodDescription
        static longcountLogTable​(org.apache.fluss.metadata.TablePath tablePath, - org.apache.fluss.config.Configuration flussConfig) 
        static voiddeleteSingleRow​(org.apache.fluss.row.GenericRow deleteRow, - org.apache.fluss.metadata.TablePath tablePath, - org.apache.fluss.config.Configuration flussConfig) 
        static List<PushdownUtils.FieldEqual>extractFieldEquals​(List<org.apache.flink.table.expressions.ResolvedExpression> expressions, - Map<Integer,​org.apache.flink.table.types.logical.LogicalType> fieldIndexToType, - List<org.apache.flink.table.expressions.ResolvedExpression> acceptedFiltersResult, - List<org.apache.flink.table.expressions.ResolvedExpression> remainingFiltersResult, - PushdownUtils.ValueConversion valueConversion) -
        Extract field equality information from expressions.
        -
        static Collection<org.apache.flink.table.data.RowData>limitScan​(org.apache.fluss.metadata.TablePath tablePath, - org.apache.fluss.config.Configuration flussConfig, - org.apache.flink.table.types.logical.RowType sourceOutputType, - int[] projectedFields, - long limitRowNum) 
        static Collection<org.apache.flink.table.data.RowData>querySingleRow​(org.apache.flink.table.data.GenericRowData lookupRow, - org.apache.fluss.metadata.TablePath tablePath, - org.apache.fluss.config.Configuration flussConfig, - org.apache.flink.table.types.logical.RowType sourceOutputType, - int[] primaryKeyIndexes, - int lookupMaxRetryTimes, - int[] projectedFields) 
        - -
      • -
      -
      -
    • -
    -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Constructor Detail

        - - - -
          -
        • -

          PushdownUtils

          -
          public PushdownUtils()
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          extractFieldEquals

          -
          public static List<PushdownUtils.FieldEqual> extractFieldEquals​(List<org.apache.flink.table.expressions.ResolvedExpression> expressions,
          -                                                                Map<Integer,​org.apache.flink.table.types.logical.LogicalType> fieldIndexToType,
          -                                                                List<org.apache.flink.table.expressions.ResolvedExpression> acceptedFiltersResult,
          -                                                                List<org.apache.flink.table.expressions.ResolvedExpression> remainingFiltersResult,
          -                                                                PushdownUtils.ValueConversion valueConversion)
          -
          Extract field equality information from expressions.
          -
        • -
        - - - -
          -
        • -

          querySingleRow

          -
          public static Collection<org.apache.flink.table.data.RowData> querySingleRow​(org.apache.flink.table.data.GenericRowData lookupRow,
          -                                                                             org.apache.fluss.metadata.TablePath tablePath,
          -                                                                             org.apache.fluss.config.Configuration flussConfig,
          -                                                                             org.apache.flink.table.types.logical.RowType sourceOutputType,
          -                                                                             int[] primaryKeyIndexes,
          -                                                                             int lookupMaxRetryTimes,
          -                                                                             @Nullable
          -                                                                             int[] projectedFields)
          -
        • -
        - - - -
          -
        • -

          deleteSingleRow

          -
          public static void deleteSingleRow​(org.apache.fluss.row.GenericRow deleteRow,
          -                                   org.apache.fluss.metadata.TablePath tablePath,
          -                                   org.apache.fluss.config.Configuration flussConfig)
          -
        • -
        - - - -
          -
        • -

          limitScan

          -
          public static Collection<org.apache.flink.table.data.RowData> limitScan​(org.apache.fluss.metadata.TablePath tablePath,
          -                                                                        org.apache.fluss.config.Configuration flussConfig,
          -                                                                        org.apache.flink.table.types.logical.RowType sourceOutputType,
          -                                                                        @Nullable
          -                                                                        int[] projectedFields,
          -                                                                        long limitRowNum)
          -
        • -
        - - - -
          -
        • -

          countLogTable

          -
          public static long countLogTable​(org.apache.fluss.metadata.TablePath tablePath,
          -                                 org.apache.fluss.config.Configuration flussConfig)
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/TimestampFormat.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/TimestampFormat.html deleted file mode 100644 index 042594e097..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/TimestampFormat.html +++ /dev/null @@ -1,390 +0,0 @@ - - - - - -TimestampFormat (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    - -
    -
    - -

    Enum TimestampFormat

    -
    -
    - -
    - -
    -
    - -
    -
    -
      -
    • - -
      -
        -
      • - - -

        Enum Constant Detail

        - - - -
          -
        • -

          SQL

          -
          public static final TimestampFormat SQL
          -
          Options to specify TIMESTAMP/TIMESTAMP_WITH_LOCAL_ZONE format. It will parse TIMESTAMP in - "yyyy-MM-dd HH:mm:ss.s{precision}" format, TIMESTAMP_WITH_LOCAL_TIMEZONE in "yyyy-MM-dd - HH:mm:ss.s{precision}'Z'" and output in the same format.
          -
        • -
        - - - -
          -
        • -

          ISO_8601

          -
          public static final TimestampFormat ISO_8601
          -
          Options to specify TIMESTAMP/TIMESTAMP_WITH_LOCAL_ZONE format. It will parse TIMESTAMP in - "yyyy-MM-ddTHH:mm:ss.s{precision}" format, TIMESTAMP_WITH_LOCAL_TIMEZONE in - "yyyy-MM-ddTHH:mm:ss.s{precision}'Z'" and output in the same format.
          -
        • -
        -
      • -
      -
      - -
      -
        -
      • - - -

        Method Detail

        - - - -
          -
        • -

          values

          -
          public static TimestampFormat[] values()
          -
          Returns an array containing the constants of this enum type, in -the order they are declared. This method may be used to iterate -over the constants as follows: -
          -for (TimestampFormat c : TimestampFormat.values())
          -    System.out.println(c);
          -
          -
          -
          Returns:
          -
          an array containing the constants of this enum type, in the order they are declared
          -
          -
        • -
        - - - -
          -
        • -

          valueOf

          -
          public static TimestampFormat valueOf​(String name)
          -
          Returns the enum constant of this type with the specified name. -The string must match exactly an identifier used to declare an -enum constant in this type. (Extraneous whitespace characters are -not permitted.)
          -
          -
          Parameters:
          -
          name - the name of the enum constant to be returned.
          -
          Returns:
          -
          the enum constant with the specified name
          -
          Throws:
          -
          IllegalArgumentException - if this enum type has no constant with the specified name
          -
          NullPointerException - if the argument is null
          -
          -
        • -
        -
      • -
      -
      -
    • -
    -
    -
    -
    - - - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/CatalogExceptionUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/CatalogExceptionUtils.html deleted file mode 100644 index 13e558da9b..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/CatalogExceptionUtils.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.utils.CatalogExceptionUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.utils.CatalogExceptionUtils

    -
    -
    No usage of org.apache.fluss.flink.utils.CatalogExceptionUtils
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/CatalogPropertiesUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/CatalogPropertiesUtils.html deleted file mode 100644 index 3e4c6b3afc..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/CatalogPropertiesUtils.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.utils.CatalogPropertiesUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.utils.CatalogPropertiesUtils

    -
    -
    No usage of org.apache.fluss.flink.utils.CatalogPropertiesUtils
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/DataLakeUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/DataLakeUtils.html deleted file mode 100644 index 3473ba3ab3..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/DataLakeUtils.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.utils.DataLakeUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.utils.DataLakeUtils

    -
    -
    No usage of org.apache.fluss.flink.utils.DataLakeUtils
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConnectorOptionsUtils.StartupOptions.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConnectorOptionsUtils.StartupOptions.html deleted file mode 100644 index 89bc5031e9..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConnectorOptionsUtils.StartupOptions.html +++ /dev/null @@ -1,231 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils.StartupOptions (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils.StartupOptions

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConnectorOptionsUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConnectorOptionsUtils.html deleted file mode 100644 index 6526d73f12..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConnectorOptionsUtils.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils

    -
    -
    No usage of org.apache.fluss.flink.utils.FlinkConnectorOptionsUtils
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConversions.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConversions.html deleted file mode 100644 index e57305db49..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkConversions.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.utils.FlinkConversions (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.utils.FlinkConversions

    -
    -
    No usage of org.apache.fluss.flink.utils.FlinkConversions
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkRowToFlussRowConverter.FlussSerializationConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkRowToFlussRowConverter.FlussSerializationConverter.html deleted file mode 100644 index 1f1d9df228..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkRowToFlussRowConverter.FlussSerializationConverter.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.flink.utils.FlinkRowToFlussRowConverter.FlussSerializationConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.flink.utils.FlinkRowToFlussRowConverter.FlussSerializationConverter

    -
    -
    No usage of org.apache.fluss.flink.utils.FlinkRowToFlussRowConverter.FlussSerializationConverter
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkRowToFlussRowConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkRowToFlussRowConverter.html deleted file mode 100644 index ae401c9c5f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkRowToFlussRowConverter.html +++ /dev/null @@ -1,204 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.utils.FlinkRowToFlussRowConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.utils.FlinkRowToFlussRowConverter

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkUtils.html deleted file mode 100644 index e66391c0b1..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlinkUtils.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.utils.FlinkUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.utils.FlinkUtils

    -
    -
    No usage of org.apache.fluss.flink.utils.FlinkUtils
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToFlinkRowConverter.FlussDeserializationConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToFlinkRowConverter.FlussDeserializationConverter.html deleted file mode 100644 index f6c68bd6a7..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToFlinkRowConverter.FlussDeserializationConverter.html +++ /dev/null @@ -1,214 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter.FlussDeserializationConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter.FlussDeserializationConverter

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToFlinkRowConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToFlinkRowConverter.html deleted file mode 100644 index 52d36a7010..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToFlinkRowConverter.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter

    -
    -
    No usage of org.apache.fluss.flink.utils.FlussRowToFlinkRowConverter
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToJsonConverters.FlussRowToJsonConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToJsonConverters.FlussRowToJsonConverter.html deleted file mode 100644 index 6ca52dd26f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToJsonConverters.FlussRowToJsonConverter.html +++ /dev/null @@ -1,192 +0,0 @@ - - - - - -Uses of Interface org.apache.fluss.flink.utils.FlussRowToJsonConverters.FlussRowToJsonConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Interface
    org.apache.fluss.flink.utils.FlussRowToJsonConverters.FlussRowToJsonConverter

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToJsonConverters.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToJsonConverters.html deleted file mode 100644 index 7041db0bc6..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/FlussRowToJsonConverters.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.utils.FlussRowToJsonConverters (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.utils.FlussRowToJsonConverters

    -
    -
    No usage of org.apache.fluss.flink.utils.FlussRowToJsonConverters
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/LakeSourceUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/LakeSourceUtils.html deleted file mode 100644 index 7fc2e1ba97..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/LakeSourceUtils.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.utils.LakeSourceUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.utils.LakeSourceUtils

    -
    -
    No usage of org.apache.fluss.flink.utils.LakeSourceUtils
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PojoToRowConverter.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PojoToRowConverter.html deleted file mode 100644 index fe9bf965b7..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PojoToRowConverter.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.utils.PojoToRowConverter (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.utils.PojoToRowConverter

    -
    -
    No usage of org.apache.fluss.flink.utils.PojoToRowConverter
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.FieldEqual.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.FieldEqual.html deleted file mode 100644 index b4b1e60743..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.FieldEqual.html +++ /dev/null @@ -1,292 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.utils.PushdownUtils.FieldEqual (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.utils.PushdownUtils.FieldEqual

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.ValueConversion.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.ValueConversion.html deleted file mode 100644 index eb102e9728..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.ValueConversion.html +++ /dev/null @@ -1,223 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.utils.PushdownUtils.ValueConversion (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.utils.PushdownUtils.ValueConversion

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.html deleted file mode 100644 index 7e9d3b75f9..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/PushdownUtils.html +++ /dev/null @@ -1,148 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.utils.PushdownUtils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.utils.PushdownUtils

    -
    -
    No usage of org.apache.fluss.flink.utils.PushdownUtils
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/TimestampFormat.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/TimestampFormat.html deleted file mode 100644 index 2d3fcaf69c..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/class-use/TimestampFormat.html +++ /dev/null @@ -1,215 +0,0 @@ - - - - - -Uses of Class org.apache.fluss.flink.utils.TimestampFormat (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Class
    org.apache.fluss.flink.utils.TimestampFormat

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-summary.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-summary.html deleted file mode 100644 index e847293228..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-summary.html +++ /dev/null @@ -1,300 +0,0 @@ - - - - - -org.apache.fluss.flink.utils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Package org.apache.fluss.flink.utils

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-tree.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-tree.html deleted file mode 100644 index 1a9951226e..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-tree.html +++ /dev/null @@ -1,203 +0,0 @@ - - - - - -org.apache.fluss.flink.utils Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For Package org.apache.fluss.flink.utils

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -

    Interface Hierarchy

    - -
    -
    -

    Enum Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-use.html b/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-use.html deleted file mode 100644 index c401e778b1..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/org/apache/fluss/flink/utils/package-use.html +++ /dev/null @@ -1,276 +0,0 @@ - - - - - -Uses of Package org.apache.fluss.flink.utils (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Uses of Package
    org.apache.fluss.flink.utils

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/overview-summary.html b/website/static/javadoc/0.8-SNAPSHOT/overview-summary.html deleted file mode 100644 index f78f7d818f..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/overview-summary.html +++ /dev/null @@ -1,23 +0,0 @@ - - - - - -fluss-javadoc-temp 0.8-SNAPSHOT API - - - - - - - -
    - -

    index.html

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/overview-tree.html b/website/static/javadoc/0.8-SNAPSHOT/overview-tree.html deleted file mode 100644 index 430dcc2918..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/overview-tree.html +++ /dev/null @@ -1,581 +0,0 @@ - - - - - -Class Hierarchy (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Hierarchy For All Packages

    -Package Hierarchies: - -
    -
    -
    -

    Class Hierarchy

    - -
    -
    -

    Interface Hierarchy

    - -
    -
    -

    Enum Hierarchy

    - -
    -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/package-search-index.js b/website/static/javadoc/0.8-SNAPSHOT/package-search-index.js deleted file mode 100644 index 9c696ae856..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/package-search-index.js +++ /dev/null @@ -1 +0,0 @@ -packageSearchIndex = [{"l":"All Packages","url":"allpackages-index.html"},{"l":"org.apache.fluss.client"},{"l":"org.apache.fluss.client.admin"},{"l":"org.apache.fluss.client.lookup"},{"l":"org.apache.fluss.client.metadata"},{"l":"org.apache.fluss.client.metrics"},{"l":"org.apache.fluss.client.table"},{"l":"org.apache.fluss.client.table.getter"},{"l":"org.apache.fluss.client.table.scanner"},{"l":"org.apache.fluss.client.table.scanner.batch"},{"l":"org.apache.fluss.client.table.scanner.log"},{"l":"org.apache.fluss.client.table.writer"},{"l":"org.apache.fluss.client.token"},{"l":"org.apache.fluss.client.utils"},{"l":"org.apache.fluss.client.write"},{"l":"org.apache.fluss.flink"},{"l":"org.apache.fluss.flink.adapter"},{"l":"org.apache.fluss.flink.catalog"},{"l":"org.apache.fluss.flink.lake"},{"l":"org.apache.fluss.flink.lake.reader"},{"l":"org.apache.fluss.flink.lake.split"},{"l":"org.apache.fluss.flink.lake.state"},{"l":"org.apache.fluss.flink.metrics"},{"l":"org.apache.fluss.flink.procedure"},{"l":"org.apache.fluss.flink.row"},{"l":"org.apache.fluss.flink.sink"},{"l":"org.apache.fluss.flink.sink.serializer"},{"l":"org.apache.fluss.flink.sink.writer"},{"l":"org.apache.fluss.flink.source"},{"l":"org.apache.fluss.flink.source.deserializer"},{"l":"org.apache.fluss.flink.source.emitter"},{"l":"org.apache.fluss.flink.source.enumerator"},{"l":"org.apache.fluss.flink.source.enumerator.initializer"},{"l":"org.apache.fluss.flink.source.event"},{"l":"org.apache.fluss.flink.source.lookup"},{"l":"org.apache.fluss.flink.source.metrics"},{"l":"org.apache.fluss.flink.source.reader"},{"l":"org.apache.fluss.flink.source.reader.fetcher"},{"l":"org.apache.fluss.flink.source.split"},{"l":"org.apache.fluss.flink.source.state"},{"l":"org.apache.fluss.flink.tiering"},{"l":"org.apache.fluss.flink.tiering.committer"},{"l":"org.apache.fluss.flink.tiering.event"},{"l":"org.apache.fluss.flink.tiering.source"},{"l":"org.apache.fluss.flink.tiering.source.enumerator"},{"l":"org.apache.fluss.flink.tiering.source.split"},{"l":"org.apache.fluss.flink.tiering.source.state"},{"l":"org.apache.fluss.flink.utils"}] \ No newline at end of file diff --git a/website/static/javadoc/0.8-SNAPSHOT/package-search-index.zip b/website/static/javadoc/0.8-SNAPSHOT/package-search-index.zip deleted file mode 100644 index 605b6f06024075e1f995de828ea2f7451e1493e0..0000000000000000000000000000000000000000 GIT binary patch literal 0 HcmV?d00001 literal 489 zcmWIWW@Zs#;Nak3Sj%l3&42_Xf$W0BR3-ORj{)C#?<;{3d+Q+)l7 z81S^NZ`vhS#87sp!8Ylk=iO_oW`}ehRuCv({>6Qfa^UXSdXwIpo;$rce!tebd8^W& zwTpkS-7;gfcj%?}N(TL}Zz;D**~h3ehu@#>DjfVefyt_bclw;U>4y|6y)@FMJruqP3r&EEvk6! z@7(@B^Rs3+t{J-)pW25oWL8EG+2E4~J&oAmpz&;2l6B;fCb!@Cob zyX0O^y7GT%#pMY(740_l27(5)ug;y>A+}SrSGDr16!+6Co4~_L2OImM%l+Gp5|}!l zv-UBaJbCLL|MC}oSC>8EGh3FiG&l49!F2~Sqh`7;Q@a`AJ)6VVk|XQz`Wc@KHylk6 zo)MSUoGWypvl3&5a7+oB*K7*4P-e` aY@h<1aTDOp$_7%(2!!rHx(8@H0|Nl=0=Q5B diff --git a/website/static/javadoc/0.8-SNAPSHOT/packages b/website/static/javadoc/0.8-SNAPSHOT/packages deleted file mode 100644 index 07a069f6f9..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/packages +++ /dev/null @@ -1,185 +0,0 @@ -org.apache.fluss.metrics -org.apache.fluss.metrics.reporter -org.apache.fluss.metrics.utils -org.apache.fluss.metrics.groups -org.apache.fluss.metrics.registry -org.apache.fluss.rocksdb -org.apache.fluss.cluster -org.apache.fluss.predicate -org.apache.fluss.types -org.apache.fluss.record -org.apache.fluss.record.bytesview -org.apache.fluss.record.send -org.apache.fluss.memory -org.apache.fluss.config -org.apache.fluss.security.auth -org.apache.fluss.security.auth.sasl.plain -org.apache.fluss.security.auth.sasl.authenticator -org.apache.fluss.security.auth.sasl.jaas -org.apache.fluss.security.acl -org.apache.fluss.plugin -org.apache.fluss.annotation -org.apache.fluss.utils -org.apache.fluss.utils.types -org.apache.fluss.utils.crc -org.apache.fluss.utils.function -org.apache.fluss.utils.json -org.apache.fluss.utils.concurrent -org.apache.fluss.utils.log -org.apache.fluss.utils.clock -org.apache.fluss.classloading -org.apache.fluss.row -org.apache.fluss.row.columnar -org.apache.fluss.row.compacted -org.apache.fluss.row.arrow.writers -org.apache.fluss.row.arrow.vectors -org.apache.fluss.row.arrow -org.apache.fluss.row.encode -org.apache.fluss.row.encode.paimon -org.apache.fluss.row.encode.iceberg -org.apache.fluss.row.indexed -org.apache.fluss.row.decode -org.apache.fluss.lake.lakestorage -org.apache.fluss.lake.serializer -org.apache.fluss.lake.source -org.apache.fluss.lake.writer -org.apache.fluss.lake.committer -org.apache.fluss.lake.batch -org.apache.fluss.compression -org.apache.fluss.bucketing -org.apache.fluss.rpc.messages -org.apache.fluss.fs.token -org.apache.fluss.fs -org.apache.fluss.fs.utils -org.apache.fluss.fs.local -org.apache.fluss.exception -org.apache.fluss.metadata -org.apache.fluss.remote -org.apache.fluss.metrics.prometheus -org.apache.fluss.metrics.jmx -org.apache.fluss.client.write -org.apache.fluss.client.token -org.apache.fluss.client.metrics -org.apache.fluss.client -org.apache.fluss.client.admin -org.apache.fluss.client.lookup -org.apache.fluss.client.utils -org.apache.fluss.client.table.getter -org.apache.fluss.client.table -org.apache.fluss.client.table.writer -org.apache.fluss.client.table.scanner -org.apache.fluss.client.table.scanner.batch -org.apache.fluss.client.table.scanner.log -org.apache.fluss.client.metadata -org.apache.fluss.rpc.metrics -org.apache.fluss.rpc -org.apache.fluss.rpc.util -org.apache.fluss.rpc.entity -org.apache.fluss.rpc.protocol -org.apache.fluss.rpc.netty -org.apache.fluss.rpc.netty.server -org.apache.fluss.rpc.netty.client -org.apache.fluss.rpc.gateway -org.apache.fluss.dist -org.apache.fluss.fs.hdfs -org.apache.fluss.fs.hdfs.utils -org.apache.fluss.fs.oss.token -org.apache.fluss.fs.oss -org.apache.fluss.fs.s3.token -org.apache.fluss.fs.s3 -com.amazonaws.services.s3.model.transform -org.apache.fluss.fs.gs -org.apache.fluss.fs.obs.token -org.apache.fluss.fs.obs -org.apache.fluss.testutils.junit.parameterized -org.apache.fluss.testutils.common -org.apache.fluss.server.metrics.group -org.apache.fluss.server.metrics -org.apache.fluss.server.tablet -org.apache.fluss.server.zk -org.apache.fluss.server.zk.data -org.apache.fluss.server.coordinator -org.apache.fluss.server.coordinator.statemachine -org.apache.fluss.server.coordinator.event -org.apache.fluss.server.coordinator.event.watcher -org.apache.fluss.server.kv.rocksdb -org.apache.fluss.server.kv.snapshot -org.apache.fluss.server.kv.prewrite -org.apache.fluss.server.kv -org.apache.fluss.server.kv.wal -org.apache.fluss.server.kv.partialupdate -org.apache.fluss.server.kv.rowmerger -org.apache.fluss.server -org.apache.fluss.server.entity -org.apache.fluss.server.utils -org.apache.fluss.server.utils.timer -org.apache.fluss.server.cli -org.apache.fluss.server.authorizer -org.apache.fluss.server.log -org.apache.fluss.server.log.checkpoint -org.apache.fluss.server.log.remote -org.apache.fluss.server.replica -org.apache.fluss.server.replica.fetcher -org.apache.fluss.server.replica.delay -org.apache.fluss.server.exception -org.apache.fluss.server.metadata -org.apache.fluss.flink.metrics -org.apache.fluss.flink.catalog -org.apache.fluss.flink.source.reader -org.apache.fluss.flink.source.reader.fetcher -org.apache.fluss.flink.source.split -org.apache.fluss.flink.source.metrics -org.apache.fluss.flink.source -org.apache.fluss.flink.source.deserializer -org.apache.fluss.flink.source.emitter -org.apache.fluss.flink.source.lookup -org.apache.fluss.flink.source.state -org.apache.fluss.flink.source.event -org.apache.fluss.flink.source.enumerator.initializer -org.apache.fluss.flink.source.enumerator -org.apache.fluss.flink.tiering.source.split -org.apache.fluss.flink.tiering.source -org.apache.fluss.flink.tiering.source.state -org.apache.fluss.flink.tiering.source.enumerator -org.apache.fluss.flink.tiering -org.apache.fluss.flink.tiering.committer -org.apache.fluss.flink.tiering.event -org.apache.fluss.flink.sink -org.apache.fluss.flink.sink.serializer -org.apache.fluss.flink.sink.writer -org.apache.fluss.flink.utils -org.apache.fluss.flink -org.apache.fluss.flink.adapter -org.apache.fluss.flink.procedure -org.apache.fluss.flink.row -org.apache.fluss.flink.lake.reader -org.apache.fluss.flink.lake.split -org.apache.fluss.flink.lake -org.apache.fluss.flink.lake.state -org.apache.flink.streaming.api.connector.sink2 -org.apache.flink.api.connector.sink2 -org.apache.fluss.lake.paimon -org.apache.fluss.lake.paimon.source -org.apache.fluss.lake.paimon.tiering -org.apache.fluss.lake.paimon.tiering.append -org.apache.fluss.lake.paimon.tiering.mergetree -org.apache.fluss.lake.paimon.utils -org.apache.fluss.lake.iceberg -org.apache.fluss.lake.iceberg.source -org.apache.fluss.lake.iceberg.tiering -org.apache.fluss.lake.iceberg.tiering.writer -org.apache.fluss.lake.iceberg.utils -org.apache.fluss.lake.iceberg.maintenance -org.apache.fluss.lake.iceberg.conf -org.apache.iceberg.data -org.apache.fluss.lake.lance -org.apache.fluss.lake.lance.writers -org.apache.fluss.lake.lance.tiering -org.apache.fluss.lake.lance.utils -org.apache.fluss.kafka -org.apache.fluss.tools.ci.utils.deploy -org.apache.fluss.tools.ci.utils.dependency -org.apache.fluss.tools.ci.utils.shared -org.apache.fluss.tools.ci.utils.notice -org.apache.fluss.tools.ci.utils.shade -org.apache.fluss.tools.ci.licensecheck \ No newline at end of file diff --git a/website/static/javadoc/0.8-SNAPSHOT/resources/glass.png b/website/static/javadoc/0.8-SNAPSHOT/resources/glass.png deleted file mode 100644 index a7f591f467a1c0c949bbc510156a0c1afb860a6e..0000000000000000000000000000000000000000 GIT binary patch literal 0 HcmV?d00001 literal 499 zcmVJoRsvExf%rEN>jUL}qZ_~k#FbE+Q;{`;0FZwVNX2n-^JoI; zP;4#$8DIy*Yk-P>VN(DUKmPse7mx+ExD4O|;?E5D0Z5($mjO3`*anwQU^s{ZDK#Lz zj>~{qyaIx5K!t%=G&2IJNzg!ChRpyLkO7}Ry!QaotAHAMpbB3AF(}|_f!G-oI|uK6 z`id_dumai5K%C3Y$;tKS_iqMPHg<*|-@e`liWLAggVM!zAP#@l;=c>S03;{#04Z~5 zN_+ss=Yg6*hTr59mzMwZ@+l~q!+?ft!fF66AXT#wWavHt30bZWFCK%!BNk}LN?0Hg z1VF_nfs`Lm^DjYZ1(1uD0u4CSIr)XAaqW6IT{!St5~1{i=i}zAy76p%_|w8rh@@c0Axr!ns=D-X+|*sY6!@wacG9%)Qn*O zl0sa739kT-&_?#oVxXF6tOnqTD)cZ}2vi$`ZU8RLAlo8=_z#*P3xI~i!lEh+Pdu-L zx{d*wgjtXbnGX_Yf@Tc7Q3YhLhPvc8noGJs2DA~1DySiA&6V{5JzFt ojAY1KXm~va;tU{v7C?Xj0BHw!K;2aXV*mgE07*qoM6N<$f;4TDA^-pY diff --git a/website/static/javadoc/0.8-SNAPSHOT/script.js b/website/static/javadoc/0.8-SNAPSHOT/script.js deleted file mode 100644 index 7dc93c48e3..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/script.js +++ /dev/null @@ -1,149 +0,0 @@ -/* - * Copyright (c) 2013, 2018, Oracle and/or its affiliates. All rights reserved. - * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. - * - * This code is free software; you can redistribute it and/or modify it - * under the terms of the GNU General Public License version 2 only, as - * published by the Free Software Foundation. Oracle designates this - * particular file as subject to the "Classpath" exception as provided - * by Oracle in the LICENSE file that accompanied this code. - * - * This code is distributed in the hope that it will be useful, but WITHOUT - * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or - * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License - * version 2 for more details (a copy is included in the LICENSE file that - * accompanied this code). - * - * You should have received a copy of the GNU General Public License version - * 2 along with this work; if not, write to the Free Software Foundation, - * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. - * - * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA - * or visit www.oracle.com if you need additional information or have any - * questions. - */ - -var moduleSearchIndex; -var packageSearchIndex; -var typeSearchIndex; -var memberSearchIndex; -var tagSearchIndex; -function loadScripts(doc, tag) { - createElem(doc, tag, 'jquery/jszip/dist/jszip.js'); - createElem(doc, tag, 'jquery/jszip-utils/dist/jszip-utils.js'); - if (window.navigator.userAgent.indexOf('MSIE ') > 0 || window.navigator.userAgent.indexOf('Trident/') > 0 || - window.navigator.userAgent.indexOf('Edge/') > 0) { - createElem(doc, tag, 'jquery/jszip-utils/dist/jszip-utils-ie.js'); - } - createElem(doc, tag, 'search.js'); - - $.get(pathtoroot + "module-search-index.zip") - .done(function() { - JSZipUtils.getBinaryContent(pathtoroot + "module-search-index.zip", function(e, data) { - JSZip.loadAsync(data).then(function(zip){ - zip.file("module-search-index.json").async("text").then(function(content){ - moduleSearchIndex = JSON.parse(content); - }); - }); - }); - }); - $.get(pathtoroot + "package-search-index.zip") - .done(function() { - JSZipUtils.getBinaryContent(pathtoroot + "package-search-index.zip", function(e, data) { - JSZip.loadAsync(data).then(function(zip){ - zip.file("package-search-index.json").async("text").then(function(content){ - packageSearchIndex = JSON.parse(content); - }); - }); - }); - }); - $.get(pathtoroot + "type-search-index.zip") - .done(function() { - JSZipUtils.getBinaryContent(pathtoroot + "type-search-index.zip", function(e, data) { - JSZip.loadAsync(data).then(function(zip){ - zip.file("type-search-index.json").async("text").then(function(content){ - typeSearchIndex = JSON.parse(content); - }); - }); - }); - }); - $.get(pathtoroot + "member-search-index.zip") - .done(function() { - JSZipUtils.getBinaryContent(pathtoroot + "member-search-index.zip", function(e, data) { - JSZip.loadAsync(data).then(function(zip){ - zip.file("member-search-index.json").async("text").then(function(content){ - memberSearchIndex = JSON.parse(content); - }); - }); - }); - }); - $.get(pathtoroot + "tag-search-index.zip") - .done(function() { - JSZipUtils.getBinaryContent(pathtoroot + "tag-search-index.zip", function(e, data) { - JSZip.loadAsync(data).then(function(zip){ - zip.file("tag-search-index.json").async("text").then(function(content){ - tagSearchIndex = JSON.parse(content); - }); - }); - }); - }); - if (!moduleSearchIndex) { - createElem(doc, tag, 'module-search-index.js'); - } - if (!packageSearchIndex) { - createElem(doc, tag, 'package-search-index.js'); - } - if (!typeSearchIndex) { - createElem(doc, tag, 'type-search-index.js'); - } - if (!memberSearchIndex) { - createElem(doc, tag, 'member-search-index.js'); - } - if (!tagSearchIndex) { - createElem(doc, tag, 'tag-search-index.js'); - } - $(window).resize(function() { - $('.navPadding').css('padding-top', $('.fixedNav').css("height")); - }); -} - -function createElem(doc, tag, path) { - var script = doc.createElement(tag); - var scriptElement = doc.getElementsByTagName(tag)[0]; - script.src = pathtoroot + path; - scriptElement.parentNode.insertBefore(script, scriptElement); -} - -function show(type) { - count = 0; - for (var key in data) { - var row = document.getElementById(key); - if ((data[key] & type) !== 0) { - row.style.display = ''; - row.className = (count++ % 2) ? rowColor : altColor; - } - else - row.style.display = 'none'; - } - updateTabs(type); -} - -function updateTabs(type) { - for (var value in tabs) { - var sNode = document.getElementById(tabs[value][0]); - var spanNode = sNode.firstChild; - if (value == type) { - sNode.className = activeTableTab; - spanNode.innerHTML = tabs[value][1]; - } - else { - sNode.className = tableTab; - spanNode.innerHTML = "" + tabs[value][1] + ""; - } - } -} - -function updateModuleFrame(pFrame, cFrame) { - top.packageFrame.location = pFrame; - top.classFrame.location = cFrame; -} diff --git a/website/static/javadoc/0.8-SNAPSHOT/search.js b/website/static/javadoc/0.8-SNAPSHOT/search.js deleted file mode 100644 index 8492271e71..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/search.js +++ /dev/null @@ -1,326 +0,0 @@ -/* - * Copyright (c) 2015, 2018, Oracle and/or its affiliates. All rights reserved. - * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. - * - * This code is free software; you can redistribute it and/or modify it - * under the terms of the GNU General Public License version 2 only, as - * published by the Free Software Foundation. Oracle designates this - * particular file as subject to the "Classpath" exception as provided - * by Oracle in the LICENSE file that accompanied this code. - * - * This code is distributed in the hope that it will be useful, but WITHOUT - * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or - * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License - * version 2 for more details (a copy is included in the LICENSE file that - * accompanied this code). - * - * You should have received a copy of the GNU General Public License version - * 2 along with this work; if not, write to the Free Software Foundation, - * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. - * - * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA - * or visit www.oracle.com if you need additional information or have any - * questions. - */ - -var noResult = {l: "No results found"}; -var catModules = "Modules"; -var catPackages = "Packages"; -var catTypes = "Types"; -var catMembers = "Members"; -var catSearchTags = "SearchTags"; -var highlight = "$&"; -var camelCaseRegexp = ""; -var secondaryMatcher = ""; -function getHighlightedText(item) { - var ccMatcher = new RegExp(camelCaseRegexp); - var label = item.replace(ccMatcher, highlight); - if (label === item) { - label = item.replace(secondaryMatcher, highlight); - } - return label; -} -function getURLPrefix(ui) { - var urlPrefix=""; - if (useModuleDirectories) { - var slash = "/"; - if (ui.item.category === catModules) { - return ui.item.l + slash; - } else if (ui.item.category === catPackages && ui.item.m) { - return ui.item.m + slash; - } else if ((ui.item.category === catTypes && ui.item.p) || ui.item.category === catMembers) { - $.each(packageSearchIndex, function(index, item) { - if (item.m && ui.item.p == item.l) { - urlPrefix = item.m + slash; - } - }); - return urlPrefix; - } else { - return urlPrefix; - } - } - return urlPrefix; -} -var watermark = 'Search'; -$(function() { - $("#search").val(''); - $("#search").prop("disabled", false); - $("#reset").prop("disabled", false); - $("#search").val(watermark).addClass('watermark'); - $("#search").blur(function() { - if ($(this).val().length == 0) { - $(this).val(watermark).addClass('watermark'); - } - }); - $("#search").on('click keydown', function() { - if ($(this).val() == watermark) { - $(this).val('').removeClass('watermark'); - } - }); - $("#reset").click(function() { - $("#search").val(''); - $("#search").focus(); - }); - $("#search").focus(); - $("#search")[0].setSelectionRange(0, 0); -}); -$.widget("custom.catcomplete", $.ui.autocomplete, { - _create: function() { - this._super(); - this.widget().menu("option", "items", "> :not(.ui-autocomplete-category)"); - }, - _renderMenu: function(ul, items) { - var rMenu = this, - currentCategory = ""; - rMenu.menu.bindings = $(); - $.each(items, function(index, item) { - var li; - if (item.l !== noResult.l && item.category !== currentCategory) { - ul.append("
  • " + item.category + "
  • "); - currentCategory = item.category; - } - li = rMenu._renderItemData(ul, item); - if (item.category) { - li.attr("aria-label", item.category + " : " + item.l); - li.attr("class", "resultItem"); - } else { - li.attr("aria-label", item.l); - li.attr("class", "resultItem"); - } - }); - }, - _renderItem: function(ul, item) { - var label = ""; - if (item.category === catModules) { - label = getHighlightedText(item.l); - } else if (item.category === catPackages) { - label = (item.m) - ? getHighlightedText(item.m + "/" + item.l) - : getHighlightedText(item.l); - } else if (item.category === catTypes) { - label = (item.p) - ? getHighlightedText(item.p + "." + item.l) - : getHighlightedText(item.l); - } else if (item.category === catMembers) { - label = getHighlightedText(item.p + "." + (item.c + "." + item.l)); - } else if (item.category === catSearchTags) { - label = getHighlightedText(item.l); - } else { - label = item.l; - } - var li = $("
  • ").appendTo(ul); - var div = $("
    ").appendTo(li); - if (item.category === catSearchTags) { - if (item.d) { - div.html(label + " (" + item.h + ")
    " - + item.d + "
    "); - } else { - div.html(label + " (" + item.h + ")"); - } - } else { - div.html(label); - } - return li; - } -}); -$(function() { - $("#search").catcomplete({ - minLength: 1, - delay: 100, - source: function(request, response) { - var result = new Array(); - var presult = new Array(); - var tresult = new Array(); - var mresult = new Array(); - var tgresult = new Array(); - var secondaryresult = new Array(); - var displayCount = 0; - var exactMatcher = new RegExp("^" + $.ui.autocomplete.escapeRegex(request.term) + "$", "i"); - camelCaseRegexp = ($.ui.autocomplete.escapeRegex(request.term)).split(/(?=[A-Z])/).join("([a-z0-9_$]*?)"); - var camelCaseMatcher = new RegExp("^" + camelCaseRegexp); - secondaryMatcher = new RegExp($.ui.autocomplete.escapeRegex(request.term), "i"); - - // Return the nested innermost name from the specified object - function nestedName(e) { - return e.l.substring(e.l.lastIndexOf(".") + 1); - } - - function concatResults(a1, a2) { - a1 = a1.concat(a2); - a2.length = 0; - return a1; - } - - if (moduleSearchIndex) { - var mdleCount = 0; - $.each(moduleSearchIndex, function(index, item) { - item.category = catModules; - if (exactMatcher.test(item.l)) { - result.push(item); - mdleCount++; - } else if (camelCaseMatcher.test(item.l)) { - result.push(item); - } else if (secondaryMatcher.test(item.l)) { - secondaryresult.push(item); - } - }); - displayCount = mdleCount; - result = concatResults(result, secondaryresult); - } - if (packageSearchIndex) { - var pCount = 0; - var pkg = ""; - $.each(packageSearchIndex, function(index, item) { - item.category = catPackages; - pkg = (item.m) - ? (item.m + "/" + item.l) - : item.l; - if (exactMatcher.test(item.l)) { - presult.push(item); - pCount++; - } else if (camelCaseMatcher.test(pkg)) { - presult.push(item); - } else if (secondaryMatcher.test(pkg)) { - secondaryresult.push(item); - } - }); - result = result.concat(concatResults(presult, secondaryresult)); - displayCount = (pCount > displayCount) ? pCount : displayCount; - } - if (typeSearchIndex) { - var tCount = 0; - $.each(typeSearchIndex, function(index, item) { - item.category = catTypes; - var s = nestedName(item); - if (exactMatcher.test(s)) { - tresult.push(item); - tCount++; - } else if (camelCaseMatcher.test(s)) { - tresult.push(item); - } else if (secondaryMatcher.test(item.p + "." + item.l)) { - secondaryresult.push(item); - } - }); - result = result.concat(concatResults(tresult, secondaryresult)); - displayCount = (tCount > displayCount) ? tCount : displayCount; - } - if (memberSearchIndex) { - var mCount = 0; - $.each(memberSearchIndex, function(index, item) { - item.category = catMembers; - var s = nestedName(item); - if (exactMatcher.test(s)) { - mresult.push(item); - mCount++; - } else if (camelCaseMatcher.test(s)) { - mresult.push(item); - } else if (secondaryMatcher.test(item.c + "." + item.l)) { - secondaryresult.push(item); - } - }); - result = result.concat(concatResults(mresult, secondaryresult)); - displayCount = (mCount > displayCount) ? mCount : displayCount; - } - if (tagSearchIndex) { - var tgCount = 0; - $.each(tagSearchIndex, function(index, item) { - item.category = catSearchTags; - if (exactMatcher.test(item.l)) { - tgresult.push(item); - tgCount++; - } else if (secondaryMatcher.test(item.l)) { - secondaryresult.push(item); - } - }); - result = result.concat(concatResults(tgresult, secondaryresult)); - displayCount = (tgCount > displayCount) ? tgCount : displayCount; - } - displayCount = (displayCount > 500) ? displayCount : 500; - var counter = function() { - var count = {Modules: 0, Packages: 0, Types: 0, Members: 0, SearchTags: 0}; - var f = function(item) { - count[item.category] += 1; - return (count[item.category] <= displayCount); - }; - return f; - }(); - response(result.filter(counter)); - }, - response: function(event, ui) { - if (!ui.content.length) { - ui.content.push(noResult); - } else { - $("#search").empty(); - } - }, - autoFocus: true, - position: { - collision: "flip" - }, - select: function(event, ui) { - if (ui.item.l !== noResult.l) { - var url = getURLPrefix(ui); - if (ui.item.category === catModules) { - if (useModuleDirectories) { - url += "module-summary.html"; - } else { - url = ui.item.l + "-summary.html"; - } - } else if (ui.item.category === catPackages) { - if (ui.item.url) { - url = ui.item.url; - } else { - url += ui.item.l.replace(/\./g, '/') + "/package-summary.html"; - } - } else if (ui.item.category === catTypes) { - if (ui.item.url) { - url = ui.item.url; - } else if (ui.item.p === "") { - url += ui.item.l + ".html"; - } else { - url += ui.item.p.replace(/\./g, '/') + "/" + ui.item.l + ".html"; - } - } else if (ui.item.category === catMembers) { - if (ui.item.p === "") { - url += ui.item.c + ".html" + "#"; - } else { - url += ui.item.p.replace(/\./g, '/') + "/" + ui.item.c + ".html" + "#"; - } - if (ui.item.url) { - url += ui.item.url; - } else { - url += ui.item.l; - } - } else if (ui.item.category === catSearchTags) { - url += ui.item.u; - } - if (top !== window) { - parent.classFrame.location = pathtoroot + url; - } else { - window.location.href = pathtoroot + url; - } - $("#search").focus(); - } - } - }); -}); diff --git a/website/static/javadoc/0.8-SNAPSHOT/serialized-form.html b/website/static/javadoc/0.8-SNAPSHOT/serialized-form.html deleted file mode 100644 index 691102e404..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/serialized-form.html +++ /dev/null @@ -1,899 +0,0 @@ - - - - - -Serialized Form (fluss-javadoc-temp 0.8-SNAPSHOT API) - - - - - - - - - - - - - - - -
    - -
    -
    -
    -

    Serialized Form

    -
    -
    - -
    -
    -
    - -

    Copyright © 2025. All rights reserved.

    -
    - - diff --git a/website/static/javadoc/0.8-SNAPSHOT/stylesheet.css b/website/static/javadoc/0.8-SNAPSHOT/stylesheet.css deleted file mode 100644 index de945eda26..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/stylesheet.css +++ /dev/null @@ -1,910 +0,0 @@ -/* - * Javadoc style sheet - */ - -@import url('resources/fonts/dejavu.css'); - -/* - * Styles for individual HTML elements. - * - * These are styles that are specific to individual HTML elements. Changing them affects the style of a particular - * HTML element throughout the page. - */ - -body { - background-color:#ffffff; - color:#353833; - font-family:'DejaVu Sans', Arial, Helvetica, sans-serif; - font-size:14px; - margin:0; - padding:0; - height:100%; - width:100%; -} -iframe { - margin:0; - padding:0; - height:100%; - width:100%; - overflow-y:scroll; - border:none; -} -a:link, a:visited { - text-decoration:none; - color:#4A6782; -} -a[href]:hover, a[href]:focus { - text-decoration:none; - color:#bb7a2a; -} -a[name] { - color:#353833; -} -a[name]:before, a[name]:target, a[id]:before, a[id]:target { - content:""; - display:inline-block; - position:relative; - padding-top:129px; - margin-top:-129px; -} -pre { - font-family:'DejaVu Sans Mono', monospace; - font-size:14px; -} -h1 { - font-size:20px; -} -h2 { - font-size:18px; -} -h3 { - font-size:16px; - font-style:italic; -} -h4 { - font-size:13px; -} -h5 { - font-size:12px; -} -h6 { - font-size:11px; -} -ul { - list-style-type:disc; -} -code, tt { - font-family:'DejaVu Sans Mono', monospace; - font-size:14px; - padding-top:4px; - margin-top:8px; - line-height:1.4em; -} -dt code { - font-family:'DejaVu Sans Mono', monospace; - font-size:14px; - padding-top:4px; -} -table tr td dt code { - font-family:'DejaVu Sans Mono', monospace; - font-size:14px; - vertical-align:top; - padding-top:4px; -} -sup { - font-size:8px; -} - -/* - * Styles for HTML generated by javadoc. - * - * These are style classes that are used by the standard doclet to generate HTML documentation. - */ - -/* - * Styles for document title and copyright. - */ -.clear { - clear:both; - height:0px; - overflow:hidden; -} -.aboutLanguage { - float:right; - padding:0px 21px; - font-size:11px; - z-index:200; - margin-top:-9px; -} -.legalCopy { - margin-left:.5em; -} -.bar a, .bar a:link, .bar a:visited, .bar a:active { - color:#FFFFFF; - text-decoration:none; -} -.bar a:hover, .bar a:focus { - color:#bb7a2a; -} -.tab { - background-color:#0066FF; - color:#ffffff; - padding:8px; - width:5em; - font-weight:bold; -} -/* - * Styles for navigation bar. - */ -.bar { - background-color:#4D7A97; - color:#FFFFFF; - padding:.8em .5em .4em .8em; - height:auto;/*height:1.8em;*/ - font-size:11px; - margin:0; -} -.navPadding { - padding-top: 107px; -} -.fixedNav { - position:fixed; - width:100%; - z-index:999; - background-color:#ffffff; -} -.topNav { - background-color:#4D7A97; - color:#FFFFFF; - float:left; - padding:0; - width:100%; - clear:right; - height:2.8em; - padding-top:10px; - overflow:hidden; - font-size:12px; -} -.bottomNav { - margin-top:10px; - background-color:#4D7A97; - color:#FFFFFF; - float:left; - padding:0; - width:100%; - clear:right; - height:2.8em; - padding-top:10px; - overflow:hidden; - font-size:12px; -} -.subNav { - background-color:#dee3e9; - float:left; - width:100%; - overflow:hidden; - font-size:12px; -} -.subNav div { - clear:left; - float:left; - padding:0 0 5px 6px; - text-transform:uppercase; -} -ul.navList, ul.subNavList { - float:left; - margin:0 25px 0 0; - padding:0; -} -ul.navList li{ - list-style:none; - float:left; - padding: 5px 6px; - text-transform:uppercase; -} -ul.navListSearch { - float:right; - margin:0 0 0 0; - padding:0; -} -ul.navListSearch li { - list-style:none; - float:right; - padding: 5px 6px; - text-transform:uppercase; -} -ul.navListSearch li label { - position:relative; - right:-16px; -} -ul.subNavList li { - list-style:none; - float:left; -} -.topNav a:link, .topNav a:active, .topNav a:visited, .bottomNav a:link, .bottomNav a:active, .bottomNav a:visited { - color:#FFFFFF; - text-decoration:none; - text-transform:uppercase; -} -.topNav a:hover, .bottomNav a:hover { - text-decoration:none; - color:#bb7a2a; - text-transform:uppercase; -} -.navBarCell1Rev { - background-color:#F8981D; - color:#253441; - margin: auto 5px; -} -.skipNav { - position:absolute; - top:auto; - left:-9999px; - overflow:hidden; -} -/* - * Styles for page header and footer. - */ -.header, .footer { - clear:both; - margin:0 20px; - padding:5px 0 0 0; -} -.indexNav { - position:relative; - font-size:12px; - background-color:#dee3e9; -} -.indexNav ul { - margin-top:0; - padding:5px; -} -.indexNav ul li { - display:inline; - list-style-type:none; - padding-right:10px; - text-transform:uppercase; -} -.indexNav h1 { - font-size:13px; -} -.title { - color:#2c4557; - margin:10px 0; -} -.subTitle { - margin:5px 0 0 0; -} -.header ul { - margin:0 0 15px 0; - padding:0; -} -.footer ul { - margin:20px 0 5px 0; -} -.header ul li, .footer ul li { - list-style:none; - font-size:13px; -} -/* - * Styles for headings. - */ -div.details ul.blockList ul.blockList ul.blockList li.blockList h4, div.details ul.blockList ul.blockList ul.blockListLast li.blockList h4 { - background-color:#dee3e9; - border:1px solid #d0d9e0; - margin:0 0 6px -8px; - padding:7px 5px; -} -ul.blockList ul.blockList ul.blockList li.blockList h3 { - background-color:#dee3e9; - border:1px solid #d0d9e0; - margin:0 0 6px -8px; - padding:7px 5px; -} -ul.blockList ul.blockList li.blockList h3 { - padding:0; - margin:15px 0; -} -ul.blockList li.blockList h2 { - padding:0px 0 20px 0; -} -/* - * Styles for page layout containers. - */ -.contentContainer, .sourceContainer, .classUseContainer, .serializedFormContainer, .constantValuesContainer, -.allClassesContainer, .allPackagesContainer { - clear:both; - padding:10px 20px; - position:relative; -} -.indexContainer { - margin:10px; - position:relative; - font-size:12px; -} -.indexContainer h2 { - font-size:13px; - padding:0 0 3px 0; -} -.indexContainer ul { - margin:0; - padding:0; -} -.indexContainer ul li { - list-style:none; - padding-top:2px; -} -.contentContainer .description dl dt, .contentContainer .details dl dt, .serializedFormContainer dl dt { - font-size:12px; - font-weight:bold; - margin:10px 0 0 0; - color:#4E4E4E; -} -.contentContainer .description dl dd, .contentContainer .details dl dd, .serializedFormContainer dl dd { - margin:5px 0 10px 0px; - font-size:14px; - font-family:'DejaVu Serif', Georgia, "Times New Roman", Times, serif; -} -.serializedFormContainer dl.nameValue dt { - margin-left:1px; - font-size:1.1em; - display:inline; - font-weight:bold; -} -.serializedFormContainer dl.nameValue dd { - margin:0 0 0 1px; - font-size:1.1em; - display:inline; -} -/* - * Styles for lists. - */ -li.circle { - list-style:circle; -} -ul.horizontal li { - display:inline; - font-size:0.9em; -} -ul.inheritance { - margin:0; - padding:0; -} -ul.inheritance li { - display:inline; - list-style:none; -} -ul.inheritance li ul.inheritance { - margin-left:15px; - padding-left:15px; - padding-top:1px; -} -ul.blockList, ul.blockListLast { - margin:10px 0 10px 0; - padding:0; -} -ul.blockList li.blockList, ul.blockListLast li.blockList { - list-style:none; - margin-bottom:15px; - line-height:1.4; -} -ul.blockList ul.blockList li.blockList, ul.blockList ul.blockListLast li.blockList { - padding:0px 20px 5px 10px; - border:1px solid #ededed; - background-color:#f8f8f8; -} -ul.blockList ul.blockList ul.blockList li.blockList, ul.blockList ul.blockList ul.blockListLast li.blockList { - padding:0 0 5px 8px; - background-color:#ffffff; - border:none; -} -ul.blockList ul.blockList ul.blockList ul.blockList li.blockList { - margin-left:0; - padding-left:0; - padding-bottom:15px; - border:none; -} -ul.blockList ul.blockList ul.blockList ul.blockList li.blockListLast { - list-style:none; - border-bottom:none; - padding-bottom:0; -} -table tr td dl, table tr td dl dt, table tr td dl dd { - margin-top:0; - margin-bottom:1px; -} -/* - * Styles for tables. - */ -.overviewSummary, .memberSummary, .typeSummary, .useSummary, .constantsSummary, .deprecatedSummary, -.requiresSummary, .packagesSummary, .providesSummary, .usesSummary { - width:100%; - border-spacing:0; - border-left:1px solid #EEE; - border-right:1px solid #EEE; - border-bottom:1px solid #EEE; -} -.overviewSummary, .memberSummary, .requiresSummary, .packagesSummary, .providesSummary, .usesSummary { - padding:0px; -} -.overviewSummary caption, .memberSummary caption, .typeSummary caption, -.useSummary caption, .constantsSummary caption, .deprecatedSummary caption, -.requiresSummary caption, .packagesSummary caption, .providesSummary caption, .usesSummary caption { - position:relative; - text-align:left; - background-repeat:no-repeat; - color:#253441; - font-weight:bold; - clear:none; - overflow:hidden; - padding:0px; - padding-top:10px; - padding-left:1px; - margin:0px; - white-space:pre; -} -.constantsSummary caption a:link, .constantsSummary caption a:visited, -.useSummary caption a:link, .useSummary caption a:visited { - color:#1f389c; -} -.overviewSummary caption a:link, .memberSummary caption a:link, .typeSummary caption a:link, -.deprecatedSummary caption a:link, -.requiresSummary caption a:link, .packagesSummary caption a:link, .providesSummary caption a:link, -.usesSummary caption a:link, -.overviewSummary caption a:hover, .memberSummary caption a:hover, .typeSummary caption a:hover, -.useSummary caption a:hover, .constantsSummary caption a:hover, .deprecatedSummary caption a:hover, -.requiresSummary caption a:hover, .packagesSummary caption a:hover, .providesSummary caption a:hover, -.usesSummary caption a:hover, -.overviewSummary caption a:active, .memberSummary caption a:active, .typeSummary caption a:active, -.useSummary caption a:active, .constantsSummary caption a:active, .deprecatedSummary caption a:active, -.requiresSummary caption a:active, .packagesSummary caption a:active, .providesSummary caption a:active, -.usesSummary caption a:active, -.overviewSummary caption a:visited, .memberSummary caption a:visited, .typeSummary caption a:visited, -.deprecatedSummary caption a:visited, -.requiresSummary caption a:visited, .packagesSummary caption a:visited, .providesSummary caption a:visited, -.usesSummary caption a:visited { - color:#FFFFFF; -} -.overviewSummary caption span, .memberSummary caption span, .typeSummary caption span, -.useSummary caption span, .constantsSummary caption span, .deprecatedSummary caption span, -.requiresSummary caption span, .packagesSummary caption span, .providesSummary caption span, -.usesSummary caption span { - white-space:nowrap; - padding-top:5px; - padding-left:12px; - padding-right:12px; - padding-bottom:7px; - display:inline-block; - float:left; - background-color:#F8981D; - border: none; - height:16px; -} -.memberSummary caption span.activeTableTab span, .packagesSummary caption span.activeTableTab span, -.overviewSummary caption span.activeTableTab span, .typeSummary caption span.activeTableTab span { - white-space:nowrap; - padding-top:5px; - padding-left:12px; - padding-right:12px; - margin-right:3px; - display:inline-block; - float:left; - background-color:#F8981D; - height:16px; -} -.memberSummary caption span.tableTab span, .packagesSummary caption span.tableTab span, -.overviewSummary caption span.tableTab span, .typeSummary caption span.tableTab span { - white-space:nowrap; - padding-top:5px; - padding-left:12px; - padding-right:12px; - margin-right:3px; - display:inline-block; - float:left; - background-color:#4D7A97; - height:16px; -} -.memberSummary caption span.tableTab, .memberSummary caption span.activeTableTab, -.packagesSummary caption span.tableTab, .packagesSummary caption span.activeTableTab, -.overviewSummary caption span.tableTab, .overviewSummary caption span.activeTableTab, -.typeSummary caption span.tableTab, .typeSummary caption span.activeTableTab { - padding-top:0px; - padding-left:0px; - padding-right:0px; - background-image:none; - float:none; - display:inline; -} -.overviewSummary .tabEnd, .memberSummary .tabEnd, .typeSummary .tabEnd, -.useSummary .tabEnd, .constantsSummary .tabEnd, .deprecatedSummary .tabEnd, -.requiresSummary .tabEnd, .packagesSummary .tabEnd, .providesSummary .tabEnd, .usesSummary .tabEnd { - display:none; - width:5px; - position:relative; - float:left; - background-color:#F8981D; -} -.memberSummary .activeTableTab .tabEnd, .packagesSummary .activeTableTab .tabEnd, -.overviewSummary .activeTableTab .tabEnd, .typeSummary .activeTableTab .tabEnd { - display:none; - width:5px; - margin-right:3px; - position:relative; - float:left; - background-color:#F8981D; -} -.memberSummary .tableTab .tabEnd, .packagesSummary .tableTab .tabEnd, -.overviewSummary .tableTab .tabEnd, .typeSummary .tableTab .tabEnd { - display:none; - width:5px; - margin-right:3px; - position:relative; - background-color:#4D7A97; - float:left; -} -.rowColor th, .altColor th { - font-weight:normal; -} -.overviewSummary td, .memberSummary td, .typeSummary td, -.useSummary td, .constantsSummary td, .deprecatedSummary td, -.requiresSummary td, .packagesSummary td, .providesSummary td, .usesSummary td { - text-align:left; - padding:0px 0px 12px 10px; -} -th.colFirst, th.colSecond, th.colLast, th.colConstructorName, th.colDeprecatedItemName, .useSummary th, -.constantsSummary th, .packagesSummary th, td.colFirst, td.colSecond, td.colLast, .useSummary td, -.constantsSummary td { - vertical-align:top; - padding-right:0px; - padding-top:8px; - padding-bottom:3px; -} -th.colFirst, th.colSecond, th.colLast, th.colConstructorName, th.colDeprecatedItemName, .constantsSummary th, -.packagesSummary th { - background:#dee3e9; - text-align:left; - padding:8px 3px 3px 7px; -} -td.colFirst, th.colFirst { - font-size:13px; -} -td.colSecond, th.colSecond, td.colLast, th.colConstructorName, th.colDeprecatedItemName, th.colLast { - font-size:13px; -} -.constantsSummary th, .packagesSummary th { - font-size:13px; -} -.providesSummary th.colFirst, .providesSummary th.colLast, .providesSummary td.colFirst, -.providesSummary td.colLast { - white-space:normal; - font-size:13px; -} -.overviewSummary td.colFirst, .overviewSummary th.colFirst, -.requiresSummary td.colFirst, .requiresSummary th.colFirst, -.packagesSummary td.colFirst, .packagesSummary td.colSecond, .packagesSummary th.colFirst, .packagesSummary th, -.usesSummary td.colFirst, .usesSummary th.colFirst, -.providesSummary td.colFirst, .providesSummary th.colFirst, -.memberSummary td.colFirst, .memberSummary th.colFirst, -.memberSummary td.colSecond, .memberSummary th.colSecond, .memberSummary th.colConstructorName, -.typeSummary td.colFirst, .typeSummary th.colFirst { - vertical-align:top; -} -.packagesSummary th.colLast, .packagesSummary td.colLast { - white-space:normal; -} -td.colFirst a:link, td.colFirst a:visited, -td.colSecond a:link, td.colSecond a:visited, -th.colFirst a:link, th.colFirst a:visited, -th.colSecond a:link, th.colSecond a:visited, -th.colConstructorName a:link, th.colConstructorName a:visited, -th.colDeprecatedItemName a:link, th.colDeprecatedItemName a:visited, -.constantValuesContainer td a:link, .constantValuesContainer td a:visited, -.allClassesContainer td a:link, .allClassesContainer td a:visited, -.allPackagesContainer td a:link, .allPackagesContainer td a:visited { - font-weight:bold; -} -.tableSubHeadingColor { - background-color:#EEEEFF; -} -.altColor, .altColor th { - background-color:#FFFFFF; -} -.rowColor, .rowColor th { - background-color:#EEEEEF; -} -/* - * Styles for contents. - */ -.description pre { - margin-top:0; -} -.deprecatedContent { - margin:0; - padding:10px 0; -} -.docSummary { - padding:0; -} -ul.blockList ul.blockList ul.blockList li.blockList h3 { - font-style:normal; -} -div.block { - font-size:14px; - font-family:'DejaVu Serif', Georgia, "Times New Roman", Times, serif; -} -td.colLast div { - padding-top:0px; -} -td.colLast a { - padding-bottom:3px; -} -/* - * Styles for formatting effect. - */ -.sourceLineNo { - color:green; - padding:0 30px 0 0; -} -h1.hidden { - visibility:hidden; - overflow:hidden; - font-size:10px; -} -.block { - display:block; - margin:3px 10px 2px 0px; - color:#474747; -} -.deprecatedLabel, .descfrmTypeLabel, .implementationLabel, .memberNameLabel, .memberNameLink, -.moduleLabelInPackage, .moduleLabelInType, .overrideSpecifyLabel, .packageLabelInType, -.packageHierarchyLabel, .paramLabel, .returnLabel, .seeLabel, .simpleTagLabel, -.throwsLabel, .typeNameLabel, .typeNameLink, .searchTagLink { - font-weight:bold; -} -.deprecationComment, .emphasizedPhrase, .interfaceName { - font-style:italic; -} -.deprecationBlock { - font-size:14px; - font-family:'DejaVu Serif', Georgia, "Times New Roman", Times, serif; - border-style:solid; - border-width:thin; - border-radius:10px; - padding:10px; - margin-bottom:10px; - margin-right:10px; - display:inline-block; -} -div.block div.deprecationComment, div.block div.block span.emphasizedPhrase, -div.block div.block span.interfaceName { - font-style:normal; -} -div.contentContainer ul.blockList li.blockList h2 { - padding-bottom:0px; -} -/* - * Styles for IFRAME. - */ -.mainContainer { - margin:0 auto; - padding:0; - height:100%; - width:100%; - position:fixed; - top:0; - left:0; -} -.leftContainer { - height:100%; - position:fixed; - width:320px; -} -.leftTop { - position:relative; - float:left; - width:315px; - top:0; - left:0; - height:30%; - border-right:6px solid #ccc; - border-bottom:6px solid #ccc; -} -.leftBottom { - position:relative; - float:left; - width:315px; - bottom:0; - left:0; - height:70%; - border-right:6px solid #ccc; - border-top:1px solid #000; -} -.rightContainer { - position:absolute; - left:320px; - top:0; - bottom:0; - height:100%; - right:0; - border-left:1px solid #000; -} -.rightIframe { - margin:0; - padding:0; - height:100%; - right:30px; - width:100%; - overflow:visible; - margin-bottom:30px; -} -/* - * Styles specific to HTML5 elements. - */ -main, nav, header, footer, section { - display:block; -} -/* - * Styles for javadoc search. - */ -.ui-autocomplete-category { - font-weight:bold; - font-size:15px; - padding:7px 0 7px 3px; - background-color:#4D7A97; - color:#FFFFFF; -} -.resultItem { - font-size:13px; -} -.ui-autocomplete { - max-height:85%; - max-width:65%; - overflow-y:scroll; - overflow-x:scroll; - white-space:nowrap; - box-shadow: 0 3px 6px rgba(0,0,0,0.16), 0 3px 6px rgba(0,0,0,0.23); -} -ul.ui-autocomplete { - position:fixed; - z-index:999999; - background-color: #FFFFFF; -} -ul.ui-autocomplete li { - float:left; - clear:both; - width:100%; -} -.resultHighlight { - font-weight:bold; -} -.ui-autocomplete .result-item { - font-size: inherit; -} -#search { - background-image:url('resources/glass.png'); - background-size:13px; - background-repeat:no-repeat; - background-position:2px 3px; - padding-left:20px; - position:relative; - right:-18px; -} -#reset { - background-color: rgb(255,255,255); - background-image:url('resources/x.png'); - background-position:center; - background-repeat:no-repeat; - background-size:12px; - border:0 none; - width:16px; - height:17px; - position:relative; - left:-4px; - top:-4px; - font-size:0px; -} -.watermark { - color:#545454; -} -.searchTagDescResult { - font-style:italic; - font-size:11px; -} -.searchTagHolderResult { - font-style:italic; - font-size:12px; -} -.searchTagResult:before, .searchTagResult:target { - color:red; -} -.moduleGraph span { - display:none; - position:absolute; -} -.moduleGraph:hover span { - display:block; - margin: -100px 0 0 100px; - z-index: 1; -} -.methodSignature { - white-space:normal; -} - -/* - * Styles for user-provided tables. - * - * borderless: - * No borders, vertical margins, styled caption. - * This style is provided for use with existing doc comments. - * In general, borderless tables should not be used for layout purposes. - * - * plain: - * Plain borders around table and cells, vertical margins, styled caption. - * Best for small tables or for complex tables for tables with cells that span - * rows and columns, when the "striped" style does not work well. - * - * striped: - * Borders around the table and vertical borders between cells, striped rows, - * vertical margins, styled caption. - * Best for tables that have a header row, and a body containing a series of simple rows. - */ - -table.borderless, -table.plain, -table.striped { - margin-top: 10px; - margin-bottom: 10px; -} -table.borderless > caption, -table.plain > caption, -table.striped > caption { - font-weight: bold; - font-size: smaller; -} -table.borderless th, table.borderless td, -table.plain th, table.plain td, -table.striped th, table.striped td { - padding: 2px 5px; -} -table.borderless, -table.borderless > thead > tr > th, table.borderless > tbody > tr > th, table.borderless > tr > th, -table.borderless > thead > tr > td, table.borderless > tbody > tr > td, table.borderless > tr > td { - border: none; -} -table.borderless > thead > tr, table.borderless > tbody > tr, table.borderless > tr { - background-color: transparent; -} -table.plain { - border-collapse: collapse; - border: 1px solid black; -} -table.plain > thead > tr, table.plain > tbody tr, table.plain > tr { - background-color: transparent; -} -table.plain > thead > tr > th, table.plain > tbody > tr > th, table.plain > tr > th, -table.plain > thead > tr > td, table.plain > tbody > tr > td, table.plain > tr > td { - border: 1px solid black; -} -table.striped { - border-collapse: collapse; - border: 1px solid black; -} -table.striped > thead { - background-color: #E3E3E3; -} -table.striped > thead > tr > th, table.striped > thead > tr > td { - border: 1px solid black; -} -table.striped > tbody > tr:nth-child(even) { - background-color: #EEE -} -table.striped > tbody > tr:nth-child(odd) { - background-color: #FFF -} -table.striped > tbody > tr > th, table.striped > tbody > tr > td { - border-left: 1px solid black; - border-right: 1px solid black; -} -table.striped > tbody > tr > th { - font-weight: normal; -} diff --git a/website/static/javadoc/0.8-SNAPSHOT/type-search-index.js b/website/static/javadoc/0.8-SNAPSHOT/type-search-index.js deleted file mode 100644 index f0ec13f7c8..0000000000 --- a/website/static/javadoc/0.8-SNAPSHOT/type-search-index.js +++ /dev/null @@ -1 +0,0 @@ -typeSearchIndex = [{"p":"org.apache.fluss.flink.procedure","l":"AbstractAclProcedure"},{"p":"org.apache.fluss.client.lookup","l":"AbstractLookupBatch"},{"p":"org.apache.fluss.client.lookup","l":"AbstractLookupQuery"},{"p":"org.apache.fluss.client.table.writer","l":"AbstractTableWriter"},{"p":"org.apache.fluss.flink.procedure","l":"AddAclProcedure"},{"p":"org.apache.fluss.client.admin","l":"Admin"},{"l":"All Classes","url":"allclasses-index.html"},{"p":"org.apache.fluss.client.table.writer","l":"Append"},{"p":"org.apache.fluss.client.table.writer","l":"AppendResult"},{"p":"org.apache.fluss.flink.sink.writer","l":"AppendSinkWriter"},{"p":"org.apache.fluss.client.table.writer","l":"AppendWriter"},{"p":"org.apache.fluss.client.write","l":"ArrowLogWriteBatch"},{"p":"org.apache.fluss.client.table.scanner.batch","l":"BatchScanner"},{"p":"org.apache.fluss.client.table.scanner.batch","l":"BatchScanUtils"},{"p":"org.apache.fluss.flink.source.reader","l":"BoundedSplitReader"},{"p":"org.apache.fluss.flink.source.enumerator.initializer","l":"OffsetsInitializer.BucketOffsetsRetriever"},{"p":"org.apache.fluss.flink.source.enumerator.initializer","l":"BucketOffsetsRetrieverImpl"},{"p":"org.apache.fluss.flink.tiering.source","l":"TieringSource.Builder"},{"p":"org.apache.fluss.flink.utils","l":"CatalogExceptionUtils"},{"p":"org.apache.fluss.flink.utils","l":"CatalogPropertiesUtils"},{"p":"org.apache.fluss.flink.adapter","l":"CatalogTableAdapter"},{"p":"org.apache.fluss.flink.sink","l":"ChannelComputer"},{"p":"org.apache.fluss.client.utils","l":"ClientRpcMessageUtils"},{"p":"org.apache.fluss.client.utils","l":"ClientUtils"},{"p":"org.apache.fluss.flink.tiering.committer","l":"CommittableMessage"},{"p":"org.apache.fluss.flink.tiering.committer","l":"CommittableMessageTypeInfo"},{"p":"org.apache.fluss.client","l":"Connection"},{"p":"org.apache.fluss.client","l":"ConnectionFactory"},{"p":"org.apache.fluss.client.admin","l":"CreateAclsResult"},{"p":"org.apache.fluss.flink.utils","l":"DataLakeUtils"},{"p":"org.apache.fluss.client.token","l":"DefaultSecurityTokenManager"},{"p":"org.apache.fluss.client.token","l":"DefaultSecurityTokenProvider"},{"p":"org.apache.fluss.client.table.writer","l":"DeleteResult"},{"p":"org.apache.fluss.flink.source.deserializer","l":"DeserializerInitContextImpl"},{"p":"org.apache.fluss.flink.procedure","l":"DropAclProcedure"},{"p":"org.apache.fluss.client.admin","l":"DropAclsResult"},{"p":"org.apache.fluss.client.write","l":"DynamicPartitionCreator"},{"p":"org.apache.fluss.client.write","l":"DynamicWriteBatchSizeEstimator"},{"p":"org.apache.fluss.client.admin","l":"OffsetSpec.EarliestSpec"},{"p":"org.apache.fluss.flink.tiering.event","l":"FailedTieringEvent"},{"p":"org.apache.fluss.flink.utils","l":"PushdownUtils.FieldEqual"},{"p":"org.apache.fluss.client.admin","l":"DropAclsResult.FilterResult"},{"p":"org.apache.fluss.client.admin","l":"DropAclsResult.FilterResults"},{"p":"org.apache.fluss.flink.tiering.event","l":"FinishedTieringEvent"},{"p":"org.apache.fluss.flink.row","l":"FlinkAsFlussRow"},{"p":"org.apache.fluss.flink.source.lookup","l":"FlinkAsyncLookupFunction"},{"p":"org.apache.fluss.flink.catalog","l":"FlinkCatalog"},{"p":"org.apache.fluss.flink.catalog","l":"FlinkCatalogFactory"},{"p":"org.apache.fluss.flink.catalog","l":"FlinkCatalogOptions"},{"p":"org.apache.fluss.flink","l":"FlinkConnectorOptions"},{"p":"org.apache.fluss.flink.utils","l":"FlinkConnectorOptionsUtils"},{"p":"org.apache.fluss.flink.utils","l":"FlinkConversions"},{"p":"org.apache.fluss.flink.metrics","l":"FlinkCounter"},{"p":"org.apache.fluss.flink.metrics","l":"FlinkGauge"},{"p":"org.apache.fluss.flink.metrics","l":"FlinkHistogram"},{"p":"org.apache.fluss.flink.source.lookup","l":"FlinkLookupFunction"},{"p":"org.apache.fluss.flink.metrics","l":"FlinkMeter"},{"p":"org.apache.fluss.flink.metrics","l":"FlinkMetricRegistry"},{"p":"org.apache.fluss.flink.source.emitter","l":"FlinkRecordEmitter"},{"p":"org.apache.fluss.flink.source.reader","l":"FlinkRecordsWithSplitIds"},{"p":"org.apache.fluss.flink.sink","l":"FlinkRowDataChannelComputer"},{"p":"org.apache.fluss.flink.utils","l":"FlinkRowToFlussRowConverter"},{"p":"org.apache.fluss.flink.sink.writer","l":"FlinkSinkWriter"},{"p":"org.apache.fluss.flink.source","l":"FlinkSource"},{"p":"org.apache.fluss.flink.source.enumerator","l":"FlinkSourceEnumerator"},{"p":"org.apache.fluss.flink.source.reader.fetcher","l":"FlinkSourceFetcherManager"},{"p":"org.apache.fluss.flink.source.reader","l":"FlinkSourceReader"},{"p":"org.apache.fluss.flink.source.metrics","l":"FlinkSourceReaderMetrics"},{"p":"org.apache.fluss.flink.source.reader","l":"FlinkSourceSplitReader"},{"p":"org.apache.fluss.flink.sink","l":"FlinkStreamPartitioner"},{"p":"org.apache.fluss.flink.catalog","l":"FlinkTableFactory"},{"p":"org.apache.fluss.flink.sink","l":"FlinkTableSink"},{"p":"org.apache.fluss.flink.source","l":"FlinkTableSource"},{"p":"org.apache.fluss.flink.utils","l":"FlinkUtils"},{"p":"org.apache.fluss.client.admin","l":"FlussAdmin"},{"p":"org.apache.fluss.client","l":"FlussConnection"},{"p":"org.apache.fluss.flink.utils","l":"FlussRowToFlinkRowConverter.FlussDeserializationConverter"},{"p":"org.apache.fluss.flink.source.deserializer","l":"FlussDeserializationSchema"},{"p":"org.apache.fluss.flink.utils","l":"FlussRowToFlinkRowConverter"},{"p":"org.apache.fluss.flink.utils","l":"FlussRowToJsonConverters.FlussRowToJsonConverter"},{"p":"org.apache.fluss.flink.utils","l":"FlussRowToJsonConverters"},{"p":"org.apache.fluss.flink.utils","l":"FlinkRowToFlussRowConverter.FlussSerializationConverter"},{"p":"org.apache.fluss.flink.sink.serializer","l":"FlussSerializationSchema"},{"p":"org.apache.fluss.flink.sink","l":"FlussSink"},{"p":"org.apache.fluss.flink.sink","l":"FlussSinkBuilder"},{"p":"org.apache.fluss.flink.source","l":"FlussSource"},{"p":"org.apache.fluss.flink.source","l":"FlussSourceBuilder"},{"p":"org.apache.fluss.flink.source.state","l":"FlussSourceEnumeratorStateSerializer"},{"p":"org.apache.fluss.client.table","l":"FlussTable"},{"p":"org.apache.fluss.flink.tiering.committer","l":"FlussTableLakeSnapshotCommitter"},{"p":"org.apache.fluss.client.write","l":"HashBucketAssigner"},{"p":"org.apache.fluss.flink.source.split","l":"HybridSnapshotLogSplit"},{"p":"org.apache.fluss.flink.source.split","l":"HybridSnapshotLogSplitState"},{"p":"org.apache.fluss.client.write","l":"IdempotenceBucketEntry"},{"p":"org.apache.fluss.client.write","l":"IdempotenceBucketMap"},{"p":"org.apache.fluss.client.write","l":"IdempotenceManager"},{"p":"org.apache.fluss.flink.lake.reader","l":"IndexedLakeSplitRecordIterator"},{"p":"org.apache.fluss.client.write","l":"IndexedLogWriteBatch"},{"p":"org.apache.fluss.flink.sink.serializer","l":"FlussSerializationSchema.InitializationContext"},{"p":"org.apache.fluss.flink.source.deserializer","l":"FlussDeserializationSchema.InitializationContext"},{"p":"org.apache.fluss.flink.source.deserializer","l":"JsonStringDeserializationSchema"},{"p":"org.apache.fluss.flink.lake.reader","l":"KeyValueRow"},{"p":"org.apache.fluss.client.table.scanner.batch","l":"KvSnapshotBatchScanner"},{"p":"org.apache.fluss.client.metadata","l":"KvSnapshotMetadata"},{"p":"org.apache.fluss.client.metadata","l":"KvSnapshots"},{"p":"org.apache.fluss.client.write","l":"KvWriteBatch"},{"p":"org.apache.fluss.flink.lake","l":"LakeCatalog"},{"p":"org.apache.fluss.flink.lake","l":"LakeRecordRecordEmitter"},{"p":"org.apache.fluss.client.metadata","l":"LakeSnapshot"},{"p":"org.apache.fluss.flink.lake.split","l":"LakeSnapshotAndFlussLogSplit"},{"p":"org.apache.fluss.flink.lake.state","l":"LakeSnapshotAndFlussLogSplitState"},{"p":"org.apache.fluss.flink.lake.reader","l":"LakeSnapshotAndLogSplitScanner"},{"p":"org.apache.fluss.flink.lake.reader","l":"LakeSnapshotScanner"},{"p":"org.apache.fluss.flink.lake.split","l":"LakeSnapshotSplit"},{"p":"org.apache.fluss.flink.lake.state","l":"LakeSnapshotSplitState"},{"p":"org.apache.fluss.flink.utils","l":"LakeSourceUtils"},{"p":"org.apache.fluss.flink.lake","l":"LakeSplitGenerator"},{"p":"org.apache.fluss.flink.lake","l":"LakeSplitReaderGenerator"},{"p":"org.apache.fluss.flink.lake","l":"LakeSplitSerializer"},{"p":"org.apache.fluss.flink.lake","l":"LakeSplitStateInitializer"},{"p":"org.apache.fluss.flink.lake","l":"LakeTableFactory"},{"p":"org.apache.fluss.flink.tiering","l":"LakeTieringJobBuilder"},{"p":"org.apache.fluss.flink.source.enumerator.initializer","l":"LatestOffsetsInitializer"},{"p":"org.apache.fluss.client.admin","l":"OffsetSpec.LatestSpec"},{"p":"org.apache.fluss.client.table.scanner.batch","l":"LimitBatchScanner"},{"p":"org.apache.fluss.flink.procedure","l":"ListAclProcedure"},{"p":"org.apache.fluss.client.admin","l":"ListOffsetsResult"},{"p":"org.apache.fluss.client.table.scanner.log","l":"LogFetchBuffer"},{"p":"org.apache.fluss.client.table.scanner.log","l":"LogFetchCollector"},{"p":"org.apache.fluss.client.table.scanner.log","l":"LogFetcher"},{"p":"org.apache.fluss.client.table.scanner.log","l":"LogScan"},{"p":"org.apache.fluss.client.table.scanner.log","l":"LogScanner"},{"p":"org.apache.fluss.client.table.scanner.log","l":"LogScannerImpl"},{"p":"org.apache.fluss.client.table.scanner.log","l":"LogScannerStatus"},{"p":"org.apache.fluss.flink.source.split","l":"LogSplit"},{"p":"org.apache.fluss.flink.source.split","l":"LogSplitState"},{"p":"org.apache.fluss.client.lookup","l":"Lookup"},{"p":"org.apache.fluss.client.lookup","l":"LookupBatch"},{"p":"org.apache.fluss.client.lookup","l":"LookupClient"},{"p":"org.apache.fluss.client.lookup","l":"Lookuper"},{"p":"org.apache.fluss.flink.source.lookup","l":"LookupNormalizer"},{"p":"org.apache.fluss.client.lookup","l":"LookupQuery"},{"p":"org.apache.fluss.client.lookup","l":"LookupResult"},{"p":"org.apache.fluss.client.lookup","l":"LookupType"},{"p":"org.apache.fluss.client.metadata","l":"MetadataUpdater"},{"p":"org.apache.fluss.client.utils","l":"MetadataUtils"},{"p":"org.apache.fluss.flink.source.reader","l":"MutableRecordAndPos"},{"p":"org.apache.fluss.flink.source.enumerator.initializer","l":"NoStoppingOffsetsInitializer"},{"p":"org.apache.fluss.flink.source.enumerator.initializer","l":"OffsetsInitializer"},{"p":"org.apache.fluss.client.admin","l":"OffsetSpec"},{"p":"org.apache.fluss.flink.row","l":"OperationType"},{"p":"org.apache.fluss.flink.source.event","l":"PartitionBucketsUnsubscribedEvent"},{"p":"org.apache.fluss.client.table.getter","l":"PartitionGetter"},{"p":"org.apache.fluss.flink.source.event","l":"PartitionsRemovedEvent"},{"p":"org.apache.fluss.flink.utils","l":"PojoToRowConverter"},{"p":"org.apache.fluss.client.lookup","l":"PrefixLookup"},{"p":"org.apache.fluss.client.lookup","l":"PrefixLookupBatch"},{"p":"org.apache.fluss.client.lookup","l":"PrefixLookupQuery"},{"p":"org.apache.fluss.flink.procedure","l":"ProcedureBase"},{"p":"org.apache.fluss.flink.procedure","l":"ProcedureManager"},{"p":"org.apache.fluss.flink.utils","l":"PushdownUtils"},{"p":"org.apache.fluss.client.write","l":"RecordAccumulator.ReadyCheckResult"},{"p":"org.apache.fluss.client.write","l":"ReadyWriteBatch"},{"p":"org.apache.fluss.client.write","l":"RecordAccumulator"},{"p":"org.apache.fluss.flink.source.reader","l":"RecordAndPos"},{"p":"org.apache.fluss.client.write","l":"RecordAccumulator.RecordAppendResult"},{"p":"org.apache.fluss.flink.source.lookup","l":"LookupNormalizer.RemainingFilter"},{"p":"org.apache.fluss.client.table.scanner","l":"RemoteFileDownloader"},{"p":"org.apache.fluss.client.table.scanner.log","l":"RemoteLogDownloader"},{"p":"org.apache.fluss.client.table.scanner.log","l":"RemoteLogDownloadFuture"},{"p":"org.apache.fluss.client.write","l":"WriteBatch.RequestFuture"},{"p":"org.apache.fluss.client.write","l":"RoundRobinBucketAssigner"},{"p":"org.apache.fluss.flink.source.deserializer","l":"RowDataDeserializationSchema"},{"p":"org.apache.fluss.flink.sink.serializer","l":"RowDataSerializationSchema"},{"p":"org.apache.fluss.flink.row","l":"RowWithOp"},{"p":"org.apache.fluss.client.table.scanner","l":"Scan"},{"p":"org.apache.fluss.client.metrics","l":"ScannerMetricGroup"},{"p":"org.apache.fluss.client.table.scanner","l":"ScanRecord"},{"p":"org.apache.fluss.client.table.scanner.log","l":"ScanRecords"},{"p":"org.apache.fluss.flink","l":"FlinkConnectorOptions.ScanStartupMode"},{"p":"org.apache.fluss.client.token","l":"SecurityTokenManager"},{"p":"org.apache.fluss.client.token","l":"SecurityTokenProvider"},{"p":"org.apache.fluss.flink.lake.reader","l":"SeekableLakeSnapshotSplitScanner"},{"p":"org.apache.fluss.client.write","l":"Sender"},{"p":"org.apache.fluss.flink.sink.serializer","l":"SerializerInitContextImpl"},{"p":"org.apache.fluss.flink.adapter","l":"SingleThreadFetcherManagerAdapter"},{"p":"org.apache.fluss.flink.adapter","l":"SingleThreadMultiplexSourceReaderBaseAdapter"},{"p":"org.apache.fluss.flink.source.enumerator.initializer","l":"SnapshotOffsetsInitializer"},{"p":"org.apache.fluss.flink.source.split","l":"SnapshotSplit"},{"p":"org.apache.fluss.flink.source.state","l":"SourceEnumeratorState"},{"p":"org.apache.fluss.flink.source.split","l":"SourceSplitBase"},{"p":"org.apache.fluss.flink.source.split","l":"SourceSplitSerializer"},{"p":"org.apache.fluss.flink.source.split","l":"SourceSplitState"},{"p":"org.apache.fluss.flink.source.reader","l":"SplitScanner"},{"p":"org.apache.fluss.flink.utils","l":"FlinkConnectorOptionsUtils.StartupOptions"},{"p":"org.apache.fluss.client.write","l":"StickyBucketAssigner"},{"p":"org.apache.fluss.client.table","l":"Table"},{"p":"org.apache.fluss.client.table.writer","l":"TableAppend"},{"p":"org.apache.fluss.flink.tiering.source","l":"TableBucketWriteResult"},{"p":"org.apache.fluss.flink.tiering.source","l":"TableBucketWriteResultEmitter"},{"p":"org.apache.fluss.flink.tiering.source","l":"TableBucketWriteResultSerializer"},{"p":"org.apache.fluss.flink.tiering.source","l":"TableBucketWriteResultTypeInfo"},{"p":"org.apache.fluss.client.lookup","l":"TableLookup"},{"p":"org.apache.fluss.client.table.scanner","l":"TableScan"},{"p":"org.apache.fluss.client.table.writer","l":"TableUpsert"},{"p":"org.apache.fluss.client.table.writer","l":"TableWriter"},{"p":"org.apache.fluss.flink.tiering.committer","l":"TieringCommitOperator"},{"p":"org.apache.fluss.flink.tiering.committer","l":"TieringCommitOperatorFactory"},{"p":"org.apache.fluss.flink.tiering.committer","l":"TieringCommitterInitContext"},{"p":"org.apache.fluss.flink.tiering.source.split","l":"TieringLogSplit"},{"p":"org.apache.fluss.flink.tiering.event","l":"TieringRestoreEvent"},{"p":"org.apache.fluss.flink.tiering.source.split","l":"TieringSnapshotSplit"},{"p":"org.apache.fluss.flink.tiering.source","l":"TieringSource"},{"p":"org.apache.fluss.flink.tiering.source.enumerator","l":"TieringSourceEnumerator"},{"p":"org.apache.fluss.flink.tiering.source.state","l":"TieringSourceEnumeratorState"},{"p":"org.apache.fluss.flink.tiering.source.state","l":"TieringSourceEnumeratorStateSerializer"},{"p":"org.apache.fluss.flink.tiering.source","l":"TieringSourceOptions"},{"p":"org.apache.fluss.flink.tiering.source","l":"TieringSourceReader"},{"p":"org.apache.fluss.flink.tiering.source.split","l":"TieringSplit"},{"p":"org.apache.fluss.flink.tiering.source.split","l":"TieringSplitGenerator"},{"p":"org.apache.fluss.flink.tiering.source","l":"TieringSplitReader"},{"p":"org.apache.fluss.flink.tiering.source.split","l":"TieringSplitSerializer"},{"p":"org.apache.fluss.flink.tiering.source.state","l":"TieringSplitState"},{"p":"org.apache.fluss.flink.tiering.source","l":"TieringWriterInitContext"},{"p":"org.apache.fluss.flink.utils","l":"TimestampFormat"},{"p":"org.apache.fluss.flink.source.enumerator.initializer","l":"TimestampOffsetsInitializer"},{"p":"org.apache.fluss.client.admin","l":"OffsetSpec.TimestampSpec"},{"p":"org.apache.fluss.client.table.writer","l":"Upsert"},{"p":"org.apache.fluss.client.table.writer","l":"UpsertResult"},{"p":"org.apache.fluss.flink.sink.writer","l":"UpsertSinkWriter"},{"p":"org.apache.fluss.client.table.writer","l":"UpsertWrite"},{"p":"org.apache.fluss.client.table.writer","l":"UpsertWriter"},{"p":"org.apache.fluss.flink.utils","l":"PushdownUtils.ValueConversion"},{"p":"org.apache.fluss.client.write","l":"WriteBatch"},{"p":"org.apache.fluss.client.write","l":"WriteCallback"},{"p":"org.apache.fluss.client.write","l":"WriteFormat"},{"p":"org.apache.fluss.client.write","l":"WriterClient"},{"p":"org.apache.fluss.client.write","l":"WriteRecord"},{"p":"org.apache.fluss.client.metrics","l":"WriterMetricGroup"}] \ No newline at end of file diff --git a/website/static/javadoc/0.8-SNAPSHOT/type-search-index.zip b/website/static/javadoc/0.8-SNAPSHOT/type-search-index.zip deleted file mode 100644 index a787845eeaab0ee8dba81b769c96537531e7a2fc..0000000000000000000000000000000000000000 GIT binary patch literal 0 HcmV?d00001 literal 2356 zcmZveYd8}O8^_0tAq)|jIb=3ujE3gLa+=dLLdu?yQ%nwXjwdE^$Z0~(G|7@gnL~3f zhhZW)7db{HA|>VV^n7}s_qyKS{o(#!*Z;om&p)Cykmo1>3~a9e$8i0wXWmAzX{TG=4MVM^K8%yW zB{j@6uCeV0(#^a>59M>Bb#pLbH*X)WXr0f{pRjQ3A4?Uh*8*xDCFy`_X>MW0R*GQy zav5yb_vwQ3R?-rVGU*w8a)v4)iRlWQj4F`%ahIVSI$JInPqq)3ZE6V8SatWCiio%}UXx-|oh#+N7@(#+FHodpl+qKvC1K z1b1-@N)x2yrKa88lsPVL(q?)?gEKO*8o8^Y8&tgV6f7dJtxrbDHE*$o_W8>e{bn8` zhXo1>gh74z0zF^x_O%@t+;?_!Yeh?zd7v=v@7zB>!Vgrj#l`miz>$W06FTmg%gQdg zxNowt^vQ#}rXY4HEfi_RY$&X&15aIs0^L+~J%>RmIN~>}v?g1%Vt9zdU~p{*F5LmU zH%F38CKfkcimv73cKEAhfmM82Bbl z6X6*y+xA8m7aoz~T_>^O`ts1@K8e~+?{~uG<0oB^zhp%930rcML)it>#L5LVCxCLo z300)~CFl2+X@=a=>zbux&3@3xTmiKvO6=N(UQ)hXpK%6a+D%%o9vI}#+gig8S>-*s zruu%wL&j@b*j1+@OJ+4i_`Y$xq(Rshpk7RKJDAI)bNwf?#{OAQm^NL$ME|L@lwi}3 z8#;9!u>RQb5$-I(RB~!>e(nVs;3fw<>Vbcic4~n~<^U-a(RJUx6?eWc5<27)@F)Xu z0yx>(o;`~>XR~CBkJUS?EM#sHlU0yt-Xagw4?Yp+Jkub>kSHSCppW+v=TO5z+VPrhF9VNu^EY z?fyn=0!heXGkl~PMj<)ayt6N9z4)fJSSbpU46nY-6=y>o7=xOx@a6vMYccc|&y^s= zKn$@nq>sx)zS@+H=Mr2r3Ub=1$Y57)^vzhnWZ04fF5Z0(kV{eO(E>zl@7ER`L zox@s#RgEpOlPTBIb zV1AazvjmRUM@2ALx**F*-+LSYL>t%ot3Rl(@TG#*`#6W8%z-)$nh7{yB9g$GIooG zDoVrVSg$;1@sJ>6yQ-RLS|51}g^%m4L<-26&+zfP-xkcErCfn)#ov#UJ;5J-PI&rp z4Jz()zk(1sngZs2G^g-DVEwL5wMWt8++RmN*{pea6nW*PHwv48;Nwx=#C$emJGV04 zF*nUgfwqzzP~0C0dMs<`=%=$?tNmy1bBuHDpy4z*fc=ZP%qjH2Tv&Q@t{1vjuimHy zHCV$#iB7?`*1gZ-+oIktGIXDIObgzC%l?tKLs}Aw%KYNn=$YX2w8x;mo1=B@T5wYH zdANBZ)iQHbEOx~4Ddy`*a_UXLCHvxi&DjNixzQVU6M-{)h$^wA4z~=!X3>313w?xq zkmVG&C7)9OVbes)WgX$7NreO2?^M}NhqBow(MR&{$r%C3qhRpc# zt|#_{GPOo!FNf?2+DE^^7j`MA2*WzFG^8rDBI^Z_xrykg3SAdXZWbX|shj9XjAdeg zdu#Q&GQe9C%o8u*LTV}Tv|^Sdf$7=eI{4{Q`>i4_APdm9ZK=}xbL!w+10$X5i>ap~cB2q$bK+MSs3 z(k>|^@7?9q;7`GEBSpSEkrSL7;&u6Vdw7-u-3`c=A5B}+gHX4{m1gD2X#;jLyE$fk ztR%ws*7tp==hYv!s(7sCY)E}qrR)UaxOI|2^4(z?fTJ^LBc-JKl0vNYW;$cNjp{Q$ zocR|s+Yowf{6EK%b{Xm8OS3$UDxF0?8_IV1sEI2RRp=ghM&kN#XDpL3Lc05Kqa`ux z8RAaTne(BC3Gs^Qhe>8saqLG^Xk1sc=-El%#BNbjA|xfD?e+B3NT(#(2vUO=E!SK~ z)e-=r76~YkIVyj#)2AP&C3uFM?@wGI1;inqtPcvR+#AH1F^T;A$dY0umcjC_-%SHg z+;kA6XtAP@mUggPOKbl$$>KE`bVb6-jBP^2hri$@Xg`XA>ZR$S=}%IG*=Wl$<4pE= zr`1G(yk<0ds5Jk)a5f_LK!j+`2X5TZ=gR{F00bNWK(yxKJ_`7++x)xx$8G+X|8F`( UYY^{0cU-^6^!J|m{Tl%I8?JFQb^rhX diff --git a/website/static/javadoc/index.html b/website/static/javadoc/index.html deleted file mode 100644 index b9c9f98d2b..0000000000 --- a/website/static/javadoc/index.html +++ /dev/null @@ -1,3 +0,0 @@ - - -Apache Fluss 0.8-SNAPSHOT API From 64c5c5942e298d1198685cf52e13570fc821f290 Mon Sep 17 00:00:00 2001 From: Mehul Batra Date: Sun, 28 Sep 2025 19:10:48 +0530 Subject: [PATCH 4/4] ignore javadocs directory to get added --- .gitignore | 1 + pom.xml | 56 ------------------------------------------------------ 2 files changed, 1 insertion(+), 56 deletions(-) diff --git a/.gitignore b/.gitignore index f0c5b601e9..0801dfe947 100644 --- a/.gitignore +++ b/.gitignore @@ -42,3 +42,4 @@ website/versioned_docs website/versioned_sidebars website/versions.json website/pnpm-lock.yaml +website/static/javadoc diff --git a/pom.xml b/pom.xml index 67df7a52ca..16ebed5021 100644 --- a/pom.xml +++ b/pom.xml @@ -162,11 +162,6 @@ -Djdk.reflect.useDirectMethodHandle=false -Dio.netty.tryReflectionSetAccessible=true - - false - false - true - 3.6.3 @@ -536,57 +531,6 @@ - - javadoc-aggregate - - - javadoc.aggregate - - - - - fluss-client - fluss-flink - - - - - org.apache.maven.plugins - maven-javadoc-plugin - ${javadoc.version} - - - 11 - UTF-8 - Apache Fluss ${project.version} API - Apache Fluss ${project.version} API - protected - true - true - true - true - true - none - false - false - true - false - - -J-Xmx2g - -J--illegal-access=permit - - *.internal.*:*.impl.*:*.test.*:*.tests.*:*.generated.*:*.benchmark.*:*.protogen.* - - - - https://docs.oracle.com/en/java/javase/11/docs/api/ - https://nightlies.apache.org/flink/flink-docs-master/api/java/ - - - - - -