package cloud
- Alphabetic
- Public
- All
Type Members
- class AbortableStreamBasedCheckpointFileManager extends AbstractFileContextBasedCheckpointFileManager with Logging
-
class
BindingParquetOutputCommitter extends ParquetOutputCommitter with Logging with StreamCapabilities
This Parquet Committer subclass dynamically binds to the factory-configured output committer, and is intended to allow callers to use any 'PathOutputCommitter', even if not a subclass of 'ParquetOutputCommitter'.
This Parquet Committer subclass dynamically binds to the factory-configured output committer, and is intended to allow callers to use any 'PathOutputCommitter', even if not a subclass of 'ParquetOutputCommitter'.
The Parquet
parquet.enable.summary-metadataoption will only be supported if the instantiated committer itself supports it. -
class
PathOutputCommitProtocol extends HadoopMapReduceCommitProtocol with Serializable
Spark Commit protocol for Path Output Committers.
Spark Commit protocol for Path Output Committers. This committer will work with the
FileOutputCommitterand subclasses. All implementations *must* be serializable.Rather than ask the
FileOutputFormatfor a committer, it uses theorg.apache.hadoop.mapreduce.lib.output.PathOutputCommitterFactoryfactory API to create the committer.In
setupCommitterthe factory is identified and instantiated; this factory then creates the actual committer implementation.Dynamic Partition support will be determined once the committer is instantiated in the setupJob/setupTask methods. If this class was instantiated with
dynamicPartitionOverwriteset to true, then the instantiated committer must either be an instance ofFileOutputCommitteror it must implement theStreamCapabilitiesinterface and declare that it has the capabilitymapreduce.job.committer.dynamic.partitioning. That feature is available on Hadoop releases with the Intermediate Manifest Committer for GCS and ABFS; it is not supported by the S3A committers.
Value Members
- object PathOutputCommitProtocol extends Serializable