org.apache.spark.internal.io.cloud
PathOutputCommitProtocol
Companion object PathOutputCommitProtocol
class PathOutputCommitProtocol extends HadoopMapReduceCommitProtocol with Serializable
Spark Commit protocol for Path Output Committers.
This committer will work with the FileOutputCommitter and subclasses.
All implementations *must* be serializable.
Rather than ask the FileOutputFormat for a committer, it uses the
org.apache.hadoop.mapreduce.lib.output.PathOutputCommitterFactory factory
API to create the committer.
In setupCommitter the factory is identified and instantiated;
this factory then creates the actual committer implementation.
Dynamic Partition support will be determined once the committer is
instantiated in the setupJob/setupTask methods. If this
class was instantiated with dynamicPartitionOverwrite set to true,
then the instantiated committer must either be an instance of
FileOutputCommitter or it must implement the StreamCapabilities
interface and declare that it has the capability
mapreduce.job.committer.dynamic.partitioning.
That feature is available on Hadoop releases with the Intermediate
Manifest Committer for GCS and ABFS; it is not supported by the
S3A committers.
- Alphabetic
- By Inheritance
- PathOutputCommitProtocol
- HadoopMapReduceCommitProtocol
- Serializable
- FileCommitProtocol
- Logging
- AnyRef
- Any
- Hide All
- Show All
- Public
- Protected
Instance Constructors
- new PathOutputCommitProtocol(jobId: String, dest: String, dynamicPartitionOverwrite: Boolean = false)
Instantiate.
Instantiate.
- jobId
job
- dest
destination
- dynamicPartitionOverwrite
does the caller want support for dynamic partition overwrite?
Value Members
- final def !=(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
- final def ##: Int
- Definition Classes
- AnyRef → Any
- final def ==(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
- def abortJob(jobContext: JobContext): Unit
- Definition Classes
- HadoopMapReduceCommitProtocol → FileCommitProtocol
- def abortTask(taskContext: TaskAttemptContext): Unit
- Definition Classes
- HadoopMapReduceCommitProtocol → FileCommitProtocol
- final def asInstanceOf[T0]: T0
- Definition Classes
- Any
- def clone(): AnyRef
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.CloneNotSupportedException]) @native()
- def commitJob(jobContext: JobContext, taskCommits: Seq[TaskCommitMessage]): Unit
- Definition Classes
- HadoopMapReduceCommitProtocol → FileCommitProtocol
- def commitTask(taskContext: TaskAttemptContext): TaskCommitMessage
- Definition Classes
- HadoopMapReduceCommitProtocol → FileCommitProtocol
- def deleteWithJob(fs: FileSystem, path: Path, recursive: Boolean): Boolean
- Definition Classes
- FileCommitProtocol
- final def eq(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
- def equals(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef → Any
- def finalize(): Unit
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.Throwable])
- final def getClass(): Class[_ <: AnyRef]
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
- def getFilename(taskContext: TaskAttemptContext, spec: FileNameSpec): String
- Attributes
- protected
- Definition Classes
- HadoopMapReduceCommitProtocol
- def hashCode(): Int
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
- def initializeLogIfNecessary(isInterpreter: Boolean, silent: Boolean): Boolean
- Attributes
- protected
- Definition Classes
- Logging
- def initializeLogIfNecessary(isInterpreter: Boolean): Unit
- Attributes
- protected
- Definition Classes
- Logging
- final def isInstanceOf[T0]: Boolean
- Definition Classes
- Any
- def isTraceEnabled(): Boolean
- Attributes
- protected
- Definition Classes
- Logging
- def log: Logger
- Attributes
- protected
- Definition Classes
- Logging
- def logDebug(msg: => String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logDebug(msg: => String): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logError(msg: => String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logError(msg: => String): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logInfo(msg: => String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logInfo(msg: => String): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logName: String
- Attributes
- protected
- Definition Classes
- Logging
- def logTrace(msg: => String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logTrace(msg: => String): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logWarning(msg: => String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logWarning(msg: => String): Unit
- Attributes
- protected
- Definition Classes
- Logging
- final def ne(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
- def newTaskTempFile(taskContext: TaskAttemptContext, dir: Option[String], spec: FileNameSpec): String
Create a temporary file for a task.
Create a temporary file for a task.
- taskContext
task context
- dir
optional subdirectory
- spec
file naming specification
- returns
a path as a string
- Definition Classes
- PathOutputCommitProtocol → HadoopMapReduceCommitProtocol → FileCommitProtocol
- def newTaskTempFile(taskContext: TaskAttemptContext, dir: Option[String], ext: String): String
- Definition Classes
- HadoopMapReduceCommitProtocol → FileCommitProtocol
- def newTaskTempFileAbsPath(taskContext: TaskAttemptContext, absoluteDir: String, spec: FileNameSpec): String
Reject any requests for an absolute path file on a committer which is not compatible with it.
Reject any requests for an absolute path file on a committer which is not compatible with it.
- taskContext
task context
- absoluteDir
final directory
- spec
output filename
- returns
a path string
- Definition Classes
- PathOutputCommitProtocol → HadoopMapReduceCommitProtocol → FileCommitProtocol
- Exceptions thrown
UnsupportedOperationExceptionif incompatible
- def newTaskTempFileAbsPath(taskContext: TaskAttemptContext, absoluteDir: String, ext: String): String
- Definition Classes
- HadoopMapReduceCommitProtocol → FileCommitProtocol
- final def notify(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
- final def notifyAll(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
- def onTaskCommit(taskCommit: TaskCommitMessage): Unit
- Definition Classes
- FileCommitProtocol
- def setupCommitter(context: TaskAttemptContext): PathOutputCommitter
Set up the committer.
Set up the committer. This creates it by talking directly to the Hadoop factories, instead of the V1
mapred.FileOutputFormatmethods.- context
task attempt
- returns
the committer to use. This will always be a subclass of
PathOutputCommitter.
- Attributes
- protected
- Definition Classes
- PathOutputCommitProtocol → HadoopMapReduceCommitProtocol
- def setupJob(jobContext: JobContext): Unit
- Definition Classes
- HadoopMapReduceCommitProtocol → FileCommitProtocol
- def setupTask(taskContext: TaskAttemptContext): Unit
- Definition Classes
- HadoopMapReduceCommitProtocol → FileCommitProtocol
- lazy val stagingDir: Path
- Attributes
- protected
- Definition Classes
- HadoopMapReduceCommitProtocol
- Annotations
- @transient()
- final def synchronized[T0](arg0: => T0): T0
- Definition Classes
- AnyRef
- def toString(): String
- Definition Classes
- AnyRef → Any
- final def wait(): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.InterruptedException])
- final def wait(arg0: Long, arg1: Int): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.InterruptedException])
- final def wait(arg0: Long): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.InterruptedException]) @native()