org.apache.spark.internal.io.cloud
PathOutputCommitProtocol
Companion object PathOutputCommitProtocol
class PathOutputCommitProtocol extends HadoopMapReduceCommitProtocol with Serializable
Spark Commit protocol for Path Output Committers.
This committer will work with the FileOutputCommitter and subclasses.
All implementations *must* be serializable.
Rather than ask the FileOutputFormat for a committer, it uses the
org.apache.hadoop.mapreduce.lib.output.PathOutputCommitterFactory factory
API to create the committer.
In setupCommitter the factory is identified and instantiated;
this factory then creates the actual committer implementation.
Dynamic Partition support will be determined once the committer is
instantiated in the setupJob/setupTask methods. If this
class was instantiated with dynamicPartitionOverwrite set to true,
then the instantiated committer must either be an instance of
FileOutputCommitter or it must implement the StreamCapabilities
interface and declare that it has the capability
mapreduce.job.committer.dynamic.partitioning.
That feature is available on Hadoop releases with the Intermediate
Manifest Committer for GCS and ABFS; it is not supported by the
S3A committers.
- Alphabetic
- By Inheritance
- PathOutputCommitProtocol
- HadoopMapReduceCommitProtocol
- Serializable
- Serializable
- FileCommitProtocol
- Logging
- AnyRef
- Any
- Hide All
- Show All
- Public
- All
Instance Constructors
-
new
PathOutputCommitProtocol(jobId: String, dest: String, dynamicPartitionOverwrite: Boolean = false)
Instantiate.
Instantiate.
- jobId
job
- dest
destination
- dynamicPartitionOverwrite
does the caller want support for dynamic partition overwrite?
Value Members
-
final
def
!=(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
final
def
##(): Int
- Definition Classes
- AnyRef → Any
-
final
def
==(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
def
abortJob(jobContext: JobContext): Unit
- Definition Classes
- HadoopMapReduceCommitProtocol → FileCommitProtocol
-
def
abortTask(taskContext: TaskAttemptContext): Unit
- Definition Classes
- HadoopMapReduceCommitProtocol → FileCommitProtocol
-
final
def
asInstanceOf[T0]: T0
- Definition Classes
- Any
-
def
clone(): AnyRef
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws( ... ) @native()
-
def
commitJob(jobContext: JobContext, taskCommits: Seq[TaskCommitMessage]): Unit
- Definition Classes
- HadoopMapReduceCommitProtocol → FileCommitProtocol
-
def
commitTask(taskContext: TaskAttemptContext): TaskCommitMessage
- Definition Classes
- HadoopMapReduceCommitProtocol → FileCommitProtocol
-
def
deleteWithJob(fs: FileSystem, path: Path, recursive: Boolean): Boolean
- Definition Classes
- FileCommitProtocol
-
final
def
eq(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
-
def
equals(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
def
finalize(): Unit
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws( classOf[java.lang.Throwable] )
-
final
def
getClass(): Class[_]
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
-
def
getFilename(taskContext: TaskAttemptContext, spec: FileNameSpec): String
- Attributes
- protected
- Definition Classes
- HadoopMapReduceCommitProtocol
-
def
hashCode(): Int
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
-
def
initializeLogIfNecessary(isInterpreter: Boolean, silent: Boolean): Boolean
- Attributes
- protected
- Definition Classes
- Logging
-
def
initializeLogIfNecessary(isInterpreter: Boolean): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
final
def
isInstanceOf[T0]: Boolean
- Definition Classes
- Any
-
def
isTraceEnabled(): Boolean
- Attributes
- protected
- Definition Classes
- Logging
-
def
log: Logger
- Attributes
- protected
- Definition Classes
- Logging
-
def
logDebug(msg: ⇒ String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logDebug(msg: ⇒ String): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logError(msg: ⇒ String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logError(msg: ⇒ String): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logInfo(msg: ⇒ String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logInfo(msg: ⇒ String): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logName: String
- Attributes
- protected
- Definition Classes
- Logging
-
def
logTrace(msg: ⇒ String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logTrace(msg: ⇒ String): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logWarning(msg: ⇒ String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
def
logWarning(msg: ⇒ String): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
final
def
ne(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
-
def
newTaskTempFile(taskContext: TaskAttemptContext, dir: Option[String], spec: FileNameSpec): String
Create a temporary file for a task.
Create a temporary file for a task.
- taskContext
task context
- dir
optional subdirectory
- spec
file naming specification
- returns
a path as a string
- Definition Classes
- PathOutputCommitProtocol → HadoopMapReduceCommitProtocol → FileCommitProtocol
-
def
newTaskTempFile(taskContext: TaskAttemptContext, dir: Option[String], ext: String): String
- Definition Classes
- HadoopMapReduceCommitProtocol → FileCommitProtocol
-
def
newTaskTempFileAbsPath(taskContext: TaskAttemptContext, absoluteDir: String, spec: FileNameSpec): String
Reject any requests for an absolute path file on a committer which is not compatible with it.
Reject any requests for an absolute path file on a committer which is not compatible with it.
- taskContext
task context
- absoluteDir
final directory
- spec
output filename
- returns
a path string
- Definition Classes
- PathOutputCommitProtocol → HadoopMapReduceCommitProtocol → FileCommitProtocol
- Exceptions thrown
UnsupportedOperationExceptionif incompatible
-
def
newTaskTempFileAbsPath(taskContext: TaskAttemptContext, absoluteDir: String, ext: String): String
- Definition Classes
- HadoopMapReduceCommitProtocol → FileCommitProtocol
-
final
def
notify(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
-
final
def
notifyAll(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
-
def
onTaskCommit(taskCommit: TaskCommitMessage): Unit
- Definition Classes
- FileCommitProtocol
-
def
setupCommitter(context: TaskAttemptContext): PathOutputCommitter
Set up the committer.
Set up the committer. This creates it by talking directly to the Hadoop factories, instead of the V1
mapred.FileOutputFormatmethods.- context
task attempt
- returns
the committer to use. This will always be a subclass of
PathOutputCommitter.
- Attributes
- protected
- Definition Classes
- PathOutputCommitProtocol → HadoopMapReduceCommitProtocol
-
def
setupJob(jobContext: JobContext): Unit
- Definition Classes
- HadoopMapReduceCommitProtocol → FileCommitProtocol
-
def
setupTask(taskContext: TaskAttemptContext): Unit
- Definition Classes
- HadoopMapReduceCommitProtocol → FileCommitProtocol
-
lazy val
stagingDir: Path
- Attributes
- protected
- Definition Classes
- HadoopMapReduceCommitProtocol
- Annotations
- @transient()
-
final
def
synchronized[T0](arg0: ⇒ T0): T0
- Definition Classes
- AnyRef
-
def
toString(): String
- Definition Classes
- AnyRef → Any
-
final
def
wait(): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... )
-
final
def
wait(arg0: Long, arg1: Int): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... )
-
final
def
wait(arg0: Long): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... ) @native()