class DefaultSmartDataLakeBuilder extends SmartDataLakeBuilder
Default Smart Data Lake Command Line Application.
Implementation Note: This must be a class and not an object in order to be found by reflection in DatabricksSmartDataLakeBuilder
- Annotations
- @Scaladoc()
- Alphabetic
- By Inheritance
- DefaultSmartDataLakeBuilder
- SmartDataLakeBuilder
- SmartDataLakeLogger
- AnyRef
- Any
- Hide All
- Show All
- Public
- All
Instance Constructors
- new DefaultSmartDataLakeBuilder()
Value Members
-
final
def
!=(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
final
def
##(): Int
- Definition Classes
- AnyRef → Any
-
final
def
==(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
val
appType: String
- Definition Classes
- SmartDataLakeBuilder
-
val
appVersion: String
- Definition Classes
- SmartDataLakeBuilder
-
final
def
asInstanceOf[T0]: T0
- Definition Classes
- Any
-
def
clone(): AnyRef
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws( ... ) @native() @HotSpotIntrinsicCandidate()
-
final
def
eq(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
-
def
equals(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
final
def
execActionDAG(actionDAGRun: ActionDAGRun, actionsSelected: Seq[Action], context: ActionPipelineContext, lastStartTime: Option[LocalDateTime] = None): Seq[SubFeed]
Execute one action DAG iteration and call recursion if streaming mode Must be implemented with tail recursion to avoid stack overflow error for long running streaming jobs.
Execute one action DAG iteration and call recursion if streaming mode Must be implemented with tail recursion to avoid stack overflow error for long running streaming jobs.
- Definition Classes
- SmartDataLakeBuilder
- Annotations
- @Scaladoc() @tailrec()
-
final
def
getClass(): Class[_]
- Definition Classes
- AnyRef → Any
- Annotations
- @native() @HotSpotIntrinsicCandidate()
-
def
hashCode(): Int
- Definition Classes
- AnyRef → Any
- Annotations
- @native() @HotSpotIntrinsicCandidate()
-
def
initConfigFromEnvironment: SmartDataLakeBuilderConfig
Create a new SDL configuration.
Create a new SDL configuration.
Could be used in the future to set default values.
- returns
a new, initialized SmartDataLakeBuilderConfig.
- Definition Classes
- SmartDataLakeBuilder
- Annotations
- @Scaladoc()
-
val
instanceRegistry: InstanceRegistry
InstanceRegistry instance
InstanceRegistry instance
- Definition Classes
- SmartDataLakeBuilder
-
final
def
isInstanceOf[T0]: Boolean
- Definition Classes
- Any
-
lazy val
logger: Logger
- Attributes
- protected
- Definition Classes
- SmartDataLakeLogger
- Annotations
- @transient()
-
final
def
ne(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
-
final
def
notify(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native() @HotSpotIntrinsicCandidate()
-
final
def
notifyAll(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native() @HotSpotIntrinsicCandidate()
- def parseAndRun(args: Array[String], ignoreOverrideJars: Boolean = false): Unit
-
def
parseCommandLineArguments(args: Array[String], config: SmartDataLakeBuilderConfig): Option[SmartDataLakeBuilderConfig]
Parses the supplied (command line) arguments.
Parses the supplied (command line) arguments.
This method parses command line arguments and creates the corresponding SmartDataLakeBuilderConfig
- args
an Array of command line arguments.
- config
a configuration initialized with default values.
- returns
a new configuration with default values overwritten from the supplied command line arguments.
- Definition Classes
- SmartDataLakeBuilder
- Annotations
- @Scaladoc()
-
val
parser: OptionParser[SmartDataLakeBuilderConfig]
The Parser defines how to extract the options from the command line args.
The Parser defines how to extract the options from the command line args. Subclasses SmartDataLakeBuilder can define additional options to be extracted.
- Attributes
- protected
- Definition Classes
- SmartDataLakeBuilder
-
def
run(appConfig: SmartDataLakeBuilderConfig): Map[RuntimeEventState, Int]
Run the application with the provided configuration.
Run the application with the provided configuration.
- appConfig
Application configuration (parsed from command line).
- Definition Classes
- SmartDataLakeBuilder
- Annotations
- @Scaladoc()
-
def
startSimulation(appConfig: SmartDataLakeBuilderConfig, initialSubFeeds: Seq[SparkSubFeed], dataObjectsState: Seq[DataObjectState] = Seq())(implicit instanceRegistry: InstanceRegistry, session: SparkSession): (Seq[SparkSubFeed], Map[RuntimeEventState, Int])
Start a simulation run.
Start a simulation run. This executes the DAG and returns all subfeeds including the transformed DataFrames. Only prepare and init are executed. All initial subfeeds must be provided as input.
Note: this only works with SparkActions for now
- appConfig
application configuration
- initialSubFeeds
initial subfeeds for DataObjects at the beginning of the DAG
- dataObjectsState
state for incremental DataObjects
- returns
tuple of list of final subfeeds and statistics (action count per RuntimeEventState)
- Definition Classes
- SmartDataLakeBuilder
- Annotations
- @Scaladoc()
-
final
def
synchronized[T0](arg0: ⇒ T0): T0
- Definition Classes
- AnyRef
-
def
toString(): String
- Definition Classes
- AnyRef → Any
-
final
def
wait(arg0: Long, arg1: Int): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... )
-
final
def
wait(arg0: Long): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... ) @native()
-
final
def
wait(): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... )
Deprecated Value Members
-
def
finalize(): Unit
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws( classOf[java.lang.Throwable] ) @Deprecated
- Deprecated