Packages

case class DatasetConf(id: String, storageid: String, path: String, format: Format, loadtype: LoadType, table: Option[TableConf] = None, keys: List[String] = List(), partitionby: List[String] = List(), readoptions: Map[String, String] = Map(), writeoptions: Map[String, String] = WriteOptions.DEFAULT_OPTIONS, documentationpath: Option[String] = None, view: Option[TableConf] = None) extends Product with Serializable

Abstraction on a dataset configuration

storageid

an alias designating where the data is sitting. this can point to an object store url in the configuration like s3://my-bucket/

path

the relative path from the root of the storage to the dataset. ie, /raw/my-system/my-source

format

data format

loadtype

how the data is written

table

OPTIONAL - configuration of a table associated to the dataset

readoptions

OPTIONAL - read options to pass to spark in order to read the data into a DataFrame

writeoptions

OPTIONAL - write options to pass to spark in order to write the data into files

documentationpath

OPTIONAL - where the documentation is located.

view

OPTIONAL - schema of the view pointing to the concrete table

Linear Supertypes
Serializable, Serializable, Product, Equals, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. DatasetConf
  2. Serializable
  3. Serializable
  4. Product
  5. Equals
  6. AnyRef
  7. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Instance Constructors

  1. new DatasetConf(id: String, storageid: String, path: String, format: Format, loadtype: LoadType, table: Option[TableConf] = None, keys: List[String] = List(), partitionby: List[String] = List(), readoptions: Map[String, String] = Map(), writeoptions: Map[String, String] = WriteOptions.DEFAULT_OPTIONS, documentationpath: Option[String] = None, view: Option[TableConf] = None)

    storageid

    an alias designating where the data is sitting. this can point to an object store url in the configuration like s3://my-bucket/

    path

    the relative path from the root of the storage to the dataset. ie, /raw/my-system/my-source

    format

    data format

    loadtype

    how the data is written

    table

    OPTIONAL - configuration of a table associated to the dataset

    readoptions

    OPTIONAL - read options to pass to spark in order to read the data into a DataFrame

    writeoptions

    OPTIONAL - write options to pass to spark in order to write the data into files

    documentationpath

    OPTIONAL - where the documentation is located.

    view

    OPTIONAL - schema of the view pointing to the concrete table

Value Members

  1. final def !=(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  4. final def asInstanceOf[T0]: T0
    Definition Classes
    Any
  5. def buid: String

    A dataset BUID is the column representing the business identifier.

    A dataset BUID is the column representing the business identifier. If the dataset defines a table the buid column name is the table name followed by '_buid'.

  6. def clone(): AnyRef
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... ) @native() @HotSpotIntrinsicCandidate()
  7. val documentationpath: Option[String]
  8. final def eq(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  9. val format: Format
  10. final def getClass(): Class[_]
    Definition Classes
    AnyRef → Any
    Annotations
    @native() @HotSpotIntrinsicCandidate()
  11. val id: String
  12. final def isInstanceOf[T0]: Boolean
    Definition Classes
    Any
  13. val keys: List[String]
  14. val loadtype: LoadType
  15. def location(implicit config: Configuration): String

    the absolute path where the dataset is stored

    the absolute path where the dataset is stored

    config

    configuration currently loaded

  16. final def ne(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  17. final def notify(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native() @HotSpotIntrinsicCandidate()
  18. final def notifyAll(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native() @HotSpotIntrinsicCandidate()
  19. def oid: String

    A dataset OID is the column representing the object identifier, in other words: it is a hash representing all the column expect the BUID.

    A dataset OID is the column representing the object identifier, in other words: it is a hash representing all the column expect the BUID. If the dataset defines a table the oid column name is the table name followed by '_oid'.

  20. val partitionby: List[String]
  21. val path: String
  22. def read(implicit config: Configuration, spark: SparkSession): DataFrame

    Using an instance of Spark and the current configuration, reads the dataset from either the tableName or from the location.

    Using an instance of Spark and the current configuration, reads the dataset from either the tableName or from the location.

    config

    configuration currently loaded

    spark

    instance of SparkSession

  23. val readoptions: Map[String, String]
  24. def rootPath(implicit config: Configuration): String

    The absolute path of the root of the storage where the dataset is stored.

    The absolute path of the root of the storage where the dataset is stored.

    config

    configuration currently loaded

  25. val storageid: String
  26. final def synchronized[T0](arg0: ⇒ T0): T0
    Definition Classes
    AnyRef
  27. val table: Option[TableConf]
  28. def uid: String

    A dataset UID is the column representing the unique identifier.

    A dataset UID is the column representing the unique identifier. If the dataset defines a table the uid column name is the table name followed by '_uid'. In most cases the uid is the hash of the keys.

  29. val view: Option[TableConf]
  30. final def wait(arg0: Long, arg1: Int): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  31. final def wait(arg0: Long): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... ) @native()
  32. final def wait(): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  33. val writeoptions: Map[String, String]

Deprecated Value Members

  1. def finalize(): Unit
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] ) @Deprecated
    Deprecated

Inherited from Serializable

Inherited from Serializable

Inherited from Product

Inherited from Equals

Inherited from AnyRef

Inherited from Any

Ungrouped