com.alpine.plugin.core.spark.utils

SparkSchemaUtils

object SparkSchemaUtils extends SparkSchemaUtils with Product with Serializable

Linear Supertypes
Serializable, Serializable, Product, Equals, SparkSchemaUtils, AnyRef, Any
Ordering
  1. Alphabetic
  2. By inheritance
Inherited
  1. SparkSchemaUtils
  2. Serializable
  3. Serializable
  4. Product
  5. Equals
  6. SparkSchemaUtils
  7. AnyRef
  8. Any
  1. Hide All
  2. Show all
Learn more about member selection
Visibility
  1. Public
  2. All

Value Members

  1. final def !=(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  2. final def !=(arg0: Any): Boolean

    Definition Classes
    Any
  3. final def ##(): Int

    Definition Classes
    AnyRef → Any
  4. final def ==(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  5. final def ==(arg0: Any): Boolean

    Definition Classes
    Any
  6. final def asInstanceOf[T0]: T0

    Definition Classes
    Any
  7. def clone(): AnyRef

    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  8. def convertColumnTypeToSparkSQLDataType(columnType: TypeValue, keepDatesAsStrings: Boolean): (DataType, Option[String])

    Convert from Alpine's 'ColumnType' to the corresponding Spark Sql Typ.

    Convert from Alpine's 'ColumnType' to the corresponding Spark Sql Typ. DateTime Behavior: Converts all DateTime columns to TimeStampType. If there is a format string will add that to the metadata. If there is no format string, will use the ISO format ("yyyy-mm-dd hh:mm:ss")

    Definition Classes
    SparkSchemaUtils
  9. def convertSparkSQLDataTypeToColumnType(structField: StructField): ColumnDef

    Converts from a Spark SQL Structfield to an Alpine-specific ColumnDef.

    Converts from a Spark SQL Structfield to an Alpine-specific ColumnDef.

    Definition Classes
    SparkSchemaUtils
  10. def convertSparkSQLSchemaToTabularSchema(schema: StructType): TabularSchema

    Converts from a Spark SQL schema to the Alpine 'TabularSchema' type.

    Converts from a Spark SQL schema to the Alpine 'TabularSchema' type. The 'TabularSchema' object this method returns can be used to create any of the tabular Alpine IO types (HDFSTabular dataset, dataTable etc.)

    Date format behavior: If the column def has not metadata stored at the DATE_METADATA_KEY constant, it wll convert DateType objects to ColumnType(DateTime, "yyyy-mm-dd") and TimeStampType objects to ColumnType(DateTime, "yyyy-mm-dd hh:mm:ss") otherwise will create a column type of ColumnType(DateTime, custom_date_format) where custom_date_format is whatever date format was specified by the column metadata.

    schema

    -a Spark SQL DataFrame schema

    returns

    the equivalent Alpine schema for that dataset

    Definition Classes
    SparkSchemaUtils
  11. def convertTabularSchemaToSparkSQLSchema(tabularSchema: TabularSchema, keepDatesAsStrings: Boolean): StructType

    Definition Classes
    SparkSchemaUtils
  12. def convertTabularSchemaToSparkSQLSchema(tabularSchema: TabularSchema): StructType

    Convert the Alpine 'TabularSchema' with column names and types to the equivalent Spark SQL data frame header.

    Convert the Alpine 'TabularSchema' with column names and types to the equivalent Spark SQL data frame header.

    Date/Time behavior: The same as convertTabularSchemaToSparkSQLSchema(tabularSchema, false). Will NOT convert special date formats to String. Instead will render Alpine date formats as Spark SQL TimeStampType. The original date format will be stored as metadata in the StructFiled object for that column definition.

    tabularSchema

    An Alpine 'TabularSchemaOutline' object with fixed column definitions containing a name and Alpine specific type.

    returns

    Definition Classes
    SparkSchemaUtils
  13. final def eq(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  14. def equals(arg0: Any): Boolean

    Definition Classes
    AnyRef → Any
  15. def finalize(): Unit

    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  16. final def getClass(): Class[_]

    Definition Classes
    AnyRef → Any
  17. def getDateMap(tabularSchema: TabularSchema): Map[String, String]

    Definition Classes
    SparkSchemaUtils
  18. final def isInstanceOf[T0]: Boolean

    Definition Classes
    Any
  19. final def ne(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  20. final def notify(): Unit

    Definition Classes
    AnyRef
  21. final def notifyAll(): Unit

    Definition Classes
    AnyRef
  22. final def synchronized[T0](arg0: ⇒ T0): T0

    Definition Classes
    AnyRef
  23. def toStructField(columnDef: ColumnDef, nullable: Boolean = true): StructField

    Definition Classes
    SparkSchemaUtils
  24. final def wait(): Unit

    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  25. final def wait(arg0: Long, arg1: Int): Unit

    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  26. final def wait(arg0: Long): Unit

    Definition Classes
    AnyRef
    Annotations
    @throws( ... )

Deprecated Value Members

  1. def convertColumnTypeToSparkSQLDataType(columnType: TypeValue): DataType

    Definition Classes
    SparkSchemaUtils
    Annotations
    @deprecated
    Deprecated

    Will not properly handle data formats. Use toStructField

  2. def convertSparkSQLDataTypeToColumnType(dataType: DataType): TypeValue

    Converts from a Spark SQL data type to an Alpine-specific ColumnType

    Converts from a Spark SQL data type to an Alpine-specific ColumnType

    Definition Classes
    SparkSchemaUtils
    Annotations
    @deprecated
    Deprecated

    This doesn't properly handle date formats. Use convertColumnTypeToSparkSQLDataType instead

Inherited from Serializable

Inherited from Serializable

Inherited from Product

Inherited from Equals

Inherited from SparkSchemaUtils

Inherited from AnyRef

Inherited from Any

Ungrouped