Control the GUI of your Spark job, through this you can specify any visualization for the output of your job, and what params the user will need to specify.
A class for plugins which will use Schema inference
Control the GUI of your Spark job, through this you can specify any visualization for the output of your job, and what params the user will need to specify.
Job base for non-inferred Spark plugin jobs taking and returning DataFrames.
Job base for non-inferred Spark plugin jobs taking and returning DataFrames. Note: This WILL NOT work with hive.
A class controlling the runtime behavior of your plugin.
A class controlling the runtime behavior of your plugin. To use the default implementation, which launches a Spark job according to the default Spark settings you will not need to add any code beyond the class definition with the appropriate type parameters.
your implementation of SparkDataFrameJob
Control the GUI of your Spark job, through this you can specify any visualization for the output of your job, and what params the user will need to specify.
Templated base for Spark plugin jobs operating on DataFrames.
Templated base for Spark plugin jobs operating on DataFrames. Most jobs will want to use SparkDataFrameJob which takes and returns Spark DataFrames. This version does not support schema inference.
The return type of the transformation method (most commonly a DataFrame)
The return type of the actual operator, extending IOBase. Most commonly will be an HDFS dataset of some flavor (see SparkDataFrame)
Control the GUI of your Spark job, through this you can specify any visualization for the output of your job, and what params the user will need to specify. Uses the provided operator to generate an updated schema, this should work for most operators but if not (e.g. your operator doesn't handle empty data or output schema depends on input data) then you will have to perform your own schema update.