2.0.6

Spark

Package: flyteplugins.spark

Use this to configure a SparkContext for a your task. Task’s marked with this will automatically execute natively onto K8s as a distributed execution of spark

Attributes: spark_conf (Optional[Dict[str, str]]): Spark configuration dictionary. hadoop_conf (Optional[Dict[str, str]]): Hadoop configuration dictionary. executor_path (Optional[str]): Path to the Python binary for PySpark execution. applications_path (Optional[str]): Path to the main application file. driver_pod (Optional[PodTemplate]): Pod template for the driver pod. executor_pod (Optional[PodTemplate]): Pod template for the executor pods.

class Spark(
    spark_conf: typing.Optional[typing.Dict[str, str]],
    hadoop_conf: typing.Optional[typing.Dict[str, str]],
    executor_path: typing.Optional[str],
    applications_path: typing.Optional[str],
    driver_pod: typing.Optional[flyte._pod.PodTemplate],
    executor_pod: typing.Optional[flyte._pod.PodTemplate],
)
Parameter Type Description
spark_conf typing.Optional[typing.Dict[str, str]]
hadoop_conf typing.Optional[typing.Dict[str, str]]
executor_path typing.Optional[str]
applications_path typing.Optional[str]
driver_pod typing.Optional[flyte._pod.PodTemplate]
executor_pod typing.Optional[flyte._pod.PodTemplate]