PySparkBatchArgs

data class PySparkBatchArgs(val archiveUris: Output<List<String>>? = null, val args: Output<List<String>>? = null, val fileUris: Output<List<String>>? = null, val jarFileUris: Output<List<String>>? = null, val mainPythonFileUri: Output<String>, val pythonFileUris: Output<List<String>>? = null) : ConvertibleToJava<PySparkBatchArgs>

A configuration for running an Apache PySpark (https://spark.apache.org/docs/latest/api/python/getting_started/quickstart.html) batch workload.

Constructors

Link copied to clipboard
fun PySparkBatchArgs(archiveUris: Output<List<String>>? = null, args: Output<List<String>>? = null, fileUris: Output<List<String>>? = null, jarFileUris: Output<List<String>>? = null, mainPythonFileUri: Output<String>, pythonFileUris: Output<List<String>>? = null)

Functions

Link copied to clipboard
open override fun toJava(): PySparkBatchArgs

Properties

Link copied to clipboard
val archiveUris: Output<List<String>>? = null

Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.

Link copied to clipboard
val args: Output<List<String>>? = null

Optional. The arguments to pass to the driver. Do not include arguments that can be set as batch properties, such as --conf, since a collision can occur that causes an incorrect batch submission.

Link copied to clipboard
val fileUris: Output<List<String>>? = null

Optional. HCFS URIs of files to be placed in the working directory of each executor.

Link copied to clipboard
val jarFileUris: Output<List<String>>? = null

Optional. HCFS URIs of jar files to add to the classpath of the Spark driver and tasks.

Link copied to clipboard

The HCFS URI of the main Python file to use as the Spark driver. Must be a .py file.

Link copied to clipboard
val pythonFileUris: Output<List<String>>? = null

Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.