The following list are the properties that you can use to fine-tune Spark Structured Streaming applications.
You can set them in a SparkSession upon instantiation using config method.
import org.apache.spark.sql.SparkSession
val spark: SparkSession = SparkSession.builder
.master("local[*]")
.appName("My Spark Application")
.config("spark.sql.streaming.metricsEnabled", true)
.getOrCreate
Name | Default | Description |
---|---|---|
(empty) |
Default checkpoint directory for storing checkpoint data for streaming queries |
|
|
Flag whether Dropwizard CodaHale metrics will be reported for active streaming queries |
|
|
(internal) The minimum number of batches that must be retained and made recoverable. Used…FIXME |
|
|
Number of progress updates to retain for a streaming query |
|
|
(internal) Time delay before |
|
|
The initial delay and how often to execute StateStore’s maintenance task. |
|
org.apache.spark.sql.execution.streaming.state.HDFSBackedStateStoreProvider |
(internal) The fully-qualified class name to manage state data in stateful streaming queries. This class must be a subclass of StateStoreProvider, and must have a zero-arg constructor. |
|
|
(internal) When enabled (i.e. |