pyspark.sql.streaming.DataStreamWriter.trigger¶
-
DataStreamWriter.
trigger
(processingTime=None, once=None, continuous=None)[source]¶ Set the trigger for the stream query. If this is not set it will run the query as fast as possible, which is equivalent to setting the trigger to
processingTime='0 seconds'
.Note
Evolving.
- Parameters
processingTime – a processing time interval as a string, e.g. ‘5 seconds’, ‘1 minute’. Set a trigger that runs a microbatch query periodically based on the processing time. Only one trigger can be set.
once – if set to True, set a trigger that processes only one batch of data in a streaming query then terminates the query. Only one trigger can be set.
continuous – a time interval as a string, e.g. ‘5 seconds’, ‘1 minute’. Set a trigger that runs a continuous query with a given checkpoint interval. Only one trigger can be set.
>>> # trigger the query for execution every 5 seconds >>> writer = sdf.writeStream.trigger(processingTime='5 seconds') >>> # trigger the query for just once batch of data >>> writer = sdf.writeStream.trigger(once=True) >>> # trigger the query for execution every 5 seconds >>> writer = sdf.writeStream.trigger(continuous='5 seconds')
New in version 2.0.