diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/Triggers.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/Triggers.scala index f29970d5de192..ebd237bfc0c19 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/Triggers.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/Triggers.scala @@ -50,17 +50,17 @@ private object Triggers { * A [[Trigger]] that processes only one batch of data in a streaming query then terminates * the query. */ -private[sql] case object OneTimeTrigger extends Trigger +case object OneTimeTrigger extends Trigger /** * A [[Trigger]] that runs a query periodically based on the processing time. If `interval` is 0, * the query will run as fast as possible. */ -private[sql] case class ProcessingTimeTrigger(intervalMs: Long) extends Trigger { +case class ProcessingTimeTrigger(intervalMs: Long) extends Trigger { Triggers.validate(intervalMs) } -private[sql] object ProcessingTimeTrigger { +object ProcessingTimeTrigger { import Triggers._ def apply(interval: String): ProcessingTimeTrigger = { @@ -84,11 +84,11 @@ private[sql] object ProcessingTimeTrigger { * A [[Trigger]] that continuously processes streaming data, asynchronously checkpointing at * the specified interval. */ -private[sql] case class ContinuousTrigger(intervalMs: Long) extends Trigger { +case class ContinuousTrigger(intervalMs: Long) extends Trigger { Triggers.validate(intervalMs) } -private[sql] object ContinuousTrigger { +object ContinuousTrigger { import Triggers._ def apply(interval: String): ContinuousTrigger = {