[SPARK-27649][SS] Unify the way use 'spark.network.timeout'
## What changes were proposed in this pull request? For historical reasons, structured streaming still has some old way of use `spark.network.timeout` , even though `org.apache.spark.internal.config.Network.NETWORK_TIMEOUT` is now available. ## How was this patch tested? Exists UT. Closes #24545 from beliefer/unify-spark-network-timeout. Authored-by: gengjiaan <gengjiaan@360.cn> Signed-off-by: Wenchen Fan <wenchen@databricks.com>
This commit is contained in:
parent
83f628b57d
commit
8329e7debd
|
@ -26,6 +26,7 @@ import org.apache.kafka.clients.consumer.ConsumerConfig
|
|||
|
||||
import org.apache.spark.SparkEnv
|
||||
import org.apache.spark.internal.Logging
|
||||
import org.apache.spark.internal.config.Network.NETWORK_TIMEOUT
|
||||
import org.apache.spark.scheduler.ExecutorCacheTaskLocation
|
||||
import org.apache.spark.sql.SparkSession
|
||||
import org.apache.spark.sql.catalyst.InternalRow
|
||||
|
@ -64,7 +65,7 @@ private[kafka010] class KafkaMicroBatchStream(
|
|||
|
||||
private val pollTimeoutMs = options.getLong(
|
||||
"kafkaConsumer.pollTimeoutMs",
|
||||
SparkEnv.get.conf.getTimeAsSeconds("spark.network.timeout", "120s") * 1000L)
|
||||
SparkEnv.get.conf.get(NETWORK_TIMEOUT) * 1000L)
|
||||
|
||||
private val maxOffsetsPerTrigger = Option(options.get("maxOffsetsPerTrigger")).map(_.toLong)
|
||||
|
||||
|
|
|
@ -22,6 +22,7 @@ import java.util.UUID
|
|||
import org.apache.kafka.common.TopicPartition
|
||||
|
||||
import org.apache.spark.internal.Logging
|
||||
import org.apache.spark.internal.config.Network.NETWORK_TIMEOUT
|
||||
import org.apache.spark.rdd.RDD
|
||||
import org.apache.spark.sql.{Row, SQLContext}
|
||||
import org.apache.spark.sql.catalyst.InternalRow
|
||||
|
@ -47,9 +48,7 @@ private[kafka010] class KafkaRelation(
|
|||
|
||||
private val pollTimeoutMs = sourceOptions.getOrElse(
|
||||
"kafkaConsumer.pollTimeoutMs",
|
||||
(sqlContext.sparkContext.conf.getTimeAsSeconds(
|
||||
"spark.network.timeout",
|
||||
"120s") * 1000L).toString
|
||||
(sqlContext.sparkContext.conf.get(NETWORK_TIMEOUT) * 1000L).toString
|
||||
).toLong
|
||||
|
||||
override def schema: StructType = KafkaOffsetReader.kafkaSchema
|
||||
|
|
|
@ -27,6 +27,7 @@ import org.apache.kafka.common.TopicPartition
|
|||
|
||||
import org.apache.spark.SparkContext
|
||||
import org.apache.spark.internal.Logging
|
||||
import org.apache.spark.internal.config.Network.NETWORK_TIMEOUT
|
||||
import org.apache.spark.scheduler.ExecutorCacheTaskLocation
|
||||
import org.apache.spark.sql._
|
||||
import org.apache.spark.sql.catalyst.InternalRow
|
||||
|
@ -85,7 +86,7 @@ private[kafka010] class KafkaSource(
|
|||
|
||||
private val pollTimeoutMs = sourceOptions.getOrElse(
|
||||
"kafkaConsumer.pollTimeoutMs",
|
||||
(sc.conf.getTimeAsSeconds("spark.network.timeout", "120s") * 1000L).toString
|
||||
(sc.conf.get(NETWORK_TIMEOUT) * 1000L).toString
|
||||
).toLong
|
||||
|
||||
private val maxOffsetsPerTrigger =
|
||||
|
|
Loading…
Reference in a new issue