SQLConf¶
SQLConf is an internal configuration store of the configuration properties and hints used in Spark SQL.
Important
SQLConf is an internal part of Spark SQL and is not supposed to be used directly. Spark SQL configuration is available through the developer-facing RuntimeConfig.
SQLConf offers methods to get, set, unset or clear values of the configuration properties and hints as well as to read the current values.
Accessing SQLConf¶
You can access a SQLConf using:
-
SQLConf.get(preferred) - theSQLConfof the current activeSparkSession -
SessionState - direct access through SessionState of the
SparkSessionof your choice (that gives more flexibility on whatSparkSessionis used that can be different from the current activeSparkSession)
import org.apache.spark.sql.internal.SQLConf
// Use type-safe access to configuration properties
// using SQLConf.get.getConf
val parallelFileListingInStatsComputation = SQLConf.get.getConf(SQLConf.PARALLEL_FILE_LISTING_IN_STATS_COMPUTATION)
// or even simpler
SQLConf.get.parallelFileListingInStatsComputation
scala> :type spark
org.apache.spark.sql.SparkSession
// Direct access to the session SQLConf
val sqlConf = spark.sessionState.conf
scala> :type sqlConf
org.apache.spark.sql.internal.SQLConf
scala> println(sqlConf.offHeapColumnVectorEnabled)
false
// Or simply import the conf value
import spark.sessionState.conf
// accessing properties through accessor methods
scala> conf.numShufflePartitions
res1: Int = 200
// Prefer SQLConf.get (over direct access)
import org.apache.spark.sql.internal.SQLConf
val cc = SQLConf.get
scala> cc == conf
res4: Boolean = true
// setting properties using aliases
import org.apache.spark.sql.internal.SQLConf.SHUFFLE_PARTITIONS
conf.setConf(SHUFFLE_PARTITIONS, 2)
scala> conf.numShufflePartitions
res2: Int = 2
// unset aka reset properties to the default value
conf.unsetConf(SHUFFLE_PARTITIONS)
scala> conf.numShufflePartitions
res3: Int = 200
ADAPTIVE_AUTO_BROADCASTJOIN_THRESHOLD¶
spark.sql.adaptive.autoBroadcastJoinThreshold
Used when:
JoinSelectionHelperis requested to canBroadcastBySize
ADAPTIVE_EXECUTION_FORCE_APPLY¶
spark.sql.adaptive.forceApply configuration property
Used when:
- InsertAdaptiveSparkPlan physical optimization is executed
adaptiveExecutionEnabled¶
The value of spark.sql.adaptive.enabled configuration property
Used when:
- InsertAdaptiveSparkPlan physical optimization is executed
SQLConfis requested for the numShufflePartitions
adaptiveExecutionLogLevel¶
The value of spark.sql.adaptive.logLevel configuration property
Used when AdaptiveSparkPlanExec physical operator is executed
ADAPTIVE_MAX_SHUFFLE_HASH_JOIN_LOCAL_MAP_THRESHOLD¶
spark.sql.adaptive.maxShuffledHashJoinLocalMapThreshold configuration property
Used when:
DynamicJoinSelectionis requested to preferShuffledHashJoin
ADAPTIVE_OPTIMIZER_EXCLUDED_RULES¶
spark.sql.adaptive.optimizer.excludedRules
ADVISORY_PARTITION_SIZE_IN_BYTES¶
spark.sql.adaptive.advisoryPartitionSizeInBytes configuration property
Used when:
- CoalesceShufflePartitions and OptimizeSkewedJoin physical optimizations are executed
autoBroadcastJoinThreshold¶
The value of spark.sql.autoBroadcastJoinThreshold configuration property
Used when:
- JoinSelection execution planning strategy is executed
autoBucketedScanEnabled¶
The value of spark.sql.sources.bucketing.autoBucketedScan.enabled configuration property
Used when:
- DisableUnnecessaryBucketedScan physical optimization is executed
allowStarWithSingleTableIdentifierInCount¶
spark.sql.legacy.allowStarWithSingleTableIdentifierInCount
Used when:
ResolveReferenceslogical resolution rule is executed
arrowPySparkSelfDestructEnabled¶
spark.sql.execution.arrow.pyspark.selfDestruct.enabled
Used when:
PandasConversionMixinis requested totoPandas
allowAutoGeneratedAliasForView¶
spark.sql.legacy.allowAutoGeneratedAliasForView
Used when:
ViewHelperutility is used toverifyAutoGeneratedAliasesNotExists
allowNonEmptyLocationInCTAS¶
spark.sql.legacy.allowNonEmptyLocationInCTAS
Used when:
DataWritingCommandutility is used to assertEmptyRootPath
allowNonEmptyLocationInCTAS¶
spark.sql.adaptive.optimizeSkewsInRebalancePartitions.enabled
Used when:
OptimizeSkewInRebalancePartitionsphysical optimization is executed
ADAPTIVE_CUSTOM_COST_EVALUATOR_CLASS¶
spark.sql.adaptive.customCostEvaluatorClass
autoSizeUpdateEnabled¶
The value of spark.sql.statistics.size.autoUpdate.enabled configuration property
Used when:
CommandUtilsis requested for updating existing table statisticsAlterTableAddPartitionCommandlogical command is executed
avroCompressionCodec¶
The value of spark.sql.avro.compression.codec configuration property
Used when AvroOptions is requested for the compression configuration property (and it was not set explicitly)
broadcastTimeout¶
The value of spark.sql.broadcastTimeout configuration property
Used in BroadcastExchangeExec (for broadcasting a table to executors)
bucketingEnabled¶
The value of spark.sql.sources.bucketing.enabled configuration property
Used when FileSourceScanExec physical operator is requested for the input RDD and to determine output partitioning and ordering
cacheVectorizedReaderEnabled¶
The value of spark.sql.inMemoryColumnarStorage.enableVectorizedReader configuration property
Used when InMemoryTableScanExec physical operator is requested for supportsBatch flag.
CAN_CHANGE_CACHED_PLAN_OUTPUT_PARTITIONING¶
spark.sql.optimizer.canChangeCachedPlanOutputPartitioning
Used when:
CacheManageris requested to getOrCloneSessionWithConfigsOff
caseSensitiveAnalysis¶
The value of spark.sql.caseSensitive configuration property
cboEnabled¶
The value of spark.sql.cbo.enabled configuration property
Used in:
- ReorderJoin logical plan optimization (and indirectly in
StarSchemaDetectionforreorderStarJoins) - CostBasedJoinReorder logical plan optimization
cliPrintHeader¶
Used when:
SparkSQLCLIDriveris requested toprocessCmd
coalesceBucketsInJoinEnabled¶
The value of spark.sql.bucketing.coalesceBucketsInJoin.enabled configuration property
Used when:
- CoalesceBucketsInJoin physical optimization is executed
COALESCE_PARTITIONS_MIN_PARTITION_SIZE¶
spark.sql.adaptive.coalescePartitions.minPartitionSize configuration property
Used when:
- CoalesceShufflePartitions physical optimization is executed
COALESCE_PARTITIONS_PARALLELISM_FIRST¶
spark.sql.adaptive.coalescePartitions.parallelismFirst configuration property
Used when:
- CoalesceShufflePartitions physical optimization is executed
coalesceShufflePartitionsEnabled¶
The value of spark.sql.adaptive.coalescePartitions.enabled configuration property
Used when:
- CoalesceShufflePartitions and EnsureRequirements physical optimizations are executed
codegenCacheMaxEntries¶
spark.sql.codegen.cache.maxEntries
columnBatchSize¶
The value of spark.sql.inMemoryColumnarStorage.batchSize configuration property
Used when:
CacheManageris requested to cache a structured queryRowToColumnarExecphysical operator is requested to doExecuteColumnar
constraintPropagationEnabled¶
The value of spark.sql.constraintPropagation.enabled configuration property
Used when:
- InferFiltersFromConstraints logical optimization is executed
QueryPlanConstraintsis requested for the constraints
CONVERT_METASTORE_ORC¶
The value of spark.sql.hive.convertMetastoreOrc configuration property
Used when RelationConversions logical post-hoc evaluation rule is executed (and requested to isConvertible)
CONVERT_METASTORE_PARQUET¶
The value of spark.sql.hive.convertMetastoreParquet configuration property
Used when RelationConversions logical post-hoc evaluation rule is executed (and requested to isConvertible)
csvExpressionOptimization¶
spark.sql.optimizer.enableCsvExpressionOptimization
Used when:
OptimizeCsvJsonExprslogical optimization is executed
dataFramePivotMaxValues¶
The value of spark.sql.pivotMaxValues configuration property
Used in pivot operator.
dataFrameRetainGroupColumns¶
decorrelateInnerQueryEnabled¶
spark.sql.optimizer.decorrelateInnerQuery.enabled
Used when:
CheckAnalysisis requested to checkCorrelationsInSubquery (with a Project unary logical operator)- PullupCorrelatedPredicates logical optimization is executed
DEFAULT_CATALOG¶
The value of spark.sql.defaultCatalog configuration property
Used when CatalogManager is requested for the current CatalogPlugin
defaultDataSourceName¶
defaultSizeInBytes¶
Used when:
DetermineTableStatslogical resolution rule could not compute the table size or spark.sql.statistics.fallBackToHdfs is disabled- ExternalRDD, LogicalRDD and DataSourceV2Relation are requested to compute stats
- (Spark Structured Streaming)
StreamingRelation,StreamingExecutionRelation,StreamingRelationV2andContinuousExecutionRelationare requested for statistics (i.e.computeStats) DataSourcecreates a HadoopFsRelation for FileFormat data source (and builds a CatalogFileIndex when no table statistics are available)BaseRelationis requested for an estimated size of this relation (in bytes)
dynamicPartitionPruningEnabled¶
spark.sql.optimizer.dynamicPartitionPruning.enabled
dynamicPartitionPruningFallbackFilterRatio¶
The value of spark.sql.optimizer.dynamicPartitionPruning.fallbackFilterRatio configuration property
Used when:
- PartitionPruning logical optimization rule is executed
dynamicPartitionPruningPruningSideExtraFilterRatio¶
The value of spark.sql.optimizer.dynamicPartitionPruning.pruningSideExtraFilterRatio configuration property
Used when:
- PartitionPruning logical optimization rule is executed
dynamicPartitionPruningReuseBroadcastOnly¶
spark.sql.optimizer.dynamicPartitionPruning.reuseBroadcastOnly
dynamicPartitionPruningUseStats¶
spark.sql.optimizer.dynamicPartitionPruning.useStats
ENABLE_FULL_OUTER_SHUFFLED_HASH_JOIN_CODEGEN¶
spark.sql.codegen.join.fullOuterShuffledHashJoin.enabled
enableDefaultColumns¶
spark.sql.defaultColumn.enabled
enableRadixSort¶
spark.sql.sort.enableRadixSort
Used when:
SortExecphysical operator is requested to create an UnsafeExternalRowSorter.
enableTwoLevelAggMap¶
spark.sql.codegen.aggregate.map.twolevel.enabled
enableVectorizedHashMap¶
spark.sql.codegen.aggregate.map.vectorized.enable
exchangeReuseEnabled¶
Used when:
-
AdaptiveSparkPlanExec physical operator is requested to createQueryStages
-
PartitionPruning logical optimization rule is executed.
-
PlanDynamicPruningFiltersand ReuseExchange physical optimizations are executed
fallBackToHdfsForStatsEnabled¶
spark.sql.statistics.fallBackToHdfs
Used when DetermineTableStats logical resolution rule is executed.
fastHashAggregateRowMaxCapacityBit¶
spark.sql.codegen.aggregate.fastHashMap.capacityBit
fetchShuffleBlocksInBatch¶
The value of spark.sql.adaptive.fetchShuffleBlocksInBatch configuration property
Used when ShuffledRowRDD is created
fileCommitProtocolClass¶
spark.sql.sources.commitProtocolClass
fileCompressionFactor¶
The value of spark.sql.sources.fileCompressionFactor configuration property
Used when:
HadoopFsRelationis requested for a sizeFileScanis requested to estimate statistics
filesMaxPartitionBytes¶
spark.sql.files.maxPartitionBytes
filesMinPartitionNum¶
spark.sql.files.minPartitionNum
filesOpenCostInBytes¶
spark.sql.files.openCostInBytes
filesourcePartitionFileCacheSize¶
spark.sql.hive.filesourcePartitionFileCacheSize
histogramEnabled¶
The value of spark.sql.statistics.histogram.enabled configuration property
Used when AnalyzeColumnCommand logical command is executed.
histogramNumBins¶
spark.sql.statistics.histogram.numBins
Used when AnalyzeColumnCommand is AnalyzeColumnCommand.md#run[executed] with configuration-properties.md#spark.sql.statistics.histogram.enabled[spark.sql.statistics.histogram.enabled] turned on (and AnalyzeColumnCommand.md#computePercentiles[calculates percentiles]).
HIVE_TABLE_PROPERTY_LENGTH_THRESHOLD¶
spark.sql.hive.tablePropertyLengthThreshold
Used when:
CatalogTableis requested to splitLargeTableProp
hugeMethodLimit¶
spark.sql.codegen.hugeMethodLimit
ignoreCorruptFiles¶
The value of spark.sql.files.ignoreCorruptFiles configuration property
Used when:
AvroUtilsutility is requested toinferSchemaOrcFileFormatis requested toinferSchemaandbuildReaderFileScanRDDis created (and then to compute a partition)SchemaMergeUtilsutility is requested tomergeSchemasInParallelOrcUtilsutility is requested toreadSchemaFilePartitionReaderis requested toignoreCorruptFiles
ignoreMissingFiles¶
The value of spark.sql.files.ignoreMissingFiles configuration property
Used when:
FileScanRDDis created (and then to compute a partition)InMemoryFileIndexutility is requested to bulkListLeafFilesFilePartitionReaderis requested toignoreMissingFiles
inMemoryPartitionPruning¶
spark.sql.inMemoryColumnarStorage.partitionPruning
isParquetBinaryAsString¶
spark.sql.parquet.binaryAsString
isParquetINT96AsTimestamp¶
spark.sql.parquet.int96AsTimestamp
isParquetINT96TimestampConversion¶
spark.sql.parquet.int96TimestampConversion
Used when ParquetFileFormat is requested to build a data reader with partition column values appended.
isParquetSchemaMergingEnabled¶
isParquetSchemaRespectSummaries¶
spark.sql.parquet.respectSummaryFiles
Used when:
ParquetUtilsis used to inferSchema
joinReorderEnabled¶
spark.sql.cbo.joinReorder.enabled
Used in CostBasedJoinReorder logical plan optimization
legacyIntervalEnabled¶
spark.sql.legacy.interval.enabled
Used when:
SubtractTimestampsexpression is createdSubtractDatesexpression is createdAstBuilderis requested to visitTypeConstructor and visitInterval
limitScaleUpFactor¶
Used when a physical operator is requested the first n rows as an array.
LOCAL_SHUFFLE_READER_ENABLED¶
spark.sql.adaptive.localShuffleReader.enabled
Used when:
- OptimizeShuffleWithLocalRead adaptive physical optimization is executed
manageFilesourcePartitions¶
spark.sql.hive.manageFilesourcePartitions
maxConcurrentOutputFileWriters¶
The value of spark.sql.maxConcurrentOutputFileWriters configuration property
Used when:
FileFormatWriteris requested to write out a query result
maxMetadataStringLength¶
spark.sql.maxMetadataStringLength
Used when:
DataSourceScanExecis requested for simpleStringFileScanis requested for description and metadataHiveTableRelationis requested for simpleString
maxRecordsPerFile¶
spark.sql.files.maxRecordsPerFile
Used when:
FileFormatWriterutility is used to write out a query resultFileWriteis requested for a BatchWrite
maxToStringFields¶
The value of spark.sql.debug.maxToStringFields configuration property
metastorePartitionPruning¶
spark.sql.hive.metastorePartitionPruning
Used when HiveTableScanExec physical operator is executed with a partitioned table (and requested for rawPartitions)
methodSplitThreshold¶
spark.sql.codegen.methodSplitThreshold
Used when:
Expressionis requested to reduceCodeSizeCodegenContextis requested to buildCodeBlocks and subexpressionEliminationForWholeStageCodegenExpandExecphysical operator is requested todoConsumeHashAggregateExecphysical operator is requested to generateEvalCodeForAggFuncs
minNumPostShufflePartitions¶
spark.sql.adaptive.minNumPostShufflePartitions
Used when EnsureRequirements physical optimization is executed (for Adaptive Query Execution).
nestedSchemaPruningEnabled¶
The value of spark.sql.optimizer.nestedSchemaPruning.enabled configuration property
Used when SchemaPruning, ColumnPruning and V2ScanRelationPushDown logical optimizations are executed
nonEmptyPartitionRatioForBroadcastJoin¶
The value of spark.sql.adaptive.nonEmptyPartitionRatioForBroadcastJoin configuration property
Used when:
- DynamicJoinSelection adaptive logical optimization is executed (and shouldDemoteBroadcastHashJoin)
numShufflePartitions¶
offHeapColumnVectorEnabled¶
spark.sql.columnVector.offheap.enabled
rangeExchangeSampleSizePerPartition¶
The value of spark.sql.execution.rangeExchange.sampleSizePerPartition configuration property
Used when:
- ShuffleExchangeExec physical operator is executed
REMOVE_REDUNDANT_SORTS_ENABLED¶
The value of spark.sql.execution.removeRedundantSorts configuration property
Used when:
- RemoveRedundantSorts physical optimization is executed
REPLACE_HASH_WITH_SORT_AGG_ENABLED¶
spark.sql.execution.replaceHashWithSortAgg
runtimeFilterBloomFilterEnabled¶
spark.sql.optimizer.runtime.bloomFilter.enabled
RUNTIME_BLOOM_FILTER_MAX_NUM_BITS¶
spark.sql.optimizer.runtime.bloomFilter.maxNumBits
RUNTIME_FILTER_NUMBER_THRESHOLD¶
spark.sql.optimizer.runtimeFilter.number.threshold
runtimeFilterSemiJoinReductionEnabled¶
spark.sql.optimizer.runtimeFilter.semiJoinReduction.enabled
SKEW_JOIN_SKEWED_PARTITION_FACTOR¶
spark.sql.adaptive.skewJoin.skewedPartitionFactor configuration property
Used when:
- OptimizeSkewedJoin physical optimization is executed
SKEW_JOIN_SKEWED_PARTITION_THRESHOLD¶
spark.sql.adaptive.skewJoin.skewedPartitionThresholdInBytes configuration property
Used when:
- OptimizeSkewedJoin physical optimization is executed
SKEW_JOIN_ENABLED¶
spark.sql.adaptive.skewJoin.enabled configuration property
Used when:
- OptimizeSkewedJoin physical optimization is executed
objectAggSortBasedFallbackThreshold¶
spark.sql.objectHashAggregate.sortBased.fallbackThreshold
offHeapColumnVectorEnabled¶
spark.sql.columnVector.offheap.enabled
Used when:
InMemoryTableScanExecis requested for the vectorTypes and the input RDDOrcFileFormatis requested tobuildReaderWithPartitionValuesParquetFileFormatis requested for vectorTypes and build a data reader with partition column values appended
OPTIMIZE_ONE_ROW_RELATION_SUBQUERY¶
spark.sql.optimizer.optimizeOneRowRelationSubquery
Used when:
OptimizeOneRowRelationSubquerylogical optimization is executed
optimizeNullAwareAntiJoin¶
spark.sql.optimizeNullAwareAntiJoin configuration property
Used when:
- ExtractSingleColumnNullAwareAntiJoin Scala extractor is executed
optimizerExcludedRules¶
The value of spark.sql.optimizer.excludedRules configuration property
Used when Optimizer is requested for the batches
optimizerInSetConversionThreshold¶
spark.sql.optimizer.inSetConversionThreshold
Used when OptimizeIn logical query optimization is executed
orcVectorizedReaderNestedColumnEnabled¶
spark.sql.orc.enableNestedColumnVectorizedReader
Used when:
OrcFileFormatis requested tosupportBatchForNestedColumn
OUTPUT_COMMITTER_CLASS¶
spark.sql.sources.outputCommitterClass
Used when:
SQLHadoopMapReduceCommitProtocolis requested to setupCommitterParquetFileFormatis requested to prepareWriteParquetWriteis requested to prepareWrite
parallelFileListingInStatsComputation¶
spark.sql.statistics.parallelFileListingInStatsComputation.enabled
Used when CommandUtils helper object is requested to calculate the total size of a table (with partitions) (for AnalyzeColumnCommand and AnalyzeTableCommand commands)
parquetAggregatePushDown¶
spark.sql.parquet.aggregatePushdown
parquetCompressionCodec¶
spark.sql.parquet.compression.codec
Used when:
ParquetOptionsis requested for compressionCodecClassName
parquetFilterPushDown¶
spark.sql.parquet.filterPushdown
parquetFilterPushDownDate¶
spark.sql.parquet.filterPushdown.date
Used when:
ParquetFileFormatis requested to build a data reader (with partition column values appended)
parquetFilterPushDownDecimal¶
spark.sql.parquet.filterPushdown.decimal
Used when:
ParquetFileFormatis requested to build a data reader (with partition column values appended)ParquetPartitionReaderFactoryis requested to buildReaderBaseParquetScanBuilderis requested for pushedParquetFilters
parquetFilterPushDownInFilterThreshold¶
spark.sql.parquet.pushdown.inFilterThreshold
Used when:
ParquetFileFormatis requested to build a data reader (with partition column values appended)ParquetPartitionReaderFactoryis requested to buildReaderBaseParquetScanBuilderis requested for pushedParquetFilters
parquetFilterPushDownStringPredicate¶
spark.sql.parquet.filterPushdown.stringPredicate
parquetFilterPushDownStringStartWith¶
spark.sql.parquet.filterPushdown.string.startsWith
parquetFilterPushDownTimestamp¶
spark.sql.parquet.filterPushdown.timestamp
Used when:
ParquetFileFormatis requested to build a data reader (with partition column values appended)ParquetPartitionReaderFactoryis requested to buildReaderBaseParquetScanBuilderis requested for pushedParquetFilters
parquetOutputCommitterClass¶
spark.sql.parquet.output.committer.class
Used when:
ParquetFileFormatis requested to prepareWriteParquetWriteis requested to prepareWrite
parquetOutputTimestampType¶
spark.sql.parquet.outputTimestampType
Used when:
ParquetFileFormatis requested to prepareWriteSparkToParquetSchemaConverteris createdParquetWriteSupportis requested to initParquetWriteis requested to prepareWrite
parquetRecordFilterEnabled¶
spark.sql.parquet.recordLevelFilter.enabled
Used when ParquetFileFormat is requested to build a data reader (with partition column values appended).
parquetVectorizedReaderBatchSize¶
spark.sql.parquet.columnarReaderBatchSize
parquetVectorizedReaderEnabled¶
spark.sql.parquet.enableVectorizedReader
Used when:
FileSourceScanExecis requested for needsUnsafeRowConversion flagParquetFileFormatis requested for supportBatch flag and build a data reader with partition column values appended
parquetVectorizedReaderNestedColumnEnabled¶
spark.sql.parquet.enableNestedColumnVectorizedReader
partitionOverwriteMode¶
The value of spark.sql.sources.partitionOverwriteMode configuration property
Used when InsertIntoHadoopFsRelationCommand logical command is executed
planChangeLogLevel¶
The value of spark.sql.planChangeLog.level configuration property
Used when:
- PlanChangeLogger is created
planChangeBatches¶
The value of spark.sql.planChangeLog.batches configuration property
Used when:
PlanChangeLoggeris requested to logBatch
planChangeRules¶
The value of spark.sql.planChangeLog.rules configuration property
Used when:
PlanChangeLoggeris requested to logRule
preferSortMergeJoin¶
spark.sql.join.preferSortMergeJoin
Used in JoinSelection execution planning strategy to prefer sort merge join over shuffle hash join.
LEAF_NODE_DEFAULT_PARALLELISM¶
spark.sql.leafNodeDefaultParallelism
Used when:
SparkSessionis requested for the leafNodeDefaultParallelism
LEGACY_CTE_PRECEDENCE_POLICY¶
spark.sql.legacy.ctePrecedencePolicy
PROPAGATE_DISTINCT_KEYS_ENABLED¶
spark.sql.optimizer.propagateDistinctKeys.enabled
replaceDatabricksSparkAvroEnabled¶
spark.sql.legacy.replaceDatabricksSparkAvro.enabled
replaceExceptWithFilter¶
spark.sql.optimizer.replaceExceptWithFilter
Used when ReplaceExceptWithFilter logical optimization is executed
runSQLonFile¶
Used when:
ResolveSQLOnFileis requested to maybeSQLFile
RUNTIME_BLOOM_FILTER_EXPECTED_NUM_ITEMS¶
spark.sql.optimizer.runtime.bloomFilter.expectedNumItems
runtimeRowLevelOperationGroupFilterEnabled¶
spark.sql.optimizer.runtime.rowLevelOperationGroupFilter.enabled
sessionLocalTimeZone¶
sessionWindowBufferInMemoryThreshold¶
spark.sql.sessionWindow.buffer.in.memory.threshold
Used when:
UpdatingSessionsExecunary physical operator is executed
sessionWindowBufferSpillThreshold¶
spark.sql.sessionWindow.buffer.spill.threshold
Used when:
UpdatingSessionsExecunary physical operator is executed
sortBeforeRepartition¶
The value of spark.sql.execution.sortBeforeRepartition configuration property
Used when ShuffleExchangeExec physical operator is executed
starSchemaDetection¶
spark.sql.cbo.starSchemaDetection
Used in ReorderJoin logical optimization (and indirectly in StarSchemaDetection)
stringRedactionPattern¶
spark.sql.redaction.string.regex
Used when:
DataSourceScanExecis requested to redact sensitive information (in text representations)QueryExecutionis requested to redact sensitive information (in text representations)
subexpressionEliminationEnabled¶
spark.sql.subexpressionElimination.enabled
Used when SparkPlan is requested for subexpressionEliminationEnabled flag.
subqueryReuseEnabled¶
spark.sql.execution.reuseSubquery
Used when:
- ReuseAdaptiveSubquery adaptive physical optimization is executed
- ReuseExchangeAndSubquery physical optimization is executed
supportQuotedRegexColumnName¶
spark.sql.parser.quotedRegexColumnNames
Used when:
- Dataset.col operator is used
AstBuilderis requested to parse a dereference and column reference in a SQL statement
targetPostShuffleInputSize¶
spark.sql.adaptive.shuffle.targetPostShuffleInputSize
Used when EnsureRequirements physical optimization is executed (for Adaptive Query Execution)
THRIFTSERVER_FORCE_CANCEL¶
spark.sql.thriftServer.interruptOnCancel
Used when:
SparkExecuteStatementOperationis created (forceCancel)
truncateTableIgnorePermissionAcl¶
spark.sql.truncateTable.ignorePermissionAcl.enabled
Used when TruncateTableCommand logical command is executed
useCompression¶
The value of spark.sql.inMemoryColumnarStorage.compressed configuration property
Used when CacheManager is requested to cache a structured query
useObjectHashAggregation¶
spark.sql.execution.useObjectHashAggregateExec
Used when Aggregation execution planning strategy is executed (and uses AggUtils to create an aggregation physical operator).
v2BucketingPartiallyClusteredDistributionEnabled¶
spark.sql.sources.v2.bucketing.partiallyClusteredDistribution.enabled
v2BucketingPushPartValuesEnabled¶
spark.sql.sources.v2.bucketing.pushPartValues.enabled
variableSubstituteEnabled¶
Used when:
VariableSubstitutionis requested to substitute variables in a SQL command
wholeStageEnabled¶
Used in:
- CollapseCodegenStages to control codegen
- ParquetFileFormat to control row batch reading
wholeStageFallback¶
wholeStageMaxNumFields¶
Used in:
- CollapseCodegenStages to control codegen
- ParquetFileFormat to control row batch reading
wholeStageSplitConsumeFuncByOperator¶
spark.sql.codegen.splitConsumeFuncByOperator
Used when CodegenSupport is requested to consume
wholeStageUseIdInClassName¶
spark.sql.codegen.useIdInClassName
Used when WholeStageCodegenExec is requested to generate the Java source code for the child physical plan subtree (when created)
windowExecBufferInMemoryThreshold¶
spark.sql.windowExec.buffer.in.memory.threshold
Used when:
- WindowExec unary physical operator is executed
windowExecBufferSpillThreshold¶
spark.sql.windowExec.buffer.spill.threshold
Used when:
- WindowExec unary physical operator is executed