Skip to content

Commit 60588cd

Browse files
committed
match
1 parent 8b55edf commit 60588cd

File tree

1 file changed

+21
-20
lines changed
  • paimon-spark/paimon-spark3-common/src/main/scala/org/apache/spark/sql/paimon/shims

1 file changed

+21
-20
lines changed

paimon-spark/paimon-spark3-common/src/main/scala/org/apache/spark/sql/paimon/shims/Spark3Shim.scala

Lines changed: 21 additions & 20 deletions
Original file line numberDiff line numberDiff line change
@@ -18,19 +18,20 @@
1818

1919
package org.apache.spark.sql.paimon.shims
2020

21-
import org.apache.hadoop.fs.Path
2221
import org.apache.paimon.data.variant.Variant
2322
import org.apache.paimon.spark.catalyst.analysis.Spark3ResolutionRules
2423
import org.apache.paimon.spark.catalyst.parser.extensions.PaimonSpark3SqlExtensionsParser
2524
import org.apache.paimon.spark.data.{Spark3ArrayData, Spark3InternalRow, Spark3InternalRowWithBlob, SparkArrayData, SparkInternalRow}
2625
import org.apache.paimon.types.{DataType, RowType}
26+
27+
import org.apache.hadoop.fs.Path
2728
import org.apache.spark.sql.SparkSession
2829
import org.apache.spark.sql.catalyst.InternalRow
2930
import org.apache.spark.sql.catalyst.expressions.{Attribute, Expression}
3031
import org.apache.spark.sql.catalyst.expressions.aggregate.AggregateExpression
3132
import org.apache.spark.sql.catalyst.parser.ParserInterface
32-
import org.apache.spark.sql.catalyst.plans.logical.MergeRows.Instruction
3333
import org.apache.spark.sql.catalyst.plans.logical._
34+
import org.apache.spark.sql.catalyst.plans.logical.MergeRows.Instruction
3435
import org.apache.spark.sql.catalyst.rules.Rule
3536
import org.apache.spark.sql.catalyst.util.ArrayData
3637
import org.apache.spark.sql.connector.catalog.{Identifier, Table, TableCatalog}
@@ -136,29 +137,29 @@ class Spark3Shim extends SparkShim {
136137
throw new UnsupportedOperationException()
137138

138139
def createFileIndex(
139-
options: CaseInsensitiveStringMap,
140-
sparkSession: SparkSession,
141-
paths: Seq[String],
142-
userSpecifiedSchema: Option[StructType],
143-
partitionSchema: StructType): PartitioningAwareFileIndex = {
140+
options: CaseInsensitiveStringMap,
141+
sparkSession: SparkSession,
142+
paths: Seq[String],
143+
userSpecifiedSchema: Option[StructType],
144+
partitionSchema: StructType): PartitioningAwareFileIndex = {
144145

145146
class PartitionedMetadataLogFileIndex(
146-
sparkSession: SparkSession,
147-
path: Path,
148-
parameters: Map[String, String],
149-
userSpecifiedSchema: Option[StructType],
150-
override val partitionSchema: StructType)
147+
sparkSession: SparkSession,
148+
path: Path,
149+
parameters: Map[String, String],
150+
userSpecifiedSchema: Option[StructType],
151+
override val partitionSchema: StructType)
151152
extends MetadataLogFileIndex(sparkSession, path, parameters, userSpecifiedSchema)
152153

153154
class PartitionedInMemoryFileIndex(
154-
sparkSession: SparkSession,
155-
rootPathsSpecified: Seq[Path],
156-
parameters: Map[String, String],
157-
userSpecifiedSchema: Option[StructType],
158-
fileStatusCache: FileStatusCache = NoopCache,
159-
userSpecifiedPartitionSpec: Option[PartitionSpec] = None,
160-
metadataOpsTimeNs: Option[Long] = None,
161-
override val partitionSchema: StructType)
155+
sparkSession: SparkSession,
156+
rootPathsSpecified: Seq[Path],
157+
parameters: Map[String, String],
158+
userSpecifiedSchema: Option[StructType],
159+
fileStatusCache: FileStatusCache = NoopCache,
160+
userSpecifiedPartitionSpec: Option[PartitionSpec] = None,
161+
metadataOpsTimeNs: Option[Long] = None,
162+
override val partitionSchema: StructType)
162163
extends InMemoryFileIndex(
163164
sparkSession,
164165
rootPathsSpecified,

0 commit comments

Comments
 (0)