diff --git a/spark/v3.4/spark/src/main/java/org/apache/iceberg/spark/data/SparkParquetWriters.java b/spark/v3.4/spark/src/main/java/org/apache/iceberg/spark/data/SparkParquetWriters.java index 6a99912e1eb9..f4ae6114c8ab 100644 --- a/spark/v3.4/spark/src/main/java/org/apache/iceberg/spark/data/SparkParquetWriters.java +++ b/spark/v3.4/spark/src/main/java/org/apache/iceberg/spark/data/SparkParquetWriters.java @@ -77,15 +77,6 @@ public static ParquetValueWriter buildWriter( new WriteBuilder(type)); } - public static ParquetValueWriter buildWriter( - StructType dfSchema, MessageType type, Schema icebergSchema) { - return (ParquetValueWriter) - ParquetWithSparkSchemaVisitor.visit( - dfSchema != null ? dfSchema : SparkSchemaUtil.convert(icebergSchema), - type, - new WriteBuilder(type)); - } - private static class WriteBuilder extends ParquetWithSparkSchemaVisitor> { private final MessageType type; diff --git a/spark/v3.4/spark/src/main/java/org/apache/iceberg/spark/source/SparkFileWriterFactory.java b/spark/v3.4/spark/src/main/java/org/apache/iceberg/spark/source/SparkFileWriterFactory.java index 2b3bf73d56b3..39110f0b0597 100644 --- a/spark/v3.4/spark/src/main/java/org/apache/iceberg/spark/source/SparkFileWriterFactory.java +++ b/spark/v3.4/spark/src/main/java/org/apache/iceberg/spark/source/SparkFileWriterFactory.java @@ -60,7 +60,7 @@ class SparkFileWriterFactory extends RegistryBasedFileWriterFactory writeProperties; /** @@ -100,7 +100,7 @@ class SparkFileWriterFactory extends RegistryBasedFileWriterFactory newPositionDeleteWriter( : MetricsConfig.forPositionDelete(table); try { - return switch (format) { + return switch (deleteFormat) { case AVRO -> Avro.writeDeletes(file) .createWriterFunc( @@ -215,14 +215,13 @@ public PositionDeleteWriter newPositionDeleteWriter( .metricsConfig(metricsConfig) .withPartition(partition) .overwrite() - .metricsConfig(metricsConfig) .rowSchema(positionDeleteRowSchema) .withSpec(spec) .withKeyMetadata(file.keyMetadata()) .buildPositionWriter(); default -> throw new UnsupportedOperationException( - "Cannot write pos-deletes for unsupported file format: " + format); + "Cannot write pos-deletes for unsupported file format: " + deleteFormat); }; } catch (IOException e) { throw new UncheckedIOException("Failed to create new position delete writer", e); diff --git a/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/data/SparkParquetWriters.java b/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/data/SparkParquetWriters.java index dda634a46fda..58be7f610c81 100644 --- a/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/data/SparkParquetWriters.java +++ b/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/data/SparkParquetWriters.java @@ -77,15 +77,6 @@ public static ParquetValueWriter buildWriter( new WriteBuilder(type)); } - public static ParquetValueWriter buildWriter( - StructType dfSchema, MessageType type, Schema icebergSchema) { - return (ParquetValueWriter) - ParquetWithSparkSchemaVisitor.visit( - dfSchema != null ? dfSchema : SparkSchemaUtil.convert(icebergSchema), - type, - new WriteBuilder(type)); - } - private static class WriteBuilder extends ParquetWithSparkSchemaVisitor> { private final MessageType type; diff --git a/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/source/SparkFileWriterFactory.java b/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/source/SparkFileWriterFactory.java index 2b3bf73d56b3..39110f0b0597 100644 --- a/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/source/SparkFileWriterFactory.java +++ b/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/source/SparkFileWriterFactory.java @@ -60,7 +60,7 @@ class SparkFileWriterFactory extends RegistryBasedFileWriterFactory writeProperties; /** @@ -100,7 +100,7 @@ class SparkFileWriterFactory extends RegistryBasedFileWriterFactory newPositionDeleteWriter( : MetricsConfig.forPositionDelete(table); try { - return switch (format) { + return switch (deleteFormat) { case AVRO -> Avro.writeDeletes(file) .createWriterFunc( @@ -215,14 +215,13 @@ public PositionDeleteWriter newPositionDeleteWriter( .metricsConfig(metricsConfig) .withPartition(partition) .overwrite() - .metricsConfig(metricsConfig) .rowSchema(positionDeleteRowSchema) .withSpec(spec) .withKeyMetadata(file.keyMetadata()) .buildPositionWriter(); default -> throw new UnsupportedOperationException( - "Cannot write pos-deletes for unsupported file format: " + format); + "Cannot write pos-deletes for unsupported file format: " + deleteFormat); }; } catch (IOException e) { throw new UncheckedIOException("Failed to create new position delete writer", e); diff --git a/spark/v4.0/spark/src/main/java/org/apache/iceberg/spark/data/SparkParquetWriters.java b/spark/v4.0/spark/src/main/java/org/apache/iceberg/spark/data/SparkParquetWriters.java index 3ff5ef9c577d..ba816efc0ac8 100644 --- a/spark/v4.0/spark/src/main/java/org/apache/iceberg/spark/data/SparkParquetWriters.java +++ b/spark/v4.0/spark/src/main/java/org/apache/iceberg/spark/data/SparkParquetWriters.java @@ -91,15 +91,6 @@ public static ParquetValueWriter buildWriter( new WriteBuilder(type)); } - public static ParquetValueWriter buildWriter( - StructType dfSchema, MessageType type, Schema icebergSchema) { - return (ParquetValueWriter) - ParquetWithSparkSchemaVisitor.visit( - dfSchema != null ? dfSchema : SparkSchemaUtil.convert(icebergSchema), - type, - new WriteBuilder(type)); - } - private static class WriteBuilder extends ParquetWithSparkSchemaVisitor> { private final MessageType type; diff --git a/spark/v4.0/spark/src/main/java/org/apache/iceberg/spark/source/SparkFileWriterFactory.java b/spark/v4.0/spark/src/main/java/org/apache/iceberg/spark/source/SparkFileWriterFactory.java index 2b3bf73d56b3..39110f0b0597 100644 --- a/spark/v4.0/spark/src/main/java/org/apache/iceberg/spark/source/SparkFileWriterFactory.java +++ b/spark/v4.0/spark/src/main/java/org/apache/iceberg/spark/source/SparkFileWriterFactory.java @@ -60,7 +60,7 @@ class SparkFileWriterFactory extends RegistryBasedFileWriterFactory writeProperties; /** @@ -100,7 +100,7 @@ class SparkFileWriterFactory extends RegistryBasedFileWriterFactory newPositionDeleteWriter( : MetricsConfig.forPositionDelete(table); try { - return switch (format) { + return switch (deleteFormat) { case AVRO -> Avro.writeDeletes(file) .createWriterFunc( @@ -215,14 +215,13 @@ public PositionDeleteWriter newPositionDeleteWriter( .metricsConfig(metricsConfig) .withPartition(partition) .overwrite() - .metricsConfig(metricsConfig) .rowSchema(positionDeleteRowSchema) .withSpec(spec) .withKeyMetadata(file.keyMetadata()) .buildPositionWriter(); default -> throw new UnsupportedOperationException( - "Cannot write pos-deletes for unsupported file format: " + format); + "Cannot write pos-deletes for unsupported file format: " + deleteFormat); }; } catch (IOException e) { throw new UncheckedIOException("Failed to create new position delete writer", e);