Skip to content

Commit d96abce

Browse files
committed
Revert "Add Iceberg as a dep (apache#916)"
This reverts commit 8a7c1a9.
1 parent 89690dd commit d96abce

File tree

8 files changed

+2
-271
lines changed

8 files changed

+2
-271
lines changed

pom.xml

Lines changed: 0 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -194,7 +194,6 @@
194194
<selenium.version>3.141.59</selenium.version>
195195
<htmlunit.version>2.40.0</htmlunit.version>
196196
<acispark.callhome.version>0.2.0</acispark.callhome.version>
197-
<iceberg.version>0.12.0-apple-preview-SNAPSHOT</iceberg.version>
198197
<!--
199198
Managed up from older version from Avro; sync with jackson-module-paranamer dependency version
200199
-->
@@ -294,10 +293,6 @@
294293
<id>apple</id>
295294
<url>https://artifacts.apple.com/libs-release</url>
296295
</repository>
297-
<repository>
298-
<id>apple-snapshot</id>
299-
<url>https://artifacts.apple.com/libs-snapshot</url>
300-
</repository>
301296
</repositories>
302297
<pluginRepositories>
303298
<pluginRepository>
@@ -2405,11 +2400,6 @@
24052400
</exclusion>
24062401
</exclusions>
24072402
</dependency>
2408-
<dependency>
2409-
<groupId>org.apache.iceberg</groupId>
2410-
<artifactId>iceberg-spark3-runtime</artifactId>
2411-
<version>${iceberg.version}</version>
2412-
</dependency>
24132403
</dependencies>
24142404
</dependencyManagement>
24152405

project/SparkBuild.scala

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -270,7 +270,6 @@ object SparkBuild extends PomBuild {
270270
"gcs-maven-central-mirror" at "https://maven-central.storage-download.googleapis.com/maven2/",
271271
DefaultMavenRepository,
272272
"apple" at "https://artifacts.apple.com/libs-release",
273-
"apple-snapshot" at "https://artifacts.apple.com/libs-snapshot",
274273
Resolver.mavenLocal,
275274
Resolver.file("ivyLocal", file(Path.userHome.absolutePath + "/.ivy2/local"))(Resolver.ivyStylePatterns)
276275
),

sql/catalyst/src/main/scala/org/apache/spark/sql/connector/expressions/expressions.scala

Lines changed: 0 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -116,22 +116,6 @@ private[sql] final case class BucketTransform(
116116
}
117117
}
118118

119-
private[sql] object TruncateTransform {
120-
def unapply(expr: Expression): Option[(Int, FieldReference)] = expr match {
121-
case transform: Transform =>
122-
transform match {
123-
case NamedTransform("truncate", Seq(Ref(seq: Seq[String]), Lit(value: Int, IntegerType))) =>
124-
Some((value, FieldReference(seq)))
125-
case NamedTransform("truncate", Seq(Lit(value: Int, IntegerType), Ref(seq: Seq[String]))) =>
126-
Some((value, FieldReference(seq)))
127-
case _ =>
128-
None
129-
}
130-
case _ =>
131-
None
132-
}
133-
}
134-
135119
private[sql] object BucketTransform {
136120
def unapply(expr: Expression): Option[(Int, FieldReference)] = expr match {
137121
case transform: Transform =>

sql/core/pom.xml

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -117,10 +117,6 @@
117117
<groupId>org.apache.xbean</groupId>
118118
<artifactId>xbean-asm7-shaded</artifactId>
119119
</dependency>
120-
<dependency>
121-
<groupId>org.apache.iceberg</groupId>
122-
<artifactId>iceberg-spark3-runtime</artifactId>
123-
</dependency>
124120
<dependency>
125121
<groupId>org.scalacheck</groupId>
126122
<artifactId>scalacheck_${scala.binary.version}</artifactId>

sql/core/src/main/scala/org/apache/spark/sql/catalyst/analysis/IcebergSupport.scala

Lines changed: 0 additions & 37 deletions
This file was deleted.

sql/core/src/main/scala/org/apache/spark/sql/catalyst/expressions/icebergExpressions.scala

Lines changed: 0 additions & 152 deletions
This file was deleted.

sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DistributionAndOrderingUtils.scala

Lines changed: 2 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -19,10 +19,10 @@ package org.apache.spark.sql.execution.datasources.v2
1919

2020
import org.apache.spark.sql.AnalysisException
2121
import org.apache.spark.sql.catalyst.analysis.Resolver
22-
import org.apache.spark.sql.catalyst.expressions.{Ascending, Descending, Expression, IcebergBucketTransform, IcebergDayTransform, IcebergHourTransform, IcebergMonthTransform, IcebergTruncateTransform, IcebergYearTransform, NamedExpression, NullOrdering, NullsFirst, NullsLast, SortDirection, SortOrder}
22+
import org.apache.spark.sql.catalyst.expressions.{Ascending, Descending, Expression, NamedExpression, NullOrdering, NullsFirst, NullsLast, SortDirection, SortOrder}
2323
import org.apache.spark.sql.catalyst.plans.logical.{LogicalPlan, RepartitionByExpression, Sort}
2424
import org.apache.spark.sql.connector.distributions.{ClusteredDistribution, OrderedDistribution, UnspecifiedDistribution}
25-
import org.apache.spark.sql.connector.expressions.{BucketTransform, DaysTransform, Expression => V2Expression, FieldReference, HoursTransform, IdentityTransform, MonthsTransform, NullOrdering => V2NullOrdering, SortDirection => V2SortDirection, SortValue, TruncateTransform, YearsTransform}
25+
import org.apache.spark.sql.connector.expressions.{Expression => V2Expression, FieldReference, IdentityTransform, NullOrdering => V2NullOrdering, SortDirection => V2SortDirection, SortValue}
2626
import org.apache.spark.sql.connector.write.{RequiresDistributionAndOrdering, Write}
2727
import org.apache.spark.sql.internal.SQLConf
2828

@@ -87,18 +87,6 @@ object DistributionAndOrderingUtils {
8787
SortOrder(catalystChild, toCatalyst(direction), toCatalyst(nullOrdering), Seq.empty)
8888
case IdentityTransform(ref) =>
8989
resolve(ref)
90-
case BucketTransform(numBuckets, ref) =>
91-
IcebergBucketTransform(numBuckets, resolve(ref))
92-
case TruncateTransform(length, ref) =>
93-
IcebergTruncateTransform(resolve(ref), length)
94-
case YearsTransform(ref) =>
95-
IcebergYearTransform(resolve(ref))
96-
case MonthsTransform(ref) =>
97-
IcebergMonthTransform(resolve(ref))
98-
case DaysTransform(ref) =>
99-
IcebergDayTransform(resolve(ref))
100-
case HoursTransform(ref) =>
101-
IcebergHourTransform(resolve(ref))
10290
case ref: FieldReference =>
10391
resolve(ref)
10492
case _ =>

sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/v2/IcebergImplicits.scala

Lines changed: 0 additions & 37 deletions
This file was deleted.

0 commit comments

Comments
 (0)