Skip to content

Commit c1b5932

Browse files
committed
move
1 parent 9567cc7 commit c1b5932

21 files changed

Lines changed: 31 additions & 35 deletions

File tree

sql/catalyst/pom.xml

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -114,6 +114,10 @@
114114
<version>2.7.3</version>
115115
<type>jar</type>
116116
</dependency>
117+
<dependency>
118+
<groupId>org.apache.arrow</groupId>
119+
<artifactId>arrow-vector</artifactId>
120+
</dependency>
117121
</dependencies>
118122
<build>
119123
<outputDirectory>target/scala-${scala.binary.version}/classes</outputDirectory>

sql/core/src/main/java/org/apache/spark/sql/sources/v2/reader/SupportsPushDownFilters.java renamed to sql/catalyst/src/main/java/org/apache/spark/sql/sources/v2/reader/SupportsPushDownFilters.java

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,6 @@
2525
* push down filters to the data source and reduce the size of the data to be read.
2626
*/
2727
@Evolving
28-
// TODO: create data source v2 filter API and move this interface to the catalyst module.
2928
public interface SupportsPushDownFilters extends ScanBuilder {
3029

3130
/**

sql/catalyst/src/main/java/org/apache/spark/sql/sources/v2/writer/SupportsDynamicOverwrite.java

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@
2525
* write does not contain data will remain unchanged.
2626
* <p>
2727
* This is provided to implement SQL compatible with Hive table operations but is not recommended.
28-
* Instead, use the {@code SupportsOverwrite overwrite by filter API} to explicitly replace data.
28+
* Instead, use the {@link SupportsOverwrite overwrite by filter API} to explicitly replace data.
2929
*/
3030
public interface SupportsDynamicOverwrite extends WriteBuilder {
3131
/**

sql/core/src/main/java/org/apache/spark/sql/sources/v2/writer/SupportsOverwrite.java renamed to sql/catalyst/src/main/java/org/apache/spark/sql/sources/v2/writer/SupportsOverwrite.java

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,6 @@
2626
* Overwriting data by filter will delete any data that matches the filter and replace it with data
2727
* that is committed in the write.
2828
*/
29-
// TODO: create data source v2 filter API and move this interface to the catalyst module.
3029
public interface SupportsOverwrite extends WriteBuilder, SupportsTruncate {
3130
/**
3231
* Configures a write to replace data matching the filters with data committed in the write.

sql/core/src/main/java/org/apache/spark/sql/execution/ArrowColumnVector.java renamed to sql/catalyst/src/main/java/org/apache/spark/sql/vectorized/ArrowColumnVector.java

Lines changed: 2 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -15,19 +15,16 @@
1515
* limitations under the License.
1616
*/
1717

18-
package org.apache.spark.sql.execution;
18+
package org.apache.spark.sql.vectorized;
1919

2020
import io.netty.buffer.ArrowBuf;
2121
import org.apache.arrow.vector.*;
2222
import org.apache.arrow.vector.complex.*;
2323
import org.apache.arrow.vector.holders.NullableVarCharHolder;
2424

2525
import org.apache.spark.annotation.Evolving;
26-
import org.apache.spark.sql.execution.arrow.ArrowUtils;
26+
import org.apache.spark.sql.util.ArrowUtils;
2727
import org.apache.spark.sql.types.*;
28-
import org.apache.spark.sql.vectorized.ColumnVector;
29-
import org.apache.spark.sql.vectorized.ColumnarArray;
30-
import org.apache.spark.sql.vectorized.ColumnarMap;
3128
import org.apache.spark.unsafe.types.UTF8String;
3229

3330
/**

sql/core/src/main/scala/org/apache/spark/sql/sources/filters.scala renamed to sql/catalyst/src/main/scala/org/apache/spark/sql/sources/filters.scala

File renamed without changes.

sql/core/src/main/scala/org/apache/spark/sql/execution/arrow/ArrowUtils.scala renamed to sql/catalyst/src/main/scala/org/apache/spark/sql/util/ArrowUtils.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@
1515
* limitations under the License.
1616
*/
1717

18-
package org.apache.spark.sql.execution.arrow
18+
package org.apache.spark.sql.util
1919

2020
import scala.collection.JavaConverters._
2121

sql/core/src/test/scala/org/apache/spark/sql/execution/arrow/ArrowUtilsSuite.scala renamed to sql/catalyst/src/test/scala/org/apache/spark/sql/util/ArrowUtilsSuite.scala

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -15,9 +15,9 @@
1515
* limitations under the License.
1616
*/
1717

18-
package org.apache.spark.sql.execution.arrow
18+
package org.apache.spark.sql.util
1919

20-
import org.apache.arrow.vector.types.pojo.{ArrowType, Field, FieldType, Schema}
20+
import org.apache.arrow.vector.types.pojo.ArrowType
2121

2222
import org.apache.spark.SparkFunSuite
2323
import org.apache.spark.sql.catalyst.util.DateTimeUtils

sql/core/pom.xml

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -112,10 +112,6 @@
112112
<groupId>com.fasterxml.jackson.core</groupId>
113113
<artifactId>jackson-databind</artifactId>
114114
</dependency>
115-
<dependency>
116-
<groupId>org.apache.arrow</groupId>
117-
<artifactId>arrow-vector</artifactId>
118-
</dependency>
119115
<dependency>
120116
<groupId>org.apache.xbean</groupId>
121117
<artifactId>xbean-asm7-shaded</artifactId>

sql/core/src/main/scala/org/apache/spark/sql/execution/arrow/ArrowConverters.scala

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -33,9 +33,9 @@ import org.apache.spark.api.java.JavaRDD
3333
import org.apache.spark.network.util.JavaUtils
3434
import org.apache.spark.sql.{DataFrame, SQLContext}
3535
import org.apache.spark.sql.catalyst.InternalRow
36-
import org.apache.spark.sql.execution.ArrowColumnVector
3736
import org.apache.spark.sql.types._
38-
import org.apache.spark.sql.vectorized.{ColumnarBatch, ColumnVector}
37+
import org.apache.spark.sql.util.ArrowUtils
38+
import org.apache.spark.sql.vectorized.{ArrowColumnVector, ColumnarBatch, ColumnVector}
3939
import org.apache.spark.util.{ByteBufferOutputStream, Utils}
4040

4141

0 commit comments

Comments
 (0)