|
17 | 17 |
|
18 | 18 | package org.apache.spark.sql.parquet |
19 | 19 |
|
20 | | -import java.sql.{Date, Timestamp} |
| 20 | +import java.sql.Timestamp |
21 | 21 | import java.util.{TimeZone, Calendar} |
22 | 22 |
|
23 | 23 | import scala.collection.mutable.{Buffer, ArrayBuffer, HashMap} |
@@ -127,6 +127,12 @@ private[sql] object CatalystConverter { |
127 | 127 | parent.updateByte(fieldIndex, value.asInstanceOf[ByteType.JvmType]) |
128 | 128 | } |
129 | 129 | } |
| 130 | + case DateType => { |
| 131 | + new CatalystPrimitiveConverter(parent, fieldIndex) { |
| 132 | + override def addInt(value: Int): Unit = |
| 133 | + parent.updateDate(fieldIndex, value.asInstanceOf[DateType.JvmType]) |
| 134 | + } |
| 135 | + } |
130 | 136 | case d: DecimalType => { |
131 | 137 | new CatalystPrimitiveConverter(parent, fieldIndex) { |
132 | 138 | override def addBinary(value: Binary): Unit = |
@@ -193,7 +199,7 @@ private[parquet] abstract class CatalystConverter extends GroupConverter { |
193 | 199 | updateField(fieldIndex, value) |
194 | 200 |
|
195 | 201 | protected[parquet] def updateDate(fieldIndex: Int, value: Int): Unit = |
196 | | - updateField(fieldIndex, new Date(value)) |
| 202 | + updateField(fieldIndex, value) |
197 | 203 |
|
198 | 204 | protected[parquet] def updateLong(fieldIndex: Int, value: Long): Unit = |
199 | 205 | updateField(fieldIndex, value) |
@@ -392,7 +398,7 @@ private[parquet] class CatalystPrimitiveRowConverter( |
392 | 398 | current.setInt(fieldIndex, value) |
393 | 399 |
|
394 | 400 | override protected[parquet] def updateDate(fieldIndex: Int, value: Int): Unit = |
395 | | - current.update(fieldIndex, new Date(value)) |
| 401 | + current.update(fieldIndex, value) |
396 | 402 |
|
397 | 403 | override protected[parquet] def updateLong(fieldIndex: Int, value: Long): Unit = |
398 | 404 | current.setLong(fieldIndex, value) |
|
0 commit comments