File tree Expand file tree Collapse file tree
sql/core/src/main/scala/org/apache/spark/sql Expand file tree Collapse file tree Original file line number Diff line number Diff line change @@ -375,17 +375,15 @@ class SQLContext(@transient val sparkContext: SparkContext)
375375 import scala .collection .JavaConversions ._
376376 def typeOfComplexValue : PartialFunction [Any , DataType ] = {
377377 case c : java.util.List [_] =>
378- ArrayType (ScalaReflection . typeOfObject(c.head))
378+ ArrayType (typeOfObject(c.head))
379379 case c : java.util.Set [_] =>
380- ArrayType (ScalaReflection . typeOfObject(c.head))
380+ ArrayType (typeOfObject(c.head))
381381 case c : java.util.Map [_, _] =>
382382 val (key, value) = c.head
383- MapType (
384- ScalaReflection .typeOfObject(key),
385- ScalaReflection .typeOfObject(value))
383+ MapType (typeOfObject(key), typeOfObject(value))
386384 case c if c.getClass.isArray =>
387385 val elem = c.asInstanceOf [Array [_]].head
388- ArrayType (ScalaReflection . typeOfObject(elem))
386+ ArrayType (typeOfObject(elem))
389387 case c => throw new Exception (s " Object of type $c cannot be used " )
390388 }
391389
Original file line number Diff line number Diff line change @@ -31,7 +31,7 @@ import org.apache.spark.sql.Logging
3131
3232private [sql] object JsonRDD extends Logging {
3333
34- def jsonStringToRow (schema : StructType , jsonIter : Iterator [String ]): Iterator [Row ] = {
34+ private [sql] def jsonStringToRow (schema : StructType , jsonIter : Iterator [String ]): Iterator [Row ] = {
3535 parseJson(jsonIter).map(parsed => asRow(parsed, schema))
3636 }
3737
You can’t perform that action at this time.
0 commit comments