Just my "few cents" on this.
I having the same problems with v 1.0.1 but this bug is sporadic and looks
like is relayed to object initialization.
Even more, i'm not using any SQL or something. I just have utility class
like this:
object DataTypeDescriptor {
type DataType = String
val BOOLEAN = "BOOLEAN"
val STRING = "STRING"
val TIMESTAMP = "TIMESTAMP"
val LONG = "LONG"
val INT = "INT"
val SHORT = "SHORT"
val BYTE = "BYTE"
val DECIMAL = "DECIMAL"
val DOUBLE = "DOUBLE"
val FLOAT = "FLOAT"
def $$(name: String, format: Option[String] = None) =
DataTypeDescriptor(name, format)
private lazy val nativeTypes: Map[String, NativeType] = Map(
BOOLEAN -> BooleanType, STRING -> StringType, TIMESTAMP ->
TimestampType, LONG -> LongType, INT -> IntegerType,
SHORT -> ShortType, BYTE -> ByteType, DECIMAL -> DecimalType, DOUBLE ->
DoubleType, FLOAT -> FloatType
)
lazy val defaultValues: Map[String, Any] = Map(
BOOLEAN -> false, STRING -> "", TIMESTAMP -> null, LONG -> 0L, INT -> 0,
SHORT -> 0.toShort, BYTE -> 0.toByte,
DECIMAL -> BigDecimal(0d), DOUBLE -> 0d, FLOAT -> 0f
)
def apply(dataType: String): DataTypeDescriptor = {
DataTypeDescriptor(dataType.toUpperCase, None)
}
def apply(dataType: SparkDataType): DataTypeDescriptor = {
nativeTypes
.find { case (_, descriptor) => descriptor == dataType }
.map { case (name, descriptor) => DataTypeDescriptor(name, None) }
.get
}
.....
and some test that check each of this methods.
The problem is that this test fails randomly with this error.
P.S.: I did not have this problem in Spark 1.0.0
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Error-while-running-Spark-SQL-join-when-using-Spark-1-0-1-tp9776p9817.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.