I'd suggest looking at the reference in the programming guide:
http://spark.apache.org/docs/latest/sql-programming-guide.html#spark-sql-datatype-reference


On Thu, Dec 11, 2014 at 6:45 PM, Alessandro Baretta <alexbare...@gmail.com>
wrote:

> Thanks. This is useful.
>
> Alex
>
> On Thu, Dec 11, 2014 at 4:35 PM, Cheng, Hao <hao.ch...@intel.com> wrote:
>>
>> Part of it can be found at:
>>
>> https://github.com/apache/spark/pull/3429/files#diff-f88c3e731fcb17b1323b778807c35b38R34
>>
>> Sorry it's a TO BE reviewed PR, but still should be informative.
>>
>> Cheng Hao
>>
>> -----Original Message-----
>> From: Alessandro Baretta [mailto:alexbare...@gmail.com]
>> Sent: Friday, December 12, 2014 6:37 AM
>> To: Michael Armbrust; dev@spark.apache.org
>> Subject: Where are the docs for the SparkSQL DataTypes?
>>
>> Michael & other Spark SQL junkies,
>>
>> As I read through the Spark API docs, in particular those for the
>> org.apache.spark.sql package, I can't seem to find details about the Scala
>> classes representing the various SparkSQL DataTypes, for instance
>> DecimalType. I find DataType classes in org.apache.spark.sql.api.java, but
>> they don't seem to match the similarly named scala classes. For instance,
>> DecimalType is documented as having a nullary constructor, but if I try to
>> construct an instance of org.apache.spark.sql.DecimalType without any
>> parameters, the compiler complains about the lack of a precisionInfo field,
>> which I have discovered can be passed in as None. Where is all this stuff
>> documented?
>>
>> Alex
>>
>

Reply via email to