Re: [ANNOUNCE] Announcing Apache Spark 2.2.3

2019-01-15 Thread Jiaan Geng
Glad to hear this. -- Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/ - To unsubscribe e-mail: user-unsubscr...@spark.apache.org

Re: Async action in Dataframe

2018-12-23 Thread Jiaan Geng
RDD have not the method `collectAsync`.There exists a implicit conversion from RDD to AsyncRDDActions in object RDD. The implicit conversion is : implicit def rddToAsyncRDDActions[T: ClassTag](rdd: RDD[T]): AsyncRDDActions[T] = { new AsyncRDDActions(rdd) } The method collect of RDD use the

Re: Spark not working with Hadoop 4mc compression

2018-12-20 Thread Jiaan Geng
I think com.hadoop.compression.lzo.LzoCodec not in spark classpaths,please put suitable hadoop-lzo.jar into directory spark_home/jars/. -- Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/ - To unsubscribe e-mail

Re: [Spark SQL]use zstd, No enum constant parquet.hadoop.metadata.CompressionCodecName.ZSTD

2018-12-20 Thread Jiaan Geng
I think your hive table using CompressionCodecName, but your parquet-hadoop-bundle.jar in spark classpaths is not a correct version. -- Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/ - To unsubscribe e-mail: us

Re: running updates using SPARK

2018-12-20 Thread Jiaan Geng
I think Spark is a Calculation engine design for OLAP or Ad-hoc.Spark is not a traditional relational database,UPDATE need some mandatory constraint like transaction and lock. -- Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/

Re: Multiple sessions in one application?

2018-12-20 Thread Jiaan Geng
This scene is rare. When you provide a web server for spark. maybe you need it. -- Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/ - To unsubscribe e-mail: user-unsubscr...@spark.apache.org

Re: Read Time from a remote data source

2018-12-19 Thread Jiaan Geng
First, Spark worker not have the ability to compute.In fact,executor is responsible for computation. Executor running tasks is distributed by driver. Each Task just read some section of data in normal, but the stage have only one partition. IF your operators not contains the operator that will pull

Re: Spark 2.2.1 - Operation not allowed: alter table replace columns

2018-12-19 Thread Jiaan Geng
This SQL syntax is not supported now!Please use ALTER TABLE ... CHANGE COLUMN . -- Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/ - To unsubscribe e-mail: user-unsubscr...@spark.apache.org