Hi Srinivas,
You can subscribe the mail list in the community page by yourself
http://spark.apache.org/community.html
On Tue, Nov 29, 2016 at 9:59 AM, Srinivas Potluri
wrote:
> Hi,
>
> I am interested to contribute code on Spark. Could you please add me into
> the mailing list / DL.
>
> Thanks,
Hi everyone,
Kubernetes is a technology that is a key player in the cluster computing world.
Currently, running Spark applications on Kubernetes requires deploying a
standalone Spark cluster on the Kubernetes cluster, and then running the jobs
against the standalone Spark cluster. However, ther
Hi all,
To explain the scenario a bit more.
We need to retain the order when writing to the RDBMS tables.
The way we found was to execute the DB Write *job* for each partition which
is really costly.
One reason being that the partition count is really high( 200) and it seems
we cannot control the
Hello Spark dev community,
I took this the following jira item
(https://github.com/apache/spark/pull/15848) and am looking for some general
pointers, it seems that I am running into issues where things work successfully
doing local development on my macbook pro but fail on jenkins for a multiti
Hi devs:
I think it's unnecessary to use c1._1 += c2.1 in combOp operation, I
think it's the same if we use c1._1+c2._1, see the code below :
in GradientDescent.scala
val (gradientSum, lossSum, miniBatchSize) = data.sample(false,
miniBatchFraction, 42 + i)
.treeAggregate((BDV.zeros[D
This is not an issue with all tables created in Spark 2.1, though I'm not sure
why some work and some do not. I have found that a table created as such
sql("create table test stored as parquet as select 1")
in Spark 2.1 cannot be read in previous versions of Spark.
Michael
> On Nov 29, 2016,
Hello,
When I try to read from a Hive table created by Spark 2.1 in Spark 2.0 or
earlier, I get an error:
java.lang.ClassNotFoundException: Failed to load class for data source: hive.
Is there a way to get previous versions of Spark to read tables written with
Spark 2.1?
Cheers,
Michael
Hi,
I am interested to contribute code on Spark. Could you please add me into
the mailing list / DL.
Thanks,
*Srinivas Potluri*
I'll send a -1 because of SPARK-18546. Haven't looked at anything else yet.
On Mon, Nov 28, 2016 at 5:25 PM, Reynold Xin wrote:
> Please vote on releasing the following candidate as Apache Spark version
> 2.1.0. The vote is open until Thursday, December 1, 2016 at 18:00 UTC and
> passes if a majo
We still have several blockers for 2.1, so I imagine at least one will mean
this won't be the final RC:
SPARK-18318 ML, Graph 2.1 QA: API: New Scala APIs, docs
SPARK-18319 ML, Graph 2.1 QA: API: Experimental, DeveloperApi, final,
sealed audit
SPARK-18326 SparkR 2.1 QA: New R APIs and API docs
SPAR
10 matches
Mail list logo