g Runtime is only available to Databricks
customers. This seems in conflict with the community efforts described here.
Can you comment on behalf of Databricks?
Look forward to your response, joseph.
See you all soon!
—
Chris Fregly
Founder @ PipelineAI <https://pipeline.ai/> (100,000 User
@reynold:
Databricks runs their proprietary product on Kubernetes. how about
contributing some of that work back to the Open Source Community?
—
Chris Fregly
Founder and Research Engineer @ PipelineAI <http://pipeline.io/>
Founder @ Advanced Spark and TensorFlow Meetup
<http://www.m
they’re presented on stage - especially the proprietary demos involving
dbutils() and display().
Chris Fregly
Research Scientist @ PipelineIO
Founder @ Advanced Spark and TensorFlow Meetup
San Francisco - Chicago - Washington DC - London
On Feb 15, 2017, 12:14 PM -0800, Nicholas Chammas ,
wrote
i seem to remember a large spark user (tencent, i believe) chiming in late
during these discussions 6-12 months ago and squashing any sort of deprecation
given the massive effort that would be required to upgrade their environment.
i just want to make sure these convos take into consideration la
ection and reinventing the wheel. not quite sure
where that project will go. doesn't seem like it will have a long shelf-life
in my opinion.
check out pipeline.io. some cool stuff in there.
> On Aug 11, 2016, at 9:35 AM, Chris Fregly wrote:
>
> this is exactly what my http://pipeline.
this is exactly what my http://pipeline.io project is addressing. check it out
and send me feedback or create issues at that github location.
> On Aug 11, 2016, at 7:42 AM, Nicholas Chammas
> wrote:
>
> Thanks Michael for the reference, and thanks Nick for the comprehensive
> overview of exi
nditions placed on
>> the package. If you find that the general public are downloading such test
>> packages, then remove them.
>>
>
> On Tue, Aug 9, 2016 at 11:32 AM, Chris Fregly wrote:
>
>> this is a valid question. there are many people building products and
>&
, also, there is no python support, no samples on the pr
> > demonstrating how to use security capabilities and no documentation
> updates.
> >
> > Thanks
> >
> > --
> > Luciano Resende
> > http://twitter.com/lresende1975
> > http://lresende.blogspot.c
+1 on bringing it back. causing all sorts of problems on my end that was not
obvious without digging in
I was having problems building spark, as well, with the --hive-thriftserver
flag. also thought I was doing something wrong on my end.
> On Jun 13, 2016, at 9:11 PM, Reynold Xin wrote:
>
>
>
> >> Project Name
> >>
> >> It would be ideal if we could have a project name that shows close ties
> to
> >> Spark (e.g. Spark Extras or Spark Connectors) but we will need
> permission
> >> and support from whoever is going to evaluate the project proposal (e.g.
> >> Apache Board)
> >>
> >>
> >> Thoughts ?
> >>
> >> Does anyone have any big disagreement or objection to moving into this
> >> direction ?
> >>
> >> Otherwise, who would be interested in joining the project, so I can
> start
> >> working on some concrete proposal ?
> >>
> >>
> >
> >
> >
> >
> > --
> > Luciano Resende
> > http://twitter.com/lresende1975
> > http://lresende.blogspot.com/
>
> -
> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
> For additional commands, e-mail: dev-h...@spark.apache.org
>
>
--
*Chris Fregly*
Principal Data Solutions Engineer
IBM Spark Technology Center, San Francisco, CA
http://spark.tc | http://advancedspark.com
perhaps renaming to Spark ML would actually clear up code and documentation
confusion?
+1 for rename
> On Apr 5, 2016, at 7:00 PM, Reynold Xin wrote:
>
> +1
>
> This is a no brainer IMO.
>
>
>> On Tue, Apr 5, 2016 at 7:32 PM, Joseph Bradley wrote:
>> +1 By the way, the JIRA for tracking
hey vadim-
sorry for the delay.
if you're interested in trying to get Kinesis working one-on-one, shoot me
a direct email and we'll get it going off-list.
we can circle back and summarize our findings here.
lots of people are using Spark Streaming+Kinesis successfully.
would love to help you t
and even the same process where the data might be cached.
these are the different locality levels:
PROCESS_LOCAL
NODE_LOCAL
RACK_LOCAL
ANY
relevant code:
https://github.com/apache/spark/blob/7712e724ad69dd0b83754e938e9799d13a4d43b9/core/src/test/scala/org/apache/spark/scheduler/TaskSetManagerSu
13 matches
Mail list logo