Hi,
I am looking for approaches to compare a row with the next one to
determine, e.g., differences in event-times/timestamps. I just found a
couple of solutions that use Window class, but that does not seem to
work on streaming data, such as
https://stackoverflow.com/questions/38229659/pyspark-co
d
dependency: org.apache.spark#spark-streaming-kafka-0-10_2.11_2.11;2.1.0:
not found Where do I find that a library? Thanks and best regards,
karamba PS: Does anybody know when python support becomes available in
spark-streaming-kafka-0-10
iner ips, it depends on your
> networking setup. You might want to try host networking so that the
> containers share the ip with the host.
>
> On Wed, Dec 28, 2016 at 1:46 AM, Karamba wrote:
>> Hi Sun Rui,
>>
>> thanks for answering!
>>
>>
>>> Al
ores.max to limit the cores to acquire,
> which means executors are available on a subset of the cluster nodes?
>
>> On Dec 27, 2016, at 01:39, Karamba wrote:
>>
>> Hi,
>>
>> I am running a couple of docker hosts, each with an HDFS and a spark
>> worker i
Hi,
I am running a couple of docker hosts, each with an HDFS and a spark
worker in a spark standalone cluster.
In order to get data locality awareness, I would like to configure Racks
for each host, so that a spark worker container knows from which hdfs
node container it should load its data. Does