Following Hadoop conventions, Spark won't overwrite an existing directory.
You need to provide a unique output path every time you run the program, or
delete or rename the target directory before you run the job.

dean

Dean Wampler, Ph.D.
Author: Programming Scala, 2nd Edition
<http://shop.oreilly.com/product/0636920033073.do> (O'Reilly)
Typesafe <http://typesafe.com>
@deanwampler <http://twitter.com/deanwampler>
http://polyglotprogramming.com

On Mon, Aug 10, 2015 at 7:08 AM, Yasemin Kaya <godo...@gmail.com> wrote:

> Hi,
>
> I have EC2 cluster, and am using spark 1.3, yarn and HDFS . When i submit
> at local there is no problem , but i run at cluster, saveAsTextFile doesn't
> work."*It says me User class threw exception: Output directory
> hdfs://172.31.42.10:54310/...../weblogReadResult
> <http://172.31.42.10:54310/...../weblogReadResult> already exists*"
>
> Is there anyone can help me about this issue ?
>
> Best,
> yasemin
>
>
>
> --
> hiç ender hiç
>

Reply via email to