Neha:

thanks for your response. I figured it out. it seems that the new
hadoop-1.x default port is changed from 8020 to 9000, so I had to use port
9000 in my default fs setting

I have to put a line

defaultFS=hdfs://localhost:9000/

something like that into my camus.properties.


camus_...@googlegroups.com seems less active :)


thanks
Yang


On Tue, Sep 3, 2013 at 6:31 PM, Neha Narkhede <neha.narkh...@gmail.com>wrote:

> I'm sorry I'm not too familiar with the camus code base. Have you tried
> pinging camus_...@googlegroups.com ?
>
> Thanks,
> Neha
>
>
> On Tue, Sep 3, 2013 at 1:40 PM, Yang <teddyyyy...@gmail.com> wrote:
>
> > I tried to run camus with the properties file in the examples dir:
> >
> > java -cp ...... com.linkedin.camus.etl.kafka.CamusJob -P
> > myproperties.properties
> >
> >
> > then it says that my output dir does not exist:
> > ~/tools/camus/camus-etl-kafka$ java -cp
> > target/camus-etl-kafka-0.1.0-SNAPSHOT.jar
> > com.linkedin.camus.etl.kafka.CamusJob -P camus.properties
> > Starting Kafka ETL Job
> > The blacklisted topics: []
> > The whitelisted topics: []
> > Dir Destination set to: /camus/out
> > Getting the base paths.
> > The execution base path does not exist. Creating the directory
> > The history base path does not exist. Creating the directory.
> > Exception in thread "main" java.io.FileNotFoundException: File
> /camus/exec
> > does not exist.
> > at
> >
> >
> org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:397)
> >  at
> >
> >
> org.apache.hadoop.fs.FilterFileSystem.getFileStatus(FilterFileSystem.java:251)
> > at org.apache.hadoop.fs.FileSystem.getContentSummary(FileSystem.java:801)
> >  at com.linkedin.camus.etl.kafka.CamusJob.run(CamusJob.java:223)
> > at com.linkedin.camus.etl.kafka.CamusJob.run(CamusJob.java:556)
> >  at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
> > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79)
> >
> >
> >
> > I tried changing the dir to hdfs://localhost/camus/out   it says:
> > $ java -cp target/camus-etl-kafka-0.1.0-SNAPSHOT.jar
> > com.linkedin.camus.etl.kafka.CamusJob -P camus.properties
> > Starting Kafka ETL Job
> > The blacklisted topics: []
> > The whitelisted topics: []
> > Dir Destination set to: hdfs://localhost/camus/out
> > Getting the base paths.
> > Exception in thread "main" java.lang.IllegalArgumentException: Wrong FS:
> > hdfs://localhost/camus/exec, expected: file:///
> >  at org.apache.hadoop.fs.FileSystem.checkPath(FileSystem.java:381)
> > at
> >
> >
> org.apache.hadoop.fs.RawLocalFileSystem.pathToFile(RawLocalFileSystem.java:55)
> >  at
> > org.apache.hadoop.fs.LocalFileSystem.pathToFile(LocalFileSystem.java:61)
> > at org.apache.hadoop.fs.LocalFileSystem.exists(LocalFileSystem.java:51)
> >  at com.linkedin.camus.etl.kafka.CamusJob.run(CamusJob.java:211)
> > at com.linkedin.camus.etl.kafka.CamusJob.run(CamusJob.java:556)
> >  at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
> > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79)
> >
> >
> > so how can I let Camus know that this is in a hdfs environment, and the
> > code in CamusJob.java:140
> >
> >     private Job createJob(Properties props) throws IOException {
> >         Job job = new Job(getConf());
> >
> > get a conf that points to a hdfs setup?
> >
> > I already set my env HADOOP_CONF_DIR to my running hadoop
> >
> >
> > thanks
> > Yang
> >
>

Reply via email to