Can you not create SparkContext inside the scheduler code? If you are
looking just to access hdfs then you can use the following object with it,
you can create/read/write files.

val hdfs = org.apache.hadoop.fs.FileSystem.get(new
URI("hdfs://localhost:9000"), hadoopConf)



Thanks
Best Regards

On Fri, Nov 14, 2014 at 9:12 PM, rapelly kartheek <kartheek.m...@gmail.com>
wrote:

> No. I am not accessing hdfs from either shell or a spark application. I
> want to access from spark "Scheduler code".
>
> I face an error when I use sc.textFile() as SparkContext wouldn't have
> been created yet. So, error says: "sc not found".
>
> On Fri, Nov 14, 2014 at 9:07 PM, Akhil Das <ak...@sigmoidanalytics.com>
> wrote:
>
>> like this?
>>
>> val file = sc.textFile("hdfs://localhost:9000/sigmoid/input.txt")
>>
>> Thanks
>> Best Regards
>>
>> On Fri, Nov 14, 2014 at 9:02 PM, rapelly kartheek <
>> kartheek.m...@gmail.com> wrote:
>>
>>> Hi,
>>> I am trying to read a HDFS file from Spark "scheduler code". I could
>>> find how to write hdfs read/writes in java.
>>>
>>> But I  need to access hdfs from spark using scala. Can someone please
>>> help me in this regard.
>>>
>>
>>
>

Reply via email to