Hi Alek,

Just a follow up question. This is what I did in sparkR shell:

lines <- SparkR:::textFile(sc, "./README.md")
head(lines)

And I am getting error:

"Error in x[seq_len(n)] : object of type 'S4' is not subsettable"

I'm wondering what did I do wrong. Thanks in advance.

Wei

2015-06-25 13:44 GMT-07:00 Wei Zhou <zhweisop...@gmail.com>:

> Hi Alek,
>
> Thanks for the explanation, it is very helpful.
>
> Cheers,
> Wei
>
> 2015-06-25 13:40 GMT-07:00 Eskilson,Aleksander <alek.eskil...@cerner.com>:
>
>>  Hi there,
>>
>>  The tutorial you’re reading there was written before the merge of
>> SparkR for Spark 1.4.0
>> For the merge, the RDD API (which includes the textFile() function) was
>> made private, as the devs felt many of its functions were too low level.
>> They focused instead on finishing the DataFrame API which supports local,
>> HDFS, and Hive/HBase file reads. In the meantime, the devs are trying to
>> determine which functions of the RDD API, if any, should be made public
>> again. You can see the rationale behind this decision on the issue’s JIRA
>> [1].
>>
>>  You can still make use of those now private RDD functions by prepending
>> the function call with the SparkR private namespace, for example, you’d use
>> SparkR:::textFile(…).
>>
>>  Hope that helps,
>> Alek
>>
>>  [1] -- https://issues.apache.org/jira/browse/SPARK-7230
>>
>>   From: Wei Zhou <zhweisop...@gmail.com>
>> Date: Thursday, June 25, 2015 at 3:33 PM
>> To: "user@spark.apache.org" <user@spark.apache.org>
>> Subject: sparkR could not find function "textFile"
>>
>>   Hi all,
>>
>>  I am exploring sparkR by activating the shell and following the
>> tutorial here https://amplab-extras.github.io/SparkR-pkg/
>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__amplab-2Dextras.github.io_SparkR-2Dpkg_&d=AwMFaQ&c=NRtzTzKNaCCmhN_9N2YJR-XrNU1huIgYP99yDsEzaJo&r=0vZw1rBdgaYvDJYLyKglbrax9kvQfRPdzxLUyWSyxPM&m=aL4A2Pv9tHbhgJUX-EnuYx2HntTnrqVpegm6Ag-FwnQ&s=qfOET1UvP0ECAKgnTJw8G13sFTi_PhiJ8Q89fMSgH_Q&e=>
>>
>>  And when I tried to read in a local file with textFile(sc,
>> "file_location"), it gives an error could not find function "textFile".
>>
>>  By reading through sparkR doc for 1.4, it seems that we need sqlContext
>> to import data, for example.
>>
>> people <- read.df(sqlContext, "./examples/src/main/resources/people.json", 
>> "json"
>>
>> )
>> And we need to specify the file type.
>>
>>  My question is does sparkR stop supporting general type file importing?
>> If not, would appreciate any help on how to do this.
>>
>>  PS, I am trying to recreate the word count example in sparkR, and want
>> to import README.md file, or just any file into sparkR.
>>
>>  Thanks in advance.
>>
>>  Best,
>> Wei
>>
>>    CONFIDENTIALITY NOTICE This message and any included attachments are
>> from Cerner Corporation and are intended only for the addressee. The
>> information contained in this message is confidential and may constitute
>> inside or non-public information under international, federal, or state
>> securities laws. Unauthorized forwarding, printing, copying, distribution,
>> or use of such information is strictly prohibited and may be unlawful. If
>> you are not the addressee, please promptly delete this message and notify
>> the sender of the delivery error by e-mail or you may call Cerner's
>> corporate offices in Kansas City, Missouri, U.S.A at (+1) (816)221-1024.
>>
>
>

Reply via email to