Hi Alek, Just a follow up question. This is what I did in sparkR shell:
lines <- SparkR:::textFile(sc, "./README.md") head(lines) And I am getting error: "Error in x[seq_len(n)] : object of type 'S4' is not subsettable" I'm wondering what did I do wrong. Thanks in advance. Wei 2015-06-25 13:44 GMT-07:00 Wei Zhou <zhweisop...@gmail.com>: > Hi Alek, > > Thanks for the explanation, it is very helpful. > > Cheers, > Wei > > 2015-06-25 13:40 GMT-07:00 Eskilson,Aleksander <alek.eskil...@cerner.com>: > >> Hi there, >> >> The tutorial you’re reading there was written before the merge of >> SparkR for Spark 1.4.0 >> For the merge, the RDD API (which includes the textFile() function) was >> made private, as the devs felt many of its functions were too low level. >> They focused instead on finishing the DataFrame API which supports local, >> HDFS, and Hive/HBase file reads. In the meantime, the devs are trying to >> determine which functions of the RDD API, if any, should be made public >> again. You can see the rationale behind this decision on the issue’s JIRA >> [1]. >> >> You can still make use of those now private RDD functions by prepending >> the function call with the SparkR private namespace, for example, you’d use >> SparkR:::textFile(…). >> >> Hope that helps, >> Alek >> >> [1] -- https://issues.apache.org/jira/browse/SPARK-7230 >> >> From: Wei Zhou <zhweisop...@gmail.com> >> Date: Thursday, June 25, 2015 at 3:33 PM >> To: "user@spark.apache.org" <user@spark.apache.org> >> Subject: sparkR could not find function "textFile" >> >> Hi all, >> >> I am exploring sparkR by activating the shell and following the >> tutorial here https://amplab-extras.github.io/SparkR-pkg/ >> <https://urldefense.proofpoint.com/v2/url?u=https-3A__amplab-2Dextras.github.io_SparkR-2Dpkg_&d=AwMFaQ&c=NRtzTzKNaCCmhN_9N2YJR-XrNU1huIgYP99yDsEzaJo&r=0vZw1rBdgaYvDJYLyKglbrax9kvQfRPdzxLUyWSyxPM&m=aL4A2Pv9tHbhgJUX-EnuYx2HntTnrqVpegm6Ag-FwnQ&s=qfOET1UvP0ECAKgnTJw8G13sFTi_PhiJ8Q89fMSgH_Q&e=> >> >> And when I tried to read in a local file with textFile(sc, >> "file_location"), it gives an error could not find function "textFile". >> >> By reading through sparkR doc for 1.4, it seems that we need sqlContext >> to import data, for example. >> >> people <- read.df(sqlContext, "./examples/src/main/resources/people.json", >> "json" >> >> ) >> And we need to specify the file type. >> >> My question is does sparkR stop supporting general type file importing? >> If not, would appreciate any help on how to do this. >> >> PS, I am trying to recreate the word count example in sparkR, and want >> to import README.md file, or just any file into sparkR. >> >> Thanks in advance. >> >> Best, >> Wei >> >> CONFIDENTIALITY NOTICE This message and any included attachments are >> from Cerner Corporation and are intended only for the addressee. The >> information contained in this message is confidential and may constitute >> inside or non-public information under international, federal, or state >> securities laws. Unauthorized forwarding, printing, copying, distribution, >> or use of such information is strictly prohibited and may be unlawful. If >> you are not the addressee, please promptly delete this message and notify >> the sender of the delivery error by e-mail or you may call Cerner's >> corporate offices in Kansas City, Missouri, U.S.A at (+1) (816)221-1024. >> > >