Yes. Kind regards, Michał Michalski, michal.michal...@boxever.com
On 24 April 2015 at 17:12, Jeetendra Gangele <gangele...@gmail.com> wrote: > you used ZipWithUniqueID? > > On 24 April 2015 at 21:28, Michal Michalski <michal.michal...@boxever.com> > wrote: > >> I somehow missed zipWithIndex (and Sean's email), thanks for hint. I mean >> - I saw it before, but I just thought it's not doing what I want. I've >> re-read the description now and it looks like it might be actually what I >> need. Thanks. >> >> Kind regards, >> Michał Michalski, >> michal.michal...@boxever.com >> >> On 24 April 2015 at 16:26, Ganelin, Ilya <ilya.gane...@capitalone.com> >> wrote: >> >>> To maintain the order you can use zipWithIndex as Sean Owen pointed >>> out. This is the same as zipWithUniqueId except the assigned number is the >>> index of the data in the RDD which I believe matches the order of data as >>> it's stored on HDFS. >>> >>> >>> >>> Sent with Good (www.good.com) >>> >>> >>> -----Original Message----- >>> *From: *Michal Michalski [michal.michal...@boxever.com] >>> *Sent: *Friday, April 24, 2015 11:18 AM Eastern Standard Time >>> *To: *Ganelin, Ilya >>> *Cc: *Spico Florin; user >>> *Subject: *Re: Does HadoopRDD.zipWithIndex method preserve the order of >>> the input data from Hadoop? >>> >>> I read it one by one as I need to maintain the order, but it doesn't >>> mean that I process them one by one later. Input lines refer to different >>> entities I update, so once I read them in order, I group them by the id of >>> the entity I want to update, sort the updates on per-entity basis and >>> process them further in parallel (including writing data to C* and Kafka at >>> the very end). That's what I use Spark for - the first step I ask about is >>> just a requirement related to the input format I get and need to support. >>> Everything what happens after that is just a normal data processing job >>> that you want to distribute. >>> >>> Kind regards, >>> Michał Michalski, >>> michal.michal...@boxever.com >>> >>> On 24 April 2015 at 16:10, Ganelin, Ilya <ilya.gane...@capitalone.com> >>> wrote: >>> >>>> If you're reading a file one by line then you should simply use Java's >>>> Hadoop FileSystem class to read the file with a BuffereInputStream. I don't >>>> think you need an RDD here. >>>> >>>> >>>> >>>> Sent with Good (www.good.com) >>>> >>>> >>>> -----Original Message----- >>>> *From: *Michal Michalski [michal.michal...@boxever.com] >>>> *Sent: *Friday, April 24, 2015 11:04 AM Eastern Standard Time >>>> *To: *Ganelin, Ilya >>>> *Cc: *Spico Florin; user >>>> *Subject: *Re: Does HadoopRDD.zipWithIndex method preserve the order >>>> of the input data from Hadoop? >>>> >>>> The problem I'm facing is that I need to process lines from input file >>>> in the order they're stored in the file, as they define the order of >>>> updates I need to apply on some data and these updates are not commutative >>>> so that order matters. Unfortunately the input is purely order-based, >>>> theres no timestamp per line etc. in the file and I'd prefer to avoid >>>> preparing the file in advance by adding ordinals before / after each line. >>>> From the approaches you suggested first two won't work as there's nothing I >>>> could sort by. I'm not sure about the third one - I'm just not sure what >>>> you meant there to be honest :-) >>>> >>>> Kind regards, >>>> Michał Michalski, >>>> michal.michal...@boxever.com >>>> >>>> On 24 April 2015 at 15:48, Ganelin, Ilya <ilya.gane...@capitalone.com> >>>> wrote: >>>> >>>>> Michael - you need to sort your RDD. Check out the shuffle >>>>> documentation on the Spark Programming Guide. It talks about this >>>>> specifically. You can resolve this in a couple of ways - either by >>>>> collecting your RDD and sorting it, using sortBy, or not worrying about >>>>> the >>>>> internal ordering. You can still extract elements in order by using a >>>>> filter with the zip if e.g RDD.filter(s => s._2 < 50).sortBy(_._1) >>>>> >>>>> >>>>> >>>>> Sent with Good (www.good.com) >>>>> >>>>> >>>>> >>>>> -----Original Message----- >>>>> *From: *Michal Michalski [michal.michal...@boxever.com] >>>>> *Sent: *Friday, April 24, 2015 10:41 AM Eastern Standard Time >>>>> *To: *Spico Florin >>>>> *Cc: *user >>>>> *Subject: *Re: Does HadoopRDD.zipWithIndex method preserve the order >>>>> of the input data from Hadoop? >>>>> >>>>> Of course after you do it, you probably want to call >>>>> repartition(somevalue) on your RDD to "get your paralellism back". >>>>> >>>>> Kind regards, >>>>> Michał Michalski, >>>>> michal.michal...@boxever.com >>>>> >>>>> On 24 April 2015 at 15:28, Michal Michalski < >>>>> michal.michal...@boxever.com> wrote: >>>>> >>>>>> I did a quick test as I was curious about it too. I created a file >>>>>> with numbers from 0 to 999, in order, line by line. Then I did: >>>>>> >>>>>> scala> val numbers = sc.textFile("./numbers.txt") >>>>>> scala> val zipped = numbers.zipWithUniqueId >>>>>> scala> zipped.foreach(i => println(i)) >>>>>> >>>>>> Expected result if the order was preserved would be something like: >>>>>> (0, 0), (1, 1) etc. >>>>>> Unfortunately, the output looks like this: >>>>>> >>>>>> (126,1) >>>>>> (223,2) >>>>>> (320,3) >>>>>> (1,0) >>>>>> (127,11) >>>>>> (2,10) >>>>>> (...) >>>>>> >>>>>> The workaround I found that works for me for my specific use case >>>>>> (relatively small input files) is setting explicitly the number of >>>>>> partitions to 1 when reading a single *text* file: >>>>>> >>>>>> scala> val numbers_sp = sc.textFile("./numbers.txt", 1) >>>>>> >>>>>> Than the output is exactly as I would expect. >>>>>> >>>>>> I didn't dive into the code too much, but I took a very quick look at >>>>>> it and figured out - correct me if I missed something, it's Friday >>>>>> afternoon! ;-) - that this workaround will work fine for all the input >>>>>> formats inheriting from org.apache.hadoop.mapred.FileInputFormat >>>>>> including >>>>>> TextInputFormat, of course - see the implementation of getSplits() method >>>>>> there ( >>>>>> http://grepcode.com/file/repo1.maven.org/maven2/org.jvnet.hudson.hadoop/hadoop-core/0.19.1-hudson-2/org/apache/hadoop/mapred/FileInputFormat.java#FileInputFormat.getSplits%28org.apache.hadoop.mapred.JobConf%2Cint%29 >>>>>> ). >>>>>> The numSplits variable passed there is exactly the same value as you >>>>>> provide as a second argument to textFile, which is minPartitions. >>>>>> However, >>>>>> while *min* suggests that we can only define a minimal number of >>>>>> partitions, while we have no control over the max, from what I can see in >>>>>> the code, that value specifies the *exact* number of partitions per the >>>>>> FileInputFormat.getSplits implementation. Of course it can differ for >>>>>> other >>>>>> input formats, but in this case it should work just fine. >>>>>> >>>>>> >>>>>> Kind regards, >>>>>> Michał Michalski, >>>>>> michal.michal...@boxever.com >>>>>> >>>>>> On 24 April 2015 at 14:05, Spico Florin <spicoflo...@gmail.com> >>>>>> wrote: >>>>>> >>>>>>> Hello! >>>>>>> I know that HadoopRDD partitions are built based on the number of >>>>>>> splits in HDFS. I'm wondering if these partitions preserve the initial >>>>>>> order of data in file. >>>>>>> As an example, if I have an HDFS (myTextFile) file that has these >>>>>>> splits: >>>>>>> >>>>>>> split 0-> line 1, ..., line k >>>>>>> split 1->line k+1,..., line k+n >>>>>>> splt 2->line k+n, line k+n+m >>>>>>> >>>>>>> and the code >>>>>>> val lines=sc.textFile("hdfs://mytextFile") >>>>>>> lines.zipWithIndex() >>>>>>> >>>>>>> will the order of lines preserved? >>>>>>> (line 1, zipIndex 1) , .. (line k, zipIndex k), and so one. >>>>>>> >>>>>>> I found this question on stackoverflow ( >>>>>>> http://stackoverflow.com/questions/26046410/how-can-i-obtain-an-element-position-in-sparks-rdd) >>>>>>> whose answer intrigued me: >>>>>>> "Essentially, RDD's zipWithIndex() method seems to do this, but it >>>>>>> won't preserve the original ordering of the data the RDD was created >>>>>>> from" >>>>>>> >>>>>>> Can you please confirm that is this the correct answer? >>>>>>> >>>>>>> Thanks. >>>>>>> Florin >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>> >>>>> ------------------------------ >>>>> >>>>> The information contained in this e-mail is confidential and/or >>>>> proprietary to Capital One and/or its affiliates. The information >>>>> transmitted herewith is intended only for use by the individual or entity >>>>> to which it is addressed. If the reader of this message is not the >>>>> intended recipient, you are hereby notified that any review, >>>>> retransmission, dissemination, distribution, copying or other use of, or >>>>> taking of any action in reliance upon this information is strictly >>>>> prohibited. If you have received this communication in error, please >>>>> contact the sender and delete the material from your computer. >>>>> >>>> >>>> >>>> ------------------------------ >>>> >>>> The information contained in this e-mail is confidential and/or >>>> proprietary to Capital One and/or its affiliates. The information >>>> transmitted herewith is intended only for use by the individual or entity >>>> to which it is addressed. If the reader of this message is not the >>>> intended recipient, you are hereby notified that any review, >>>> retransmission, dissemination, distribution, copying or other use of, or >>>> taking of any action in reliance upon this information is strictly >>>> prohibited. If you have received this communication in error, please >>>> contact the sender and delete the material from your computer. >>>> >>> >>> >>> ------------------------------ >>> >>> The information contained in this e-mail is confidential and/or >>> proprietary to Capital One and/or its affiliates. The information >>> transmitted herewith is intended only for use by the individual or entity >>> to which it is addressed. If the reader of this message is not the >>> intended recipient, you are hereby notified that any review, >>> retransmission, dissemination, distribution, copying or other use of, or >>> taking of any action in reliance upon this information is strictly >>> prohibited. If you have received this communication in error, please >>> contact the sender and delete the material from your computer. >>> >> >> > >