Hi I am trying to implement custom partitioner using this link http://stackoverflow.com/questions/23127329/how-to-define-custom-partitioner-for-spark-rdds-of-equally-sized-partition-where ( in link example key value is from 0 to (noOfElement - 1))
but not able to understand how i implement custom partitioner in my case: my parent RDD have 4 partition and RDD key is : TimeStamp and Value is JPEG Byte Array Regards Prateek On Tue, Nov 17, 2015 at 9:28 AM, Ted Yu <yuzhih...@gmail.com> wrote: > Please take a look at the following for example: > > ./core/src/main/scala/org/apache/spark/api/python/PythonPartitioner.scala > ./core/src/main/scala/org/apache/spark/Partitioner.scala > > Cheers > > On Tue, Nov 17, 2015 at 9:24 AM, prateek arora <prateek.arora...@gmail.com > > wrote: > >> Hi >> Thanks >> I am new in spark development so can you provide some help to write a >> custom partitioner to achieve this. >> if you have and link or example to write custom partitioner please >> provide to me. >> >> On Mon, Nov 16, 2015 at 6:13 PM, Sabarish Sasidharan < >> sabarish.sasidha...@manthan.com> wrote: >> >>> You can write your own custom partitioner to achieve this >>> >>> Regards >>> Sab >>> On 17-Nov-2015 1:11 am, "prateek arora" <prateek.arora...@gmail.com> >>> wrote: >>> >>>> Hi >>>> >>>> I have a RDD with 30 record ( Key/value pair ) and running 30 executor >>>> . i >>>> want to reparation this RDD in to 30 partition so every partition get >>>> one >>>> record and assigned to one executor . >>>> >>>> when i used rdd.repartition(30) its repartition my rdd in 30 partition >>>> but >>>> some partition get 2 record , some get 1 record and some not getting any >>>> record . >>>> >>>> is there any way in spark so i can evenly distribute my record in all >>>> partition . >>>> >>>> Regards >>>> Prateek >>>> >>>> >>>> >>>> -- >>>> View this message in context: >>>> http://apache-spark-user-list.1001560.n3.nabble.com/how-can-evenly-distribute-my-records-in-all-partition-tp25394.html >>>> Sent from the Apache Spark User List mailing list archive at Nabble.com. >>>> >>>> --------------------------------------------------------------------- >>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >>>> For additional commands, e-mail: user-h...@spark.apache.org >>>> >>>> >> >