Satish Can you post the SQL query you are using?
The SQL query must have 2 placeholders and both of them should be an inclusive range (<= and >=).. e.g. select title, author from books where ? <= id and id <= ? Are you doing this? Deenar On 23 September 2015 at 20:18, Deenar Toraskar < deenar.toras...@thinkreactive.co.uk> wrote: > Satish > > Can you post the SQL query you are using? > > The SQL query must have 2 placeholders and both of them should be an > inclusive range (<= and >=).. > > e.g. select title, author from books where ? <= id and id <= ? > > Are you doing this? > > Deenar > > > > > *Think Reactive Ltd* > deenar.toras...@thinkreactive.co.uk > 07714140812 > > > > On 23 September 2015 at 13:47, satish chandra j <jsatishchan...@gmail.com> > wrote: > >> HI, >> Could anybody provide inputs if they have came across similar issue >> >> @Rishitesh >> Could you provide if any sample code to use JdbcRDDSuite >> >> >> Regards, >> Satish Chandra >> >> On Wed, Sep 23, 2015 at 5:14 PM, Rishitesh Mishra < >> rishi80.mis...@gmail.com> wrote: >> >>> I am using Spark 1.5. I always get count = 100, irrespective of num >>> partitions. >>> >>> On Wed, Sep 23, 2015 at 5:00 PM, satish chandra j < >>> jsatishchan...@gmail.com> wrote: >>> >>>> HI, >>>> Currently using Spark 1.2.2, could you please let me know correct >>>> results output count which you got it by using JdbcRDDSuite >>>> >>>> Regards, >>>> Satish Chandra >>>> >>>> On Wed, Sep 23, 2015 at 4:02 PM, Rishitesh Mishra < >>>> rishi80.mis...@gmail.com> wrote: >>>> >>>>> Which version of Spark you are using ?? I can get correct results >>>>> using JdbcRDD. Infact there is a test suite precisely for this ( >>>>> JdbcRDDSuite) . >>>>> I changed according to your input and got correct results from this >>>>> test suite. >>>>> >>>>> On Wed, Sep 23, 2015 at 11:00 AM, satish chandra j < >>>>> jsatishchan...@gmail.com> wrote: >>>>> >>>>>> HI All, >>>>>> >>>>>> JdbcRDD constructor has following parameters, >>>>>> >>>>>> *JdbcRDD >>>>>> <https://spark.apache.org/docs/1.2.0/api/java/org/apache/spark/rdd/JdbcRDD.html#JdbcRDD(org.apache.spark.SparkContext,%20scala.Function0,%20java.lang.String,%20long,%20long,%20int,%20scala.Function1,%20scala.reflect.ClassTag)>* >>>>>> (SparkContext >>>>>> <https://spark.apache.org/docs/1.2.0/api/java/org/apache/spark/SparkContext.html> >>>>>> sc, >>>>>> scala.Function0<java.sql.Connection> getConnection, String sql, *long >>>>>> lowerBound, >>>>>> long upperBound, int numPartitions*, >>>>>> scala.Function1<java.sql.ResultSet,T >>>>>> <https://spark.apache.org/docs/1.2.0/api/java/org/apache/spark/rdd/JdbcRDD.html>> >>>>>> mapRow, >>>>>> scala.reflect.ClassTag<T >>>>>> <https://spark.apache.org/docs/1.2.0/api/java/org/apache/spark/rdd/JdbcRDD.html> >>>>>> > evidence$1) >>>>>> >>>>>> where the below parameters *lowerBound* refers to Lower boundary of >>>>>> entire data, *upperBound *refers to Upper boundary of entire data >>>>>> and *numPartitions *refer to Number of partitions >>>>>> >>>>>> Source table to which JbdcRDD is fetching data from Oracle DB has >>>>>> more than 500 records but its confusing when I tried several executions >>>>>> by >>>>>> changing "numPartitions" parameter >>>>>> >>>>>> LowerBound,UpperBound,numPartitions: Output Count >>>>>> >>>>>> 0 ,100 ,1 : 100 >>>>>> >>>>>> 0 ,100 ,2 : 151 >>>>>> >>>>>> 0 ,100 ,3 : 201 >>>>>> >>>>>> >>>>>> Please help me in understanding the why Output count is 151 if >>>>>> numPartitions is 2 and Output count is 201 if numPartitions is 3 >>>>>> >>>>>> Regards, >>>>>> >>>>>> Satish Chandra >>>>>> >>>>> >>>>> >>>> >>> >> >