Yes, absolutely. Take a look at: https://spark.apache.org/docs/1.4.1/mllib-statistics.html#summary-statistics
On Fri, Aug 28, 2015 at 8:39 AM, ashensw <as...@wso2.com> wrote: > Hi all, > > I have a dataset which consist of large number of features(columns). It is > in csv format. So I loaded it into a spark dataframe. Then I converted it > into a JavaRDD<Row> Then using a spark transformation I converted that into > JavaRDD<String[]>. Then again converted it into a JavaRDD<double[]>. So now > I have a JavaRDD<double[]>. So is there any method to calculate max and min > values of each columns in this JavaRDD<double[]> ? > > Or Is there any way to access the array if I store max and min values to a > array inside the spark transformation class? > > Thanks. > > > > -- > View this message in context: > http://apache-spark-user-list.1001560.n3.nabble.com/Calculating-Min-and-Max-Values-using-Spark-Transformations-tp24491.html > Sent from the Apache Spark User List mailing list archive at Nabble.com. > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org > >