Hi All, I want to access a particular column of a DB table stored in a CSV format and perform some aggregate queries over it. I wrote the following query in scala as a first step.
*var add=(x:String)=>x.split("\\s+)(2).toInt* *var result=List[Int]()* *input.split("\\n").foreach(x=>result::=add(x)) * *[Queries:]result.max/min/filter/sum...* But is there an efficient way/in-built function to access a particular column value or entire column in Spark ? Because built-in implementation might be efficient ! Thanks.