Hi Jamal, One nice feature of PySpark is that you can easily use existing functions from NumPy and SciPy inside your Spark code. For a simple example, the following uses Spark's cartesian operation (which combines pairs of vectors into tuples), followed by NumPy's corrcoef to compute the pearson correlation coefficient between every pair of a set of vectors. The vectors are an RDD of numpy arrays.
>> from numpy import array, corrcoef >> data = sc.parallelize([array([1,2,3]),array([2,4,6.1]),array([3,2,1.1])]) >> corrs = data.cartesian(data).map(lambda (x,y): corrcoef(x,y)[0,1]).collect() >> corrs [1.0, 0.99990086740991746, -0.99953863896044948 ... This just returns a list of the correlation coefficients, you could also add a key to each array, to keep track of which pair is which >> data_with_keys = sc.parallelize([(0,array([1,2,3])),(1,array([2,4,6.1])),(2,array([3,2,1.1]))]) >> corrs_with_keys = data_with_keys.cartesian(data_with_keys).map(lambda ((k1,v1),(k2,v2)): ((k1,k2),corrcoef(v1,v2)[0,1])).collect() >> corrs_with_keys [((0, 0), 1.0), ((0, 1), 0.99990086740991746), ((0, 2), -0.99953863896044948) ... Finally, you could just replace corrcoef in either of the above with scipy.spatial.distance.cosine to get your cosine similarity. Hope that's useful, as Andrei said, the answer partly depends on exactly what you're trying to do. -- Jeremy On Fri, May 23, 2014 at 2:41 PM, Andrei <faithlessfri...@gmail.com> wrote: > Do you need cosine distance and correlation between vectors or between > variables (elements of vector)? It would be helpful if you could tell us > details of your task. > > > On Thu, May 22, 2014 at 5:49 PM, jamal sasha <jamalsha...@gmail.com>wrote: > >> Hi, >> I have bunch of vectors like >> [0.1234,-0.231,0.23131] >> .... and so on. >> >> and I want to compute cosine similarity and pearson correlation using >> pyspark.. >> How do I do this? >> Any ideas? >> Thanks >> > >