Hi,

Currently the TakeOrderedAndProject operator in spark sql uses RDD’s 
takeOrdered method. When we pass a large limit to operator, however, it will 
return partitionNum*limit number of records to the driver which may cause an 
OOM.

Are there any plans to deal with the problem in the community? 


Thanks.


Yang

Reply via email to