Is TorrentBroadcastFactory out of beta? IS it preferred over
HttpBroadcastFactory for large broadcasts?

What are the benefits of HttpBroadcastFactory as the default factory?


On Wed, Mar 12, 2014 at 7:09 PM, Stephen Boesch <java...@gmail.com> wrote:

> Hi Josh,
>   So then   2^31 (2.2Bilion) * 2^6  (length of double)  = 128GB  would be
> max array byte length with Doubles?
>
>
> 2014-03-12 11:30 GMT-07:00 Josh Marcus <jmar...@meetup.com>:
>
> Aureliano,
>>
>> Just to answer your second question (unrelated to Spark), arrays in java
>> and scala can't be larger than the maximum value of an Integer
>> (Integer.MAX_VALUE), which means that arrays are limited to about 2.2
>> billion elements.
>>
>> --j
>>
>>
>>
>> On Wed, Mar 12, 2014 at 1:08 PM, Aureliano Buendia 
>> <buendia...@gmail.com>wrote:
>>
>>> Hi,
>>>
>>> I asked a similar question a while ago, didn't get any answers.
>>>
>>> I'd like to share a 10 gb double array between 50 to 100 workers. The
>>> physical memory of workers is over 40 gb, so it can fit in each memory. The
>>> reason I'm sharing this array is that a cartesian operation is applied to
>>> this array, and I want to avoid network shuffling.
>>>
>>> 1. Is Spark broadcast built for pushing variables of gb size? Does it
>>> need special configurations (eg akka config, etc) to work under this
>>> condition?
>>>
>>> 2. (Not directly related to spark) Is the an upper limit for scala/java
>>> arrays other than the physical memory? Do they stop working when the array
>>> elements count exceeds a certain number?
>>>
>>
>>
>

Reply via email to