Hi, Is it possible to maintain state inside a Spark map function? With Hadoop MapReduce, Mappers and Reducers are classes that can have their own state using instance variables. Can this be done with Spark? Are there any examples?
Most examples I have seen do a simple operating on the value passed into the map function and then pass it along to the reduce function. Thanks in advance. -Kevin