Not quite sure which geo processing you're doing are they raster, vector? More
info will be appreciated for me to help you further.

Meanwhile I can try to give some hints, for instance, did you considered
GeoMesa <http://www.geomesa.org/2014/08/05/spark/>?
Since you need a WMS (or alike), did you considered GeoTrellis
<http://geotrellis.io/> (go to the batch processing)?

When you say SQLite, you mean that you're using Spatialite? Or your db is
not a geo one, and it's simple SQLite. In case you need an r-tree (or
related) index, you're headaches will come from congestion within your
database transaction... unless you go to a dedicated database like Vertica
(just mentioning)

kr,
andy



On Mon Dec 01 2014 at 2:49:44 PM Stadin, Benjamin <
benjamin.sta...@heidelberg-mobil.com> wrote:

> Hi all,
>
> I need some advise whether Spark is the right tool for my zoo. My
> requirements share commonalities with „big data“, workflow coordination and
> „reactive“ event driven data processing (as in for example Haskell Arrows),
> which doesn’t make it any easier to decide on a tool set.
>
> NB: I have asked a similar question on the Storm mailing list, but have
> been deferred to Spark. I previously thought Storm was closer to my needs –
> but maybe neither is.
>
> To explain my needs it’s probably best to give an example scenario:
>
>    - A user uploads small files (typically 1-200 files, file size
>    typically 2-10MB per file)
>    - Files should be converted in parallel and on available nodes. The
>    conversion is actually done via native tools, so there is not so much big
>    data processing required, but dynamic parallelization (so for example to
>    split the conversion step into as many conversion tasks as files are
>    available). The conversion typically takes between several minutes and a
>    few hours.
>    - The converted files gathered and are stored in a single database
>    (containing geometries for rendering)
>    - Once the db is ready, a web map server is (re-)configured and the
>    user can make small updates to the data set via a web UI.
>    - … Some other data processing steps which I leave away for brevity …
>    - There will be initially only a few concurrent users, but the system
>    shall be able to scale if needed
>
> My current thoughts:
>
>    - I should avoid to upload files into the distributed storage during
>    conversion, but probably should rather have each conversion filter download
>    the file it is actually converting from a shared place. Other wise it’s bad
>    for scalability reasons (too many redundant copies of same temporary files
>    if there are many concurrent users and many cluster nodes).
>    - Apache Oozie seems an option to chain together my pipes into a
>    workflow. But is it a good fit with Spark? What options do I have with
>    Spark to chain a workflow from pipes?
>    - Apache Crunch seems to make it easy to dynamically parallelize tasks
>    (Oozie itself can’t do this). But I may not need crunch after all if I have
>    Spark, and it also doesn’t seem to fit to my last problem following.
>    - The part that causes me the most headache is the user interactive db
>    update: I consider to use Kafka as message bus to broker between the web UI
>    and a custom db handler (nb, the db is a SQLite file). But how about
>    update responsiveness, isn’t it that Spark will cause some lags (as opposed
>    to Storm)?
>    - The db handler probably has to be implemented as a long running
>    continuing task, so when a user sends some changes the handler writes these
>    to the db file. However, I want this to be decoupled from the job. So file
>    these updates should be done locally only on the machine that started the
>    job for the whole lifetime of this user interaction. Does Spark allow to
>    create such long running tasks dynamically, so that when another (web) user
>    starts a new task a new long–running task is created and run on the same
>    node, which eventually ends and triggers the next task? Also, is it
>    possible to identify a running task, so that a long running task can be
>    bound to a session (db handler working on local db updates, until task
>    done), and eventually restarted / recreated on failure?
>
>
> ~Ben
>

Reply via email to