Below is a link to an example which Silvio Fiorito put together
demonstrating how to link Zeppelin with Spark Stream for real-time charts.
I think the original thread was pack in early November 2015, subject: Real
time chart in Zeppelin, if you care to try to find it.

https://gist.github.com/granturing/a09aed4a302a7367be92

HTH.

-Todd

On Sat, Mar 12, 2016 at 6:21 AM, Chris Miller <cmiller11...@gmail.com>
wrote:

> I'm pretty new to all of this stuff, so bare with me.
>
> Zeppelin isn't really intended for realtime dashboards as far as I know.
> Its reporting features (tables, graphs, etc.) are more for displaying the
> results from the output of something. As far as I know, there isn't really
> anything to "watch" a dataset and have updates pushed to the Zeppelin UI.
>
> As for Spark, unless you're doing a lot of processing that you didn't
> mention here, I don't think it's a good fit just for this.
>
> If it were me (just off the top of my head), I'd just build a simple web
> service that uses websockets to push updates to the client which could then
> be used to update graphs, tables, etc. The data itself -- that is, the
> accumulated totals -- you could store in something like Redis. When an
> order comes in, just add that quantity and price to the existing value and
> trigger your code to push out an updated value to any clients via the
> websocket. You could use something like a Redis pub/sub channel to trigger
> the web app to notify clients of an update.
>
> There are about 5 million other ways you could design this, but I would
> just keep it as simple as possible. I just threw one idea out...
>
> Good luck.
>
>
> --
> Chris Miller
>
> On Sat, Mar 12, 2016 at 6:58 PM, trung kien <kient...@gmail.com> wrote:
>
>> Thanks Chris and Mich for replying.
>>
>> Sorry for not explaining my problem clearly.  Yes i am talking about a
>> flexibke dashboard when mention Zeppelin.
>>
>> Here is the problem i am having:
>>
>> I am running a comercial website where we selle many products and we have
>> many branchs in many place. We have a lots of realtime transactions and
>> want to anaylyze it in realtime.
>>
>> We dont want every time doing analytics we have to aggregate every single
>> transactions ( each transaction have BranchID, ProductID, Qty, Price). So,
>> we maintain intermediate data which contains : BranchID, ProducrID,
>> totalQty, totalDollar
>>
>> Ideally, we have 2 tables:
>>    Transaction ( BranchID, ProducrID, Qty, Price, Timestamp)
>>
>> And intermediate table Stats is just sum of every transaction group by
>> BranchID and ProductID( i am using Sparkstreaming to calculate this table
>> realtime)
>>
>> My thinking is that doing statistics ( realtime dashboard)  on Stats
>> table is much easier, this table is also not enough for maintain.
>>
>> I'm just wondering, whats the best way to store Stats table( a database
>> or parquet file?)
>> What exactly are you trying to do? Zeppelin is for interactive analysis
>> of a dataset. What do you mean "realtime analytics" -- do you mean build a
>> report or dashboard that automatically updates as new data comes in?
>>
>>
>> --
>> Chris Miller
>>
>> On Sat, Mar 12, 2016 at 3:13 PM, trung kien <kient...@gmail.com> wrote:
>>
>>> Hi all,
>>>
>>> I've just viewed some Zeppenlin's videos. The intergration between
>>> Zeppenlin and Spark is really amazing and i want to use it for my
>>> application.
>>>
>>> In my app, i will have a Spark streaming app to do some basic realtime
>>> aggregation ( intermediate data). Then i want to use Zeppenlin to do some
>>> realtime analytics on the intermediate data.
>>>
>>> My question is what's the most efficient storage engine to store
>>> realtime intermediate data? Is parquet file somewhere is suitable?
>>>
>>
>>
>

Reply via email to