Hi guys

We're using Zeppelin to do some analysis of log files (Cloudera Cluster, 
Zeppelin 0.7.1 currently) and we're experiencing that zeppelin tends to get 
really slow when notebooks / queries return large datasets.


*         Is there a best practice on what amounts of data / query results 
zeppelin can handle?

*         And is there a way to increase the performance?

o   (This may even be actually browser specific?)

As an example we'd like to be able to save a simple select timestamp, hostname, 
etc.. query, displayed in a table as a csv file. This will work fine, as long 
as the resultset is "small enough". Once a certain size is exceeded, it takes 
veeery long until the "save as" popup window appears (if it appears at all)

We experience the same extremely slow behavior when large resultsets are used 
for charts - the notebooks become unusable (too slow, browser becomes 
irresponsive)

How are you guys dealing with this?

Thanks in advance
Alex

Reply via email to