Hi guys We're using Zeppelin to do some analysis of log files (Cloudera Cluster, Zeppelin 0.7.1 currently) and we're experiencing that zeppelin tends to get really slow when notebooks / queries return large datasets.
* Is there a best practice on what amounts of data / query results zeppelin can handle? * And is there a way to increase the performance? o (This may even be actually browser specific?) As an example we'd like to be able to save a simple select timestamp, hostname, etc.. query, displayed in a table as a csv file. This will work fine, as long as the resultset is "small enough". Once a certain size is exceeded, it takes veeery long until the "save as" popup window appears (if it appears at all) We experience the same extremely slow behavior when large resultsets are used for charts - the notebooks become unusable (too slow, browser becomes irresponsive) How are you guys dealing with this? Thanks in advance Alex