Re: scaling question

2015-06-19 Thread Fabian Hueske
Hi Bill, no worry, questions are the purpose of this mailing list. The number network buffers is a parameter that needs to be scaled with your setup. The reason for that is Flink's pipelined data transfer, which requires a certain number of network buffers to be available at the same time during

Re: scaling question

2015-06-19 Thread Bill Sparks
To clarify Š it's 64HT cores per node, 16 nodes each with 128GB. Well actually I have 48 nodes Š but trying to limit it so we have a comparison with Spark/MPI/MapReduce all at the same node count. Thanks for the information. -- Jonathan (Bill) Sparks Software Architecture Cray Inc. On 6/19/

Re: scaling question

2015-06-19 Thread Ufuk Celebi
PS: I've read your last email as 64 HT cores per machine. If it was in total over the 16 nodes, you have to adjust my response accordingly. ;) On 19 Jun 2015, at 16:42, Fabian Hueske wrote: > Hi Bill, > > no worry, questions are the purpose of this mailing list. > > The number network buffers

Re: scaling question

2015-06-19 Thread Ufuk Celebi
Hey Bill! On 19 Jun 2015, at 16:24, Bill Sparks wrote: > Sorry for the post again. I guess I'm not understanding this… Thanks for posting again, not sorry! ;-) Regarding the broken link: where did you get this link? I think it should be referring here: http://ci.apache.org/projects/flink/f

Re: scaling question

2015-06-19 Thread Maximilian Michels
Hi Bill, You're right. Simply increasing the task manager slots doesn't do anything. It is correct to set the parallelism to taskManagers*slots. Simply increase the number of network buffers in the flink-conf.yaml, e.g. to 4096. In the future, we will configure this setting dynamically. Let us kn

scaling question

2015-06-19 Thread Bill Sparks
Sorry for the post again. I guess I'm not understanding this… The question is how to scale up/increase the execution of a problem. What I'm trying to do, is get the best out of the available processors for a given node count and compare this against spark, using KMeans. For spark, one method

Re: Log messages - redirect

2015-06-19 Thread Till Rohrmann
If I’m not mistaken from the shown output, you’re talking about the stdout output of the client, right? This output is not controlled by the log4j.properties or logback.xml file. However, you can use any command line tool available on your platform to redirect the stdout. For example on a Linux sy

Re: Log messages - redirect

2015-06-19 Thread Chiwan Park
Hi, Flink uses slf4j and log4j for logging. You can override default log4j configuration programmatically. [1] Or you can use logback as a logging backend and override default logback configuration also. [2][3] [1] http://stackoverflow.com/a/9003191 [2] http://ci.apache.org/projects/flink/flink-

Log messages - redirect

2015-06-19 Thread Juan Fumero
Hi, is there any option (from API level) to redirect the log messages to a file instead of stdout? Thanks log4j:WARN No appenders could be found for logger (org.apache.flink.api.java.ExecutionEnvironment). log4j:WARN Please initialize the log4j system properly. log4j:WARN See http://logging.