I think this is a bug in the scala API. def writeToSocket(hostname : scala.Predef.String, port : java.lang.Integer, schema : org.apache.flink.streaming.util.serialization.SerializationSchema[T, scala.Array[scala.Byte]]) : org.apache.flink.streaming.api.datastream.DataStreamSink[T] = { /* compiled code */ }
On Tue, Jan 19, 2016 at 2:16 PM, Matthias J. Sax <mj...@apache.org> wrote: > It should work. > > Your error message indicates, that your DataStream is of type > [String,Array[Byte]] and not of type [String]. > > > Type mismatch, expected: SerializationSchema[String, Array[Byte]], > actual: SimpleStringSchema > > Can you maybe share your code? > > -Matthias > > On 01/19/2016 01:57 PM, Saiph Kappa wrote: > > It's DataStream[String]. So it seems that SimpleStringSchema cannot be > > used in writeToSocket regardless of the type of the DataStream. Right? > > > > On Tue, Jan 19, 2016 at 1:32 PM, Matthias J. Sax <mj...@apache.org > > <mailto:mj...@apache.org>> wrote: > > > > What type is your DataStream? It must be DataStream[String] to work > with > > SimpleStringSchema. > > > > If you have a different type, just implement a customized > > SerializationSchema. > > > > -Matthias > > > > > > On 01/19/2016 11:26 AM, Saiph Kappa wrote: > > > When I use SimpleStringSchema I get the error: Type mismatch, > expected: > > > SerializationSchema[String, Array[Byte]], actual: > SimpleStringSchema. I > > > think SimpleStringSchema extends SerializationSchema[String], and > > > therefore cannot be used as argument of writeToSocket. Can you > confirm > > > this please? > > > > > > s.writeToSocket(host, port.toInt, new SimpleStringSchema()) > > > > > > > > > Thanks. > > > > > > On Tue, Jan 19, 2016 at 10:34 AM, Matthias J. Sax < > mj...@apache.org <mailto:mj...@apache.org> > > > <mailto:mj...@apache.org <mailto:mj...@apache.org>>> wrote: > > > > > > There is SimpleStringSchema. > > > > > > -Matthias > > > > > > On 01/18/2016 11:21 PM, Saiph Kappa wrote: > > > > Hi Matthias, > > > > > > > > Thanks for your response. The method .writeToSocket seems to > be what I > > > > was looking for. Can you tell me what kind of serialization > schema > > > > should I use assuming my socket server receives strings. I > have > > > > something like this in scala: > > > > > > > > |val server =newServerSocket(9999)while(true){val s > =server.accept()val > > > > > in=newBufferedSource(s.getInputStream()).getLines()println(in.next())s.close()} > > > > > > > > | > > > > > > > > Thanks| > > > > | > > > > > > > > > > > > > > > > On Mon, Jan 18, 2016 at 8:46 PM, Matthias J. Sax < > mj...@apache.org <mailto:mj...@apache.org> <mailto:mj...@apache.org > > <mailto:mj...@apache.org>> > > > > <mailto:mj...@apache.org <mailto:mj...@apache.org> > > <mailto:mj...@apache.org <mailto:mj...@apache.org>>>> wrote: > > > > > > > > Hi Saiph, > > > > > > > > you can use AllWindowFunction via .apply(...) to get an > > > .collect method: > > > > > > > > From: > > > > > > > > > > https://ci.apache.org/projects/flink/flink-docs-release-0.10/apis/streaming_guide.html > > > > > > > > > // applying an AllWindowFunction on non-keyed window > > stream > > > > > allWindowedStream.apply (new > > > > AllWindowFunction<Tuple2<String,Integer>, Integer, > > Window>() { > > > > > public void apply (Window window, > > > > > Iterable<Tuple2<String, Integer>> values, > > > > > Collector<Integer> out) throws Exception { > > > > > int sum = 0; > > > > > for (value t: values) { > > > > > sum += t.f1; > > > > > } > > > > > out.collect (new Integer(sum)); > > > > > } > > > > > }); > > > > > > > > If you consume all those value via an sink, the sink > > will run > > > an the > > > > cluster. You can use .writeToSocket(...) as sink: > > > > > > > > > > https://ci.apache.org/projects/flink/flink-docs-release-0.10/apis/streaming_guide.html#data-sinks > > > > > > > > -Matthias > > > > > > > > > > > > On 01/18/2016 06:30 PM, Saiph Kappa wrote: > > > > > Hi, > > > > > > > > > > After performing a windowAll() on a > DataStream[String], is > > > there any > > > > > method to collect and return an array with all Strings > > > within a window > > > > > (similar to .collect in Spark). > > > > > > > > > > I basically want to ship all strings in a window to a > > remote > > > server > > > > > through a socket, and want to use the same socket > > connection > > > for all > > > > > strings that I send. The method .addSink iterates over > all > > > > records, but > > > > > does the provided function runs on the flink client or > on > > > the server? > > > > > > > > > > Thanks. > > > > > > > > > > > > > > > > > > > >