Flink on YARN: frequent "container released on a *lost* node"

2019-11-21 Thread amran dean
Hello, I am frequently seeing this error in my jobmanager logs: 2019-11-18 09:07:08,863 INFO org.apache.flink.runtime.executiongraph.ExecutionGraph- Source: kdi_g2 -> (Sink: s_id_metadata, Sink: t_id_metadata) (23/24) (5e10e88814fe4ab0be5f554ec59bd93d) switched from RUNNING to FAILED. jav

Flink configuration at runtime

2019-11-18 Thread amran dean
Is it possible to configure certain settings at runtime, on a per-job basis rather than globally within flink-conf.yaml? For example, I have a job where it's desirable to retain a large number of checkpoints via state.checkpoints.num-retained. The checkpoints are cheap, and it's low cost. For oth

Does Flink merge streams deterministically?

2019-11-06 Thread amran dean
The Flink Kafka consumer's watermark extractor claims:. "The watermark extractor will run per Kafka partition, watermarks will be merged across partitions in the same way as in the Flink runtime, when streams are merged." Suppose we have simple situation where two streams merge into one. Is the o

Flink checkpointing behavior

2019-10-29 Thread amran dean
Hello, Exact semantics for checkpointing/task recovery are still a little confusing to me after parsing docs: so a few questions. - What does Flink consider a task failure? Is it any exception that the job does not handle? - Do the failure recovery strategies mentioned in https://ci.apache.org/pr

Flink Kafka->S3 exactly once guarantees

2019-10-23 Thread amran dean
Hello, Suppose I am using a *nondeterministic* time based partitioning scheme (e.g Flink processing time) to bucket S3 objects via the *BucketAssigner*, designated using *BulkFormatBuilder* for StreamingFileSink. Suppose that after an S3 MPU has completed, but *before* Flink internally commits (w

Flink StreamingFileSink part file behavior

2019-10-23 Thread amran dean
Hello, I am using StreamingFileSink, KafkaConsumer010 as a Kafka -> S3 connector (Flink 1.8.1, Kafka 0.10.1). The setup is simple: Data is written first bucketed by datetime (granularity of 1 day), then by kafka partition. I am using *event time* (Kafka timestamp, recorded at the time of creation

Re: Issue with BulkWriter

2019-10-22 Thread amran dean
); > } > > > https://ci.apache.org/projects/flink/flink-docs-master/api/java/org/apache/flink/api/common/serialization/BulkWriter.html#finish-- > > Regards, > Ravi > > On Tue, Oct 22, 2019 at 4:10 AM amran dean wrote: > >> Hello, >> I'm using BulkWri

Issue with BulkWriter

2019-10-21 Thread amran dean
Hello, I'm using BulkWriter to write newline-delimited, LZO-compressed files. The logic is very straightforward (See code below). I am experiencing an issue decompressing the created files created in this manner, consistently getting "lzop: unexpected end of file". Is this an issue with caller of

Customize Part file naming (Flink 1.9.0)

2019-10-18 Thread amran dean
Hello, StreamingFileSink's part file naming convention is not adjustable. It has form: *part--. * My use case for StreamingFileSink is a Kafka -> S3 pipeline, and files are read and processed from S3 using spark. In almost all cases, I want to compress raw data before writing to S3 using the BulkF

Flink S3 sink unable to compress data

2019-10-18 Thread amran dean
Hello, I am writing a custom S3 object encoder (code here: https://pastebin.com/raw/9Ag4ZVpX) used via: StreamingFileSink dataSink = StreamingFileSink .forRowFormat(new Path("s3a://some_path"), new RecordDataSerializer()) During execution, it does not produce any data in S3. This pipeline

Verifying correctness of StreamingFileSink (Kafka -> S3)

2019-10-15 Thread amran dean
I am evaluating StreamingFileSink (Kafka 0.10.11) as a production-ready alternative to a current Kafka -> S3 solution. Is there any way to verify the integrity of data written in S3? I'm confused how the file names (e.g part-1-17) map to Kafka partitions, and further unsure how to ensure that no K