Have you tried by using the command?

flink-1.12.2 ./bin/flink run examples/streaming/WordCount.jar --input
file://(file location) --output file://(file location)

for example:

./bin/flink run
/Users/abubakarsiddiqurrahman/Documents/flink/examples/streaming/WordCount.jar
--input 
file:///Users/abubakarsiddiqurrahman/Documents/flink/examples/streaming/input.txt
--output 
file:///Users/abubakarsiddiqurrahman/Documents/flink/examples/streaming/streamingoutput

I think it will be worked in that case.

Thank you




On Sat, Apr 17, 2021 at 1:17 PM Robert Metzger <rmetz...@apache.org> wrote:

> Thanks a lot for the logs. I filed a ticket to track the issue:
> https://issues.apache.org/jira/browse/FLINK-22331
> I hope somebody with M1 hardware will soon have time to look into it.
>
> On Fri, Apr 16, 2021 at 11:02 AM Klemens Muthmann <
> klemens.muthm...@cyface.de> wrote:
>
>> Hi,
>>
>> I’ve appended you two log files. One is from a run without the Rosetta 2
>> compatibility layer while the other is with. As I said it would be great if
>> everything works without Rosetta 2, but at first it might be sufficient to
>> make it work with the compatibility layer.
>>
>> Regards
>>      Klemens
>>
>>
>> Am 15.04.2021 um 21:29 schrieb Robert Metzger <rmetz...@apache.org>:
>>
>> Hi,
>>
>> a DEBUG log of the client would indeed be nice.
>> Can you adjust this file:
>>
>> conf/log4j-cli.properties
>>
>> to the following contents: (basically TRACE logging with netty logs
>> enabled)
>>
>>
>>
>> ################################################################################
>> #  Licensed to the Apache Software Foundation (ASF) under one
>> #  or more contributor license agreements.  See the NOTICE file
>> #  distributed with this work for additional information
>> #  regarding copyright ownership.  The ASF licenses this file
>> #  to you under the Apache License, Version 2.0 (the
>> #  "License"); you may not use this file except in compliance
>> #  with the License.  You may obtain a copy of the License at
>> #
>> #      http://www.apache.org/licenses/LICENSE-2.0
>> #
>> #  Unless required by applicable law or agreed to in writing, software
>> #  distributed under the License is distributed on an "AS IS" BASIS,
>> #  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
>> implied.
>> #  See the License for the specific language governing permissions and
>> # limitations under the License.
>>
>> ################################################################################
>>
>> # Allows this configuration to be modified at runtime. The file will be
>> checked every 30 seconds.
>> monitorInterval=30
>>
>> rootLogger.level = TRACE
>> rootLogger.appenderRef.file.ref = FileAppender
>>
>> # Log all infos in the given file
>> appender.file.name = FileAppender
>> appender.file.type = FILE
>> appender.file.append = false
>> appender.file.fileName = ${sys:log.file}
>> appender.file.layout.type = PatternLayout
>> appender.file.layout.pattern = %d{yyyy-MM-dd HH:mm:ss,SSS} %-5p %-60c %x
>> - %m%n
>>
>> # Log output from org.apache.flink.yarn to the console. This is used by
>> the
>> # CliFrontend class when using a per-job YARN cluster.
>> logger.yarn.name = org.apache.flink.yarn
>> logger.yarn.level = INFO
>> logger.yarn.appenderRef.console.ref = ConsoleAppender
>> logger.yarncli.name = org.apache.flink.yarn.cli.FlinkYarnSessionCli
>> logger.yarncli.level = INFO
>> logger.yarncli.appenderRef.console.ref = ConsoleAppender
>> logger.hadoop.name = org.apache.hadoop
>> logger.hadoop.level = INFO
>> logger.hadoop.appenderRef.console.ref = ConsoleAppender
>>
>> # Make sure hive logs go to the file.
>> logger.hive.name = org.apache.hadoop.hive
>> logger.hive.level = INFO
>> logger.hive.additivity = false
>> logger.hive.appenderRef.file.ref = FileAppender
>>
>> # Log output from org.apache.flink.kubernetes to the console.
>> logger.kubernetes.name = org.apache.flink.kubernetes
>> logger.kubernetes.level = INFO
>> logger.kubernetes.appenderRef.console.ref = ConsoleAppender
>>
>> appender.console.name = ConsoleAppender
>> appender.console.type = CONSOLE
>> appender.console.layout.type = PatternLayout
>> appender.console.layout.pattern = %d{yyyy-MM-dd HH:mm:ss,SSS} %-5p %-60c
>> %x - %m%n
>>
>> # suppress the warning that hadoop native libraries are not loaded
>> (irrelevant for the client)
>> logger.hadoopnative.name = org.apache.hadoop.util.NativeCodeLoader
>> logger.hadoopnative.level = OFF
>>
>> # Suppress the irrelevant (wrong) warnings from the Netty channel handler
>> #logger.netty.name =
>> org.apache.flink.shaded.akka.org.jboss.netty.channel.DefaultChannelPipeline
>> #logger.netty.level = OFF
>>
>>
>>
>> And then submit a job locally, and send me the respective log file
>> (containing the "client" string in the file name).
>>
>> Thanks a lot, and stay healthy through the pandemic!
>>
>> Best,
>> Robert
>>
>>
>> On Thu, Apr 15, 2021 at 9:12 PM Klemens Muthmann <
>> klemens.muthm...@cyface.de> wrote:
>>
>>> Hi,
>>>
>>> Since kindergarden time is shortened due to the pandemic I only get four
>>> hours of work into each day and I am supposed to do eight. So unfortunately
>>> I will not be able to develop a fix at the moment. -.- I am happy to
>>> provide any debug log you need or test adaptations and provide fixes as
>>> pull requests. But I will sadly have no time to do any research into the
>>> problem. :( So for now I guess I will be using one of our Linux servers to
>>> test the Flink Pipelines until Silicon is supported.
>>>
>>> Nevertheless, thanks for your answer. If there is anything I can provide
>>> you to narrow down the problem, I am happy to help.
>>>
>>> Regards
>>>      Klemens
>>>
>>> Am 15.04.2021 um 20:59 schrieb Robert Metzger <rmetz...@apache.org>:
>>>
>>> Hey Klemens,
>>>
>>> I'm sorry that you are running into this. Looks like you are the first
>>> (of probably many people) who use Flink on a M1 chip.
>>>
>>> If you are up for it, we would really appreciate a fix for this issue,
>>> as a contribution to Flink.
>>> Maybe you can distill the problem into an integration test, so that you
>>> can look into fixes right from your IDE. (It seems that the RestClient is
>>> causing the problems. The client is used by the command line interface to
>>> upload the job to the cluster (that's not happening when executing the job
>>> from the IDE))
>>> My first guess is that a newer netty version might be required? Or maybe
>>> there's some DEBUG log output that's helpful in understanding the issue?
>>>
>>>
>>>
>>>
>>> On Tue, Apr 13, 2021 at 5:34 PM Klemens Muthmann <
>>> klemens.muthm...@cyface.de> wrote:
>>>
>>>> Hi,
>>>>
>>>> I've just tried to run the basic example for Apache Flink
>>>> <https://ci.apache.org/projects/flink/flink-docs-release-1.12/try-flink/local_installation.html>
>>>>  on
>>>> an Apple Mac Pro with the new M1 Processor. I only need this for
>>>> development purposes. The actual thing is going to run on a Linux server
>>>> later on, so I would not mind if it only runs using the Rosetta
>>>> compatibility layer. Unfortunately it failed with the following Stack 
>>>> Trace:
>>>>
>>>> flink-1.12.2 ./bin/flink run examples/streaming/WordCount.jar
>>>> Executing WordCount example with default input data set.
>>>> Use --input to specify file input.
>>>> Printing result to stdout. Use --output to specify output path.
>>>> WARNING: An illegal reflective access operation has occurred
>>>> WARNING: Illegal reflective access by 
>>>> org.apache.flink.api.java.ClosureCleaner 
>>>> (file:/Users/muthmann/Development/Flink/flink-1.12.2/lib/flink-dist_2.11-1.12.2.jar)
>>>>  to field java.lang.String.value
>>>> WARNING: Please consider reporting this to the maintainers of 
>>>> org.apache.flink.api.java.ClosureCleaner
>>>> WARNING: Use --illegal-access=warn to enable warnings of further illegal 
>>>> reflective access operations
>>>> WARNING: All illegal access operations will be denied in a future release
>>>>
>>>> ------------------------------------------------------------
>>>>  The program finished with the following exception:
>>>>
>>>> org.apache.flink.client.program.ProgramInvocationException: The main 
>>>> method caused an error: Failed to execute job 'Streaming WordCount'.
>>>>     at 
>>>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:366)
>>>>     at 
>>>> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:219)
>>>>     at 
>>>> org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:114)
>>>>     at 
>>>> org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:812)
>>>>     at org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:246)
>>>>     at 
>>>> org.apache.flink.client.cli.CliFrontend.parseAndRun(CliFrontend.java:1054)
>>>>     at 
>>>> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:1132)
>>>>     at 
>>>> org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:28)
>>>>     at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1132)
>>>> Caused by: org.apache.flink.util.FlinkException: Failed to execute job 
>>>> 'Streaming WordCount'.
>>>>     at 
>>>> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.executeAsync(StreamExecutionEnvironment.java:1918)
>>>>     at 
>>>> org.apache.flink.client.program.StreamContextEnvironment.executeAsync(StreamContextEnvironment.java:135)
>>>>     at 
>>>> org.apache.flink.client.program.StreamContextEnvironment.execute(StreamContextEnvironment.java:76)
>>>>     at 
>>>> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:1782)
>>>>     at 
>>>> org.apache.flink.streaming.examples.wordcount.WordCount.main(WordCount.java:97)
>>>>     at 
>>>> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native 
>>>> Method)
>>>>     at 
>>>> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>>>     at 
>>>> java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>>     at java.base/java.lang.reflect.Method.invoke(Method.java:566)
>>>>     at 
>>>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:349)
>>>>     ... 8 more
>>>> Caused by: org.apache.flink.runtime.client.JobSubmissionException: Failed 
>>>> to submit JobGraph.
>>>>     at 
>>>> org.apache.flink.client.program.rest.RestClusterClient.lambda$submitJob$7(RestClusterClient.java:400)
>>>>     at 
>>>> java.base/java.util.concurrent.CompletableFuture.uniExceptionally(CompletableFuture.java:986)
>>>>     at 
>>>> java.base/java.util.concurrent.CompletableFuture$UniExceptionally.tryFire(CompletableFuture.java:970)
>>>>     at 
>>>> java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
>>>>     at 
>>>> java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
>>>>     at 
>>>> org.apache.flink.runtime.concurrent.FutureUtils.lambda$retryOperationWithDelay$9(FutureUtils.java:390)
>>>>     at 
>>>> java.base/java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:859)
>>>>     at 
>>>> java.base/java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:837)
>>>>     at 
>>>> java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
>>>>     at 
>>>> java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
>>>>     at 
>>>> org.apache.flink.runtime.rest.RestClient$ClientHandler.channelInactive(RestClient.java:588)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:262)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:248)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:241)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.ChannelInboundHandlerAdapter.channelInactive(ChannelInboundHandlerAdapter.java:81)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.handler.timeout.IdleStateHandler.channelInactive(IdleStateHandler.java:277)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:262)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:248)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:241)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.handler.stream.ChunkedWriteHandler.channelInactive(ChunkedWriteHandler.java:138)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:262)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:248)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:241)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.ChannelInboundHandlerAdapter.channelInactive(ChannelInboundHandlerAdapter.java:81)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.handler.codec.MessageAggregator.channelInactive(MessageAggregator.java:438)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:262)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:248)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:241)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.CombinedChannelDuplexHandler$DelegatingChannelHandlerContext.fireChannelInactive(CombinedChannelDuplexHandler.java:418)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.handler.codec.ByteToMessageDecoder.channelInputClosed(ByteToMessageDecoder.java:389)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.handler.codec.ByteToMessageDecoder.channelInactive(ByteToMessageDecoder.java:354)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.handler.codec.http.HttpClientCodec$Decoder.channelInactive(HttpClientCodec.java:288)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.CombinedChannelDuplexHandler.channelInactive(CombinedChannelDuplexHandler.java:221)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:262)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:248)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:241)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline$HeadContext.channelInactive(DefaultChannelPipeline.java:1405)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:262)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:248)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline.fireChannelInactive(DefaultChannelPipeline.java:901)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannel$AbstractUnsafe$8.run(AbstractChannel.java:818)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:164)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:472)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:500)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989)
>>>>     at 
>>>> org.apache.flink.shaded.netty4.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
>>>>     at java.base/java.lang.Thread.run(Thread.java:834)
>>>> Caused by: org.apache.flink.runtime.concurrent.FutureUtils$RetryException: 
>>>> Could not complete the operation. Number of retries has been exhausted.
>>>>     at 
>>>> org.apache.flink.runtime.concurrent.FutureUtils.lambda$retryOperationWithDelay$9(FutureUtils.java:386)
>>>>     ... 41 more
>>>> Caused by: java.util.concurrent.CompletionException: 
>>>> org.apache.flink.runtime.rest.ConnectionClosedException: Channel became 
>>>> inactive.
>>>>     at 
>>>> java.base/java.util.concurrent.CompletableFuture.encodeRelay(CompletableFuture.java:367)
>>>>     at 
>>>> java.base/java.util.concurrent.CompletableFuture.completeRelay(CompletableFuture.java:376)
>>>>     at 
>>>> java.base/java.util.concurrent.CompletableFuture$UniRelay.tryFire(CompletableFuture.java:1019)
>>>>     ... 39 more
>>>> Caused by: org.apache.flink.runtime.rest.ConnectionClosedException: 
>>>> Channel became inactive.
>>>>     ... 37 more
>>>>
>>>> Without Rosetta I get:
>>>>
>>>> ./bin/flink run ./examples/streaming/WordCount.jar
>>>> Executing WordCount example with default input data set.
>>>> Use --input to specify file input.
>>>> Printing result to stdout. Use --output to specify output path.
>>>> WARNING: An illegal reflective access operation has occurred
>>>> WARNING: Illegal reflective access by
>>>> org.apache.flink.api.java.ClosureCleaner
>>>> (file:/Users/muthmann/Development/Flink/flink-1.12.2/lib/flink-dist_2.11-1.12.2.jar)
>>>> to field java.lang.String.value
>>>> WARNING: Please consider reporting this to the maintainers of
>>>> org.apache.flink.api.java.ClosureCleaner
>>>> WARNING: Use --illegal-access=warn to enable warnings of further
>>>> illegal reflective access operations
>>>> WARNING: All illegal access operations will be denied in a future
>>>> release
>>>>
>>>> ------------------------------------------------------------
>>>>  The program finished with the following exception:
>>>>
>>>> org.apache.flink.client.program.ProgramInvocationException: The main
>>>> method caused an error: Failed to execute job 'Streaming WordCount'.
>>>> at
>>>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:366)
>>>> at
>>>> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:219)
>>>> at
>>>> org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:114)
>>>> at
>>>> org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:812)
>>>> at org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:246)
>>>> at
>>>> org.apache.flink.client.cli.CliFrontend.parseAndRun(CliFrontend.java:1054)
>>>> at
>>>> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:1132)
>>>> at
>>>> org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:28)
>>>> at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1132)
>>>> Caused by: org.apache.flink.util.FlinkException: Failed to execute job
>>>> 'Streaming WordCount'.
>>>> at
>>>> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.executeAsync(StreamExecutionEnvironment.java:1918)
>>>> at
>>>> org.apache.flink.client.program.StreamContextEnvironment.executeAsync(StreamContextEnvironment.java:135)
>>>> at
>>>> org.apache.flink.client.program.StreamContextEnvironment.execute(StreamContextEnvironment.java:76)
>>>> at
>>>> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:1782)
>>>> at
>>>> org.apache.flink.streaming.examples.wordcount.WordCount.main(WordCount.java:97)
>>>> at
>>>> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native
>>>> Method)
>>>> at
>>>> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>>> at
>>>> java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>> at java.base/java.lang.reflect.Method.invoke(Method.java:566)
>>>> at
>>>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:349)
>>>> ... 8 more
>>>> Caused by: org.apache.flink.runtime.client.JobSubmissionException:
>>>> Failed to submit JobGraph.
>>>> at
>>>> org.apache.flink.client.program.rest.RestClusterClient.lambda$submitJob$7(RestClusterClient.java:400)
>>>> at
>>>> java.base/java.util.concurrent.CompletableFuture.uniExceptionally(CompletableFuture.java:986)
>>>> at
>>>> java.base/java.util.concurrent.CompletableFuture$UniExceptionally.tryFire(CompletableFuture.java:970)
>>>> at
>>>> java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
>>>> at
>>>> java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
>>>> at
>>>> org.apache.flink.runtime.concurrent.FutureUtils.lambda$retryOperationWithDelay$9(FutureUtils.java:390)
>>>> at
>>>> java.base/java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:859)
>>>> at
>>>> java.base/java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:837)
>>>> at
>>>> java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
>>>> at
>>>> java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
>>>> at
>>>> org.apache.flink.runtime.rest.RestClient$ClientHandler.exceptionCaught(RestClient.java:613)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeExceptionCaught(AbstractChannelHandlerContext.java:302)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeExceptionCaught(AbstractChannelHandlerContext.java:281)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireExceptionCaught(AbstractChannelHandlerContext.java:273)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.channel.CombinedChannelDuplexHandler$DelegatingChannelHandlerContext.fireExceptionCaught(CombinedChannelDuplexHandler.java:424)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.channel.ChannelHandlerAdapter.exceptionCaught(ChannelHandlerAdapter.java:92)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.channel.CombinedChannelDuplexHandler$1.fireExceptionCaught(CombinedChannelDuplexHandler.java:145)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.channel.ChannelInboundHandlerAdapter.exceptionCaught(ChannelInboundHandlerAdapter.java:143)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.channel.CombinedChannelDuplexHandler.exceptionCaught(CombinedChannelDuplexHandler.java:231)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeExceptionCaught(AbstractChannelHandlerContext.java:302)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeExceptionCaught(AbstractChannelHandlerContext.java:281)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireExceptionCaught(AbstractChannelHandlerContext.java:273)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline$HeadContext.exceptionCaught(DefaultChannelPipeline.java:1377)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeExceptionCaught(AbstractChannelHandlerContext.java:302)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeExceptionCaught(AbstractChannelHandlerContext.java:281)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline.fireExceptionCaught(DefaultChannelPipeline.java:907)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.handleReadException(AbstractNioByteChannel.java:125)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:174)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:714)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:650)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:576)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:493)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
>>>> at java.base/java.lang.Thread.run(Thread.java:834)
>>>> Caused by:
>>>> org.apache.flink.runtime.concurrent.FutureUtils$RetryException: Could not
>>>> complete the operation. Number of retries has been exhausted.
>>>> at
>>>> org.apache.flink.runtime.concurrent.FutureUtils.lambda$retryOperationWithDelay$9(FutureUtils.java:386)
>>>> ... 29 more
>>>> Caused by: java.util.concurrent.CompletionException:
>>>> java.io.IOException: Connection reset by peer
>>>> at
>>>> java.base/java.util.concurrent.CompletableFuture.encodeRelay(CompletableFuture.java:367)
>>>> at
>>>> java.base/java.util.concurrent.CompletableFuture.completeRelay(CompletableFuture.java:376)
>>>> at
>>>> java.base/java.util.concurrent.CompletableFuture$UniRelay.tryFire(CompletableFuture.java:1019)
>>>> ... 27 more
>>>> Caused by: java.io.IOException: Connection reset by peer
>>>> at java.base/sun.nio.ch.FileDispatcherImpl.read0(Native Method)
>>>> at java.base/sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:39)
>>>> at java.base/sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:276)
>>>> at java.base/sun.nio.ch.IOUtil.read(IOUtil.java:233)
>>>> at java.base/sun.nio.ch.IOUtil.read(IOUtil.java:223)
>>>> at
>>>> java.base/sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:358)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.buffer.PooledByteBuf.setBytes(PooledByteBuf.java:253)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.buffer.AbstractByteBuf.writeBytes(AbstractByteBuf.java:1133)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.channel.socket.nio.NioSocketChannel.doReadBytes(NioSocketChannel.java:350)
>>>> at
>>>> org.apache.flink.shaded.netty4.io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:148)
>>>> ... 7 more
>>>>
>>>> Funny thing is, I am able to execute a Flink Pipeline from within my
>>>> IntelliJ IDE, so I thought it should be possible somehow.
>>>>
>>>>
>>>> So my question is, did anyone get this to run? Is there any chance of
>>>> executing Flink jobs on Apple Silicon?
>>>>
>>>> I’ve posted the same question on StackOverflow
>>>> <https://stackoverflow.com/questions/67071254/running-apache-flink-1-12-jobs-on-apple-m1-silicon>.
>>>> So If you’d like some points there you are free to head over and post a
>>>> reply. ;)
>>>>
>>>> Thanks and Regards
>>>>      Klemens Muthmann
>>>>
>>>>
>>>>
>>>>
>>>
>>

-- 
Regards,
Abu Bakar Siddiqur Rahman

Reply via email to