Hello Mich,
thanks for your reply.
As an engineer I can chip in. You may have partial execution and retries
meaning when spark encounters a *FetchFailedException*, it may retry
fetching the data from the unavailable (the one being rebooted) node a few
times before marking it permanently unavailab
Hi,
Your point -> "When Spark job shows FetchFailedException it creates few
duplicate data and we see few data also missing , please explain why. We
have scenario when spark job complains *FetchFailedException as one of the
data node got ** rebooted middle of job running ."*
As an engineer I ca
Hi Bhuwan et al,
Thank you for passing on the DataBricks Structured Streaming team's review
of the SPIP document. FYI, I work closely with Pawan and other members to
help deliver this piece of work. We appreciate your insights, especially
regarding the cost savings potential from the PoC.
Pavan a
+1
Nivi
Thanks Bhuwan and rest of the databricks team for the reviews,
I appreciate your reviews, was very helpful in evaluating a few options
that were overlooked earlier (especially about mixed spark apps running on
notebooks). Regarding the use-cases, It could handle multiple streaming
queries provided
+1
Nivi
Hi Pavan,
I am from the DataBricks Structured Streaming team, and we did a review of
the SPIP internally. Wanted to pass on the points discussed in the meeting.
Thanks for putting together the SPIP document. It's useful to have dynamic
resource allocation for Streaming queries, and it's exciting
Hello All,
in the list of JIRAs i didn't find anything related to fetchFailedException.
as mentioned above
"When Spark job shows FetchFailedException it creates few duplicate data
and we see few data also missing , please explain why. We have a scenario
when spark job complains FetchFailedExcepti