I just tried with v2.29.0 and use_deprecated_read but unfortunately I observed slow behavior again. Is it possible that use_deprecated_read is broken in 2.29.0 as well?
Thanks, Evan On Wed, May 12, 2021 at 3:21 PM Steve Niemitz <sniem...@apache.org> wrote: > oops sorry I was off by 10...I meant 2.29 not 2.19. > > On Wed, May 12, 2021 at 2:55 PM Evan Galpin <evan.gal...@gmail.com> wrote: > >> Thanks for the link/info. v2.19.0 and v2.21.0 did exhibit the "faster" >> behavior, as did v2.23.0. But that "fast" behavior stopped at v2.25.0 (for >> my use case at least) regardless of use_deprecated_read setting. >> >> Thanks, >> Evan >> >> >> On Wed, May 12, 2021 at 2:47 PM Steve Niemitz <sniem...@apache.org> >> wrote: >> >>> use_deprecated_read was broken in 2.19 on the direct runner and didn't >>> do anything. [1] I don't think the fix is in 2.20 either, but will be in >>> 2.21. >>> >>> [1] https://github.com/apache/beam/pull/14469 >>> >>> On Wed, May 12, 2021 at 1:41 PM Evan Galpin <evan.gal...@gmail.com> >>> wrote: >>> >>>> I forgot to also mention that in all tests I was setting >>>> --experiments=use_deprecated_read >>>> >>>> Thanks, >>>> Evan >>>> >>>> On Wed, May 12, 2021 at 1:39 PM Evan Galpin <evan.gal...@gmail.com> >>>> wrote: >>>> >>>>> Hmm, I think I spoke too soon. I'm still seeing an issue of overall >>>>> DirectRunner slowness, not just pubsub. I have a pipeline like so: >>>>> >>>>> Read pubsub | extract GCS glob patterns | FileIO.matchAll() | >>>>> FileIO.readMatches() | Read file contents | etc >>>>> >>>>> I have temporarily set up a transform between each step to log what's >>>>> going on and illustrate timing issues. I ran a series of tests changing >>>>> only the SDK version each time since I hadn't noticed this performance >>>>> issue with 2.19.0 (effectively git-bisect). Before each test, I seeded the >>>>> pubsub subscription with the exact same contents. >>>>> >>>>> SDK version 2.25.0 (I had a build issue with 2.24.0 that I couldn't >>>>> seem to resolve) and onward show a significant slowdown. >>>>> >>>>> Here is a snippet of logging from v2.25.0: >>>>> >>>>> *May 12, 2021 11:11:52 A.M.* com.myOrg.myPipeline.PipelineLeg$1 >>>>> processElement >>>>> INFO: Got file pattern: gs://my-bucket/my-dir/5004728247517184/** >>>>> May 12, 2021 11:16:59 A.M. >>>>> org.apache.beam.sdk.io.FileIO$MatchAll$MatchFn process >>>>> INFO: Matched 2 files for pattern >>>>> gs://my-bucket/my-dir/5004728247517184/** >>>>> May 12, 2021 11:23:32 A.M. com.myOrg.myPipeline.PipelineLeg$3 >>>>> processElement >>>>> INFO: Got ReadableFile: my-file1.json >>>>> May 12, 2021 11:23:32 A.M. com.myOrg.myPipeline.PipelineLeg$3 >>>>> processElement >>>>> INFO: Got ReadableFile: my-file2.json >>>>> May 12, 2021 11:24:35 A.M. com.myOrg.myPipeline.PipelineLeg$4 >>>>> processElement >>>>> INFO: Got file contents for document_id my-file1.json >>>>> *May 12, 2021 11:24:35 A.M*. com.myOrg.myPipeline.PipelineLeg$4 >>>>> processElement >>>>> INFO: Got file contents for document_id my-file2.json >>>>> >>>>> Note that end-to-end, these steps took about *13 minutes*. With SDK >>>>> 2.23.0 and identical user code, the same section of the pipeline took *2 >>>>> seconds*: >>>>> >>>>> *May 12, 2021 11:03:39 A.M.* com.myOrg.myPipeline.PipelineLeg$1 >>>>> processElement >>>>> INFO: Got file pattern: gs://my-bucket/my-dir/5004728247517184/** >>>>> May 12, 2021 11:03:40 A.M. >>>>> org.apache.beam.sdk.io.FileIO$MatchAll$MatchFn process >>>>> INFO: Matched 2 files for pattern >>>>> gs://my-bucket/my-dir/5004728247517184/** >>>>> May 12, 2021 11:03:40 A.M. com.myOrg.myPipeline.PipelineLeg$3 >>>>> processElement >>>>> INFO: Got ReadableFile: my-file1.json >>>>> May 12, 2021 11:03:40 A.M. com.myOrg.myPipeline.PipelineLeg$3 >>>>> processElement >>>>> INFO: Got ReadableFile: my-file2.json >>>>> May 12, 2021 11:03:41 A.M. com.myOrg.myPipeline.PipelineLeg$4 >>>>> processElement >>>>> INFO: Got file contents for document_id my-file1.json >>>>> *May 12, 2021 11:03:41 A.M.* com.myOrg.myPipeline.PipelineLeg$4 >>>>> processElement >>>>> INFO: Got file contents for document_id my-file2.json >>>>> >>>>> Any thoughts on what could be causing this? >>>>> >>>>> Thanks, >>>>> Evan >>>>> >>>>> On Wed, May 12, 2021 at 9:53 AM Evan Galpin <evan.gal...@gmail.com> >>>>> wrote: >>>>> >>>>>> >>>>>> >>>>>> On Mon, May 10, 2021 at 2:09 PM Boyuan Zhang <boyu...@google.com> >>>>>> wrote: >>>>>> >>>>>>> Hi Evan, >>>>>>> >>>>>>> What do you mean startup delay? Is it the time that from you start >>>>>>> the pipeline to the time that you notice the first output record from >>>>>>> PubSub? >>>>>>> >>>>>> >>>>>> Yes that's what I meant, the seemingly idle system waiting for pubsub >>>>>> output despite data being in the subscription at pipeline start time. >>>>>> >>>>>> On Sat, May 8, 2021 at 12:50 AM Ismaël Mejía <ieme...@gmail.com> >>>>>>> wrote: >>>>>>> >>>>>>>> Can you try running direct runner with the option >>>>>>>> `--experiments=use_deprecated_read` >>>>>>>> >>>>>>> >>>>>> This seems to work for me, thanks for this! 👍 >>>>>> >>>>>> >>>>>>>> Seems like an instance of >>>>>>>> https://issues.apache.org/jira/browse/BEAM-10670?focusedCommentId=17316858&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-17316858 >>>>>>>> also reported in >>>>>>>> https://lists.apache.org/thread.html/re6b0941a8b4951293a0327ce9b25e607cafd6e45b69783f65290edee%40%3Cdev.beam.apache.org%3E >>>>>>>> >>>>>>>> We should rollback using the SDF wrapper by default because of the >>>>>>>> usability and performance issues reported. >>>>>>>> >>>>>>>> >>>>>>>> On Sat, May 8, 2021 at 12:57 AM Evan Galpin <evan.gal...@gmail.com> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Hi all, >>>>>>>>> >>>>>>>>> I’m experiencing very slow performance and startup delay when >>>>>>>>> testing a pipeline locally. I’m reading data from a Google PubSub >>>>>>>>> subscription as the data source, and before each pipeline execution I >>>>>>>>> ensure that data is present in the subscription (readable from GCP >>>>>>>>> console). >>>>>>>>> >>>>>>>>> I’m seeing startup delay on the order of minutes with DirectRunner >>>>>>>>> (5-10 min). Is that expected? I did find a Jira ticket[1] that at >>>>>>>>> first >>>>>>>>> seemed related, but I think it has more to do with BQ than >>>>>>>>> DirectRunner. >>>>>>>>> >>>>>>>>> I’ve run the pipeline with a debugger connected and confirmed that >>>>>>>>> it’s minutes before the first DoFn in my pipeline receives any data. >>>>>>>>> Is >>>>>>>>> there a way I can profile the direct runner to see what it’s churning >>>>>>>>> on? >>>>>>>>> >>>>>>>>> Thanks, >>>>>>>>> Evan >>>>>>>>> >>>>>>>>> [1] >>>>>>>>> >>>>>>>>> https://issues.apache.org/jira/plugins/servlet/mobile#issue/BEAM-4548 >>>>>>>>> >>>>>>>>