Got it. Since I am not applying any windowing strategy to the side input, does beam automatically pickup the windowing strategy for the side inputs from the main input? By that I mean the scope of the side input would be a per window one and it would be different for every window. Is that correct?
Regards, Harsh On Tue, May 15, 2018 at 17:54 Lukasz Cwik <lc...@google.com> wrote: > Using deduplicate + side inputs will allow you to have a consistent view > of the account information for the entire window which can be nice since it > gives consistent processing semantics but using a simple in memory cache to > reduce the amount of lookups will likely be much easier to debug and > simpler to implement and maintain. > > On Tue, May 15, 2018 at 2:31 PM Harshvardhan Agrawal < > harshvardhan.ag...@gmail.com> wrote: > >> Thanks Raghu! >> >> Lukasz, >> >> Do you think lookups would be a better option than side inputs in my case? >> >> >> On Tue, May 15, 2018 at 16:33 Raghu Angadi <rang...@google.com> wrote: >> >>> It should work. I think you need apply Distinct before looking up >>> account info : >>> billingDataPairs.apply(Keys.create()).apply(Distinct.create()).apply("LookupAccounts", >>> ...). >>> Note that all of the accounts are stored in single in-memory map. It >>> should be small enough for that. >>> >>> On Tue, May 15, 2018 at 1:15 PM Harshvardhan Agrawal < >>> harshvardhan.ag...@gmail.com> wrote: >>> >>>> Well ideally, I actually made the example a little easy. In the actual >>>> example I have multiple reference datasets. Say, I have a tuple of Account >>>> and Product as the key. The reason we don’t do the lookup in the DoFn >>>> directly is that we don’t want to lookup the data for the same account or >>>> same product multiple times across workers in a window. >>>> >>>> What I was thinking was that it might be better to perform the lookup >>>> only once for each account and product in a window and then supply them as >>>> side inputs to the main input. >>>> >>>> On Tue, May 15, 2018 at 16:03 Lukasz Cwik <lc...@google.com> wrote: >>>> >>>>> Is there a reason you don't want to read the accounting information >>>>> within the DoFn directly from the datastore, it seems like that would be >>>>> your simplest approach. >>>>> >>>>> On Tue, May 15, 2018 at 12:43 PM Harshvardhan Agrawal < >>>>> harshvardhan.ag...@gmail.com> wrote: >>>>> >>>>>> Hi, >>>>>> >>>>>> No we don’t receive any such information from Kafka. >>>>>> >>>>>> The account information in the external store does change. Every time >>>>>> we have a change in the account information we will have to recompute all >>>>>> the billing info. Our source systems will make sure that they publish >>>>>> messages for those accounts again. >>>>>> >>>>>> >>>>>> On Tue, May 15, 2018 at 15:11 Lukasz Cwik <lc...@google.com> wrote: >>>>>> >>>>>>> For each BillingModel you receive over Kafka, how "fresh" should the >>>>>>> account information be? >>>>>>> Does the account information in the external store change? >>>>>>> >>>>>>> On Tue, May 15, 2018 at 11:22 AM Harshvardhan Agrawal < >>>>>>> harshvardhan.ag...@gmail.com> wrote: >>>>>>> >>>>>>>> Hi, >>>>>>>> >>>>>>>> We have certain billing data that arrives to us from Kafka. The >>>>>>>> billing data is in json and it contains an account ID. In order for us >>>>>>>> to >>>>>>>> generate the final report we need to use some account data associated >>>>>>>> with >>>>>>>> the account id and is stored in an external database. >>>>>>>> >>>>>>>> It is possible that we get multiple billing info messages for the >>>>>>>> same account. We want to be able to lookup the account information for >>>>>>>> the >>>>>>>> messages in a window and then supply that as a side input to the next >>>>>>>> PTransform. >>>>>>>> >>>>>>>> Is it possible to achieve that in Beam? >>>>>>>> >>>>>>>> Here is my attempt: >>>>>>>> >>>>>>>> PCollection<KV<Integer, BillingModel>> billingDataPairs = >>>>>>>> p.apply("ReadBillingInfo", KafkaIO.<String, String>read() >>>>>>>> .withBootstrapServers(KAFKA_BOOTSTRAP_SERVER) >>>>>>>> .withTopic(KAFKA_TOPIC) >>>>>>>> .withKeyDeserializer(StringDeserializer.class) >>>>>>>> .withValueDeserializer(StringDeserializer.class) >>>>>>>> ) >>>>>>>> .apply("Window", >>>>>>>> Window.into(FixedWindows.of(Duration.standardSeconds(30)))) >>>>>>>> .apply("ProcessKafkaMessages",new KafkaProcessor()); >>>>>>>> >>>>>>>> PCollection<KV<Integer, Iterable<BillingModel>> billingData = >>>>>>>> billingDataPairs.apply(GroupByKey.<Integer, BillingModel>create()); >>>>>>>> >>>>>>>> PCollectionView<Map<Integer, Account>> accountData = >>>>>>>> billingDataPairs.apply("LookupAccounts",new >>>>>>>> AccountLookupClient()).apply(View.asMap()); >>>>>>>> >>>>>>>> billingDataPairs.apply(ParDo.of(new DoFn<KV<Integer, >>>>>>>> BillingModel>>(){ >>>>>>>> @ProcessElement >>>>>>>> public void processElement(ProcessContext ctx) { >>>>>>>> Integer accountId = ctx.element().getKey(); >>>>>>>> Iterable<BillingModel> billingModel = ctx.element().getValue(); >>>>>>>> Account account = ctx.sideinput(accountData).get(accountId); >>>>>>>> } >>>>>>>> })); >>>>>>>> >>>>>>>> Regards, >>>>>>>> Harsh >>>>>>>> -- >>>>>>>> >>>>>>>> *Regards,Harshvardhan Agrawal* >>>>>>>> *267.991.6618 | LinkedIn >>>>>>>> <https://www.linkedin.com/in/harshvardhanagr/>* >>>>>>>> >>>>>>> -- >>>>>> >>>>>> *Regards,Harshvardhan Agrawal* >>>>>> *267.991.6618 | LinkedIn >>>>>> <https://www.linkedin.com/in/harshvardhanagr/>* >>>>>> >>>>> -- >>>> >>>> *Regards,Harshvardhan Agrawal* >>>> *267.991.6618 | LinkedIn <https://www.linkedin.com/in/harshvardhanagr/>* >>>> >>> -- >> >> *Regards,Harshvardhan Agrawal* >> *267.991.6618 | LinkedIn <https://www.linkedin.com/in/harshvardhanagr/>* >> > -- *Regards,Harshvardhan Agrawal* *267.991.6618 | LinkedIn <https://www.linkedin.com/in/harshvardhanagr/>*