Hi All,

I am using structured streaming in Databricks by using foreach
functionality to do my transformation and action and finally need to write
the data into a delta table my data soruce is either (Eventhub or delta
table or azure cosmos changefeed).

Whenever there are huge changes in source(Delta table, Azure SQL, cosmos),
in the streaming process all the data is processed at the same micro-batch
due to which with our existing cluster we are not able to process all the
data at the same time.

So we need to chunk the data in forarch microbatch while reading change
data from source(delta table, Azure SQL, Azure cosmos). we need to set the
foreach microbatch count limit in the structured streaming flow.

Could you please suggest a way to do the same.

Thanks,

Reply via email to