Hi Russell,
This is super helpful. Thank you so much.
Can you elaborate on the differences between structured streaming vs
dstreams? How would the number of receivers required etc change?
On Sat, 8 Aug, 2020, 10:28 pm Russell Spitzer,
wrote:
> Note, none of this applies to Direct streaming appr
Hello, Sir!
What about process and group the data first then write grouped data to Kafka
topics A and B. Then read topic A or B from another Spark Application and
process it more. Like the term ETL's mean.
TianlangStudio
Some of the biggest lies: I will start tomorrow/Others are better
Hi,
I have a scenario where a kafka topic is being written with different types
of json records.
I have to regroup the records based on the type and then fetch the schema
and parse and write as parquet.
I have tried structured programming. But dynamic schema is a constraint.
So I have used DStream
Hi All,
I have a following info.in the data column.
<1000> date=2020-08-01 time=20:50:04 name=processing id=123 session=new
packt=20 orgin=null address=null dest=fgjglgl
here I want to create a separate column for the above key value pairs after
the integer <1000> separated by spaces.
Is there a