Hey Team,
We're trying to implement an aggregation which involves *several trillions
of rows *using apache beam sql.
However I'm getting an exception
Exception in thread "main" java.lang.UnsupportedOperationException: Does
not support COUNT DISTINCT
Here's the code for doing the aggregation:
PCo
Does Apache Beam version (2.41.0) or latest (2.51.0) support Spark 3.5
environment for spark runner ?
Apache Beam - Spark Runner Documentation states -
The Spark runner currently supports Spark’s 3.2.x branch
Thanks
Sri Ganesh V