Hi, Do we support distinct count in the over clause in spark sql?
I ran a sql like this: select a, count(distinct b) over ( order by a rows between unbounded preceding and current row) from table limit 10 Currently, it return an error says: expression ‘a' is neither present in the group by, nor is it an aggregate function. Add to group by or wrap in first() if you don't care which value you get.; Yang