Thank you for the clarification.I would like to how can I  proceed for this
kind of scenario in pyspark

I have a scenario subtracting the total number of days with the number of
holidays in pyspark by using dataframes

I have a table with dates  date1  date2 in one table and number of holidays
in another table
df1 = select date1,date2 ,ctry ,unixtimestamp(date2-date1)
totalnumberofdays  - df2.holidays  from table A;

df2 = select count(holiays)
from table B
where holidate >= 'date1'(table A)
and holidate < = date2(table A)
and country = A.ctry(table A)

Except country no other column is not a unique key




On Wed, Sep 30, 2020 at 6:05 PM Sean Owen <sro...@gmail.com> wrote:

> No, you can't use the SparkSession from within a function executed by
> Spark tasks.
>
> On Wed, Sep 30, 2020 at 7:29 AM Lakshmi Nivedita <klnived...@gmail.com>
> wrote:
>
>> Here is a spark udf structure as an example
>>
>> Def sampl_fn(x):
>>            Spark.sql(“select count(Id) from sample Where Id = x ”)
>>
>>
>> Spark.udf.register(“sample_fn”, sample_fn)
>>
>> Spark.sql(“select id, sampl_fn(Id) from example”)
>>
>> Advance Thanks for the help
>> --
>> k.Lakshmi Nivedita
>>
>>
>>
>>
>
>

Reply via email to