apologies, the code should read as below from threading import Thread
context = pyspark.sql.SparkSession.builder.appName("spark").getOrCreate() t1 = Thread(target=my_func, args=(context,)) t1.start() t2 = Thread(target=my_func, args=(context,)) t2.start() On Tue, Dec 13, 2022 at 4:10 PM Jack Goodson <jackagood...@gmail.com> wrote: > Hi Kevin, > > I had a similar use case (see below code) but with something that wasn’t > spark related. I think the below should work for you, you may need to edit > the context variable to suit your needs but hopefully it gives the general > idea of sharing a single object between multiple threads. > > Thanks > > > from threading import Thread > > context = pyspark.sql.SparkSession.builder.appName("spark").getOrCreate() > > t1 = Thread(target=order_creator, args=(app_id, sleep_time,)) > t1.start(target=my_func, args=(context,)) > > t2 = Thread(target=order_creator, args=(app_id, sleep_time,)) > t2.start(target=my_func, args=(context,)) >