The installation happens in a separate process and so there should be no direct relationship between them. Could you try to take a look at whether the GC runs normally. Besides, you could also try to increase the heartbeat timeout.
Regards, Dian On Fri, Dec 24, 2021 at 12:02 PM Paul Lam <paullin3...@gmail.com> wrote: > Hi Dian, > > Thanks lot for your pointer! > > I tried `-pyreq` and it worked as expected in general. But another problem > arises when I tried to specified more dependencies in the requirements.txt. > > The taskmanagers are often lost(unreachable from jobmanager) during the > process of pip install. > > I wonder if the installation affects the heartbeats between taskmanager > and jobmanager? > > Best, > Paul Lam > > 2021年12月23日 19:16,Dian Fu <dian0511...@gmail.com> 写道: > > Hi Paul, > > Currently, you need to build venv in an environment where you want to > execute the PyFlink jobs. > > >> Also, I wonder if it’s possible for pyflink to optionally provide an > automatically created venv for each pyflink job? > Do you mean to create the venv during executing the job? If this is your > requirement, maybe you could try to specify a requirements.txt file [1]. It > will prepare the Python environment according to > the requirements.txt during executing the job. > > Regards, > Dian > > [1] > https://nightlies.apache.org/flink/flink-docs-release-1.14/docs/dev/python/dependency_management/#requirementstxt > > > On Thu, Dec 23, 2021 at 5:15 PM Paul Lam <paullin3...@gmail.com> wrote: > >> Hi, >> >> The document says we could use `-pyarch` to upload python venv, but I >> found this is often problematic because users may not have the python >> binaries that fits the flink runtime environment. >> >> For example, a user may upload a venv for macOS within the project, but >> the Flink cluster is using Debian, causing pyflink to fail. >> >> Is there some good practice to avoid this? Also, I wonder if it’s >> possible for pyflink to optionally provide an automatically created venv >> for each pyflink job? >> >> Thanks! >> >> Best, >> Paul Lam >> >> >