Thanks, creating the venv in a container did the trick. May I suggest adding virtualenv to the Python 3.7 image? python -m venv works, but virtualenv is more commonly used.
On Mon, Jan 13, 2020 at 2:09 AM Brooke Storm <bst...@wikimedia.org> wrote: > You have to create the venv in a container using 'webservice shell of the > right runtime'. We support Python versions from Debian Jessie, Stretch and > Buster by building in containers, so we cannot sync more than one of those > to the bastion. We have moved a lot of Python tools back and forth without > issue, but we have rebuilt the container image, which can introduce > issues. Tomorrow, I will try a few things to be sure, but a venv can easily > cause problems. > > On Sun, Jan 12, 2020, 16:30 Alex Monk <kren...@gmail.com> wrote: > >> Interesting, uwsgi had Python 3.7.3 but `./www/python/venv/bin/python >> --version` says 3.7.6. Is that a big enough difference to cause problems? >> >> On Sun, 12 Jan 2020 at 23:19, Chico Venancio <chicocvenan...@gmail.com> >> wrote: >> >>> Maybe a venv created in a different python version? >>> >>> Chico Venancio >>> >>> Em dom, 12 de jan de 2020 20:14, Alex Monk <kren...@gmail.com> escreveu: >>> >>>> I think I've seen that particular error that you see in stdout/stderr >>>> (via kubectl logs) before - on pods that in fact were working. >>>> >>>> Meanwhile, uwsgi.log says: >>>> >>>> Python version: 3.7.3 (default, Apr 3 2019, 05:39:12) [GCC 8.3.0] >>>> Set PythonHome to /data/project/countcounttest/www/python/venv >>>> Fatal Python error: initfsencoding: Unable to get the locale encoding >>>> ModuleNotFoundError: No module named 'encodings' >>>> >>>> Current thread 0x00007fe50490e780 (most recent call first): >>>> !!! uWSGI process 1 got Segmentation Fault !!! >>>> >>>> followed by a backtrace. Suggests the problem is related to something >>>> inside the image/application code rather than the cluster itself anyway. >>>> I notice the pod on the new cluster seems to be using the sssd variant >>>> of the toolforge-python37-web image, which pods in the old cluster are not >>>> using. I doubt it's the source problem as uwsgi shouldn't be segfaulting >>>> over some problem talking to LDAP... >>>> Needs further investigation by someone during the week I think. >>>> >>>> On Sun, 12 Jan 2020 at 23:00, Count Count < >>>> countvoncount123...@gmail.com> wrote: >>>> >>>>> Your pod started and container and it crashed, I see a uwsgi.log file >>>>>> with a python module problem and a uwsgi segfault. >>>>>> >>>>> >>>>> Yes. It was working fine with the legacy cluster. >>>>> The service ist started via webservice --backend=kubernetes >>>>> python3.7 start >>>>> >>>>> Apparently it cannot load the uwsgi shared library if deployed on the >>>>> new cluster? >>>>> tools.countcounttest@tools-sgebastion-07:~$ kubectl logs >>>>> countcounttest-6b58f5c547-785mr >>>>> open("/usr/lib/uwsgi/plugins/python_plugin.so"): No such file or >>>>> directory [core/utils.c line 3724] >>>>> !!! UNABLE to load uWSGI plugin: >>>>> /usr/lib/uwsgi/plugins/python_plugin.so: cannot open shared object file: >>>>> No >>>>> such file or directory !!! >>>>> >>>>> On Sun, Jan 12, 2020 at 11:42 PM Alex Monk <kren...@gmail.com> wrote: >>>>> >>>>>> Hi Count Count, I believe I may have sorted out an issue that >>>>>> prevented some pods (depending partially on luck) from creating >>>>>> containers. >>>>>> Your pod started and container and it crashed, I see a uwsgi.log file >>>>>> with >>>>>> a python module problem and a uwsgi segfault. >>>>>> >>>>>> On Sun, 12 Jan 2020 at 22:12, Alex Monk <kren...@gmail.com> wrote: >>>>>> >>>>>>> Thanks Count Count. I have identified a new issue with the new k8s >>>>>>> cluster and am looking into it. >>>>>>> >>>>>>> On Sun, 12 Jan 2020 at 21:43, Count Count < >>>>>>> countvoncount123...@gmail.com> wrote: >>>>>>> >>>>>>>> Yes, I switched back to the old cluster. This is a new tool that >>>>>>>> was used in production even if only rarely. I can't leave it offline >>>>>>>> for >>>>>>>> hours. >>>>>>>> >>>>>>>> I have created a test tool as a copy with which I can reproduce the >>>>>>>> issue: >>>>>>>> tools.countcounttest@tools-sgebastion-07:~$ kubectl get pods >>>>>>>> NAME READY STATUS >>>>>>>> RESTARTS AGE >>>>>>>> countcounttest-6b58f5c547-mf4jx 0/1 ContainerCreating 0 >>>>>>>> 77s >>>>>>>> >>>>>>>> I will leave that running. If the container gets created I might >>>>>>>> also be able to reproduce the segfault. >>>>>>>> >>>>>>>> Best regards, >>>>>>>> >>>>>>>> Count Count >>>>>>>> >>>>>>>> On Sun, Jan 12, 2020 at 10:20 PM Alex Monk <kren...@gmail.com> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Hi Count Count, >>>>>>>>> >>>>>>>>> I'm afraid you seem to have no pods on the new cluster to look at: >>>>>>>>> >>>>>>>>> # kubectl get -n tool-flaggedrevspromotioncheck pod >>>>>>>>> No resources found. >>>>>>>>> >>>>>>>>> Alex >>>>>>>>> >>>>>>>>> On Sun, 12 Jan 2020 at 21:07, Count Count < >>>>>>>>> countvoncount123...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> Hi! >>>>>>>>>> >>>>>>>>>> I don't have much luck with a webservice based on the python3.7 >>>>>>>>>> image. It is running fine on the legacy K8s cluster. >>>>>>>>>> >>>>>>>>>> On the new cluster I got a segfault. After stopping the >>>>>>>>>> webservice and trying again to get an empty log the pod is now stuck >>>>>>>>>> in >>>>>>>>>> ContainerCreating. >>>>>>>>>> >>>>>>>>>> A few minutes ago: >>>>>>>>>> tools.flaggedrevspromotioncheck@tools-sgebastion-08:~$ kubectl >>>>>>>>>> get pods >>>>>>>>>> NAME READY STATUS >>>>>>>>>> RESTARTS AGE >>>>>>>>>> flaggedrevspromotioncheck-7cbfff44fc-jnhmq 0/1 >>>>>>>>>> ContainerCreating 0 2m48s >>>>>>>>>> >>>>>>>>>> ...and just now: >>>>>>>>>> tools.flaggedrevspromotioncheck@tools-sgebastion-08:~$ kubectl >>>>>>>>>> get pods >>>>>>>>>> NAME READY STATUS >>>>>>>>>> RESTARTS AGE >>>>>>>>>> flaggedrevspromotioncheck-7cbfff44fc-q55gm 0/1 >>>>>>>>>> ContainerCreating 0 5m18s >>>>>>>>>> >>>>>>>>>> Best regards, >>>>>>>>>> >>>>>>>>>> Count Count >>>>>>>>>> >>>>>>>>>> On Thu, Jan 9, 2020 at 10:58 PM Bryan Davis <bd...@wikimedia.org> >>>>>>>>>> wrote: >>>>>>>>>> >>>>>>>>>>> I am happy to announce that a new and improved Kubernetes >>>>>>>>>>> cluster is >>>>>>>>>>> now available for use by beta testers on an opt-in basis. A page >>>>>>>>>>> has >>>>>>>>>>> been created on Wikitech [0] outlining the self-service migration >>>>>>>>>>> process. >>>>>>>>>>> >>>>>>>>>>> Timeline: >>>>>>>>>>> * 2020-01-09: 2020 Kubernetes cluster available for beta testers >>>>>>>>>>> on an >>>>>>>>>>> opt-in basis >>>>>>>>>>> * 2020-01-23: 2020 Kubernetes cluster general availability for >>>>>>>>>>> migration on an opt-in basis >>>>>>>>>>> * 2020-02-10: Automatic migration of remaining workloads from >>>>>>>>>>> 2016 >>>>>>>>>>> cluster to 2020 cluster by Toolforge admins >>>>>>>>>>> >>>>>>>>>>> This new cluster has been a work in progress for more than a year >>>>>>>>>>> within the Wikimedia Cloud Services team, and a top priority >>>>>>>>>>> project >>>>>>>>>>> for the past six months. About 35 tools, including >>>>>>>>>>> https://tools.wmflabs.org/admin/, are currently running on what >>>>>>>>>>> we are >>>>>>>>>>> calling the "2020 Kubernetes cluster". This new cluster is >>>>>>>>>>> running >>>>>>>>>>> Kubernetes v1.15.6 and Docker 19.03.4. It is also using a newer >>>>>>>>>>> authentication and authorization method (RBAC), a new ingress >>>>>>>>>>> routing >>>>>>>>>>> service, and a different method of integrating with the Developer >>>>>>>>>>> account LDAP service. We have built a new tool [1] which makes >>>>>>>>>>> the >>>>>>>>>>> state of the Kubernetes cluster more transparent and on par with >>>>>>>>>>> the >>>>>>>>>>> information that we already expose for the grid engine cluster >>>>>>>>>>> [2]. >>>>>>>>>>> >>>>>>>>>>> With a significant number of tools managed by Toolforge >>>>>>>>>>> administrators >>>>>>>>>>> already migrated to the new cluster, we are fairly confident >>>>>>>>>>> that the >>>>>>>>>>> basic features used by most Kubernetes tools are covered. It is >>>>>>>>>>> likely >>>>>>>>>>> that a few outlying issues remain to be found as more tools >>>>>>>>>>> move, but >>>>>>>>>>> we have confidence that we can address them quickly. This has >>>>>>>>>>> led us >>>>>>>>>>> to propose a fairly short period of voluntary beta testing, >>>>>>>>>>> followed >>>>>>>>>>> by a short general availability opt-in migration period, and >>>>>>>>>>> finally a >>>>>>>>>>> complete migration of all remaining tools which will be done by >>>>>>>>>>> the >>>>>>>>>>> Toolforge administration team for anyone who has not migrated >>>>>>>>>>> their >>>>>>>>>>> self. >>>>>>>>>>> >>>>>>>>>>> Please help with beta testing if you have some time and are >>>>>>>>>>> willing to >>>>>>>>>>> get help on irc, Phabricator, and the cloud@lists.wikimedia.org >>>>>>>>>>> mailing list for early adopter issues you may encounter. >>>>>>>>>>> >>>>>>>>>>> I want to publicly praise Brooke Storm and Arturo Borrero >>>>>>>>>>> González for >>>>>>>>>>> the hours that they have put into reading docs, building proof of >>>>>>>>>>> concept clusters, and improving automation and processes to make >>>>>>>>>>> the >>>>>>>>>>> 2020 Kubernetes cluster possible. The Toolforge community can >>>>>>>>>>> look >>>>>>>>>>> forward to more frequent and less disruptive software upgrades >>>>>>>>>>> in this >>>>>>>>>>> cluster as a direct result of this work. We have some other >>>>>>>>>>> feature >>>>>>>>>>> improvements in planning now that I think you will all be >>>>>>>>>>> excited to >>>>>>>>>>> see and use later this year! >>>>>>>>>>> >>>>>>>>>>> [0]: >>>>>>>>>>> https://wikitech.wikimedia.org/wiki/News/2020_Kubernetes_cluster_migration >>>>>>>>>>> [1]: https://tools.wmflabs.org/k8s-status/ >>>>>>>>>>> [2]: https://tools.wmflabs.org/sge-status/ >>>>>>>>>>> >>>>>>>>>>> Bryan (on behalf of the Toolforge admins and the Cloud Services >>>>>>>>>>> team) >>>>>>>>>>> -- >>>>>>>>>>> Bryan Davis Technical Engagement Wikimedia >>>>>>>>>>> Foundation >>>>>>>>>>> Principal Software Engineer Boise, >>>>>>>>>>> ID USA >>>>>>>>>>> [[m:User:BDavis_(WMF)]] >>>>>>>>>>> irc: bd808 >>>>>>>>>>> >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> Wikimedia Cloud Services announce mailing list >>>>>>>>>>> cloud-annou...@lists.wikimedia.org (formerly >>>>>>>>>>> labs-annou...@lists.wikimedia.org) >>>>>>>>>>> https://lists.wikimedia.org/mailman/listinfo/cloud-announce >>>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>>> Wikimedia Cloud Services mailing list >>>>>>>>>> Cloud@lists.wikimedia.org (formerly lab...@lists.wikimedia.org) >>>>>>>>>> https://lists.wikimedia.org/mailman/listinfo/cloud >>>>>>>>> >>>>>>>>> _______________________________________________ >>>>>>>>> Wikimedia Cloud Services mailing list >>>>>>>>> Cloud@lists.wikimedia.org (formerly lab...@lists.wikimedia.org) >>>>>>>>> https://lists.wikimedia.org/mailman/listinfo/cloud >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> Wikimedia Cloud Services mailing list >>>>>>>> Cloud@lists.wikimedia.org (formerly lab...@lists.wikimedia.org) >>>>>>>> https://lists.wikimedia.org/mailman/listinfo/cloud >>>>>>> >>>>>>> _______________________________________________ >>>>>> Wikimedia Cloud Services mailing list >>>>>> Cloud@lists.wikimedia.org (formerly lab...@lists.wikimedia.org) >>>>>> https://lists.wikimedia.org/mailman/listinfo/cloud >>>>> >>>>> _______________________________________________ >>>>> Wikimedia Cloud Services mailing list >>>>> Cloud@lists.wikimedia.org (formerly lab...@lists.wikimedia.org) >>>>> https://lists.wikimedia.org/mailman/listinfo/cloud >>>> >>>> _______________________________________________ >>>> Wikimedia Cloud Services mailing list >>>> Cloud@lists.wikimedia.org (formerly lab...@lists.wikimedia.org) >>>> https://lists.wikimedia.org/mailman/listinfo/cloud >>> >>> _______________________________________________ >>> Wikimedia Cloud Services mailing list >>> Cloud@lists.wikimedia.org (formerly lab...@lists.wikimedia.org) >>> https://lists.wikimedia.org/mailman/listinfo/cloud >> >> _______________________________________________ >> Wikimedia Cloud Services mailing list >> Cloud@lists.wikimedia.org (formerly lab...@lists.wikimedia.org) >> https://lists.wikimedia.org/mailman/listinfo/cloud > > _______________________________________________ > Wikimedia Cloud Services mailing list > Cloud@lists.wikimedia.org (formerly lab...@lists.wikimedia.org) > https://lists.wikimedia.org/mailman/listinfo/cloud
_______________________________________________ Wikimedia Cloud Services mailing list Cloud@lists.wikimedia.org (formerly lab...@lists.wikimedia.org) https://lists.wikimedia.org/mailman/listinfo/cloud