Hi All - we need a LP bug for this, please. The gerrit review will need to have its commit message updated with the lp bug info so that CI can do bug management on it. Then, once it lands in the charm's master, we can cherry-pick it for a stable backport to the stable charm. Thanks, Ryan
On Fri, Oct 21, 2016 at 9:54 AM, Chuck Short <chuck.sh...@canonical.com> wrote: > Hi, > > I proposed a fix: > > https://review.openstack.org/#/c/389740/ > > chuck > > On Fri, Oct 21, 2016 at 10:27 AM, Adam Stokes <adam.sto...@canonical.com> > wrote: > >> This looks like it's due to the way we deploy OpenStack with NovaLXD in >> all containers, this effectively breaks anyone wanting to do an all-in-one >> install on their system. >> >> On Fri, Oct 21, 2016 at 10:22 AM Adam Stokes <adam.sto...@canonical.com> >> wrote: >> >>> So it looks like a recent change to the LXD charm, see here: >>> >>> https://github.com/openstack/charm-lxd/commit/017246768e097c >>> 5fcd5283e23f19f075ff9f9d4e >>> >>> Chuck, are you aware of this issue? >>> >>> On Fri, Oct 21, 2016 at 10:19 AM Heather Lanigan <hmlani...@gmail.com> >>> wrote: >>> >>> Adam, >>> >>> The entire container is not readonly. Just /sys, the mount point for >>> /dev/.lxc/sys. I choose another charm (neutron-api) to look at, /sys on >>> that unit is readonly as well. Is that normal? >>> >>> What would be different in my config? My Xenial install is on a VM, but >>> I’ve been running that way for weeks. I did have the openstack-novalxd >>> bundle successfully deployed on it previously using juju 2.0_rc1. >>> >>> -Heather >>> >>> On Oct 20, 2016, at 11:30 PM, Adam Stokes <adam.sto...@canonical.com> >>> wrote: >>> >>> Odd it looks like the container has a read only file system? I ran >>> through a full openstack-novalxd deployment today and one of the upstream >>> maintainers ran through the same deployment and didn't run into any issues. >>> >>> On Thu, Oct 20, 2016, 10:02 PM Heather Lanigan <hmlani...@gmail.com> >>> wrote: >>> >>> >>> I used conjure-up to deploy openstack-novalxd on a Xenial system. >>> Before deploying, the operating system was updated. LXD init was setup >>> with dir, not xfs. All but one of the charms has a status of “unit is >>> ready" >>> >>> The lxd/0 subordinate charm has a status of: hook failed: >>> "config-changed”. See details below. >>> >>> I can boot an instance within this OpenStack deployment. However >>> deleting the instance fails. A side effect of the lxd/0 issues? >>> >>> Juju version 2.0.0-xenial-amd64 >>> conjure-up version 2.0.2 >>> lxd charm version 2.0.5 >>> >>> Any ideas? >>> >>> Thanks in advance, >>> Heather >>> >>> ++++++++++++++++++++++++++++++++++++++++++++++ >>> >>> The /var/log/juju/unit-lxd-0.log on the unit reports: >>> >>> 2016-10-21 01:09:33 INFO config-changed Traceback (most recent call >>> last): >>> 2016-10-21 01:09:33 INFO config-changed File >>> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/config-changed", line 140, >>> in <module> >>> 2016-10-21 01:09:33 INFO config-changed main() >>> 2016-10-21 01:09:33 INFO config-changed File >>> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/config-changed", line 134, >>> in main >>> 2016-10-21 01:09:33 INFO config-changed hooks.execute(sys.argv) >>> 2016-10-21 01:09:33 INFO config-changed File >>> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/charmhelpers/core/hookenv.py", >>> line 715, in execute >>> 2016-10-21 01:09:33 INFO config-changed self._hooks[hook_name]() >>> 2016-10-21 01:09:33 INFO config-changed File >>> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/config-changed", line 78, >>> in config_changed >>> 2016-10-21 01:09:33 INFO config-changed configure_lxd_host() >>> 2016-10-21 01:09:33 INFO config-changed File >>> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/charmhelpers/core/decorators.py", >>> line 40, in _retry_on_exception_inner_2 >>> 2016-10-21 01:09:33 INFO config-changed return f(*args, **kwargs) >>> 2016-10-21 01:09:33 INFO config-changed File >>> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/lxd_utils.py", line 429, >>> in configure_lxd_host >>> 2016-10-21 01:09:33 INFO config-changed with >>> open(EXT4_USERNS_MOUNTS, 'w') as userns_mounts: >>> 2016-10-21 01:09:33 INFO config-changed IOError: [Errno 30] Read-only >>> file system: '/sys/module/ext4/parameters/userns_mounts' >>> 2016-10-21 01:09:33 ERROR juju.worker.uniter.operation runhook.go:107 >>> hook "config-changed" failed: exit status 1 >>> >>> >>> root@juju-456efd-13:~# touch /sys/module/ext4/parameters/temp-file >>> touch: cannot touch '/sys/module/ext4/parameters/temp-file': Read-only >>> file system >>> root@juju-456efd-13:~# df -h /sys/module/ext4/parameters/userns_mounts >>> Filesystem Size Used Avail Use% Mounted on >>> sys 0 0 0 - /dev/.lxc/sys >>> root@juju-456efd-13:~# touch /home/ubuntu/temp-file >>> root@juju-456efd-13:~# ls /home/ubuntu/temp-file >>> /home/ubuntu/temp-file >>> root@juju-456efd-13:~# df -h >>> Filesystem Size Used Avail Use% Mounted on >>> /dev/mapper/mitaka--vg-root 165G 47G 110G 30% / >>> none 492K 0 492K 0% /dev >>> udev 16G 0 16G 0% /dev/fuse >>> tmpfs 16G 0 16G 0% /dev/shm >>> tmpfs 16G 49M 16G 1% /run >>> tmpfs 5.0M 0 5.0M 0% /run/lock >>> tmpfs 16G 0 16G 0% /sys/fs/cgroup >>> tmpfs 3.2G 0 3.2G 0% /run/user/112 >>> tmpfs 3.2G 0 3.2G 0% /run/user/1000 >>> >>> +++++++++++++++++++++++++++++++++++++++++ >>> >>> heather@mitaka:~$ nova boot --image d2eba22a-e1b1-4a2b-aa87-450ee9d9e492 >>> --flavor d --nic net-name=ubuntu-net --key-name keypair-admin >>> xenial-instance >>> heather@mitaka:~/goose-work/src/gopkg.in/goose.v1$ nova list >>> +--------------------------------------+-----------------+-- >>> ------+------------+-------------+-----------------------+ >>> | ID | Name | Status | Task >>> State | Power State | Networks | >>> +--------------------------------------+-----------------+-- >>> ------+------------+-------------+-----------------------+ >>> | 80424b94-f24d-45ff-a330-7b67a911fbc6 | xenial-instance | ACTIVE | - >>> | Running | ubuntu-net=10.101.0.8 | >>> +--------------------------------------+-----------------+-- >>> ------+------------+-------------+-----------------------+ >>> >>> heather@mitaka:~$ nova delete 80424b94-f24d-45ff-a330-7b67a911fbc6 >>> Request to delete server 80424b94-f24d-45ff-a330-7b67a911fbc6 has been >>> accepted. >>> heather@mitaka:~$ nova list >>> +--------------------------------------+-----------------+-- >>> ------+------------+-------------+----------+ >>> | ID | Name | Status | Task >>> State | Power State | Networks | >>> +--------------------------------------+-----------------+-- >>> ------+------------+-------------+----------+ >>> | 80424b94-f24d-45ff-a330-7b67a911fbc6 | xenial-instance | ERROR | - >>> | Running | | >>> +--------------------------------------+-----------------+-- >>> ------+------------+-------------+----------+ >>> heather@mitaka:~$ nova show 80424b94-f24d-45ff-a330-7b67a911fbc6 >>> … >>> | fault | {"message": "Failed to >>> communicate with LXD API instance-00000006: Error 400 - Profile is >>> currently in use.", "code": 500, "details": " File >>> \"/usr/lib/python2.7/dist-packages/nova/compute/manager.py\", line 375, >>> in decorated_function | >>> | >>> ... >>> >>> >>> CONFIDENTIAL This electronic message contains information from >>> TransCirrus, Inc. which may be confidential, privileged or otherwise >>> protected from disclosure, and is to be used solely by the intended >>> recipient. Any unauthorized review, disclosure, copying, distribution or >>> use of this transmission or its contents is prohibited. If you have >>> received this transmission in error, please notify the sender immediately >>> by reply email to i...@transcirrus.com, and destroy all copies of the >>> original message. Thank you. >>> >>> >>> > > -- > Juju mailing list > Juju@lists.ubuntu.com > Modify settings or unsubscribe at: https://lists.ubuntu.com/ > mailman/listinfo/juju > >
-- Juju mailing list Juju@lists.ubuntu.com Modify settings or unsubscribe at: https://lists.ubuntu.com/mailman/listinfo/juju