Update based on IRC conversation: I tried the fix. Turns out that (my) Xenial install doesn’t have the systemd-virt-detect command.
-Heather > On Oct 21, 2016, at 10:54 AM, Chuck Short <chuck.sh...@canonical.com> wrote: > > Hi, > > I proposed a fix: > > https://review.openstack.org/#/c/389740/ > <https://review.openstack.org/#/c/389740/> > > chuck > > On Fri, Oct 21, 2016 at 10:27 AM, Adam Stokes <adam.sto...@canonical.com > <mailto:adam.sto...@canonical.com>> wrote: > This looks like it's due to the way we deploy OpenStack with NovaLXD in all > containers, this effectively breaks anyone wanting to do an all-in-one > install on their system. > > On Fri, Oct 21, 2016 at 10:22 AM Adam Stokes <adam.sto...@canonical.com > <mailto:adam.sto...@canonical.com>> wrote: > So it looks like a recent change to the LXD charm, see here: > > https://github.com/openstack/charm-lxd/commit/017246768e097c5fcd5283e23f19f075ff9f9d4e > > <https://github.com/openstack/charm-lxd/commit/017246768e097c5fcd5283e23f19f075ff9f9d4e> > > Chuck, are you aware of this issue? > > On Fri, Oct 21, 2016 at 10:19 AM Heather Lanigan <hmlani...@gmail.com > <mailto:hmlani...@gmail.com>> wrote: > Adam, > > The entire container is not readonly. Just /sys, the mount point for > /dev/.lxc/sys. I choose another charm (neutron-api) to look at, /sys on that > unit is readonly as well. Is that normal? > > What would be different in my config? My Xenial install is on a VM, but I’ve > been running that way for weeks. I did have the openstack-novalxd bundle > successfully deployed on it previously using juju 2.0_rc1. > > -Heather > >> On Oct 20, 2016, at 11:30 PM, Adam Stokes <adam.sto...@canonical.com >> <mailto:adam.sto...@canonical.com>> wrote: >> >> Odd it looks like the container has a read only file system? I ran through a >> full openstack-novalxd deployment today and one of the upstream maintainers >> ran through the same deployment and didn't run into any issues. >> >> >> On Thu, Oct 20, 2016, 10:02 PM Heather Lanigan <hmlani...@gmail.com >> <mailto:hmlani...@gmail.com>> wrote: >> >> I used conjure-up to deploy openstack-novalxd on a Xenial system. Before >> deploying, the operating system was updated. LXD init was setup with dir, >> not xfs. All but one of the charms has a status of “unit is ready" >> >> The lxd/0 subordinate charm has a status of: hook failed: "config-changed”. >> See details below. >> >> I can boot an instance within this OpenStack deployment. However deleting >> the instance fails. A side effect of the lxd/0 issues? >> >> Juju version 2.0.0-xenial-amd64 >> conjure-up version 2.0.2 >> lxd charm version 2.0.5 >> >> Any ideas? >> >> Thanks in advance, >> Heather >> >> ++++++++++++++++++++++++++++++++++++++++++++++ >> >> The /var/log/juju/unit-lxd-0.log on the unit reports: >> 2016-10-21 01:09:33 INFO config-changed Traceback (most recent call last): >> 2016-10-21 01:09:33 INFO config-changed File >> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/config-changed", line 140, in >> <module> >> 2016-10-21 01:09:33 INFO config-changed main() >> 2016-10-21 01:09:33 INFO config-changed File >> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/config-changed", line 134, in >> main >> 2016-10-21 01:09:33 INFO config-changed hooks.execute(sys.argv) >> 2016-10-21 01:09:33 INFO config-changed File >> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/charmhelpers/core/hookenv.py", >> line 715, in execute >> 2016-10-21 01:09:33 INFO config-changed self._hooks[hook_name]() >> 2016-10-21 01:09:33 INFO config-changed File >> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/config-changed", line 78, in >> config_changed >> 2016-10-21 01:09:33 INFO config-changed configure_lxd_host() >> 2016-10-21 01:09:33 INFO config-changed File >> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/charmhelpers/core/decorators.py", >> line 40, in _retry_on_exception_inner_2 >> 2016-10-21 01:09:33 INFO config-changed return f(*args, **kwargs) >> 2016-10-21 01:09:33 INFO config-changed File >> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/lxd_utils.py", line 429, in >> configure_lxd_host >> 2016-10-21 01:09:33 INFO config-changed with open(EXT4_USERNS_MOUNTS, >> 'w') as userns_mounts: >> 2016-10-21 01:09:33 INFO config-changed IOError: [Errno 30] Read-only file >> system: '/sys/module/ext4/parameters/userns_mounts' >> 2016-10-21 01:09:33 ERROR juju.worker.uniter.operation runhook.go:107 hook >> "config-changed" failed: exit status 1 >> >> root@juju-456efd-13:~# touch /sys/module/ext4/parameters/temp-file >> touch: cannot touch '/sys/module/ext4/parameters/temp-file': Read-only file >> system >> root@juju-456efd-13:~# df -h /sys/module/ext4/parameters/userns_mounts >> Filesystem Size Used Avail Use% Mounted on >> sys 0 0 0 - /dev/.lxc/sys >> root@juju-456efd-13:~# touch /home/ubuntu/temp-file >> root@juju-456efd-13:~# ls /home/ubuntu/temp-file >> /home/ubuntu/temp-file >> root@juju-456efd-13:~# df -h >> Filesystem Size Used Avail Use% Mounted on >> /dev/mapper/mitaka--vg-root 165G 47G 110G 30% / >> none 492K 0 492K 0% /dev >> udev 16G 0 16G 0% /dev/fuse >> tmpfs 16G 0 16G 0% /dev/shm >> tmpfs 16G 49M 16G 1% /run >> tmpfs 5.0M 0 5.0M 0% /run/lock >> tmpfs 16G 0 16G 0% /sys/fs/cgroup >> tmpfs 3.2G 0 3.2G 0% /run/user/112 >> tmpfs 3.2G 0 3.2G 0% /run/user/1000 >> >> +++++++++++++++++++++++++++++++++++++++++ >> >> heather@mitaka:~$ nova boot --image d2eba22a-e1b1-4a2b-aa87-450ee9d9e492 >> --flavor d --nic net-name=ubuntu-net --key-name keypair-admin xenial-instance >> heather@mitaka:~/goose-work/src/gopkg.in/goose.v1$ >> <http://gopkg.in/goose.v1$> nova list >> +--------------------------------------+-----------------+--------+------------+-------------+-----------------------+ >> | ID | Name | Status | Task >> State | Power State | Networks | >> +--------------------------------------+-----------------+--------+------------+-------------+-----------------------+ >> | 80424b94-f24d-45ff-a330-7b67a911fbc6 | xenial-instance | ACTIVE | - >> | Running | ubuntu-net=10.101.0.8 | >> +--------------------------------------+-----------------+--------+------------+-------------+-----------------------+ >> >> heather@mitaka:~$ nova delete 80424b94-f24d-45ff-a330-7b67a911fbc6 >> Request to delete server 80424b94-f24d-45ff-a330-7b67a911fbc6 has been >> accepted. >> heather@mitaka:~$ nova list >> +--------------------------------------+-----------------+--------+------------+-------------+----------+ >> | ID | Name | Status | Task >> State | Power State | Networks | >> +--------------------------------------+-----------------+--------+------------+-------------+----------+ >> | 80424b94-f24d-45ff-a330-7b67a911fbc6 | xenial-instance | ERROR | - >> | Running | | >> +--------------------------------------+-----------------+--------+------------+-------------+----------+ >> heather@mitaka:~$ nova show 80424b94-f24d-45ff-a330-7b67a911fbc6 >> … >> | fault | {"message": "Failed to communicate >> with LXD API instance-00000006: Error 400 - Profile is currently in use.", >> "code": 500, "details": " File >> \"/usr/lib/python2.7/dist-packages/nova/compute/manager.py\", line 375, in >> decorated_function | >> | >> ... >> >> >> CONFIDENTIAL This electronic message contains information from TransCirrus, >> Inc. which may be confidential, privileged or otherwise protected from >> disclosure, and is to be used solely by the intended recipient. Any >> unauthorized review, disclosure, copying, distribution or use of this >> transmission or its contents is prohibited. If you have received this >> transmission in error, please notify the sender immediately by reply email >> to i...@transcirrus.com <mailto:i...@transcirrus.com>, and destroy all >> copies of the original message. Thank you. > >
-- Juju mailing list Juju@lists.ubuntu.com Modify settings or unsubscribe at: https://lists.ubuntu.com/mailman/listinfo/juju