Update based on IRC conversation:  I tried the fix.  Turns out that (my) Xenial 
install doesn’t have the systemd-virt-detect command.

-Heather

> On Oct 21, 2016, at 10:54 AM, Chuck Short <chuck.sh...@canonical.com> wrote:
> 
> Hi,
> 
> I proposed a fix:
> 
> https://review.openstack.org/#/c/389740/ 
> <https://review.openstack.org/#/c/389740/>
> 
> chuck
> 
> On Fri, Oct 21, 2016 at 10:27 AM, Adam Stokes <adam.sto...@canonical.com 
> <mailto:adam.sto...@canonical.com>> wrote:
> This looks like it's due to the way we deploy OpenStack with NovaLXD in all 
> containers, this effectively breaks anyone wanting to do an all-in-one 
> install on their system.
> 
> On Fri, Oct 21, 2016 at 10:22 AM Adam Stokes <adam.sto...@canonical.com 
> <mailto:adam.sto...@canonical.com>> wrote:
> So it looks like a recent change to the LXD charm, see here:
> 
> https://github.com/openstack/charm-lxd/commit/017246768e097c5fcd5283e23f19f075ff9f9d4e
>  
> <https://github.com/openstack/charm-lxd/commit/017246768e097c5fcd5283e23f19f075ff9f9d4e>
> 
> Chuck, are you aware of this issue?
> 
> On Fri, Oct 21, 2016 at 10:19 AM Heather Lanigan <hmlani...@gmail.com 
> <mailto:hmlani...@gmail.com>> wrote:
> Adam,
> 
> The entire container is not readonly.  Just /sys, the mount point for 
> /dev/.lxc/sys.  I choose another charm (neutron-api) to look at, /sys on that 
> unit is readonly as well.    Is that normal?
> 
> What would be different in my config?  My Xenial install is on a VM, but I’ve 
> been running that way for weeks.  I did have the openstack-novalxd bundle 
> successfully deployed on it previously using juju 2.0_rc1.
> 
> -Heather
> 
>> On Oct 20, 2016, at 11:30 PM, Adam Stokes <adam.sto...@canonical.com 
>> <mailto:adam.sto...@canonical.com>> wrote:
>> 
>> Odd it looks like the container has a read only file system? I ran through a 
>> full openstack-novalxd deployment today and one of the upstream maintainers 
>> ran through the same deployment and didn't run into any issues.
>> 
>> 
>> On Thu, Oct 20, 2016, 10:02 PM Heather Lanigan <hmlani...@gmail.com 
>> <mailto:hmlani...@gmail.com>> wrote:
>> 
>> I used conjure-up to deploy openstack-novalxd on a Xenial system.  Before 
>> deploying, the operating system was updated.  LXD init was setup with dir, 
>> not xfs.  All but one of the charms has a status of “unit is ready"
>> 
>> The lxd/0 subordinate charm has a status of: hook failed: "config-changed”.  
>> See details below.
>> 
>> I can boot an instance within this OpenStack deployment.  However deleting 
>> the instance fails. A side effect of the lxd/0 issues?
>> 
>> Juju version 2.0.0-xenial-amd64
>> conjure-up version 2.0.2
>> lxd charm version 2.0.5
>> 
>> Any ideas?
>> 
>> Thanks in advance,
>> Heather
>> 
>> ++++++++++++++++++++++++++++++++++++++++++++++
>> 
>> The /var/log/juju/unit-lxd-0.log on the unit reports:
>> 2016-10-21 01:09:33 INFO config-changed Traceback (most recent call last):
>> 2016-10-21 01:09:33 INFO config-changed   File 
>> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/config-changed", line 140, in 
>> <module>
>> 2016-10-21 01:09:33 INFO config-changed     main()
>> 2016-10-21 01:09:33 INFO config-changed   File 
>> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/config-changed", line 134, in 
>> main
>> 2016-10-21 01:09:33 INFO config-changed     hooks.execute(sys.argv)
>> 2016-10-21 01:09:33 INFO config-changed   File 
>> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/charmhelpers/core/hookenv.py", 
>> line 715, in execute
>> 2016-10-21 01:09:33 INFO config-changed     self._hooks[hook_name]()
>> 2016-10-21 01:09:33 INFO config-changed   File 
>> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/config-changed", line 78, in 
>> config_changed
>> 2016-10-21 01:09:33 INFO config-changed     configure_lxd_host()
>> 2016-10-21 01:09:33 INFO config-changed   File 
>> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/charmhelpers/core/decorators.py",
>>  line 40, in _retry_on_exception_inner_2
>> 2016-10-21 01:09:33 INFO config-changed     return f(*args, **kwargs)
>> 2016-10-21 01:09:33 INFO config-changed   File 
>> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/lxd_utils.py", line 429, in 
>> configure_lxd_host
>> 2016-10-21 01:09:33 INFO config-changed     with open(EXT4_USERNS_MOUNTS, 
>> 'w') as userns_mounts:
>> 2016-10-21 01:09:33 INFO config-changed IOError: [Errno 30] Read-only file 
>> system: '/sys/module/ext4/parameters/userns_mounts'
>> 2016-10-21 01:09:33 ERROR juju.worker.uniter.operation runhook.go:107 hook 
>> "config-changed" failed: exit status 1
>> 
>> root@juju-456efd-13:~# touch /sys/module/ext4/parameters/temp-file
>> touch: cannot touch '/sys/module/ext4/parameters/temp-file': Read-only file 
>> system
>> root@juju-456efd-13:~# df -h /sys/module/ext4/parameters/userns_mounts
>> Filesystem      Size  Used Avail Use% Mounted on
>> sys                0     0     0    - /dev/.lxc/sys
>> root@juju-456efd-13:~# touch /home/ubuntu/temp-file
>> root@juju-456efd-13:~# ls /home/ubuntu/temp-file
>> /home/ubuntu/temp-file
>> root@juju-456efd-13:~# df -h
>> Filesystem                   Size  Used Avail Use% Mounted on
>> /dev/mapper/mitaka--vg-root  165G   47G  110G  30% /
>> none                         492K     0  492K   0% /dev
>> udev                          16G     0   16G   0% /dev/fuse
>> tmpfs                         16G     0   16G   0% /dev/shm
>> tmpfs                         16G   49M   16G   1% /run
>> tmpfs                        5.0M     0  5.0M   0% /run/lock
>> tmpfs                         16G     0   16G   0% /sys/fs/cgroup
>> tmpfs                        3.2G     0  3.2G   0% /run/user/112
>> tmpfs                        3.2G     0  3.2G   0% /run/user/1000
>> 
>> +++++++++++++++++++++++++++++++++++++++++
>> 
>> heather@mitaka:~$ nova boot --image d2eba22a-e1b1-4a2b-aa87-450ee9d9e492 
>> --flavor d --nic net-name=ubuntu-net --key-name keypair-admin xenial-instance
>> heather@mitaka:~/goose-work/src/gopkg.in/goose.v1$ 
>> <http://gopkg.in/goose.v1$> nova list
>> +--------------------------------------+-----------------+--------+------------+-------------+-----------------------+
>> | ID                                   | Name            | Status | Task 
>> State | Power State | Networks              |
>> +--------------------------------------+-----------------+--------+------------+-------------+-----------------------+
>> | 80424b94-f24d-45ff-a330-7b67a911fbc6 | xenial-instance | ACTIVE | -        
>>   | Running     | ubuntu-net=10.101.0.8 |
>> +--------------------------------------+-----------------+--------+------------+-------------+-----------------------+
>> 
>> heather@mitaka:~$ nova delete 80424b94-f24d-45ff-a330-7b67a911fbc6
>> Request to delete server 80424b94-f24d-45ff-a330-7b67a911fbc6 has been 
>> accepted.
>> heather@mitaka:~$ nova list
>> +--------------------------------------+-----------------+--------+------------+-------------+----------+
>> | ID                                   | Name            | Status | Task 
>> State | Power State | Networks |
>> +--------------------------------------+-----------------+--------+------------+-------------+----------+
>> | 80424b94-f24d-45ff-a330-7b67a911fbc6 | xenial-instance | ERROR  | -        
>>   | Running     |          |
>> +--------------------------------------+-----------------+--------+------------+-------------+----------+
>> heather@mitaka:~$ nova show 80424b94-f24d-45ff-a330-7b67a911fbc6
>> …
>> | fault                                | {"message": "Failed to communicate 
>> with LXD API instance-00000006: Error 400 - Profile is currently in use.", 
>> "code": 500, "details": "  File 
>> \"/usr/lib/python2.7/dist-packages/nova/compute/manager.py\", line 375, in 
>> decorated_function |
>> |                                     
>> ...
>> 
>> 
>> CONFIDENTIAL This electronic message contains information from TransCirrus, 
>> Inc. which may be confidential, privileged or otherwise protected from 
>> disclosure, and is to be used solely by the intended recipient. Any 
>> unauthorized review, disclosure, copying, distribution or use of this 
>> transmission or its contents is prohibited. If you have received this 
>> transmission in error, please notify the sender immediately by reply email 
>> to i...@transcirrus.com <mailto:i...@transcirrus.com>, and destroy all 
>> copies of the original message. Thank you.
> 
> 

-- 
Juju mailing list
Juju@lists.ubuntu.com
Modify settings or unsubscribe at: 
https://lists.ubuntu.com/mailman/listinfo/juju

Reply via email to