Hi All - we need a LP bug for this, please.  The gerrit review will need to
have its commit message updated with the lp bug info so that CI can do bug
management on it.  Then, once it lands in the charm's master, we can
cherry-pick it for a stable backport to the stable charm.
Thanks,
Ryan

On Fri, Oct 21, 2016 at 9:54 AM, Chuck Short <chuck.sh...@canonical.com>
wrote:

> Hi,
>
> I proposed a fix:
>
> https://review.openstack.org/#/c/389740/
>
> chuck
>
> On Fri, Oct 21, 2016 at 10:27 AM, Adam Stokes <adam.sto...@canonical.com>
> wrote:
>
>> This looks like it's due to the way we deploy OpenStack with NovaLXD in
>> all containers, this effectively breaks anyone wanting to do an all-in-one
>> install on their system.
>>
>> On Fri, Oct 21, 2016 at 10:22 AM Adam Stokes <adam.sto...@canonical.com>
>> wrote:
>>
>>> So it looks like a recent change to the LXD charm, see here:
>>>
>>> https://github.com/openstack/charm-lxd/commit/017246768e097c
>>> 5fcd5283e23f19f075ff9f9d4e
>>>
>>> Chuck, are you aware of this issue?
>>>
>>> On Fri, Oct 21, 2016 at 10:19 AM Heather Lanigan <hmlani...@gmail.com>
>>> wrote:
>>>
>>> Adam,
>>>
>>> The entire container is not readonly.  Just /sys, the mount point for
>>> /dev/.lxc/sys.  I choose another charm (neutron-api) to look at, /sys on
>>> that unit is readonly as well.    Is that normal?
>>>
>>> What would be different in my config?  My Xenial install is on a VM, but
>>> I’ve been running that way for weeks.  I did have the openstack-novalxd
>>> bundle successfully deployed on it previously using juju 2.0_rc1.
>>>
>>> -Heather
>>>
>>> On Oct 20, 2016, at 11:30 PM, Adam Stokes <adam.sto...@canonical.com>
>>> wrote:
>>>
>>> Odd it looks like the container has a read only file system? I ran
>>> through a full openstack-novalxd deployment today and one of the upstream
>>> maintainers ran through the same deployment and didn't run into any issues.
>>>
>>> On Thu, Oct 20, 2016, 10:02 PM Heather Lanigan <hmlani...@gmail.com>
>>> wrote:
>>>
>>>
>>> I used conjure-up to deploy openstack-novalxd on a Xenial system.
>>> Before deploying, the operating system was updated.  LXD init was setup
>>> with dir, not xfs.  All but one of the charms has a status of “unit is
>>> ready"
>>>
>>> The lxd/0 subordinate charm has a status of: hook failed:
>>> "config-changed”.  See details below.
>>>
>>> I can boot an instance within this OpenStack deployment.  However
>>> deleting the instance fails. A side effect of the lxd/0 issues?
>>>
>>> Juju version 2.0.0-xenial-amd64
>>> conjure-up version 2.0.2
>>> lxd charm version 2.0.5
>>>
>>> Any ideas?
>>>
>>> Thanks in advance,
>>> Heather
>>>
>>> ++++++++++++++++++++++++++++++++++++++++++++++
>>>
>>> The /var/log/juju/unit-lxd-0.log on the unit reports:
>>>
>>> 2016-10-21 01:09:33 INFO config-changed Traceback (most recent call
>>> last):
>>> 2016-10-21 01:09:33 INFO config-changed   File
>>> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/config-changed", line 140,
>>> in <module>
>>> 2016-10-21 01:09:33 INFO config-changed     main()
>>> 2016-10-21 01:09:33 INFO config-changed   File
>>> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/config-changed", line 134,
>>> in main
>>> 2016-10-21 01:09:33 INFO config-changed     hooks.execute(sys.argv)
>>> 2016-10-21 01:09:33 INFO config-changed   File
>>> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/charmhelpers/core/hookenv.py",
>>> line 715, in execute
>>> 2016-10-21 01:09:33 INFO config-changed     self._hooks[hook_name]()
>>> 2016-10-21 01:09:33 INFO config-changed   File
>>> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/config-changed", line 78,
>>> in config_changed
>>> 2016-10-21 01:09:33 INFO config-changed     configure_lxd_host()
>>> 2016-10-21 01:09:33 INFO config-changed   File
>>> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/charmhelpers/core/decorators.py",
>>> line 40, in _retry_on_exception_inner_2
>>> 2016-10-21 01:09:33 INFO config-changed     return f(*args, **kwargs)
>>> 2016-10-21 01:09:33 INFO config-changed   File
>>> "/var/lib/juju/agents/unit-lxd-0/charm/hooks/lxd_utils.py", line 429,
>>> in configure_lxd_host
>>> 2016-10-21 01:09:33 INFO config-changed     with
>>> open(EXT4_USERNS_MOUNTS, 'w') as userns_mounts:
>>> 2016-10-21 01:09:33 INFO config-changed IOError: [Errno 30] Read-only
>>> file system: '/sys/module/ext4/parameters/userns_mounts'
>>> 2016-10-21 01:09:33 ERROR juju.worker.uniter.operation runhook.go:107
>>> hook "config-changed" failed: exit status 1
>>>
>>>
>>> root@juju-456efd-13:~# touch /sys/module/ext4/parameters/temp-file
>>> touch: cannot touch '/sys/module/ext4/parameters/temp-file': Read-only
>>> file system
>>> root@juju-456efd-13:~# df -h /sys/module/ext4/parameters/userns_mounts
>>> Filesystem      Size  Used Avail Use% Mounted on
>>> sys                0     0     0    - /dev/.lxc/sys
>>> root@juju-456efd-13:~# touch /home/ubuntu/temp-file
>>> root@juju-456efd-13:~# ls /home/ubuntu/temp-file
>>> /home/ubuntu/temp-file
>>> root@juju-456efd-13:~# df -h
>>> Filesystem                   Size  Used Avail Use% Mounted on
>>> /dev/mapper/mitaka--vg-root  165G   47G  110G  30% /
>>> none                         492K     0  492K   0% /dev
>>> udev                          16G     0   16G   0% /dev/fuse
>>> tmpfs                         16G     0   16G   0% /dev/shm
>>> tmpfs                         16G   49M   16G   1% /run
>>> tmpfs                        5.0M     0  5.0M   0% /run/lock
>>> tmpfs                         16G     0   16G   0% /sys/fs/cgroup
>>> tmpfs                        3.2G     0  3.2G   0% /run/user/112
>>> tmpfs                        3.2G     0  3.2G   0% /run/user/1000
>>>
>>> +++++++++++++++++++++++++++++++++++++++++
>>>
>>> heather@mitaka:~$ nova boot --image d2eba22a-e1b1-4a2b-aa87-450ee9d9e492
>>> --flavor d --nic net-name=ubuntu-net --key-name keypair-admin
>>> xenial-instance
>>> heather@mitaka:~/goose-work/src/gopkg.in/goose.v1$ nova list
>>> +--------------------------------------+-----------------+--
>>> ------+------------+-------------+-----------------------+
>>> | ID                                   | Name            | Status | Task
>>> State | Power State | Networks              |
>>> +--------------------------------------+-----------------+--
>>> ------+------------+-------------+-----------------------+
>>> | 80424b94-f24d-45ff-a330-7b67a911fbc6 | xenial-instance | ACTIVE | -
>>>        | Running     | ubuntu-net=10.101.0.8 |
>>> +--------------------------------------+-----------------+--
>>> ------+------------+-------------+-----------------------+
>>>
>>> heather@mitaka:~$ nova delete 80424b94-f24d-45ff-a330-7b67a911fbc6
>>> Request to delete server 80424b94-f24d-45ff-a330-7b67a911fbc6 has been
>>> accepted.
>>> heather@mitaka:~$ nova list
>>> +--------------------------------------+-----------------+--
>>> ------+------------+-------------+----------+
>>> | ID                                   | Name            | Status | Task
>>> State | Power State | Networks |
>>> +--------------------------------------+-----------------+--
>>> ------+------------+-------------+----------+
>>> | 80424b94-f24d-45ff-a330-7b67a911fbc6 | xenial-instance | ERROR  | -
>>>        | Running     |          |
>>> +--------------------------------------+-----------------+--
>>> ------+------------+-------------+----------+
>>> heather@mitaka:~$ nova show 80424b94-f24d-45ff-a330-7b67a911fbc6
>>> …
>>> | fault                                | {"message": "Failed to
>>> communicate with LXD API instance-00000006: Error 400 - Profile is
>>> currently in use.", "code": 500, "details": "  File
>>> \"/usr/lib/python2.7/dist-packages/nova/compute/manager.py\", line 375,
>>> in decorated_function |
>>> |
>>> ...
>>>
>>>
>>> CONFIDENTIAL This electronic message contains information from
>>> TransCirrus, Inc. which may be confidential, privileged or otherwise
>>> protected from disclosure, and is to be used solely by the intended
>>> recipient. Any unauthorized review, disclosure, copying, distribution or
>>> use of this transmission or its contents is prohibited. If you have
>>> received this transmission in error, please notify the sender immediately
>>> by reply email to i...@transcirrus.com, and destroy all copies of the
>>> original message. Thank you.
>>>
>>>
>>>
>
> --
> Juju mailing list
> Juju@lists.ubuntu.com
> Modify settings or unsubscribe at: https://lists.ubuntu.com/
> mailman/listinfo/juju
>
>
-- 
Juju mailing list
Juju@lists.ubuntu.com
Modify settings or unsubscribe at: 
https://lists.ubuntu.com/mailman/listinfo/juju

Reply via email to