Ran the job manually on rax VM, provided by Jeremy. (Thank you Jeremy).

After running 971 test cases VM inaccessible for 569 ticks, then continues... (Look at the console.log [1])
And also have a look at dstat log. [2]

The summary is:
======
Totals
======
Ran: 1125 tests in 5835.0000 sec.
 - Passed: 960
 - Skipped: 88
 - Expected Fail: 0
 - Unexpected Success: 0
 - Failed: 77
Sum of execute time for each test: 13603.6755 sec.


[1] https://etherpad.openstack.org/p/rax_console.txt
[2] https://etherpad.openstack.org/p/rax_dstat.log

On 02/24/2015 07:03 PM, Deepak Shetty wrote:
FWIW, we tried to run our job in a rax provider VM (provided by ianw from his personal account) and we ran the tempest tests twice, but the OOM did not re-create. Of the 2 runs, one of the run used the same PYTHONHASHSEED as we had in one of the failed runs, still no oom.

Jeremy graciously agreed to provide us 2 VMs , one each from rax and hpcloud provider
to see if provider platform has anything to do with it.

So we plan to run again wtih the VMs given from Jeremy , post which i will send
next update here.

thanx,
deepak


On Tue, Feb 24, 2015 at 4:50 AM, Jeremy Stanley <fu...@yuggoth.org <mailto:fu...@yuggoth.org>> wrote:

    Due to an image setup bug (I have a fix proposed currently), I was
    able to rerun this on a VM in HPCloud with 30GB memory and it
    completed in about an hour with a couple of tempest tests failing.
    Logs at: http://fungi.yuggoth.org/tmp/logs3.tar

    Rerunning again on another 8GB Rackspace VM with the job timeout
    increased to 5 hours, I was able to recreate the network
    connectivity issues exhibited previously. The job itself seems to
    have run for roughly 3 hours while failing 15 tests, and the worker
    was mostly unreachable for a while at the end (I don't know exactly
    how long) until around the time it completed. The OOM condition is
    present this time too according to the logs, occurring right near
    the end of the job. Collected logs are available at:
    http://fungi.yuggoth.org/tmp/logs4.tar

    Given the comparison between these two runs, I suspect this is
    either caused by memory constraints or block device I/O performance
    differences (or perhaps an unhappy combination of the two).
    Hopefully a close review of the logs will indicate which.
    --
    Jeremy Stanley

    __________________________________________________________________________
    OpenStack Development Mailing List (not for usage questions)
    Unsubscribe:
    openstack-dev-requ...@lists.openstack.org?subject:unsubscribe
    <http://openstack-dev-requ...@lists.openstack.org?subject:unsubscribe>
    http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev




__________________________________________________________________________
OpenStack Development Mailing List (not for usage questions)
Unsubscribe: openstack-dev-requ...@lists.openstack.org?subject:unsubscribe
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

--
Warm Regards,
Bharat Kumar Kobagana
Software Engineer
OpenStack Storage – RedHat India
Mobile - +91 9949278005

__________________________________________________________________________
OpenStack Development Mailing List (not for usage questions)
Unsubscribe: openstack-dev-requ...@lists.openstack.org?subject:unsubscribe
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

Reply via email to