[Canonical-ubuntu-qa] [Bug 1943841] Re: svm from ubuntu_kvm_unit_tests interrupt with "Unhandled exception 13 #GP at ip 00000000004027e3" on F-intel-5.13
There is a good chance that the error I mentioned, "Unhandled exception 6 #UD" (undefined instruction exception) is different and architecture specific (AMD EPYC 7713). However, other kernels referenced this bug for the same error. Rerunning this test to see if failure rate is 100%... 20:29:08 DEBUG| [stdout] FAIL: MSR_IA32_LASTBRANCHFROMIP, expected=0x401f4d, actual=0x401f4d 20:29:08 DEBUG| [stdout] PASS: Test that without LBRV enabled, guest LBR state does 'leak' to the host(1) 20:29:08 DEBUG| [stdout] Unhandled exception 6 #UD at ip 00401750 20:29:08 DEBUG| [stdout] error_code= rflags=00010086 cs=0008 20:29:08 DEBUG| [stdout] rax=004016db rcx=01dc rdx=804016db rbx=80010015 20:29:08 DEBUG| [stdout] rbp=0042fb68 rsi=0041776f rdi=00414d40 20:29:08 DEBUG| [stdout] r8=0041776f r9=03f8 r10=000d r11= 20:29:08 DEBUG| [stdout] r12= r13= r14= r15= 20:29:08 DEBUG| [stdout] cr0=80010011 cr2= cr3=010bf000 cr4=00040020 20:29:08 DEBUG| [stdout] cr8= 20:29:08 DEBUG| [stdout] STACK: @401750 4001d6 414df0 40bf1e 40bb72 4001d6 414df0 40bf1e 40bb72 4001d6 414df0 40bf1e 40bb72 4001d6 414df0 40bf1e 40bb72 4001d6 414df0 40bf1e 20:29:08 DEBUG| [stdout] FAIL svm -- You received this bug notification because you are a member of Canonical Platform QA Team, which is subscribed to ubuntu-kernel-tests. https://bugs.launchpad.net/bugs/1943841 Title: svm from ubuntu_kvm_unit_tests interrupt with "Unhandled exception 13 #GP at ip 004027e3" on F-intel-5.13 Status in ubuntu-kernel-tests: New Status in linux-intel-5.13 package in Ubuntu: New Bug description: Failing with Focal Intel 5.13.0-1004.4 on node gonzo A bit like bug 1934939, but this time it looks like this has passed through more cases than that bug. Running '/home/ubuntu/autotest/client/tmp/ubuntu_kvm_unit_tests/src/kvm-unit-tests/tests/svm' BUILD_HEAD=1593e88a timeout -k 1s --foreground 90s /usr/bin/qemu-system-x86_64 --no-reboot -nodefaults -device pc-testdev -device isa-debug-exit,iobase=0xf4,iosize=0x4 -vnc none -serial stdio -device pci-testdev -machine accel=kvm -kernel /tmp/tmp.8QIXdsXIyJ -smp 2 -cpu max,+svm -m 4g # -initrd /tmp/tmp.VvuuwXHRuX enabling apic enabling apic paging enabled cr0 = 80010011 cr3 = 10bf000 cr4 = 20 NPT detected - running all tests with NPT enabled PASS: null PASS: vmrun PASS: ioio PASS: vmrun intercept check PASS: rsm PASS: cr3 read intercept PASS: cr3 read nointercept PASS: cr3 read intercept emulate PASS: dr intercept check PASS: next_rip PASS: msr intercept check PASS: mode_switch PASS: asid_zero PASS: sel_cr0_bug PASS: npt_nx PASS: npt_np PASS: npt_us PASS: npt_rw PASS: npt_rw_pfwalk PASS: npt_l1mmio PASS: npt_rw_l1mmio PASS: tsc_adjust Latency VMRUN : max: 322434 min: 21233 avg: 36460 Latency VMEXIT: max: 298726 min: 16889 avg: 17717 PASS: latency_run_exit Latency VMRUN : max: 334040 min: 23446 avg: 36313 Latency VMEXIT: max: 310725 min: 16935 avg: 17618 PASS: latency_run_exit_clean Latency VMLOAD: max: 732466 min: 4675 avg: 4879 Latency VMSAVE: max: 60901 min: 4565 avg: 4850 Latency STGI: max: 43057 min: 3726 avg: 3862 Latency CLGI: max: 723675 min: 3644 avg: 3724 PASS: latency_svm_insn PASS: exception with vector 2 not injected PASS: divide overflow exception injected PASS: eventinj.VALID cleared PASS: exc_inject PASS: pending_event PASS: pending_event_cli PASS: direct interrupt while running guest PASS: intercepted interrupt while running guest PASS: direct interrupt + hlt PASS: intercepted interrupt + hlt PASS: interrupt PASS: direct NMI while running guest PASS: NMI intercept while running guest PASS: nmi PASS: direct NMI + hlt PASS: NMI intercept while running guest PASS: intercepted NMI + hlt PASS: nmi_hlt PASS: virq_inject PASS: No RIP corruption detected after 1 timer interrupts PASS: reg_corruption enabling apic PASS: svm_init_startup_test PASS: host_rflags PASS: CPUID.01H:ECX.XSAVE set before VMRUN PASS: svm_cr4_osxsave_test_guest finished with VMMCALL PASS: CPUID.01H:ECX.XSAVE set after VMRUN PASS: EFER.SVME: 1500 PASS: EFER.SVME: 500 PASS: Test EFER 9:8: 1700 PASS: Test EFER 63:16: 11500 PASS: Test EFER 63:16: 101500 PASS: Test EFER 63:16: 1001500 PASS: Test EFER 63:16: 10001500 PASS: Test EFER 63:16: 11500 PASS: Test EFER 63:16: 101500 PASS: Test EFER 63:16: 1001500 PASS: Test EFER 63:16: 10001500 PASS: Test EFER 63:16: 11500 PASS: Test EFER 63:16: 101500 PASS: Test EFER 63:16: 1001500 PASS: Test EFE
[Canonical-ubuntu-qa] [Merge] ~andersson123/qa-jenkins-jobs:fix-bad-substitution into qa-jenkins-jobs:master
The proposal to merge ~andersson123/qa-jenkins-jobs:fix-bad-substitution into qa-jenkins-jobs:master has been updated. Status: Needs review => Merged For more details, see: https://code.launchpad.net/~andersson123/qa-jenkins-jobs/+git/qa-jenkins-jobs/+merge/475506 -- Your team Canonical Platform QA Team is requested to review the proposed merge of ~andersson123/qa-jenkins-jobs:fix-bad-substitution into qa-jenkins-jobs:master. -- Mailing list: https://launchpad.net/~canonical-ubuntu-qa Post to : canonical-ubuntu-qa@lists.launchpad.net Unsubscribe : https://launchpad.net/~canonical-ubuntu-qa More help : https://help.launchpad.net/ListHelp
[Canonical-ubuntu-qa] [Merge] ~andersson123/qa-jenkins-jobs:fix-bad-substitution into qa-jenkins-jobs:master
Tim Andersson has proposed merging ~andersson123/qa-jenkins-jobs:fix-bad-substitution into qa-jenkins-jobs:master. Requested reviews: Canonical Platform QA Team (canonical-platform-qa) For more details, see: https://code.launchpad.net/~andersson123/qa-jenkins-jobs/+git/qa-jenkins-jobs/+merge/475506 Quick fix! I missed this in previous MPs as I manually configured the job, and forgot I did so. It then got overwritten by deploy-qa-jobs. -- Your team Canonical Platform QA Team is requested to review the proposed merge of ~andersson123/qa-jenkins-jobs:fix-bad-substitution into qa-jenkins-jobs:master. diff --git a/jobs/autopkgtest-cloud/jobs.yaml b/jobs/autopkgtest-cloud/jobs.yaml index 487efba..707e8ed 100644 --- a/jobs/autopkgtest-cloud/jobs.yaml +++ b/jobs/autopkgtest-cloud/jobs.yaml @@ -42,12 +42,12 @@ #!/bin/bash set -eu -RTD_LATEST_COMMIT=$(curl -H "Authorization: Token ${{RTD_TOKEN}}" https://readthedocs.org/api/v3/projects/autopkgtest-cloud/builds/ | jq -r '.results[0]' | jq -r '.commit') +RTD_LATEST_COMMIT=$(curl -H "Authorization: Token ${RTD_TOKEN}" https://readthedocs.org/api/v3/projects/autopkgtest-cloud/builds/ | jq -r '.results[0]' | jq -r '.commit') git clone https://git.launchpad.net/autopkgtest-cloud cd autopkgtest-cloud || exit 1 if ! git diff --quiet "${RTD_LATEST_COMMIT}" docs/ .readthedocs.yaml; then printf "Changes have been made to docs since last built version, building on readthedocs...\n" -curl -sS -X POST -H "Authorization: Token ${{RTD_TOKEN}}" https://readthedocs.org/api/v3/projects/autopkgtest-cloud/versions/latest/builds/ +curl -sS -X POST -H "Authorization: Token ${RTD_TOKEN}" https://readthedocs.org/api/v3/projects/autopkgtest-cloud/versions/latest/builds/ printf "readthedocs build request submitted\n" else printf "No changes to docs/ since latest build, not building\n" -- Mailing list: https://launchpad.net/~canonical-ubuntu-qa Post to : canonical-ubuntu-qa@lists.launchpad.net Unsubscribe : https://launchpad.net/~canonical-ubuntu-qa More help : https://help.launchpad.net/ListHelp
[Canonical-ubuntu-qa] [Bug 2071590] Re: test_vxlan_vnifiltering.sh from ubuntu_kselftests_net failed on linux-oem-6.8 (with ipv6 default rdst)
Encountered this with noble:linux-lowlatency 6.8.0-48.48.3 (both lowlatency and lowlatency-64k flavours) on openstack:cpu2-ram4-disk20 (arm). -- You received this bug notification because you are a member of Canonical Platform QA Team, which is subscribed to ubuntu-kernel-tests. https://bugs.launchpad.net/bugs/2071590 Title: test_vxlan_vnifiltering.sh from ubuntu_kselftests_net failed on linux- oem-6.8 (with ipv6 default rdst) Status in ubuntu-kernel-tests: New Bug description: Test complains about: TAP version 13 1..1 # timeout set to 0 # selftests: net: test_vxlan_vnifiltering.sh # TEST: Create traditional vxlan device [ OK ] # TEST: Cannot create vnifilter device without external flag [ OK ] # TEST: Creating external vxlan device with vnifilter flag[ OK ] # TEST: Cannot set in-use vni id on vnifiltering device [ OK ] # TEST: Set new vni id on vnifiltering device [ OK ] # TEST: Create second external vxlan device with vnifilter flag [ OK ] # TEST: Cannot set in-use vni id on vnifiltering device [ OK ] # TEST: Set new vni id on vnifiltering device [ OK ] # TEST: Update vni id on vnifiltering device [ OK ] # TEST: Add new vni id on vnifiltering device [ OK ] # TEST: Set multicast group on existing vni [ OK ] # TEST: Set multicast group on existing vni [ OK ] # TEST: Cannot set vnifilter flag on a device [ OK ] # TEST: Cannot unset vnifilter flag on a device [ OK ] # TEST: VM connectivity over vnifiltering vxlan (ipv4 default rdst) [ OK ] # TEST: VM connectivity over vnifiltering vxlan (ipv6 default rdst) [ OK ] # TEST: VM connectivity over vnifiltering vxlan pervni remote (ipv4 default rdst) [ OK ] # TEST: VM connectivity over vnifiltering vxlan pervni remote (ipv6 default rdst) [FAIL] # TEST: VM connectivity over vnifiltering vxlan mgroup (ipv4 default rdst) [ OK ] # TEST: VM connectivity over vnifiltering vxlan mgroup (ipv6 default rdst) [ OK ] # TEST: VM connectivity over vnifiltering vxlan pervni mgroup (ipv4 default rdst) [ OK ] # TEST: VM connectivity over vnifiltering vxlan pervni mgroup (ipv6 default rdst) [ OK ] # TEST: VM connectivity over vnifiltering vxlan pervni remote mix (ipv4 default rdst) [ OK ] # TEST: VM connectivity over vnifiltering vxlan pervni remote mix (ipv6 default rdst) [ OK ] # TEST: VM connectivity over traditional vxlan (ipv4 default rdst)[ OK ] # TEST: VM connectivity over traditional vxlan (ipv6 default rdst)[ OK ] # TEST: VM connectivity over metadata nonfiltering vxlan (ipv4 default rdst) [ OK ] # # Tests passed: 26 # Tests failed: 1 not ok 1 selftests: net: test_vxlan_vnifiltering.sh # exit=1 To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu-kernel-tests/+bug/2071590/+subscriptions -- Mailing list: https://launchpad.net/~canonical-ubuntu-qa Post to : canonical-ubuntu-qa@lists.launchpad.net Unsubscribe : https://launchpad.net/~canonical-ubuntu-qa More help : https://help.launchpad.net/ListHelp
Re: [Canonical-ubuntu-qa] [Merge] ~hyask/autopkgtest-cloud:skia/publish_db_copy_url_last_checked into autopkgtest-cloud:master
LGTM! Nice catch. -- https://code.launchpad.net/~hyask/autopkgtest-cloud/+git/autopkgtest-cloud/+merge/475520 Your team Canonical's Ubuntu QA is requested to review the proposed merge of ~hyask/autopkgtest-cloud:skia/publish_db_copy_url_last_checked into autopkgtest-cloud:master. -- Mailing list: https://launchpad.net/~canonical-ubuntu-qa Post to : canonical-ubuntu-qa@lists.launchpad.net Unsubscribe : https://launchpad.net/~canonical-ubuntu-qa More help : https://help.launchpad.net/ListHelp
[Canonical-ubuntu-qa] [Merge] ~hyask/autopkgtest-cloud:skia/publish_db_copy_url_last_checked into autopkgtest-cloud:master
Skia has proposed merging ~hyask/autopkgtest-cloud:skia/publish_db_copy_url_last_checked into autopkgtest-cloud:master. Requested reviews: Canonical's Ubuntu QA (canonical-ubuntu-qa) For more details, see: https://code.launchpad.net/~hyask/autopkgtest-cloud/+git/autopkgtest-cloud/+merge/475520 Quick win bringing nice performance improvement to `publish-db`. -- Your team Canonical's Ubuntu QA is requested to review the proposed merge of ~hyask/autopkgtest-cloud:skia/publish_db_copy_url_last_checked into autopkgtest-cloud:master. diff --git a/charms/focal/autopkgtest-web/webcontrol/publish-db b/charms/focal/autopkgtest-web/webcontrol/publish-db index 4149637..78b4b92 100755 --- a/charms/focal/autopkgtest-web/webcontrol/publish-db +++ b/charms/focal/autopkgtest-web/webcontrol/publish-db @@ -78,7 +78,7 @@ def init_db(path, path_current, path_rw): except sqlite3.OperationalError as e2: if "duplicate column name" not in str(e2): raise -logging.debug("table already exists") +logging.debug("table 'current_version' already exists") db.execute( "CREATE INDEX IF NOT EXISTS current_version_pocket_ix " "ON current_version(pocket, component)" @@ -110,21 +110,22 @@ def init_db(path, path_current, path_rw): " PRIMARY KEY(url))" ) logging.debug("database table url_last_checked created") -try: -if current_version_copied: -db.execute( -"INSERT INTO url_last_checked (url, timestamp) " -"SELECT url, timestamp FROM current.url_last_checked" -) -logging.debug("Old url_last_checked copied over") -except sqlite3.OperationalError as e: -if "no such table" not in str(e): -raise -logging.debug("no url_last_checked yet, first run probably") except sqlite3.OperationalError as e: if "already exists" not in str(e): raise -logging.debug("table already exists") +logging.debug("table 'url_last_checked' already exists") + +try: +if current_version_copied: +db.execute( +"INSERT OR REPLACE INTO url_last_checked (url, timestamp) " +"SELECT url, timestamp FROM current.url_last_checked" +) +logging.debug("Old url_last_checked copied over") +except sqlite3.OperationalError as e: +if "no such table" not in str(e): +raise +logging.debug("no url_last_checked yet, first run probably") db.commit() db.execute("DETACH DATABASE current") -- Mailing list: https://launchpad.net/~canonical-ubuntu-qa Post to : canonical-ubuntu-qa@lists.launchpad.net Unsubscribe : https://launchpad.net/~canonical-ubuntu-qa More help : https://help.launchpad.net/ListHelp
[Canonical-ubuntu-qa] [Bug 2009714] Re: 7 failures reported in net:fcnal-test.sh on Kinetic
Seen this with noble:linux-lowlatency 6.8.0-48.48.3 on openstack:cpu2-ram4-disk20 (arm64). SRU cycle: 2024.09.30. 20:00:21 DEBUG| [stdout] # # 20:00:21 DEBUG| [stdout] # Ping LLA with multiple interfaces 20:00:21 DEBUG| [stdout] # 20:00:26 DEBUG| [stdout] # TEST: Pre cycle, ping out ns-B - multicast IP [FAIL] 20:00:27 DEBUG| [stdout] # TEST: Pre cycle, ping out ns-C - multicast IP [ OK ] 20:00:30 DEBUG| [stdout] # TEST: Post cycle nsa-RcZfpc eth1, ping out ns-B - multicast IP[ OK ] 20:00:31 DEBUG| [stdout] # TEST: Post cycle nsa-RcZfpc eth1, ping out ns-C - multicast IP[ OK ] 20:00:34 DEBUG| [stdout] # TEST: Post cycle nsa-RcZfpc eth2, ping out ns-B - multicast IP[ OK ] 20:00:35 DEBUG| [stdout] # TEST: Post cycle nsa-RcZfpc eth2, ping out ns-C - multicast IP[ OK ] 20:00:36 DEBUG| [stdout] # -- You received this bug notification because you are a member of Canonical Platform QA Team, which is subscribed to ubuntu-kernel-tests. https://bugs.launchpad.net/bugs/2009714 Title: 7 failures reported in net:fcnal-test.sh on Kinetic Status in ubuntu-kernel-tests: New Bug description: This is not a regression, the test was skipped in the old days due to the lack of `nettest` command (bug 2006391) Unlike bug 2006716 ("5 failures reported in net:fcnal-test.sh on Jammy") there are 7 common failures reported on Kinetic # # IPv4 ping # # With VRF # SYSCTL: net.ipv4.raw_l3mdev_accept=1 # TEST: ping local, VRF bind - VRF IP [FAIL] # TEST: ping local, device bind - ns-A IP [FAIL] # SYSCTL: net.ipv4.ping_group_range=0 2147483647 # SYSCTL: net.ipv4.raw_l3mdev_accept=1 # TEST: ping local, VRF bind - VRF IP [FAIL] # TEST: ping local, device bind - ns-A IP [FAIL] # # IPv6 ping # # With VRF # SYSCTL: net.ipv4.ping_group_range=0 2147483647 # SYSCTL: net.ipv4.raw_l3mdev_accept=1 # TEST: ping out, vrf device+address bind - ns-B IPv6 LLA [FAIL] # # IPv6/UDP # # No VRF # udp_l3mdev_accept disabled # SYSCTL: net.ipv4.udp_l3mdev_accept=0 # TEST: Global server, device client via IP_UNICAST_IF, local connection, with connect() - ns-A loopback IPv6 [FAIL] # udp_l3mdev_accept enabled # SYSCTL: net.ipv4.udp_l3mdev_accept=1 # TEST: Global server, device client via IP_UNICAST_IF, local connection, with connect() - ns-A loopback IPv6 [FAIL] Sometimes it will fail with 8 failures, but the 8th failure looks flaky and can be failing on different test cases: With K-aws c6g.8xlarge (arm64) and r5.large (amd64) the 8th failure is: # # Use cases # # Ping LLA with multiple interfaces # TEST: Pre cycle, ping out ns-B - multicast IP[FAIL] With K-azure Standard_D2pds_v5 (arm64) the 8th failure is: # # Use cases # # TEST: Post cycle ns-A eth1, ping out ns-B - multicast IP [FAIL] To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu-kernel-tests/+bug/2009714/+subscriptions -- Mailing list: https://launchpad.net/~canonical-ubuntu-qa Post to : canonical-ubuntu-qa@lists.launchpad.net Unsubscribe : https://launchpad.net/~canonical-ubuntu-qa More help : https://help.launchpad.net/ListHelp
[Canonical-ubuntu-qa] [Bug 1931390] Re: cpuhotplug03 in cpuhotplug from ubuntu_ltp failed after successful CPU1 offline
I have seen this issue for noble:linux-lowlatency 6.8.0-48.48.3 on openstack:cpu2-ram4-disk20 (arm64). (SRU Cycle: 2024.09.30) I think this test is flaky. To check if "cpuhotplug_do_spin_loop" it uses "ps psr" command where "r" option lists the currently running processes only. Even if a "cpuhotplug_do_spin_loop" process did migrate to target CPU, this does not mean that it is currently running there (as it can be preempted by availability of other tasks). And for architectures with lower number of cpus (in my case 2), test is likely to fail more. Moreover, test doesn't seem to adjust "spin" tasks' priority (I am not sure about this but it doesn't seem to do anything in this regard prior to running the test either). Test needs to either; - run in isolated environment where OS threads run on a dedicated core. - sleep some time (as done currently) and check if the spin process is migrated by the spin process itself. Here is the log: FO: Test start time: Fri Oct 18 20:06:59 UTC 2024 ND:/opt/ltp/bin/ltp-pan -q -e -S -a 166236 -n 166236 -f /tmp/ltp-9srhcvY2Gx/alltests -l /dev/null -C /dev/null -T /dev/null OG File: /dev/null ED COMMAND File: /dev/null NF COMMAND File: /dev/null ng tests... me: cpuhotplug03 te: Fri Oct 18 20:06:59 UTC 2024 sc: Do tasks get scheduled to a newly on-lined CPU? 08 CPU is 1 ER PID %CPU %MEMVSZ RSS TTY STAT START TIME COMMAND ot 166323 12.8 0.0 2380 1408 ?R20:06 0:00 /bin/sh /opt/ltp/testcases/bin/cpuhotplug_do_spin_loop ot 166324 12.8 0.0 2380 1408 ?R20:06 0:00 /bin/sh /opt/ltp/testcases/bin/cpuhotplug_do_spin_loop ot 166325 12.8 0.0 2380 1408 ?R20:06 0:00 /bin/sh /opt/ltp/testcases/bin/cpuhotplug_do_spin_loop ot 166326 12.8 0.0 2380 1408 ?R20:06 0:00 /bin/sh /opt/ltp/testcases/bin/cpuhotplug_do_spin_loop ot 166331 0.0 0.0 6680 1920 ?S20:06 0:00 grep cpuhotplug_do_spin_loop 03 1 TINFO: Onlining CPU 1 0 /bin/sh /opt/ltp/testcases/bin/cpuhotplug_do_spin_loop 0 /bin/sh /opt/ltp/testcases/bin/cpuhotplug_do_spin_loop 0 /bin/sh /opt/ltp/testcases/bin/cpuhotplug_do_spin_loop 0 /bin/sh /opt/ltp/testcases/bin/cpuhotplug_do_spin_loop 03 1 TFAIL: No cpuhotplug_do_spin_loop processes found on CPU1 FO: ltp-pan reported some tests FAIL TP Version: 20230929-874-gba610da01 FO: Test end time: Fri Oct 18 20:07:01 UTC 2024 -- You received this bug notification because you are a member of Canonical Platform QA Team, which is subscribed to ubuntu-kernel-tests. https://bugs.launchpad.net/bugs/1931390 Title: cpuhotplug03 in cpuhotplug from ubuntu_ltp failed after successful CPU1 offline Status in ubuntu-kernel-tests: New Status in linux package in Ubuntu: Incomplete Bug description: In cpuhotplug (ubuntu_ltp) tests, offline and online of CPU1 succeeds, but later the test fails with: utils:0153| [stdout] Name: cpuhotplug03 utils:0153| [stdout] Date: Wed Jun 9 03:52:16 UTC 2021 utils:0153| [stdout] Desc: Do tasks get scheduled to a newly on-lined CPU? utils:0153| [stdout] utils:0153| [stdout] CPU is 1 utils:0153| [stdout] USER PID %CPU %MEM VSZ RSS TTY STAT START TIME COMMAND utils:0153| [stdout] root 858611 0.0 0.0 2616 540 ? R 03:52 0:00 /bin/sh /opt/ltp/testcases/bin/cpuhotplug_do_spin_loop ... ... ... utils:0153| [stdout] 49 /bin/sh /opt/ltp/testcases/bin/cpuhotplug_do_spin_loop utils:0153| [stdout] 82 /bin/sh /opt/ltp/testcases/bin/cpuhotplug_do_spin_loop utils:0153| [stdout] 18 /bin/sh /opt/ltp/testcases/bin/cpuhotplug_do_spin_loop utils:0153| [stdout] cpuhotplug03 1 TFAIL: No cpuhotplug_do_spin_loop processes found on CPU1 utils:0153| [stdout] tag=cpuhotplug03 stime=1623210736 dur=3 exit=exited stat=1 core=no cu=22260 cs=78 IMPORTANT: The CPU1 got properly offlined and onlined. For failures of CPU1 offline, please see lp:1836167 (reported previously as part of lp:1836167). Po-Hsu Lin (cypressyew) wrote on 2020-09-25: On F-oem-5.6 with node glameow: cpuhotplug03 1 TFAIL: No cpuhotplug_do_spin_loop processes found on CPU1 Kelsey Skunberg (kelsey-skunberg) wrote on 2020-10-06: spotted on Focal aws : 5.4.0-1026.26 : amd64 cpuhotplug03 1 TFAIL: No cpuhotplug_do_spin_loop processes found on CPU1 Po-Hsu Lin (cypressyew) wrote on 2021-06-09: Found on 5.11.0-1005.5 - intel Node spitfire, passed on node bavor. cpuhotplug03 1 TFAIL: No cpuhotplug_do_spin_loop processes found on CPU1 To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu-kernel-tests/+bug/1931390/+subscriptions -- Mailing list: https://launchpad.net/~canonical-ubuntu-qa Post to : canonical-ubuntu-qa@lists.launchpad.net Unsubscribe : https://launchpad.net/~canonical-ubuntu-qa More help : https://help.launchpad.net/ListHelp
[Canonical-ubuntu-qa] [Bug 1943841] Re: svm from ubuntu_kvm_unit_tests interrupt with "Unhandled exception 13 #GP at ip 00000000004027e3" on F-intel-5.13
Similar error happens with noble:linux-lowlatency 6.8.0-48.48.3 on riccioli (SRU cycle: 2024.09.30) 20:28:53 INFO | STARTubuntu_kvm_unit_tests.svm ubuntu_kvm_unit_tests.svmtimeout=1800timestamp=172928 localtime=Oct 18 20:28:53 20:28:53 DEBUG| Persistent state client._record_indent now set to 2 20:28:53 DEBUG| Persistent state client.unexpected_reboot now set to ('ubuntu_kvm_unit_tests.svm', 'ubuntu_kvm_unit_tests.svm') 20:28:53 DEBUG| Waiting for pid 18157 for 1800 seconds 20:28:53 WARNI| System python is too old, crash handling disabled 20:28:53 DEBUG| Running 'kvm-ok' 20:28:53 DEBUG| [stdout] INFO: /dev/kvm exists 20:28:53 DEBUG| [stdout] KVM acceleration can be used 20:28:53 DEBUG| Running '/home/ubuntu/autotest/client/tmp/ubuntu_kvm_unit_tests/src/kvm-unit-tests/tests/svm' 20:28:53 DEBUG| [stdout] BUILD_HEAD=b04954c9 20:28:53 DEBUG| [stdout] timeout -k 1s --foreground 90s /usr/bin/qemu-system-x86_64 --no-reboot -nodefaults -device pc-testdev -device isa-debug-exit,iobase=0xf4,iosize=0x4 -vnc none -serial stdio -device pci-testdev -machine accel=kvm -kernel /tmp/tmp.jgXFu3YdzN -smp 2 -cpu max,+svm -m 4g -append -pause_filter_test # -initrd /tmp/tmp.jZ1NEkg5BI 20:28:54 DEBUG| [stdout] enabling apic 20:28:54 DEBUG| [stdout] smp: waiting for 1 APs 20:28:54 DEBUG| [stdout] enabling apic 20:28:54 DEBUG| [stdout] setup: CPU 1 online 20:28:54 DEBUG| [stdout] paging enabled 20:28:54 DEBUG| [stdout] cr0 = 80010011 20:28:54 DEBUG| [stdout] cr3 = 10bf000 20:28:54 DEBUG| [stdout] cr4 = 20 20:28:54 DEBUG| [stdout] NPT detected - running all tests with NPT enabled 20:28:54 DEBUG| [stdout] PASS: null 20:28:54 DEBUG| [stdout] PASS: vmrun 20:28:54 DEBUG| [stdout] PASS: ioio 20:28:54 DEBUG| [stdout] PASS: vmrun intercept check 20:28:54 DEBUG| [stdout] PASS: rsm 20:28:54 DEBUG| [stdout] PASS: cr3 read intercept 20:28:54 DEBUG| [stdout] PASS: cr3 read nointercept 20:28:54 DEBUG| [stdout] PASS: cr3 read intercept emulate 20:28:54 DEBUG| [stdout] PASS: dr intercept check 20:28:54 DEBUG| [stdout] PASS: next_rip 20:28:54 DEBUG| [stdout] PASS: msr intercept check 20:28:54 DEBUG| [stdout] PASS: mode_switch 20:28:54 DEBUG| [stdout] PASS: asid_zero 20:28:54 DEBUG| [stdout] PASS: sel_cr0_bug 20:28:54 DEBUG| [stdout] PASS: tsc_adjust 20:28:58 DEBUG| [stdout] Latency VMRUN : max: 4793660 min: 6460 avg: 8899 20:28:58 DEBUG| [stdout] Latency VMEXIT: max: 4766600 min: 3580 avg: 3774 20:28:58 DEBUG| [stdout] PASS: latency_run_exit 20:29:03 DEBUG| [stdout] Latency VMRUN : max: 2145480 min: 8480 avg: 8898 20:29:03 DEBUG| [stdout] Latency VMEXIT: max: 2133480 min: 3580 avg: 3788 20:29:03 DEBUG| [stdout] PASS: latency_run_exit_clean 20:29:03 DEBUG| [stdout] Latency VMLOAD: max: 2443180 min: 260 avg: 275 20:29:03 DEBUG| [stdout] Latency VMSAVE: max: 1971420 min: 240 avg: 263 20:29:03 DEBUG| [stdout] Latency STGI: max: 9740 min: 40 avg: 49 20:29:03 DEBUG| [stdout] Latency CLGI: max: 46560 min: 40 avg: 53 20:29:03 DEBUG| [stdout] PASS: latency_svm_insn 20:29:03 DEBUG| [stdout] PASS: exception with vector 2 not injected 20:29:03 DEBUG| [stdout] PASS: divide overflow exception injected 20:29:03 DEBUG| [stdout] PASS: eventinj.VALID cleared 20:29:03 DEBUG| [stdout] PASS: exc_inject 20:29:03 DEBUG| [stdout] PASS: pending_event 20:29:03 DEBUG| [stdout] PASS: pending_event_cli 20:29:03 DEBUG| [stdout] PASS: direct interrupt while running guest 20:29:03 DEBUG| [stdout] PASS: intercepted interrupt while running guest 20:29:03 DEBUG| [stdout] PASS: direct interrupt + hlt 20:29:03 DEBUG| [stdout] PASS: intercepted interrupt + hlt 20:29:03 DEBUG| [stdout] PASS: interrupt 20:29:03 DEBUG| [stdout] PASS: direct NMI while running guest 20:29:03 DEBUG| [stdout] PASS: NMI intercept while running guest 20:29:03 DEBUG| [stdout] PASS: nmi 20:29:03 DEBUG| [stdout] PASS: direct NMI + hlt 20:29:03 DEBUG| [stdout] PASS: NMI intercept while running guest 20:29:03 DEBUG| [stdout] PASS: intercepted NMI + hlt 20:29:03 DEBUG| [stdout] PASS: nmi_hlt 20:29:03 DEBUG| [stdout] PASS: virq_inject 20:29:05 DEBUG| [stdout] PASS: No RIP corruption detected after 1 timer interrupts 20:29:05 DEBUG| [stdout] PASS: reg_corruption 20:29:05 DEBUG| [stdout] enabling apic 20:29:05 DEBUG| [stdout] setup: CPU 1 online 20:29:05 DEBUG| [stdout] PASS: svm_init_startup_test 20:29:05 DEBUG| [stdout] PASS: host_rflags 20:29:05 DEBUG| [stdout] PASS: STGI set VGIF bit. 20:29:05 DEBUG| [stdout] PASS: CLGI cleared VGIF bit. 20:29:05 DEBUG| [stdout] PASS: vgif 20:29:05 DEBUG| [stdout] PASS: CPUID.01H:ECX.XSAVE set before VMRUN 20:29:05 DEBUG| [stdout] PASS: svm_cr4_osxsave_test_guest finished with VMMCALL 20:29:05 DEBUG| [stdout] PASS: CPUID.01H:ECX.XSAVE set after VMRUN 20:29:05 DEBUG| [stdout] PASS: EFER.SVME: 1500 20:29:05 DEBUG| [stdout] PASS: EFER.SVME: 500 20:29:05 DEBUG| [stdout] PASS: Test EFER 9:8: 1700 20:29:05 DEBUG| [stdout] PASS: Test EFER 63:16: 11500 20:29:05 DEBUG| [stdout] PASS: Test EFER 63:16: 101500 20:2
[Canonical-ubuntu-qa] [Bug 1955529] Re: cpuset_memory_pressure in ubuntu_ltp_controllers failed on B-oracle BM.Standard2.52 (root group's memory_pressure didn't have memory pressure rate.)
Seen this on n/lowlatency 6.8.0-48.48.3 on fili (SRU cycle: 2024.09.30) -- You received this bug notification because you are a member of Canonical Platform QA Team, which is subscribed to ubuntu-kernel-tests. https://bugs.launchpad.net/bugs/1955529 Title: cpuset_memory_pressure in ubuntu_ltp_controllers failed on B-oracle BM.Standard2.52 (root group's memory_pressure didn't have memory pressure rate.) Status in ubuntu-kernel-tests: New Bug description: Issue found on Oracle instance BM.Standard2.52 only with kernel 4.15.0-1085.93 This failure can be found in 4.15.0-1084.92 / 4.15.0-1083.91 as well, thus it's not a regression. Test failed with: cpuset_memory_pressure 11 TFAIL: root group's memory_pressure didn't have memory pressure rate. Test log: Checking for required user/group ids 'nobody' user id and group found. 'bin' user id and group found. 'daemon' user id and group found. Users group found. Sys group found. Required users/groups exist. no big block device was specified on commandline. Tests which require a big block device are disabled. You can specify it with option -z INFO: Test start time: Tue Dec 21 15:07:18 UTC 2021 COMMAND:/opt/ltp/bin/ltp-pan -q -e -S -a 57071 -n 57071 -f /tmp/ltp-9znr7PBmp3/alltests -l /dev/null -C /dev/null -T /dev/null LOG File: /dev/null FAILED COMMAND File: /dev/null TCONF COMMAND File: /dev/null Running tests... cpuset_memory_pressure 1 TINFO: CPUs are numbered continuously starting at 0 (0-103) cpuset_memory_pressure 1 TINFO: Nodes are numbered continuously starting at 0 (0-1) cpuset_memory_pressure 1 TPASS: Cpuset memory pressure test succeeded. cpuset_memory_pressure 3 TPASS: Cpuset memory pressure test succeeded. cpuset_memory_pressure 5 TPASS: Cpuset memory pressure test succeeded. Killed cpuset_memory_pressure 7 TPASS: Cpuset memory pressure test succeeded. cpuset_memory_pressure 9 TFAIL: root group's memory_pressure didn't have memory pressure rate. Killed cpuset_memory_pressure 11 TFAIL: root group's memory_pressure didn't have memory pressure rate. INFO: ltp-pan reported some tests FAIL LTP Version: 20210927 INFO: Test end time: Tue Dec 21 15:48:45 UTC 2021 To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu-kernel-tests/+bug/1955529/+subscriptions -- Mailing list: https://launchpad.net/~canonical-ubuntu-qa Post to : canonical-ubuntu-qa@lists.launchpad.net Unsubscribe : https://launchpad.net/~canonical-ubuntu-qa More help : https://help.launchpad.net/ListHelp
[Canonical-ubuntu-qa] [Bug 2083687] Re: cachestat03 from ubuntu_ltp_syscalls failed on B-KVM
** Tags added: 5.15 5.4 focal jammy -- You received this bug notification because you are a member of Canonical Platform QA Team, which is subscribed to ubuntu-kernel-tests. https://bugs.launchpad.net/bugs/2083687 Title: cachestat03 from ubuntu_ltp_syscalls failed on B-KVM Status in ubuntu-kernel-tests: New Bug description: This is a new test case added on Jul. 23 https://github.com/linux-test-project/ltp/commit/1564e8c4700c0bd4b17282f5560d6f1774798005 And we got our fork updated recently in order to cover Oracular. So this is not a regression. B-4.15, B-hwe-5.4 are good. It's just failing for B-KVM for now. Test failed with: TBROK: Failed to open FILE '/proc/sys/vm/compact_memory' for writing: ENOENT (2) Test log: COMMAND:/opt/ltp/bin/ltp-pan -e -S -a 8849 -n 8849 -p -f /tmp/ltp-8M7ll8HANh/alltests -l /opt/ltp/results/LTP_RUN_ON-2024_10_04-11h_49m_21s.log -C /opt/ltp/output/LTP_RUN_ON-2024_10_04-11h_49m_21s.failed -T /opt/ltp/output/LTP_RUN_ON-2024_10_04-11h_49m_21s.tconf LOG File: /opt/ltp/results/LTP_RUN_ON-2024_10_04-11h_49m_21s.log FAILED COMMAND File: /opt/ltp/output/LTP_RUN_ON-2024_10_04-11h_49m_21s.failed TCONF COMMAND File: /opt/ltp/output/LTP_RUN_ON-2024_10_04-11h_49m_21s.tconf Running tests... <<>> tag=cachestat03 stime=1728042561 cmdline="cachestat03" contacts="" analysis=exit <<>> incrementing stop tst_hugepage.c:50: TBROK: Failed to open FILE '/proc/sys/vm/compact_memory' for writing: ENOENT (2) <<>> initiation_status="ok" duration=1 termination_type=exited termination_id=2 corefile=no cutime=0 cstime=37 <<>> INFO: ltp-pan reported some tests FAIL To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu-kernel-tests/+bug/2083687/+subscriptions -- Mailing list: https://launchpad.net/~canonical-ubuntu-qa Post to : canonical-ubuntu-qa@lists.launchpad.net Unsubscribe : https://launchpad.net/~canonical-ubuntu-qa More help : https://help.launchpad.net/ListHelp
[Canonical-ubuntu-qa] [Bug 2083687] Re: cachestat03 from ubuntu_ltp_syscalls failed on B-KVM
** Tags added: sru-20240930 -- You received this bug notification because you are a member of Canonical Platform QA Team, which is subscribed to ubuntu-kernel-tests. https://bugs.launchpad.net/bugs/2083687 Title: cachestat03 from ubuntu_ltp_syscalls failed on B-KVM Status in ubuntu-kernel-tests: New Bug description: This is a new test case added on Jul. 23 https://github.com/linux-test-project/ltp/commit/1564e8c4700c0bd4b17282f5560d6f1774798005 And we got our fork updated recently in order to cover Oracular. So this is not a regression. B-4.15, B-hwe-5.4 are good. It's just failing for B-KVM for now. Test failed with: TBROK: Failed to open FILE '/proc/sys/vm/compact_memory' for writing: ENOENT (2) Test log: COMMAND:/opt/ltp/bin/ltp-pan -e -S -a 8849 -n 8849 -p -f /tmp/ltp-8M7ll8HANh/alltests -l /opt/ltp/results/LTP_RUN_ON-2024_10_04-11h_49m_21s.log -C /opt/ltp/output/LTP_RUN_ON-2024_10_04-11h_49m_21s.failed -T /opt/ltp/output/LTP_RUN_ON-2024_10_04-11h_49m_21s.tconf LOG File: /opt/ltp/results/LTP_RUN_ON-2024_10_04-11h_49m_21s.log FAILED COMMAND File: /opt/ltp/output/LTP_RUN_ON-2024_10_04-11h_49m_21s.failed TCONF COMMAND File: /opt/ltp/output/LTP_RUN_ON-2024_10_04-11h_49m_21s.tconf Running tests... <<>> tag=cachestat03 stime=1728042561 cmdline="cachestat03" contacts="" analysis=exit <<>> incrementing stop tst_hugepage.c:50: TBROK: Failed to open FILE '/proc/sys/vm/compact_memory' for writing: ENOENT (2) <<>> initiation_status="ok" duration=1 termination_type=exited termination_id=2 corefile=no cutime=0 cstime=37 <<>> INFO: ltp-pan reported some tests FAIL To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu-kernel-tests/+bug/2083687/+subscriptions -- Mailing list: https://launchpad.net/~canonical-ubuntu-qa Post to : canonical-ubuntu-qa@lists.launchpad.net Unsubscribe : https://launchpad.net/~canonical-ubuntu-qa More help : https://help.launchpad.net/ListHelp
[Canonical-ubuntu-qa] [Bug 2056461] Re: autopkgtest-virt-qemu on noble images sometimes hangs doing copydown
Performing the verification on Noble. First, we check to see if we can reproduce the problem. $ apt policy autopkgtest autopkgtest: Installed: 5.34ubuntu2 Candidate: 5.34ubuntu2 Version table: 5.38ubuntu1~24.04.1 100 100 http://archive.ubuntu.com/ubuntu noble-proposed/main amd64 Packages *** 5.34ubuntu2 500 500 http://archive.ubuntu.com/ubuntu noble/main amd64 Packages 100 /var/lib/dpkg/status After creating a Noble image we run the reproducer: $ autopkgtest -U -ddd hello*.dsc -- qemu ../autopkgtest-noble-amd64.img autopkgtest: DBG: autopkgtest options: Namespace(override_control=None, only_tests=[], skip_tests=None, built_binaries=True, architecture=None, packages=['hello_2.10-3build2.dsc'], output_dir=None) autopkgtest: DBG: virt-runner arguments: ['qemu', '../autopkgtest-noble-amd64.img'] autopkgtest: DBG: actions: [('source', 'hello_2.10-3build2.dsc', True)] autopkgtest: DBG: build binaries: True autopkgtest: DBG: testbed init autopkgtest [12:09:07]: starting date and time: 2024-10-21 12:09:07-0700 autopkgtest [12:09:07]: version 5.34ubuntu2 autopkgtest [12:09:07]: host clean-noble-amd64; command line: /usr/bin/autopkgtest -U -ddd hello_2.10-3build2.dsc -- qemu ../autopkgtest-noble-amd64.img (truncated for brevity) autopkgtest: DBG: sending command to testbed: copydown hello_2.10.orig.tar.gz /tmp/autopkgtest.gHgSJl/hello_2.10.orig.tar.gz autopkgtest: DBG: got reply from testbed: timeout autopkgtest: DBG: sending command to testbed: auxverb_debug_fail autopkgtest: DBG: got reply from testbed: ok autopkgtest: DBG: TestbedFailure sent `auxverb_debug_fail', got `timeout', expected `ok...' autopkgtest: DBG: testbed stop autopkgtest: DBG: testbed close, scratch=/tmp/autopkgtest.gHgSJl autopkgtest: DBG: sending command to testbed: close qemu-system-x86_64: terminating on signal 15 from pid 7523 (/usr/bin/python3) autopkgtest: DBG: got reply from testbed: ok autopkgtest: DBG: sending command to testbed: quit autopkgtest [12:14:36]: ERROR: testbed failure: sent `auxverb_debug_fail', got `timeout', expected `ok...' The testbed hung until timeout during copydown, so the bug is replicated. Now we verify the fix. $ apt policy autopkgtest autopkgtest: Installed: 5.38ubuntu1~24.04.1 Candidate: 5.38ubuntu1~24.04.1 Version table: *** 5.38ubuntu1~24.04.1 100 100 http://archive.ubuntu.com/ubuntu noble-proposed/main amd64 Packages 100 /var/lib/dpkg/status 5.34ubuntu2 500 500 http://archive.ubuntu.com/ubuntu noble/main amd64 Packages We create a fresh Noble image and rerun the reproducer: $ autopkgtest -U hello*.dsc -- qemu ../autopkgtest-noble-amd64.img autopkgtest [12:35:03]: starting date and time: 2024-10-21 12:35:03-0700 autopkgtest [12:35:03]: version 5.38ubuntu1~24.04.1 autopkgtest [12:35:03]: host clean-noble-amd64; command line: /usr/bin/autopkgtest -U hello_2.10-3build2.dsc -- qemu ../autopkgtest-noble-amd64.img autopkgtest [12:35:19]: testbed dpkg architecture: amd64 autopkgtest [12:35:20]: testbed apt version: 2.7.14build2 autopkgtest [12:35:20]: test bed setup (truncated for brevity) autopkgtest [12:37:04]: test upstream-tests: [--- Testing greeting-1 ... Testing hello-1 ... Testing last-1 ... Testing traditional-1 ... autopkgtest [12:37:05]: test upstream-tests: ---] autopkgtest [12:37:06]: test upstream-tests: - - - - - - - - - - results - - - - - - - - - - upstream-tests PASS autopkgtest [12:37:06]: summary command1 PASS upstream-tests PASS So the test was completed without a timeout. This concludes the verification for Noble. -- You received this bug notification because you are a member of Canonical's Ubuntu QA, which is subscribed to autopkgtest in Ubuntu. Matching subscriptions: ubuntu-qa-bugs https://bugs.launchpad.net/bugs/2056461 Title: autopkgtest-virt-qemu on noble images sometimes hangs doing copydown Status in Linux: Confirmed Status in autopkgtest package in Ubuntu: Fix Released Status in linux package in Ubuntu: In Progress Status in autopkgtest source package in Jammy: In Progress Status in linux source package in Jammy: New Status in autopkgtest source package in Noble: In Progress Status in linux source package in Noble: New Status in autopkgtest package in Debian: Fix Released Bug description: [Impact] It seems that kernel 6.8 introduced a regression in the 9pfs related to caching and netfslib, that can cause some user-space apps to read content from files that is not up-to-date (when they are used in a producer/consumer fashion). It seems that the offending commit is this one: 80105ed2fd27 ("9p: Use netfslib read/write_iter") Reverting the commit seems to fix the problem. However the actual bug might be in netfslib or how netfslib is used in the 9p context. The regression has been reported upstream and we are still investigating (https://lore.
[Canonical-ubuntu-qa] [Bug 2056461] Re: autopkgtest-virt-qemu on noble images sometimes hangs doing copydown
Performing the verification on Jammy. First, we check if we can reproduce the problem. $ apt policy autopkgtest autopkgtest: Installed: 5.32ubuntu3~22.04.1 Candidate: 5.38ubuntu1~22.04.1 Version table: 5.38ubuntu1~22.04.1 500 500 http://archive.ubuntu.com/ubuntu jammy-proposed/main amd64 Packages 500 http://archive.ubuntu.com/ubuntu jammy-proposed/main i386 Packages *** 5.32ubuntu3~22.04.1 500 500 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 Packages 500 http://archive.ubuntu.com/ubuntu jammy-updates/main i386 Packages 100 /var/lib/dpkg/status 5.20 500 500 http://archive.ubuntu.com/ubuntu jammy/main amd64 Packages 500 http://archive.ubuntu.com/ubuntu jammy/main i386 Packages After creating a Noble image we run the reproducer: $ autopkgtest -U -ddd hello*.dsc -- qemu ../autopkgtest-noble-amd64.img autopkgtest: DBG: autopkgtest options: Namespace(override_control=None, only_tests=[], skip_tests=None, built_binaries=True, architecture=None, packages=['hello_2.10-3build2.dsc'], output_dir=None) autopkgtest: DBG: virt-runner arguments: ['qemu', '../autopkgtest-noble-amd64.img'] autopkgtest: DBG: actions: [('source', 'hello_2.10-3build2.dsc', True)] autopkgtest: DBG: build binaries: True autopkgtest: DBG: testbed init autopkgtest [20:52:54]: starting date and time: 2024-10-21 20:52:54+0100 autopkgtest [20:52:54]: version 5.32ubuntu3~22.04.1 autopkgtest [20:52:54]: host clean-jammy-amd64; command line: /usr/bin/autopkgtest -U -ddd hello_2.10-3build2.dsc -- qemu ../autopkgtest-noble-amd64.img (truncated for brevity) autopkgtest: DBG: sending command to testbed: copydown /tmp/autopkgtest.output.cbb3dd47/tests-tree/ /tmp/autopkgtest.ShBCzZ/build.DY9/src/ autopkgtest: DBG: got reply from testbed: timeout autopkgtest: DBG: sending command to testbed: auxverb_debug_fail autopkgtest: DBG: got reply from testbed: ok autopkgtest: DBG: TestbedFailure sent `auxverb_debug_fail', got `timeout', expected `ok...' autopkgtest: DBG: testbed stop autopkgtest: DBG: testbed close, scratch=/tmp/autopkgtest.ShBCzZ autopkgtest: DBG: sending command to testbed: close qemu-system-x86_64: terminating on signal 15 from pid 4744 (/usr/bin/python3) autopkgtest: DBG: got reply from testbed: ok autopkgtest: DBG: sending command to testbed: quit autopkgtest [21:02:35]: ERROR: testbed failure: sent `auxverb_debug_fail', got `timeout', expected `ok...' The testbed hung until timeout during copydown, so the bug is verified. Now we verify the fix. $ apt policy autopkgtest autopkgtest: Installed: 5.38ubuntu1~22.04.1 Candidate: 5.38ubuntu1~22.04.1 Version table: *** 5.38ubuntu1~22.04.1 500 500 http://archive.ubuntu.com/ubuntu jammy-proposed/main amd64 Packages 500 http://archive.ubuntu.com/ubuntu jammy-proposed/main i386 Packages 100 /var/lib/dpkg/status 5.32ubuntu3~22.04.1 500 500 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 Packages 500 http://archive.ubuntu.com/ubuntu jammy-updates/main i386 Packages 5.20 500 500 http://archive.ubuntu.com/ubuntu jammy/main amd64 Packages 500 http://archive.ubuntu.com/ubuntu jammy/main i386 Packages $ autopkgtest -U hello*.dsc -- qemu ../autopkgtest-noble-amd64.img autopkgtest [21:29:25]: starting date and time: 2024-10-21 21:29:25+0100 autopkgtest [21:29:25]: version 5.38ubuntu1~22.04.1 autopkgtest [21:29:25]: host clean-jammy-amd64; command line: /usr/bin/autopkgtest -U hello_2.10-3build2.dsc -- qemu ../autopkgtest-noble-amd64.img autopkgtest [21:29:41]: testbed dpkg architecture: amd64 autopkgtest [21:29:41]: testbed apt version: 2.7.14build2 autopkgtest [21:29:41]: test bed setup (truncated for brevity) autopkgtest [21:31:32]: test upstream-tests: [--- Testing greeting-1 ... Testing hello-1 ... Testing last-1 ... Testing traditional-1 ... autopkgtest [21:31:33]: test upstream-tests: ---] autopkgtest [21:31:34]: test upstream-tests: - - - - - - - - - - results - - - - - - - - - - upstream-tests PASS autopkgtest [21:31:34]: summary command1 PASS upstream-tests PASS So the test was completed without a timeout. This concludes the verification for Jammy. ** Tags added: verification-done-jammy verification-done-noble -- You received this bug notification because you are a member of Canonical's Ubuntu QA, which is subscribed to autopkgtest in Ubuntu. Matching subscriptions: ubuntu-qa-bugs https://bugs.launchpad.net/bugs/2056461 Title: autopkgtest-virt-qemu on noble images sometimes hangs doing copydown Status in Linux: Confirmed Status in autopkgtest package in Ubuntu: Fix Released Status in linux package in Ubuntu: In Progress Status in autopkgtest source package in Jammy: In Progress Status in linux source package in Jammy: New Status in autopkgtest source package in Noble: In Progress Status in linux