[Canonical-ubuntu-qa] [Bug 1943841] Re: svm from ubuntu_kvm_unit_tests interrupt with "Unhandled exception 13 #GP at ip 00000000004027e3" on F-intel-5.13

2024-10-21 Thread Mehmet Basaran
There is a good chance that the error I mentioned, "Unhandled exception
6 #UD" (undefined instruction exception) is different and architecture
specific (AMD EPYC 7713). However, other kernels referenced this bug for
the same error.

Rerunning this test to see if failure rate is 100%...


20:29:08 DEBUG| [stdout] FAIL: MSR_IA32_LASTBRANCHFROMIP, expected=0x401f4d, 
actual=0x401f4d
20:29:08 DEBUG| [stdout] PASS: Test that without LBRV enabled, guest LBR state 
does 'leak' to the host(1)
20:29:08 DEBUG| [stdout] Unhandled exception 6 #UD at ip 00401750
20:29:08 DEBUG| [stdout] error_code=  rflags=00010086  cs=0008
20:29:08 DEBUG| [stdout] rax=004016db rcx=01dc 
rdx=804016db rbx=80010015
20:29:08 DEBUG| [stdout] rbp=0042fb68 rsi=0041776f 
rdi=00414d40
20:29:08 DEBUG| [stdout]  r8=0041776f  r9=03f8 
r10=000d r11=
20:29:08 DEBUG| [stdout] r12= r13= 
r14= r15=
20:29:08 DEBUG| [stdout] cr0=80010011 cr2= 
cr3=010bf000 cr4=00040020
20:29:08 DEBUG| [stdout] cr8=
20:29:08 DEBUG| [stdout] STACK: @401750 4001d6 414df0 40bf1e 40bb72 4001d6 
414df0 40bf1e 40bb72 4001d6 414df0 40bf1e 40bb72 4001d6 414df0 40bf1e 40bb72 
4001d6 414df0 40bf1e
20:29:08 DEBUG| [stdout] FAIL svm

-- 
You received this bug notification because you are a member of Canonical
Platform QA Team, which is subscribed to ubuntu-kernel-tests.
https://bugs.launchpad.net/bugs/1943841

Title:
  svm from ubuntu_kvm_unit_tests interrupt with "Unhandled exception 13
  #GP at ip 004027e3" on F-intel-5.13

Status in ubuntu-kernel-tests:
  New
Status in linux-intel-5.13 package in Ubuntu:
  New

Bug description:
  Failing with Focal Intel 5.13.0-1004.4 on node gonzo

  A bit like bug 1934939, but this time it looks like this has passed
  through more cases than that bug.

  Running 
'/home/ubuntu/autotest/client/tmp/ubuntu_kvm_unit_tests/src/kvm-unit-tests/tests/svm'
   BUILD_HEAD=1593e88a
   timeout -k 1s --foreground 90s /usr/bin/qemu-system-x86_64 --no-reboot 
-nodefaults -device pc-testdev -device isa-debug-exit,iobase=0xf4,iosize=0x4 
-vnc none -serial stdio -device pci-testdev -machine accel=kvm -kernel 
/tmp/tmp.8QIXdsXIyJ -smp 2 -cpu max,+svm -m 4g # -initrd /tmp/tmp.VvuuwXHRuX
   enabling apic
   enabling apic
   paging enabled
   cr0 = 80010011
   cr3 = 10bf000
   cr4 = 20
   NPT detected - running all tests with NPT enabled
   PASS: null
   PASS: vmrun
   PASS: ioio
   PASS: vmrun intercept check
   PASS: rsm
   PASS: cr3 read intercept
   PASS: cr3 read nointercept
   PASS: cr3 read intercept emulate
   PASS: dr intercept check
   PASS: next_rip
   PASS: msr intercept check
   PASS: mode_switch
   PASS: asid_zero
   PASS: sel_cr0_bug
   PASS: npt_nx
   PASS: npt_np
   PASS: npt_us
   PASS: npt_rw
   PASS: npt_rw_pfwalk
   PASS: npt_l1mmio
   PASS: npt_rw_l1mmio
   PASS: tsc_adjust
   Latency VMRUN : max: 322434 min: 21233 avg: 36460
   Latency VMEXIT: max: 298726 min: 16889 avg: 17717
   PASS: latency_run_exit
   Latency VMRUN : max: 334040 min: 23446 avg: 36313
   Latency VMEXIT: max: 310725 min: 16935 avg: 17618
   PASS: latency_run_exit_clean
   Latency VMLOAD: max: 732466 min: 4675 avg: 4879
   Latency VMSAVE: max: 60901 min: 4565 avg: 4850
   Latency STGI:   max: 43057 min: 3726 avg: 3862
   Latency CLGI:   max: 723675 min: 3644 avg: 3724
   PASS: latency_svm_insn
   PASS: exception with vector 2 not injected
   PASS: divide overflow exception injected
   PASS: eventinj.VALID cleared
   PASS: exc_inject
   PASS: pending_event
   PASS: pending_event_cli
   PASS: direct interrupt while running guest
   PASS: intercepted interrupt while running guest
   PASS: direct interrupt + hlt
   PASS: intercepted interrupt + hlt
   PASS: interrupt
   PASS: direct NMI while running guest
   PASS: NMI intercept while running guest
   PASS: nmi
   PASS: direct NMI + hlt
   PASS: NMI intercept while running guest
   PASS: intercepted NMI + hlt
   PASS: nmi_hlt
   PASS: virq_inject
   PASS: No RIP corruption detected after 1 timer interrupts
   PASS: reg_corruption
   enabling apic
   PASS: svm_init_startup_test
   PASS: host_rflags
   PASS: CPUID.01H:ECX.XSAVE set before VMRUN
   PASS: svm_cr4_osxsave_test_guest finished with VMMCALL
   PASS: CPUID.01H:ECX.XSAVE set after VMRUN
   PASS: EFER.SVME: 1500
   PASS: EFER.SVME: 500
   PASS: Test EFER 9:8: 1700
   PASS: Test EFER 63:16: 11500
   PASS: Test EFER 63:16: 101500
   PASS: Test EFER 63:16: 1001500
   PASS: Test EFER 63:16: 10001500
   PASS: Test EFER 63:16: 11500
   PASS: Test EFER 63:16: 101500
   PASS: Test EFER 63:16: 1001500
   PASS: Test EFER 63:16: 10001500
   PASS: Test EFER 63:16: 11500
   PASS: Test EFER 63:16: 101500
   PASS: Test EFER 63:16: 1001500
   PASS: Test EFE

[Canonical-ubuntu-qa] [Merge] ~andersson123/qa-jenkins-jobs:fix-bad-substitution into qa-jenkins-jobs:master

2024-10-21 Thread mp+475506
The proposal to merge ~andersson123/qa-jenkins-jobs:fix-bad-substitution into 
qa-jenkins-jobs:master has been updated.

Status: Needs review => Merged

For more details, see:
https://code.launchpad.net/~andersson123/qa-jenkins-jobs/+git/qa-jenkins-jobs/+merge/475506
-- 
Your team Canonical Platform QA Team is requested to review the proposed merge 
of ~andersson123/qa-jenkins-jobs:fix-bad-substitution into 
qa-jenkins-jobs:master.


-- 
Mailing list: https://launchpad.net/~canonical-ubuntu-qa
Post to : canonical-ubuntu-qa@lists.launchpad.net
Unsubscribe : https://launchpad.net/~canonical-ubuntu-qa
More help   : https://help.launchpad.net/ListHelp


[Canonical-ubuntu-qa] [Merge] ~andersson123/qa-jenkins-jobs:fix-bad-substitution into qa-jenkins-jobs:master

2024-10-21 Thread Tim Andersson
Tim Andersson has proposed merging 
~andersson123/qa-jenkins-jobs:fix-bad-substitution into qa-jenkins-jobs:master.

Requested reviews:
  Canonical Platform QA Team (canonical-platform-qa)

For more details, see:
https://code.launchpad.net/~andersson123/qa-jenkins-jobs/+git/qa-jenkins-jobs/+merge/475506

Quick fix! I missed this in previous MPs as I manually configured the job, and 
forgot I did so. It then got overwritten by deploy-qa-jobs.
-- 
Your team Canonical Platform QA Team is requested to review the proposed merge 
of ~andersson123/qa-jenkins-jobs:fix-bad-substitution into 
qa-jenkins-jobs:master.
diff --git a/jobs/autopkgtest-cloud/jobs.yaml b/jobs/autopkgtest-cloud/jobs.yaml
index 487efba..707e8ed 100644
--- a/jobs/autopkgtest-cloud/jobs.yaml
+++ b/jobs/autopkgtest-cloud/jobs.yaml
@@ -42,12 +42,12 @@
 #!/bin/bash
 set -eu
 
-RTD_LATEST_COMMIT=$(curl -H "Authorization: Token ${{RTD_TOKEN}}" https://readthedocs.org/api/v3/projects/autopkgtest-cloud/builds/ | jq -r '.results[0]' | jq -r '.commit')
+RTD_LATEST_COMMIT=$(curl -H "Authorization: Token ${RTD_TOKEN}" https://readthedocs.org/api/v3/projects/autopkgtest-cloud/builds/ | jq -r '.results[0]' | jq -r '.commit')
 git clone https://git.launchpad.net/autopkgtest-cloud
 cd autopkgtest-cloud || exit 1
 if ! git diff --quiet "${RTD_LATEST_COMMIT}" docs/ .readthedocs.yaml; then
 printf "Changes have been made to docs since last built version, building on readthedocs...\n"
-curl -sS -X POST -H "Authorization: Token ${{RTD_TOKEN}}" https://readthedocs.org/api/v3/projects/autopkgtest-cloud/versions/latest/builds/
+curl -sS -X POST -H "Authorization: Token ${RTD_TOKEN}" https://readthedocs.org/api/v3/projects/autopkgtest-cloud/versions/latest/builds/
 printf "readthedocs build request submitted\n"
 else
 printf "No changes to docs/ since latest build, not building\n"
-- 
Mailing list: https://launchpad.net/~canonical-ubuntu-qa
Post to : canonical-ubuntu-qa@lists.launchpad.net
Unsubscribe : https://launchpad.net/~canonical-ubuntu-qa
More help   : https://help.launchpad.net/ListHelp


[Canonical-ubuntu-qa] [Bug 2071590] Re: test_vxlan_vnifiltering.sh from ubuntu_kselftests_net failed on linux-oem-6.8 (with ipv6 default rdst)

2024-10-21 Thread Mehmet Basaran
Encountered this with noble:linux-lowlatency 6.8.0-48.48.3 (both
lowlatency and lowlatency-64k flavours) on openstack:cpu2-ram4-disk20
(arm).

-- 
You received this bug notification because you are a member of Canonical
Platform QA Team, which is subscribed to ubuntu-kernel-tests.
https://bugs.launchpad.net/bugs/2071590

Title:
  test_vxlan_vnifiltering.sh from ubuntu_kselftests_net failed on linux-
  oem-6.8 (with ipv6 default rdst)

Status in ubuntu-kernel-tests:
  New

Bug description:
  Test complains about:
  TAP version 13
  1..1
  # timeout set to 0
  # selftests: net: test_vxlan_vnifiltering.sh
  # TEST: Create traditional vxlan device   [ 
OK ]
  # TEST: Cannot create vnifilter device without external flag  [ 
OK ]
  # TEST: Creating external vxlan device with vnifilter flag[ 
OK ]
  # TEST: Cannot set in-use vni id on vnifiltering device   [ 
OK ]
  # TEST: Set new vni id on vnifiltering device [ 
OK ]
  # TEST: Create second external vxlan device with vnifilter flag   [ 
OK ]
  # TEST: Cannot set in-use vni id on vnifiltering device   [ 
OK ]
  # TEST: Set new vni id on vnifiltering device [ 
OK ]
  # TEST: Update vni id on vnifiltering device  [ 
OK ]
  # TEST: Add new vni id on vnifiltering device [ 
OK ]
  # TEST: Set multicast group on existing vni   [ 
OK ]
  # TEST: Set multicast group on existing vni   [ 
OK ]
  # TEST: Cannot set vnifilter flag on a device [ 
OK ]
  # TEST: Cannot unset vnifilter flag on a device   [ 
OK ]
  # TEST: VM connectivity over vnifiltering vxlan (ipv4 default rdst)   [ 
OK ]
  # TEST: VM connectivity over vnifiltering vxlan (ipv6 default rdst)   [ 
OK ]
  # TEST: VM connectivity over vnifiltering vxlan pervni remote (ipv4 
default rdst)  [ OK ]
  # TEST: VM connectivity over vnifiltering vxlan pervni remote (ipv6 
default rdst)  [FAIL]
  # TEST: VM connectivity over vnifiltering vxlan mgroup (ipv4 default 
rdst)  [ OK ]
  # TEST: VM connectivity over vnifiltering vxlan mgroup (ipv6 default 
rdst)  [ OK ]
  # TEST: VM connectivity over vnifiltering vxlan pervni mgroup (ipv4 
default rdst)  [ OK ]
  # TEST: VM connectivity over vnifiltering vxlan pervni mgroup (ipv6 
default rdst)  [ OK ]
  # TEST: VM connectivity over vnifiltering vxlan pervni remote mix (ipv4 
default rdst)  [ OK ]
  # TEST: VM connectivity over vnifiltering vxlan pervni remote mix (ipv6 
default rdst)  [ OK ]
  # TEST: VM connectivity over traditional vxlan (ipv4 default rdst)[ 
OK ]
  # TEST: VM connectivity over traditional vxlan (ipv6 default rdst)[ 
OK ]
  # TEST: VM connectivity over metadata nonfiltering vxlan (ipv4 default 
rdst)  [ OK ]
  # 
  # Tests passed:  26
  # Tests failed:   1
  not ok 1 selftests: net: test_vxlan_vnifiltering.sh # exit=1

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-kernel-tests/+bug/2071590/+subscriptions


-- 
Mailing list: https://launchpad.net/~canonical-ubuntu-qa
Post to : canonical-ubuntu-qa@lists.launchpad.net
Unsubscribe : https://launchpad.net/~canonical-ubuntu-qa
More help   : https://help.launchpad.net/ListHelp


Re: [Canonical-ubuntu-qa] [Merge] ~hyask/autopkgtest-cloud:skia/publish_db_copy_url_last_checked into autopkgtest-cloud:master

2024-10-21 Thread Tim Andersson
LGTM! Nice catch.
-- 
https://code.launchpad.net/~hyask/autopkgtest-cloud/+git/autopkgtest-cloud/+merge/475520
Your team Canonical's Ubuntu QA is requested to review the proposed merge of 
~hyask/autopkgtest-cloud:skia/publish_db_copy_url_last_checked into 
autopkgtest-cloud:master.


-- 
Mailing list: https://launchpad.net/~canonical-ubuntu-qa
Post to : canonical-ubuntu-qa@lists.launchpad.net
Unsubscribe : https://launchpad.net/~canonical-ubuntu-qa
More help   : https://help.launchpad.net/ListHelp


[Canonical-ubuntu-qa] [Merge] ~hyask/autopkgtest-cloud:skia/publish_db_copy_url_last_checked into autopkgtest-cloud:master

2024-10-21 Thread Skia
Skia has proposed merging 
~hyask/autopkgtest-cloud:skia/publish_db_copy_url_last_checked into 
autopkgtest-cloud:master.

Requested reviews:
  Canonical's Ubuntu QA (canonical-ubuntu-qa)

For more details, see:
https://code.launchpad.net/~hyask/autopkgtest-cloud/+git/autopkgtest-cloud/+merge/475520

Quick win bringing nice performance improvement to `publish-db`.
-- 
Your team Canonical's Ubuntu QA is requested to review the proposed merge of 
~hyask/autopkgtest-cloud:skia/publish_db_copy_url_last_checked into 
autopkgtest-cloud:master.
diff --git a/charms/focal/autopkgtest-web/webcontrol/publish-db b/charms/focal/autopkgtest-web/webcontrol/publish-db
index 4149637..78b4b92 100755
--- a/charms/focal/autopkgtest-web/webcontrol/publish-db
+++ b/charms/focal/autopkgtest-web/webcontrol/publish-db
@@ -78,7 +78,7 @@ def init_db(path, path_current, path_rw):
 except sqlite3.OperationalError as e2:
 if "duplicate column name" not in str(e2):
 raise
-logging.debug("table already exists")
+logging.debug("table 'current_version' already exists")
 db.execute(
 "CREATE INDEX IF NOT EXISTS current_version_pocket_ix "
 "ON current_version(pocket, component)"
@@ -110,21 +110,22 @@ def init_db(path, path_current, path_rw):
 "  PRIMARY KEY(url))"
 )
 logging.debug("database table url_last_checked created")
-try:
-if current_version_copied:
-db.execute(
-"INSERT INTO url_last_checked (url, timestamp) "
-"SELECT url, timestamp FROM current.url_last_checked"
-)
-logging.debug("Old url_last_checked copied over")
-except sqlite3.OperationalError as e:
-if "no such table" not in str(e):
-raise
-logging.debug("no url_last_checked yet, first run probably")
 except sqlite3.OperationalError as e:
 if "already exists" not in str(e):
 raise
-logging.debug("table already exists")
+logging.debug("table 'url_last_checked' already exists")
+
+try:
+if current_version_copied:
+db.execute(
+"INSERT OR REPLACE INTO url_last_checked (url, timestamp) "
+"SELECT url, timestamp FROM current.url_last_checked"
+)
+logging.debug("Old url_last_checked copied over")
+except sqlite3.OperationalError as e:
+if "no such table" not in str(e):
+raise
+logging.debug("no url_last_checked yet, first run probably")
 
 db.commit()
 db.execute("DETACH DATABASE current")
-- 
Mailing list: https://launchpad.net/~canonical-ubuntu-qa
Post to : canonical-ubuntu-qa@lists.launchpad.net
Unsubscribe : https://launchpad.net/~canonical-ubuntu-qa
More help   : https://help.launchpad.net/ListHelp


[Canonical-ubuntu-qa] [Bug 2009714] Re: 7 failures reported in net:fcnal-test.sh on Kinetic

2024-10-21 Thread Mehmet Basaran
Seen this with noble:linux-lowlatency 6.8.0-48.48.3 on
openstack:cpu2-ram4-disk20 (arm64). SRU cycle: 2024.09.30.

20:00:21 DEBUG| [stdout] # 
#
20:00:21 DEBUG| [stdout] # Ping LLA with multiple interfaces
20:00:21 DEBUG| [stdout] # 
20:00:26 DEBUG| [stdout] # TEST: Pre cycle, ping out ns-B - multicast IP
 [FAIL]
20:00:27 DEBUG| [stdout] # TEST: Pre cycle, ping out ns-C - multicast IP
 [ OK ]
20:00:30 DEBUG| [stdout] # TEST: Post cycle nsa-RcZfpc eth1, ping out ns-B - 
multicast IP[ OK ]
20:00:31 DEBUG| [stdout] # TEST: Post cycle nsa-RcZfpc eth1, ping out ns-C - 
multicast IP[ OK ]
20:00:34 DEBUG| [stdout] # TEST: Post cycle nsa-RcZfpc eth2, ping out ns-B - 
multicast IP[ OK ]
20:00:35 DEBUG| [stdout] # TEST: Post cycle nsa-RcZfpc eth2, ping out ns-C - 
multicast IP[ OK ]
20:00:36 DEBUG| [stdout] #

-- 
You received this bug notification because you are a member of Canonical
Platform QA Team, which is subscribed to ubuntu-kernel-tests.
https://bugs.launchpad.net/bugs/2009714

Title:
  7 failures reported in net:fcnal-test.sh on Kinetic

Status in ubuntu-kernel-tests:
  New

Bug description:
  This is not a regression, the test was skipped in the old days due to
  the lack of `nettest` command (bug 2006391)

  Unlike bug 2006716 ("5 failures reported in net:fcnal-test.sh on
  Jammy") there are 7 common failures reported on Kinetic

   #
   # IPv4 ping
   #
   # With VRF
   # SYSCTL: net.ipv4.raw_l3mdev_accept=1
   # TEST: ping local, VRF bind - VRF IP  [FAIL]
   # TEST: ping local, device bind - ns-A IP  [FAIL]

   # SYSCTL: net.ipv4.ping_group_range=0 2147483647
   # SYSCTL: net.ipv4.raw_l3mdev_accept=1
   # TEST: ping local, VRF bind - VRF IP  [FAIL]
   # TEST: ping local, device bind - ns-A IP  [FAIL]

   #
   # IPv6 ping
   #
   # With VRF
   # SYSCTL: net.ipv4.ping_group_range=0 2147483647
   # SYSCTL: net.ipv4.raw_l3mdev_accept=1
   # TEST: ping out, vrf device+address bind - ns-B IPv6 LLA  [FAIL]

   #
   # IPv6/UDP
   #
   # No VRF
   # udp_l3mdev_accept disabled
   # SYSCTL: net.ipv4.udp_l3mdev_accept=0
   # TEST: Global server, device client via IP_UNICAST_IF, local connection, 
with connect() - ns-A loopback IPv6  [FAIL]

   # udp_l3mdev_accept enabled
   # SYSCTL: net.ipv4.udp_l3mdev_accept=1
   # TEST: Global server, device client via IP_UNICAST_IF, local connection, 
with connect() - ns-A loopback IPv6  [FAIL]

  Sometimes it will fail with 8 failures, but the 8th failure looks
  flaky and can be failing on different test cases:

  With K-aws c6g.8xlarge (arm64) and r5.large (amd64) the 8th failure is:
   #
   # Use cases
   #
   # Ping LLA with multiple interfaces
   # TEST: Pre cycle, ping out ns-B - multicast IP[FAIL]

  With K-azure Standard_D2pds_v5 (arm64) the 8th failure is:
   #
   # Use cases
   #
   # TEST: Post cycle ns-A eth1, ping out ns-B - multicast IP [FAIL]

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-kernel-tests/+bug/2009714/+subscriptions


-- 
Mailing list: https://launchpad.net/~canonical-ubuntu-qa
Post to : canonical-ubuntu-qa@lists.launchpad.net
Unsubscribe : https://launchpad.net/~canonical-ubuntu-qa
More help   : https://help.launchpad.net/ListHelp


[Canonical-ubuntu-qa] [Bug 1931390] Re: cpuhotplug03 in cpuhotplug from ubuntu_ltp failed after successful CPU1 offline

2024-10-21 Thread Mehmet Basaran
I have seen this issue for noble:linux-lowlatency 6.8.0-48.48.3 on
openstack:cpu2-ram4-disk20 (arm64). (SRU Cycle: 2024.09.30)

I think this test is flaky. To check if "cpuhotplug_do_spin_loop" it
uses "ps psr" command where "r" option lists the currently running
processes only. Even if a "cpuhotplug_do_spin_loop" process did migrate
to target CPU, this does not mean that it is currently running there (as
it can be preempted by availability of other tasks). And for
architectures with lower number of cpus (in my case 2), test is likely
to fail more.

Moreover, test doesn't seem to adjust "spin" tasks' priority (I am not
sure about this but it doesn't seem to do anything in this regard prior
to running the test either).

Test needs to either;
 - run in isolated environment where OS threads run on a dedicated core.
 - sleep some time (as done currently) and check if the spin process is 
migrated by the spin process itself.

Here is the log:
FO: Test start time: Fri Oct 18 20:06:59 UTC 2024
ND:/opt/ltp/bin/ltp-pan -q  -e -S   -a 166236 -n 166236  -f 
/tmp/ltp-9srhcvY2Gx/alltests -l /dev/null  -C /dev/null -T /dev/null
OG File: /dev/null
ED COMMAND File: /dev/null
NF COMMAND File: /dev/null
ng tests...
me:   cpuhotplug03
te:   Fri Oct 18 20:06:59 UTC 2024
sc:   Do tasks get scheduled to a newly on-lined CPU?
08  CPU is 1
ER PID %CPU %MEMVSZ   RSS TTY  STAT START   TIME COMMAND
ot  166323 12.8  0.0   2380  1408 ?R20:06   0:00 /bin/sh 
/opt/ltp/testcases/bin/cpuhotplug_do_spin_loop
ot  166324 12.8  0.0   2380  1408 ?R20:06   0:00 /bin/sh 
/opt/ltp/testcases/bin/cpuhotplug_do_spin_loop
ot  166325 12.8  0.0   2380  1408 ?R20:06   0:00 /bin/sh 
/opt/ltp/testcases/bin/cpuhotplug_do_spin_loop
ot  166326 12.8  0.0   2380  1408 ?R20:06   0:00 /bin/sh 
/opt/ltp/testcases/bin/cpuhotplug_do_spin_loop
ot  166331  0.0  0.0   6680  1920 ?S20:06   0:00 grep 
cpuhotplug_do_spin_loop
03 1 TINFO: Onlining CPU 1
 0 /bin/sh /opt/ltp/testcases/bin/cpuhotplug_do_spin_loop
 0 /bin/sh /opt/ltp/testcases/bin/cpuhotplug_do_spin_loop
 0 /bin/sh /opt/ltp/testcases/bin/cpuhotplug_do_spin_loop
 0 /bin/sh /opt/ltp/testcases/bin/cpuhotplug_do_spin_loop
03 1 TFAIL: No cpuhotplug_do_spin_loop processes found on CPU1
FO: ltp-pan reported some tests FAIL
TP Version: 20230929-874-gba610da01
FO: Test end time: Fri Oct 18 20:07:01 UTC 2024

-- 
You received this bug notification because you are a member of Canonical
Platform QA Team, which is subscribed to ubuntu-kernel-tests.
https://bugs.launchpad.net/bugs/1931390

Title:
  cpuhotplug03 in cpuhotplug from ubuntu_ltp failed after successful
  CPU1 offline

Status in ubuntu-kernel-tests:
  New
Status in linux package in Ubuntu:
  Incomplete

Bug description:
  In cpuhotplug (ubuntu_ltp) tests, offline and online of CPU1 succeeds,
  but later the test fails with:

  utils:0153| [stdout] Name: cpuhotplug03
  utils:0153| [stdout] Date: Wed Jun 9 03:52:16 UTC 2021
  utils:0153| [stdout] Desc: Do tasks get scheduled to a newly on-lined CPU?
  utils:0153| [stdout]
  utils:0153| [stdout] CPU is 1
  utils:0153| [stdout] USER PID %CPU %MEM VSZ RSS TTY STAT START TIME COMMAND
  utils:0153| [stdout] root 858611 0.0 0.0 2616 540 ? R 03:52 0:00 /bin/sh 
/opt/ltp/testcases/bin/cpuhotplug_do_spin_loop
  ...
  ...
  ...
  utils:0153| [stdout] 49 /bin/sh /opt/ltp/testcases/bin/cpuhotplug_do_spin_loop
  utils:0153| [stdout] 82 /bin/sh /opt/ltp/testcases/bin/cpuhotplug_do_spin_loop
  utils:0153| [stdout] 18 /bin/sh /opt/ltp/testcases/bin/cpuhotplug_do_spin_loop
  utils:0153| [stdout] cpuhotplug03 1 TFAIL: No cpuhotplug_do_spin_loop 
processes found on CPU1
  utils:0153| [stdout] tag=cpuhotplug03 stime=1623210736 dur=3 exit=exited 
stat=1 core=no cu=22260 cs=78

  IMPORTANT: The CPU1 got properly offlined and onlined. For failures of
  CPU1 offline, please see lp:1836167 (reported previously as part of
  lp:1836167).

  Po-Hsu Lin (cypressyew) wrote on 2020-09-25:
  On F-oem-5.6 with node glameow:
  cpuhotplug03 1 TFAIL: No cpuhotplug_do_spin_loop processes found on CPU1

  Kelsey Skunberg (kelsey-skunberg) wrote on 2020-10-06:
  spotted on Focal aws : 5.4.0-1026.26 : amd64
  cpuhotplug03 1 TFAIL: No cpuhotplug_do_spin_loop processes found on CPU1

  Po-Hsu Lin (cypressyew) wrote on 2021-06-09:
  Found on 5.11.0-1005.5 - intel
  Node spitfire, passed on node bavor.
  cpuhotplug03 1 TFAIL: No cpuhotplug_do_spin_loop processes found on CPU1

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-kernel-tests/+bug/1931390/+subscriptions


-- 
Mailing list: https://launchpad.net/~canonical-ubuntu-qa
Post to : canonical-ubuntu-qa@lists.launchpad.net
Unsubscribe : https://launchpad.net/~canonical-ubuntu-qa
More help   : https://help.launchpad.net/ListHelp


[Canonical-ubuntu-qa] [Bug 1943841] Re: svm from ubuntu_kvm_unit_tests interrupt with "Unhandled exception 13 #GP at ip 00000000004027e3" on F-intel-5.13

2024-10-21 Thread Mehmet Basaran
Similar error happens with noble:linux-lowlatency 6.8.0-48.48.3 on
riccioli (SRU cycle: 2024.09.30)


20:28:53 INFO | STARTubuntu_kvm_unit_tests.svm
ubuntu_kvm_unit_tests.svmtimeout=1800timestamp=172928
localtime=Oct 18 20:28:53
20:28:53 DEBUG| Persistent state client._record_indent now set to 2
20:28:53 DEBUG| Persistent state client.unexpected_reboot now set to 
('ubuntu_kvm_unit_tests.svm', 'ubuntu_kvm_unit_tests.svm')
20:28:53 DEBUG| Waiting for pid 18157 for 1800 seconds
20:28:53 WARNI| System python is too old, crash handling disabled
20:28:53 DEBUG| Running 'kvm-ok'
20:28:53 DEBUG| [stdout] INFO: /dev/kvm exists
20:28:53 DEBUG| [stdout] KVM acceleration can be used
20:28:53 DEBUG| Running 
'/home/ubuntu/autotest/client/tmp/ubuntu_kvm_unit_tests/src/kvm-unit-tests/tests/svm'
20:28:53 DEBUG| [stdout] BUILD_HEAD=b04954c9
20:28:53 DEBUG| [stdout] timeout -k 1s --foreground 90s 
/usr/bin/qemu-system-x86_64 --no-reboot -nodefaults -device pc-testdev -device 
isa-debug-exit,iobase=0xf4,iosize=0x4 -vnc none -serial stdio -device 
pci-testdev -machine accel=kvm -kernel /tmp/tmp.jgXFu3YdzN -smp 2 -cpu max,+svm 
-m 4g -append -pause_filter_test # -initrd /tmp/tmp.jZ1NEkg5BI
20:28:54 DEBUG| [stdout] enabling apic
20:28:54 DEBUG| [stdout] smp: waiting for 1 APs
20:28:54 DEBUG| [stdout] enabling apic
20:28:54 DEBUG| [stdout] setup: CPU 1 online
20:28:54 DEBUG| [stdout] paging enabled
20:28:54 DEBUG| [stdout] cr0 = 80010011
20:28:54 DEBUG| [stdout] cr3 = 10bf000
20:28:54 DEBUG| [stdout] cr4 = 20
20:28:54 DEBUG| [stdout] NPT detected - running all tests with NPT enabled
20:28:54 DEBUG| [stdout] PASS: null
20:28:54 DEBUG| [stdout] PASS: vmrun
20:28:54 DEBUG| [stdout] PASS: ioio
20:28:54 DEBUG| [stdout] PASS: vmrun intercept check
20:28:54 DEBUG| [stdout] PASS: rsm
20:28:54 DEBUG| [stdout] PASS: cr3 read intercept
20:28:54 DEBUG| [stdout] PASS: cr3 read nointercept
20:28:54 DEBUG| [stdout] PASS: cr3 read intercept emulate
20:28:54 DEBUG| [stdout] PASS: dr intercept check
20:28:54 DEBUG| [stdout] PASS: next_rip
20:28:54 DEBUG| [stdout] PASS: msr intercept check
20:28:54 DEBUG| [stdout] PASS: mode_switch
20:28:54 DEBUG| [stdout] PASS: asid_zero
20:28:54 DEBUG| [stdout] PASS: sel_cr0_bug
20:28:54 DEBUG| [stdout] PASS: tsc_adjust
20:28:58 DEBUG| [stdout] Latency VMRUN : max: 4793660 min: 6460 avg: 8899
20:28:58 DEBUG| [stdout] Latency VMEXIT: max: 4766600 min: 3580 avg: 3774
20:28:58 DEBUG| [stdout] PASS: latency_run_exit
20:29:03 DEBUG| [stdout] Latency VMRUN : max: 2145480 min: 8480 avg: 8898
20:29:03 DEBUG| [stdout] Latency VMEXIT: max: 2133480 min: 3580 avg: 3788
20:29:03 DEBUG| [stdout] PASS: latency_run_exit_clean
20:29:03 DEBUG| [stdout] Latency VMLOAD: max: 2443180 min: 260 avg: 275
20:29:03 DEBUG| [stdout] Latency VMSAVE: max: 1971420 min: 240 avg: 263
20:29:03 DEBUG| [stdout] Latency STGI:   max: 9740 min: 40 avg: 49
20:29:03 DEBUG| [stdout] Latency CLGI:   max: 46560 min: 40 avg: 53
20:29:03 DEBUG| [stdout] PASS: latency_svm_insn
20:29:03 DEBUG| [stdout] PASS: exception with vector 2 not injected
20:29:03 DEBUG| [stdout] PASS: divide overflow exception injected
20:29:03 DEBUG| [stdout] PASS: eventinj.VALID cleared
20:29:03 DEBUG| [stdout] PASS: exc_inject
20:29:03 DEBUG| [stdout] PASS: pending_event
20:29:03 DEBUG| [stdout] PASS: pending_event_cli
20:29:03 DEBUG| [stdout] PASS: direct interrupt while running guest
20:29:03 DEBUG| [stdout] PASS: intercepted interrupt while running guest
20:29:03 DEBUG| [stdout] PASS: direct interrupt + hlt
20:29:03 DEBUG| [stdout] PASS: intercepted interrupt + hlt
20:29:03 DEBUG| [stdout] PASS: interrupt
20:29:03 DEBUG| [stdout] PASS: direct NMI while running guest
20:29:03 DEBUG| [stdout] PASS: NMI intercept while running guest
20:29:03 DEBUG| [stdout] PASS: nmi
20:29:03 DEBUG| [stdout] PASS: direct NMI + hlt
20:29:03 DEBUG| [stdout] PASS: NMI intercept while running guest
20:29:03 DEBUG| [stdout] PASS: intercepted NMI + hlt
20:29:03 DEBUG| [stdout] PASS: nmi_hlt
20:29:03 DEBUG| [stdout] PASS: virq_inject
20:29:05 DEBUG| [stdout] PASS: No RIP corruption detected after 1 timer 
interrupts
20:29:05 DEBUG| [stdout] PASS: reg_corruption
20:29:05 DEBUG| [stdout] enabling apic
20:29:05 DEBUG| [stdout] setup: CPU 1 online
20:29:05 DEBUG| [stdout] PASS: svm_init_startup_test
20:29:05 DEBUG| [stdout] PASS: host_rflags
20:29:05 DEBUG| [stdout] PASS: STGI set VGIF bit.
20:29:05 DEBUG| [stdout] PASS: CLGI cleared VGIF bit.
20:29:05 DEBUG| [stdout] PASS: vgif
20:29:05 DEBUG| [stdout] PASS: CPUID.01H:ECX.XSAVE set before VMRUN
20:29:05 DEBUG| [stdout] PASS: svm_cr4_osxsave_test_guest finished with VMMCALL
20:29:05 DEBUG| [stdout] PASS: CPUID.01H:ECX.XSAVE set after VMRUN
20:29:05 DEBUG| [stdout] PASS: EFER.SVME: 1500
20:29:05 DEBUG| [stdout] PASS: EFER.SVME: 500
20:29:05 DEBUG| [stdout] PASS: Test EFER 9:8: 1700
20:29:05 DEBUG| [stdout] PASS: Test EFER 63:16: 11500
20:29:05 DEBUG| [stdout] PASS: Test EFER 63:16: 101500
20:2

[Canonical-ubuntu-qa] [Bug 1955529] Re: cpuset_memory_pressure in ubuntu_ltp_controllers failed on B-oracle BM.Standard2.52 (root group's memory_pressure didn't have memory pressure rate.)

2024-10-21 Thread Mehmet Basaran
Seen this on n/lowlatency 6.8.0-48.48.3 on fili (SRU cycle: 2024.09.30)

-- 
You received this bug notification because you are a member of Canonical
Platform QA Team, which is subscribed to ubuntu-kernel-tests.
https://bugs.launchpad.net/bugs/1955529

Title:
  cpuset_memory_pressure in ubuntu_ltp_controllers failed on B-oracle
  BM.Standard2.52 (root group's memory_pressure didn't have memory
  pressure rate.)

Status in ubuntu-kernel-tests:
  New

Bug description:
  Issue found on Oracle instance BM.Standard2.52 only with kernel
  4.15.0-1085.93

  This failure can be found in 4.15.0-1084.92 / 4.15.0-1083.91 as well,
  thus it's not a regression.

  Test failed with:
  cpuset_memory_pressure 11 TFAIL: root group's memory_pressure didn't have 
memory pressure rate.

  Test log:
  Checking for required user/group ids

  'nobody' user id and group found.
  'bin' user id and group found.
  'daemon' user id and group found.
  Users group found.
  Sys group found.
  Required users/groups exist.
  no big block device was specified on commandline.
  Tests which require a big block device are disabled.
  You can specify it with option -z
  INFO: Test start time: Tue Dec 21 15:07:18 UTC 2021
  COMMAND:/opt/ltp/bin/ltp-pan -q  -e -S   -a 57071 -n 57071  -f 
/tmp/ltp-9znr7PBmp3/alltests -l /dev/null  -C /dev/null -T /dev/null
  LOG File: /dev/null
  FAILED COMMAND File: /dev/null
  TCONF COMMAND File: /dev/null
  Running tests...
  cpuset_memory_pressure 1 TINFO: CPUs are numbered continuously starting at 0 
(0-103)
  cpuset_memory_pressure 1 TINFO: Nodes are numbered continuously starting at 0 
(0-1)
  cpuset_memory_pressure 1 TPASS: Cpuset memory pressure test succeeded.
  cpuset_memory_pressure 3 TPASS: Cpuset memory pressure test succeeded.
  cpuset_memory_pressure 5 TPASS: Cpuset memory pressure test succeeded.
  Killed
  cpuset_memory_pressure 7 TPASS: Cpuset memory pressure test succeeded.
  cpuset_memory_pressure 9 TFAIL: root group's memory_pressure didn't have 
memory pressure rate.
  Killed
  cpuset_memory_pressure 11 TFAIL: root group's memory_pressure didn't have 
memory pressure rate.
  INFO: ltp-pan reported some tests FAIL
  LTP Version: 20210927
  INFO: Test end time: Tue Dec 21 15:48:45 UTC 2021

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-kernel-tests/+bug/1955529/+subscriptions


-- 
Mailing list: https://launchpad.net/~canonical-ubuntu-qa
Post to : canonical-ubuntu-qa@lists.launchpad.net
Unsubscribe : https://launchpad.net/~canonical-ubuntu-qa
More help   : https://help.launchpad.net/ListHelp


[Canonical-ubuntu-qa] [Bug 2083687] Re: cachestat03 from ubuntu_ltp_syscalls failed on B-KVM

2024-10-21 Thread Po-Hsu Lin
** Tags added: 5.15 5.4 focal jammy

-- 
You received this bug notification because you are a member of Canonical
Platform QA Team, which is subscribed to ubuntu-kernel-tests.
https://bugs.launchpad.net/bugs/2083687

Title:
  cachestat03 from ubuntu_ltp_syscalls failed on B-KVM

Status in ubuntu-kernel-tests:
  New

Bug description:
  This is a new test case added on Jul. 23
  
https://github.com/linux-test-project/ltp/commit/1564e8c4700c0bd4b17282f5560d6f1774798005

  And we got our fork updated recently in order to cover Oracular.
  So this is not a regression.

  B-4.15, B-hwe-5.4 are good. It's just failing for B-KVM for now.

  Test failed with:
    TBROK: Failed to open FILE '/proc/sys/vm/compact_memory' for writing: 
ENOENT (2)

  Test log:
  COMMAND:/opt/ltp/bin/ltp-pan   -e -S   -a 8849 -n 8849 -p -f 
/tmp/ltp-8M7ll8HANh/alltests -l 
/opt/ltp/results/LTP_RUN_ON-2024_10_04-11h_49m_21s.log  -C 
/opt/ltp/output/LTP_RUN_ON-2024_10_04-11h_49m_21s.failed -T 
/opt/ltp/output/LTP_RUN_ON-2024_10_04-11h_49m_21s.tconf
  LOG File: /opt/ltp/results/LTP_RUN_ON-2024_10_04-11h_49m_21s.log
  FAILED COMMAND File: /opt/ltp/output/LTP_RUN_ON-2024_10_04-11h_49m_21s.failed
  TCONF COMMAND File: /opt/ltp/output/LTP_RUN_ON-2024_10_04-11h_49m_21s.tconf
  Running tests...
  <<>>
  tag=cachestat03 stime=1728042561
  cmdline="cachestat03"
  contacts=""
  analysis=exit
  <<>>
  incrementing stop
  tst_hugepage.c:50: TBROK: Failed to open FILE '/proc/sys/vm/compact_memory' 
for writing: ENOENT (2)
  <<>>
  initiation_status="ok"
  duration=1 termination_type=exited termination_id=2 corefile=no
  cutime=0 cstime=37
  <<>>
  INFO: ltp-pan reported some tests FAIL

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-kernel-tests/+bug/2083687/+subscriptions


-- 
Mailing list: https://launchpad.net/~canonical-ubuntu-qa
Post to : canonical-ubuntu-qa@lists.launchpad.net
Unsubscribe : https://launchpad.net/~canonical-ubuntu-qa
More help   : https://help.launchpad.net/ListHelp


[Canonical-ubuntu-qa] [Bug 2083687] Re: cachestat03 from ubuntu_ltp_syscalls failed on B-KVM

2024-10-21 Thread Po-Hsu Lin
** Tags added: sru-20240930

-- 
You received this bug notification because you are a member of Canonical
Platform QA Team, which is subscribed to ubuntu-kernel-tests.
https://bugs.launchpad.net/bugs/2083687

Title:
  cachestat03 from ubuntu_ltp_syscalls failed on B-KVM

Status in ubuntu-kernel-tests:
  New

Bug description:
  This is a new test case added on Jul. 23
  
https://github.com/linux-test-project/ltp/commit/1564e8c4700c0bd4b17282f5560d6f1774798005

  And we got our fork updated recently in order to cover Oracular.
  So this is not a regression.

  B-4.15, B-hwe-5.4 are good. It's just failing for B-KVM for now.

  Test failed with:
    TBROK: Failed to open FILE '/proc/sys/vm/compact_memory' for writing: 
ENOENT (2)

  Test log:
  COMMAND:/opt/ltp/bin/ltp-pan   -e -S   -a 8849 -n 8849 -p -f 
/tmp/ltp-8M7ll8HANh/alltests -l 
/opt/ltp/results/LTP_RUN_ON-2024_10_04-11h_49m_21s.log  -C 
/opt/ltp/output/LTP_RUN_ON-2024_10_04-11h_49m_21s.failed -T 
/opt/ltp/output/LTP_RUN_ON-2024_10_04-11h_49m_21s.tconf
  LOG File: /opt/ltp/results/LTP_RUN_ON-2024_10_04-11h_49m_21s.log
  FAILED COMMAND File: /opt/ltp/output/LTP_RUN_ON-2024_10_04-11h_49m_21s.failed
  TCONF COMMAND File: /opt/ltp/output/LTP_RUN_ON-2024_10_04-11h_49m_21s.tconf
  Running tests...
  <<>>
  tag=cachestat03 stime=1728042561
  cmdline="cachestat03"
  contacts=""
  analysis=exit
  <<>>
  incrementing stop
  tst_hugepage.c:50: TBROK: Failed to open FILE '/proc/sys/vm/compact_memory' 
for writing: ENOENT (2)
  <<>>
  initiation_status="ok"
  duration=1 termination_type=exited termination_id=2 corefile=no
  cutime=0 cstime=37
  <<>>
  INFO: ltp-pan reported some tests FAIL

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-kernel-tests/+bug/2083687/+subscriptions


-- 
Mailing list: https://launchpad.net/~canonical-ubuntu-qa
Post to : canonical-ubuntu-qa@lists.launchpad.net
Unsubscribe : https://launchpad.net/~canonical-ubuntu-qa
More help   : https://help.launchpad.net/ListHelp


[Canonical-ubuntu-qa] [Bug 2056461] Re: autopkgtest-virt-qemu on noble images sometimes hangs doing copydown

2024-10-21 Thread Ural Tunaboyu
Performing the verification on Noble.

First, we check to see if we can reproduce the problem.

$ apt policy autopkgtest
autopkgtest:
  Installed: 5.34ubuntu2
  Candidate: 5.34ubuntu2
  Version table:
 5.38ubuntu1~24.04.1 100
100 http://archive.ubuntu.com/ubuntu noble-proposed/main amd64 Packages
 *** 5.34ubuntu2 500
500 http://archive.ubuntu.com/ubuntu noble/main amd64 Packages
100 /var/lib/dpkg/status

After creating a Noble image we run the reproducer:

$ autopkgtest -U -ddd hello*.dsc -- qemu ../autopkgtest-noble-amd64.img
autopkgtest: DBG: autopkgtest options: Namespace(override_control=None, 
only_tests=[], skip_tests=None, built_binaries=True, architecture=None, 
packages=['hello_2.10-3build2.dsc'], output_dir=None)
autopkgtest: DBG: virt-runner arguments: ['qemu', 
'../autopkgtest-noble-amd64.img']
autopkgtest: DBG: actions: [('source', 'hello_2.10-3build2.dsc', True)]
autopkgtest: DBG: build binaries: True
autopkgtest: DBG: testbed init
autopkgtest [12:09:07]: starting date and time: 2024-10-21 12:09:07-0700
autopkgtest [12:09:07]: version 5.34ubuntu2
autopkgtest [12:09:07]: host clean-noble-amd64; command line: 
/usr/bin/autopkgtest -U -ddd hello_2.10-3build2.dsc -- qemu 
../autopkgtest-noble-amd64.img
(truncated for brevity)
autopkgtest: DBG: sending command to testbed: copydown hello_2.10.orig.tar.gz 
/tmp/autopkgtest.gHgSJl/hello_2.10.orig.tar.gz
autopkgtest: DBG: got reply from testbed: timeout
autopkgtest: DBG: sending command to testbed: auxverb_debug_fail
autopkgtest: DBG: got reply from testbed: ok
autopkgtest: DBG: TestbedFailure sent `auxverb_debug_fail', got `timeout', 
expected `ok...'
autopkgtest: DBG: testbed stop
autopkgtest: DBG: testbed close, scratch=/tmp/autopkgtest.gHgSJl
autopkgtest: DBG: sending command to testbed: close
qemu-system-x86_64: terminating on signal 15 from pid 7523 (/usr/bin/python3)
autopkgtest: DBG: got reply from testbed: ok
autopkgtest: DBG: sending command to testbed: quit
autopkgtest [12:14:36]: ERROR: testbed failure: sent `auxverb_debug_fail', got 
`timeout', expected `ok...'

The testbed hung until timeout during copydown, so the bug is replicated.
Now we verify the fix.

$ apt policy autopkgtest
autopkgtest:
  Installed: 5.38ubuntu1~24.04.1
  Candidate: 5.38ubuntu1~24.04.1
  Version table:
 *** 5.38ubuntu1~24.04.1 100
100 http://archive.ubuntu.com/ubuntu noble-proposed/main amd64 Packages
100 /var/lib/dpkg/status
 5.34ubuntu2 500
500 http://archive.ubuntu.com/ubuntu noble/main amd64 Packages

We create a fresh Noble image and rerun the reproducer:

$ autopkgtest -U hello*.dsc -- qemu ../autopkgtest-noble-amd64.img
autopkgtest [12:35:03]: starting date and time: 2024-10-21 12:35:03-0700
autopkgtest [12:35:03]: version 5.38ubuntu1~24.04.1
autopkgtest [12:35:03]: host clean-noble-amd64; command line: 
/usr/bin/autopkgtest -U hello_2.10-3build2.dsc -- qemu 
../autopkgtest-noble-amd64.img
autopkgtest [12:35:19]: testbed dpkg architecture: amd64
autopkgtest [12:35:20]: testbed apt version: 2.7.14build2
autopkgtest [12:35:20]:  test bed setup
(truncated for brevity)
autopkgtest [12:37:04]: test upstream-tests: [---
Testing greeting-1 ...
Testing hello-1 ...
Testing last-1 ...
Testing traditional-1 ...
autopkgtest [12:37:05]: test upstream-tests: ---]
autopkgtest [12:37:06]: test upstream-tests:  - - - - - - - - - - results - - - 
- - - - - - -
upstream-tests   PASS
autopkgtest [12:37:06]:  summary
command1 PASS
upstream-tests   PASS

So the test was completed without a timeout. 
This concludes the verification for Noble.

-- 
You received this bug notification because you are a member of
Canonical's Ubuntu QA, which is subscribed to autopkgtest in Ubuntu.
Matching subscriptions: ubuntu-qa-bugs
https://bugs.launchpad.net/bugs/2056461

Title:
  autopkgtest-virt-qemu on noble images sometimes hangs doing copydown

Status in Linux:
  Confirmed
Status in autopkgtest package in Ubuntu:
  Fix Released
Status in linux package in Ubuntu:
  In Progress
Status in autopkgtest source package in Jammy:
  In Progress
Status in linux source package in Jammy:
  New
Status in autopkgtest source package in Noble:
  In Progress
Status in linux source package in Noble:
  New
Status in autopkgtest package in Debian:
  Fix Released

Bug description:
  [Impact]

  It seems that kernel 6.8 introduced a regression in the 9pfs related
  to caching and netfslib, that can cause some user-space apps to read
  content from files that is not up-to-date (when they are used in a
  producer/consumer fashion).

  It seems that the offending commit is this one:

   80105ed2fd27 ("9p: Use netfslib read/write_iter")

  Reverting the commit seems to fix the problem. However the actual bug
  might be in netfslib or how netfslib is used in the 9p context.

  The regression has been reported upstream and we are still
  investigating (https://lore.

[Canonical-ubuntu-qa] [Bug 2056461] Re: autopkgtest-virt-qemu on noble images sometimes hangs doing copydown

2024-10-21 Thread Ural Tunaboyu
Performing the verification on Jammy.

First, we check if we can reproduce the problem.

$ apt policy autopkgtest
autopkgtest:
  Installed: 5.32ubuntu3~22.04.1
  Candidate: 5.38ubuntu1~22.04.1
  Version table:
 5.38ubuntu1~22.04.1 500
500 http://archive.ubuntu.com/ubuntu jammy-proposed/main amd64 Packages
500 http://archive.ubuntu.com/ubuntu jammy-proposed/main i386 Packages
 *** 5.32ubuntu3~22.04.1 500
500 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 Packages
500 http://archive.ubuntu.com/ubuntu jammy-updates/main i386 Packages
100 /var/lib/dpkg/status
 5.20 500
500 http://archive.ubuntu.com/ubuntu jammy/main amd64 Packages
500 http://archive.ubuntu.com/ubuntu jammy/main i386 Packages

After creating a Noble image we run the reproducer:
$ autopkgtest -U -ddd hello*.dsc -- qemu ../autopkgtest-noble-amd64.img
autopkgtest: DBG: autopkgtest options: Namespace(override_control=None, 
only_tests=[], skip_tests=None, built_binaries=True, architecture=None, 
packages=['hello_2.10-3build2.dsc'], output_dir=None)
autopkgtest: DBG: virt-runner arguments: ['qemu', 
'../autopkgtest-noble-amd64.img']
autopkgtest: DBG: actions: [('source', 'hello_2.10-3build2.dsc', True)]
autopkgtest: DBG: build binaries: True
autopkgtest: DBG: testbed init
autopkgtest [20:52:54]: starting date and time: 2024-10-21 20:52:54+0100
autopkgtest [20:52:54]: version 5.32ubuntu3~22.04.1
autopkgtest [20:52:54]: host clean-jammy-amd64; command line: 
/usr/bin/autopkgtest -U -ddd hello_2.10-3build2.dsc -- qemu 
../autopkgtest-noble-amd64.img
(truncated for brevity)
autopkgtest: DBG: sending command to testbed: copydown 
/tmp/autopkgtest.output.cbb3dd47/tests-tree/ 
/tmp/autopkgtest.ShBCzZ/build.DY9/src/
autopkgtest: DBG: got reply from testbed: timeout
autopkgtest: DBG: sending command to testbed: auxverb_debug_fail
autopkgtest: DBG: got reply from testbed: ok
autopkgtest: DBG: TestbedFailure sent `auxverb_debug_fail', got `timeout', 
expected `ok...'
autopkgtest: DBG: testbed stop
autopkgtest: DBG: testbed close, scratch=/tmp/autopkgtest.ShBCzZ
autopkgtest: DBG: sending command to testbed: close
qemu-system-x86_64: terminating on signal 15 from pid 4744 (/usr/bin/python3)
autopkgtest: DBG: got reply from testbed: ok
autopkgtest: DBG: sending command to testbed: quit
autopkgtest [21:02:35]: ERROR: testbed failure: sent `auxverb_debug_fail', got 
`timeout', expected `ok...'

The testbed hung until timeout during copydown, so the bug is verified. 
Now we verify the fix.

$ apt policy autopkgtest
autopkgtest:
  Installed: 5.38ubuntu1~22.04.1
  Candidate: 5.38ubuntu1~22.04.1
  Version table:
 *** 5.38ubuntu1~22.04.1 500
500 http://archive.ubuntu.com/ubuntu jammy-proposed/main amd64 Packages
500 http://archive.ubuntu.com/ubuntu jammy-proposed/main i386 Packages
100 /var/lib/dpkg/status
 5.32ubuntu3~22.04.1 500
500 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 Packages
500 http://archive.ubuntu.com/ubuntu jammy-updates/main i386 Packages
 5.20 500
500 http://archive.ubuntu.com/ubuntu jammy/main amd64 Packages
500 http://archive.ubuntu.com/ubuntu jammy/main i386 Packages

$ autopkgtest -U  hello*.dsc -- qemu ../autopkgtest-noble-amd64.img
autopkgtest [21:29:25]: starting date and time: 2024-10-21 21:29:25+0100
autopkgtest [21:29:25]: version 5.38ubuntu1~22.04.1
autopkgtest [21:29:25]: host clean-jammy-amd64; command line: 
/usr/bin/autopkgtest -U hello_2.10-3build2.dsc -- qemu 
../autopkgtest-noble-amd64.img
autopkgtest [21:29:41]: testbed dpkg architecture: amd64
autopkgtest [21:29:41]: testbed apt version: 2.7.14build2
autopkgtest [21:29:41]:  test bed setup
(truncated for brevity)
autopkgtest [21:31:32]: test upstream-tests: [---
Testing greeting-1 ...
Testing hello-1 ...
Testing last-1 ...
Testing traditional-1 ...
autopkgtest [21:31:33]: test upstream-tests: ---]
autopkgtest [21:31:34]: test upstream-tests:  - - - - - - - - - - results - - - 
- - - - - - -
upstream-tests   PASS
autopkgtest [21:31:34]:  summary
command1 PASS
upstream-tests   PASS

So the test was completed without a timeout.
This concludes the verification for Jammy.



** Tags added: verification-done-jammy verification-done-noble

-- 
You received this bug notification because you are a member of
Canonical's Ubuntu QA, which is subscribed to autopkgtest in Ubuntu.
Matching subscriptions: ubuntu-qa-bugs
https://bugs.launchpad.net/bugs/2056461

Title:
  autopkgtest-virt-qemu on noble images sometimes hangs doing copydown

Status in Linux:
  Confirmed
Status in autopkgtest package in Ubuntu:
  Fix Released
Status in linux package in Ubuntu:
  In Progress
Status in autopkgtest source package in Jammy:
  In Progress
Status in linux source package in Jammy:
  New
Status in autopkgtest source package in Noble:
  In Progress
Status in linux