Hi Steven, Thank you for your help, I removed sock1.sock and sock2.sock, restarted vpp, atleast interfaces get created. However, when I start dpdk/testpmd inside the container it crashes as well. Below are some details. I am using vpp code from latest repo.
(1) On host show interface Name Idx State Counter Count VhostEthernet2 3 down VhostEthernet3 4 down VirtualFunctionEthernet4/10/4 1 down VirtualFunctionEthernet4/10/6 2 down local0 0 down vpp# vpp# set interface state VhostEthernet2 up vpp# set interface state VhostEthernet3 up vpp# vpp# set interface l2 bridge VhostEthernet2 1 vpp# set interface l2 bridge VhostEthernet3 1 vpp# (2) Run tespmd inside the container docker run -it --privileged -v /var/run/vpp/sock1.sock:/var/run/usvhost1 -v /var/run/vpp/sock2.sock:/var/run/usvhost2 -v /dev/hugepages:/dev/hugepages dpdk-app-testpmd ./bin/testpmd -l 16-19 -n 4 --log-level=8 -m 64 --no-pci --vdev=virtio_user0,path=/var/run/usvhost1,mac=54:00:00:01:01:01 --vdev=virtio_user1,path=/var/run/usvhost2,mac=54:00:00:01:01:02 -- -i EAL: Detected 28 lcore(s) EAL: Detected 2 NUMA nodes EAL: Multi-process socket /var/run/dpdk/rte/mp_socket EAL: 8192 hugepages of size 2097152 reserved, but no mounted hugetlbfs found for that size EAL: Probing VFIO support... EAL: VFIO support initialized EAL: Setting up physically contiguous memory... EAL: locking hot plug lock memory... EAL: primary init32... Interactive-mode selected Warning: NUMA should be configured manually by using --port-numa-config and --ring-numa-config parameters along with --numa. testpmd: create a new mbuf pool <mbuf_pool_socket_0>: n=171456, size=2176, socket=0 testpmd: preferred mempool ops selected: ring_mp_mc testpmd: create a new mbuf pool <mbuf_pool_socket_1>: n=171456, size=2176, socket=1 testpmd: preferred mempool ops selected: ring_mp_mc Port 0 is now not stopped Port 1 is now not stopped Please stop the ports first Done testpmd> (3) VPP crashes with the same issue but inside dpdk code (gdb) cont Continuing. Program received signal SIGSEGV, Segmentation fault. [Switching to Thread 0x7ffd0d08e700 (LWP 41257)] rte_vhost_dequeue_burst (vid=<optimized out>, queue_id=<optimized out>, mbuf_pool=0x7fe17fc883c0, pkts=pkts@entry=0x7fffb671ebc0, count=count@entry=32) at /var/venom/vpp/build-root/build-vpp-native/dpdk/dpdk-stable-18.02.1/lib/librte_vhost/virtio_net.c:1504 1504 free_entries = *((volatile uint16_t *)&vq->avail->idx) - (gdb) bt #0 rte_vhost_dequeue_burst (vid=<optimized out>, queue_id=<optimized out>, mbuf_pool=0x7fe17fc883c0, pkts=pkts@entry=0x7fffb671ebc0, count=count@entry=32) at /var/venom/vpp/build-root/build-vpp-native/dpdk/dpdk-stable-18.02.1/lib/librte_vhost/virtio_net.c:1504 #1 0x00007fffb4718e6f in eth_vhost_rx (q=0x7fe17fbbdd80, bufs=0x7fffb671ebc0, nb_bufs=<optimized out>) at /var/venom/vpp/build-root/build-vpp-native/dpdk/dpdk-stable-18.02.1/drivers/net/vhost/rte_eth_vhost.c:410 #2 0x00007fffb441cb7c in rte_eth_rx_burst (nb_pkts=256, rx_pkts=0x7fffb671ebc0, queue_id=0, port_id=3) at /var/venom/vpp/build-root/install-vpp-native/dpdk/include/dpdk/rte_ethdev.h:3635 #3 dpdk_device_input (queue_id=0, thread_index=<optimized out>, node=0x7fffb732c700, xd=0x7fffb7337240, dm=<optimized out>, vm=0x7fffb6703340) at /var/venom/vpp/build-data/../src/plugins/dpdk/device/node.c:477 #4 dpdk_input_node_fn_avx2 (vm=<optimized out>, node=<optimized out>, f=<optimized out>) at /var/venom/vpp/build-data/../src/plugins/dpdk/device/node.c:658 #5 0x00007ffff7954d35 in dispatch_node (last_time_stamp=12531752723928016, frame=0x0, dispatch_state=VLIB_NODE_STATE_POLLING, type=VLIB_NODE_TYPE_INPUT, node=0x7fffb732c700, vm=0x7fffb6703340) at /var/venom/vpp/build-data/../src/vlib/main.c:988 #6 vlib_main_or_worker_loop (is_main=0, vm=0x7fffb6703340) at /var/venom/vpp/build-data/../src/vlib/main.c:1507 #7 vlib_worker_loop (vm=0x7fffb6703340) at /var/venom/vpp/build-data/../src/vlib/main.c:1641 #8 0x00007ffff6ad25d8 in clib_calljmp () at /var/venom/vpp/build-data/../src/vppinfra/longjmp.S:110 #9 0x00007ffd0d08ddb0 in ?? () #10 0x00007fffb4436edd in eal_thread_loop (arg=<optimized out>) at /var/venom/vpp/build-root/build-vpp-native/dpdk/dpdk-stable-18.02.1/lib/librte_eal/linuxapp/eal/eal_thread.c:153 #11 0x0000000000000000 in ?? () (gdb) frame 0 #0 rte_vhost_dequeue_burst (vid=<optimized out>, queue_id=<optimized out>, mbuf_pool=0x7fe17fc883c0, pkts=pkts@entry=0x7fffb671ebc0, count=count@entry=32) at /var/venom/vpp/build-root/build-vpp-native/dpdk/dpdk-stable-18.02.1/lib/librte_vhost/virtio_net.c:1504 1504 free_entries = *((volatile uint16_t *)&vq->avail->idx) - (gdb) p vq $1 = (struct vhost_virtqueue *) 0x7fc3ffc84b00 (gdb) p vq->avail $2 = (struct vring_avail *) 0x7ffbfff98000 (gdb) p *$2 Cannot access memory at address 0x7ffbfff98000 (gdb) Thanks. On Thu, May 31, 2018 at 12:09 AM, Steven Luong (sluong) <slu...@cisco.com> wrote: > Sorry, I was expecting to see two VhostEthernet interfaces like this. Those > VirtualFunctionEthernet are your physical interfaces. > > sh int > Name Idx State Counter > Count > VhostEthernet0 1 up > VhostEthernet1 2 up > local0 0 down > DBGvpp# > > You have to first manually remove /var/run/vpp/sock1.sock and > /var/run/vpp/sock2.sock before you start vpp on the host. dpdk does not like > it if they already existed. If you successfully create VhostEthernet > interface, try to send some traffic through it to see if it crashes or not. > > Steven > > On 5/30/18, 9:17 PM, "vpp-dev@lists.fd.io on behalf of Steven Luong > (sluong)" <vpp-dev@lists.fd.io on behalf of slu...@cisco.com> wrote: > > Ravi, > > I don't think you can declare (2) works fine yet. Please bring up the > dpdk vhost-user interfaces and try to send some traffic between them to > exercise the shared memory region from dpdk virtio-user which may be > "questionable". > > VirtualFunctionEthernet4/10/4 1 down > VirtualFunctionEthernet4/10/6 2 down > > Steven > > On 5/30/18, 4:41 PM, "Ravi Kerur" <rke...@gmail.com> wrote: > > Hi Steve, > > Thank you for your inputs, I added feature-mask to see if it helps in > setting up queues correctly, it didn't so I will remove it. I have > tried following combination > > (1) VPP->vhost-user (on host) and DPDK/testpmd->virtio-user (in a > container) -- VPP crashes > (2) DPDK/testpmd->vhost-user (on host) and DPDK/testpmd->virtio-user > (in a container) -- works fine > > To use DPDK vhost-user inside VPP, I defined configuration in > startup.conf as mentioned by you and it looks as follows > > unix { > nodaemon > log /var/log/vpp/vpp.log > full-coredump > cli-listen /run/vpp/cli.sock > gid vpp > } > > api-segment { > gid vpp > } > > cpu { > main-core 1 > corelist-workers 6-9 > } > > dpdk { > dev 0000:04:10.4 > dev 0000:04:10.6 > uio-driver vfio-pci > vdev net_vhost0,iface=/var/run/vpp/sock1.sock > vdev net_vhost1,iface=/var/run/vpp/sock2.sock > huge-dir /dev/hugepages_1GB > socket-mem 2048,2048 > } > > From VPP logs > dpdk: EAL init args: -c 3c2 -n 4 --vdev > net_vhost0,iface=/var/run/vpp/sock1.sock --vdev > net_vhost1,iface=/var/run/vpp/sock2.sock --huge-dir /dev/hugepages_1GB > -w 0000:04:10.4 -w 0000:04:10.6 --master-lcore 1 --socket-mem > 2048,2048 > > However, VPP doesn't create interface at all > > vpp# show interface > Name Idx State Counter > Count > VirtualFunctionEthernet4/10/4 1 down > VirtualFunctionEthernet4/10/6 2 down > local0 0 down > > since it is a static mapping I am assuming it should be created, > correct? > > Thanks. > > On Wed, May 30, 2018 at 3:43 PM, Steven Luong (sluong) > <slu...@cisco.com> wrote: > > Ravi, > > > > First and foremost, get rid of the feature-mask option. I don't > know what 0x40400000 does for you. If that does not help, try testing it with > dpdk based vhost-user instead of VPP native vhost-user to make sure that they > can work well with each other first. To use dpdk vhost-user, add a vdev > command in the startup.conf for each vhost-user device that you have. > > > > dpdk { vdev net_vhost0,iface=/var/run/vpp/sock1.sock } > > > > dpdk based vhost-user interface is named VhostEthernet0, > VhostEthernet1, etc. Make sure you use the right interface name to set the > state to up. > > > > If dpdk based vhost-user does not work with testpmd either, it > looks like some problem with the way that you invoke testpmd. > > > > If dpdk based vhost-user works well with the same testpmd device > driver and not vpp native vhost-user, I can set up something similar to yours > to look into it. > > > > The device driver, testpmd, is supposed to pass the shared memory > region to VPP for TX/RX queues. It looks like VPP vhost-user might have run > into a bump there with using the shared memory (txvq->avail). > > > > Steven > > > > PS. vhost-user is not an optimum interface for containers. You may > want to look into using memif if you don't already know about it. > > > > > > On 5/30/18, 2:06 PM, "Ravi Kerur" <rke...@gmail.com> wrote: > > > > I am not sure what is wrong with the setup or a bug in vpp, vpp > > crashes with vhost<-->virtio communication. > > > > (1) Vhost-interfaces are created and attached to bridge-domain > as follows > > > > create vhost socket /var/run/vpp/sock1.sock server feature-mask > 0x40400000 > > create vhost socket /var/run/vpp/sock2.sock server feature-mask > 0x40400000 > > set interface state VirtualEthernet0/0/0 up > > set interface state VirtualEthernet0/0/1 up > > > > set interface l2 bridge VirtualEthernet0/0/0 1 > > set interface l2 bridge VirtualEthernet0/0/1 1 > > > > > > (2) DPDK/testpmd is started in a container to talk to > vpp/vhost-user > > interface as follows > > > > docker run -it --privileged -v > > /var/run/vpp/sock1.sock:/var/run/usvhost1 -v > > /var/run/vpp/sock2.sock:/var/run/usvhost2 -v > > /dev/hugepages:/dev/hugepages dpdk-app-testpmd ./bin/testpmd -c > 0x3 -n > > 4 --log-level=9 -m 64 --no-pci --single-file-segments > > --vdev=virtio_user0,path=/var/run/usvhost1,mac=54:00:00:01:01:01 > > > --vdev=virtio_user1,path=/var/run/usvhost2,mac=54:00:00:01:01:02 -- > > -i > > > > (3) show vhost-user VirtualEthernet0/0/1 > > Virtio vhost-user interfaces > > Global: > > coalesce frames 32 time 1e-3 > > number of rx virtqueues in interrupt mode: 0 > > Interface: VirtualEthernet0/0/1 (ifindex 4) > > virtio_net_hdr_sz 10 > > features mask (0x40400000): > > features (0x0): > > protocol features (0x0) > > > > socket filename /var/run/vpp/sock2.sock type server errno > "Success" > > > > rx placement: > > tx placement: spin-lock > > thread 0 on vring 0 > > thread 1 on vring 0 > > thread 2 on vring 0 > > thread 3 on vring 0 > > thread 4 on vring 0 > > > > Memory regions (total 1) > > region fd guest_phys_addr memory_size > userspace_addr > > mmap_offset mmap_addr > > ====== ===== ================== ================== > ================== > > ================== ================== > > 0 55 0x00007ff7c0000000 0x0000000040000000 > 0x00007ff7c0000000 > > 0x0000000000000000 0x00007ffbc0000000 > > > > vpp# show vhost-user VirtualEthernet0/0/0 > > Virtio vhost-user interfaces > > Global: > > coalesce frames 32 time 1e-3 > > number of rx virtqueues in interrupt mode: 0 > > Interface: VirtualEthernet0/0/0 (ifindex 3) > > virtio_net_hdr_sz 10 > > features mask (0x40400000): > > features (0x0): > > protocol features (0x0) > > > > socket filename /var/run/vpp/sock1.sock type server errno > "Success" > > > > rx placement: > > tx placement: spin-lock > > thread 0 on vring 0 > > thread 1 on vring 0 > > thread 2 on vring 0 > > thread 3 on vring 0 > > thread 4 on vring 0 > > > > Memory regions (total 1) > > region fd guest_phys_addr memory_size > userspace_addr > > mmap_offset mmap_addr > > ====== ===== ================== ================== > ================== > > ================== ================== > > 0 51 0x00007ff7c0000000 0x0000000040000000 > 0x00007ff7c0000000 > > 0x0000000000000000 0x00007ffc00000000 > > > > (4) vpp stack trace > > Program received signal SIGSEGV, Segmentation fault. > > [Switching to Thread 0x7ffd0e090700 (LWP 46570)] > > 0x00007ffff7414642 in vhost_user_if_input > > (mode=VNET_HW_INTERFACE_RX_MODE_POLLING, > > node=0x7fffb76bab00, qid=<optimized out>, > vui=0x7fffb6739700, > > vum=0x7ffff78f4480 <vhost_user_main>, vm=0x7fffb672a9c0) > > at > /var/venom/vpp/build-data/../src/vnet/devices/virtio/vhost-user.c:1596 > > 1596 if (PREDICT_FALSE (txvq->avail->flags & 0xFFFE)) > > (gdb) bt > > #0 0x00007ffff7414642 in vhost_user_if_input > > (mode=VNET_HW_INTERFACE_RX_MODE_POLLING, > > node=0x7fffb76bab00, qid=<optimized out>, > vui=0x7fffb6739700, > > vum=0x7ffff78f4480 <vhost_user_main>, vm=0x7fffb672a9c0) > > at > /var/venom/vpp/build-data/../src/vnet/devices/virtio/vhost-user.c:1596 > > #1 vhost_user_input (f=<optimized out>, node=<optimized out>, > > vm=<optimized out>) > > at > /var/venom/vpp/build-data/../src/vnet/devices/virtio/vhost-user.c:1947 > > #2 vhost_user_input_avx2 (vm=<optimized out>, node=<optimized > out>, > > frame=<optimized out>) > > at > /var/venom/vpp/build-data/../src/vnet/devices/virtio/vhost-user.c:1972 > > #3 0x00007ffff7954d35 in dispatch_node > > (last_time_stamp=12391212490024174, frame=0x0, > > dispatch_state=VLIB_NODE_STATE_POLLING, > type=VLIB_NODE_TYPE_INPUT, > > node=0x7fffb76bab00, > > vm=0x7fffb672a9c0) at > /var/venom/vpp/build-data/../src/vlib/main.c:988 > > #4 vlib_main_or_worker_loop (is_main=0, vm=0x7fffb672a9c0) > > at /var/venom/vpp/build-data/../src/vlib/main.c:1507 > > #5 vlib_worker_loop (vm=0x7fffb672a9c0) at > > /var/venom/vpp/build-data/../src/vlib/main.c:1641 > > #6 0x00007ffff6ad25d8 in clib_calljmp () > > at /var/venom/vpp/build-data/../src/vppinfra/longjmp.S:110 > > #7 0x00007ffd0e08fdb0 in ?? () > > #8 0x00007fffb4436edd in eal_thread_loop (arg=<optimized out>) > > at > /var/venom/vpp/build-root/build-vpp-native/dpdk/dpdk-stable-18.02.1/lib/librte_eal/linuxapp/eal/eal_thread.c:153 > > #9 0x0000000000000000 in ?? () > > (gdb) frame 0 > > #0 0x00007ffff7414642 in vhost_user_if_input > > (mode=VNET_HW_INTERFACE_RX_MODE_POLLING, > > node=0x7fffb76bab00, qid=<optimized out>, > vui=0x7fffb6739700, > > vum=0x7ffff78f4480 <vhost_user_main>, vm=0x7fffb672a9c0) > > at > /var/venom/vpp/build-data/../src/vnet/devices/virtio/vhost-user.c:1596 > > 1596 if (PREDICT_FALSE (txvq->avail->flags & 0xFFFE)) > > (gdb) p txvq > > $1 = (vhost_user_vring_t *) 0x7fffb6739ac0 > > (gdb) p *txvq > > $2 = {cacheline0 = 0x7fffb6739ac0 "?", qsz_mask = 255, > last_avail_idx > > = 0, last_used_idx = 0, > > n_since_last_int = 0, desc = 0x7ffbfff97000, avail = > 0x7ffbfff98000, > > used = 0x7ffbfff99000, > > int_deadline = 0, started = 1 '\001', enabled = 0 '\000', > log_used = 0 '\000', > > cacheline1 = 0x7fffb6739b00 "????\n", errfd = -1, callfd_idx > = 10, > > kickfd_idx = 14, > > log_guest_addr = 0, mode = 1} > > (gdb) p *(txvq->avail) > > Cannot access memory at address 0x7ffbfff98000 > > (gdb) > > > > On Tue, May 29, 2018 at 10:47 AM, Ravi Kerur <rke...@gmail.com> > wrote: > > > Steve, > > > > > > Thanks for inputs on debugs and gdb. I am using gdb on my > development > > > system to debug the issue. I would like to have reliable core > > > generation on the system on which I don't have access to > install gdb. > > > I installed corekeeper and it still doesn't generate core. I > am > > > running vpp inside a VM (VirtualBox/vagrant), not sure if I > need to > > > set something inside vagrant config file. > > > > > > dpkg -l corekeeper > > > Desired=Unknown/Install/Remove/Purge/Hold > > > | > Status=Not/Inst/Conf-files/Unpacked/halF-conf/Half-inst/trig-aWait/Trig-pend > > > |/ Err?=(none)/Reinst-required (Status,Err: uppercase=bad) > > > ||/ Name Version Architecture > Description > > > > +++-====================-===============-===============-============================================== > > > ii corekeeper 1.6 amd64 > enable core > > > files and report crashes to the system > > > > > > Thanks. > > > > > > On Tue, May 29, 2018 at 9:38 AM, Steven Luong (sluong) > <slu...@cisco.com> wrote: > > >> Ravi, > > >> > > >> I install corekeeper and the core file is kept in > /var/crash. But why not use gdb to attach to the VPP process? > > >> To turn on VPP vhost-user debug, type "debug vhost-user on" > at the VPP prompt. > > >> > > >> Steven > > >> > > >> On 5/29/18, 9:10 AM, "vpp-dev@lists.fd.io on behalf of Ravi > Kerur" <vpp-dev@lists.fd.io on behalf of rke...@gmail.com> wrote: > > >> > > >> Hi Marco, > > >> > > >> > > >> On Tue, May 29, 2018 at 6:30 AM, Marco Varlese > <mvarl...@suse.de> wrote: > > >> > Ravi, > > >> > > > >> > On Sun, 2018-05-27 at 12:20 -0700, Ravi Kerur wrote: > > >> >> Hello, > > >> >> > > >> >> I have a VM(16.04.4 Ubuntu x86_64) with 2 cores and > 4G RAM. I have > > >> >> installed VPP successfully on it. Later I have > created vhost-user > > >> >> interfaces via > > >> >> > > >> >> create vhost socket /var/run/vpp/sock1.sock server > > >> >> create vhost socket /var/run/vpp/sock2.sock server > > >> >> set interface state VirtualEthernet0/0/0 up > > >> >> set interface state VirtualEthernet0/0/1 up > > >> >> > > >> >> set interface l2 bridge VirtualEthernet0/0/0 1 > > >> >> set interface l2 bridge VirtualEthernet0/0/1 1 > > >> >> > > >> >> I then run 'DPDK/testpmd' inside a container which > will use > > >> >> virtio-user interfaces using the following command > > >> >> > > >> >> docker run -it --privileged -v > > >> >> /var/run/vpp/sock1.sock:/var/run/usvhost1 -v > > >> >> /var/run/vpp/sock2.sock:/var/run/usvhost2 -v > > >> >> /dev/hugepages:/dev/hugepages dpdk-app-testpmd > ./bin/testpmd -c 0x3 -n > > >> >> 4 --log-level=9 -m 64 --no-pci --single-file-segments > > >> >> > --vdev=virtio_user0,path=/var/run/usvhost1,mac=54:01:00:01:01:01 > > >> >> > --vdev=virtio_user1,path=/var/run/usvhost2,mac=54:01:00:01:01:02 -- > > >> >> -i > > >> >> > > >> >> VPP Vnet crashes with following message > > >> >> > > >> >> May 27 11:44:00 localhost vnet[6818]: received signal > SIGSEGV, PC > > >> >> 0x7fcca4620187, faulting address 0x7fcb317ac000 > > >> >> > > >> >> Questions: > > >> >> I have 'ulimit -c unlimited' and > /etc/vpp/startup.conf has > > >> >> unix { > > >> >> nodaemon > > >> >> log /var/log/vpp/vpp.log > > >> >> full-coredump > > >> >> cli-listen /run/vpp/cli.sock > > >> >> gid vpp > > >> >> } > > >> >> > > >> >> But I couldn't locate corefile? > > >> > The location of the coredump file depends on your > system configuration. > > >> > > > >> > Please, check "cat /proc/sys/kernel/core_pattern" > > >> > > > >> > If you have systemd-coredump in the output of the > above command, then likely the > > >> > location of the coredump files is > "/var/lib/systemd/coredump/" > > >> > > > >> > You can also change the location of where your system > places the coredump files: > > >> > echo '/PATH_TO_YOU_LOCATION/core_%e.%p' | sudo tee > /proc/sys/kernel/core_pattern > > >> > > > >> > See if that helps... > > >> > > > >> > > >> Initially '/proc/sys/kernel/core_pattern' was set to > 'core'. I changed > > >> it to 'systemd-coredump'. Still no core generated. VPP > crashes > > >> > > >> May 29 08:54:34 localhost vnet[4107]: received signal > SIGSEGV, PC > > >> 0x7f0167751187, faulting address 0x7efff43ac000 > > >> May 29 08:54:34 localhost systemd[1]: vpp.service: Main > process > > >> exited, code=killed, status=6/ABRT > > >> May 29 08:54:34 localhost systemd[1]: vpp.service: Unit > entered failed state. > > >> May 29 08:54:34 localhost systemd[1]: vpp.service: > Failed with result 'signal'. > > >> > > >> > > >> cat /proc/sys/kernel/core_pattern > > >> systemd-coredump > > >> > > >> > > >> ulimit -a > > >> core file size (blocks, -c) unlimited > > >> data seg size (kbytes, -d) unlimited > > >> scheduling priority (-e) 0 > > >> file size (blocks, -f) unlimited > > >> pending signals (-i) 15657 > > >> max locked memory (kbytes, -l) 64 > > >> max memory size (kbytes, -m) unlimited > > >> open files (-n) 1024 > > >> pipe size (512 bytes, -p) 8 > > >> POSIX message queues (bytes, -q) 819200 > > >> real-time priority (-r) 0 > > >> stack size (kbytes, -s) 8192 > > >> cpu time (seconds, -t) unlimited > > >> max user processes (-u) 15657 > > >> virtual memory (kbytes, -v) unlimited > > >> file locks (-x) unlimited > > >> > > >> cd /var/lib/systemd/coredump/ > > >> root@localhost:/var/lib/systemd/coredump# ls > > >> root@localhost:/var/lib/systemd/coredump# > > >> > > >> >> > > >> >> (2) How to enable debugs? I have used 'make build' > but no additional > > >> >> logs other than those shown below > > >> >> > > >> >> > > >> >> VPP logs from /var/log/syslog is shown below > > >> >> cat /var/log/syslog > > >> >> May 27 11:40:28 localhost vpp[6818]: > vlib_plugin_early_init:361: > > >> >> plugin path > /usr/lib/vpp_plugins:/usr/lib64/vpp_plugins > > >> >> May 27 11:40:28 localhost vpp[6818]: > load_one_plugin:189: Loaded > > >> >> plugin: abf_plugin.so (ACL based Forwarding) > > >> >> May 27 11:40:28 localhost vpp[6818]: > load_one_plugin:189: Loaded > > >> >> plugin: acl_plugin.so (Access Control Lists) > > >> >> May 27 11:40:28 localhost vpp[6818]: > load_one_plugin:189: Loaded > > >> >> plugin: avf_plugin.so (Intel Adaptive Virtual > Function (AVF) Device > > >> >> Plugin) > > >> >> May 27 11:40:28 localhost vpp[6818]: > load_one_plugin:191: Loaded > > >> >> plugin: cdp_plugin.so > > >> >> May 27 11:40:28 localhost vpp[6818]: > load_one_plugin:189: Loaded > > >> >> plugin: dpdk_plugin.so (Data Plane Development Kit > (DPDK)) > > >> >> May 27 11:40:28 localhost vpp[6818]: > load_one_plugin:189: Loaded > > >> >> plugin: flowprobe_plugin.so (Flow per Packet) > > >> >> May 27 11:40:28 localhost vpp[6818]: > load_one_plugin:189: Loaded > > >> >> plugin: gbp_plugin.so (Group Based Policy) > > >> >> May 27 11:40:28 localhost vpp[6818]: > load_one_plugin:189: Loaded > > >> >> plugin: gtpu_plugin.so (GTPv1-U) > > >> >> May 27 11:40:28 localhost vpp[6818]: > load_one_plugin:189: Loaded > > >> >> plugin: igmp_plugin.so (IGMP messaging) > > >> >> May 27 11:40:28 localhost vpp[6818]: > load_one_plugin:189: Loaded > > >> >> plugin: ila_plugin.so (Identifier-locator addressing > for IPv6) > > >> >> May 27 11:40:28 localhost vpp[6818]: > load_one_plugin:189: Loaded > > >> >> plugin: ioam_plugin.so (Inbound OAM) > > >> >> May 27 11:40:28 localhost vpp[6818]: > load_one_plugin:117: Plugin > > >> >> disabled (default): ixge_plugin.so > > >> >> May 27 11:40:28 localhost vpp[6818]: > load_one_plugin:189: Loaded > > >> >> plugin: l2e_plugin.so (L2 Emulation) > > >> >> May 27 11:40:28 localhost vpp[6818]: > load_one_plugin:189: Loaded > > >> >> plugin: lacp_plugin.so (Link Aggregation Control > Protocol) > > >> >> May 27 11:40:28 localhost vpp[6818]: > load_one_plugin:189: Loaded > > >> >> plugin: lb_plugin.so (Load Balancer) > > >> >> May 27 11:40:28 localhost vpp[6818]: > load_one_plugin:189: Loaded > > >> >> plugin: memif_plugin.so (Packet Memory Interface > (experimetal)) > > >> >> May 27 11:40:28 localhost vpp[6818]: > load_one_plugin:189: Loaded > > >> >> plugin: nat_plugin.so (Network Address Translation) > > >> >> May 27 11:40:28 localhost vpp[6818]: > load_one_plugin:189: Loaded > > >> >> plugin: pppoe_plugin.so (PPPoE) > > >> >> May 27 11:40:28 localhost vpp[6818]: > load_one_plugin:189: Loaded > > >> >> plugin: srv6ad_plugin.so (Dynamic SRv6 proxy) > > >> >> May 27 11:40:28 localhost vpp[6818]: > load_one_plugin:189: Loaded > > >> >> plugin: srv6am_plugin.so (Masquerading SRv6 proxy) > > >> >> May 27 11:40:28 localhost vpp[6818]: > load_one_plugin:189: Loaded > > >> >> plugin: srv6as_plugin.so (Static SRv6 proxy) > > >> >> May 27 11:40:29 localhost vpp[6818]: > load_one_plugin:189: Loaded > > >> >> plugin: stn_plugin.so (VPP Steals the NIC for > Container integration) > > >> >> May 27 11:40:29 localhost vpp[6818]: > load_one_plugin:189: Loaded > > >> >> plugin: tlsmbedtls_plugin.so (mbedtls based TLS > Engine) > > >> >> May 27 11:40:29 localhost vpp[6818]: > load_one_plugin:189: Loaded > > >> >> plugin: tlsopenssl_plugin.so (openssl based TLS > Engine) > > >> >> May 27 11:40:29 localhost vpp[6818]: > /usr/bin/vpp[6818]: > > >> >> load_one_vat_plugin:67: Loaded plugin: > dpdk_test_plugin.so > > >> >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: > load_one_vat_plugin:67: > > >> >> Loaded plugin: dpdk_test_plugin.so > > >> >> May 27 11:40:29 localhost vpp[6818]: > /usr/bin/vpp[6818]: > > >> >> load_one_vat_plugin:67: Loaded plugin: > lb_test_plugin.so > > >> >> May 27 11:40:29 localhost vpp[6818]: > /usr/bin/vpp[6818]: > > >> >> load_one_vat_plugin:67: Loaded plugin: > flowprobe_test_plugin.so > > >> >> May 27 11:40:29 localhost vpp[6818]: > /usr/bin/vpp[6818]: > > >> >> load_one_vat_plugin:67: Loaded plugin: > stn_test_plugin.so > > >> >> May 27 11:40:29 localhost vpp[6818]: > /usr/bin/vpp[6818]: > > >> >> load_one_vat_plugin:67: Loaded plugin: > nat_test_plugin.so > > >> >> May 27 11:40:29 localhost vpp[6818]: > /usr/bin/vpp[6818]: > > >> >> load_one_vat_plugin:67: Loaded plugin: > udp_ping_test_plugin.so > > >> >> May 27 11:40:29 localhost vpp[6818]: > /usr/bin/vpp[6818]: > > >> >> load_one_vat_plugin:67: Loaded plugin: > pppoe_test_plugin.so > > >> >> May 27 11:40:29 localhost vpp[6818]: > /usr/bin/vpp[6818]: > > >> >> load_one_vat_plugin:67: Loaded plugin: > lacp_test_plugin.so > > >> >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: > load_one_vat_plugin:67: > > >> >> Loaded plugin: lb_test_plugin.so > > >> >> May 27 11:40:29 localhost vpp[6818]: > /usr/bin/vpp[6818]: > > >> >> load_one_vat_plugin:67: Loaded plugin: > acl_test_plugin.so > > >> >> May 27 11:40:29 localhost vpp[6818]: > /usr/bin/vpp[6818]: > > >> >> load_one_vat_plugin:67: Loaded plugin: > ioam_export_test_plugin.so > > >> >> May 27 11:40:29 localhost vpp[6818]: > /usr/bin/vpp[6818]: > > >> >> load_one_vat_plugin:67: Loaded plugin: > ioam_trace_test_plugin.so > > >> >> May 27 11:40:29 localhost vpp[6818]: > /usr/bin/vpp[6818]: > > >> >> load_one_vat_plugin:67: Loaded plugin: > > >> >> vxlan_gpe_ioam_export_test_plugin.so > > >> >> May 27 11:40:29 localhost vpp[6818]: > /usr/bin/vpp[6818]: > > >> >> load_one_vat_plugin:67: Loaded plugin: > gtpu_test_plugin.so > > >> >> May 27 11:40:29 localhost vpp[6818]: > /usr/bin/vpp[6818]: > > >> >> load_one_vat_plugin:67: Loaded plugin: > cdp_test_plugin.so > > >> >> May 27 11:40:29 localhost vpp[6818]: > /usr/bin/vpp[6818]: > > >> >> load_one_vat_plugin:67: Loaded plugin: > ioam_vxlan_gpe_test_plugin.so > > >> >> May 27 11:40:29 localhost vpp[6818]: > /usr/bin/vpp[6818]: > > >> >> load_one_vat_plugin:67: Loaded plugin: > memif_test_plugin.so > > >> >> May 27 11:40:29 localhost vpp[6818]: > /usr/bin/vpp[6818]: > > >> >> load_one_vat_plugin:67: Loaded plugin: > ioam_pot_test_plugin.so > > >> >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: > load_one_vat_plugin:67: > > >> >> Loaded plugin: flowprobe_test_plugin.so > > >> >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: > load_one_vat_plugin:67: > > >> >> Loaded plugin: stn_test_plugin.so > > >> >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: > load_one_vat_plugin:67: > > >> >> Loaded plugin: nat_test_plugin.so > > >> >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: > load_one_vat_plugin:67: > > >> >> Loaded plugin: udp_ping_test_plugin.so > > >> >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: > load_one_vat_plugin:67: > > >> >> Loaded plugin: pppoe_test_plugin.so > > >> >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: > load_one_vat_plugin:67: > > >> >> Loaded plugin: lacp_test_plugin.so > > >> >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: > load_one_vat_plugin:67: > > >> >> Loaded plugin: acl_test_plugin.so > > >> >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: > load_one_vat_plugin:67: > > >> >> Loaded plugin: ioam_export_test_plugin.so > > >> >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: > load_one_vat_plugin:67: > > >> >> Loaded plugin: ioam_trace_test_plugin.so > > >> >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: > load_one_vat_plugin:67: > > >> >> Loaded plugin: vxlan_gpe_ioam_export_test_plugin.so > > >> >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: > load_one_vat_plugin:67: > > >> >> Loaded plugin: gtpu_test_plugin.so > > >> >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: > load_one_vat_plugin:67: > > >> >> Loaded plugin: cdp_test_plugin.so > > >> >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: > load_one_vat_plugin:67: > > >> >> Loaded plugin: ioam_vxlan_gpe_test_plugin.so > > >> >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: > load_one_vat_plugin:67: > > >> >> Loaded plugin: memif_test_plugin.so > > >> >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: > load_one_vat_plugin:67: > > >> >> Loaded plugin: ioam_pot_test_plugin.so > > >> >> May 27 11:40:29 localhost vpp[6818]: > /usr/bin/vpp[6818]: dpdk: EAL > > >> >> init args: -c 1 -n 4 --no-pci --huge-dir > /dev/hugepages --master-lcore > > >> >> 0 --socket-mem 256,0 > > >> >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: dpdk: > EAL init args: -c > > >> >> 1 -n 4 --no-pci --huge-dir /dev/hugepages > --master-lcore 0 > > >> >> --socket-mem 256,0 > > >> >> May 27 11:40:29 localhost vnet[6818]: > dpdk_ipsec_process:1019: not > > >> >> enough DPDK crypto resources, default to OpenSSL > > >> >> May 27 11:43:19 localhost vnet[6818]: show > vhost-user: unknown input `detail > > >> >> May 27 11:44:00 localhost vnet[6818]: received signal > SIGSEGV, PC > > >> >> 0x7fcca4620187, faulting address 0x7fcb317ac000 > > >> >> May 27 11:44:00 localhost systemd[1]: vpp.service: > Main process > > >> >> exited, code=killed, status=6/ABRT > > >> >> May 27 11:44:00 localhost systemd[1]: vpp.service: > Unit entered failed state. > > >> >> May 27 11:44:00 localhost systemd[1]: vpp.service: > Failed with result > > >> >> 'signal'. > > >> >> May 27 11:44:00 localhost systemd[1]: vpp.service: > Service hold-off > > >> >> time over, scheduling restart > > >> >> > > >> > > >> Thanks, > > >> Ravi > > >> > > >> >> > > >> >> > > >> >> Thanks. > > >> > Cheers, > > >> > Marco > > >> >> > > >> >> > > >> >> > > >> > -- > > >> > Marco V > > >> > > > >> > SUSE LINUX GmbH | GF: Felix Imendörffer, Jane > Smithard, Graham Norton > > >> > HRB 21284 (AG Nürnberg) Maxfeldstr. 5, D-90409, > Nürnberg > > >> > > >> > > >> > > >> > > >> > > > > > > > > > > > > > > > > > > > > -=-=-=-=-=-=-=-=-=-=-=- Links: You receive all messages sent to this group. View/Reply Online (#9478): https://lists.fd.io/g/vpp-dev/message/9478 View All Messages In Topic (12): https://lists.fd.io/g/vpp-dev/topic/20346431 Mute This Topic: https://lists.fd.io/mt/20346431/21656 New Topic: https://lists.fd.io/g/vpp-dev/post Change Your Subscription: https://lists.fd.io/g/vpp-dev/editsub/21656 Group Home: https://lists.fd.io/g/vpp-dev Contact Group Owner: vpp-dev+ow...@lists.fd.io Terms of Service: https://lists.fd.io/static/tos Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub -=-=-=-=-=-=-=-=-=-=-=-