Dear Slava,

Your guess is corrrect.
When I put flow into Connect-X5, it was successful.

General question.
Are there any way to input flow to ConnectX-4?
In another word, are there any way to activate Verb?
And which type of flow is supported in Verb?

-----------------------------------------------------------
tx_h-yamashita@R730n10:~/dpdk-next-net/x86_64-native-linuxapp-gcc/app$ sudo 
./te          stpmd -c 0xF -n 4 -w 04:00.0,dv_flow_en=1 --socket-mem 512,512 
--huge-dir=/mnt/h
uge1G --log-level port:8 -- -i --portmask=0x1 --nb-cores=2 --txq=16 --rxq=16
[sudo] password for tx_h-yamashita:
EAL: Detected 48 lcore(s)
EAL: Detected 2 NUMA nodes
EAL: Multi-process socket /var/run/dpdk/rte/mp_socket
EAL: Selected IOVA mode 'PA'
EAL: Probing VFIO support...
EAL: PCI device 0000:04:00.0 on NUMA socket 0
EAL:   probe driver: 15b3:1017 net_mlx5
net_mlx5: mlx5.c:1852: mlx5_dev_spawn(): can't query devx port 1 on device 
mlx5_          1

Interactive-mode selected

testpmd: create a new mbuf pool <mbuf_pool_socket_0>: n=171456, size=2176, 
socke          t=0
testpmd: preferred mempool ops selected: ring_mp_mc
testpmd: create a new mbuf pool <mbuf_pool_socket_1>: n=171456, size=2176, 
socke          t=1
testpmd: preferred mempool ops selected: ring_mp_mc

Warning! port-topology=paired and odd forward ports number, the last port will 
p          air with itself.

Configuring Port 0 (socket 0)
Port 0: B8:59:9F:C1:4A:CE
Checking link statuses...
Done
testpmd>
testpmd>  flow create 0 ingress group 1 priority 0 pattern eth dst is 
00:16:3e:2          e:7b:6a / vlan vid is 1480 / end actions of_pop_vlan  / 
queue index 0 / end
Flow rule #0 created
testpmd>
--------------------------------------------------------------------------------------------------------------

BR,
Hideyuki Yamashita
NTT TechnoCross

> Hi, Hideyuki
> 
> > -----Original Message-----
> > From: Hideyuki Yamashita <yamashita.hidey...@ntt-tx.co.jp>
> > Sent: Wednesday, October 30, 2019 12:46
> > To: Slava Ovsiienko <viachesl...@mellanox.com>
> > Cc: dev@dpdk.org
> > Subject: Re: [dpdk-dev] [PATCH 0/7] net/mlx5: support for flow action on
> > VLAN header
> > 
> > Hello Slava,
> > 
> > Thanks for your help.
> > I added magic phrase. with chaging PCI number with proper one in my env.
> 
> > It changes situation but still result in error.
> > 
> > I used /usertools/dpdk-setup.sh to allocate hugepage dynamically.
> > Your help is appreciated.
> > 
> > I think it is getting closer.
> > tx_h-yamashita@R730n10:~/dpdk-next-net/x86_64-native-linuxapp-
> > gcc/app$
> > sudo ./testpmd -c 0xF -n 4 -w 03:00.0,dv_flow_en=1 --socket-mem 512,512 -
> > -huge-dir=/mnt/h uge1G --log-level port:8 -- -i --portmask=0x1 --nb-cores=2
> 
> mlx5 PMD supports two flow engines:
> - Verbs, this is legacy one, almost no new features are being added, just bug 
> fixes,
>   provides slow rule insertion rate, etc.
> - Direct Rules, the new one, all new features are being added here.
> 
> (We had one more intermediate engine  - Direct Verbs, it was dropped,
> but prefix dv in dv_flow_en remains ??)
> 
> Verbs are supported over all NICs - ConnectX-4,ConnectX-4LX, ConnectX-5, 
> ConnectX-6, etc.
> Direct Rules is supported for NICs starting from ConnectX-5.
> "dv_flow_en=1" partameter engages Direct Rules, but I see you run testpmd
> over 03:00.0 which is ConnectX-4, not  supporting Direct Rules.
> Please, run over ConnectX-5 you have on your host.
> 
> As for error - it is not related to memory, rdma core just failed to create 
> the group table,
> because ConnectX-4 does not support DR.
> 
> With best regards, Slava
> 
> > --txq=16 --rxq=16
> > EAL: Detected 48 lcore(s)
> > EAL: Detected 2 NUMA nodes
> > EAL: Multi-process socket /var/run/dpdk/rte/mp_socket
> > EAL: Selected IOVA mode 'PA'
> > EAL: Probing VFIO support...
> > EAL: PCI device 0000:03:00.0 on NUMA socket 0
> > EAL:   probe driver: 15b3:1015 net_mlx5
> > net_mlx5: mlx5.c:1852: mlx5_dev_spawn(): can't query devx port 1 on device
> > mlx5_3
> > 
> > Interactive-mode selected
> > testpmd: create a new mbuf pool <mbuf_pool_socket_0>: n=171456,
> > size=2176, socket=0
> > testpmd: preferred mempool ops selected: ring_mp_mc
> > testpmd: create a new mbuf pool <mbuf_pool_socket_1>: n=171456,
> > size=2176, socket=1
> > testpmd: preferred mempool ops selected: ring_mp_mc
> > 
> > Warning! port-topology=paired and odd forward ports number, the last port
> > will pair with itself.
> > 
> > Configuring Port 0 (socket 0)
> > Port 0: B8:59:9F:DB:22:20
> > Checking link statuses...
> > Done
> > testpmd> flow create 0 ingress group 1 priority 0 pattern eth dst is
> > testpmd> 00:16:3e:2e:7b:6a / vlan vid is 1480 / end actions of_pop_vlan
> > testpmd> / queue index 0 / end
> > Caught error type 1 (cause unspecified): cannot create table: Cannot 
> > allocate
> > memory
> > 
> > 
> > BR,
> > Hideyuki Yamashita
> 


Reply via email to