When cast from void pointer to other pointer type, there is not need
to use '()' for force cast.

Signed-off-by: Chaoyong He <chaoyong...@corigine.com>
Reviewed-by: Niklas Söderlund <niklas.soderl...@corigine.com>
---
 drivers/net/nfp/flower/nfp_flower.c           | 20 +++----
 .../net/nfp/flower/nfp_flower_representor.c   | 30 +++++------
 drivers/net/nfp/nfp_common.c                  | 12 ++---
 drivers/net/nfp/nfp_ethdev.c                  |  2 +-
 drivers/net/nfp/nfp_ethdev_vf.c               |  2 +-
 drivers/net/nfp/nfp_flow.c                    | 54 +++++++++----------
 drivers/net/nfp/nfp_rxtx.c                    |  2 +-
 drivers/net/nfp/nfpcore/nfp_cpp_pcie_ops.c    |  2 +-
 8 files changed, 60 insertions(+), 64 deletions(-)

diff --git a/drivers/net/nfp/flower/nfp_flower.c 
b/drivers/net/nfp/flower/nfp_flower.c
index 9a08ae3b75..72933e55d0 100644
--- a/drivers/net/nfp/flower/nfp_flower.c
+++ b/drivers/net/nfp/flower/nfp_flower.c
@@ -60,7 +60,7 @@ nfp_pf_repr_disable_queues(struct rte_eth_dev *dev)
        uint32_t update = 0;
        struct nfp_flower_representor *repr;
 
-       repr = (struct nfp_flower_representor *)dev->data->dev_private;
+       repr = dev->data->dev_private;
        hw = repr->app_fw_flower->pf_hw;
 
        nn_cfg_writeq(hw, NFP_NET_CFG_TXRS_ENABLE, 0);
@@ -89,7 +89,7 @@ nfp_flower_pf_start(struct rte_eth_dev *dev)
        struct nfp_net_hw *hw;
        struct nfp_flower_representor *repr;
 
-       repr = (struct nfp_flower_representor *)dev->data->dev_private;
+       repr = dev->data->dev_private;
        hw = repr->app_fw_flower->pf_hw;
 
        /* Disabling queues just in case... */
@@ -149,19 +149,19 @@ nfp_flower_pf_stop(struct rte_eth_dev *dev)
        struct nfp_net_rxq *this_rx_q;
        struct nfp_flower_representor *repr;
 
-       repr = (struct nfp_flower_representor *)dev->data->dev_private;
+       repr = dev->data->dev_private;
        hw = repr->app_fw_flower->pf_hw;
 
        nfp_pf_repr_disable_queues(dev);
 
        /* Clear queues */
        for (i = 0; i < dev->data->nb_tx_queues; i++) {
-               this_tx_q = (struct nfp_net_txq *)dev->data->tx_queues[i];
+               this_tx_q = dev->data->tx_queues[i];
                nfp_net_reset_tx_queue(this_tx_q);
        }
 
        for (i = 0; i < dev->data->nb_rx_queues; i++) {
-               this_rx_q = (struct nfp_net_rxq *)dev->data->rx_queues[i];
+               this_rx_q = dev->data->rx_queues[i];
                nfp_net_reset_rx_queue(this_rx_q);
        }
 
@@ -189,7 +189,7 @@ nfp_flower_pf_close(struct rte_eth_dev *dev)
        if (rte_eal_process_type() != RTE_PROC_PRIMARY)
                return 0;
 
-       repr = (struct nfp_flower_representor *)dev->data->dev_private;
+       repr = dev->data->dev_private;
        hw = repr->app_fw_flower->pf_hw;
        pf_dev = hw->pf_dev;
        app_fw_flower = NFP_PRIV_TO_APP_FW_FLOWER(pf_dev->app_fw_priv);
@@ -204,12 +204,12 @@ nfp_flower_pf_close(struct rte_eth_dev *dev)
 
        /* Clear queues */
        for (i = 0; i < dev->data->nb_tx_queues; i++) {
-               this_tx_q = (struct nfp_net_txq *)dev->data->tx_queues[i];
+               this_tx_q = dev->data->tx_queues[i];
                nfp_net_reset_tx_queue(this_tx_q);
        }
 
        for (i = 0; i < dev->data->nb_rx_queues; i++) {
-               this_rx_q = (struct nfp_net_rxq *)dev->data->rx_queues[i];
+               this_rx_q = dev->data->rx_queues[i];
                nfp_net_reset_rx_queue(this_rx_q);
        }
 
@@ -814,7 +814,7 @@ nfp_flower_init_ctrl_vnic(struct nfp_net_hw *hw)
 
                /* Saving physical and virtual addresses for the RX ring */
                rxq->dma = (uint64_t)tz->iova;
-               rxq->rxds = (struct nfp_net_rx_desc *)tz->addr;
+               rxq->rxds = tz->addr;
 
                /* Mbuf pointers array for referencing mbufs linked to RX 
descriptors */
                rxq->rxbufs = rte_zmalloc_socket("rxq->rxbufs",
@@ -877,7 +877,7 @@ nfp_flower_init_ctrl_vnic(struct nfp_net_hw *hw)
 
                /* Saving physical and virtual addresses for the TX ring */
                txq->dma = (uint64_t)tz->iova;
-               txq->txds = (struct nfp_net_nfd3_tx_desc *)tz->addr;
+               txq->txds = tz->addr;
 
                /* Mbuf pointers array for referencing mbufs linked to TX 
descriptors */
                txq->txbufs = rte_zmalloc_socket("txq->txbufs",
diff --git a/drivers/net/nfp/flower/nfp_flower_representor.c 
b/drivers/net/nfp/flower/nfp_flower_representor.c
index 0479eb4792..86b7d1a220 100644
--- a/drivers/net/nfp/flower/nfp_flower_representor.c
+++ b/drivers/net/nfp/flower/nfp_flower_representor.c
@@ -32,7 +32,7 @@ nfp_pf_repr_rx_queue_setup(struct rte_eth_dev *dev,
        const struct rte_memzone *tz;
        struct nfp_flower_representor *repr;
 
-       repr = (struct nfp_flower_representor *)dev->data->dev_private;
+       repr = dev->data->dev_private;
        hw = repr->app_fw_flower->pf_hw;
 
        /* Allocating rx queue data structure */
@@ -78,7 +78,7 @@ nfp_pf_repr_rx_queue_setup(struct rte_eth_dev *dev,
 
        /* Saving physical and virtual addresses for the RX ring */
        rxq->dma = (uint64_t)tz->iova;
-       rxq->rxds = (struct nfp_net_rx_desc *)tz->addr;
+       rxq->rxds = tz->addr;
 
        /* mbuf pointers array for referencing mbufs linked to RX descriptors */
        rxq->rxbufs = rte_zmalloc_socket("rxq->rxbufs",
@@ -116,7 +116,7 @@ nfp_pf_repr_tx_queue_setup(struct rte_eth_dev *dev,
        const struct rte_memzone *tz;
        struct nfp_flower_representor *repr;
 
-       repr = (struct nfp_flower_representor *)dev->data->dev_private;
+       repr = dev->data->dev_private;
        hw = repr->app_fw_flower->pf_hw;
 
        tx_free_thresh = (tx_conf->tx_free_thresh) ? tx_conf->tx_free_thresh :
@@ -161,7 +161,7 @@ nfp_pf_repr_tx_queue_setup(struct rte_eth_dev *dev,
 
        /* Saving physical and virtual addresses for the TX ring */
        txq->dma = (uint64_t)tz->iova;
-       txq->txds = (struct nfp_net_nfd3_tx_desc *)tz->addr;
+       txq->txds = tz->addr;
 
        /* mbuf pointers array for referencing mbufs linked to TX descriptors */
        txq->txbufs = rte_zmalloc_socket("txq->txbufs",
@@ -207,7 +207,7 @@ nfp_flower_repr_link_update(struct rte_eth_dev *dev,
                [NFP_NET_CFG_STS_LINK_RATE_100G]        = 
RTE_ETH_SPEED_NUM_100G,
        };
 
-       repr = (struct nfp_flower_representor *)dev->data->dev_private;
+       repr = dev->data->dev_private;
        link = &repr->link;
        pf_hw = repr->app_fw_flower->pf_hw;
 
@@ -273,7 +273,7 @@ nfp_flower_repr_dev_configure(struct rte_eth_dev *dev)
        struct rte_eth_rxmode *rxmode;
        struct nfp_flower_representor *repr;
 
-       repr = (struct nfp_flower_representor *)dev->data->dev_private;
+       repr = dev->data->dev_private;
        pf_hw = repr->app_fw_flower->pf_hw;
 
        dev_conf = &dev->data->dev_conf;
@@ -295,7 +295,7 @@ nfp_flower_repr_dev_start(struct rte_eth_dev *dev)
        struct nfp_flower_representor *repr;
        struct nfp_app_fw_flower *app_fw_flower;
 
-       repr = (struct nfp_flower_representor *)dev->data->dev_private;
+       repr = dev->data->dev_private;
        app_fw_flower = repr->app_fw_flower;
 
        if (repr->repr_type == NFP_REPR_TYPE_PHYS_PORT) {
@@ -314,7 +314,7 @@ nfp_flower_repr_dev_stop(struct rte_eth_dev *dev)
        struct nfp_flower_representor *repr;
        struct nfp_app_fw_flower *app_fw_flower;
 
-       repr = (struct nfp_flower_representor *)dev->data->dev_private;
+       repr = dev->data->dev_private;
        app_fw_flower = repr->app_fw_flower;
 
        nfp_flower_cmsg_port_mod(app_fw_flower, repr->port_id, false);
@@ -339,7 +339,7 @@ nfp_flower_repr_rx_queue_setup(struct rte_eth_dev *dev,
        struct nfp_net_hw *pf_hw;
        struct nfp_flower_representor *repr;
 
-       repr = (struct nfp_flower_representor *)dev->data->dev_private;
+       repr = dev->data->dev_private;
        pf_hw = repr->app_fw_flower->pf_hw;
 
        /* Allocating rx queue data structure */
@@ -367,7 +367,7 @@ nfp_flower_repr_tx_queue_setup(struct rte_eth_dev *dev,
        struct nfp_net_hw *pf_hw;
        struct nfp_flower_representor *repr;
 
-       repr = (struct nfp_flower_representor *)dev->data->dev_private;
+       repr = dev->data->dev_private;
        pf_hw = repr->app_fw_flower->pf_hw;
 
        /* Allocating tx queue data structure */
@@ -390,7 +390,7 @@ nfp_flower_repr_stats_get(struct rte_eth_dev *ethdev,
 {
        struct nfp_flower_representor *repr;
 
-       repr = (struct nfp_flower_representor *)ethdev->data->dev_private;
+       repr = ethdev->data->dev_private;
        rte_memcpy(stats, &repr->repr_stats, sizeof(struct rte_eth_stats));
 
        return 0;
@@ -401,7 +401,7 @@ nfp_flower_repr_stats_reset(struct rte_eth_dev *ethdev)
 {
        struct nfp_flower_representor *repr;
 
-       repr = (struct nfp_flower_representor *)ethdev->data->dev_private;
+       repr = ethdev->data->dev_private;
        memset(&repr->repr_stats, 0, sizeof(struct rte_eth_stats));
 
        return 0;
@@ -413,7 +413,7 @@ nfp_flower_repr_mac_addr_set(struct rte_eth_dev *ethdev,
 {
        struct nfp_flower_representor *repr;
 
-       repr = (struct nfp_flower_representor *)ethdev->data->dev_private;
+       repr = ethdev->data->dev_private;
        rte_ether_addr_copy(mac_addr, &repr->mac_addr);
        rte_ether_addr_copy(mac_addr, ethdev->data->mac_addrs);
 
@@ -584,7 +584,7 @@ nfp_flower_pf_repr_init(struct rte_eth_dev *eth_dev,
        struct nfp_flower_representor *init_repr_data;
 
        /* Cast the input representor data to the correct struct here */
-       init_repr_data = (struct nfp_flower_representor *)init_params;
+       init_repr_data = init_params;
 
        /* Memory has been allocated in the eth_dev_create() function */
        repr = eth_dev->data->dev_private;
@@ -639,7 +639,7 @@ nfp_flower_repr_init(struct rte_eth_dev *eth_dev,
        struct nfp_flower_representor *init_repr_data;
 
        /* Cast the input representor data to the correct struct here */
-       init_repr_data = (struct nfp_flower_representor *)init_params;
+       init_repr_data = init_params;
        app_fw_flower = init_repr_data->app_fw_flower;
 
        /* Memory has been allocated in the eth_dev_create() function */
diff --git a/drivers/net/nfp/nfp_common.c b/drivers/net/nfp/nfp_common.c
index 634afbf1b5..b7a7296999 100644
--- a/drivers/net/nfp/nfp_common.c
+++ b/drivers/net/nfp/nfp_common.c
@@ -1404,7 +1404,7 @@ nfp_net_irq_unmask(struct rte_eth_dev *dev)
 void
 nfp_net_dev_interrupt_delayed_handler(void *param)
 {
-       struct rte_eth_dev *dev = (struct rte_eth_dev *)param;
+       struct rte_eth_dev *dev = param;
 
        nfp_net_link_update(dev, 0);
        rte_eth_dev_callback_process(dev, RTE_ETH_EVENT_INTR_LSC, NULL);
@@ -1420,7 +1420,7 @@ nfp_net_dev_interrupt_handler(void *param)
 {
        int64_t timeout;
        struct rte_eth_link link;
-       struct rte_eth_dev *dev = (struct rte_eth_dev *)param;
+       struct rte_eth_dev *dev = param;
 
        PMD_DRV_LOG(DEBUG, "We got a LSC interrupt!!!");
 
@@ -1840,7 +1840,7 @@ nfp_net_stop_rx_queue(struct rte_eth_dev *dev)
        struct nfp_net_rxq *this_rx_q;
 
        for (i = 0; i < dev->data->nb_rx_queues; i++) {
-               this_rx_q = (struct nfp_net_rxq *)dev->data->rx_queues[i];
+               this_rx_q = dev->data->rx_queues[i];
                nfp_net_reset_rx_queue(this_rx_q);
        }
 }
@@ -1852,7 +1852,7 @@ nfp_net_close_rx_queue(struct rte_eth_dev *dev)
        struct nfp_net_rxq *this_rx_q;
 
        for (i = 0; i < dev->data->nb_rx_queues; i++) {
-               this_rx_q = (struct nfp_net_rxq *)dev->data->rx_queues[i];
+               this_rx_q = dev->data->rx_queues[i];
                nfp_net_reset_rx_queue(this_rx_q);
                nfp_net_rx_queue_release(dev, i);
        }
@@ -1865,7 +1865,7 @@ nfp_net_stop_tx_queue(struct rte_eth_dev *dev)
        struct nfp_net_txq *this_tx_q;
 
        for (i = 0; i < dev->data->nb_tx_queues; i++) {
-               this_tx_q = (struct nfp_net_txq *)dev->data->tx_queues[i];
+               this_tx_q = dev->data->tx_queues[i];
                nfp_net_reset_tx_queue(this_tx_q);
        }
 }
@@ -1877,7 +1877,7 @@ nfp_net_close_tx_queue(struct rte_eth_dev *dev)
        struct nfp_net_txq *this_tx_q;
 
        for (i = 0; i < dev->data->nb_tx_queues; i++) {
-               this_tx_q = (struct nfp_net_txq *)dev->data->tx_queues[i];
+               this_tx_q = dev->data->tx_queues[i];
                nfp_net_reset_tx_queue(this_tx_q);
                nfp_net_tx_queue_release(dev, i);
        }
diff --git a/drivers/net/nfp/nfp_ethdev.c b/drivers/net/nfp/nfp_ethdev.c
index 3a56726388..06146050ff 100644
--- a/drivers/net/nfp/nfp_ethdev.c
+++ b/drivers/net/nfp/nfp_ethdev.c
@@ -542,7 +542,7 @@ nfp_net_init(struct rte_eth_dev *eth_dev)
                     pci_dev->addr.domain, pci_dev->addr.bus,
                     pci_dev->addr.devid, pci_dev->addr.function);
 
-       hw->ctrl_bar = (uint8_t *)pci_dev->mem_resource[0].addr;
+       hw->ctrl_bar = pci_dev->mem_resource[0].addr;
        if (hw->ctrl_bar == NULL) {
                PMD_DRV_LOG(ERR,
                        "hw->ctrl_bar is NULL. BAR0 not configured");
diff --git a/drivers/net/nfp/nfp_ethdev_vf.c b/drivers/net/nfp/nfp_ethdev_vf.c
index f8135fa0c8..bac8df6142 100644
--- a/drivers/net/nfp/nfp_ethdev_vf.c
+++ b/drivers/net/nfp/nfp_ethdev_vf.c
@@ -289,7 +289,7 @@ nfp_netvf_init(struct rte_eth_dev *eth_dev)
 
        hw = NFP_NET_DEV_PRIVATE_TO_HW(eth_dev->data->dev_private);
 
-       hw->ctrl_bar = (uint8_t *)pci_dev->mem_resource[0].addr;
+       hw->ctrl_bar = pci_dev->mem_resource[0].addr;
        if (hw->ctrl_bar == NULL) {
                PMD_DRV_LOG(ERR,
                        "hw->ctrl_bar is NULL. BAR0 not configured");
diff --git a/drivers/net/nfp/nfp_flow.c b/drivers/net/nfp/nfp_flow.c
index fdf5c0f40e..b46acf6a39 100644
--- a/drivers/net/nfp/nfp_flow.c
+++ b/drivers/net/nfp/nfp_flow.c
@@ -93,7 +93,7 @@ nfp_flow_dev_to_priv(struct rte_eth_dev *dev)
 {
        struct nfp_flower_representor *repr;
 
-       repr = (struct nfp_flower_representor *)dev->data->dev_private;
+       repr = dev->data->dev_private;
        return repr->app_fw_flower->flow_priv;
 }
 
@@ -726,8 +726,7 @@ nfp_flow_key_layers_calculate_items(const struct 
rte_flow_item items[],
                        if (port_id->id >= RTE_MAX_ETHPORTS)
                                return -ERANGE;
                        ethdev = &rte_eth_devices[port_id->id];
-                       representor = (struct nfp_flower_representor *)
-                                       ethdev->data->dev_private;
+                       representor = ethdev->data->dev_private;
                        key_ls->port = rte_cpu_to_be_32(representor->port_id);
                        break;
                case RTE_FLOW_ITEM_TYPE_VLAN:
@@ -2047,7 +2046,7 @@ nfp_flow_action_output(char *act_data,
                return -ERANGE;
 
        ethdev = &rte_eth_devices[port_id->id];
-       representor = (struct nfp_flower_representor 
*)ethdev->data->dev_private;
+       representor = ethdev->data->dev_private;
        act_size = sizeof(struct nfp_fl_act_output);
 
        output = (struct nfp_fl_act_output *)act_data;
@@ -2083,7 +2082,7 @@ nfp_flow_action_set_mac(char *act_data,
        set_eth->head.len_lw  = act_size >> NFP_FL_LW_SIZ;
        set_eth->reserved     = 0;
 
-       set_mac = (const struct rte_flow_action_set_mac *)action->conf;
+       set_mac = action->conf;
        if (mac_src_flag) {
                rte_memcpy(&set_eth->eth_addr[RTE_ETHER_ADDR_LEN],
                                set_mac->mac_addr, RTE_ETHER_ADDR_LEN);
@@ -2133,7 +2132,7 @@ nfp_flow_action_set_ip(char *act_data,
        set_ip->head.len_lw  = act_size >> NFP_FL_LW_SIZ;
        set_ip->reserved     = 0;
 
-       set_ipv4 = (const struct rte_flow_action_set_ipv4 *)action->conf;
+       set_ipv4 = action->conf;
        if (ip_src_flag) {
                set_ip->ipv4_src = set_ipv4->ipv4_addr;
                set_ip->ipv4_src_mask = RTE_BE32(0xffffffff);
@@ -2154,7 +2153,7 @@ nfp_flow_action_set_ipv6(char *act_data,
        const struct rte_flow_action_set_ipv6 *set_ipv6;
 
        set_ip = (struct nfp_fl_act_set_ipv6_addr *)act_data;
-       set_ipv6 = (const struct rte_flow_action_set_ipv6 *)action->conf;
+       set_ipv6 = action->conf;
 
        if (ip_src_flag)
                set_ip->head.jump_id = NFP_FL_ACTION_OPCODE_SET_IPV6_SRC;
@@ -2191,7 +2190,7 @@ nfp_flow_action_set_tp(char *act_data,
        set_tp->head.len_lw  = act_size >> NFP_FL_LW_SIZ;
        set_tp->reserved     = 0;
 
-       set_tp_conf = (const struct rte_flow_action_set_tp *)action->conf;
+       set_tp_conf = action->conf;
        if (tp_src_flag) {
                set_tp->src_port = set_tp_conf->port;
                set_tp->src_port_mask = RTE_BE16(0xffff);
@@ -2223,12 +2222,9 @@ nfp_flow_action_push_vlan(char *act_data,
        push_vlan->head.len_lw  = act_size >> NFP_FL_LW_SIZ;
        push_vlan->reserved     = 0;
 
-       push_vlan_conf = (const struct rte_flow_action_of_push_vlan *)
-                       action->conf;
-       vlan_pcp_conf  = (const struct rte_flow_action_of_set_vlan_pcp *)
-                       (action + 1)->conf;
-       vlan_vid_conf  = (const struct rte_flow_action_of_set_vlan_vid *)
-                       (action + 2)->conf;
+       push_vlan_conf = action->conf;
+       vlan_pcp_conf  = (action + 1)->conf;
+       vlan_vid_conf  = (action + 2)->conf;
 
        vid = rte_be_to_cpu_16(vlan_vid_conf->vlan_vid) & 0x0fff;
        pcp = vlan_pcp_conf->vlan_pcp & 0x07;
@@ -2256,7 +2252,7 @@ nfp_flow_action_set_ttl(char *act_data,
        ttl_tos->head.jump_id = NFP_FL_ACTION_OPCODE_SET_IPV4_TTL_TOS;
        ttl_tos->head.len_lw = act_size >> NFP_FL_LW_SIZ;
 
-       ttl_conf = (const struct rte_flow_action_set_ttl *)action->conf;
+       ttl_conf = action->conf;
        ttl_tos->ipv4_ttl = ttl_conf->ttl_value;
        ttl_tos->ipv4_ttl_mask = 0xff;
        ttl_tos->reserved = 0;
@@ -2280,7 +2276,7 @@ nfp_flow_action_set_hl(char *act_data,
        tc_hl->head.jump_id = NFP_FL_ACTION_OPCODE_SET_IPV6_TC_HL_FL;
        tc_hl->head.len_lw = act_size >> NFP_FL_LW_SIZ;
 
-       ttl_conf = (const struct rte_flow_action_set_ttl *)action->conf;
+       ttl_conf = action->conf;
        tc_hl->ipv6_hop_limit = ttl_conf->ttl_value;
        tc_hl->ipv6_hop_limit_mask = 0xff;
        tc_hl->reserved = 0;
@@ -2304,7 +2300,7 @@ nfp_flow_action_set_tos(char *act_data,
        ttl_tos->head.jump_id = NFP_FL_ACTION_OPCODE_SET_IPV4_TTL_TOS;
        ttl_tos->head.len_lw = act_size >> NFP_FL_LW_SIZ;
 
-       tos_conf = (const struct rte_flow_action_set_dscp *)action->conf;
+       tos_conf = action->conf;
        ttl_tos->ipv4_tos = tos_conf->dscp;
        ttl_tos->ipv4_tos_mask = 0xff;
        ttl_tos->reserved = 0;
@@ -2328,7 +2324,7 @@ nfp_flow_action_set_tc(char *act_data,
        tc_hl->head.jump_id = NFP_FL_ACTION_OPCODE_SET_IPV6_TC_HL_FL;
        tc_hl->head.len_lw = act_size >> NFP_FL_LW_SIZ;
 
-       tos_conf = (const struct rte_flow_action_set_dscp *)action->conf;
+       tos_conf = action->conf;
        tc_hl->ipv6_tc = tos_conf->dscp;
        tc_hl->ipv6_tc_mask = 0xff;
        tc_hl->reserved = 0;
@@ -2719,9 +2715,9 @@ nfp_flow_action_vxlan_encap_v4(struct nfp_app_fw_flower 
*app_fw_flower,
        size_t act_pre_size = sizeof(struct nfp_fl_act_pre_tun);
        size_t act_set_size = sizeof(struct nfp_fl_act_set_tun);
 
-       eth   = (const struct rte_flow_item_eth *)vxlan_data->items[0].spec;
-       ipv4  = (const struct rte_flow_item_ipv4 *)vxlan_data->items[1].spec;
-       vxlan = (const struct rte_flow_item_vxlan *)vxlan_data->items[3].spec;
+       eth   = vxlan_data->items[0].spec;
+       ipv4  = vxlan_data->items[1].spec;
+       vxlan = vxlan_data->items[3].spec;
 
        pre_tun = (struct nfp_fl_act_pre_tun *)actions;
        memset(pre_tun, 0, act_pre_size);
@@ -2756,9 +2752,9 @@ nfp_flow_action_vxlan_encap_v6(struct nfp_app_fw_flower 
*app_fw_flower,
        size_t act_pre_size = sizeof(struct nfp_fl_act_pre_tun);
        size_t act_set_size = sizeof(struct nfp_fl_act_set_tun);
 
-       eth   = (const struct rte_flow_item_eth *)vxlan_data->items[0].spec;
-       ipv6  = (const struct rte_flow_item_ipv6 *)vxlan_data->items[1].spec;
-       vxlan = (const struct rte_flow_item_vxlan *)vxlan_data->items[3].spec;
+       eth   = vxlan_data->items[0].spec;
+       ipv6  = vxlan_data->items[1].spec;
+       vxlan = vxlan_data->items[3].spec;
 
        pre_tun = (struct nfp_fl_act_pre_tun *)actions;
        memset(pre_tun, 0, act_pre_size);
@@ -3626,7 +3622,7 @@ nfp_flow_process(struct nfp_flower_representor 
*representor,
        nfp_flower_update_meta_tci(nfp_flow->payload.unmasked_data, 
new_mask_id);
 
        /* Calculate and store the hash_key for later use */
-       hash_data = (char *)(nfp_flow->payload.unmasked_data);
+       hash_data = nfp_flow->payload.unmasked_data;
        nfp_flow->hash_key = rte_jhash(hash_data, nfp_flow->length, 
priv->hash_seed);
 
        /* Find the flow in hash table */
@@ -3716,7 +3712,7 @@ nfp_flow_validate(struct rte_eth_dev *dev,
        struct nfp_flow_priv *priv;
        struct nfp_flower_representor *representor;
 
-       representor = (struct nfp_flower_representor *)dev->data->dev_private;
+       representor = dev->data->dev_private;
        priv = representor->app_fw_flower->flow_priv;
 
        nfp_flow = nfp_flow_setup(representor, attr, items, actions, error, 
true);
@@ -3751,7 +3747,7 @@ nfp_flow_create(struct rte_eth_dev *dev,
        struct nfp_app_fw_flower *app_fw_flower;
        struct nfp_flower_representor *representor;
 
-       representor = (struct nfp_flower_representor *)dev->data->dev_private;
+       representor = dev->data->dev_private;
        app_fw_flower = representor->app_fw_flower;
        priv = app_fw_flower->flow_priv;
 
@@ -3813,7 +3809,7 @@ nfp_flow_destroy(struct rte_eth_dev *dev,
        struct nfp_app_fw_flower *app_fw_flower;
        struct nfp_flower_representor *representor;
 
-       representor = (struct nfp_flower_representor *)dev->data->dev_private;
+       representor = dev->data->dev_private;
        app_fw_flower = representor->app_fw_flower;
        priv = app_fw_flower->flow_priv;
 
@@ -3949,7 +3945,7 @@ nfp_flow_stats_get(struct rte_eth_dev *dev,
                return;
        }
 
-       query = (struct rte_flow_query_count *)data;
+       query = data;
        reset = query->reset;
        memset(query, 0, sizeof(*query));
 
diff --git a/drivers/net/nfp/nfp_rxtx.c b/drivers/net/nfp/nfp_rxtx.c
index 2409f63205..3c78557221 100644
--- a/drivers/net/nfp/nfp_rxtx.c
+++ b/drivers/net/nfp/nfp_rxtx.c
@@ -630,7 +630,7 @@ nfp_net_rx_queue_setup(struct rte_eth_dev *dev,
 
        /* Saving physical and virtual addresses for the RX ring */
        rxq->dma = (uint64_t)tz->iova;
-       rxq->rxds = (struct nfp_net_rx_desc *)tz->addr;
+       rxq->rxds = tz->addr;
 
        /* mbuf pointers array for referencing mbufs linked to RX descriptors */
        rxq->rxbufs = rte_zmalloc_socket("rxq->rxbufs",
diff --git a/drivers/net/nfp/nfpcore/nfp_cpp_pcie_ops.c 
b/drivers/net/nfp/nfpcore/nfp_cpp_pcie_ops.c
index edf4088747..9d63e0ee73 100644
--- a/drivers/net/nfp/nfpcore/nfp_cpp_pcie_ops.c
+++ b/drivers/net/nfp/nfpcore/nfp_cpp_pcie_ops.c
@@ -835,7 +835,7 @@ nfp6000_init(struct nfp_cpp *cpp, struct rte_pci_device 
*dev)
        if (nfp6000_set_barsz(dev, desc) < 0)
                goto error;
 
-       desc->cfg = (char *)dev->mem_resource[0].addr;
+       desc->cfg = dev->mem_resource[0].addr;
 
        nfp_enable_bars(desc);
 
-- 
2.39.1

Reply via email to