From: Aaron Williams <awilli...@marvell.com>

Import cvmx-ipd.c from 2013 U-Boot. It will be used by the later
added drivers to support networking on the MIPS Octeon II / III
platforms.

Signed-off-by: Aaron Williams <awilli...@marvell.com>
Signed-off-by: Stefan Roese <s...@denx.de>
---
 arch/mips/mach-octeon/cvmx-ipd.c | 690 +++++++++++++++++++++++++++++++
 1 file changed, 690 insertions(+)
 create mode 100644 arch/mips/mach-octeon/cvmx-ipd.c

diff --git a/arch/mips/mach-octeon/cvmx-ipd.c b/arch/mips/mach-octeon/cvmx-ipd.c
new file mode 100644
index 000000000000..0b5aca56e399
--- /dev/null
+++ b/arch/mips/mach-octeon/cvmx-ipd.c
@@ -0,0 +1,690 @@
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * Copyright (C) 2018-2022 Marvell International Ltd.
+ *
+ * IPD Support.
+ */
+
+#include <errno.h>
+#include <log.h>
+#include <time.h>
+#include <linux/delay.h>
+
+#include <mach/cvmx-regs.h>
+#include <mach/cvmx-csr.h>
+#include <mach/cvmx-bootmem.h>
+#include <mach/octeon-model.h>
+#include <mach/cvmx-fuse.h>
+#include <mach/octeon-feature.h>
+#include <mach/cvmx-qlm.h>
+#include <mach/octeon_qlm.h>
+#include <mach/cvmx-pcie.h>
+#include <mach/cvmx-coremask.h>
+#include <mach/cvmx-range.h>
+#include <mach/cvmx-global-resources.h>
+
+#include <mach/cvmx-agl-defs.h>
+#include <mach/cvmx-bgxx-defs.h>
+#include <mach/cvmx-ciu-defs.h>
+#include <mach/cvmx-gmxx-defs.h>
+#include <mach/cvmx-gserx-defs.h>
+#include <mach/cvmx-ilk-defs.h>
+#include <mach/cvmx-ipd-defs.h>
+#include <mach/cvmx-pcsx-defs.h>
+#include <mach/cvmx-pcsxx-defs.h>
+#include <mach/cvmx-pki-defs.h>
+#include <mach/cvmx-pko-defs.h>
+#include <mach/cvmx-xcv-defs.h>
+
+#include <mach/cvmx-hwpko.h>
+#include <mach/cvmx-ilk.h>
+#include <mach/cvmx-ipd.h>
+#include <mach/cvmx-pki.h>
+#include <mach/cvmx-pko3.h>
+#include <mach/cvmx-pko3-queue.h>
+#include <mach/cvmx-pko3-resources.h>
+
+#include <mach/cvmx-helper.h>
+#include <mach/cvmx-helper-board.h>
+#include <mach/cvmx-helper-cfg.h>
+
+#include <mach/cvmx-helper-bgx.h>
+#include <mach/cvmx-helper-cfg.h>
+#include <mach/cvmx-helper-util.h>
+#include <mach/cvmx-helper-pki.h>
+
+cvmx_ipd_config_t cvmx_ipd_cfg = {
+       .first_mbuf_skip = 184,
+       .ipd_enable = 1,
+       .cache_mode = CVMX_IPD_OPC_MODE_STT,
+       .packet_pool = { 0, 2048, 0 },
+       .wqe_pool = { 1, 128, 0 },
+       .port_config = { CVMX_PIP_PORT_CFG_MODE_SKIPL2,
+                        CVMX_POW_TAG_TYPE_ORDERED, CVMX_PIP_TAG_MODE_TUPLE,
+                        .tag_fields = { 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1 } }
+};
+
+#define IPD_RED_AVG_DLY 1000
+#define IPD_RED_PRB_DLY 1000
+
+void cvmx_ipd_convert_to_newcfg(cvmx_ipd_config_t ipd_config)
+{
+       int pkind;
+       unsigned int node = cvmx_get_node_num();
+
+       /*Set all the styles to same parameters since old config does not have 
per port config*/
+       pki_dflt_style[node].parm_cfg.cache_mode = ipd_config.cache_mode;
+       pki_dflt_style[node].parm_cfg.first_skip = ipd_config.first_mbuf_skip;
+       pki_dflt_style[node].parm_cfg.later_skip =
+               ipd_config.not_first_mbuf_skip;
+       pki_dflt_style[node].parm_cfg.mbuff_size =
+               ipd_config.packet_pool.buffer_size;
+       pki_dflt_style[node].parm_cfg.tag_type =
+               ipd_config.port_config.tag_type;
+
+       pki_dflt_style[node].tag_cfg.tag_fields.layer_c_src =
+               ipd_config.port_config.tag_fields.ipv6_src_ip |
+               ipd_config.port_config.tag_fields.ipv4_src_ip;
+       pki_dflt_style[node].tag_cfg.tag_fields.layer_c_dst =
+               ipd_config.port_config.tag_fields.ipv6_dst_ip |
+               ipd_config.port_config.tag_fields.ipv4_dst_ip;
+       pki_dflt_style[node].tag_cfg.tag_fields.ip_prot_nexthdr =
+               ipd_config.port_config.tag_fields.ipv6_next_header |
+               ipd_config.port_config.tag_fields.ipv4_protocol;
+       pki_dflt_style[node].tag_cfg.tag_fields.layer_f_src =
+               ipd_config.port_config.tag_fields.ipv6_src_port |
+               ipd_config.port_config.tag_fields.ipv4_src_port;
+       pki_dflt_style[node].tag_cfg.tag_fields.layer_f_dst =
+               ipd_config.port_config.tag_fields.ipv6_dst_port |
+               ipd_config.port_config.tag_fields.ipv4_dst_port;
+       pki_dflt_style[node].tag_cfg.tag_fields.input_port =
+               ipd_config.port_config.tag_fields.input_port;
+
+       if (ipd_config.port_config.parse_mode == 0x1)
+               pki_dflt_pkind[node].initial_parse_mode =
+                       CVMX_PKI_PARSE_LA_TO_LG;
+       else if (ipd_config.port_config.parse_mode == 0x2)
+               pki_dflt_pkind[node].initial_parse_mode =
+                       CVMX_PKI_PARSE_LC_TO_LG;
+       else
+               pki_dflt_pkind[node].initial_parse_mode =
+                       CVMX_PKI_PARSE_NOTHING;
+
+       /* For compatibility make style = pkind so old software can modify 
style */
+       for (pkind = 0; pkind < CVMX_PKI_NUM_PKIND; pkind++)
+               pkind_style_map[node][pkind] = pkind;
+       /*setup packet pool*/
+       cvmx_helper_pki_set_dflt_pool(node, ipd_config.packet_pool.pool_num,
+                                     ipd_config.packet_pool.buffer_size,
+                                     ipd_config.packet_pool.buffer_count);
+       cvmx_helper_pki_set_dflt_aura(node, ipd_config.packet_pool.pool_num,
+                                     ipd_config.packet_pool.pool_num,
+                                     ipd_config.packet_pool.buffer_count);
+}
+
+int cvmx_ipd_set_config(cvmx_ipd_config_t ipd_config)
+{
+       cvmx_ipd_cfg = ipd_config;
+       if (octeon_has_feature(OCTEON_FEATURE_PKI))
+               cvmx_ipd_convert_to_newcfg(ipd_config);
+       return 0;
+}
+
+void cvmx_ipd_get_config(cvmx_ipd_config_t *ipd_config)
+{
+       *ipd_config = cvmx_ipd_cfg;
+}
+
+void cvmx_ipd_set_packet_pool_buffer_count(u64 buffer_count)
+{
+       cvmx_ipd_cfg.packet_pool.buffer_count = buffer_count;
+}
+
+void cvmx_ipd_set_packet_pool_config(s64 pool, u64 buffer_size,
+                                    u64 buffer_count)
+{
+       cvmx_ipd_cfg.packet_pool.pool_num = pool;
+       cvmx_ipd_cfg.packet_pool.buffer_size = buffer_size;
+       cvmx_ipd_cfg.packet_pool.buffer_count = buffer_count;
+       if (octeon_has_feature(OCTEON_FEATURE_PKI)) {
+               int node = cvmx_get_node_num();
+               s64 aura = pool;
+
+               cvmx_helper_pki_set_dflt_pool(node, pool, buffer_size,
+                                             buffer_count);
+               cvmx_helper_pki_set_dflt_aura(node, aura, pool, buffer_count);
+       }
+}
+
+void cvmx_ipd_set_wqe_pool_buffer_count(u64 buffer_count)
+{
+       cvmx_ipd_cfg.wqe_pool.buffer_count = buffer_count;
+}
+
+void cvmx_ipd_set_wqe_pool_config(s64 pool, u64 buffer_size, u64 buffer_count)
+{
+       cvmx_ipd_cfg.wqe_pool.pool_num = pool;
+       cvmx_ipd_cfg.wqe_pool.buffer_size = buffer_size;
+       cvmx_ipd_cfg.wqe_pool.buffer_count = buffer_count;
+}
+
+static void __cvmx_ipd_free_ptr_v1(void)
+{
+       unsigned int wqe_pool = cvmx_fpa_get_wqe_pool();
+       int i;
+       union cvmx_ipd_ptr_count ptr_count;
+       union cvmx_ipd_prc_port_ptr_fifo_ctl prc_port_fifo;
+       int packet_pool = (int)cvmx_fpa_get_packet_pool();
+
+       ptr_count.u64 = csr_rd(CVMX_IPD_PTR_COUNT);
+
+       /* Handle Work Queue Entry in cn56xx and cn52xx */
+       if (octeon_has_feature(OCTEON_FEATURE_NO_WPTR)) {
+               union cvmx_ipd_ctl_status ctl_status;
+
+               ctl_status.u64 = csr_rd(CVMX_IPD_CTL_STATUS);
+               if (ctl_status.s.no_wptr)
+                       wqe_pool = packet_pool;
+       }
+
+       /* Free the prefetched WQE */
+       if (ptr_count.s.wqev_cnt) {
+               union cvmx_ipd_wqe_ptr_valid wqe_ptr_valid;
+
+               wqe_ptr_valid.u64 = csr_rd(CVMX_IPD_WQE_PTR_VALID);
+               cvmx_fpa1_free(cvmx_phys_to_ptr((u64)wqe_ptr_valid.s.ptr << 7),
+                              wqe_pool, 0);
+       }
+
+       /* Free all WQE in the fifo */
+       if (ptr_count.s.wqe_pcnt) {
+               int i;
+               union cvmx_ipd_pwp_ptr_fifo_ctl pwp_fifo;
+
+               pwp_fifo.u64 = csr_rd(CVMX_IPD_PWP_PTR_FIFO_CTL);
+               for (i = 0; i < ptr_count.s.wqe_pcnt; i++) {
+                       pwp_fifo.s.cena = 0;
+                       pwp_fifo.s.raddr =
+                               pwp_fifo.s.max_cnts +
+                               (pwp_fifo.s.wraddr + i) % pwp_fifo.s.max_cnts;
+                       csr_wr(CVMX_IPD_PWP_PTR_FIFO_CTL, pwp_fifo.u64);
+                       pwp_fifo.u64 = csr_rd(CVMX_IPD_PWP_PTR_FIFO_CTL);
+                       cvmx_fpa1_free(cvmx_phys_to_ptr((u64)pwp_fifo.s.ptr
+                                                       << 7),
+                                      wqe_pool, 0);
+               }
+               pwp_fifo.s.cena = 1;
+               csr_wr(CVMX_IPD_PWP_PTR_FIFO_CTL, pwp_fifo.u64);
+       }
+
+       /* Free the prefetched packet */
+       if (ptr_count.s.pktv_cnt) {
+               union cvmx_ipd_pkt_ptr_valid pkt_ptr_valid;
+
+               pkt_ptr_valid.u64 = csr_rd(CVMX_IPD_PKT_PTR_VALID);
+               cvmx_fpa1_free(cvmx_phys_to_ptr((u64)pkt_ptr_valid.s.ptr << 7),
+                              packet_pool, 0);
+       }
+
+       /* Free the per port prefetched packets */
+       prc_port_fifo.u64 = csr_rd(CVMX_IPD_PRC_PORT_PTR_FIFO_CTL);
+
+       for (i = 0; i < prc_port_fifo.s.max_pkt; i++) {
+               prc_port_fifo.s.cena = 0;
+               prc_port_fifo.s.raddr = i % prc_port_fifo.s.max_pkt;
+               csr_wr(CVMX_IPD_PRC_PORT_PTR_FIFO_CTL, prc_port_fifo.u64);
+               prc_port_fifo.u64 = csr_rd(CVMX_IPD_PRC_PORT_PTR_FIFO_CTL);
+               cvmx_fpa1_free(cvmx_phys_to_ptr((u64)prc_port_fifo.s.ptr << 7),
+                              packet_pool, 0);
+       }
+       prc_port_fifo.s.cena = 1;
+       csr_wr(CVMX_IPD_PRC_PORT_PTR_FIFO_CTL, prc_port_fifo.u64);
+
+       /* Free all packets in the holding fifo */
+       if (ptr_count.s.pfif_cnt) {
+               int i;
+               union cvmx_ipd_prc_hold_ptr_fifo_ctl prc_hold_fifo;
+
+               prc_hold_fifo.u64 = csr_rd(CVMX_IPD_PRC_HOLD_PTR_FIFO_CTL);
+
+               for (i = 0; i < ptr_count.s.pfif_cnt; i++) {
+                       prc_hold_fifo.s.cena = 0;
+                       prc_hold_fifo.s.raddr = (prc_hold_fifo.s.praddr + i) %
+                                               prc_hold_fifo.s.max_pkt;
+                       csr_wr(CVMX_IPD_PRC_HOLD_PTR_FIFO_CTL,
+                              prc_hold_fifo.u64);
+                       prc_hold_fifo.u64 =
+                               csr_rd(CVMX_IPD_PRC_HOLD_PTR_FIFO_CTL);
+                       cvmx_fpa1_free(cvmx_phys_to_ptr((u64)prc_hold_fifo.s.ptr
+                                                       << 7),
+                                      packet_pool, 0);
+               }
+               prc_hold_fifo.s.cena = 1;
+               csr_wr(CVMX_IPD_PRC_HOLD_PTR_FIFO_CTL, prc_hold_fifo.u64);
+       }
+
+       /* Free all packets in the fifo */
+       if (ptr_count.s.pkt_pcnt) {
+               int i;
+               union cvmx_ipd_pwp_ptr_fifo_ctl pwp_fifo;
+
+               pwp_fifo.u64 = csr_rd(CVMX_IPD_PWP_PTR_FIFO_CTL);
+
+               for (i = 0; i < ptr_count.s.pkt_pcnt; i++) {
+                       pwp_fifo.s.cena = 0;
+                       pwp_fifo.s.raddr =
+                               (pwp_fifo.s.praddr + i) % pwp_fifo.s.max_cnts;
+                       csr_wr(CVMX_IPD_PWP_PTR_FIFO_CTL, pwp_fifo.u64);
+                       pwp_fifo.u64 = csr_rd(CVMX_IPD_PWP_PTR_FIFO_CTL);
+                       cvmx_fpa1_free(cvmx_phys_to_ptr((u64)pwp_fifo.s.ptr
+                                                       << 7),
+                                      packet_pool, 0);
+               }
+               pwp_fifo.s.cena = 1;
+               csr_wr(CVMX_IPD_PWP_PTR_FIFO_CTL, pwp_fifo.u64);
+       }
+}
+
+static void __cvmx_ipd_free_ptr_v2(void)
+{
+       int no_wptr = 0;
+       int i;
+       union cvmx_ipd_port_ptr_fifo_ctl port_ptr_fifo;
+       union cvmx_ipd_ptr_count ptr_count;
+       int packet_pool = (int)cvmx_fpa_get_packet_pool();
+       int wqe_pool = (int)cvmx_fpa_get_wqe_pool();
+
+       ptr_count.u64 = csr_rd(CVMX_IPD_PTR_COUNT);
+
+       /* Handle Work Queue Entry in cn68xx */
+       if (octeon_has_feature(OCTEON_FEATURE_NO_WPTR)) {
+               union cvmx_ipd_ctl_status ctl_status;
+
+               ctl_status.u64 = csr_rd(CVMX_IPD_CTL_STATUS);
+               if (ctl_status.s.no_wptr)
+                       no_wptr = 1;
+       }
+
+       /* Free the prefetched WQE */
+       if (ptr_count.s.wqev_cnt) {
+               union cvmx_ipd_next_wqe_ptr next_wqe_ptr;
+
+               next_wqe_ptr.u64 = csr_rd(CVMX_IPD_NEXT_WQE_PTR);
+               if (no_wptr)
+                       cvmx_fpa1_free(cvmx_phys_to_ptr((u64)next_wqe_ptr.s.ptr
+                                                       << 7),
+                                      packet_pool, 0);
+               else
+                       cvmx_fpa1_free(cvmx_phys_to_ptr((u64)next_wqe_ptr.s.ptr
+                                                       << 7),
+                                      wqe_pool, 0);
+       }
+
+       /* Free all WQE in the fifo */
+       if (ptr_count.s.wqe_pcnt) {
+               union cvmx_ipd_free_ptr_fifo_ctl free_fifo;
+               union cvmx_ipd_free_ptr_value free_ptr_value;
+
+               free_fifo.u64 = csr_rd(CVMX_IPD_FREE_PTR_FIFO_CTL);
+               for (i = 0; i < ptr_count.s.wqe_pcnt; i++) {
+                       free_fifo.s.cena = 0;
+                       free_fifo.s.raddr =
+                               free_fifo.s.max_cnts +
+                               (free_fifo.s.wraddr + i) % free_fifo.s.max_cnts;
+                       csr_wr(CVMX_IPD_FREE_PTR_FIFO_CTL, free_fifo.u64);
+                       free_fifo.u64 = csr_rd(CVMX_IPD_FREE_PTR_FIFO_CTL);
+                       free_ptr_value.u64 = csr_rd(CVMX_IPD_FREE_PTR_VALUE);
+                       if (no_wptr)
+                               cvmx_fpa1_free(cvmx_phys_to_ptr(
+                                                      (u64)free_ptr_value.s.ptr
+                                                      << 7),
+                                              packet_pool, 0);
+                       else
+                               cvmx_fpa1_free(cvmx_phys_to_ptr(
+                                                      (u64)free_ptr_value.s.ptr
+                                                      << 7),
+                                              wqe_pool, 0);
+               }
+               free_fifo.s.cena = 1;
+               csr_wr(CVMX_IPD_FREE_PTR_FIFO_CTL, free_fifo.u64);
+       }
+
+       /* Free the prefetched packet */
+       if (ptr_count.s.pktv_cnt) {
+               union cvmx_ipd_next_pkt_ptr next_pkt_ptr;
+
+               next_pkt_ptr.u64 = csr_rd(CVMX_IPD_NEXT_PKT_PTR);
+               cvmx_fpa1_free(cvmx_phys_to_ptr((u64)next_pkt_ptr.s.ptr << 7),
+                              packet_pool, 0);
+       }
+
+       /* Free the per port prefetched packets */
+       port_ptr_fifo.u64 = csr_rd(CVMX_IPD_PORT_PTR_FIFO_CTL);
+
+       for (i = 0; i < port_ptr_fifo.s.max_pkt; i++) {
+               port_ptr_fifo.s.cena = 0;
+               port_ptr_fifo.s.raddr = i % port_ptr_fifo.s.max_pkt;
+               csr_wr(CVMX_IPD_PORT_PTR_FIFO_CTL, port_ptr_fifo.u64);
+               port_ptr_fifo.u64 = csr_rd(CVMX_IPD_PORT_PTR_FIFO_CTL);
+               cvmx_fpa1_free(cvmx_phys_to_ptr((u64)port_ptr_fifo.s.ptr << 7),
+                              packet_pool, 0);
+       }
+       port_ptr_fifo.s.cena = 1;
+       csr_wr(CVMX_IPD_PORT_PTR_FIFO_CTL, port_ptr_fifo.u64);
+
+       /* Free all packets in the holding fifo */
+       if (ptr_count.s.pfif_cnt) {
+               union cvmx_ipd_hold_ptr_fifo_ctl hold_ptr_fifo;
+
+               hold_ptr_fifo.u64 = csr_rd(CVMX_IPD_HOLD_PTR_FIFO_CTL);
+
+               for (i = 0; i < ptr_count.s.pfif_cnt; i++) {
+                       hold_ptr_fifo.s.cena = 0;
+                       hold_ptr_fifo.s.raddr = (hold_ptr_fifo.s.praddr + i) %
+                                               hold_ptr_fifo.s.max_pkt;
+                       csr_wr(CVMX_IPD_HOLD_PTR_FIFO_CTL, hold_ptr_fifo.u64);
+                       hold_ptr_fifo.u64 = csr_rd(CVMX_IPD_HOLD_PTR_FIFO_CTL);
+                       cvmx_fpa1_free(cvmx_phys_to_ptr((u64)hold_ptr_fifo.s.ptr
+                                                       << 7),
+                                      packet_pool, 0);
+               }
+               hold_ptr_fifo.s.cena = 1;
+               csr_wr(CVMX_IPD_HOLD_PTR_FIFO_CTL, hold_ptr_fifo.u64);
+       }
+
+       /* Free all packets in the fifo */
+       if (ptr_count.s.pkt_pcnt) {
+               union cvmx_ipd_free_ptr_fifo_ctl free_fifo;
+               union cvmx_ipd_free_ptr_value free_ptr_value;
+
+               free_fifo.u64 = csr_rd(CVMX_IPD_FREE_PTR_FIFO_CTL);
+
+               for (i = 0; i < ptr_count.s.pkt_pcnt; i++) {
+                       free_fifo.s.cena = 0;
+                       free_fifo.s.raddr =
+                               (free_fifo.s.praddr + i) % free_fifo.s.max_cnts;
+                       csr_wr(CVMX_IPD_FREE_PTR_FIFO_CTL, free_fifo.u64);
+                       free_fifo.u64 = csr_rd(CVMX_IPD_FREE_PTR_FIFO_CTL);
+                       free_ptr_value.u64 = csr_rd(CVMX_IPD_FREE_PTR_VALUE);
+                       cvmx_fpa1_free(cvmx_phys_to_ptr(
+                                              (u64)free_ptr_value.s.ptr << 7),
+                                      packet_pool, 0);
+               }
+               free_fifo.s.cena = 1;
+               csr_wr(CVMX_IPD_FREE_PTR_FIFO_CTL, free_fifo.u64);
+       }
+}
+
+/**
+ * @INTERNAL
+ * This function is called by cvmx_helper_shutdown() to extract
+ * all FPA buffers out of the IPD and PIP. After this function
+ * completes, all FPA buffers that were prefetched by IPD and PIP
+ * wil be in the appropriate FPA pool. This functions does not reset
+ * PIP or IPD as FPA pool zero must be empty before the reset can
+ * be performed. WARNING: It is very important that IPD and PIP be
+ * reset soon after a call to this function.
+ */
+void __cvmx_ipd_free_ptr(void)
+{
+       if (octeon_has_feature(OCTEON_FEATURE_PKND))
+               __cvmx_ipd_free_ptr_v2();
+       else
+               __cvmx_ipd_free_ptr_v1();
+}
+
+void cvmx_ipd_config(u64 mbuff_size, u64 first_mbuff_skip,
+                    u64 not_first_mbuff_skip, u64 first_back, u64 second_back,
+                    u64 wqe_fpa_pool, cvmx_ipd_mode_t cache_mode,
+                    u64 back_pres_enable_flag)
+{
+       cvmx_ipd_1st_mbuff_skip_t first_skip;
+       cvmx_ipd_mbuff_not_first_skip_t not_first_skip;
+       cvmx_ipd_packet_mbuff_size_t size;
+       cvmx_ipd_1st_next_ptr_back_t first_back_struct;
+       cvmx_ipd_second_next_ptr_back_t second_back_struct;
+       cvmx_ipd_wqe_fpa_queue_t wqe_pool;
+       cvmx_ipd_ctl_status_t ipd_ctl_reg;
+
+       /* Enforce 1st skip minimum if WQE shares the buffer with packet */
+       if (octeon_has_feature(OCTEON_FEATURE_NO_WPTR)) {
+               union cvmx_ipd_ctl_status ctl_status;
+
+               ctl_status.u64 = csr_rd(CVMX_IPD_CTL_STATUS);
+               if (ctl_status.s.no_wptr != 0 && first_mbuff_skip < 16)
+                       first_mbuff_skip = 16;
+       }
+
+       first_skip.u64 = 0;
+       first_skip.s.skip_sz = first_mbuff_skip;
+       csr_wr(CVMX_IPD_1ST_MBUFF_SKIP, first_skip.u64);
+
+       not_first_skip.u64 = 0;
+       not_first_skip.s.skip_sz = not_first_mbuff_skip;
+       csr_wr(CVMX_IPD_NOT_1ST_MBUFF_SKIP, not_first_skip.u64);
+
+       size.u64 = 0;
+       size.s.mb_size = mbuff_size;
+       csr_wr(CVMX_IPD_PACKET_MBUFF_SIZE, size.u64);
+
+       first_back_struct.u64 = 0;
+       first_back_struct.s.back = first_back;
+       csr_wr(CVMX_IPD_1st_NEXT_PTR_BACK, first_back_struct.u64);
+
+       second_back_struct.u64 = 0;
+       second_back_struct.s.back = second_back;
+       csr_wr(CVMX_IPD_2nd_NEXT_PTR_BACK, second_back_struct.u64);
+
+       wqe_pool.u64 = 0;
+       wqe_pool.s.wqe_pool = wqe_fpa_pool;
+       csr_wr(CVMX_IPD_WQE_FPA_QUEUE, wqe_pool.u64);
+
+       ipd_ctl_reg.u64 = csr_rd(CVMX_IPD_CTL_STATUS);
+       ipd_ctl_reg.s.opc_mode = cache_mode;
+       ipd_ctl_reg.s.pbp_en = back_pres_enable_flag;
+       csr_wr(CVMX_IPD_CTL_STATUS, ipd_ctl_reg.u64);
+
+       /* Note: the example RED code is below */
+}
+
+/**
+ * Setup Random Early Drop on a specific input queue
+ *
+ * @param queue  Input queue to setup RED on (0-7)
+ * @param pass_thresh
+ *               Packets will begin slowly dropping when there are less than
+ *               this many packet buffers free in FPA 0.
+ * @param drop_thresh
+ *               All incoming packets will be dropped when there are less
+ *               than this many free packet buffers in FPA 0.
+ * @return Zero on success. Negative on failure
+ */
+int cvmx_ipd_setup_red_queue(int queue, int pass_thresh, int drop_thresh)
+{
+       union cvmx_ipd_qosx_red_marks red_marks;
+       union cvmx_ipd_red_quex_param red_param;
+
+       /*
+        * Set RED to begin dropping packets when there are
+        * pass_thresh buffers left. It will linearly drop more
+        * packets until reaching drop_thresh buffers.
+        */
+       red_marks.u64 = 0;
+       red_marks.s.drop = drop_thresh;
+       red_marks.s.pass = pass_thresh;
+       csr_wr(CVMX_IPD_QOSX_RED_MARKS(queue), red_marks.u64);
+
+       /* Use the actual queue 0 counter, not the average */
+       red_param.u64 = 0;
+       red_param.s.prb_con =
+               (255ul << 24) / (red_marks.s.pass - red_marks.s.drop);
+       red_param.s.avg_con = 1;
+       red_param.s.new_con = 255;
+       red_param.s.use_pcnt = 1;
+       csr_wr(CVMX_IPD_RED_QUEX_PARAM(queue), red_param.u64);
+       return 0;
+}
+
+/**
+ * Setup Random Early Drop to automatically begin dropping packets.
+ *
+ * @param pass_thresh
+ *               Packets will begin slowly dropping when there are less than
+ *               this many packet buffers free in FPA 0.
+ * @param drop_thresh
+ *               All incoming packets will be dropped when there are less
+ *               than this many free packet buffers in FPA 0.
+ * @return Zero on success. Negative on failure
+ */
+int cvmx_ipd_setup_red(int pass_thresh, int drop_thresh)
+{
+       int queue;
+       int interface;
+       int port;
+
+       if (octeon_has_feature(OCTEON_FEATURE_PKI))
+               return -1;
+       /*
+        * Disable backpressure based on queued buffers. It needs SW support
+        */
+       if (octeon_has_feature(OCTEON_FEATURE_PKND)) {
+               int bpid;
+
+               for (interface = 0; interface < CVMX_HELPER_MAX_GMX;
+                    interface++) {
+                       int num_ports;
+
+                       num_ports = cvmx_helper_ports_on_interface(interface);
+                       for (port = 0; port < num_ports; port++) {
+                               bpid = cvmx_helper_get_bpid(interface, port);
+                               if (bpid == CVMX_INVALID_BPID)
+                                       debug("setup_red: 
cvmx_helper_get_bpid(%d, %d) = %d\n",
+                                             interface, port,
+                                             cvmx_helper_get_bpid(interface,
+                                                                  port));
+                               else
+                                       csr_wr(CVMX_IPD_BPIDX_MBUF_TH(bpid), 0);
+                       }
+               }
+       } else {
+               union cvmx_ipd_portx_bp_page_cnt page_cnt;
+
+               page_cnt.u64 = 0;
+               page_cnt.s.bp_enb = 0;
+               page_cnt.s.page_cnt = 100;
+               for (interface = 0; interface < CVMX_HELPER_MAX_GMX;
+                    interface++) {
+                       for (port = cvmx_helper_get_first_ipd_port(interface);
+                            port < cvmx_helper_get_last_ipd_port(interface);
+                            port++)
+                               csr_wr(CVMX_IPD_PORTX_BP_PAGE_CNT(port),
+                                      page_cnt.u64);
+               }
+       }
+
+       for (queue = 0; queue < 8; queue++)
+               cvmx_ipd_setup_red_queue(queue, pass_thresh, drop_thresh);
+
+       /*
+        * Shutoff the dropping based on the per port page count. SW isn't
+        * decrementing it right now
+        */
+       if (octeon_has_feature(OCTEON_FEATURE_PKND))
+               csr_wr(CVMX_IPD_ON_BP_DROP_PKTX(0), 0);
+       else
+               csr_wr(CVMX_IPD_BP_PRT_RED_END, 0);
+
+       /*
+        * Setting up avg_dly and prb_dly, enable bits
+        */
+       if (octeon_has_feature(OCTEON_FEATURE_PKND)) {
+               union cvmx_ipd_red_delay red_delay;
+               union cvmx_ipd_red_bpid_enablex red_bpid_enable;
+
+               red_delay.u64 = 0;
+               red_delay.s.avg_dly = IPD_RED_AVG_DLY;
+               red_delay.s.prb_dly = IPD_RED_PRB_DLY;
+               csr_wr(CVMX_IPD_RED_DELAY, red_delay.u64);
+
+               /*
+                * Only enable the gmx ports
+                */
+               red_bpid_enable.u64 = 0;
+               for (interface = 0; interface < CVMX_HELPER_MAX_GMX;
+                    interface++) {
+                       int num_ports =
+                               cvmx_helper_ports_on_interface(interface);
+                       for (port = 0; port < num_ports; port++)
+                               red_bpid_enable.u64 |=
+                                       (((u64)1) << cvmx_helper_get_bpid(
+                                                interface, port));
+               }
+               csr_wr(CVMX_IPD_RED_BPID_ENABLEX(0), red_bpid_enable.u64);
+       } else {
+               union cvmx_ipd_red_port_enable red_port_enable;
+
+               red_port_enable.u64 = 0;
+               red_port_enable.s.prt_enb = 0xfffffffffull;
+               red_port_enable.s.avg_dly = IPD_RED_AVG_DLY;
+               red_port_enable.s.prb_dly = IPD_RED_PRB_DLY;
+               csr_wr(CVMX_IPD_RED_PORT_ENABLE, red_port_enable.u64);
+
+               /*
+                * Shutoff the dropping of packets based on RED for SRIO ports
+                */
+               if (octeon_has_feature(OCTEON_FEATURE_SRIO)) {
+                       union cvmx_ipd_red_port_enable2 red_port_enable2;
+
+                       red_port_enable2.u64 = 0;
+                       red_port_enable2.s.prt_enb = 0xf0;
+                       csr_wr(CVMX_IPD_RED_PORT_ENABLE2, red_port_enable2.u64);
+               }
+       }
+
+       return 0;
+}
+
+/**
+ * Enable IPD
+ */
+void cvmx_ipd_enable(void)
+{
+       cvmx_ipd_ctl_status_t ipd_reg;
+
+       ipd_reg.u64 = csr_rd(CVMX_IPD_CTL_STATUS);
+
+       /*
+        * busy-waiting for rst_done in o68
+        */
+       if (OCTEON_IS_MODEL(OCTEON_CN68XX))
+               while (ipd_reg.s.rst_done != 0)
+                       ipd_reg.u64 = csr_rd(CVMX_IPD_CTL_STATUS);
+
+       if (ipd_reg.s.ipd_en)
+               debug("Warning: Enabling IPD when IPD already enabled.\n");
+
+       ipd_reg.s.ipd_en = 1;
+
+       if (cvmx_ipd_cfg.enable_len_M8_fix)
+               ipd_reg.s.len_m8 = 1;
+
+       csr_wr(CVMX_IPD_CTL_STATUS, ipd_reg.u64);
+}
+
+/**
+ * Disable IPD
+ */
+void cvmx_ipd_disable(void)
+{
+       cvmx_ipd_ctl_status_t ipd_reg;
+
+       if (octeon_has_feature(OCTEON_FEATURE_PKI)) {
+               unsigned int node = cvmx_get_node_num();
+
+               cvmx_pki_disable(node);
+               return;
+       }
+       ipd_reg.u64 = csr_rd(CVMX_IPD_CTL_STATUS);
+       ipd_reg.s.ipd_en = 0;
+       csr_wr(CVMX_IPD_CTL_STATUS, ipd_reg.u64);
+}
-- 
2.35.1

Reply via email to