From: Mohsin Bashir <mohsin.bashr@gmail.com>
To: "illusion.wang" <illusion.wang@nebula-matrix.com>,
dimon.zhao@nebula-matrix.com, alvin.wang@nebula-matrix.com,
sam.chen@nebula-matrix.com, netdev@vger.kernel.org
Cc: andrew+netdev@lunn.ch, corbet@lwn.net, kuba@kernel.org,
linux-doc@vger.kernel.org, lorenzo@kernel.org, pabeni@redhat.com,
horms@kernel.org, vadim.fedorenko@linux.dev,
lukas.bulwahn@redhat.com, edumazet@google.com,
enelsonmoore@gmail.com, skhan@linuxfoundation.org,
hkallweit1@gmail.com, jani.nikula@intel.com,
open list <linux-kernel@vger.kernel.org>
Subject: Re: [PATCH v10 net-next 08/11] net/nebula-matrix: add vsi resource implementation
Date: Thu, 2 Apr 2026 02:06:11 -0700 [thread overview]
Message-ID: <db9e681f-3a87-4ea3-a8b3-cfa96e328c42@gmail.com> (raw)
In-Reply-To: <20260401022318.28550-9-illusion.wang@nebula-matrix.com>
On 3/31/26 7:23 PM, illusion.wang wrote:
> The HW (Hardware) layer code can have a quick review since it is
> highly chip-specific.
> Chip initialization includes the initialization of the DP module, the
> intf module, and the P4 registers.
> The initialization of the DP module encompasses the initialization of
> the dped(downstream pkt edit), uped(upstream pkt edit), dsch(downstream
> schedule), ustore, dstore, dvn, uvn, and uqm modules.
>
> Signed-off-by: illusion.wang <illusion.wang@nebula-matrix.com>
> ---
> .../net/ethernet/nebula-matrix/nbl/Makefile | 1 +
> .../nbl/nbl_hw/nbl_hw_leonis/nbl_hw_leonis.c | 406 ++++++++++++++++++
> .../nbl/nbl_hw/nbl_hw_leonis/nbl_hw_leonis.h | 12 +-
> .../nbl_hw_leonis/nbl_resource_leonis.c | 4 +
> .../nebula-matrix/nbl/nbl_hw/nbl_resource.h | 1 +
> .../nebula-matrix/nbl/nbl_hw/nbl_vsi.c | 51 +++
> .../nebula-matrix/nbl/nbl_hw/nbl_vsi.h | 11 +
> .../nbl/nbl_include/nbl_def_hw.h | 4 +
> .../nbl/nbl_include/nbl_include.h | 31 ++
> 9 files changed, 520 insertions(+), 1 deletion(-)
> create mode 100644 drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_vsi.c
> create mode 100644 drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_vsi.h
>
> diff --git a/drivers/net/ethernet/nebula-matrix/nbl/Makefile b/drivers/net/ethernet/nebula-matrix/nbl/Makefile
> index a56e722a5ac7..241bbb572b5e 100644
> --- a/drivers/net/ethernet/nebula-matrix/nbl/Makefile
> +++ b/drivers/net/ethernet/nebula-matrix/nbl/Makefile
> @@ -10,6 +10,7 @@ nbl-objs += nbl_common/nbl_common.o \
> nbl_hw/nbl_hw_leonis/nbl_hw_leonis_regs.o \
> nbl_hw/nbl_resource.o \
> nbl_hw/nbl_interrupt.o \
> + nbl_hw/nbl_vsi.o \
> nbl_core/nbl_dispatch.o \
> nbl_core/nbl_dev.o \
> nbl_main.o
> diff --git a/drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_hw_leonis/nbl_hw_leonis.c b/drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_hw_leonis/nbl_hw_leonis.c
> index aa5e91c2b278..37cfd322d422 100644
> --- a/drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_hw_leonis/nbl_hw_leonis.c
> +++ b/drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_hw_leonis/nbl_hw_leonis.c
> @@ -8,6 +8,7 @@
> #include <linux/io.h>
> #include <linux/spinlock.h>
> #include "nbl_hw_leonis.h"
> +#include "nbl_hw_leonis_regs.h"
>
> static void nbl_hw_read_mbx_regs(struct nbl_hw_mgt *hw_mgt, u64 reg,
> u32 *data, u32 len)
> @@ -63,6 +64,408 @@ static void nbl_hw_wr_regs(struct nbl_hw_mgt *hw_mgt, u64 reg, const u32 *data,
> spin_unlock(&hw_mgt->reg_lock);
> }
>
> +static u32 nbl_hw_get_quirks(struct nbl_hw_mgt *hw_mgt)
> +{
> + u32 quirks;
> +
> + nbl_hw_read_mbx_regs(hw_mgt, NBL_LEONIS_QUIRKS_OFFSET, &quirks,
> + sizeof(u32));
> +
> + if (quirks == NBL_LEONIS_ILLEGAL_REG_VALUE)
> + return 0;
> +
> + return quirks;
> +}
> +
> +static void nbl_configure_dped_checksum(struct nbl_hw_mgt *hw_mgt)
> +{
> + union dped_l4_ck_cmd_40_u l4_ck_cmd_40;
> +
> + /* DPED dped_l4_ck_cmd_40 for sctp */
> + nbl_hw_rd_regs(hw_mgt, NBL_DPED_L4_CK_CMD_40_ADDR, (u32 *)&l4_ck_cmd_40,
> + sizeof(l4_ck_cmd_40));
> + l4_ck_cmd_40.info.en = 1;
> + nbl_hw_wr_regs(hw_mgt, NBL_DPED_L4_CK_CMD_40_ADDR, (u32 *)&l4_ck_cmd_40,
> + sizeof(l4_ck_cmd_40));
> +}
> +
> +static int nbl_dped_init(struct nbl_hw_mgt *hw_mgt)
> +{
> + nbl_hw_wr32(hw_mgt, NBL_DPED_VLAN_OFFSET, 0xC);
> + nbl_hw_wr32(hw_mgt, NBL_DPED_DSCP_OFFSET_0, 0x8);
> + nbl_hw_wr32(hw_mgt, NBL_DPED_DSCP_OFFSET_1, 0x4);
> +
> + /* dped checksum offload */
> + nbl_configure_dped_checksum(hw_mgt);
> +
> + return 0;
> +}
> +
> +static int nbl_uped_init(struct nbl_hw_mgt *hw_mgt)
> +{
> + struct ped_hw_edit_profile hw_edit;
> +
> + nbl_hw_rd_regs(hw_mgt, NBL_UPED_HW_EDT_PROF_TABLE(NBL_DPED_V4_TCP_IDX),
> + (u32 *)&hw_edit, sizeof(hw_edit));
> + hw_edit.l3_len = 0;
> + nbl_hw_wr_regs(hw_mgt, NBL_UPED_HW_EDT_PROF_TABLE(NBL_DPED_V4_TCP_IDX),
> + (u32 *)&hw_edit, sizeof(hw_edit));
> +
> + nbl_hw_rd_regs(hw_mgt, NBL_UPED_HW_EDT_PROF_TABLE(NBL_DPED_V6_TCP_IDX),
> + (u32 *)&hw_edit, sizeof(hw_edit));
> + hw_edit.l3_len = 1;
> + nbl_hw_wr_regs(hw_mgt, NBL_UPED_HW_EDT_PROF_TABLE(NBL_DPED_V6_TCP_IDX),
> + (u32 *)&hw_edit, sizeof(hw_edit));
> +
> + return 0;
> +}
> +
> +static void nbl_shaping_eth_init(struct nbl_hw_mgt *hw_mgt, u8 eth_id, u8 speed)
> +{
> + struct nbl_shaping_dvn_dport dvn_dport = { 0 };
> + struct nbl_shaping_dport dport = { 0 };
> + u32 rate, half_rate;
> +
> + if (speed == NBL_FW_PORT_SPEED_100G) {
> + rate = NBL_SHAPING_DPORT_100G_RATE;
> + half_rate = NBL_SHAPING_DPORT_HALF_100G_RATE;
> + } else {
> + rate = NBL_SHAPING_DPORT_25G_RATE;
> + half_rate = NBL_SHAPING_DPORT_HALF_25G_RATE;
> + }
> +
> + dport.cir = rate;
> + dport.pir = rate;
> + dport.depth = max(dport.cir * 2, NBL_LR_LEONIS_NET_BUCKET_DEPTH);
> + dport.cbs = dport.depth;
> + dport.pbs = dport.depth;
> + dport.valid = 1;
> +
> + dvn_dport.cir = half_rate;
> + dvn_dport.pir = rate;
> + dvn_dport.depth = dport.depth;
> + dvn_dport.cbs = dvn_dport.depth;
> + dvn_dport.pbs = dvn_dport.depth;
> + dvn_dport.valid = 1;
> +
> + nbl_hw_wr_regs(hw_mgt, NBL_SHAPING_DPORT_REG(eth_id), (u32 *)&dport,
> + sizeof(dport));
> + nbl_hw_wr_regs(hw_mgt, NBL_SHAPING_DVN_DPORT_REG(eth_id),
> + (u32 *)&dvn_dport, sizeof(dvn_dport));
> +}
> +
> +static int nbl_shaping_init(struct nbl_hw_mgt *hw_mgt, u8 speed)
> +{
> +#define NBL_SHAPING_FLUSH_INTERVAL 128
> + struct nbl_shaping_net net_shaping = { 0 };
> + struct dsch_psha_en psha_en = { 0 };
> + int i;
> +
> + for (i = 0; i < NBL_MAX_ETHERNET; i++)
> + nbl_shaping_eth_init(hw_mgt, i, speed);
> +
> + psha_en.en = 0xF;
> + nbl_hw_wr_regs(hw_mgt, NBL_DSCH_PSHA_EN_ADDR, (u32 *)&psha_en,
> + sizeof(psha_en));
> +
> + for (i = 0; i < NBL_MAX_FUNC; i++) {
> + nbl_hw_wr_regs(hw_mgt, NBL_SHAPING_NET_REG(i),
> + (u32 *)&net_shaping, sizeof(net_shaping));
> + if ((i % NBL_SHAPING_FLUSH_INTERVAL) == 0)
> + nbl_flush_writes(hw_mgt);
> + }
> + nbl_flush_writes(hw_mgt);
> + return 0;
> +}
> +
> +static int nbl_dsch_qid_max_init(struct nbl_hw_mgt *hw_mgt)
> +{
> + struct dsch_vn_quanta quanta = { 0 };
> +
> + quanta.h_qua = NBL_HOST_QUANTA;
> + quanta.e_qua = NBL_ECPU_QUANTA;
> + nbl_hw_wr_regs(hw_mgt, NBL_DSCH_VN_QUANTA_ADDR, (u32 *)&quanta,
> + sizeof(quanta));
> + nbl_hw_wr32(hw_mgt, NBL_DSCH_HOST_QID_MAX, NBL_MAX_QUEUE_ID);
> +
> + nbl_hw_wr32(hw_mgt, NBL_DVN_ECPU_QUEUE_NUM, 0);
> + nbl_hw_wr32(hw_mgt, NBL_UVN_ECPU_QUEUE_NUM, 0);
> +
> + return 0;
> +}
> +
> +static int nbl_ustore_init(struct nbl_hw_mgt *hw_mgt, u8 eth_num)
> +{
> + struct nbl_ustore_port_drop_th drop_th = { 0 };
> + struct ustore_pkt_len pkt_len;
> + int i;
> +
> + nbl_hw_rd_regs(hw_mgt, NBL_USTORE_PKT_LEN_ADDR, (u32 *)&pkt_len,
> + sizeof(pkt_len));
Why are we issuing read while ignoring return values? Issuing read to
clear registers? A comment here would help.
> + /* min arp packet length 42 (14 + 28) */
> + pkt_len.min = 42;
> + nbl_hw_wr_regs(hw_mgt, NBL_USTORE_PKT_LEN_ADDR, (u32 *)&pkt_len,
> + sizeof(pkt_len));
> +
> + drop_th.en = 1;
> + if (eth_num == 1)
> + drop_th.disc_th = NBL_USTORE_SINGLE_ETH_DROP_TH;
> + else if (eth_num == 2)
> + drop_th.disc_th = NBL_USTORE_DUAL_ETH_DROP_TH;
> + else
> + drop_th.disc_th = NBL_USTORE_QUAD_ETH_DROP_TH;
> +
> + for (i = 0; i < NBL_MAX_ETHERNET; i++)
> + nbl_hw_wr_regs(hw_mgt, NBL_USTORE_PORT_DROP_TH_REG_ARR(i),
> + (u32 *)&drop_th, sizeof(drop_th));
> +
> + for (i = 0; i < NBL_MAX_ETHERNET; i++) {
> + nbl_hw_rd32(hw_mgt, NBL_USTORE_BUF_PORT_DROP_PKT(i));
> + nbl_hw_rd32(hw_mgt, NBL_USTORE_BUF_PORT_TRUN_PKT(i));
> + }
> +
> + return 0;
> +}
> +
> +static int nbl_dstore_init(struct nbl_hw_mgt *hw_mgt, u8 speed)
> +{
> + struct dstore_port_drop_th drop_th;
> + struct dstore_d_dport_fc_th fc_th;
> + struct dstore_disc_bp_th bp_th;
> + int i;
> +
> + for (i = 0; i < NBL_DSTORE_PORT_DROP_TH_DEPTH; i++) {
> + nbl_hw_rd_regs(hw_mgt, NBL_DSTORE_PORT_DROP_TH_REG(i),
> + (u32 *)&drop_th, sizeof(drop_th));
> + drop_th.en = 0;
> + nbl_hw_wr_regs(hw_mgt, NBL_DSTORE_PORT_DROP_TH_REG(i),
> + (u32 *)&drop_th, sizeof(drop_th));
> + }
> +
> + nbl_hw_rd_regs(hw_mgt, NBL_DSTORE_DISC_BP_TH, (u32 *)&bp_th,
> + sizeof(bp_th));
> + bp_th.en = 1;
> + nbl_hw_wr_regs(hw_mgt, NBL_DSTORE_DISC_BP_TH, (u32 *)&bp_th,
> + sizeof(bp_th));
> +
> + for (i = 0; i < NBL_MAX_ETHERNET; i++) {
> + nbl_hw_rd_regs(hw_mgt, NBL_DSTORE_D_DPORT_FC_TH_REG(i),
> + (u32 *)&fc_th, sizeof(fc_th));
> + if (speed == NBL_FW_PORT_SPEED_100G) {
> + fc_th.xoff_th = NBL_DSTORE_DROP_XOFF_TH_100G;
> + fc_th.xon_th = NBL_DSTORE_DROP_XON_TH_100G;
> + } else {
> + fc_th.xoff_th = NBL_DSTORE_DROP_XOFF_TH;
> + fc_th.xon_th = NBL_DSTORE_DROP_XON_TH;
> + }
> +
> + fc_th.fc_en = 1;
> + nbl_hw_wr_regs(hw_mgt, NBL_DSTORE_D_DPORT_FC_TH_REG(i),
> + (u32 *)&fc_th, sizeof(fc_th));
> + }
> +
> + return 0;
> +}
> +
> +static void nbl_dvn_descreq_num_cfg(struct nbl_hw_mgt *hw_mgt, u32 descreq_num)
> +{
> + u32 split_ring_num = (descreq_num >> 16) & 0xffff;
> + struct nbl_dvn_descreq_num_cfg num_cfg = { 0 };
> + u32 packet_ring_num = descreq_num & 0xffff;
> +
> + packet_ring_num =
> + clamp(packet_ring_num, PACKET_RING_MIN, PACKET_RING_MAX);
> + num_cfg.packed_l1_num =
> + (packet_ring_num - PACKET_RING_BASE) / PACKET_RING_DIV;
> +
> + split_ring_num = clamp(split_ring_num, SPLIT_RING_MIN,
> + SPLIT_RING_MAX);
> + num_cfg.avring_cfg_num = split_ring_num > SPLIT_RING_MIN ?
> + SPLIT_RING_CFG_16 :
> + SPLIT_RING_CFG_8;
> +
> + nbl_hw_wr_regs(hw_mgt, NBL_DVN_DESCREQ_NUM_CFG, (u32 *)&num_cfg,
> + sizeof(num_cfg));
> +}
> +
> +static int nbl_dvn_init(struct nbl_hw_mgt *hw_mgt, u8 speed)
> +{
> + struct nbl_dvn_desc_wr_merge_timeout timeout = { 0 };
> + struct nbl_dvn_dif_req_rd_ro_flag ro_flag = { 0 };
> +
> + timeout.cfg_cycle = DEFAULT_DVN_DESC_WR_MERGE_TIMEOUT_MAX;
> + nbl_hw_wr_regs(hw_mgt, NBL_DVN_DESC_WR_MERGE_TIMEOUT, (u32 *)&timeout,
> + sizeof(timeout));
> +
> + ro_flag.rd_desc_ro_en = 1;
> + ro_flag.rd_data_ro_en = 1;
> + ro_flag.rd_avring_ro_en = 1;
> + nbl_hw_wr_regs(hw_mgt, NBL_DVN_DIF_REQ_RD_RO_FLAG, (u32 *)&ro_flag,
> + sizeof(ro_flag));
> +
> + if (speed == NBL_FW_PORT_SPEED_100G)
> + nbl_dvn_descreq_num_cfg(hw_mgt,
> + DEFAULT_DVN_100G_DESCREQ_NUMCFG);
> + else
> + nbl_dvn_descreq_num_cfg(hw_mgt, DEFAULT_DVN_DESCREQ_NUMCFG);
> +
> + return 0;
> +}
> +
> +static int nbl_uvn_init(struct nbl_hw_mgt *hw_mgt)
> +{
> + struct uvn_desc_prefetch_init prefetch_init = { 0 };
> + struct uvn_desc_wr_timeout desc_wr_timeout = { 0 };
> + struct uvn_dif_req_ro_flag flag = { 0 };
> + struct uvn_queue_err_mask mask = { 0 };
> + u16 wr_timeout = 0x12c;
> + u32 timeout = 119760; /* 200us 200000/1.67 */
> + u32 quirks;
> +
> + nbl_hw_wr32(hw_mgt, NBL_UVN_DESC_RD_WAIT, timeout);
> +
> + desc_wr_timeout.num = wr_timeout;
> + nbl_hw_wr_regs(hw_mgt, NBL_UVN_DESC_WR_TIMEOUT, (u32 *)&desc_wr_timeout,
> + sizeof(desc_wr_timeout));
> +
> + flag.avail_rd = 1;
> + flag.desc_rd = 1;
> + flag.pkt_wr = 1;
> + flag.desc_wr = 0;
> + nbl_hw_wr_regs(hw_mgt, NBL_UVN_DIF_REQ_RO_FLAG, (u32 *)&flag,
> + sizeof(flag));
> +
> + nbl_hw_rd_regs(hw_mgt, NBL_UVN_QUEUE_ERR_MASK, (u32 *)&mask,
> + sizeof(mask));
> + mask.dif_err = 1;
> + nbl_hw_wr_regs(hw_mgt, NBL_UVN_QUEUE_ERR_MASK, (u32 *)&mask,
> + sizeof(mask));
> +
> + prefetch_init.num = NBL_UVN_DESC_PREFETCH_NUM;
> + prefetch_init.sel = 0;
> + quirks = nbl_hw_get_quirks(hw_mgt);
> + if (!(quirks & BIT(NBL_QUIRKS_UVN_PREFETCH_ALIGN)))
> + prefetch_init.sel = 1;
> + nbl_hw_wr_regs(hw_mgt, NBL_UVN_DESC_PREFETCH_INIT,
> + (u32 *)&prefetch_init, sizeof(prefetch_init));
> +
> + return 0;
> +}
> +
> +static int nbl_uqm_init(struct nbl_hw_mgt *hw_mgt)
> +{
> + struct nbl_uqm_que_type que_type = { 0 };
> + u32 cnt = 0;
> + int i;
> +
> + nbl_hw_wr_regs(hw_mgt, NBL_UQM_FWD_DROP_CNT, &cnt, sizeof(cnt));
> +
> + nbl_hw_wr_regs(hw_mgt, NBL_UQM_DROP_PKT_CNT, &cnt, sizeof(cnt));
> + nbl_hw_wr_regs(hw_mgt, NBL_UQM_DROP_PKT_SLICE_CNT, &cnt,
> + sizeof(cnt));
> + nbl_hw_wr_regs(hw_mgt, NBL_UQM_DROP_PKT_LEN_ADD_CNT, &cnt,
> + sizeof(cnt));
> + nbl_hw_wr_regs(hw_mgt, NBL_UQM_DROP_HEAD_PNTR_ADD_CNT, &cnt,
> + sizeof(cnt));
> + nbl_hw_wr_regs(hw_mgt, NBL_UQM_DROP_WEIGHT_ADD_CNT, &cnt,
> + sizeof(cnt));
> +
> + for (i = 0; i < NBL_UQM_PORT_DROP_DEPTH; i++) {
> + nbl_hw_wr_regs(hw_mgt,
> + NBL_UQM_PORT_DROP_PKT_CNT + (sizeof(cnt) * i),
> + &cnt, sizeof(cnt));
> + nbl_hw_wr_regs(hw_mgt,
> + NBL_UQM_PORT_DROP_PKT_SLICE_CNT +
> + (sizeof(cnt) * i),
> + &cnt, sizeof(cnt));
> + nbl_hw_wr_regs(hw_mgt,
> + NBL_UQM_PORT_DROP_PKT_LEN_ADD_CNT +
> + (sizeof(cnt) * i),
> + &cnt, sizeof(cnt));
> + nbl_hw_wr_regs(hw_mgt,
> + NBL_UQM_PORT_DROP_HEAD_PNTR_ADD_CNT +
> + (sizeof(cnt) * i),
> + &cnt, sizeof(cnt));
> + nbl_hw_wr_regs(hw_mgt,
> + NBL_UQM_PORT_DROP_WEIGHT_ADD_CNT +
> + (sizeof(cnt) * i),
> + &cnt, sizeof(cnt));
> + }
> +
> + for (i = 0; i < NBL_UQM_DPORT_DROP_DEPTH; i++)
> + nbl_hw_wr_regs(hw_mgt,
> + NBL_UQM_DPORT_DROP_CNT + (sizeof(cnt) * i),
> + &cnt, sizeof(cnt));
> +
> + que_type.bp_drop = 0;
> + nbl_hw_wr_regs(hw_mgt, NBL_UQM_QUE_TYPE, (u32 *)&que_type,
> + sizeof(que_type));
> +
> + return 0;
> +}
> +
> +static int nbl_dp_init(struct nbl_hw_mgt *hw_mgt, u8 speed, u8 eth_num)
> +{
> + nbl_dped_init(hw_mgt);
> + nbl_uped_init(hw_mgt);
> + nbl_shaping_init(hw_mgt, speed);
> + nbl_dsch_qid_max_init(hw_mgt);
> + nbl_ustore_init(hw_mgt, eth_num);
> + nbl_dstore_init(hw_mgt, speed);
> + nbl_dvn_init(hw_mgt, speed);
> + nbl_uvn_init(hw_mgt);
> + nbl_uqm_init(hw_mgt);
> +
> + return 0;
> +}
> +
> +static int nbl_host_padpt_init(struct nbl_hw_mgt *hw_mgt)
> +{
> + /* padpt flow control register */
> + nbl_hw_wr32(hw_mgt, NBL_HOST_PADPT_HOST_CFG_FC_CPLH_UP, 0x10400);
> + nbl_hw_wr32(hw_mgt, NBL_HOST_PADPT_HOST_CFG_FC_PD_DN, 0x10080);
> + nbl_hw_wr32(hw_mgt, NBL_HOST_PADPT_HOST_CFG_FC_PH_DN, 0x10010);
> + nbl_hw_wr32(hw_mgt, NBL_HOST_PADPT_HOST_CFG_FC_NPH_DN, 0x10010);
> +
> + return 0;
> +}
> +
> +static int nbl_intf_init(struct nbl_hw_mgt *hw_mgt)
> +{
> + nbl_host_padpt_init(hw_mgt);
> + return 0;
> +}
> +
> +static void nbl_hw_set_driver_status(struct nbl_hw_mgt *hw_mgt, bool active)
> +{
> + u32 status;
> +
> + status = nbl_hw_rd32(hw_mgt, NBL_DRIVER_STATUS_REG);
> +
> + status = (status & ~(1 << NBL_DRIVER_STATUS_BIT)) |
> + (active << NBL_DRIVER_STATUS_BIT);
> +
> + nbl_hw_wr32(hw_mgt, NBL_DRIVER_STATUS_REG, status);
> +}
> +
> +static void nbl_hw_deinit_chip_module(struct nbl_hw_mgt *hw_mgt)
> +{
> + nbl_hw_set_driver_status(hw_mgt, false);
> +}
> +
> +static int nbl_hw_init_chip_module(struct nbl_hw_mgt *hw_mgt, u8 eth_speed,
> + u8 eth_num)
> +{
> + nbl_dp_init(hw_mgt, eth_speed, eth_num);
> + nbl_intf_init(hw_mgt);
> +
> + nbl_write_all_regs(hw_mgt);
> + nbl_hw_set_driver_status(hw_mgt, true);
> + hw_mgt->version = nbl_hw_rd32(hw_mgt, NBL_HW_DUMMY_REG);
> +
> + return 0;
> +}
> +
> static void nbl_hw_enable_mailbox_irq(struct nbl_hw_mgt *hw_mgt, u16 func_id,
> bool enable_msix, u16 global_vec_id)
> {
> @@ -262,6 +665,9 @@ static u32 nbl_hw_get_fw_eth_map(struct nbl_hw_mgt *hw_mgt)
> }
>
> static struct nbl_hw_ops hw_ops = {
> + .init_chip_module = nbl_hw_init_chip_module,
> + .deinit_chip_module = nbl_hw_deinit_chip_module,
> +
> .configure_msix_map = nbl_hw_configure_msix_map,
> .configure_msix_info = nbl_hw_configure_msix_info,
>
> diff --git a/drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_hw_leonis/nbl_hw_leonis.h b/drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_hw_leonis/nbl_hw_leonis.h
> index 8831394ed11b..7487d0e757e3 100644
> --- a/drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_hw_leonis/nbl_hw_leonis.h
> +++ b/drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_hw_leonis/nbl_hw_leonis.h
> @@ -128,7 +128,8 @@ struct nbl_function_msix_map {
> #define NBL_DPED_VLAN_OFFSET (NBL_DP_DPED_BASE + 0x000003F4)
> #define NBL_DPED_DSCP_OFFSET_0 (NBL_DP_DPED_BASE + 0x000003F8)
> #define NBL_DPED_DSCP_OFFSET_1 (NBL_DP_DPED_BASE + 0x000003FC)
> -
> +#define NBL_DPED_V4_TCP_IDX 5
> +#define NBL_DPED_V6_TCP_IDX 6
> /* DPED hw_edt_prof/ UPED hw_edt_prof */
> struct ped_hw_edit_profile {
> u32 l4_len:2;
> @@ -268,6 +269,15 @@ struct dsch_vn_quanta {
>
> #define DEFAULT_DVN_DESC_WR_MERGE_TIMEOUT_MAX 0x3FF
>
> +#define PACKET_RING_MIN 8U
> +#define PACKET_RING_MAX 32U
> +#define SPLIT_RING_MIN 8U
> +#define SPLIT_RING_MAX 16U
> +#define PACKET_RING_BASE 8U
> +#define PACKET_RING_DIV 4U
> +#define SPLIT_RING_CFG_8 0U
> +#define SPLIT_RING_CFG_16 1U
> +
> struct nbl_dvn_descreq_num_cfg {
> u32 avring_cfg_num:1; /* spilit ring descreq_num 0:8,1:16 */
> u32 rsv0:3;
> diff --git a/drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_hw_leonis/nbl_resource_leonis.c b/drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_hw_leonis/nbl_resource_leonis.c
> index 770a255ab06f..403e74b9b1d0 100644
> --- a/drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_hw_leonis/nbl_resource_leonis.c
> +++ b/drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_hw_leonis/nbl_resource_leonis.c
> @@ -46,6 +46,10 @@ nbl_res_setup_ops(struct device *dev, struct nbl_resource_mgt *res_mgt)
>
> if (!is_ops_inited) {
> ret = nbl_intr_setup_ops(&res_ops);
> + if (ret)
> + return ERR_PTR(-ENOMEM);
> +
> + ret = nbl_vsi_setup_ops(&res_ops);
> if (ret)
> return ERR_PTR(-ENOMEM);
> is_ops_inited = true;
> diff --git a/drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_resource.h b/drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_resource.h
> index 1b80676cf19a..675649ffb271 100644
> --- a/drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_resource.h
> +++ b/drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_resource.h
> @@ -110,6 +110,7 @@ void nbl_res_get_eth_id(struct nbl_resource_mgt *res_mgt, u16 vsi_id,
> u8 *eth_mode, u8 *eth_id, u8 *logic_eth_id);
> int nbl_intr_mgt_start(struct nbl_resource_mgt *res_mgt);
> int nbl_intr_setup_ops(struct nbl_resource_ops *resource_ops);
> +int nbl_vsi_setup_ops(struct nbl_resource_ops *resource_ops);
> void nbl_res_pf_dev_vsi_type_to_hw_vsi_type(u16 src_type,
> enum nbl_vsi_serv_type *dst_type);
> #endif
> diff --git a/drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_vsi.c b/drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_vsi.c
> new file mode 100644
> index 000000000000..67b9b23ad012
> --- /dev/null
> +++ b/drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_vsi.c
> @@ -0,0 +1,51 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + * Copyright (c) 2025 Nebula Matrix Limited.
> + */
> +#include <linux/device.h>
> +#include "nbl_vsi.h"
> +
> +static void nbl_res_vsi_deinit_chip_module(struct nbl_resource_mgt *res_mgt)
> +{
> + struct nbl_hw_ops *hw_ops = res_mgt->hw_ops_tbl->ops;
> +
> + hw_ops->deinit_chip_module(res_mgt->hw_ops_tbl->priv);
> +}
> +
> +static int nbl_res_vsi_init_chip_module(struct nbl_resource_mgt *res_mgt)
> +{
> + u8 eth_speed = res_mgt->resource_info->board_info.eth_speed;
> + u8 eth_num = res_mgt->resource_info->board_info.eth_num;
> + struct nbl_hw_ops *hw_ops = res_mgt->hw_ops_tbl->ops;
> + struct nbl_hw_mgt *p = res_mgt->hw_ops_tbl->priv;
> + int ret;
> +
> + ret = hw_ops->init_chip_module(p, eth_speed, eth_num);
> +
> + return ret;
> +}
> +
> +/* NBL_VSI_SET_OPS(ops_name, func)
> + *
> + * Use X Macros to reduce setup and remove codes.
> + */
> +#define NBL_VSI_OPS_TBL \
> +do { \
> + NBL_VSI_SET_OPS(init_chip_module, \
> + nbl_res_vsi_init_chip_module); \
> + NBL_VSI_SET_OPS(deinit_chip_module, \
> + nbl_res_vsi_deinit_chip_module); \
> +} while (0)
> +
> +int nbl_vsi_setup_ops(struct nbl_resource_ops *res_ops)
> +{
> +#define NBL_VSI_SET_OPS(name, func) \
> + do { \
> + res_ops->NBL_NAME(name) = func; \
> + ; \
> + } while (0)
> + NBL_VSI_OPS_TBL;
> +#undef NBL_VSI_SET_OPS
> +
> + return 0;
> +}
> diff --git a/drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_vsi.h b/drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_vsi.h
> new file mode 100644
> index 000000000000..6fd79add57d6
> --- /dev/null
> +++ b/drivers/net/ethernet/nebula-matrix/nbl/nbl_hw/nbl_vsi.h
> @@ -0,0 +1,11 @@
> +/* SPDX-License-Identifier: GPL-2.0 */
> +/*
> + * Copyright (c) 2025 Nebula Matrix Limited.
> + */
> +
> +#ifndef _NBL_VSI_H_
> +#define _NBL_VSI_H_
> +
> +#include "nbl_resource.h"
> +
> +#endif
> diff --git a/drivers/net/ethernet/nebula-matrix/nbl/nbl_include/nbl_def_hw.h b/drivers/net/ethernet/nebula-matrix/nbl/nbl_include/nbl_def_hw.h
> index f97f7a810ad0..93dfe74d3643 100644
> --- a/drivers/net/ethernet/nebula-matrix/nbl/nbl_include/nbl_def_hw.h
> +++ b/drivers/net/ethernet/nebula-matrix/nbl/nbl_include/nbl_def_hw.h
> @@ -11,6 +11,10 @@
> struct nbl_hw_mgt;
> struct nbl_adapter;
> struct nbl_hw_ops {
> + int (*init_chip_module)(struct nbl_hw_mgt *hw_mgt, u8 eth_speed,
> + u8 eth_num);
> + void (*deinit_chip_module)(struct nbl_hw_mgt *hw_mgt);
> +
> void (*configure_msix_map)(struct nbl_hw_mgt *hw_mgt, u16 func_id,
> bool valid, dma_addr_t dma_addr, u8 bus,
> u8 devid, u8 function);
> diff --git a/drivers/net/ethernet/nebula-matrix/nbl/nbl_include/nbl_include.h b/drivers/net/ethernet/nebula-matrix/nbl/nbl_include/nbl_include.h
> index e4f11e6ded94..5203bb2a9a5f 100644
> --- a/drivers/net/ethernet/nebula-matrix/nbl/nbl_include/nbl_include.h
> +++ b/drivers/net/ethernet/nebula-matrix/nbl/nbl_include/nbl_include.h
> @@ -45,4 +45,35 @@ struct nbl_init_param {
> bool pci_using_dac;
> };
>
> +enum nbl_fw_port_speed {
> + NBL_FW_PORT_SPEED_10G,
> + NBL_FW_PORT_SPEED_25G,
> + NBL_FW_PORT_SPEED_50G,
> + NBL_FW_PORT_SPEED_100G,
> +};
> +
> +#define NBL_OPS_CALL(func, para) \
> +do { \
> + typeof(func) _func = (func); \
> + if (_func) \
> + _func para; \
> +} while (0)
> +
> +#define NBL_OPS_CALL_RET(func, para) \
> +({ \
> + typeof(func) _func = (func); \
> + _func ? _func para : 0; \
> +})
> +
> +#define NBL_OPS_CALL_RET_PTR(func, para) \
> +({ \
> + typeof(func) _func = (func); \
> + _func ? _func para : NULL; \
> +})
> +
> +enum nbl_performance_mode {
> + NBL_QUIRKS_NO_TOE,
> + NBL_QUIRKS_UVN_PREFETCH_ALIGN,
> +};
> +
> #endif
next prev parent reply other threads:[~2026-04-02 9:06 UTC|newest]
Thread overview: 18+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-04-01 2:23 [PATCH v10 net-next 00/11] nbl driver for Nebulamatrix NICs illusion.wang
2026-04-01 2:23 ` [PATCH v10 net-next 01/11] net/nebula-matrix: add minimum nbl build framework illusion.wang
2026-04-01 2:23 ` [PATCH v10 net-next 02/11] net/nebula-matrix: add our driver architecture illusion.wang
2026-04-02 9:02 ` Mohsin Bashir
2026-04-01 2:23 ` [PATCH v10 net-next 03/11] net/nebula-matrix: add chip related definitions illusion.wang
2026-04-01 2:23 ` [PATCH v10 net-next 04/11] net/nebula-matrix: channel msg value and msg struct illusion.wang
2026-04-02 9:03 ` Mohsin Bashir
2026-04-01 2:23 ` [PATCH v10 net-next 05/11] net/nebula-matrix: add channel layer illusion.wang
2026-04-02 9:04 ` Mohsin Bashir
2026-04-01 2:23 ` [PATCH v10 net-next 06/11] net/nebula-matrix: add common resource implementation illusion.wang
2026-04-02 9:05 ` Mohsin Bashir
2026-04-01 2:23 ` [PATCH v10 net-next 07/11] net/nebula-matrix: add intr " illusion.wang
2026-04-02 9:05 ` Mohsin Bashir
2026-04-01 2:23 ` [PATCH v10 net-next 08/11] net/nebula-matrix: add vsi " illusion.wang
2026-04-02 9:06 ` Mohsin Bashir [this message]
2026-04-01 2:23 ` [PATCH v10 net-next 09/11] net/nebula-matrix: add Dispatch layer implementation illusion.wang
2026-04-01 2:23 ` [PATCH v10 net-next 10/11] net/nebula-matrix: add common/ctrl dev init/reinit operation illusion.wang
2026-04-01 2:23 ` [PATCH v10 net-next 11/11] net/nebula-matrix: add common dev start/stop operation illusion.wang
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=db9e681f-3a87-4ea3-a8b3-cfa96e328c42@gmail.com \
--to=mohsin.bashr@gmail.com \
--cc=alvin.wang@nebula-matrix.com \
--cc=andrew+netdev@lunn.ch \
--cc=corbet@lwn.net \
--cc=dimon.zhao@nebula-matrix.com \
--cc=edumazet@google.com \
--cc=enelsonmoore@gmail.com \
--cc=hkallweit1@gmail.com \
--cc=horms@kernel.org \
--cc=illusion.wang@nebula-matrix.com \
--cc=jani.nikula@intel.com \
--cc=kuba@kernel.org \
--cc=linux-doc@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=lorenzo@kernel.org \
--cc=lukas.bulwahn@redhat.com \
--cc=netdev@vger.kernel.org \
--cc=pabeni@redhat.com \
--cc=sam.chen@nebula-matrix.com \
--cc=skhan@linuxfoundation.org \
--cc=vadim.fedorenko@linux.dev \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox