From mboxrd@z Thu Jan 1 00:00:00 1970 Received: from mail-pl1-f170.google.com (mail-pl1-f170.google.com [209.85.214.170]) (using TLSv1.2 with cipher ECDHE-RSA-AES128-GCM-SHA256 (128/128 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 357CD3D4114 for ; Tue, 31 Mar 2026 11:18:39 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=209.85.214.170 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1774955923; cv=none; b=B8ll/QS+6oejhvqWLlCAMqNBF9hWw8oGGOrfJsFgWIz8x2sNJNi5u9g8iKT4JMJJRTO1NEoeMuguOAEQ1eh41dgXP++ZucpETWEKplRn10jsawZtARcBDbpPSjfiWWLFxu530pkUqyNU1Jjq4mJiMzBCTC6Hm6CVfSGE3dHMBfQ= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1774955923; c=relaxed/simple; bh=XOgh1lQpfMb8xq/ZeI2S8j7Mn0TIsIqZfVtAnxB89kE=; h=From:To:Cc:Subject:Date:Message-ID:In-Reply-To:References: MIME-Version; b=Spfs+8CJIzs/zKkDIVFyAGN3Rhaq2+RSF3nHkqHjykAQqWmQBQfw2bk8FxEcm+9JzUyHrgCrApSrmySw15FYvHGfnp1VS/5Zf8UKDJFsonL773Aa7PFGrRHay9ftnWzN08bN7T9mWAncqejiO505OzY3sRc9aOy8fFISLtDa6A4= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=gmail.com; spf=pass smtp.mailfrom=gmail.com; dkim=pass (2048-bit key) header.d=gmail.com header.i=@gmail.com header.b=QpHzneL1; arc=none smtp.client-ip=209.85.214.170 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=gmail.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=gmail.com Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=gmail.com header.i=@gmail.com header.b="QpHzneL1" Received: by mail-pl1-f170.google.com with SMTP id d9443c01a7336-2a9633ef0d6so3099775ad.0 for ; Tue, 31 Mar 2026 04:18:39 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20251104; t=1774955919; x=1775560719; darn=lists.linux.dev; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:from:to:cc:subject:date :message-id:reply-to; bh=QHpvoUfsACZGTgRgdNP3q3dMEYguoaXrUM/Oeck39zo=; b=QpHzneL1df2NVc3LDwvrCAHFCx3xTOfyAddSDhrfZ2hS/Gb9WTQS0QHssBr3S9+UfT JJkWcE+mWdytztrE8C1FYp2wEY90W1iocrPJi62+cP7OReYrqP2mK9fYqHEDOQpxJwKK r2qXg0lFD3yHy6l1oYRqdHwit9gD4TsH/Xv/9fmlxgal87g6/rYFGF655Q6AxlN9ZXg0 HwuE1MObqeLqToStKkGrksyBSYtgcKPCZAiTRZzu4ONfEJ9ALjdDykZRMV3+J4Ql3zPq jlOggo/GmvBc18DbyX9k53VlH0NdGafm3CbvMd9+VjnogVrOMdIR08UELCZSobkvuboR JLpA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20251104; t=1774955919; x=1775560719; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:x-gm-gg:x-gm-message-state:from :to:cc:subject:date:message-id:reply-to; bh=QHpvoUfsACZGTgRgdNP3q3dMEYguoaXrUM/Oeck39zo=; b=TpaoOjUwaTV02vCZSNtJCkuNSiK6Spqa7+PNz+iVaWX7NEWJbiemh00H61R9C/Kc2E MrUM8HFBnvI4j+dp1pyzmNz6hV1w+bCqm3sVYxgG30Bs879pZzGrGfOsk2YdVrcM9J+j Hd+iG2zRWKC9s7nqkeDMRnYfCvhUO8KblIRbQcA7jvyHNJ3r9PE5lzptabk8KZQ8Nje3 MbycCjtUaSGmPDFQ9UXKcSHTLgKpecD6lV0qiXqC84oKEtBvCW/MQiBHAEy5lISURZ8e zVabyGbcpnB+zA4LFHdLGIZdP28NF2Knd7aJqshH65uVV29OgegKl9PWp+IC4SqHMThE CNJg== X-Gm-Message-State: AOJu0YxMipF1fUp+wQKnNc3A2GlR+nUUxxyhBHRHV6npVgHRubPGM6j2 3hoDfAaA9fNFIevOV6fE8PI0SnOoa3aF+tS6C1Mn4LUL8Re7sWSDDvoW X-Gm-Gg: ATEYQzybsXG6MWwndZqpLNle4tUo0XrWWRjZEe2KH9LpN18Q+kpb7eRa0a24BGyGyP5 9TJhZMkKzXBKtDK4qSTgP1lQwyps4tjhorK/AvMC1xEt9/jurXugV4u2SONTn+7PAPB+/sbWYLa REzUoLI7J15WjInNMMmtgP9UGYvezJB2AkiwHh6jlCua3t3TLiZHramG67n/6uID3GVxQ1+NET6 VtOl+yZ1KfTdjKoM9FIUvhl16jp5IwWxjjKsiwGQX8/f1uu3v+8ZUwMX5TBSZGmpAC4y2SBRLnU l3xQIRKaaiX0RusmOU3ZnRYN4+ASfpiA+uOnlV1m1Aa45FtkfpOaWbBFZeKuFBJHsC4KXFFCgu+ cmJUbEt8UTfJYlqfvpheUetybFAtm+KGRo+94Hwc6wGWr/jZPpVISQ0vpWWAEbYINVVtGlwTVF2 IxzvJl7jwmEgo0lglkUSgmRr7LUCk= X-Received: by 2002:a17:902:f54f:b0:2b2:527d:fd with SMTP id d9443c01a7336-2b2527d0369mr44846605ad.8.1774955919382; Tue, 31 Mar 2026 04:18:39 -0700 (PDT) Received: from fedora ([2401:4900:86e0:ee11:14b4:e5e5:9799:983f]) by smtp.gmail.com with ESMTPSA id d9443c01a7336-2b242766203sm128901145ad.40.2026.03.31.04.18.36 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Tue, 31 Mar 2026 04:18:39 -0700 (PDT) From: AyushMukkanwar To: gregkh@linuxfoundation.org Cc: linux-staging@lists.linux.dev, linux-kernel@vger.kernel.org, AyushMukkanwar Subject: [PATCH v2 2/2] staging: octeon: ethernet: replace pr_* with dev_* and netdev_* Date: Tue, 31 Mar 2026 16:47:57 +0530 Message-ID: <20260331111757.110703-3-ayushmukkanwar@gmail.com> X-Mailer: git-send-email 2.53.0 In-Reply-To: <20260331111757.110703-1-ayushmukkanwar@gmail.com> References: <2026033119-endpoint-jimmy-6b83@gregkh> <20260331111757.110703-1-ayushmukkanwar@gmail.com> Precedence: bulk X-Mailing-List: linux-staging@lists.linux.dev List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Transfer-Encoding: 8bit Replace pr_*() calls with dev_*() and netdev_*() to include device information in log messages. A per-device struct octeon_ethernet_platform is introduced to pass the device pointer through work queue callbacks via container_of. The device pointer is also stored in oct_rx_group to make it available in the NAPI poll function and propagated through the memory pool functions by adding a struct device * parameter. Signed-off-by: AyushMukkanwar --- drivers/staging/octeon/ethernet-mem.c | 26 ++++++------ drivers/staging/octeon/ethernet-mem.h | 4 +- drivers/staging/octeon/ethernet-rx.c | 6 ++- drivers/staging/octeon/ethernet-rx.h | 6 +-- drivers/staging/octeon/ethernet.c | 61 ++++++++++++++++----------- 5 files changed, 59 insertions(+), 44 deletions(-) diff --git a/drivers/staging/octeon/ethernet-mem.c b/drivers/staging/octeon/ethernet-mem.c index 532594957ebc..68e1d416c22c 100644 --- a/drivers/staging/octeon/ethernet-mem.c +++ b/drivers/staging/octeon/ethernet-mem.c @@ -44,7 +44,7 @@ static int cvm_oct_fill_hw_skbuff(int pool, int size, int elements) * @size: Size of the buffer needed for the pool * @elements: Number of buffers to allocate */ -static void cvm_oct_free_hw_skbuff(int pool, int size, int elements) +static void cvm_oct_free_hw_skbuff(int pool, int size, int elements, struct device *dev) { char *memory; @@ -59,10 +59,10 @@ static void cvm_oct_free_hw_skbuff(int pool, int size, int elements) } while (memory); if (elements < 0) - pr_warn("Freeing of pool %u had too many skbuffs (%d)\n", + dev_warn(dev, "Freeing of pool %u had too many skbuffs (%d)\n", pool, elements); else if (elements > 0) - pr_warn("Freeing of pool %u is missing %d skbuffs\n", + dev_warn(dev, "Freeing of pool %u is missing %d skbuffs\n", pool, elements); } @@ -74,7 +74,7 @@ static void cvm_oct_free_hw_skbuff(int pool, int size, int elements) * * Returns the actual number of buffers allocated. */ -static int cvm_oct_fill_hw_memory(int pool, int size, int elements) +static int cvm_oct_fill_hw_memory(int pool, int size, int elements, struct device *dev) { char *memory; char *fpa; @@ -93,7 +93,7 @@ static int cvm_oct_fill_hw_memory(int pool, int size, int elements) */ memory = kmalloc(size + 256, GFP_ATOMIC); if (unlikely(!memory)) { - pr_warn("Unable to allocate %u bytes for FPA pool %d\n", + dev_warn(dev, "Unable to allocate %u bytes for FPA pool %d\n", elements * size, pool); break; } @@ -111,7 +111,7 @@ static int cvm_oct_fill_hw_memory(int pool, int size, int elements) * @size: Size of each buffer in the pool * @elements: Number of buffers that should be in the pool */ -static void cvm_oct_free_hw_memory(int pool, int size, int elements) +static void cvm_oct_free_hw_memory(int pool, int size, int elements, struct device *dev) { char *memory; char *fpa; @@ -127,28 +127,28 @@ static void cvm_oct_free_hw_memory(int pool, int size, int elements) } while (fpa); if (elements < 0) - pr_warn("Freeing of pool %u had too many buffers (%d)\n", + dev_warn(dev, "Freeing of pool %u had too many buffers (%d)\n", pool, elements); else if (elements > 0) - pr_warn("Warning: Freeing of pool %u is missing %d buffers\n", + dev_warn(dev, "Warning: Freeing of pool %u is missing %d buffers\n", pool, elements); } -int cvm_oct_mem_fill_fpa(int pool, int size, int elements) +int cvm_oct_mem_fill_fpa(int pool, int size, int elements, struct device *dev) { int freed; if (pool == CVMX_FPA_PACKET_POOL) freed = cvm_oct_fill_hw_skbuff(pool, size, elements); else - freed = cvm_oct_fill_hw_memory(pool, size, elements); + freed = cvm_oct_fill_hw_memory(pool, size, elements, dev); return freed; } -void cvm_oct_mem_empty_fpa(int pool, int size, int elements) +void cvm_oct_mem_empty_fpa(int pool, int size, int elements, struct device *dev) { if (pool == CVMX_FPA_PACKET_POOL) - cvm_oct_free_hw_skbuff(pool, size, elements); + cvm_oct_free_hw_skbuff(pool, size, elements, dev); else - cvm_oct_free_hw_memory(pool, size, elements); + cvm_oct_free_hw_memory(pool, size, elements, dev); } diff --git a/drivers/staging/octeon/ethernet-mem.h b/drivers/staging/octeon/ethernet-mem.h index 692dcdb7154d..22a38846c751 100644 --- a/drivers/staging/octeon/ethernet-mem.h +++ b/drivers/staging/octeon/ethernet-mem.h @@ -5,5 +5,5 @@ * Copyright (c) 2003-2007 Cavium Networks */ -int cvm_oct_mem_fill_fpa(int pool, int size, int elements); -void cvm_oct_mem_empty_fpa(int pool, int size, int elements); +int cvm_oct_mem_fill_fpa(int pool, int size, int elements, struct device *dev); +void cvm_oct_mem_empty_fpa(int pool, int size, int elements, struct device *dev); diff --git a/drivers/staging/octeon/ethernet-rx.c b/drivers/staging/octeon/ethernet-rx.c index d0b43d50b83c..461f9077742e 100644 --- a/drivers/staging/octeon/ethernet-rx.c +++ b/drivers/staging/octeon/ethernet-rx.c @@ -35,6 +35,7 @@ static struct oct_rx_group { int irq; int group; struct napi_struct napi; + struct device *dev; } oct_rx_group[16]; /** @@ -397,7 +398,7 @@ static int cvm_oct_poll(struct oct_rx_group *rx_group, int budget) /* Restore the scratch area */ cvmx_scratch_write64(CVMX_SCR_SCRATCH, old_scratch); } - cvm_oct_rx_refill_pool(0); + cvm_oct_rx_refill_pool(0, rx_group->dev); return rx_count; } @@ -448,7 +449,7 @@ void cvm_oct_poll_controller(struct net_device *dev) } #endif -void cvm_oct_rx_initialize(void) +void cvm_oct_rx_initialize(struct device *dev) { int i; struct net_device *dev_for_napi = NULL; @@ -475,6 +476,7 @@ void cvm_oct_rx_initialize(void) oct_rx_group[i].irq = OCTEON_IRQ_WORKQ0 + i; oct_rx_group[i].group = i; + oct_rx_group[i].dev = dev; /* Register an IRQ handler to receive POW interrupts */ ret = request_irq(oct_rx_group[i].irq, cvm_oct_do_interrupt, 0, diff --git a/drivers/staging/octeon/ethernet-rx.h b/drivers/staging/octeon/ethernet-rx.h index ff6482fa20d6..443c6208d09d 100644 --- a/drivers/staging/octeon/ethernet-rx.h +++ b/drivers/staging/octeon/ethernet-rx.h @@ -6,10 +6,10 @@ */ void cvm_oct_poll_controller(struct net_device *dev); -void cvm_oct_rx_initialize(void); +void cvm_oct_rx_initialize(struct device *dev); void cvm_oct_rx_shutdown(void); -static inline void cvm_oct_rx_refill_pool(int fill_threshold) +static inline void cvm_oct_rx_refill_pool(int fill_threshold, struct device *dev) { int number_to_free; int num_freed; @@ -22,7 +22,7 @@ static inline void cvm_oct_rx_refill_pool(int fill_threshold) -number_to_free); num_freed = cvm_oct_mem_fill_fpa(CVMX_FPA_PACKET_POOL, CVMX_FPA_PACKET_POOL_SIZE, - number_to_free); + number_to_free, dev); if (num_freed != number_to_free) { cvmx_fau_atomic_add32(FAU_NUM_PACKET_BUFFERS_TO_FREE, number_to_free - num_freed); diff --git a/drivers/staging/octeon/ethernet.c b/drivers/staging/octeon/ethernet.c index eadb74fc14c8..36fbeabe9eb2 100644 --- a/drivers/staging/octeon/ethernet.c +++ b/drivers/staging/octeon/ethernet.c @@ -104,11 +104,18 @@ struct net_device *cvm_oct_device[TOTAL_NUMBER_OF_PORTS]; u64 cvm_oct_tx_poll_interval; -static void cvm_oct_rx_refill_worker(struct work_struct *work); -static DECLARE_DELAYED_WORK(cvm_oct_rx_refill_work, cvm_oct_rx_refill_worker); +struct octeon_ethernet_platform { + struct device *dev; + struct delayed_work rx_refill_work; +}; + +static struct octeon_ethernet_platform *oct_plt; static void cvm_oct_rx_refill_worker(struct work_struct *work) { + struct octeon_ethernet_platform *plt = container_of(work, struct octeon_ethernet_platform, + rx_refill_work.work); + /* * FPA 0 may have been drained, try to refill it if we need * more than num_packet_buffers / 2, otherwise normal receive @@ -116,10 +123,11 @@ static void cvm_oct_rx_refill_worker(struct work_struct *work) * could be received so cvm_oct_napi_poll would never be * invoked to do the refill. */ - cvm_oct_rx_refill_pool(num_packet_buffers / 2); + + cvm_oct_rx_refill_pool(num_packet_buffers / 2, plt->dev); if (!atomic_read(&cvm_oct_poll_queue_stopping)) - schedule_delayed_work(&cvm_oct_rx_refill_work, HZ); + schedule_delayed_work(&plt->rx_refill_work, HZ); } static void cvm_oct_periodic_worker(struct work_struct *work) @@ -138,17 +146,17 @@ static void cvm_oct_periodic_worker(struct work_struct *work) schedule_delayed_work(&priv->port_periodic_work, HZ); } -static void cvm_oct_configure_common_hw(void) +static void cvm_oct_configure_common_hw(struct device *dev) { /* Setup the FPA */ cvmx_fpa_enable(); cvm_oct_mem_fill_fpa(CVMX_FPA_PACKET_POOL, CVMX_FPA_PACKET_POOL_SIZE, - num_packet_buffers); + num_packet_buffers, dev); cvm_oct_mem_fill_fpa(CVMX_FPA_WQE_POOL, CVMX_FPA_WQE_POOL_SIZE, - num_packet_buffers); + num_packet_buffers, dev); if (CVMX_FPA_OUTPUT_BUFFER_POOL != CVMX_FPA_PACKET_POOL) cvm_oct_mem_fill_fpa(CVMX_FPA_OUTPUT_BUFFER_POOL, - CVMX_FPA_OUTPUT_BUFFER_POOL_SIZE, 1024); + CVMX_FPA_OUTPUT_BUFFER_POOL_SIZE, 1024, dev); #ifdef __LITTLE_ENDIAN { @@ -685,11 +693,17 @@ static int cvm_oct_probe(struct platform_device *pdev) pip = pdev->dev.of_node; if (!pip) { - pr_err("Error: No 'pip' in /aliases\n"); + dev_err(&pdev->dev, "Error: No 'pip' in /aliases\n"); return -EINVAL; } - cvm_oct_configure_common_hw(); + oct_plt = devm_kzalloc(&pdev->dev, sizeof(*oct_plt), GFP_KERNEL); + if (!oct_plt) + return -ENOMEM; + oct_plt->dev = &pdev->dev; + INIT_DELAYED_WORK(&oct_plt->rx_refill_work, cvm_oct_rx_refill_worker); + + cvm_oct_configure_common_hw(&pdev->dev); cvmx_helper_initialize_packet_io_global(); @@ -783,16 +797,15 @@ static int cvm_oct_probe(struct platform_device *pdev) dev->max_mtu = OCTEON_MAX_MTU - mtu_overhead; if (register_netdev(dev) < 0) { - pr_err("Failed to register ethernet device for POW\n"); + netdev_err(dev, "Failed to register ethernet device for POW\n"); free_netdev(dev); } else { cvm_oct_device[CVMX_PIP_NUM_INPUT_PORTS] = dev; - pr_info("%s: POW send group %d, receive group %d\n", - dev->name, pow_send_group, - pow_receive_group); + netdev_info(dev, "POW send group %d, receive group %d\n", + pow_send_group, pow_receive_group); } } else { - pr_err("Failed to allocate ethernet device for POW\n"); + dev_err(&pdev->dev, "Failed to allocate ethernet device for POW\n"); } } @@ -812,8 +825,8 @@ static int cvm_oct_probe(struct platform_device *pdev) struct net_device *dev = alloc_etherdev(sizeof(struct octeon_ethernet)); if (!dev) { - pr_err("Failed to allocate ethernet device for port %d\n", - port); + dev_err(&pdev->dev, "Failed to allocate ethernet device for port %d\n", + port); continue; } @@ -897,7 +910,7 @@ static int cvm_oct_probe(struct platform_device *pdev) if (!dev->netdev_ops) { free_netdev(dev); } else if (register_netdev(dev) < 0) { - pr_err("Failed to register ethernet device for interface %d, port %d\n", + netdev_err(dev, "Failed to register ethernet device for interface %d, port %d\n", interface, priv->port); free_netdev(dev); } else { @@ -912,14 +925,14 @@ static int cvm_oct_probe(struct platform_device *pdev) } cvm_oct_tx_initialize(); - cvm_oct_rx_initialize(); + cvm_oct_rx_initialize(&pdev->dev); /* * 150 uS: about 10 1500-byte packets at 1GE. */ cvm_oct_tx_poll_interval = 150 * (octeon_get_clock_rate() / 1000000); - schedule_delayed_work(&cvm_oct_rx_refill_work, HZ); + schedule_delayed_work(&oct_plt->rx_refill_work, HZ); return 0; } @@ -931,7 +944,7 @@ static void cvm_oct_remove(struct platform_device *pdev) cvmx_ipd_disable(); atomic_inc_return(&cvm_oct_poll_queue_stopping); - cancel_delayed_work_sync(&cvm_oct_rx_refill_work); + cancel_delayed_work_sync(&oct_plt->rx_refill_work); cvm_oct_rx_shutdown(); cvm_oct_tx_shutdown(); @@ -959,12 +972,12 @@ static void cvm_oct_remove(struct platform_device *pdev) /* Free the HW pools */ cvm_oct_mem_empty_fpa(CVMX_FPA_PACKET_POOL, CVMX_FPA_PACKET_POOL_SIZE, - num_packet_buffers); + num_packet_buffers, &pdev->dev); cvm_oct_mem_empty_fpa(CVMX_FPA_WQE_POOL, CVMX_FPA_WQE_POOL_SIZE, - num_packet_buffers); + num_packet_buffers, &pdev->dev); if (CVMX_FPA_OUTPUT_BUFFER_POOL != CVMX_FPA_PACKET_POOL) cvm_oct_mem_empty_fpa(CVMX_FPA_OUTPUT_BUFFER_POOL, - CVMX_FPA_OUTPUT_BUFFER_POOL_SIZE, 128); + CVMX_FPA_OUTPUT_BUFFER_POOL_SIZE, 128, &pdev->dev); } static const struct of_device_id cvm_oct_match[] = { -- 2.53.0