Linux Tegra architecture development
 help / color / mirror / Atom feed
From: Mostafa Saleh <smostafa@google.com>
To: Jason Gunthorpe <jgg@nvidia.com>
Cc: iommu@lists.linux.dev, Jonathan Hunter <jonathanh@nvidia.com>,
	Joerg Roedel <joro@8bytes.org>,
	linux-arm-kernel@lists.infradead.org,
	linux-tegra@vger.kernel.org, Robin Murphy <robin.murphy@arm.com>,
	Thierry Reding <thierry.reding@kernel.org>,
	Krishna Reddy <vdumpa@nvidia.com>, Will Deacon <will@kernel.org>,
	David Matlack <dmatlack@google.com>,
	Pasha Tatashin <pasha.tatashin@soleen.com>,
	patches@lists.linux.dev, Samiullah Khawaja <skhawaja@google.com>
Subject: Re: [PATCH 4/9] iommu/arm-smmu-v3: Convert arm_smmu_cmdq_batch cmds to struct arm_smmu_cmd
Date: Thu, 7 May 2026 09:22:05 +0000	[thread overview]
Message-ID: <afxZvSt2uuW5OEU9@google.com> (raw)
In-Reply-To: <4-v1-b7dc0a0d4aa0+3723d-smmu_no_cmdq_ent_jgg@nvidia.com>

On Fri, May 01, 2026 at 11:29:13AM -0300, Jason Gunthorpe wrote:
> Convert the batch's type to also get the remaining helper functions to
> use the new type and complete replacing naked u64s with the new struct.
> 
> The low-level queue_write()/queue_read()/queue_remove_raw() functions
> remain u64-based since they are shared by event and PRI queues which
> have different entry sizes.
> 
> Signed-off-by: Jason Gunthorpe <jgg@nvidia.com>

Reviewed-by: Mostafa Saleh <smostafa@google.com>

Thanks,
Mostafa

> ---
>  .../arm/arm-smmu-v3/arm-smmu-v3-iommufd.c     | 24 +++---
>  drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c   | 74 ++++++++++---------
>  drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.h   |  5 +-
>  .../iommu/arm/arm-smmu-v3/tegra241-cmdqv.c    |  8 +-
>  4 files changed, 58 insertions(+), 53 deletions(-)
> 
> diff --git a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-iommufd.c b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-iommufd.c
> index ddae0b07c76b50..1e9f7d2de34414 100644
> --- a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-iommufd.c
> +++ b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-iommufd.c
> @@ -300,7 +300,7 @@ static int arm_vsmmu_vsid_to_sid(struct arm_vsmmu *vsmmu, u32 vsid, u32 *sid)
>  /* This is basically iommu_viommu_arm_smmuv3_invalidate in u64 for conversion */
>  struct arm_vsmmu_invalidation_cmd {
>  	union {
> -		u64 cmd[2];
> +		struct arm_smmu_cmd cmd;
>  		struct iommu_viommu_arm_smmuv3_invalidate ucmd;
>  	};
>  };
> @@ -316,32 +316,32 @@ static int arm_vsmmu_convert_user_cmd(struct arm_vsmmu *vsmmu,
>  				      struct arm_vsmmu_invalidation_cmd *cmd)
>  {
>  	/* Commands are le64 stored in u64 */
> -	cmd->cmd[0] = le64_to_cpu(cmd->ucmd.cmd[0]);
> -	cmd->cmd[1] = le64_to_cpu(cmd->ucmd.cmd[1]);
> +	cmd->cmd.data[0] = le64_to_cpu(cmd->ucmd.cmd[0]);
> +	cmd->cmd.data[1] = le64_to_cpu(cmd->ucmd.cmd[1]);
>  
> -	switch (cmd->cmd[0] & CMDQ_0_OP) {
> +	switch (cmd->cmd.data[0] & CMDQ_0_OP) {
>  	case CMDQ_OP_TLBI_NSNH_ALL:
>  		/* Convert to NH_ALL */
> -		cmd->cmd[0] = CMDQ_OP_TLBI_NH_ALL |
> +		cmd->cmd.data[0] = CMDQ_OP_TLBI_NH_ALL |
>  			      FIELD_PREP(CMDQ_TLBI_0_VMID, vsmmu->vmid);
> -		cmd->cmd[1] = 0;
> +		cmd->cmd.data[1] = 0;
>  		break;
>  	case CMDQ_OP_TLBI_NH_VA:
>  	case CMDQ_OP_TLBI_NH_VAA:
>  	case CMDQ_OP_TLBI_NH_ALL:
>  	case CMDQ_OP_TLBI_NH_ASID:
> -		cmd->cmd[0] &= ~CMDQ_TLBI_0_VMID;
> -		cmd->cmd[0] |= FIELD_PREP(CMDQ_TLBI_0_VMID, vsmmu->vmid);
> +		cmd->cmd.data[0] &= ~CMDQ_TLBI_0_VMID;
> +		cmd->cmd.data[0] |= FIELD_PREP(CMDQ_TLBI_0_VMID, vsmmu->vmid);
>  		break;
>  	case CMDQ_OP_ATC_INV:
>  	case CMDQ_OP_CFGI_CD:
>  	case CMDQ_OP_CFGI_CD_ALL: {
> -		u32 sid, vsid = FIELD_GET(CMDQ_CFGI_0_SID, cmd->cmd[0]);
> +		u32 sid, vsid = FIELD_GET(CMDQ_CFGI_0_SID, cmd->cmd.data[0]);
>  
>  		if (arm_vsmmu_vsid_to_sid(vsmmu, vsid, &sid))
>  			return -EIO;
> -		cmd->cmd[0] &= ~CMDQ_CFGI_0_SID;
> -		cmd->cmd[0] |= FIELD_PREP(CMDQ_CFGI_0_SID, sid);
> +		cmd->cmd.data[0] &= ~CMDQ_CFGI_0_SID;
> +		cmd->cmd.data[0] |= FIELD_PREP(CMDQ_CFGI_0_SID, sid);
>  		break;
>  	}
>  	default:
> @@ -386,7 +386,7 @@ int arm_vsmmu_cache_invalidate(struct iommufd_viommu *viommu,
>  			continue;
>  
>  		/* FIXME always uses the main cmdq rather than trying to group by type */
> -		ret = arm_smmu_cmdq_issue_cmdlist(smmu, &smmu->cmdq, last->cmd,
> +		ret = arm_smmu_cmdq_issue_cmdlist(smmu, &smmu->cmdq, &last->cmd,
>  						  cur - last, true);
>  		if (ret) {
>  			cur--;
> diff --git a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
> index 67d23e9c54804e..b3ef001ce80d23 100644
> --- a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
> +++ b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
> @@ -268,9 +268,12 @@ static int queue_remove_raw(struct arm_smmu_queue *q, u64 *ent)
>  }
>  
>  /* High-level queue accessors */
> -static int arm_smmu_cmdq_build_cmd(u64 *cmd, struct arm_smmu_cmdq_ent *ent)
> +static int arm_smmu_cmdq_build_cmd(struct arm_smmu_cmd *cmd_out,
> +				   struct arm_smmu_cmdq_ent *ent)
>  {
> -	memset(cmd, 0, 1 << CMDQ_ENT_SZ_SHIFT);
> +	u64 *cmd = cmd_out->data;
> +
> +	memset(cmd_out, 0, sizeof(*cmd_out));
>  	cmd[0] |= FIELD_PREP(CMDQ_0_OP, ent->opcode);
>  
>  	switch (ent->opcode) {
> @@ -390,7 +393,8 @@ static bool arm_smmu_cmdq_needs_busy_polling(struct arm_smmu_device *smmu,
>  	return smmu->options & ARM_SMMU_OPT_TEGRA241_CMDQV;
>  }
>  
> -static void arm_smmu_cmdq_build_sync_cmd(u64 *cmd, struct arm_smmu_device *smmu,
> +static void arm_smmu_cmdq_build_sync_cmd(struct arm_smmu_cmd *cmd,
> +					 struct arm_smmu_device *smmu,
>  					 struct arm_smmu_cmdq *cmdq, u32 prod)
>  {
>  	struct arm_smmu_queue *q = &cmdq->q;
> @@ -409,7 +413,8 @@ static void arm_smmu_cmdq_build_sync_cmd(u64 *cmd, struct arm_smmu_device *smmu,
>  
>  	arm_smmu_cmdq_build_cmd(cmd, &ent);
>  	if (arm_smmu_cmdq_needs_busy_polling(smmu, cmdq))
> -		u64p_replace_bits(cmd, CMDQ_SYNC_0_CS_NONE, CMDQ_SYNC_0_CS);
> +		u64p_replace_bits(&cmd->data[0], CMDQ_SYNC_0_CS_NONE,
> +				  CMDQ_SYNC_0_CS);
>  }
>  
>  void __arm_smmu_cmdq_skip_err(struct arm_smmu_device *smmu,
> @@ -422,9 +427,8 @@ void __arm_smmu_cmdq_skip_err(struct arm_smmu_device *smmu,
>  		[CMDQ_ERR_CERROR_ATC_INV_IDX]	= "ATC invalidate timeout",
>  	};
>  	struct arm_smmu_queue *q = &cmdq->q;
> -
>  	int i;
> -	u64 cmd[CMDQ_ENT_DWORDS];
> +	struct arm_smmu_cmd cmd;
>  	u32 cons = readl_relaxed(q->cons_reg);
>  	u32 idx = FIELD_GET(CMDQ_CONS_ERR, cons);
>  	struct arm_smmu_cmdq_ent cmd_sync = {
> @@ -457,17 +461,18 @@ void __arm_smmu_cmdq_skip_err(struct arm_smmu_device *smmu,
>  	 * We may have concurrent producers, so we need to be careful
>  	 * not to touch any of the shadow cmdq state.
>  	 */
> -	queue_read(cmd, Q_ENT(q, cons), q->ent_dwords);
> +	queue_read(cmd.data, Q_ENT(q, cons), q->ent_dwords);
>  	dev_err(smmu->dev, "skipping command in error state:\n");
> -	for (i = 0; i < ARRAY_SIZE(cmd); ++i)
> -		dev_err(smmu->dev, "\t0x%016llx\n", (unsigned long long)cmd[i]);
> +	for (i = 0; i < ARRAY_SIZE(cmd.data); ++i)
> +		dev_err(smmu->dev, "\t0x%016llx\n", (unsigned long long)cmd.data[i]);
>  
>  	/* Convert the erroneous command into a CMD_SYNC */
> -	arm_smmu_cmdq_build_cmd(cmd, &cmd_sync);
> +	arm_smmu_cmdq_build_cmd(&cmd, &cmd_sync);
>  	if (arm_smmu_cmdq_needs_busy_polling(smmu, cmdq))
> -		u64p_replace_bits(cmd, CMDQ_SYNC_0_CS_NONE, CMDQ_SYNC_0_CS);
> +		u64p_replace_bits(&cmd.data[0], CMDQ_SYNC_0_CS_NONE,
> +				  CMDQ_SYNC_0_CS);
>  
> -	queue_write(Q_ENT(q, cons), cmd, q->ent_dwords);
> +	queue_write(Q_ENT(q, cons), cmd.data, q->ent_dwords);
>  }
>  
>  static void arm_smmu_cmdq_skip_err(struct arm_smmu_device *smmu)
> @@ -767,7 +772,8 @@ static int arm_smmu_cmdq_poll_until_sync(struct arm_smmu_device *smmu,
>  	return __arm_smmu_cmdq_poll_until_consumed(smmu, cmdq, llq);
>  }
>  
> -static void arm_smmu_cmdq_write_entries(struct arm_smmu_cmdq *cmdq, u64 *cmds,
> +static void arm_smmu_cmdq_write_entries(struct arm_smmu_cmdq *cmdq,
> +					struct arm_smmu_cmd *cmds,
>  					u32 prod, int n)
>  {
>  	int i;
> @@ -777,10 +783,9 @@ static void arm_smmu_cmdq_write_entries(struct arm_smmu_cmdq *cmdq, u64 *cmds,
>  	};
>  
>  	for (i = 0; i < n; ++i) {
> -		u64 *cmd = &cmds[i * CMDQ_ENT_DWORDS];
> -
>  		prod = queue_inc_prod_n(&llq, i);
> -		queue_write(Q_ENT(&cmdq->q, prod), cmd, CMDQ_ENT_DWORDS);
> +		queue_write(Q_ENT(&cmdq->q, prod), cmds[i].data,
> +			    ARRAY_SIZE(cmds[i].data));
>  	}
>  }
>  
> @@ -801,10 +806,11 @@ static void arm_smmu_cmdq_write_entries(struct arm_smmu_cmdq *cmdq, u64 *cmds,
>   *   CPU will appear before any of the commands from the other CPU.
>   */
>  int arm_smmu_cmdq_issue_cmdlist(struct arm_smmu_device *smmu,
> -				struct arm_smmu_cmdq *cmdq, u64 *cmds, int n,
> +				struct arm_smmu_cmdq *cmdq,
> +				struct arm_smmu_cmd *cmds, int n,
>  				bool sync)
>  {
> -	u64 cmd_sync[CMDQ_ENT_DWORDS];
> +	struct arm_smmu_cmd cmd_sync;
>  	u32 prod;
>  	unsigned long flags;
>  	bool owner;
> @@ -847,8 +853,9 @@ int arm_smmu_cmdq_issue_cmdlist(struct arm_smmu_device *smmu,
>  	arm_smmu_cmdq_write_entries(cmdq, cmds, llq.prod, n);
>  	if (sync) {
>  		prod = queue_inc_prod_n(&llq, n);
> -		arm_smmu_cmdq_build_sync_cmd(cmd_sync, smmu, cmdq, prod);
> -		queue_write(Q_ENT(&cmdq->q, prod), cmd_sync, CMDQ_ENT_DWORDS);
> +		arm_smmu_cmdq_build_sync_cmd(&cmd_sync, smmu, cmdq, prod);
> +		queue_write(Q_ENT(&cmdq->q, prod), cmd_sync.data,
> +			    ARRAY_SIZE(cmd_sync.data));
>  
>  		/*
>  		 * In order to determine completion of our CMD_SYNC, we must
> @@ -925,7 +932,7 @@ static int __arm_smmu_cmdq_issue_cmd(struct arm_smmu_device *smmu,
>  				     bool sync)
>  {
>  	return arm_smmu_cmdq_issue_cmdlist(
> -		smmu, arm_smmu_get_cmdq(smmu, cmd), cmd->data, 1, sync);
> +		smmu, arm_smmu_get_cmdq(smmu, cmd), cmd, 1, sync);
>  }
>  
>  static int arm_smmu_cmdq_issue_cmd(struct arm_smmu_device *smmu,
> @@ -954,7 +961,7 @@ static void arm_smmu_cmdq_batch_init(struct arm_smmu_device *smmu,
>  {
>  	struct arm_smmu_cmd cmd;
>  
> -	arm_smmu_cmdq_build_cmd(cmd.data, ent);
> +	arm_smmu_cmdq_build_cmd(&cmd, ent);
>  	arm_smmu_cmdq_batch_init_cmd(smmu, cmds, &cmd);
>  }
>  
> @@ -966,9 +973,8 @@ static void arm_smmu_cmdq_batch_add(struct arm_smmu_device *smmu,
>  			  (smmu->options & ARM_SMMU_OPT_CMDQ_FORCE_SYNC);
>  	struct arm_smmu_cmd cmd;
>  	bool unsupported_cmd;
> -	int index;
>  
> -	if (unlikely(arm_smmu_cmdq_build_cmd(cmd.data, ent))) {
> +	if (unlikely(arm_smmu_cmdq_build_cmd(&cmd, ent))) {
>  		dev_warn(smmu->dev, "ignoring unknown CMDQ opcode 0x%x\n",
>  			 ent->opcode);
>  		return;
> @@ -987,9 +993,7 @@ static void arm_smmu_cmdq_batch_add(struct arm_smmu_device *smmu,
>  		arm_smmu_cmdq_batch_init_cmd(smmu, cmds, &cmd);
>  	}
>  
> -	index = cmds->num * CMDQ_ENT_DWORDS;
> -	memcpy(&cmds->cmds[index], cmd.data, sizeof(cmd.data));
> -	cmds->num++;
> +	cmds->cmds[cmds->num++] = cmd;
>  }
>  
>  static int arm_smmu_cmdq_batch_submit(struct arm_smmu_device *smmu,
> @@ -1025,7 +1029,7 @@ static void arm_smmu_page_response(struct device *dev, struct iopf_fault *unused
>  		break;
>  	}
>  
> -	arm_smmu_cmdq_build_cmd(hw_cmd.data, &cmd);
> +	arm_smmu_cmdq_build_cmd(&hw_cmd, &cmd);
>  	arm_smmu_cmdq_issue_cmd(master->smmu, &hw_cmd);
>  
>  	/*
> @@ -1865,7 +1869,7 @@ static void arm_smmu_ste_writer_sync_entry(struct arm_smmu_entry_writer *writer)
>  	};
>  	struct arm_smmu_cmd cmd;
>  
> -	arm_smmu_cmdq_build_cmd(cmd.data, &ent);
> +	arm_smmu_cmdq_build_cmd(&cmd, &ent);
>  	arm_smmu_cmdq_issue_cmd_with_sync(writer->master->smmu, &cmd);
>  }
>  
> @@ -1899,7 +1903,7 @@ static void arm_smmu_write_ste(struct arm_smmu_master *master, u32 sid,
>  					 } };
>  		struct arm_smmu_cmd prefetch_cmd;
>  
> -		arm_smmu_cmdq_build_cmd(prefetch_cmd.data, &prefetch_ent);
> +		arm_smmu_cmdq_build_cmd(&prefetch_cmd, &prefetch_ent);
>  		arm_smmu_cmdq_issue_cmd(smmu, &prefetch_cmd);
>  	}
>  }
> @@ -2339,7 +2343,7 @@ static void arm_smmu_handle_ppr(struct arm_smmu_device *smmu, u64 *evt)
>  		};
>  		struct arm_smmu_cmd cmd;
>  
> -		arm_smmu_cmdq_build_cmd(cmd.data, &ent);
> +		arm_smmu_cmdq_build_cmd(&cmd, &ent);
>  		arm_smmu_cmdq_issue_cmd(smmu, &cmd);
>  	}
>  }
> @@ -3462,7 +3466,7 @@ static void arm_smmu_inv_flush_iotlb_tag(struct arm_smmu_inv *inv)
>  	}
>  
>  	cmd.opcode = inv->nsize_opcode;
> -	arm_smmu_cmdq_build_cmd(hw_cmd.data, &cmd);
> +	arm_smmu_cmdq_build_cmd(&hw_cmd, &cmd);
>  	arm_smmu_cmdq_issue_cmd_with_sync(inv->smmu, &hw_cmd);
>  }
>  
> @@ -4875,18 +4879,18 @@ static int arm_smmu_device_reset(struct arm_smmu_device *smmu)
>  
>  	/* Invalidate any cached configuration */
>  	ent.opcode = CMDQ_OP_CFGI_ALL;
> -	arm_smmu_cmdq_build_cmd(cmd.data, &ent);
> +	arm_smmu_cmdq_build_cmd(&cmd, &ent);
>  	arm_smmu_cmdq_issue_cmd_with_sync(smmu, &cmd);
>  
>  	/* Invalidate any stale TLB entries */
>  	if (smmu->features & ARM_SMMU_FEAT_HYP) {
>  		ent.opcode = CMDQ_OP_TLBI_EL2_ALL;
> -		arm_smmu_cmdq_build_cmd(cmd.data, &ent);
> +		arm_smmu_cmdq_build_cmd(&cmd, &ent);
>  		arm_smmu_cmdq_issue_cmd_with_sync(smmu, &cmd);
>  	}
>  
>  	ent.opcode = CMDQ_OP_TLBI_NSNH_ALL;
> -	arm_smmu_cmdq_build_cmd(cmd.data, &ent);
> +	arm_smmu_cmdq_build_cmd(&cmd, &ent);
>  	arm_smmu_cmdq_issue_cmd_with_sync(smmu, &cmd);
>  
>  	/* Event queue */
> diff --git a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.h b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.h
> index 6d73f6b63e64a9..1fe6917448b774 100644
> --- a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.h
> +++ b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.h
> @@ -651,7 +651,7 @@ static inline bool arm_smmu_cmdq_supports_cmd(struct arm_smmu_cmdq *cmdq,
>  }
>  
>  struct arm_smmu_cmdq_batch {
> -	u64				cmds[CMDQ_BATCH_ENTRIES * CMDQ_ENT_DWORDS];
> +	struct arm_smmu_cmd		cmds[CMDQ_BATCH_ENTRIES];
>  	struct arm_smmu_cmdq		*cmdq;
>  	int				num;
>  };
> @@ -1148,7 +1148,8 @@ void arm_smmu_install_ste_for_dev(struct arm_smmu_master *master,
>  				  const struct arm_smmu_ste *target);
>  
>  int arm_smmu_cmdq_issue_cmdlist(struct arm_smmu_device *smmu,
> -				struct arm_smmu_cmdq *cmdq, u64 *cmds, int n,
> +				struct arm_smmu_cmdq *cmdq,
> +				struct arm_smmu_cmd *cmds, int n,
>  				bool sync);
>  
>  #ifdef CONFIG_ARM_SMMU_V3_SVA
> diff --git a/drivers/iommu/arm/arm-smmu-v3/tegra241-cmdqv.c b/drivers/iommu/arm/arm-smmu-v3/tegra241-cmdqv.c
> index b4d8c1f2fd3878..67be62a6e7640a 100644
> --- a/drivers/iommu/arm/arm-smmu-v3/tegra241-cmdqv.c
> +++ b/drivers/iommu/arm/arm-smmu-v3/tegra241-cmdqv.c
> @@ -427,16 +427,16 @@ tegra241_cmdqv_get_cmdq(struct arm_smmu_device *smmu,
>  static void tegra241_vcmdq_hw_flush_timeout(struct tegra241_vcmdq *vcmdq)
>  {
>  	struct arm_smmu_device *smmu = &vcmdq->cmdqv->smmu;
> -	u64 cmd_sync[CMDQ_ENT_DWORDS] = {};
> +	struct arm_smmu_cmd cmd_sync = {};
>  
> -	cmd_sync[0] = FIELD_PREP(CMDQ_0_OP, CMDQ_OP_CMD_SYNC) |
> -		      FIELD_PREP(CMDQ_SYNC_0_CS, CMDQ_SYNC_0_CS_NONE);
> +	cmd_sync.data[0] = FIELD_PREP(CMDQ_0_OP, CMDQ_OP_CMD_SYNC) |
> +			   FIELD_PREP(CMDQ_SYNC_0_CS, CMDQ_SYNC_0_CS_NONE);
>  
>  	/*
>  	 * It does not hurt to insert another CMD_SYNC, taking advantage of the
>  	 * arm_smmu_cmdq_issue_cmdlist() that waits for the CMD_SYNC completion.
>  	 */
> -	arm_smmu_cmdq_issue_cmdlist(smmu, &smmu->cmdq, cmd_sync, 1, true);
> +	arm_smmu_cmdq_issue_cmdlist(smmu, &smmu->cmdq, &cmd_sync, 1, true);
>  }
>  
>  /* This function is for LVCMDQ, so @vcmdq must not be unmapped yet */
> -- 
> 2.43.0
> 

  reply	other threads:[~2026-05-07  9:22 UTC|newest]

Thread overview: 45+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2026-05-01 14:29 [PATCH 0/9] Remove SMMUv3 struct arm_smmu_cmdq_ent Jason Gunthorpe
2026-05-01 14:29 ` [PATCH 1/9] iommu/arm-smmu-v3: Add struct arm_smmu_cmd to represent the HW format command Jason Gunthorpe
2026-05-06  6:11   ` Nicolin Chen
2026-05-06 23:41   ` Samiullah Khawaja
2026-05-07  9:19   ` Mostafa Saleh
2026-05-08  7:29   ` Pranjal Shrivastava
2026-05-01 14:29 ` [PATCH 2/9] iommu/arm-smmu-v3: Use the HW arm_smmu_cmd in cmdq selection functions Jason Gunthorpe
2026-05-07  9:21   ` Mostafa Saleh
2026-05-08 15:49     ` Jason Gunthorpe
2026-05-08  7:47   ` Pranjal Shrivastava
2026-05-08 15:54     ` Jason Gunthorpe
2026-05-08 16:58       ` Pranjal Shrivastava
2026-05-01 14:29 ` [PATCH 3/9] iommu/arm-smmu-v3: Use the HW arm_smmu_cmd in cmdq submission functions Jason Gunthorpe
2026-05-07  9:21   ` Mostafa Saleh
2026-05-08  8:27   ` Pranjal Shrivastava
2026-05-08 16:00     ` Jason Gunthorpe
2026-05-08 17:00       ` Pranjal Shrivastava
2026-05-01 14:29 ` [PATCH 4/9] iommu/arm-smmu-v3: Convert arm_smmu_cmdq_batch cmds to struct arm_smmu_cmd Jason Gunthorpe
2026-05-07  9:22   ` Mostafa Saleh [this message]
2026-05-08  9:26   ` Pranjal Shrivastava
2026-05-01 14:29 ` [PATCH 5/9] iommu/arm-smmu-v3: Remove CMDQ_OP_CFGI_CD_ALL from arm_smmu_cmdq_build_cmd() Jason Gunthorpe
2026-05-07  9:22   ` Mostafa Saleh
2026-05-08  9:45   ` Pranjal Shrivastava
2026-05-08 16:02     ` Jason Gunthorpe
2026-05-08 17:17       ` Pranjal Shrivastava
2026-05-01 14:29 ` [PATCH 6/9] iommu/arm-smmu-v3: Directly encode simple commands Jason Gunthorpe
2026-05-07  9:22   ` Mostafa Saleh
2026-05-08 11:33   ` Pranjal Shrivastava
2026-05-08 17:37     ` Jason Gunthorpe
2026-05-08 20:09       ` Pranjal Shrivastava
2026-05-08 23:36         ` Jason Gunthorpe
2026-05-10 18:59           ` Pranjal Shrivastava
2026-05-01 14:29 ` [PATCH 7/9] iommu/arm-smmu-v3: Directly encode CMDQ_OP_ATC_INV Jason Gunthorpe
2026-05-07  9:23   ` Mostafa Saleh
2026-05-08 11:46   ` Pranjal Shrivastava
2026-05-09 16:54     ` Jason Gunthorpe
2026-05-11 10:34       ` Pranjal Shrivastava
2026-05-01 14:29 ` [PATCH 8/9] iommu/arm-smmu-v3: Directly encode CMDQ_OP_SYNC Jason Gunthorpe
2026-05-07  9:23   ` Mostafa Saleh
2026-05-08 13:41   ` Pranjal Shrivastava
2026-05-01 14:29 ` [PATCH 9/9] iommu/arm-smmu-v3: Directly encode TLBI commands Jason Gunthorpe
2026-05-07  9:24   ` Mostafa Saleh
2026-05-08 14:00   ` Pranjal Shrivastava
2026-05-07  9:26 ` [PATCH 0/9] Remove SMMUv3 struct arm_smmu_cmdq_ent Mostafa Saleh
2026-05-08 14:03 ` Pranjal Shrivastava

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=afxZvSt2uuW5OEU9@google.com \
    --to=smostafa@google.com \
    --cc=dmatlack@google.com \
    --cc=iommu@lists.linux.dev \
    --cc=jgg@nvidia.com \
    --cc=jonathanh@nvidia.com \
    --cc=joro@8bytes.org \
    --cc=linux-arm-kernel@lists.infradead.org \
    --cc=linux-tegra@vger.kernel.org \
    --cc=pasha.tatashin@soleen.com \
    --cc=patches@lists.linux.dev \
    --cc=robin.murphy@arm.com \
    --cc=skhawaja@google.com \
    --cc=thierry.reding@kernel.org \
    --cc=vdumpa@nvidia.com \
    --cc=will@kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox