* Re: [PATCH 1/3] vdpa/mlx5: Remove mtu field from vdpa net device [not found] ` <20210909123635.30884-2-elic@nvidia.com> @ 2021-09-13 22:09 ` Michael S. Tsirkin 0 siblings, 0 replies; 3+ messages in thread From: Michael S. Tsirkin @ 2021-09-13 22:09 UTC (permalink / raw) To: Eli Cohen; +Cc: eperezma, virtualization On Thu, Sep 09, 2021 at 03:36:33PM +0300, Eli Cohen wrote: > No need to save the mtu int the net device struct. We can save it in the > config struct which cannot be modified. > > Moreover, move the initialization to. mlx5_vdpa_set_features() callback > is not the right place to put it. the reason it's there is the endian-ness mess. before set features VERSION_1 is not set. Does mlx5 support a transitional mode? Or modern only? If the later then cpu_to_mlx5vdpa16 should really be switched to just use LE unconfitionally. > > Signed-off-by: Eli Cohen <elic@nvidia.com> > --- > drivers/vdpa/mlx5/net/mlx5_vnet.c | 9 +++++---- > 1 file changed, 5 insertions(+), 4 deletions(-) > > diff --git a/drivers/vdpa/mlx5/net/mlx5_vnet.c b/drivers/vdpa/mlx5/net/mlx5_vnet.c > index 7784e8a5647f..08ac15b17b83 100644 > --- a/drivers/vdpa/mlx5/net/mlx5_vnet.c > +++ b/drivers/vdpa/mlx5/net/mlx5_vnet.c > @@ -159,7 +159,6 @@ struct mlx5_vdpa_net { > struct mlx5_fc *rx_counter; > struct mlx5_flow_handle *rx_rule; > bool setup; > - u16 mtu; > u32 cur_num_vqs; > }; > > @@ -1942,8 +1941,6 @@ static int mlx5_vdpa_set_features(struct vdpa_device *vdev, u64 features) > return err; > > ndev->mvdev.actual_features = features & ndev->mvdev.mlx_features; > - ndev->config.mtu = cpu_to_mlx5vdpa16(mvdev, ndev->mtu); > - ndev->config.status |= cpu_to_mlx5vdpa16(mvdev, VIRTIO_NET_S_LINK_UP); > update_cvq_info(mvdev); > return err; > } > @@ -2405,6 +2402,7 @@ static int mlx5_vdpa_dev_add(struct vdpa_mgmt_dev *v_mdev, const char *name) > struct mlx5_vdpa_net *ndev; > struct mlx5_core_dev *mdev; > u32 max_vqs; > + u16 mtu; > int err; > > if (mgtdev->ndev) > @@ -2432,10 +2430,13 @@ static int mlx5_vdpa_dev_add(struct vdpa_mgmt_dev *v_mdev, const char *name) > init_mvqs(ndev); > mutex_init(&ndev->reslock); > config = &ndev->config; > - err = query_mtu(mdev, &ndev->mtu); > + err = query_mtu(mdev, &mtu); > if (err) > goto err_mtu; > > + ndev->config.mtu = cpu_to_mlx5vdpa16(mvdev, mtu); > + ndev->config.status |= cpu_to_mlx5vdpa16(mvdev, VIRTIO_NET_S_LINK_UP); > + > err = mlx5_query_nic_vport_mac_address(mdev, 0, 0, config->mac); > if (err) > goto err_mtu; > -- > 2.31.1 _______________________________________________ Virtualization mailing list Virtualization@lists.linux-foundation.org https://lists.linuxfoundation.org/mailman/listinfo/virtualization ^ permalink raw reply [flat|nested] 3+ messages in thread
[parent not found: <20210909123635.30884-3-elic@nvidia.com>]
* Re: [PATCH 2/3] vdpa/mlx5: Rename control VQ workqueue to vdpa wq [not found] ` <20210909123635.30884-3-elic@nvidia.com> @ 2021-09-14 2:57 ` Jason Wang 0 siblings, 0 replies; 3+ messages in thread From: Jason Wang @ 2021-09-14 2:57 UTC (permalink / raw) To: Eli Cohen; +Cc: eperezma, virtualization, mst On Thu, Sep 9, 2021 at 8:36 PM Eli Cohen <elic@nvidia.com> wrote: > > A subesequent patch will use the same workqueue for executing other > work not related to control VQ. Rename the workqueue and the work queue > entry used to convey information to the workqueue. > > Signed-off-by: Eli Cohen <elic@nvidia.com> Acked-by: Jason Wang <jasowang@redhat.com> > --- > drivers/vdpa/mlx5/core/mlx5_vdpa.h | 2 +- > drivers/vdpa/mlx5/net/mlx5_vnet.c | 8 ++++---- > 2 files changed, 5 insertions(+), 5 deletions(-) > > diff --git a/drivers/vdpa/mlx5/core/mlx5_vdpa.h b/drivers/vdpa/mlx5/core/mlx5_vdpa.h > index 01a848adf590..81dc3d88d3dd 100644 > --- a/drivers/vdpa/mlx5/core/mlx5_vdpa.h > +++ b/drivers/vdpa/mlx5/core/mlx5_vdpa.h > @@ -63,7 +63,7 @@ struct mlx5_control_vq { > unsigned short head; > }; > > -struct mlx5_ctrl_wq_ent { > +struct mlx5_vdpa_wq_ent { > struct work_struct work; > struct mlx5_vdpa_dev *mvdev; > }; > diff --git a/drivers/vdpa/mlx5/net/mlx5_vnet.c b/drivers/vdpa/mlx5/net/mlx5_vnet.c > index 08ac15b17b83..59f1874648ae 100644 > --- a/drivers/vdpa/mlx5/net/mlx5_vnet.c > +++ b/drivers/vdpa/mlx5/net/mlx5_vnet.c > @@ -1556,14 +1556,14 @@ static void mlx5_cvq_kick_handler(struct work_struct *work) > { > virtio_net_ctrl_ack status = VIRTIO_NET_ERR; > struct virtio_net_ctrl_hdr ctrl; > - struct mlx5_ctrl_wq_ent *wqent; > + struct mlx5_vdpa_wq_ent *wqent; > struct mlx5_vdpa_dev *mvdev; > struct mlx5_control_vq *cvq; > struct mlx5_vdpa_net *ndev; > size_t read, write; > int err; > > - wqent = container_of(work, struct mlx5_ctrl_wq_ent, work); > + wqent = container_of(work, struct mlx5_vdpa_wq_ent, work); > mvdev = wqent->mvdev; > ndev = to_mlx5_vdpa_ndev(mvdev); > cvq = &mvdev->cvq; > @@ -1615,7 +1615,7 @@ static void mlx5_vdpa_kick_vq(struct vdpa_device *vdev, u16 idx) > struct mlx5_vdpa_dev *mvdev = to_mvdev(vdev); > struct mlx5_vdpa_net *ndev = to_mlx5_vdpa_ndev(mvdev); > struct mlx5_vdpa_virtqueue *mvq; > - struct mlx5_ctrl_wq_ent *wqent; > + struct mlx5_vdpa_wq_ent *wqent; > > if (!is_index_valid(mvdev, idx)) > return; > @@ -2466,7 +2466,7 @@ static int mlx5_vdpa_dev_add(struct vdpa_mgmt_dev *v_mdev, const char *name) > if (err) > goto err_mr; > > - mvdev->wq = create_singlethread_workqueue("mlx5_vdpa_ctrl_wq"); > + mvdev->wq = create_singlethread_workqueue("mlx5_vdpa_wq"); > if (!mvdev->wq) { > err = -ENOMEM; > goto err_res2; > -- > 2.31.1 > _______________________________________________ Virtualization mailing list Virtualization@lists.linux-foundation.org https://lists.linuxfoundation.org/mailman/listinfo/virtualization ^ permalink raw reply [flat|nested] 3+ messages in thread
[parent not found: <20210909123635.30884-4-elic@nvidia.com>]
* Re: [PATCH 3/3] vdpa/mlx5: Propagate link status from device to vdpa driver [not found] ` <20210909123635.30884-4-elic@nvidia.com> @ 2021-09-14 3:00 ` Jason Wang 0 siblings, 0 replies; 3+ messages in thread From: Jason Wang @ 2021-09-14 3:00 UTC (permalink / raw) To: Eli Cohen; +Cc: eperezma, virtualization, mst On Thu, Sep 9, 2021 at 8:36 PM Eli Cohen <elic@nvidia.com> wrote: > > Add code to register to hardware asynchronous events. Use this > mechanism to track link status events coming from the device and update > the config struct. > > After doing link status change, call the vdpa callback to notify of the > link status change. > > Signed-off-by: Eli Cohen <elic@nvidia.com> Acked-by: Jason Wang <jasowang@redhat.com> > --- > drivers/vdpa/mlx5/net/mlx5_vnet.c | 94 ++++++++++++++++++++++++++++++- > 1 file changed, 92 insertions(+), 2 deletions(-) > > diff --git a/drivers/vdpa/mlx5/net/mlx5_vnet.c b/drivers/vdpa/mlx5/net/mlx5_vnet.c > index 59f1874648ae..c2b5c62358b8 100644 > --- a/drivers/vdpa/mlx5/net/mlx5_vnet.c > +++ b/drivers/vdpa/mlx5/net/mlx5_vnet.c > @@ -160,6 +160,8 @@ struct mlx5_vdpa_net { > struct mlx5_flow_handle *rx_rule; > bool setup; > u32 cur_num_vqs; > + struct notifier_block nb; > + struct vdpa_callback config_cb; > }; > > static void free_resources(struct mlx5_vdpa_net *ndev); > @@ -1851,6 +1853,7 @@ static u64 mlx5_vdpa_get_features(struct vdpa_device *vdev) > ndev->mvdev.mlx_features |= BIT_ULL(VIRTIO_NET_F_CTRL_VQ); > ndev->mvdev.mlx_features |= BIT_ULL(VIRTIO_NET_F_CTRL_MAC_ADDR); > ndev->mvdev.mlx_features |= BIT_ULL(VIRTIO_NET_F_MQ); > + ndev->mvdev.mlx_features |= BIT_ULL(VIRTIO_NET_F_STATUS); > > print_features(mvdev, ndev->mvdev.mlx_features, false); > return ndev->mvdev.mlx_features; > @@ -1947,8 +1950,10 @@ static int mlx5_vdpa_set_features(struct vdpa_device *vdev, u64 features) > > static void mlx5_vdpa_set_config_cb(struct vdpa_device *vdev, struct vdpa_callback *cb) > { > - /* not implemented */ > - mlx5_vdpa_warn(to_mvdev(vdev), "set config callback not supported\n"); > + struct mlx5_vdpa_dev *mvdev = to_mvdev(vdev); > + struct mlx5_vdpa_net *ndev = to_mlx5_vdpa_ndev(mvdev); > + > + ndev->config_cb = *cb; > } > > #define MLX5_VDPA_MAX_VQ_ENTRIES 256 > @@ -2393,6 +2398,82 @@ struct mlx5_vdpa_mgmtdev { > struct mlx5_vdpa_net *ndev; > }; > > +static u8 query_vport_state(struct mlx5_core_dev *mdev, u8 opmod, u16 vport) > +{ > + u32 out[MLX5_ST_SZ_DW(query_vport_state_out)] = {}; > + u32 in[MLX5_ST_SZ_DW(query_vport_state_in)] = {}; > + int err; > + > + MLX5_SET(query_vport_state_in, in, opcode, MLX5_CMD_OP_QUERY_VPORT_STATE); > + MLX5_SET(query_vport_state_in, in, op_mod, opmod); > + MLX5_SET(query_vport_state_in, in, vport_number, vport); > + if (vport) > + MLX5_SET(query_vport_state_in, in, other_vport, 1); > + > + err = mlx5_cmd_exec_inout(mdev, query_vport_state, in, out); > + if (err) > + return 0; > + > + return MLX5_GET(query_vport_state_out, out, state); > +} > + > +static bool get_link_state(struct mlx5_vdpa_dev *mvdev) > +{ > + if (query_vport_state(mvdev->mdev, MLX5_VPORT_STATE_OP_MOD_VNIC_VPORT, 0) == > + VPORT_STATE_UP) > + return true; > + > + return false; > +} > + > +static void update_carrier(struct work_struct *work) > +{ > + struct mlx5_vdpa_wq_ent *wqent; > + struct mlx5_vdpa_dev *mvdev; > + struct mlx5_vdpa_net *ndev; > + > + wqent = container_of(work, struct mlx5_vdpa_wq_ent, work); > + mvdev = wqent->mvdev; > + ndev = to_mlx5_vdpa_ndev(mvdev); > + if (get_link_state(mvdev)) > + ndev->config.status |= cpu_to_mlx5vdpa16(mvdev, VIRTIO_NET_S_LINK_UP); > + else > + ndev->config.status &= cpu_to_mlx5vdpa16(mvdev, ~VIRTIO_NET_S_LINK_UP); > + > + if (ndev->config_cb.callback) > + ndev->config_cb.callback(ndev->config_cb.private); > + > + kfree(wqent); > +} > + > +static int event_handler(struct notifier_block *nb, unsigned long event, void *param) > +{ > + struct mlx5_vdpa_net *ndev = container_of(nb, struct mlx5_vdpa_net, nb); > + struct mlx5_eqe *eqe = param; > + int ret = NOTIFY_DONE; > + struct mlx5_vdpa_wq_ent *wqent; > + > + if (event == MLX5_EVENT_TYPE_PORT_CHANGE) { > + switch (eqe->sub_type) { > + case MLX5_PORT_CHANGE_SUBTYPE_DOWN: > + case MLX5_PORT_CHANGE_SUBTYPE_ACTIVE: > + wqent = kzalloc(sizeof(*wqent), GFP_ATOMIC); > + if (!wqent) > + return NOTIFY_DONE; > + > + wqent->mvdev = &ndev->mvdev; > + INIT_WORK(&wqent->work, update_carrier); > + queue_work(ndev->mvdev.wq, &wqent->work); > + ret = NOTIFY_OK; > + break; > + default: > + return NOTIFY_DONE; > + } > + return ret; > + } > + return ret; > +} > + > static int mlx5_vdpa_dev_add(struct vdpa_mgmt_dev *v_mdev, const char *name) > { > struct mlx5_vdpa_mgmtdev *mgtdev = container_of(v_mdev, struct mlx5_vdpa_mgmtdev, mgtdev); > @@ -2441,6 +2522,11 @@ static int mlx5_vdpa_dev_add(struct vdpa_mgmt_dev *v_mdev, const char *name) > if (err) > goto err_mtu; > > + if (get_link_state(mvdev)) > + ndev->config.status |= cpu_to_mlx5vdpa16(mvdev, VIRTIO_NET_S_LINK_UP); > + else > + ndev->config.status &= cpu_to_mlx5vdpa16(mvdev, ~VIRTIO_NET_S_LINK_UP); > + > if (!is_zero_ether_addr(config->mac)) { > pfmdev = pci_get_drvdata(pci_physfn(mdev->pdev)); > err = mlx5_mpfs_add_mac(pfmdev, config->mac); > @@ -2472,6 +2558,8 @@ static int mlx5_vdpa_dev_add(struct vdpa_mgmt_dev *v_mdev, const char *name) > goto err_res2; > } > > + ndev->nb.notifier_call = event_handler; > + mlx5_notifier_register(mdev, &ndev->nb); > ndev->cur_num_vqs = 2 * mlx5_vdpa_max_qps(max_vqs); > mvdev->vdev.mdev = &mgtdev->mgtdev; > err = _vdpa_register_device(&mvdev->vdev, ndev->cur_num_vqs + 1); > @@ -2502,7 +2590,9 @@ static void mlx5_vdpa_dev_del(struct vdpa_mgmt_dev *v_mdev, struct vdpa_device * > { > struct mlx5_vdpa_mgmtdev *mgtdev = container_of(v_mdev, struct mlx5_vdpa_mgmtdev, mgtdev); > struct mlx5_vdpa_dev *mvdev = to_mvdev(dev); > + struct mlx5_vdpa_net *ndev = to_mlx5_vdpa_ndev(mvdev); > > + mlx5_notifier_unregister(mvdev->mdev, &ndev->nb); > destroy_workqueue(mvdev->wq); > _vdpa_unregister_device(dev); > mgtdev->ndev = NULL; > -- > 2.31.1 > _______________________________________________ Virtualization mailing list Virtualization@lists.linux-foundation.org https://lists.linuxfoundation.org/mailman/listinfo/virtualization ^ permalink raw reply [flat|nested] 3+ messages in thread
end of thread, other threads:[~2021-09-14 3:00 UTC | newest]
Thread overview: 3+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
[not found] <20210909123635.30884-1-elic@nvidia.com>
[not found] ` <20210909123635.30884-2-elic@nvidia.com>
2021-09-13 22:09 ` [PATCH 1/3] vdpa/mlx5: Remove mtu field from vdpa net device Michael S. Tsirkin
[not found] ` <20210909123635.30884-3-elic@nvidia.com>
2021-09-14 2:57 ` [PATCH 2/3] vdpa/mlx5: Rename control VQ workqueue to vdpa wq Jason Wang
[not found] ` <20210909123635.30884-4-elic@nvidia.com>
2021-09-14 3:00 ` [PATCH 3/3] vdpa/mlx5: Propagate link status from device to vdpa driver Jason Wang
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).