diff options
Diffstat (limited to 'drivers/infiniband/hw/mana/qp.c')
| -rw-r--r-- | drivers/infiniband/hw/mana/qp.c | 506 | 
1 files changed, 506 insertions, 0 deletions
diff --git a/drivers/infiniband/hw/mana/qp.c b/drivers/infiniband/hw/mana/qp.c new file mode 100644 index 000000000000..54b61930a7fd --- /dev/null +++ b/drivers/infiniband/hw/mana/qp.c @@ -0,0 +1,506 @@ +// SPDX-License-Identifier: GPL-2.0-only +/* + * Copyright (c) 2022, Microsoft Corporation. All rights reserved. + */ + +#include "mana_ib.h" + +static int mana_ib_cfg_vport_steering(struct mana_ib_dev *dev, +				      struct net_device *ndev, +				      mana_handle_t default_rxobj, +				      mana_handle_t ind_table[], +				      u32 log_ind_tbl_size, u32 rx_hash_key_len, +				      u8 *rx_hash_key) +{ +	struct mana_port_context *mpc = netdev_priv(ndev); +	struct mana_cfg_rx_steer_req *req = NULL; +	struct mana_cfg_rx_steer_resp resp = {}; +	mana_handle_t *req_indir_tab; +	struct gdma_context *gc; +	struct gdma_dev *mdev; +	u32 req_buf_size; +	int i, err; + +	mdev = dev->gdma_dev; +	gc = mdev->gdma_context; + +	req_buf_size = +		sizeof(*req) + sizeof(mana_handle_t) * MANA_INDIRECT_TABLE_SIZE; +	req = kzalloc(req_buf_size, GFP_KERNEL); +	if (!req) +		return -ENOMEM; + +	mana_gd_init_req_hdr(&req->hdr, MANA_CONFIG_VPORT_RX, req_buf_size, +			     sizeof(resp)); + +	req->vport = mpc->port_handle; +	req->rx_enable = 1; +	req->update_default_rxobj = 1; +	req->default_rxobj = default_rxobj; +	req->hdr.dev_id = mdev->dev_id; + +	/* If there are more than 1 entries in indirection table, enable RSS */ +	if (log_ind_tbl_size) +		req->rss_enable = true; + +	req->num_indir_entries = MANA_INDIRECT_TABLE_SIZE; +	req->indir_tab_offset = sizeof(*req); +	req->update_indir_tab = true; + +	req_indir_tab = (mana_handle_t *)(req + 1); +	/* The ind table passed to the hardware must have +	 * MANA_INDIRECT_TABLE_SIZE entries. Adjust the verb +	 * ind_table to MANA_INDIRECT_TABLE_SIZE if required +	 */ +	ibdev_dbg(&dev->ib_dev, "ind table size %u\n", 1 << log_ind_tbl_size); +	for (i = 0; i < MANA_INDIRECT_TABLE_SIZE; i++) { +		req_indir_tab[i] = ind_table[i % (1 << log_ind_tbl_size)]; +		ibdev_dbg(&dev->ib_dev, "index %u handle 0x%llx\n", i, +			  req_indir_tab[i]); +	} + +	req->update_hashkey = true; +	if (rx_hash_key_len) +		memcpy(req->hashkey, rx_hash_key, rx_hash_key_len); +	else +		netdev_rss_key_fill(req->hashkey, MANA_HASH_KEY_SIZE); + +	ibdev_dbg(&dev->ib_dev, "vport handle %llu default_rxobj 0x%llx\n", +		  req->vport, default_rxobj); + +	err = mana_gd_send_request(gc, req_buf_size, req, sizeof(resp), &resp); +	if (err) { +		netdev_err(ndev, "Failed to configure vPort RX: %d\n", err); +		goto out; +	} + +	if (resp.hdr.status) { +		netdev_err(ndev, "vPort RX configuration failed: 0x%x\n", +			   resp.hdr.status); +		err = -EPROTO; +		goto out; +	} + +	netdev_info(ndev, "Configured steering vPort %llu log_entries %u\n", +		    mpc->port_handle, log_ind_tbl_size); + +out: +	kfree(req); +	return err; +} + +static int mana_ib_create_qp_rss(struct ib_qp *ibqp, struct ib_pd *pd, +				 struct ib_qp_init_attr *attr, +				 struct ib_udata *udata) +{ +	struct mana_ib_qp *qp = container_of(ibqp, struct mana_ib_qp, ibqp); +	struct mana_ib_dev *mdev = +		container_of(pd->device, struct mana_ib_dev, ib_dev); +	struct ib_rwq_ind_table *ind_tbl = attr->rwq_ind_tbl; +	struct mana_ib_create_qp_rss_resp resp = {}; +	struct mana_ib_create_qp_rss ucmd = {}; +	struct gdma_dev *gd = mdev->gdma_dev; +	mana_handle_t *mana_ind_table; +	struct mana_port_context *mpc; +	struct mana_context *mc; +	struct net_device *ndev; +	struct mana_ib_cq *cq; +	struct mana_ib_wq *wq; +	unsigned int ind_tbl_size; +	struct ib_cq *ibcq; +	struct ib_wq *ibwq; +	int i = 0; +	u32 port; +	int ret; + +	mc = gd->driver_data; + +	if (!udata || udata->inlen < sizeof(ucmd)) +		return -EINVAL; + +	ret = ib_copy_from_udata(&ucmd, udata, min(sizeof(ucmd), udata->inlen)); +	if (ret) { +		ibdev_dbg(&mdev->ib_dev, +			  "Failed copy from udata for create rss-qp, err %d\n", +			  ret); +		return ret; +	} + +	if (attr->cap.max_recv_wr > MAX_SEND_BUFFERS_PER_QUEUE) { +		ibdev_dbg(&mdev->ib_dev, +			  "Requested max_recv_wr %d exceeding limit\n", +			  attr->cap.max_recv_wr); +		return -EINVAL; +	} + +	if (attr->cap.max_recv_sge > MAX_RX_WQE_SGL_ENTRIES) { +		ibdev_dbg(&mdev->ib_dev, +			  "Requested max_recv_sge %d exceeding limit\n", +			  attr->cap.max_recv_sge); +		return -EINVAL; +	} + +	ind_tbl_size = 1 << ind_tbl->log_ind_tbl_size; +	if (ind_tbl_size > MANA_INDIRECT_TABLE_SIZE) { +		ibdev_dbg(&mdev->ib_dev, +			  "Indirect table size %d exceeding limit\n", +			  ind_tbl_size); +		return -EINVAL; +	} + +	if (ucmd.rx_hash_function != MANA_IB_RX_HASH_FUNC_TOEPLITZ) { +		ibdev_dbg(&mdev->ib_dev, +			  "RX Hash function is not supported, %d\n", +			  ucmd.rx_hash_function); +		return -EINVAL; +	} + +	/* IB ports start with 1, MANA start with 0 */ +	port = ucmd.port; +	if (port < 1 || port > mc->num_ports) { +		ibdev_dbg(&mdev->ib_dev, "Invalid port %u in creating qp\n", +			  port); +		return -EINVAL; +	} +	ndev = mc->ports[port - 1]; +	mpc = netdev_priv(ndev); + +	ibdev_dbg(&mdev->ib_dev, "rx_hash_function %d port %d\n", +		  ucmd.rx_hash_function, port); + +	mana_ind_table = kcalloc(ind_tbl_size, sizeof(mana_handle_t), +				 GFP_KERNEL); +	if (!mana_ind_table) { +		ret = -ENOMEM; +		goto fail; +	} + +	qp->port = port; + +	for (i = 0; i < ind_tbl_size; i++) { +		struct mana_obj_spec wq_spec = {}; +		struct mana_obj_spec cq_spec = {}; + +		ibwq = ind_tbl->ind_tbl[i]; +		wq = container_of(ibwq, struct mana_ib_wq, ibwq); + +		ibcq = ibwq->cq; +		cq = container_of(ibcq, struct mana_ib_cq, ibcq); + +		wq_spec.gdma_region = wq->gdma_region; +		wq_spec.queue_size = wq->wq_buf_size; + +		cq_spec.gdma_region = cq->gdma_region; +		cq_spec.queue_size = cq->cqe * COMP_ENTRY_SIZE; +		cq_spec.modr_ctx_id = 0; +		cq_spec.attached_eq = GDMA_CQ_NO_EQ; + +		ret = mana_create_wq_obj(mpc, mpc->port_handle, GDMA_RQ, +					 &wq_spec, &cq_spec, &wq->rx_object); +		if (ret) +			goto fail; + +		/* The GDMA regions are now owned by the WQ object */ +		wq->gdma_region = GDMA_INVALID_DMA_REGION; +		cq->gdma_region = GDMA_INVALID_DMA_REGION; + +		wq->id = wq_spec.queue_index; +		cq->id = cq_spec.queue_index; + +		ibdev_dbg(&mdev->ib_dev, +			  "ret %d rx_object 0x%llx wq id %llu cq id %llu\n", +			  ret, wq->rx_object, wq->id, cq->id); + +		resp.entries[i].cqid = cq->id; +		resp.entries[i].wqid = wq->id; + +		mana_ind_table[i] = wq->rx_object; +	} +	resp.num_entries = i; + +	ret = mana_ib_cfg_vport_steering(mdev, ndev, wq->rx_object, +					 mana_ind_table, +					 ind_tbl->log_ind_tbl_size, +					 ucmd.rx_hash_key_len, +					 ucmd.rx_hash_key); +	if (ret) +		goto fail; + +	ret = ib_copy_to_udata(udata, &resp, sizeof(resp)); +	if (ret) { +		ibdev_dbg(&mdev->ib_dev, +			  "Failed to copy to udata create rss-qp, %d\n", +			  ret); +		goto fail; +	} + +	kfree(mana_ind_table); + +	return 0; + +fail: +	while (i-- > 0) { +		ibwq = ind_tbl->ind_tbl[i]; +		wq = container_of(ibwq, struct mana_ib_wq, ibwq); +		mana_destroy_wq_obj(mpc, GDMA_RQ, wq->rx_object); +	} + +	kfree(mana_ind_table); + +	return ret; +} + +static int mana_ib_create_qp_raw(struct ib_qp *ibqp, struct ib_pd *ibpd, +				 struct ib_qp_init_attr *attr, +				 struct ib_udata *udata) +{ +	struct mana_ib_pd *pd = container_of(ibpd, struct mana_ib_pd, ibpd); +	struct mana_ib_qp *qp = container_of(ibqp, struct mana_ib_qp, ibqp); +	struct mana_ib_dev *mdev = +		container_of(ibpd->device, struct mana_ib_dev, ib_dev); +	struct mana_ib_cq *send_cq = +		container_of(attr->send_cq, struct mana_ib_cq, ibcq); +	struct mana_ib_ucontext *mana_ucontext = +		rdma_udata_to_drv_context(udata, struct mana_ib_ucontext, +					  ibucontext); +	struct mana_ib_create_qp_resp resp = {}; +	struct gdma_dev *gd = mdev->gdma_dev; +	struct mana_ib_create_qp ucmd = {}; +	struct mana_obj_spec wq_spec = {}; +	struct mana_obj_spec cq_spec = {}; +	struct mana_port_context *mpc; +	struct mana_context *mc; +	struct net_device *ndev; +	struct ib_umem *umem; +	int err; +	u32 port; + +	mc = gd->driver_data; + +	if (!mana_ucontext || udata->inlen < sizeof(ucmd)) +		return -EINVAL; + +	err = ib_copy_from_udata(&ucmd, udata, min(sizeof(ucmd), udata->inlen)); +	if (err) { +		ibdev_dbg(&mdev->ib_dev, +			  "Failed to copy from udata create qp-raw, %d\n", err); +		return err; +	} + +	/* IB ports start with 1, MANA Ethernet ports start with 0 */ +	port = ucmd.port; +	if (port < 1 || port > mc->num_ports) +		return -EINVAL; + +	if (attr->cap.max_send_wr > MAX_SEND_BUFFERS_PER_QUEUE) { +		ibdev_dbg(&mdev->ib_dev, +			  "Requested max_send_wr %d exceeding limit\n", +			  attr->cap.max_send_wr); +		return -EINVAL; +	} + +	if (attr->cap.max_send_sge > MAX_TX_WQE_SGL_ENTRIES) { +		ibdev_dbg(&mdev->ib_dev, +			  "Requested max_send_sge %d exceeding limit\n", +			  attr->cap.max_send_sge); +		return -EINVAL; +	} + +	ndev = mc->ports[port - 1]; +	mpc = netdev_priv(ndev); +	ibdev_dbg(&mdev->ib_dev, "port %u ndev %p mpc %p\n", port, ndev, mpc); + +	err = mana_ib_cfg_vport(mdev, port - 1, pd, mana_ucontext->doorbell); +	if (err) +		return -ENODEV; + +	qp->port = port; + +	ibdev_dbg(&mdev->ib_dev, "ucmd sq_buf_addr 0x%llx port %u\n", +		  ucmd.sq_buf_addr, ucmd.port); + +	umem = ib_umem_get(ibpd->device, ucmd.sq_buf_addr, ucmd.sq_buf_size, +			   IB_ACCESS_LOCAL_WRITE); +	if (IS_ERR(umem)) { +		err = PTR_ERR(umem); +		ibdev_dbg(&mdev->ib_dev, +			  "Failed to get umem for create qp-raw, err %d\n", +			  err); +		goto err_free_vport; +	} +	qp->sq_umem = umem; + +	err = mana_ib_gd_create_dma_region(mdev, qp->sq_umem, +					   &qp->sq_gdma_region); +	if (err) { +		ibdev_dbg(&mdev->ib_dev, +			  "Failed to create dma region for create qp-raw, %d\n", +			  err); +		goto err_release_umem; +	} + +	ibdev_dbg(&mdev->ib_dev, +		  "mana_ib_gd_create_dma_region ret %d gdma_region 0x%llx\n", +		  err, qp->sq_gdma_region); + +	/* Create a WQ on the same port handle used by the Ethernet */ +	wq_spec.gdma_region = qp->sq_gdma_region; +	wq_spec.queue_size = ucmd.sq_buf_size; + +	cq_spec.gdma_region = send_cq->gdma_region; +	cq_spec.queue_size = send_cq->cqe * COMP_ENTRY_SIZE; +	cq_spec.modr_ctx_id = 0; +	cq_spec.attached_eq = GDMA_CQ_NO_EQ; + +	err = mana_create_wq_obj(mpc, mpc->port_handle, GDMA_SQ, &wq_spec, +				 &cq_spec, &qp->tx_object); +	if (err) { +		ibdev_dbg(&mdev->ib_dev, +			  "Failed to create wq for create raw-qp, err %d\n", +			  err); +		goto err_destroy_dma_region; +	} + +	/* The GDMA regions are now owned by the WQ object */ +	qp->sq_gdma_region = GDMA_INVALID_DMA_REGION; +	send_cq->gdma_region = GDMA_INVALID_DMA_REGION; + +	qp->sq_id = wq_spec.queue_index; +	send_cq->id = cq_spec.queue_index; + +	ibdev_dbg(&mdev->ib_dev, +		  "ret %d qp->tx_object 0x%llx sq id %llu cq id %llu\n", err, +		  qp->tx_object, qp->sq_id, send_cq->id); + +	resp.sqid = qp->sq_id; +	resp.cqid = send_cq->id; +	resp.tx_vp_offset = pd->tx_vp_offset; + +	err = ib_copy_to_udata(udata, &resp, sizeof(resp)); +	if (err) { +		ibdev_dbg(&mdev->ib_dev, +			  "Failed copy udata for create qp-raw, %d\n", +			  err); +		goto err_destroy_wq_obj; +	} + +	return 0; + +err_destroy_wq_obj: +	mana_destroy_wq_obj(mpc, GDMA_SQ, qp->tx_object); + +err_destroy_dma_region: +	mana_ib_gd_destroy_dma_region(mdev, qp->sq_gdma_region); + +err_release_umem: +	ib_umem_release(umem); + +err_free_vport: +	mana_ib_uncfg_vport(mdev, pd, port - 1); + +	return err; +} + +int mana_ib_create_qp(struct ib_qp *ibqp, struct ib_qp_init_attr *attr, +		      struct ib_udata *udata) +{ +	switch (attr->qp_type) { +	case IB_QPT_RAW_PACKET: +		/* When rwq_ind_tbl is used, it's for creating WQs for RSS */ +		if (attr->rwq_ind_tbl) +			return mana_ib_create_qp_rss(ibqp, ibqp->pd, attr, +						     udata); + +		return mana_ib_create_qp_raw(ibqp, ibqp->pd, attr, udata); +	default: +		/* Creating QP other than IB_QPT_RAW_PACKET is not supported */ +		ibdev_dbg(ibqp->device, "Creating QP type %u not supported\n", +			  attr->qp_type); +	} + +	return -EINVAL; +} + +int mana_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, +		      int attr_mask, struct ib_udata *udata) +{ +	/* modify_qp is not supported by this version of the driver */ +	return -EOPNOTSUPP; +} + +static int mana_ib_destroy_qp_rss(struct mana_ib_qp *qp, +				  struct ib_rwq_ind_table *ind_tbl, +				  struct ib_udata *udata) +{ +	struct mana_ib_dev *mdev = +		container_of(qp->ibqp.device, struct mana_ib_dev, ib_dev); +	struct gdma_dev *gd = mdev->gdma_dev; +	struct mana_port_context *mpc; +	struct mana_context *mc; +	struct net_device *ndev; +	struct mana_ib_wq *wq; +	struct ib_wq *ibwq; +	int i; + +	mc = gd->driver_data; +	ndev = mc->ports[qp->port - 1]; +	mpc = netdev_priv(ndev); + +	for (i = 0; i < (1 << ind_tbl->log_ind_tbl_size); i++) { +		ibwq = ind_tbl->ind_tbl[i]; +		wq = container_of(ibwq, struct mana_ib_wq, ibwq); +		ibdev_dbg(&mdev->ib_dev, "destroying wq->rx_object %llu\n", +			  wq->rx_object); +		mana_destroy_wq_obj(mpc, GDMA_RQ, wq->rx_object); +	} + +	return 0; +} + +static int mana_ib_destroy_qp_raw(struct mana_ib_qp *qp, struct ib_udata *udata) +{ +	struct mana_ib_dev *mdev = +		container_of(qp->ibqp.device, struct mana_ib_dev, ib_dev); +	struct gdma_dev *gd = mdev->gdma_dev; +	struct ib_pd *ibpd = qp->ibqp.pd; +	struct mana_port_context *mpc; +	struct mana_context *mc; +	struct net_device *ndev; +	struct mana_ib_pd *pd; + +	mc = gd->driver_data; +	ndev = mc->ports[qp->port - 1]; +	mpc = netdev_priv(ndev); +	pd = container_of(ibpd, struct mana_ib_pd, ibpd); + +	mana_destroy_wq_obj(mpc, GDMA_SQ, qp->tx_object); + +	if (qp->sq_umem) { +		mana_ib_gd_destroy_dma_region(mdev, qp->sq_gdma_region); +		ib_umem_release(qp->sq_umem); +	} + +	mana_ib_uncfg_vport(mdev, pd, qp->port - 1); + +	return 0; +} + +int mana_ib_destroy_qp(struct ib_qp *ibqp, struct ib_udata *udata) +{ +	struct mana_ib_qp *qp = container_of(ibqp, struct mana_ib_qp, ibqp); + +	switch (ibqp->qp_type) { +	case IB_QPT_RAW_PACKET: +		if (ibqp->rwq_ind_tbl) +			return mana_ib_destroy_qp_rss(qp, ibqp->rwq_ind_tbl, +						      udata); + +		return mana_ib_destroy_qp_raw(qp, udata); + +	default: +		ibdev_dbg(ibqp->device, "Unexpected QP type %u\n", +			  ibqp->qp_type); +	} + +	return -ENOENT; +}  |