| // SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause) | 
 | /* | 
 |  * Copyright(c) 2018 Intel Corporation. | 
 |  * | 
 |  */ | 
 | #include "hfi.h" | 
 | #include "trace.h" | 
 | #include "qp.h" | 
 | #include "opfn.h" | 
 |  | 
 | #define IB_BTHE_E                 BIT(IB_BTHE_E_SHIFT) | 
 |  | 
 | #define OPFN_CODE(code) BIT((code) - 1) | 
 | #define OPFN_MASK(code) OPFN_CODE(STL_VERBS_EXTD_##code) | 
 |  | 
 | struct hfi1_opfn_type { | 
 | 	bool (*request)(struct rvt_qp *qp, u64 *data); | 
 | 	bool (*response)(struct rvt_qp *qp, u64 *data); | 
 | 	bool (*reply)(struct rvt_qp *qp, u64 data); | 
 | 	void (*error)(struct rvt_qp *qp); | 
 | }; | 
 |  | 
 | static struct hfi1_opfn_type hfi1_opfn_handlers[STL_VERBS_EXTD_MAX] = { | 
 | 	[STL_VERBS_EXTD_TID_RDMA] = { | 
 | 		.request = tid_rdma_conn_req, | 
 | 		.response = tid_rdma_conn_resp, | 
 | 		.reply = tid_rdma_conn_reply, | 
 | 		.error = tid_rdma_conn_error, | 
 | 	}, | 
 | }; | 
 |  | 
 | static struct workqueue_struct *opfn_wq; | 
 |  | 
 | static void opfn_schedule_conn_request(struct rvt_qp *qp); | 
 |  | 
 | static bool hfi1_opfn_extended(u32 bth1) | 
 | { | 
 | 	return !!(bth1 & IB_BTHE_E); | 
 | } | 
 |  | 
 | static void opfn_conn_request(struct rvt_qp *qp) | 
 | { | 
 | 	struct hfi1_qp_priv *priv = qp->priv; | 
 | 	struct ib_atomic_wr wr; | 
 | 	u16 mask, capcode; | 
 | 	struct hfi1_opfn_type *extd; | 
 | 	u64 data; | 
 | 	unsigned long flags; | 
 | 	int ret = 0; | 
 |  | 
 | 	trace_hfi1_opfn_state_conn_request(qp); | 
 | 	spin_lock_irqsave(&priv->opfn.lock, flags); | 
 | 	/* | 
 | 	 * Exit if the extended bit is not set, or if nothing is requested, or | 
 | 	 * if we have completed all requests, or if a previous request is in | 
 | 	 * progress | 
 | 	 */ | 
 | 	if (!priv->opfn.extended || !priv->opfn.requested || | 
 | 	    priv->opfn.requested == priv->opfn.completed || priv->opfn.curr) | 
 | 		goto done; | 
 |  | 
 | 	mask = priv->opfn.requested & ~priv->opfn.completed; | 
 | 	capcode = ilog2(mask & ~(mask - 1)) + 1; | 
 | 	if (capcode >= STL_VERBS_EXTD_MAX) { | 
 | 		priv->opfn.completed |= OPFN_CODE(capcode); | 
 | 		goto done; | 
 | 	} | 
 |  | 
 | 	extd = &hfi1_opfn_handlers[capcode]; | 
 | 	if (!extd || !extd->request || !extd->request(qp, &data)) { | 
 | 		/* | 
 | 		 * Either there is no handler for this capability or the request | 
 | 		 * packet could not be generated. Either way, mark it as done so | 
 | 		 * we don't keep attempting to complete it. | 
 | 		 */ | 
 | 		priv->opfn.completed |= OPFN_CODE(capcode); | 
 | 		goto done; | 
 | 	} | 
 |  | 
 | 	trace_hfi1_opfn_data_conn_request(qp, capcode, data); | 
 | 	data = (data & ~0xf) | capcode; | 
 |  | 
 | 	memset(&wr, 0, sizeof(wr)); | 
 | 	wr.wr.opcode = IB_WR_OPFN; | 
 | 	wr.remote_addr = HFI1_VERBS_E_ATOMIC_VADDR; | 
 | 	wr.compare_add = data; | 
 |  | 
 | 	priv->opfn.curr = capcode;	/* A new request is now in progress */ | 
 | 	/* Drop opfn.lock before calling ib_post_send() */ | 
 | 	spin_unlock_irqrestore(&priv->opfn.lock, flags); | 
 |  | 
 | 	ret = ib_post_send(&qp->ibqp, &wr.wr, NULL); | 
 | 	if (ret) | 
 | 		goto err; | 
 | 	trace_hfi1_opfn_state_conn_request(qp); | 
 | 	return; | 
 | err: | 
 | 	trace_hfi1_msg_opfn_conn_request(qp, "ib_ost_send failed: ret = ", | 
 | 					 (u64)ret); | 
 | 	spin_lock_irqsave(&priv->opfn.lock, flags); | 
 | 	/* | 
 | 	 * In case of an unexpected error return from ib_post_send | 
 | 	 * clear opfn.curr and reschedule to try again | 
 | 	 */ | 
 | 	priv->opfn.curr = STL_VERBS_EXTD_NONE; | 
 | 	opfn_schedule_conn_request(qp); | 
 | done: | 
 | 	spin_unlock_irqrestore(&priv->opfn.lock, flags); | 
 | } | 
 |  | 
 | void opfn_send_conn_request(struct work_struct *work) | 
 | { | 
 | 	struct hfi1_opfn_data *od; | 
 | 	struct hfi1_qp_priv *qpriv; | 
 |  | 
 | 	od = container_of(work, struct hfi1_opfn_data, opfn_work); | 
 | 	qpriv = container_of(od, struct hfi1_qp_priv, opfn); | 
 |  | 
 | 	opfn_conn_request(qpriv->owner); | 
 | } | 
 |  | 
 | /* | 
 |  * When QP s_lock is held in the caller, the OPFN request must be scheduled | 
 |  * to a different workqueue to avoid double locking QP s_lock in call to | 
 |  * ib_post_send in opfn_conn_request | 
 |  */ | 
 | static void opfn_schedule_conn_request(struct rvt_qp *qp) | 
 | { | 
 | 	struct hfi1_qp_priv *priv = qp->priv; | 
 |  | 
 | 	trace_hfi1_opfn_state_sched_conn_request(qp); | 
 | 	queue_work(opfn_wq, &priv->opfn.opfn_work); | 
 | } | 
 |  | 
 | void opfn_conn_response(struct rvt_qp *qp, struct rvt_ack_entry *e, | 
 | 			struct ib_atomic_eth *ateth) | 
 | { | 
 | 	struct hfi1_qp_priv *priv = qp->priv; | 
 | 	u64 data = be64_to_cpu(ateth->compare_data); | 
 | 	struct hfi1_opfn_type *extd; | 
 | 	u8 capcode; | 
 | 	unsigned long flags; | 
 |  | 
 | 	trace_hfi1_opfn_state_conn_response(qp); | 
 | 	capcode = data & 0xf; | 
 | 	trace_hfi1_opfn_data_conn_response(qp, capcode, data); | 
 | 	if (!capcode || capcode >= STL_VERBS_EXTD_MAX) | 
 | 		return; | 
 |  | 
 | 	extd = &hfi1_opfn_handlers[capcode]; | 
 |  | 
 | 	if (!extd || !extd->response) { | 
 | 		e->atomic_data = capcode; | 
 | 		return; | 
 | 	} | 
 |  | 
 | 	spin_lock_irqsave(&priv->opfn.lock, flags); | 
 | 	if (priv->opfn.completed & OPFN_CODE(capcode)) { | 
 | 		/* | 
 | 		 * We are receiving a request for a feature that has already | 
 | 		 * been negotiated. This may mean that the other side has reset | 
 | 		 */ | 
 | 		priv->opfn.completed &= ~OPFN_CODE(capcode); | 
 | 		if (extd->error) | 
 | 			extd->error(qp); | 
 | 	} | 
 |  | 
 | 	if (extd->response(qp, &data)) | 
 | 		priv->opfn.completed |= OPFN_CODE(capcode); | 
 | 	e->atomic_data = (data & ~0xf) | capcode; | 
 | 	trace_hfi1_opfn_state_conn_response(qp); | 
 | 	spin_unlock_irqrestore(&priv->opfn.lock, flags); | 
 | } | 
 |  | 
 | void opfn_conn_reply(struct rvt_qp *qp, u64 data) | 
 | { | 
 | 	struct hfi1_qp_priv *priv = qp->priv; | 
 | 	struct hfi1_opfn_type *extd; | 
 | 	u8 capcode; | 
 | 	unsigned long flags; | 
 |  | 
 | 	trace_hfi1_opfn_state_conn_reply(qp); | 
 | 	capcode = data & 0xf; | 
 | 	trace_hfi1_opfn_data_conn_reply(qp, capcode, data); | 
 | 	if (!capcode || capcode >= STL_VERBS_EXTD_MAX) | 
 | 		return; | 
 |  | 
 | 	spin_lock_irqsave(&priv->opfn.lock, flags); | 
 | 	/* | 
 | 	 * Either there is no previous request or the reply is not for the | 
 | 	 * current request | 
 | 	 */ | 
 | 	if (!priv->opfn.curr || capcode != priv->opfn.curr) | 
 | 		goto done; | 
 |  | 
 | 	extd = &hfi1_opfn_handlers[capcode]; | 
 |  | 
 | 	if (!extd || !extd->reply) | 
 | 		goto clear; | 
 |  | 
 | 	if (extd->reply(qp, data)) | 
 | 		priv->opfn.completed |= OPFN_CODE(capcode); | 
 | clear: | 
 | 	/* | 
 | 	 * Clear opfn.curr to indicate that the previous request is no longer in | 
 | 	 * progress | 
 | 	 */ | 
 | 	priv->opfn.curr = STL_VERBS_EXTD_NONE; | 
 | 	trace_hfi1_opfn_state_conn_reply(qp); | 
 | done: | 
 | 	spin_unlock_irqrestore(&priv->opfn.lock, flags); | 
 | } | 
 |  | 
 | void opfn_conn_error(struct rvt_qp *qp) | 
 | { | 
 | 	struct hfi1_qp_priv *priv = qp->priv; | 
 | 	struct hfi1_opfn_type *extd = NULL; | 
 | 	unsigned long flags; | 
 | 	u16 capcode; | 
 |  | 
 | 	trace_hfi1_opfn_state_conn_error(qp); | 
 | 	trace_hfi1_msg_opfn_conn_error(qp, "error. qp state ", (u64)qp->state); | 
 | 	/* | 
 | 	 * The QP has gone into the Error state. We have to invalidate all | 
 | 	 * negotiated feature, including the one in progress (if any). The RC | 
 | 	 * QP handling will clean the WQE for the connection request. | 
 | 	 */ | 
 | 	spin_lock_irqsave(&priv->opfn.lock, flags); | 
 | 	while (priv->opfn.completed) { | 
 | 		capcode = priv->opfn.completed & ~(priv->opfn.completed - 1); | 
 | 		extd = &hfi1_opfn_handlers[ilog2(capcode) + 1]; | 
 | 		if (extd->error) | 
 | 			extd->error(qp); | 
 | 		priv->opfn.completed &= ~OPFN_CODE(capcode); | 
 | 	} | 
 | 	priv->opfn.extended = 0; | 
 | 	priv->opfn.requested = 0; | 
 | 	priv->opfn.curr = STL_VERBS_EXTD_NONE; | 
 | 	spin_unlock_irqrestore(&priv->opfn.lock, flags); | 
 | } | 
 |  | 
 | void opfn_qp_init(struct rvt_qp *qp, struct ib_qp_attr *attr, int attr_mask) | 
 | { | 
 | 	struct ib_qp *ibqp = &qp->ibqp; | 
 | 	struct hfi1_qp_priv *priv = qp->priv; | 
 | 	unsigned long flags; | 
 |  | 
 | 	if (attr_mask & IB_QP_RETRY_CNT) | 
 | 		priv->s_retry = attr->retry_cnt; | 
 |  | 
 | 	spin_lock_irqsave(&priv->opfn.lock, flags); | 
 | 	if (ibqp->qp_type == IB_QPT_RC && HFI1_CAP_IS_KSET(TID_RDMA)) { | 
 | 		struct tid_rdma_params *local = &priv->tid_rdma.local; | 
 |  | 
 | 		if (attr_mask & IB_QP_TIMEOUT) | 
 | 			priv->tid_retry_timeout_jiffies = qp->timeout_jiffies; | 
 | 		if (qp->pmtu == enum_to_mtu(OPA_MTU_4096) || | 
 | 		    qp->pmtu == enum_to_mtu(OPA_MTU_8192)) { | 
 | 			tid_rdma_opfn_init(qp, local); | 
 | 			/* | 
 | 			 * We only want to set the OPFN requested bit when the | 
 | 			 * QP transitions to RTS. | 
 | 			 */ | 
 | 			if (attr_mask & IB_QP_STATE && | 
 | 			    attr->qp_state == IB_QPS_RTS) { | 
 | 				priv->opfn.requested |= OPFN_MASK(TID_RDMA); | 
 | 				/* | 
 | 				 * If the QP is transitioning to RTS and the | 
 | 				 * opfn.completed for TID RDMA has already been | 
 | 				 * set, the QP is being moved *back* into RTS. | 
 | 				 * We can now renegotiate the TID RDMA | 
 | 				 * parameters. | 
 | 				 */ | 
 | 				if (priv->opfn.completed & | 
 | 				    OPFN_MASK(TID_RDMA)) { | 
 | 					priv->opfn.completed &= | 
 | 						~OPFN_MASK(TID_RDMA); | 
 | 					/* | 
 | 					 * Since the opfn.completed bit was | 
 | 					 * already set, it is safe to assume | 
 | 					 * that the opfn.extended is also set. | 
 | 					 */ | 
 | 					opfn_schedule_conn_request(qp); | 
 | 				} | 
 | 			} | 
 | 		} else { | 
 | 			memset(local, 0, sizeof(*local)); | 
 | 		} | 
 | 	} | 
 | 	spin_unlock_irqrestore(&priv->opfn.lock, flags); | 
 | } | 
 |  | 
 | void opfn_trigger_conn_request(struct rvt_qp *qp, u32 bth1) | 
 | { | 
 | 	struct hfi1_qp_priv *priv = qp->priv; | 
 |  | 
 | 	if (!priv->opfn.extended && hfi1_opfn_extended(bth1) && | 
 | 	    HFI1_CAP_IS_KSET(OPFN)) { | 
 | 		priv->opfn.extended = 1; | 
 | 		if (qp->state == IB_QPS_RTS) | 
 | 			opfn_conn_request(qp); | 
 | 	} | 
 | } | 
 |  | 
 | int opfn_init(void) | 
 | { | 
 | 	opfn_wq = alloc_workqueue("hfi_opfn", | 
 | 				  WQ_SYSFS | WQ_HIGHPRI | WQ_CPU_INTENSIVE | | 
 | 				  WQ_MEM_RECLAIM, | 
 | 				  HFI1_MAX_ACTIVE_WORKQUEUE_ENTRIES); | 
 | 	if (!opfn_wq) | 
 | 		return -ENOMEM; | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | void opfn_exit(void) | 
 | { | 
 | 	if (opfn_wq) { | 
 | 		destroy_workqueue(opfn_wq); | 
 | 		opfn_wq = NULL; | 
 | 	} | 
 | } |