diff options
Diffstat (limited to 'drivers/infiniband/core/cm.c')
| -rw-r--r-- | drivers/infiniband/core/cm.c | 165 | 
1 files changed, 124 insertions, 41 deletions
diff --git a/drivers/infiniband/core/cm.c b/drivers/infiniband/core/cm.c index 64e0903091a..c3239170d8b 100644 --- a/drivers/infiniband/core/cm.c +++ b/drivers/infiniband/core/cm.c @@ -36,6 +36,7 @@  #include <linux/completion.h>  #include <linux/dma-mapping.h>  #include <linux/device.h> +#include <linux/module.h>  #include <linux/err.h>  #include <linux/idr.h>  #include <linux/interrupt.h> @@ -46,6 +47,7 @@  #include <linux/sysfs.h>  #include <linux/workqueue.h>  #include <linux/kdev_t.h> +#include <linux/etherdevice.h>  #include <rdma/ib_cache.h>  #include <rdma/ib_cm.h> @@ -176,6 +178,8 @@ struct cm_av {  	struct ib_ah_attr ah_attr;  	u16 pkey_index;  	u8 timeout; +	u8  valid; +	u8  smac[ETH_ALEN];  };  struct cm_work { @@ -375,26 +379,27 @@ static int cm_init_av_by_path(struct ib_sa_path_rec *path, struct cm_av *av)  	ib_init_ah_from_path(cm_dev->ib_device, port->port_num, path,  			     &av->ah_attr);  	av->timeout = path->packet_life_time + 1; +	memcpy(av->smac, path->smac, sizeof(av->smac)); + +	av->valid = 1;  	return 0;  }  static int cm_alloc_id(struct cm_id_private *cm_id_priv)  {  	unsigned long flags; -	int ret, id; -	static int next_id; +	int id; -	do { -		spin_lock_irqsave(&cm.lock, flags); -		ret = idr_get_new_above(&cm.local_id_table, cm_id_priv, -					next_id, &id); -		if (!ret) -			next_id = ((unsigned) id + 1) & MAX_ID_MASK; -		spin_unlock_irqrestore(&cm.lock, flags); -	} while( (ret == -EAGAIN) && idr_pre_get(&cm.local_id_table, GFP_KERNEL) ); +	idr_preload(GFP_KERNEL); +	spin_lock_irqsave(&cm.lock, flags); + +	id = idr_alloc_cyclic(&cm.local_id_table, cm_id_priv, 0, 0, GFP_NOWAIT); + +	spin_unlock_irqrestore(&cm.lock, flags); +	idr_preload_end();  	cm_id_priv->id.local_id = (__force __be32)id ^ cm.random_id_operand; -	return ret; +	return id < 0 ? id : 0;  }  static void cm_free_id(__be32 local_id) @@ -889,6 +894,8 @@ retest:  		break;  	case IB_CM_ESTABLISHED:  		spin_unlock_irq(&cm_id_priv->lock); +		if (cm_id_priv->qp_type == IB_QPT_XRC_TGT) +			break;  		ib_send_cm_dreq(cm_id, NULL, 0);  		goto retest;  	case IB_CM_DREQ_SENT: @@ -1008,7 +1015,6 @@ static void cm_format_req(struct cm_req_msg *req_msg,  	req_msg->service_id = param->service_id;  	req_msg->local_ca_guid = cm_id_priv->id.device->node_guid;  	cm_req_set_local_qpn(req_msg, cpu_to_be32(param->qp_num)); -	cm_req_set_resp_res(req_msg, param->responder_resources);  	cm_req_set_init_depth(req_msg, param->initiator_depth);  	cm_req_set_remote_resp_timeout(req_msg,  				       param->remote_cm_response_timeout); @@ -1017,12 +1023,16 @@ static void cm_format_req(struct cm_req_msg *req_msg,  	cm_req_set_starting_psn(req_msg, cpu_to_be32(param->starting_psn));  	cm_req_set_local_resp_timeout(req_msg,  				      param->local_cm_response_timeout); -	cm_req_set_retry_count(req_msg, param->retry_count);  	req_msg->pkey = param->primary_path->pkey;  	cm_req_set_path_mtu(req_msg, param->primary_path->mtu); -	cm_req_set_rnr_retry_count(req_msg, param->rnr_retry_count);  	cm_req_set_max_cm_retries(req_msg, param->max_cm_retries); -	cm_req_set_srq(req_msg, param->srq); + +	if (param->qp_type != IB_QPT_XRC_INI) { +		cm_req_set_resp_res(req_msg, param->responder_resources); +		cm_req_set_retry_count(req_msg, param->retry_count); +		cm_req_set_rnr_retry_count(req_msg, param->rnr_retry_count); +		cm_req_set_srq(req_msg, param->srq); +	}  	if (pri_path->hop_limit <= 1) {  		req_msg->primary_local_lid = pri_path->slid; @@ -1080,7 +1090,8 @@ static int cm_validate_req_param(struct ib_cm_req_param *param)  	if (!param->primary_path)  		return -EINVAL; -	if (param->qp_type != IB_QPT_RC && param->qp_type != IB_QPT_UC) +	if (param->qp_type != IB_QPT_RC && param->qp_type != IB_QPT_UC && +	    param->qp_type != IB_QPT_XRC_INI)  		return -EINVAL;  	if (param->private_data && @@ -1549,6 +1560,9 @@ static int cm_req_handler(struct cm_work *work)  	cm_process_routed_req(req_msg, work->mad_recv_wc->wc);  	cm_format_paths_from_req(req_msg, &work->path[0], &work->path[1]); + +	memcpy(work->path[0].dmac, cm_id_priv->av.ah_attr.dmac, ETH_ALEN); +	work->path[0].vlan_id = cm_id_priv->av.ah_attr.vlan_id;  	ret = cm_init_av_by_path(&work->path[0], &cm_id_priv->av);  	if (ret) {  		ib_get_cached_gid(work->port->cm_dev->ib_device, @@ -1601,18 +1615,24 @@ static void cm_format_rep(struct cm_rep_msg *rep_msg,  	cm_format_mad_hdr(&rep_msg->hdr, CM_REP_ATTR_ID, cm_id_priv->tid);  	rep_msg->local_comm_id = cm_id_priv->id.local_id;  	rep_msg->remote_comm_id = cm_id_priv->id.remote_id; -	cm_rep_set_local_qpn(rep_msg, cpu_to_be32(param->qp_num));  	cm_rep_set_starting_psn(rep_msg, cpu_to_be32(param->starting_psn));  	rep_msg->resp_resources = param->responder_resources; -	rep_msg->initiator_depth = param->initiator_depth;  	cm_rep_set_target_ack_delay(rep_msg,  				    cm_id_priv->av.port->cm_dev->ack_delay);  	cm_rep_set_failover(rep_msg, param->failover_accepted); -	cm_rep_set_flow_ctrl(rep_msg, param->flow_control);  	cm_rep_set_rnr_retry_count(rep_msg, param->rnr_retry_count); -	cm_rep_set_srq(rep_msg, param->srq);  	rep_msg->local_ca_guid = cm_id_priv->id.device->node_guid; +	if (cm_id_priv->qp_type != IB_QPT_XRC_TGT) { +		rep_msg->initiator_depth = param->initiator_depth; +		cm_rep_set_flow_ctrl(rep_msg, param->flow_control); +		cm_rep_set_srq(rep_msg, param->srq); +		cm_rep_set_local_qpn(rep_msg, cpu_to_be32(param->qp_num)); +	} else { +		cm_rep_set_srq(rep_msg, 1); +		cm_rep_set_local_eecn(rep_msg, cpu_to_be32(param->qp_num)); +	} +  	if (param->private_data && param->private_data_len)  		memcpy(rep_msg->private_data, param->private_data,  		       param->private_data_len); @@ -1660,7 +1680,7 @@ int ib_send_cm_rep(struct ib_cm_id *cm_id,  	cm_id_priv->initiator_depth = param->initiator_depth;  	cm_id_priv->responder_resources = param->responder_resources;  	cm_id_priv->rq_psn = cm_rep_get_starting_psn(rep_msg); -	cm_id_priv->local_qpn = cm_rep_get_local_qpn(rep_msg); +	cm_id_priv->local_qpn = cpu_to_be32(param->qp_num & 0xFFFFFF);  out:	spin_unlock_irqrestore(&cm_id_priv->lock, flags);  	return ret; @@ -1731,7 +1751,7 @@ error:	spin_unlock_irqrestore(&cm_id_priv->lock, flags);  }  EXPORT_SYMBOL(ib_send_cm_rtu); -static void cm_format_rep_event(struct cm_work *work) +static void cm_format_rep_event(struct cm_work *work, enum ib_qp_type qp_type)  {  	struct cm_rep_msg *rep_msg;  	struct ib_cm_rep_event_param *param; @@ -1740,7 +1760,7 @@ static void cm_format_rep_event(struct cm_work *work)  	param = &work->cm_event.param.rep_rcvd;  	param->remote_ca_guid = rep_msg->local_ca_guid;  	param->remote_qkey = be32_to_cpu(rep_msg->local_qkey); -	param->remote_qpn = be32_to_cpu(cm_rep_get_local_qpn(rep_msg)); +	param->remote_qpn = be32_to_cpu(cm_rep_get_qpn(rep_msg, qp_type));  	param->starting_psn = be32_to_cpu(cm_rep_get_starting_psn(rep_msg));  	param->responder_resources = rep_msg->initiator_depth;  	param->initiator_depth = rep_msg->resp_resources; @@ -1808,7 +1828,7 @@ static int cm_rep_handler(struct cm_work *work)  		return -EINVAL;  	} -	cm_format_rep_event(work); +	cm_format_rep_event(work, cm_id_priv->qp_type);  	spin_lock_irq(&cm_id_priv->lock);  	switch (cm_id_priv->id.state) { @@ -1823,7 +1843,7 @@ static int cm_rep_handler(struct cm_work *work)  	cm_id_priv->timewait_info->work.remote_id = rep_msg->local_comm_id;  	cm_id_priv->timewait_info->remote_ca_guid = rep_msg->local_ca_guid; -	cm_id_priv->timewait_info->remote_qpn = cm_rep_get_local_qpn(rep_msg); +	cm_id_priv->timewait_info->remote_qpn = cm_rep_get_qpn(rep_msg, cm_id_priv->qp_type);  	spin_lock(&cm.lock);  	/* Check for duplicate REP. */ @@ -1850,7 +1870,7 @@ static int cm_rep_handler(struct cm_work *work)  	cm_id_priv->id.state = IB_CM_REP_RCVD;  	cm_id_priv->id.remote_id = rep_msg->local_comm_id; -	cm_id_priv->remote_qpn = cm_rep_get_local_qpn(rep_msg); +	cm_id_priv->remote_qpn = cm_rep_get_qpn(rep_msg, cm_id_priv->qp_type);  	cm_id_priv->initiator_depth = rep_msg->resp_resources;  	cm_id_priv->responder_resources = rep_msg->initiator_depth;  	cm_id_priv->sq_psn = cm_rep_get_starting_psn(rep_msg); @@ -1988,6 +2008,10 @@ int ib_send_cm_dreq(struct ib_cm_id *cm_id,  		goto out;  	} +	if (cm_id->lap_state == IB_CM_LAP_SENT || +	    cm_id->lap_state == IB_CM_MRA_LAP_RCVD) +		ib_cancel_mad(cm_id_priv->av.port->mad_agent, cm_id_priv->msg); +  	ret = cm_alloc_msg(cm_id_priv, &msg);  	if (ret) {  		cm_enter_timewait(cm_id_priv); @@ -2129,6 +2153,10 @@ static int cm_dreq_handler(struct cm_work *work)  		ib_cancel_mad(cm_id_priv->av.port->mad_agent, cm_id_priv->msg);  		break;  	case IB_CM_ESTABLISHED: +		if (cm_id_priv->id.lap_state == IB_CM_LAP_SENT || +		    cm_id_priv->id.lap_state == IB_CM_MRA_LAP_RCVD) +			ib_cancel_mad(cm_id_priv->av.port->mad_agent, cm_id_priv->msg); +		break;  	case IB_CM_MRA_REP_RCVD:  		break;  	case IB_CM_TIMEWAIT: @@ -2349,9 +2377,18 @@ static int cm_rej_handler(struct cm_work *work)  		/* fall through */  	case IB_CM_REP_RCVD:  	case IB_CM_MRA_REP_SENT: -	case IB_CM_ESTABLISHED:  		cm_enter_timewait(cm_id_priv);  		break; +	case IB_CM_ESTABLISHED: +		if (cm_id_priv->id.lap_state == IB_CM_LAP_UNINIT || +		    cm_id_priv->id.lap_state == IB_CM_LAP_SENT) { +			if (cm_id_priv->id.lap_state == IB_CM_LAP_SENT) +				ib_cancel_mad(cm_id_priv->av.port->mad_agent, +					      cm_id_priv->msg); +			cm_enter_timewait(cm_id_priv); +			break; +		} +		/* fall through */  	default:  		spin_unlock_irq(&cm_id_priv->lock);  		ret = -EINVAL; @@ -2989,6 +3026,7 @@ static int cm_sidr_req_handler(struct cm_work *work)  		goto out; /* No match. */  	}  	atomic_inc(&cur_cm_id_priv->refcount); +	atomic_inc(&cm_id_priv->refcount);  	spin_unlock_irq(&cm.lock);  	cm_id_priv->id.cm_handler = cur_cm_id_priv->id.cm_handler; @@ -3471,10 +3509,37 @@ static int cm_init_qp_rtr_attr(struct cm_id_private *cm_id_priv,  		*qp_attr_mask = IB_QP_STATE | IB_QP_AV | IB_QP_PATH_MTU |  				IB_QP_DEST_QPN | IB_QP_RQ_PSN;  		qp_attr->ah_attr = cm_id_priv->av.ah_attr; +		if (!cm_id_priv->av.valid) { +			spin_unlock_irqrestore(&cm_id_priv->lock, flags); +			return -EINVAL; +		} +		if (cm_id_priv->av.ah_attr.vlan_id != 0xffff) { +			qp_attr->vlan_id = cm_id_priv->av.ah_attr.vlan_id; +			*qp_attr_mask |= IB_QP_VID; +		} +		if (!is_zero_ether_addr(cm_id_priv->av.smac)) { +			memcpy(qp_attr->smac, cm_id_priv->av.smac, +			       sizeof(qp_attr->smac)); +			*qp_attr_mask |= IB_QP_SMAC; +		} +		if (cm_id_priv->alt_av.valid) { +			if (cm_id_priv->alt_av.ah_attr.vlan_id != 0xffff) { +				qp_attr->alt_vlan_id = +					cm_id_priv->alt_av.ah_attr.vlan_id; +				*qp_attr_mask |= IB_QP_ALT_VID; +			} +			if (!is_zero_ether_addr(cm_id_priv->alt_av.smac)) { +				memcpy(qp_attr->alt_smac, +				       cm_id_priv->alt_av.smac, +				       sizeof(qp_attr->alt_smac)); +				*qp_attr_mask |= IB_QP_ALT_SMAC; +			} +		}  		qp_attr->path_mtu = cm_id_priv->path_mtu;  		qp_attr->dest_qp_num = be32_to_cpu(cm_id_priv->remote_qpn);  		qp_attr->rq_psn = be32_to_cpu(cm_id_priv->rq_psn); -		if (cm_id_priv->qp_type == IB_QPT_RC) { +		if (cm_id_priv->qp_type == IB_QPT_RC || +		    cm_id_priv->qp_type == IB_QPT_XRC_TGT) {  			*qp_attr_mask |= IB_QP_MAX_DEST_RD_ATOMIC |  					 IB_QP_MIN_RNR_TIMER;  			qp_attr->max_dest_rd_atomic = @@ -3519,15 +3584,21 @@ static int cm_init_qp_rts_attr(struct cm_id_private *cm_id_priv,  		if (cm_id_priv->id.lap_state == IB_CM_LAP_UNINIT) {  			*qp_attr_mask = IB_QP_STATE | IB_QP_SQ_PSN;  			qp_attr->sq_psn = be32_to_cpu(cm_id_priv->sq_psn); -			if (cm_id_priv->qp_type == IB_QPT_RC) { -				*qp_attr_mask |= IB_QP_TIMEOUT | IB_QP_RETRY_CNT | -						 IB_QP_RNR_RETRY | +			switch (cm_id_priv->qp_type) { +			case IB_QPT_RC: +			case IB_QPT_XRC_INI: +				*qp_attr_mask |= IB_QP_RETRY_CNT | IB_QP_RNR_RETRY |  						 IB_QP_MAX_QP_RD_ATOMIC; -				qp_attr->timeout = cm_id_priv->av.timeout;  				qp_attr->retry_cnt = cm_id_priv->retry_count;  				qp_attr->rnr_retry = cm_id_priv->rnr_retry_count; -				qp_attr->max_rd_atomic = -					cm_id_priv->initiator_depth; +				qp_attr->max_rd_atomic = cm_id_priv->initiator_depth; +				/* fall through */ +			case IB_QPT_XRC_TGT: +				*qp_attr_mask |= IB_QP_TIMEOUT; +				qp_attr->timeout = cm_id_priv->av.timeout; +				break; +			default: +				break;  			}  			if (cm_id_priv->alt_av.ah_attr.dlid) {  				*qp_attr_mask |= IB_QP_PATH_MIG_STATE; @@ -3621,8 +3692,17 @@ static struct kobj_type cm_port_obj_type = {  	.release = cm_release_port_obj  }; +static char *cm_devnode(struct device *dev, umode_t *mode) +{ +	if (mode) +		*mode = 0666; +	return kasprintf(GFP_KERNEL, "infiniband/%s", dev_name(dev)); +} +  struct class cm_class = { +	.owner   = THIS_MODULE,  	.name    = "infiniband_cm", +	.devnode = cm_devnode,  };  EXPORT_SYMBOL(cm_class); @@ -3797,28 +3877,31 @@ static int __init ib_cm_init(void)  	cm.remote_sidr_table = RB_ROOT;  	idr_init(&cm.local_id_table);  	get_random_bytes(&cm.random_id_operand, sizeof cm.random_id_operand); -	idr_pre_get(&cm.local_id_table, GFP_KERNEL);  	INIT_LIST_HEAD(&cm.timewait_list);  	ret = class_register(&cm_class); -	if (ret) -		return -ENOMEM; +	if (ret) { +		ret = -ENOMEM; +		goto error1; +	}  	cm.wq = create_workqueue("ib_cm");  	if (!cm.wq) {  		ret = -ENOMEM; -		goto error1; +		goto error2;  	}  	ret = ib_register_client(&cm_client);  	if (ret) -		goto error2; +		goto error3;  	return 0; -error2: +error3:  	destroy_workqueue(cm.wq); -error1: +error2:  	class_unregister(&cm_class); +error1: +	idr_destroy(&cm.local_id_table);  	return ret;  }  | 
