Linux Audio

Check our new training course

Loading...
v6.8
   1/*
   2 * Copyright (c) 2004, 2005 Topspin Communications.  All rights reserved.
   3 * Copyright (c) 2005 Sun Microsystems, Inc. All rights reserved.
   4 * Copyright (c) 2005, 2006 Cisco Systems.  All rights reserved.
   5 * Copyright (c) 2005 Mellanox Technologies. All rights reserved.
   6 * Copyright (c) 2004 Voltaire, Inc. All rights reserved.
   7 *
   8 * This software is available to you under a choice of one of two
   9 * licenses.  You may choose to be licensed under the terms of the GNU
  10 * General Public License (GPL) Version 2, available from the file
  11 * COPYING in the main directory of this source tree, or the
  12 * OpenIB.org BSD license below:
  13 *
  14 *     Redistribution and use in source and binary forms, with or
  15 *     without modification, are permitted provided that the following
  16 *     conditions are met:
  17 *
  18 *      - Redistributions of source code must retain the above
  19 *        copyright notice, this list of conditions and the following
  20 *        disclaimer.
  21 *
  22 *      - Redistributions in binary form must reproduce the above
  23 *        copyright notice, this list of conditions and the following
  24 *        disclaimer in the documentation and/or other materials
  25 *        provided with the distribution.
  26 *
  27 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
  28 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
  29 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
  30 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
  31 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
  32 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
  33 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
  34 * SOFTWARE.
  35 */
  36
  37#include <rdma/ib_smi.h>
  38#include <rdma/ib_umem.h>
  39#include <rdma/ib_user_verbs.h>
  40#include <rdma/uverbs_ioctl.h>
  41
  42#include <linux/sched.h>
  43#include <linux/slab.h>
  44#include <linux/stat.h>
  45#include <linux/mm.h>
  46#include <linux/export.h>
  47
  48#include "mthca_dev.h"
  49#include "mthca_cmd.h"
  50#include <rdma/mthca-abi.h>
  51#include "mthca_memfree.h"
  52
 
 
 
 
 
 
 
 
  53static int mthca_query_device(struct ib_device *ibdev, struct ib_device_attr *props,
  54			      struct ib_udata *uhw)
  55{
  56	struct ib_smp *in_mad;
  57	struct ib_smp *out_mad;
  58	int err = -ENOMEM;
  59	struct mthca_dev *mdev = to_mdev(ibdev);
  60
  61	if (uhw->inlen || uhw->outlen)
  62		return -EINVAL;
  63
  64	in_mad  = kzalloc(sizeof *in_mad, GFP_KERNEL);
  65	out_mad = kmalloc(sizeof *out_mad, GFP_KERNEL);
  66	if (!in_mad || !out_mad)
  67		goto out;
  68
  69	memset(props, 0, sizeof *props);
  70
  71	props->fw_ver              = mdev->fw_ver;
  72
  73	ib_init_query_mad(in_mad);
  74	in_mad->attr_id = IB_SMP_ATTR_NODE_INFO;
  75
  76	err = mthca_MAD_IFC(mdev, 1, 1,
  77			    1, NULL, NULL, in_mad, out_mad);
  78	if (err)
  79		goto out;
  80
  81	props->device_cap_flags    = mdev->device_cap_flags;
  82	props->vendor_id           = be32_to_cpup((__be32 *) (out_mad->data + 36)) &
  83		0xffffff;
  84	props->vendor_part_id      = be16_to_cpup((__be16 *) (out_mad->data + 30));
  85	props->hw_ver              = be32_to_cpup((__be32 *) (out_mad->data + 32));
  86	memcpy(&props->sys_image_guid, out_mad->data +  4, 8);
  87
  88	props->max_mr_size         = ~0ull;
  89	props->page_size_cap       = mdev->limits.page_size_cap;
  90	props->max_qp              = mdev->limits.num_qps - mdev->limits.reserved_qps;
  91	props->max_qp_wr           = mdev->limits.max_wqes;
  92	props->max_send_sge        = mdev->limits.max_sg;
  93	props->max_recv_sge        = mdev->limits.max_sg;
  94	props->max_sge_rd          = mdev->limits.max_sg;
  95	props->max_cq              = mdev->limits.num_cqs - mdev->limits.reserved_cqs;
  96	props->max_cqe             = mdev->limits.max_cqes;
  97	props->max_mr              = mdev->limits.num_mpts - mdev->limits.reserved_mrws;
  98	props->max_pd              = mdev->limits.num_pds - mdev->limits.reserved_pds;
  99	props->max_qp_rd_atom      = 1 << mdev->qp_table.rdb_shift;
 100	props->max_qp_init_rd_atom = mdev->limits.max_qp_init_rdma;
 101	props->max_res_rd_atom     = props->max_qp_rd_atom * props->max_qp;
 102	props->max_srq             = mdev->limits.num_srqs - mdev->limits.reserved_srqs;
 103	props->max_srq_wr          = mdev->limits.max_srq_wqes;
 104	props->max_srq_sge         = mdev->limits.max_srq_sge;
 105	props->local_ca_ack_delay  = mdev->limits.local_ca_ack_delay;
 106	props->atomic_cap          = mdev->limits.flags & DEV_LIM_FLAG_ATOMIC ?
 107					IB_ATOMIC_HCA : IB_ATOMIC_NONE;
 108	props->max_pkeys           = mdev->limits.pkey_table_len;
 109	props->max_mcast_grp       = mdev->limits.num_mgms + mdev->limits.num_amgms;
 110	props->max_mcast_qp_attach = MTHCA_QP_PER_MGM;
 111	props->max_total_mcast_qp_attach = props->max_mcast_qp_attach *
 112					   props->max_mcast_grp;
 
 
 
 
 
 
 
 
 
 
 113
 114	err = 0;
 115 out:
 116	kfree(in_mad);
 117	kfree(out_mad);
 118	return err;
 119}
 120
 121static int mthca_query_port(struct ib_device *ibdev,
 122			    u32 port, struct ib_port_attr *props)
 123{
 124	struct ib_smp *in_mad;
 125	struct ib_smp *out_mad;
 126	int err = -ENOMEM;
 127
 128	in_mad  = kzalloc(sizeof *in_mad, GFP_KERNEL);
 129	out_mad = kmalloc(sizeof *out_mad, GFP_KERNEL);
 130	if (!in_mad || !out_mad)
 131		goto out;
 132
 133	/* props being zeroed by the caller, avoid zeroing it here */
 134
 135	ib_init_query_mad(in_mad);
 136	in_mad->attr_id  = IB_SMP_ATTR_PORT_INFO;
 137	in_mad->attr_mod = cpu_to_be32(port);
 138
 139	err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1,
 140			    port, NULL, NULL, in_mad, out_mad);
 141	if (err)
 142		goto out;
 143
 144	props->lid               = be16_to_cpup((__be16 *) (out_mad->data + 16));
 145	props->lmc               = out_mad->data[34] & 0x7;
 146	props->sm_lid            = be16_to_cpup((__be16 *) (out_mad->data + 18));
 147	props->sm_sl             = out_mad->data[36] & 0xf;
 148	props->state             = out_mad->data[32] & 0xf;
 149	props->phys_state        = out_mad->data[33] >> 4;
 150	props->port_cap_flags    = be32_to_cpup((__be32 *) (out_mad->data + 20));
 151	props->gid_tbl_len       = to_mdev(ibdev)->limits.gid_table_len;
 152	props->max_msg_sz        = 0x80000000;
 153	props->pkey_tbl_len      = to_mdev(ibdev)->limits.pkey_table_len;
 154	props->bad_pkey_cntr     = be16_to_cpup((__be16 *) (out_mad->data + 46));
 155	props->qkey_viol_cntr    = be16_to_cpup((__be16 *) (out_mad->data + 48));
 156	props->active_width      = out_mad->data[31] & 0xf;
 157	props->active_speed      = out_mad->data[35] >> 4;
 158	props->max_mtu           = out_mad->data[41] & 0xf;
 159	props->active_mtu        = out_mad->data[36] >> 4;
 160	props->subnet_timeout    = out_mad->data[51] & 0x1f;
 161	props->max_vl_num        = out_mad->data[37] >> 4;
 162	props->init_type_reply   = out_mad->data[41] >> 4;
 163
 164 out:
 165	kfree(in_mad);
 166	kfree(out_mad);
 167	return err;
 168}
 169
 170static int mthca_modify_device(struct ib_device *ibdev,
 171			       int mask,
 172			       struct ib_device_modify *props)
 173{
 174	if (mask & ~IB_DEVICE_MODIFY_NODE_DESC)
 175		return -EOPNOTSUPP;
 176
 177	if (mask & IB_DEVICE_MODIFY_NODE_DESC) {
 178		if (mutex_lock_interruptible(&to_mdev(ibdev)->cap_mask_mutex))
 179			return -ERESTARTSYS;
 180		memcpy(ibdev->node_desc, props->node_desc,
 181		       IB_DEVICE_NODE_DESC_MAX);
 182		mutex_unlock(&to_mdev(ibdev)->cap_mask_mutex);
 183	}
 184
 185	return 0;
 186}
 187
 188static int mthca_modify_port(struct ib_device *ibdev,
 189			     u32 port, int port_modify_mask,
 190			     struct ib_port_modify *props)
 191{
 192	struct mthca_set_ib_param set_ib;
 193	struct ib_port_attr attr;
 194	int err;
 195
 196	if (mutex_lock_interruptible(&to_mdev(ibdev)->cap_mask_mutex))
 197		return -ERESTARTSYS;
 198
 199	err = ib_query_port(ibdev, port, &attr);
 200	if (err)
 201		goto out;
 202
 203	set_ib.set_si_guid     = 0;
 204	set_ib.reset_qkey_viol = !!(port_modify_mask & IB_PORT_RESET_QKEY_CNTR);
 205
 206	set_ib.cap_mask = (attr.port_cap_flags | props->set_port_cap_mask) &
 207		~props->clr_port_cap_mask;
 208
 209	err = mthca_SET_IB(to_mdev(ibdev), &set_ib, port);
 210	if (err)
 211		goto out;
 212out:
 213	mutex_unlock(&to_mdev(ibdev)->cap_mask_mutex);
 214	return err;
 215}
 216
 217static int mthca_query_pkey(struct ib_device *ibdev,
 218			    u32 port, u16 index, u16 *pkey)
 219{
 220	struct ib_smp *in_mad;
 221	struct ib_smp *out_mad;
 222	int err = -ENOMEM;
 223
 224	in_mad  = kzalloc(sizeof *in_mad, GFP_KERNEL);
 225	out_mad = kmalloc(sizeof *out_mad, GFP_KERNEL);
 226	if (!in_mad || !out_mad)
 227		goto out;
 228
 229	ib_init_query_mad(in_mad);
 230	in_mad->attr_id  = IB_SMP_ATTR_PKEY_TABLE;
 231	in_mad->attr_mod = cpu_to_be32(index / 32);
 232
 233	err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1,
 234			    port, NULL, NULL, in_mad, out_mad);
 235	if (err)
 236		goto out;
 237
 238	*pkey = be16_to_cpu(((__be16 *) out_mad->data)[index % 32]);
 239
 240 out:
 241	kfree(in_mad);
 242	kfree(out_mad);
 243	return err;
 244}
 245
 246static int mthca_query_gid(struct ib_device *ibdev, u32 port,
 247			   int index, union ib_gid *gid)
 248{
 249	struct ib_smp *in_mad;
 250	struct ib_smp *out_mad;
 251	int err = -ENOMEM;
 252
 253	in_mad  = kzalloc(sizeof *in_mad, GFP_KERNEL);
 254	out_mad = kmalloc(sizeof *out_mad, GFP_KERNEL);
 255	if (!in_mad || !out_mad)
 256		goto out;
 257
 258	ib_init_query_mad(in_mad);
 259	in_mad->attr_id  = IB_SMP_ATTR_PORT_INFO;
 260	in_mad->attr_mod = cpu_to_be32(port);
 261
 262	err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1,
 263			    port, NULL, NULL, in_mad, out_mad);
 264	if (err)
 265		goto out;
 266
 267	memcpy(gid->raw, out_mad->data + 8, 8);
 268
 269	ib_init_query_mad(in_mad);
 270	in_mad->attr_id  = IB_SMP_ATTR_GUID_INFO;
 271	in_mad->attr_mod = cpu_to_be32(index / 8);
 272
 273	err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1,
 274			    port, NULL, NULL, in_mad, out_mad);
 275	if (err)
 276		goto out;
 277
 278	memcpy(gid->raw + 8, out_mad->data + (index % 8) * 8, 8);
 279
 280 out:
 281	kfree(in_mad);
 282	kfree(out_mad);
 283	return err;
 284}
 285
 286static int mthca_alloc_ucontext(struct ib_ucontext *uctx,
 287				struct ib_udata *udata)
 288{
 289	struct ib_device *ibdev = uctx->device;
 290	struct mthca_alloc_ucontext_resp uresp = {};
 291	struct mthca_ucontext *context = to_mucontext(uctx);
 292	int                              err;
 293
 294	if (!(to_mdev(ibdev)->active))
 295		return -EAGAIN;
 
 
 296
 297	uresp.qp_tab_size = to_mdev(ibdev)->limits.num_qps;
 298	if (mthca_is_memfree(to_mdev(ibdev)))
 299		uresp.uarc_size = to_mdev(ibdev)->uar_table.uarc_size;
 300	else
 301		uresp.uarc_size = 0;
 302
 
 
 
 
 303	err = mthca_uar_alloc(to_mdev(ibdev), &context->uar);
 304	if (err)
 305		return err;
 
 
 306
 307	context->db_tab = mthca_init_user_db_tab(to_mdev(ibdev));
 308	if (IS_ERR(context->db_tab)) {
 309		err = PTR_ERR(context->db_tab);
 310		mthca_uar_free(to_mdev(ibdev), &context->uar);
 311		return err;
 
 312	}
 313
 314	if (ib_copy_to_udata(udata, &uresp, sizeof(uresp))) {
 315		mthca_cleanup_user_db_tab(to_mdev(ibdev), &context->uar, context->db_tab);
 316		mthca_uar_free(to_mdev(ibdev), &context->uar);
 317		return -EFAULT;
 
 318	}
 319
 320	context->reg_mr_warned = 0;
 321
 322	return 0;
 323}
 324
 325static void mthca_dealloc_ucontext(struct ib_ucontext *context)
 326{
 327	mthca_cleanup_user_db_tab(to_mdev(context->device), &to_mucontext(context)->uar,
 328				  to_mucontext(context)->db_tab);
 329	mthca_uar_free(to_mdev(context->device), &to_mucontext(context)->uar);
 
 
 
 330}
 331
 332static int mthca_mmap_uar(struct ib_ucontext *context,
 333			  struct vm_area_struct *vma)
 334{
 335	if (vma->vm_end - vma->vm_start != PAGE_SIZE)
 336		return -EINVAL;
 337
 338	vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
 339
 340	if (io_remap_pfn_range(vma, vma->vm_start,
 341			       to_mucontext(context)->uar.pfn,
 342			       PAGE_SIZE, vma->vm_page_prot))
 343		return -EAGAIN;
 344
 345	return 0;
 346}
 347
 348static int mthca_alloc_pd(struct ib_pd *ibpd, struct ib_udata *udata)
 
 
 349{
 350	struct ib_device *ibdev = ibpd->device;
 351	struct mthca_pd *pd = to_mpd(ibpd);
 352	int err;
 353
 354	err = mthca_pd_alloc(to_mdev(ibdev), !udata, pd);
 355	if (err)
 356		return err;
 
 
 
 
 
 
 357
 358	if (udata) {
 359		if (ib_copy_to_udata(udata, &pd->pd_num, sizeof (__u32))) {
 360			mthca_pd_free(to_mdev(ibdev), pd);
 361			return -EFAULT;
 
 362		}
 363	}
 364
 365	return 0;
 366}
 367
 368static int mthca_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata)
 369{
 370	mthca_pd_free(to_mdev(pd->device), to_mpd(pd));
 
 
 371	return 0;
 372}
 373
 374static int mthca_ah_create(struct ib_ah *ibah,
 375			   struct rdma_ah_init_attr *init_attr,
 376			   struct ib_udata *udata)
 377
 378{
 379	struct mthca_ah *ah = to_mah(ibah);
 
 
 
 
 
 380
 381	return mthca_create_ah(to_mdev(ibah->device), to_mpd(ibah->pd),
 382			       init_attr->ah_attr, ah);
 
 
 
 
 
 383}
 384
 385static int mthca_ah_destroy(struct ib_ah *ah, u32 flags)
 386{
 387	mthca_destroy_ah(to_mdev(ah->device), to_mah(ah));
 
 
 388	return 0;
 389}
 390
 391static int mthca_create_srq(struct ib_srq *ibsrq,
 392			    struct ib_srq_init_attr *init_attr,
 393			    struct ib_udata *udata)
 394{
 395	struct mthca_create_srq ucmd;
 396	struct mthca_ucontext *context = rdma_udata_to_drv_context(
 397		udata, struct mthca_ucontext, ibucontext);
 398	struct mthca_srq *srq = to_msrq(ibsrq);
 399	int err;
 400
 401	if (init_attr->srq_type != IB_SRQT_BASIC)
 402		return -EOPNOTSUPP;
 403
 404	if (udata) {
 405		if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd)))
 406			return -EFAULT;
 
 
 
 
 
 
 
 
 407
 408		err = mthca_map_user_db(to_mdev(ibsrq->device), &context->uar,
 409					context->db_tab, ucmd.db_index,
 410					ucmd.db_page);
 411
 412		if (err)
 413			return err;
 414
 415		srq->mr.ibmr.lkey = ucmd.lkey;
 416		srq->db_index     = ucmd.db_index;
 417	}
 418
 419	err = mthca_alloc_srq(to_mdev(ibsrq->device), to_mpd(ibsrq->pd),
 420			      &init_attr->attr, srq, udata);
 421
 422	if (err && udata)
 423		mthca_unmap_user_db(to_mdev(ibsrq->device), &context->uar,
 424				    context->db_tab, ucmd.db_index);
 425
 426	if (err)
 427		return err;
 428
 429	if (context && ib_copy_to_udata(udata, &srq->srqn, sizeof(__u32))) {
 430		mthca_free_srq(to_mdev(ibsrq->device), srq);
 431		return -EFAULT;
 
 432	}
 433
 434	return 0;
 
 
 
 
 
 435}
 436
 437static int mthca_destroy_srq(struct ib_srq *srq, struct ib_udata *udata)
 438{
 439	if (udata) {
 440		struct mthca_ucontext *context =
 441			rdma_udata_to_drv_context(
 442				udata,
 443				struct mthca_ucontext,
 444				ibucontext);
 445
 446		mthca_unmap_user_db(to_mdev(srq->device), &context->uar,
 447				    context->db_tab, to_msrq(srq)->db_index);
 448	}
 449
 450	mthca_free_srq(to_mdev(srq->device), to_msrq(srq));
 
 
 451	return 0;
 452}
 453
 454static int mthca_create_qp(struct ib_qp *ibqp,
 455			   struct ib_qp_init_attr *init_attr,
 456			   struct ib_udata *udata)
 457{
 458	struct mthca_ucontext *context = rdma_udata_to_drv_context(
 459		udata, struct mthca_ucontext, ibucontext);
 460	struct mthca_create_qp ucmd;
 461	struct mthca_qp *qp = to_mqp(ibqp);
 462	struct mthca_dev *dev = to_mdev(ibqp->device);
 463	int err;
 464
 465	if (init_attr->create_flags)
 466		return -EOPNOTSUPP;
 467
 468	switch (init_attr->qp_type) {
 469	case IB_QPT_RC:
 470	case IB_QPT_UC:
 471	case IB_QPT_UD:
 472	{
 473		if (udata) {
 474			if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd)))
 475				return -EFAULT;
 476
 477			err = mthca_map_user_db(dev, &context->uar,
 
 
 
 
 
 
 
 
 
 
 
 
 478						context->db_tab,
 479						ucmd.sq_db_index,
 480						ucmd.sq_db_page);
 481			if (err)
 482				return err;
 
 483
 484			err = mthca_map_user_db(dev, &context->uar,
 485						context->db_tab,
 486						ucmd.rq_db_index,
 487						ucmd.rq_db_page);
 488			if (err) {
 489				mthca_unmap_user_db(dev, &context->uar,
 
 490						    context->db_tab,
 491						    ucmd.sq_db_index);
 492				return err;
 
 493			}
 494
 495			qp->mr.ibmr.lkey = ucmd.lkey;
 496			qp->sq.db_index  = ucmd.sq_db_index;
 497			qp->rq.db_index  = ucmd.rq_db_index;
 498		}
 499
 500		err = mthca_alloc_qp(dev, to_mpd(ibqp->pd),
 501				     to_mcq(init_attr->send_cq),
 502				     to_mcq(init_attr->recv_cq),
 503				     init_attr->qp_type, init_attr->sq_sig_type,
 504				     &init_attr->cap, qp, udata);
 
 
 
 505
 506		if (err && udata) {
 507			mthca_unmap_user_db(dev, &context->uar, context->db_tab,
 
 508					    ucmd.sq_db_index);
 509			mthca_unmap_user_db(dev, &context->uar, context->db_tab,
 
 
 510					    ucmd.rq_db_index);
 511		}
 512
 513		qp->ibqp.qp_num = qp->qpn;
 514		break;
 515	}
 516	case IB_QPT_SMI:
 517	case IB_QPT_GSI:
 518	{
 519		qp->sqp = kzalloc(sizeof(struct mthca_sqp), GFP_KERNEL);
 520		if (!qp->sqp)
 521			return -ENOMEM;
 
 
 
 
 522
 523		qp->ibqp.qp_num = init_attr->qp_type == IB_QPT_SMI ? 0 : 1;
 524
 525		err = mthca_alloc_sqp(dev, to_mpd(ibqp->pd),
 526				      to_mcq(init_attr->send_cq),
 527				      to_mcq(init_attr->recv_cq),
 528				      init_attr->sq_sig_type, &init_attr->cap,
 529				      qp->ibqp.qp_num, init_attr->port_num, qp,
 530				      udata);
 531		break;
 532	}
 533	default:
 534		/* Don't support raw QPs */
 535		return -EOPNOTSUPP;
 536	}
 537
 538	if (err) {
 539		kfree(qp->sqp);
 540		return err;
 541	}
 542
 543	init_attr->cap.max_send_wr     = qp->sq.max;
 544	init_attr->cap.max_recv_wr     = qp->rq.max;
 545	init_attr->cap.max_send_sge    = qp->sq.max_gs;
 546	init_attr->cap.max_recv_sge    = qp->rq.max_gs;
 547	init_attr->cap.max_inline_data = qp->max_inline_data;
 548
 549	return 0;
 550}
 551
 552static int mthca_destroy_qp(struct ib_qp *qp, struct ib_udata *udata)
 553{
 554	if (udata) {
 555		struct mthca_ucontext *context =
 556			rdma_udata_to_drv_context(
 557				udata,
 558				struct mthca_ucontext,
 559				ibucontext);
 560
 561		mthca_unmap_user_db(to_mdev(qp->device),
 562				    &context->uar,
 563				    context->db_tab,
 564				    to_mqp(qp)->sq.db_index);
 565		mthca_unmap_user_db(to_mdev(qp->device),
 566				    &context->uar,
 567				    context->db_tab,
 568				    to_mqp(qp)->rq.db_index);
 569	}
 570	mthca_free_qp(to_mdev(qp->device), to_mqp(qp));
 571	kfree(to_mqp(qp)->sqp);
 572	return 0;
 573}
 574
 575static int mthca_create_cq(struct ib_cq *ibcq,
 576			   const struct ib_cq_init_attr *attr,
 577			   struct ib_udata *udata)
 
 578{
 579	struct ib_device *ibdev = ibcq->device;
 580	int entries = attr->cqe;
 581	struct mthca_create_cq ucmd;
 582	struct mthca_cq *cq;
 583	int nent;
 584	int err;
 585	struct mthca_ucontext *context = rdma_udata_to_drv_context(
 586		udata, struct mthca_ucontext, ibucontext);
 587
 588	if (attr->flags)
 589		return -EOPNOTSUPP;
 590
 591	if (entries < 1 || entries > to_mdev(ibdev)->limits.max_cqes)
 592		return -EINVAL;
 593
 594	if (udata) {
 595		if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd)))
 596			return -EFAULT;
 597
 598		err = mthca_map_user_db(to_mdev(ibdev), &context->uar,
 599					context->db_tab, ucmd.set_db_index,
 600					ucmd.set_db_page);
 601		if (err)
 602			return err;
 603
 604		err = mthca_map_user_db(to_mdev(ibdev), &context->uar,
 605					context->db_tab, ucmd.arm_db_index,
 606					ucmd.arm_db_page);
 607		if (err)
 608			goto err_unmap_set;
 609	}
 610
 611	cq = to_mcq(ibcq);
 
 
 
 
 612
 613	if (udata) {
 614		cq->buf.mr.ibmr.lkey = ucmd.lkey;
 615		cq->set_ci_db_index  = ucmd.set_db_index;
 616		cq->arm_db_index     = ucmd.arm_db_index;
 617	}
 618
 619	for (nent = 1; nent <= entries; nent <<= 1)
 620		; /* nothing */
 621
 622	err = mthca_init_cq(to_mdev(ibdev), nent, context,
 623			    udata ? ucmd.pdn : to_mdev(ibdev)->driver_pd.pd_num,
 
 624			    cq);
 625	if (err)
 626		goto err_unmap_arm;
 627
 628	if (udata && ib_copy_to_udata(udata, &cq->cqn, sizeof(__u32))) {
 629		mthca_free_cq(to_mdev(ibdev), cq);
 630		err = -EFAULT;
 631		goto err_unmap_arm;
 632	}
 633
 634	cq->resize_buf = NULL;
 635
 636	return 0;
 
 
 
 637
 638err_unmap_arm:
 639	if (udata)
 640		mthca_unmap_user_db(to_mdev(ibdev), &context->uar,
 641				    context->db_tab, ucmd.arm_db_index);
 642
 643err_unmap_set:
 644	if (udata)
 645		mthca_unmap_user_db(to_mdev(ibdev), &context->uar,
 646				    context->db_tab, ucmd.set_db_index);
 647
 648	return err;
 649}
 650
 651static int mthca_alloc_resize_buf(struct mthca_dev *dev, struct mthca_cq *cq,
 652				  int entries)
 653{
 654	int ret;
 655
 656	spin_lock_irq(&cq->lock);
 657	if (cq->resize_buf) {
 658		ret = -EBUSY;
 659		goto unlock;
 660	}
 661
 662	cq->resize_buf = kmalloc(sizeof *cq->resize_buf, GFP_ATOMIC);
 663	if (!cq->resize_buf) {
 664		ret = -ENOMEM;
 665		goto unlock;
 666	}
 667
 668	cq->resize_buf->state = CQ_RESIZE_ALLOC;
 669
 670	ret = 0;
 671
 672unlock:
 673	spin_unlock_irq(&cq->lock);
 674
 675	if (ret)
 676		return ret;
 677
 678	ret = mthca_alloc_cq_buf(dev, &cq->resize_buf->buf, entries);
 679	if (ret) {
 680		spin_lock_irq(&cq->lock);
 681		kfree(cq->resize_buf);
 682		cq->resize_buf = NULL;
 683		spin_unlock_irq(&cq->lock);
 684		return ret;
 685	}
 686
 687	cq->resize_buf->cqe = entries - 1;
 688
 689	spin_lock_irq(&cq->lock);
 690	cq->resize_buf->state = CQ_RESIZE_READY;
 691	spin_unlock_irq(&cq->lock);
 692
 693	return 0;
 694}
 695
 696static int mthca_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata)
 697{
 698	struct mthca_dev *dev = to_mdev(ibcq->device);
 699	struct mthca_cq *cq = to_mcq(ibcq);
 700	struct mthca_resize_cq ucmd;
 701	u32 lkey;
 702	int ret;
 703
 704	if (entries < 1 || entries > dev->limits.max_cqes)
 705		return -EINVAL;
 706
 707	mutex_lock(&cq->mutex);
 708
 709	entries = roundup_pow_of_two(entries + 1);
 710	if (entries == ibcq->cqe + 1) {
 711		ret = 0;
 712		goto out;
 713	}
 714
 715	if (cq->is_kernel) {
 716		ret = mthca_alloc_resize_buf(dev, cq, entries);
 717		if (ret)
 718			goto out;
 719		lkey = cq->resize_buf->buf.mr.ibmr.lkey;
 720	} else {
 721		if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) {
 722			ret = -EFAULT;
 723			goto out;
 724		}
 725		lkey = ucmd.lkey;
 726	}
 727
 728	ret = mthca_RESIZE_CQ(dev, cq->cqn, lkey, ilog2(entries));
 729
 730	if (ret) {
 731		if (cq->resize_buf) {
 732			mthca_free_cq_buf(dev, &cq->resize_buf->buf,
 733					  cq->resize_buf->cqe);
 734			kfree(cq->resize_buf);
 735			spin_lock_irq(&cq->lock);
 736			cq->resize_buf = NULL;
 737			spin_unlock_irq(&cq->lock);
 738		}
 739		goto out;
 740	}
 741
 742	if (cq->is_kernel) {
 743		struct mthca_cq_buf tbuf;
 744		int tcqe;
 745
 746		spin_lock_irq(&cq->lock);
 747		if (cq->resize_buf->state == CQ_RESIZE_READY) {
 748			mthca_cq_resize_copy_cqes(cq);
 749			tbuf         = cq->buf;
 750			tcqe         = cq->ibcq.cqe;
 751			cq->buf      = cq->resize_buf->buf;
 752			cq->ibcq.cqe = cq->resize_buf->cqe;
 753		} else {
 754			tbuf = cq->resize_buf->buf;
 755			tcqe = cq->resize_buf->cqe;
 756		}
 757
 758		kfree(cq->resize_buf);
 759		cq->resize_buf = NULL;
 760		spin_unlock_irq(&cq->lock);
 761
 762		mthca_free_cq_buf(dev, &tbuf, tcqe);
 763	} else
 764		ibcq->cqe = entries - 1;
 765
 766out:
 767	mutex_unlock(&cq->mutex);
 768
 769	return ret;
 770}
 771
 772static int mthca_destroy_cq(struct ib_cq *cq, struct ib_udata *udata)
 773{
 774	if (udata) {
 775		struct mthca_ucontext *context =
 776			rdma_udata_to_drv_context(
 777				udata,
 778				struct mthca_ucontext,
 779				ibucontext);
 780
 781		mthca_unmap_user_db(to_mdev(cq->device),
 782				    &context->uar,
 783				    context->db_tab,
 784				    to_mcq(cq)->arm_db_index);
 785		mthca_unmap_user_db(to_mdev(cq->device),
 786				    &context->uar,
 787				    context->db_tab,
 788				    to_mcq(cq)->set_ci_db_index);
 789	}
 790	mthca_free_cq(to_mdev(cq->device), to_mcq(cq));
 
 
 791	return 0;
 792}
 793
 794static inline u32 convert_access(int acc)
 795{
 796	return (acc & IB_ACCESS_REMOTE_ATOMIC ? MTHCA_MPT_FLAG_ATOMIC       : 0) |
 797	       (acc & IB_ACCESS_REMOTE_WRITE  ? MTHCA_MPT_FLAG_REMOTE_WRITE : 0) |
 798	       (acc & IB_ACCESS_REMOTE_READ   ? MTHCA_MPT_FLAG_REMOTE_READ  : 0) |
 799	       (acc & IB_ACCESS_LOCAL_WRITE   ? MTHCA_MPT_FLAG_LOCAL_WRITE  : 0) |
 800	       MTHCA_MPT_FLAG_LOCAL_READ;
 801}
 802
 803static struct ib_mr *mthca_get_dma_mr(struct ib_pd *pd, int acc)
 804{
 805	struct mthca_mr *mr;
 806	int err;
 807
 808	mr = kmalloc(sizeof *mr, GFP_KERNEL);
 809	if (!mr)
 810		return ERR_PTR(-ENOMEM);
 811
 812	err = mthca_mr_alloc_notrans(to_mdev(pd->device),
 813				     to_mpd(pd)->pd_num,
 814				     convert_access(acc), mr);
 815
 816	if (err) {
 817		kfree(mr);
 818		return ERR_PTR(err);
 819	}
 820
 821	mr->umem = NULL;
 822
 823	return &mr->ibmr;
 824}
 825
 826static struct ib_mr *mthca_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
 827				       u64 virt, int acc, struct ib_udata *udata)
 828{
 829	struct mthca_dev *dev = to_mdev(pd->device);
 830	struct ib_block_iter biter;
 831	struct mthca_ucontext *context = rdma_udata_to_drv_context(
 832		udata, struct mthca_ucontext, ibucontext);
 833	struct mthca_mr *mr;
 834	struct mthca_reg_mr ucmd;
 835	u64 *pages;
 836	int n, i;
 
 837	int err = 0;
 838	int write_mtt_size;
 839
 840	if (udata->inlen < sizeof ucmd) {
 841		if (!context->reg_mr_warned) {
 842			mthca_warn(dev, "Process '%s' did not pass in MR attrs.\n",
 843				   current->comm);
 844			mthca_warn(dev, "  Update libmthca to fix this.\n");
 845		}
 846		++context->reg_mr_warned;
 847		ucmd.mr_attrs = 0;
 848	} else if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd))
 849		return ERR_PTR(-EFAULT);
 850
 851	mr = kmalloc(sizeof *mr, GFP_KERNEL);
 852	if (!mr)
 853		return ERR_PTR(-ENOMEM);
 854
 855	mr->umem = ib_umem_get(pd->device, start, length, acc);
 
 
 856	if (IS_ERR(mr->umem)) {
 857		err = PTR_ERR(mr->umem);
 858		goto err;
 859	}
 860
 861	n = ib_umem_num_dma_blocks(mr->umem, PAGE_SIZE);
 
 862
 863	mr->mtt = mthca_alloc_mtt(dev, n);
 864	if (IS_ERR(mr->mtt)) {
 865		err = PTR_ERR(mr->mtt);
 866		goto err_umem;
 867	}
 868
 869	pages = (u64 *) __get_free_page(GFP_KERNEL);
 870	if (!pages) {
 871		err = -ENOMEM;
 872		goto err_mtt;
 873	}
 874
 875	i = n = 0;
 876
 877	write_mtt_size = min(mthca_write_mtt_size(dev), (int) (PAGE_SIZE / sizeof *pages));
 878
 879	rdma_umem_for_each_dma_block(mr->umem, &biter, PAGE_SIZE) {
 880		pages[i++] = rdma_block_iter_dma_address(&biter);
 881
 882		/*
 883		 * Be friendly to write_mtt and pass it chunks
 884		 * of appropriate size.
 885		 */
 886		if (i == write_mtt_size) {
 887			err = mthca_write_mtt(dev, mr->mtt, n, pages, i);
 888			if (err)
 889				goto mtt_done;
 890			n += i;
 891			i = 0;
 
 
 
 892		}
 893	}
 894
 895	if (i)
 896		err = mthca_write_mtt(dev, mr->mtt, n, pages, i);
 897mtt_done:
 898	free_page((unsigned long) pages);
 899	if (err)
 900		goto err_mtt;
 901
 902	err = mthca_mr_alloc(dev, to_mpd(pd)->pd_num, PAGE_SHIFT, virt, length,
 903			     convert_access(acc), mr);
 904
 905	if (err)
 906		goto err_mtt;
 907
 908	return &mr->ibmr;
 909
 910err_mtt:
 911	mthca_free_mtt(dev, mr->mtt);
 912
 913err_umem:
 914	ib_umem_release(mr->umem);
 915
 916err:
 917	kfree(mr);
 918	return ERR_PTR(err);
 919}
 920
 921static int mthca_dereg_mr(struct ib_mr *mr, struct ib_udata *udata)
 922{
 923	struct mthca_mr *mmr = to_mmr(mr);
 924
 925	mthca_free_mr(to_mdev(mr->device), mmr);
 926	ib_umem_release(mmr->umem);
 
 927	kfree(mmr);
 928
 929	return 0;
 930}
 931
 932static ssize_t hw_rev_show(struct device *device,
 933			   struct device_attribute *attr, char *buf)
 934{
 935	struct mthca_dev *dev =
 936		rdma_device_to_drv_device(device, struct mthca_dev, ib_dev);
 
 
 
 
 
 
 
 
 
 
 
 
 
 937
 938	return sysfs_emit(buf, "%x\n", dev->rev_id);
 939}
 940static DEVICE_ATTR_RO(hw_rev);
 941
 942static const char *hca_type_string(int hca_type)
 943{
 944	switch (hca_type) {
 945	case PCI_DEVICE_ID_MELLANOX_TAVOR:
 946		return "MT23108";
 947	case PCI_DEVICE_ID_MELLANOX_ARBEL_COMPAT:
 948		return "MT25208 (MT23108 compat mode)";
 949	case PCI_DEVICE_ID_MELLANOX_ARBEL:
 950		return "MT25208";
 951	case PCI_DEVICE_ID_MELLANOX_SINAI:
 952	case PCI_DEVICE_ID_MELLANOX_SINAI_OLD:
 953		return "MT25204";
 
 
 
 
 
 
 
 
 
 
 
 954	}
 955
 956	return "unknown";
 
 
 
 
 
 
 
 
 
 
 
 
 
 957}
 958
 959static ssize_t hca_type_show(struct device *device,
 960			     struct device_attribute *attr, char *buf)
 961{
 962	struct mthca_dev *dev =
 963		rdma_device_to_drv_device(device, struct mthca_dev, ib_dev);
 964
 965	return sysfs_emit(buf, "%s\n", hca_type_string(dev->pdev->device));
 966}
 967static DEVICE_ATTR_RO(hca_type);
 968
 969static ssize_t board_id_show(struct device *device,
 970			     struct device_attribute *attr, char *buf)
 971{
 972	struct mthca_dev *dev =
 973		rdma_device_to_drv_device(device, struct mthca_dev, ib_dev);
 
 
 
 
 974
 975	return sysfs_emit(buf, "%.*s\n", MTHCA_BOARD_ID_LEN, dev->board_id);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 976}
 977static DEVICE_ATTR_RO(board_id);
 978
 979static struct attribute *mthca_dev_attributes[] = {
 980	&dev_attr_hw_rev.attr,
 981	&dev_attr_hca_type.attr,
 982	&dev_attr_board_id.attr,
 983	NULL
 984};
 
 985
 986static const struct attribute_group mthca_attr_group = {
 987	.attrs = mthca_dev_attributes,
 
 
 
 
 
 
 
 
 988};
 989
 990static int mthca_init_node_data(struct mthca_dev *dev)
 991{
 992	struct ib_smp *in_mad;
 993	struct ib_smp *out_mad;
 994	int err = -ENOMEM;
 995
 996	in_mad  = kzalloc(sizeof *in_mad, GFP_KERNEL);
 997	out_mad = kmalloc(sizeof *out_mad, GFP_KERNEL);
 998	if (!in_mad || !out_mad)
 999		goto out;
1000
1001	ib_init_query_mad(in_mad);
1002	in_mad->attr_id = IB_SMP_ATTR_NODE_DESC;
1003
1004	err = mthca_MAD_IFC(dev, 1, 1,
1005			    1, NULL, NULL, in_mad, out_mad);
1006	if (err)
1007		goto out;
1008
1009	memcpy(dev->ib_dev.node_desc, out_mad->data, IB_DEVICE_NODE_DESC_MAX);
1010
1011	in_mad->attr_id = IB_SMP_ATTR_NODE_INFO;
1012
1013	err = mthca_MAD_IFC(dev, 1, 1,
1014			    1, NULL, NULL, in_mad, out_mad);
1015	if (err)
1016		goto out;
1017
1018	if (mthca_is_memfree(dev))
1019		dev->rev_id = be32_to_cpup((__be32 *) (out_mad->data + 32));
1020	memcpy(&dev->ib_dev.node_guid, out_mad->data + 12, 8);
1021
1022out:
1023	kfree(in_mad);
1024	kfree(out_mad);
1025	return err;
1026}
1027
1028static int mthca_port_immutable(struct ib_device *ibdev, u32 port_num,
1029			        struct ib_port_immutable *immutable)
1030{
1031	struct ib_port_attr attr;
1032	int err;
1033
1034	immutable->core_cap_flags = RDMA_CORE_PORT_IBA_IB;
1035
1036	err = ib_query_port(ibdev, port_num, &attr);
1037	if (err)
1038		return err;
1039
1040	immutable->pkey_tbl_len = attr.pkey_tbl_len;
1041	immutable->gid_tbl_len = attr.gid_tbl_len;
 
1042	immutable->max_mad_size = IB_MGMT_MAD_SIZE;
1043
1044	return 0;
1045}
1046
1047static void get_dev_fw_str(struct ib_device *device, char *str)
1048{
1049	struct mthca_dev *dev =
1050		container_of(device, struct mthca_dev, ib_dev);
1051	snprintf(str, IB_FW_VERSION_NAME_MAX, "%d.%d.%d",
1052		 (int) (dev->fw_ver >> 32),
1053		 (int) (dev->fw_ver >> 16) & 0xffff,
1054		 (int) dev->fw_ver & 0xffff);
1055}
1056
1057static const struct ib_device_ops mthca_dev_ops = {
1058	.owner = THIS_MODULE,
1059	.driver_id = RDMA_DRIVER_MTHCA,
1060	.uverbs_abi_ver = MTHCA_UVERBS_ABI_VERSION,
1061	.uverbs_no_driver_id_binding = 1,
1062
1063	.alloc_pd = mthca_alloc_pd,
1064	.alloc_ucontext = mthca_alloc_ucontext,
1065	.attach_mcast = mthca_multicast_attach,
1066	.create_ah = mthca_ah_create,
1067	.create_cq = mthca_create_cq,
1068	.create_qp = mthca_create_qp,
1069	.dealloc_pd = mthca_dealloc_pd,
1070	.dealloc_ucontext = mthca_dealloc_ucontext,
1071	.dereg_mr = mthca_dereg_mr,
1072	.destroy_ah = mthca_ah_destroy,
1073	.destroy_cq = mthca_destroy_cq,
1074	.destroy_qp = mthca_destroy_qp,
1075	.detach_mcast = mthca_multicast_detach,
1076	.device_group = &mthca_attr_group,
1077	.get_dev_fw_str = get_dev_fw_str,
1078	.get_dma_mr = mthca_get_dma_mr,
1079	.get_port_immutable = mthca_port_immutable,
1080	.mmap = mthca_mmap_uar,
1081	.modify_device = mthca_modify_device,
1082	.modify_port = mthca_modify_port,
1083	.modify_qp = mthca_modify_qp,
1084	.poll_cq = mthca_poll_cq,
1085	.process_mad = mthca_process_mad,
1086	.query_ah = mthca_ah_query,
1087	.query_device = mthca_query_device,
1088	.query_gid = mthca_query_gid,
1089	.query_pkey = mthca_query_pkey,
1090	.query_port = mthca_query_port,
1091	.query_qp = mthca_query_qp,
1092	.reg_user_mr = mthca_reg_user_mr,
1093	.resize_cq = mthca_resize_cq,
1094
1095	INIT_RDMA_OBJ_SIZE(ib_ah, mthca_ah, ibah),
1096	INIT_RDMA_OBJ_SIZE(ib_cq, mthca_cq, ibcq),
1097	INIT_RDMA_OBJ_SIZE(ib_pd, mthca_pd, ibpd),
1098	INIT_RDMA_OBJ_SIZE(ib_qp, mthca_qp, ibqp),
1099	INIT_RDMA_OBJ_SIZE(ib_ucontext, mthca_ucontext, ibucontext),
1100};
1101
1102static const struct ib_device_ops mthca_dev_arbel_srq_ops = {
1103	.create_srq = mthca_create_srq,
1104	.destroy_srq = mthca_destroy_srq,
1105	.modify_srq = mthca_modify_srq,
1106	.post_srq_recv = mthca_arbel_post_srq_recv,
1107	.query_srq = mthca_query_srq,
1108
1109	INIT_RDMA_OBJ_SIZE(ib_srq, mthca_srq, ibsrq),
1110};
1111
1112static const struct ib_device_ops mthca_dev_tavor_srq_ops = {
1113	.create_srq = mthca_create_srq,
1114	.destroy_srq = mthca_destroy_srq,
1115	.modify_srq = mthca_modify_srq,
1116	.post_srq_recv = mthca_tavor_post_srq_recv,
1117	.query_srq = mthca_query_srq,
1118
1119	INIT_RDMA_OBJ_SIZE(ib_srq, mthca_srq, ibsrq),
1120};
1121
1122static const struct ib_device_ops mthca_dev_arbel_ops = {
1123	.post_recv = mthca_arbel_post_receive,
1124	.post_send = mthca_arbel_post_send,
1125	.req_notify_cq = mthca_arbel_arm_cq,
1126};
1127
1128static const struct ib_device_ops mthca_dev_tavor_ops = {
1129	.post_recv = mthca_tavor_post_receive,
1130	.post_send = mthca_tavor_post_send,
1131	.req_notify_cq = mthca_tavor_arm_cq,
1132};
1133
1134int mthca_register_device(struct mthca_dev *dev)
1135{
1136	int ret;
 
1137
1138	ret = mthca_init_node_data(dev);
1139	if (ret)
1140		return ret;
1141
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1142	dev->ib_dev.node_type            = RDMA_NODE_IB_CA;
1143	dev->ib_dev.phys_port_cnt        = dev->limits.num_ports;
1144	dev->ib_dev.num_comp_vectors     = 1;
1145	dev->ib_dev.dev.parent           = &dev->pdev->dev;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1146
1147	if (dev->mthca_flags & MTHCA_FLAG_SRQ) {
 
 
 
 
 
 
 
 
 
 
1148		if (mthca_is_memfree(dev))
1149			ib_set_device_ops(&dev->ib_dev,
1150					  &mthca_dev_arbel_srq_ops);
1151		else
1152			ib_set_device_ops(&dev->ib_dev,
1153					  &mthca_dev_tavor_srq_ops);
1154	}
1155
1156	ib_set_device_ops(&dev->ib_dev, &mthca_dev_ops);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1157
1158	if (mthca_is_memfree(dev))
1159		ib_set_device_ops(&dev->ib_dev, &mthca_dev_arbel_ops);
1160	else
1161		ib_set_device_ops(&dev->ib_dev, &mthca_dev_tavor_ops);
 
 
 
 
 
 
 
 
 
1162
1163	mutex_init(&dev->cap_mask_mutex);
1164
1165	ret = ib_register_device(&dev->ib_dev, "mthca%d", &dev->pdev->dev);
1166	if (ret)
1167		return ret;
 
 
 
 
 
 
 
 
 
1168
1169	mthca_start_catas_poll(dev);
1170
1171	return 0;
1172}
1173
1174void mthca_unregister_device(struct mthca_dev *dev)
1175{
1176	mthca_stop_catas_poll(dev);
1177	ib_unregister_device(&dev->ib_dev);
1178}
v4.6
   1/*
   2 * Copyright (c) 2004, 2005 Topspin Communications.  All rights reserved.
   3 * Copyright (c) 2005 Sun Microsystems, Inc. All rights reserved.
   4 * Copyright (c) 2005, 2006 Cisco Systems.  All rights reserved.
   5 * Copyright (c) 2005 Mellanox Technologies. All rights reserved.
   6 * Copyright (c) 2004 Voltaire, Inc. All rights reserved.
   7 *
   8 * This software is available to you under a choice of one of two
   9 * licenses.  You may choose to be licensed under the terms of the GNU
  10 * General Public License (GPL) Version 2, available from the file
  11 * COPYING in the main directory of this source tree, or the
  12 * OpenIB.org BSD license below:
  13 *
  14 *     Redistribution and use in source and binary forms, with or
  15 *     without modification, are permitted provided that the following
  16 *     conditions are met:
  17 *
  18 *      - Redistributions of source code must retain the above
  19 *        copyright notice, this list of conditions and the following
  20 *        disclaimer.
  21 *
  22 *      - Redistributions in binary form must reproduce the above
  23 *        copyright notice, this list of conditions and the following
  24 *        disclaimer in the documentation and/or other materials
  25 *        provided with the distribution.
  26 *
  27 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
  28 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
  29 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
  30 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
  31 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
  32 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
  33 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
  34 * SOFTWARE.
  35 */
  36
  37#include <rdma/ib_smi.h>
  38#include <rdma/ib_umem.h>
  39#include <rdma/ib_user_verbs.h>
 
  40
  41#include <linux/sched.h>
  42#include <linux/slab.h>
  43#include <linux/stat.h>
  44#include <linux/mm.h>
  45#include <linux/export.h>
  46
  47#include "mthca_dev.h"
  48#include "mthca_cmd.h"
  49#include "mthca_user.h"
  50#include "mthca_memfree.h"
  51
  52static void init_query_mad(struct ib_smp *mad)
  53{
  54	mad->base_version  = 1;
  55	mad->mgmt_class    = IB_MGMT_CLASS_SUBN_LID_ROUTED;
  56	mad->class_version = 1;
  57	mad->method    	   = IB_MGMT_METHOD_GET;
  58}
  59
  60static int mthca_query_device(struct ib_device *ibdev, struct ib_device_attr *props,
  61			      struct ib_udata *uhw)
  62{
  63	struct ib_smp *in_mad  = NULL;
  64	struct ib_smp *out_mad = NULL;
  65	int err = -ENOMEM;
  66	struct mthca_dev *mdev = to_mdev(ibdev);
  67
  68	if (uhw->inlen || uhw->outlen)
  69		return -EINVAL;
  70
  71	in_mad  = kzalloc(sizeof *in_mad, GFP_KERNEL);
  72	out_mad = kmalloc(sizeof *out_mad, GFP_KERNEL);
  73	if (!in_mad || !out_mad)
  74		goto out;
  75
  76	memset(props, 0, sizeof *props);
  77
  78	props->fw_ver              = mdev->fw_ver;
  79
  80	init_query_mad(in_mad);
  81	in_mad->attr_id = IB_SMP_ATTR_NODE_INFO;
  82
  83	err = mthca_MAD_IFC(mdev, 1, 1,
  84			    1, NULL, NULL, in_mad, out_mad);
  85	if (err)
  86		goto out;
  87
  88	props->device_cap_flags    = mdev->device_cap_flags;
  89	props->vendor_id           = be32_to_cpup((__be32 *) (out_mad->data + 36)) &
  90		0xffffff;
  91	props->vendor_part_id      = be16_to_cpup((__be16 *) (out_mad->data + 30));
  92	props->hw_ver              = be32_to_cpup((__be32 *) (out_mad->data + 32));
  93	memcpy(&props->sys_image_guid, out_mad->data +  4, 8);
  94
  95	props->max_mr_size         = ~0ull;
  96	props->page_size_cap       = mdev->limits.page_size_cap;
  97	props->max_qp              = mdev->limits.num_qps - mdev->limits.reserved_qps;
  98	props->max_qp_wr           = mdev->limits.max_wqes;
  99	props->max_sge             = mdev->limits.max_sg;
 100	props->max_sge_rd          = props->max_sge;
 
 101	props->max_cq              = mdev->limits.num_cqs - mdev->limits.reserved_cqs;
 102	props->max_cqe             = mdev->limits.max_cqes;
 103	props->max_mr              = mdev->limits.num_mpts - mdev->limits.reserved_mrws;
 104	props->max_pd              = mdev->limits.num_pds - mdev->limits.reserved_pds;
 105	props->max_qp_rd_atom      = 1 << mdev->qp_table.rdb_shift;
 106	props->max_qp_init_rd_atom = mdev->limits.max_qp_init_rdma;
 107	props->max_res_rd_atom     = props->max_qp_rd_atom * props->max_qp;
 108	props->max_srq             = mdev->limits.num_srqs - mdev->limits.reserved_srqs;
 109	props->max_srq_wr          = mdev->limits.max_srq_wqes;
 110	props->max_srq_sge         = mdev->limits.max_srq_sge;
 111	props->local_ca_ack_delay  = mdev->limits.local_ca_ack_delay;
 112	props->atomic_cap          = mdev->limits.flags & DEV_LIM_FLAG_ATOMIC ?
 113					IB_ATOMIC_HCA : IB_ATOMIC_NONE;
 114	props->max_pkeys           = mdev->limits.pkey_table_len;
 115	props->max_mcast_grp       = mdev->limits.num_mgms + mdev->limits.num_amgms;
 116	props->max_mcast_qp_attach = MTHCA_QP_PER_MGM;
 117	props->max_total_mcast_qp_attach = props->max_mcast_qp_attach *
 118					   props->max_mcast_grp;
 119	/*
 120	 * If Sinai memory key optimization is being used, then only
 121	 * the 8-bit key portion will change.  For other HCAs, the
 122	 * unused index bits will also be used for FMR remapping.
 123	 */
 124	if (mdev->mthca_flags & MTHCA_FLAG_SINAI_OPT)
 125		props->max_map_per_fmr = 255;
 126	else
 127		props->max_map_per_fmr =
 128			(1 << (32 - ilog2(mdev->limits.num_mpts))) - 1;
 129
 130	err = 0;
 131 out:
 132	kfree(in_mad);
 133	kfree(out_mad);
 134	return err;
 135}
 136
 137static int mthca_query_port(struct ib_device *ibdev,
 138			    u8 port, struct ib_port_attr *props)
 139{
 140	struct ib_smp *in_mad  = NULL;
 141	struct ib_smp *out_mad = NULL;
 142	int err = -ENOMEM;
 143
 144	in_mad  = kzalloc(sizeof *in_mad, GFP_KERNEL);
 145	out_mad = kmalloc(sizeof *out_mad, GFP_KERNEL);
 146	if (!in_mad || !out_mad)
 147		goto out;
 148
 149	memset(props, 0, sizeof *props);
 150
 151	init_query_mad(in_mad);
 152	in_mad->attr_id  = IB_SMP_ATTR_PORT_INFO;
 153	in_mad->attr_mod = cpu_to_be32(port);
 154
 155	err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1,
 156			    port, NULL, NULL, in_mad, out_mad);
 157	if (err)
 158		goto out;
 159
 160	props->lid               = be16_to_cpup((__be16 *) (out_mad->data + 16));
 161	props->lmc               = out_mad->data[34] & 0x7;
 162	props->sm_lid            = be16_to_cpup((__be16 *) (out_mad->data + 18));
 163	props->sm_sl             = out_mad->data[36] & 0xf;
 164	props->state             = out_mad->data[32] & 0xf;
 165	props->phys_state        = out_mad->data[33] >> 4;
 166	props->port_cap_flags    = be32_to_cpup((__be32 *) (out_mad->data + 20));
 167	props->gid_tbl_len       = to_mdev(ibdev)->limits.gid_table_len;
 168	props->max_msg_sz        = 0x80000000;
 169	props->pkey_tbl_len      = to_mdev(ibdev)->limits.pkey_table_len;
 170	props->bad_pkey_cntr     = be16_to_cpup((__be16 *) (out_mad->data + 46));
 171	props->qkey_viol_cntr    = be16_to_cpup((__be16 *) (out_mad->data + 48));
 172	props->active_width      = out_mad->data[31] & 0xf;
 173	props->active_speed      = out_mad->data[35] >> 4;
 174	props->max_mtu           = out_mad->data[41] & 0xf;
 175	props->active_mtu        = out_mad->data[36] >> 4;
 176	props->subnet_timeout    = out_mad->data[51] & 0x1f;
 177	props->max_vl_num        = out_mad->data[37] >> 4;
 178	props->init_type_reply   = out_mad->data[41] >> 4;
 179
 180 out:
 181	kfree(in_mad);
 182	kfree(out_mad);
 183	return err;
 184}
 185
 186static int mthca_modify_device(struct ib_device *ibdev,
 187			       int mask,
 188			       struct ib_device_modify *props)
 189{
 190	if (mask & ~IB_DEVICE_MODIFY_NODE_DESC)
 191		return -EOPNOTSUPP;
 192
 193	if (mask & IB_DEVICE_MODIFY_NODE_DESC) {
 194		if (mutex_lock_interruptible(&to_mdev(ibdev)->cap_mask_mutex))
 195			return -ERESTARTSYS;
 196		memcpy(ibdev->node_desc, props->node_desc, 64);
 
 197		mutex_unlock(&to_mdev(ibdev)->cap_mask_mutex);
 198	}
 199
 200	return 0;
 201}
 202
 203static int mthca_modify_port(struct ib_device *ibdev,
 204			     u8 port, int port_modify_mask,
 205			     struct ib_port_modify *props)
 206{
 207	struct mthca_set_ib_param set_ib;
 208	struct ib_port_attr attr;
 209	int err;
 210
 211	if (mutex_lock_interruptible(&to_mdev(ibdev)->cap_mask_mutex))
 212		return -ERESTARTSYS;
 213
 214	err = mthca_query_port(ibdev, port, &attr);
 215	if (err)
 216		goto out;
 217
 218	set_ib.set_si_guid     = 0;
 219	set_ib.reset_qkey_viol = !!(port_modify_mask & IB_PORT_RESET_QKEY_CNTR);
 220
 221	set_ib.cap_mask = (attr.port_cap_flags | props->set_port_cap_mask) &
 222		~props->clr_port_cap_mask;
 223
 224	err = mthca_SET_IB(to_mdev(ibdev), &set_ib, port);
 225	if (err)
 226		goto out;
 227out:
 228	mutex_unlock(&to_mdev(ibdev)->cap_mask_mutex);
 229	return err;
 230}
 231
 232static int mthca_query_pkey(struct ib_device *ibdev,
 233			    u8 port, u16 index, u16 *pkey)
 234{
 235	struct ib_smp *in_mad  = NULL;
 236	struct ib_smp *out_mad = NULL;
 237	int err = -ENOMEM;
 238
 239	in_mad  = kzalloc(sizeof *in_mad, GFP_KERNEL);
 240	out_mad = kmalloc(sizeof *out_mad, GFP_KERNEL);
 241	if (!in_mad || !out_mad)
 242		goto out;
 243
 244	init_query_mad(in_mad);
 245	in_mad->attr_id  = IB_SMP_ATTR_PKEY_TABLE;
 246	in_mad->attr_mod = cpu_to_be32(index / 32);
 247
 248	err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1,
 249			    port, NULL, NULL, in_mad, out_mad);
 250	if (err)
 251		goto out;
 252
 253	*pkey = be16_to_cpu(((__be16 *) out_mad->data)[index % 32]);
 254
 255 out:
 256	kfree(in_mad);
 257	kfree(out_mad);
 258	return err;
 259}
 260
 261static int mthca_query_gid(struct ib_device *ibdev, u8 port,
 262			   int index, union ib_gid *gid)
 263{
 264	struct ib_smp *in_mad  = NULL;
 265	struct ib_smp *out_mad = NULL;
 266	int err = -ENOMEM;
 267
 268	in_mad  = kzalloc(sizeof *in_mad, GFP_KERNEL);
 269	out_mad = kmalloc(sizeof *out_mad, GFP_KERNEL);
 270	if (!in_mad || !out_mad)
 271		goto out;
 272
 273	init_query_mad(in_mad);
 274	in_mad->attr_id  = IB_SMP_ATTR_PORT_INFO;
 275	in_mad->attr_mod = cpu_to_be32(port);
 276
 277	err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1,
 278			    port, NULL, NULL, in_mad, out_mad);
 279	if (err)
 280		goto out;
 281
 282	memcpy(gid->raw, out_mad->data + 8, 8);
 283
 284	init_query_mad(in_mad);
 285	in_mad->attr_id  = IB_SMP_ATTR_GUID_INFO;
 286	in_mad->attr_mod = cpu_to_be32(index / 8);
 287
 288	err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1,
 289			    port, NULL, NULL, in_mad, out_mad);
 290	if (err)
 291		goto out;
 292
 293	memcpy(gid->raw + 8, out_mad->data + (index % 8) * 8, 8);
 294
 295 out:
 296	kfree(in_mad);
 297	kfree(out_mad);
 298	return err;
 299}
 300
 301static struct ib_ucontext *mthca_alloc_ucontext(struct ib_device *ibdev,
 302						struct ib_udata *udata)
 303{
 304	struct mthca_alloc_ucontext_resp uresp;
 305	struct mthca_ucontext           *context;
 
 306	int                              err;
 307
 308	if (!(to_mdev(ibdev)->active))
 309		return ERR_PTR(-EAGAIN);
 310
 311	memset(&uresp, 0, sizeof uresp);
 312
 313	uresp.qp_tab_size = to_mdev(ibdev)->limits.num_qps;
 314	if (mthca_is_memfree(to_mdev(ibdev)))
 315		uresp.uarc_size = to_mdev(ibdev)->uar_table.uarc_size;
 316	else
 317		uresp.uarc_size = 0;
 318
 319	context = kmalloc(sizeof *context, GFP_KERNEL);
 320	if (!context)
 321		return ERR_PTR(-ENOMEM);
 322
 323	err = mthca_uar_alloc(to_mdev(ibdev), &context->uar);
 324	if (err) {
 325		kfree(context);
 326		return ERR_PTR(err);
 327	}
 328
 329	context->db_tab = mthca_init_user_db_tab(to_mdev(ibdev));
 330	if (IS_ERR(context->db_tab)) {
 331		err = PTR_ERR(context->db_tab);
 332		mthca_uar_free(to_mdev(ibdev), &context->uar);
 333		kfree(context);
 334		return ERR_PTR(err);
 335	}
 336
 337	if (ib_copy_to_udata(udata, &uresp, sizeof uresp)) {
 338		mthca_cleanup_user_db_tab(to_mdev(ibdev), &context->uar, context->db_tab);
 339		mthca_uar_free(to_mdev(ibdev), &context->uar);
 340		kfree(context);
 341		return ERR_PTR(-EFAULT);
 342	}
 343
 344	context->reg_mr_warned = 0;
 345
 346	return &context->ibucontext;
 347}
 348
 349static int mthca_dealloc_ucontext(struct ib_ucontext *context)
 350{
 351	mthca_cleanup_user_db_tab(to_mdev(context->device), &to_mucontext(context)->uar,
 352				  to_mucontext(context)->db_tab);
 353	mthca_uar_free(to_mdev(context->device), &to_mucontext(context)->uar);
 354	kfree(to_mucontext(context));
 355
 356	return 0;
 357}
 358
 359static int mthca_mmap_uar(struct ib_ucontext *context,
 360			  struct vm_area_struct *vma)
 361{
 362	if (vma->vm_end - vma->vm_start != PAGE_SIZE)
 363		return -EINVAL;
 364
 365	vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
 366
 367	if (io_remap_pfn_range(vma, vma->vm_start,
 368			       to_mucontext(context)->uar.pfn,
 369			       PAGE_SIZE, vma->vm_page_prot))
 370		return -EAGAIN;
 371
 372	return 0;
 373}
 374
 375static struct ib_pd *mthca_alloc_pd(struct ib_device *ibdev,
 376				    struct ib_ucontext *context,
 377				    struct ib_udata *udata)
 378{
 379	struct mthca_pd *pd;
 
 380	int err;
 381
 382	pd = kmalloc(sizeof *pd, GFP_KERNEL);
 383	if (!pd)
 384		return ERR_PTR(-ENOMEM);
 385
 386	err = mthca_pd_alloc(to_mdev(ibdev), !context, pd);
 387	if (err) {
 388		kfree(pd);
 389		return ERR_PTR(err);
 390	}
 391
 392	if (context) {
 393		if (ib_copy_to_udata(udata, &pd->pd_num, sizeof (__u32))) {
 394			mthca_pd_free(to_mdev(ibdev), pd);
 395			kfree(pd);
 396			return ERR_PTR(-EFAULT);
 397		}
 398	}
 399
 400	return &pd->ibpd;
 401}
 402
 403static int mthca_dealloc_pd(struct ib_pd *pd)
 404{
 405	mthca_pd_free(to_mdev(pd->device), to_mpd(pd));
 406	kfree(pd);
 407
 408	return 0;
 409}
 410
 411static struct ib_ah *mthca_ah_create(struct ib_pd *pd,
 412				     struct ib_ah_attr *ah_attr)
 
 
 413{
 414	int err;
 415	struct mthca_ah *ah;
 416
 417	ah = kmalloc(sizeof *ah, GFP_ATOMIC);
 418	if (!ah)
 419		return ERR_PTR(-ENOMEM);
 420
 421	err = mthca_create_ah(to_mdev(pd->device), to_mpd(pd), ah_attr, ah);
 422	if (err) {
 423		kfree(ah);
 424		return ERR_PTR(err);
 425	}
 426
 427	return &ah->ibah;
 428}
 429
 430static int mthca_ah_destroy(struct ib_ah *ah)
 431{
 432	mthca_destroy_ah(to_mdev(ah->device), to_mah(ah));
 433	kfree(ah);
 434
 435	return 0;
 436}
 437
 438static struct ib_srq *mthca_create_srq(struct ib_pd *pd,
 439				       struct ib_srq_init_attr *init_attr,
 440				       struct ib_udata *udata)
 441{
 442	struct mthca_create_srq ucmd;
 443	struct mthca_ucontext *context = NULL;
 444	struct mthca_srq *srq;
 
 445	int err;
 446
 447	if (init_attr->srq_type != IB_SRQT_BASIC)
 448		return ERR_PTR(-ENOSYS);
 449
 450	srq = kmalloc(sizeof *srq, GFP_KERNEL);
 451	if (!srq)
 452		return ERR_PTR(-ENOMEM);
 453
 454	if (pd->uobject) {
 455		context = to_mucontext(pd->uobject->context);
 456
 457		if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) {
 458			err = -EFAULT;
 459			goto err_free;
 460		}
 461
 462		err = mthca_map_user_db(to_mdev(pd->device), &context->uar,
 463					context->db_tab, ucmd.db_index,
 464					ucmd.db_page);
 465
 466		if (err)
 467			goto err_free;
 468
 469		srq->mr.ibmr.lkey = ucmd.lkey;
 470		srq->db_index     = ucmd.db_index;
 471	}
 472
 473	err = mthca_alloc_srq(to_mdev(pd->device), to_mpd(pd),
 474			      &init_attr->attr, srq);
 475
 476	if (err && pd->uobject)
 477		mthca_unmap_user_db(to_mdev(pd->device), &context->uar,
 478				    context->db_tab, ucmd.db_index);
 479
 480	if (err)
 481		goto err_free;
 482
 483	if (context && ib_copy_to_udata(udata, &srq->srqn, sizeof (__u32))) {
 484		mthca_free_srq(to_mdev(pd->device), srq);
 485		err = -EFAULT;
 486		goto err_free;
 487	}
 488
 489	return &srq->ibsrq;
 490
 491err_free:
 492	kfree(srq);
 493
 494	return ERR_PTR(err);
 495}
 496
 497static int mthca_destroy_srq(struct ib_srq *srq)
 498{
 499	struct mthca_ucontext *context;
 500
 501	if (srq->uobject) {
 502		context = to_mucontext(srq->uobject->context);
 
 
 503
 504		mthca_unmap_user_db(to_mdev(srq->device), &context->uar,
 505				    context->db_tab, to_msrq(srq)->db_index);
 506	}
 507
 508	mthca_free_srq(to_mdev(srq->device), to_msrq(srq));
 509	kfree(srq);
 510
 511	return 0;
 512}
 513
 514static struct ib_qp *mthca_create_qp(struct ib_pd *pd,
 515				     struct ib_qp_init_attr *init_attr,
 516				     struct ib_udata *udata)
 517{
 
 
 518	struct mthca_create_qp ucmd;
 519	struct mthca_qp *qp;
 
 520	int err;
 521
 522	if (init_attr->create_flags)
 523		return ERR_PTR(-EINVAL);
 524
 525	switch (init_attr->qp_type) {
 526	case IB_QPT_RC:
 527	case IB_QPT_UC:
 528	case IB_QPT_UD:
 529	{
 530		struct mthca_ucontext *context;
 
 
 531
 532		qp = kmalloc(sizeof *qp, GFP_KERNEL);
 533		if (!qp)
 534			return ERR_PTR(-ENOMEM);
 535
 536		if (pd->uobject) {
 537			context = to_mucontext(pd->uobject->context);
 538
 539			if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) {
 540				kfree(qp);
 541				return ERR_PTR(-EFAULT);
 542			}
 543
 544			err = mthca_map_user_db(to_mdev(pd->device), &context->uar,
 545						context->db_tab,
 546						ucmd.sq_db_index, ucmd.sq_db_page);
 547			if (err) {
 548				kfree(qp);
 549				return ERR_PTR(err);
 550			}
 551
 552			err = mthca_map_user_db(to_mdev(pd->device), &context->uar,
 553						context->db_tab,
 554						ucmd.rq_db_index, ucmd.rq_db_page);
 
 555			if (err) {
 556				mthca_unmap_user_db(to_mdev(pd->device),
 557						    &context->uar,
 558						    context->db_tab,
 559						    ucmd.sq_db_index);
 560				kfree(qp);
 561				return ERR_PTR(err);
 562			}
 563
 564			qp->mr.ibmr.lkey = ucmd.lkey;
 565			qp->sq.db_index  = ucmd.sq_db_index;
 566			qp->rq.db_index  = ucmd.rq_db_index;
 567		}
 568
 569		err = mthca_alloc_qp(to_mdev(pd->device), to_mpd(pd),
 570				     to_mcq(init_attr->send_cq),
 571				     to_mcq(init_attr->recv_cq),
 572				     init_attr->qp_type, init_attr->sq_sig_type,
 573				     &init_attr->cap, qp);
 574
 575		if (err && pd->uobject) {
 576			context = to_mucontext(pd->uobject->context);
 577
 578			mthca_unmap_user_db(to_mdev(pd->device),
 579					    &context->uar,
 580					    context->db_tab,
 581					    ucmd.sq_db_index);
 582			mthca_unmap_user_db(to_mdev(pd->device),
 583					    &context->uar,
 584					    context->db_tab,
 585					    ucmd.rq_db_index);
 586		}
 587
 588		qp->ibqp.qp_num = qp->qpn;
 589		break;
 590	}
 591	case IB_QPT_SMI:
 592	case IB_QPT_GSI:
 593	{
 594		/* Don't allow userspace to create special QPs */
 595		if (pd->uobject)
 596			return ERR_PTR(-EINVAL);
 597
 598		qp = kmalloc(sizeof (struct mthca_sqp), GFP_KERNEL);
 599		if (!qp)
 600			return ERR_PTR(-ENOMEM);
 601
 602		qp->ibqp.qp_num = init_attr->qp_type == IB_QPT_SMI ? 0 : 1;
 603
 604		err = mthca_alloc_sqp(to_mdev(pd->device), to_mpd(pd),
 605				      to_mcq(init_attr->send_cq),
 606				      to_mcq(init_attr->recv_cq),
 607				      init_attr->sq_sig_type, &init_attr->cap,
 608				      qp->ibqp.qp_num, init_attr->port_num,
 609				      to_msqp(qp));
 610		break;
 611	}
 612	default:
 613		/* Don't support raw QPs */
 614		return ERR_PTR(-ENOSYS);
 615	}
 616
 617	if (err) {
 618		kfree(qp);
 619		return ERR_PTR(err);
 620	}
 621
 622	init_attr->cap.max_send_wr     = qp->sq.max;
 623	init_attr->cap.max_recv_wr     = qp->rq.max;
 624	init_attr->cap.max_send_sge    = qp->sq.max_gs;
 625	init_attr->cap.max_recv_sge    = qp->rq.max_gs;
 626	init_attr->cap.max_inline_data = qp->max_inline_data;
 627
 628	return &qp->ibqp;
 629}
 630
 631static int mthca_destroy_qp(struct ib_qp *qp)
 632{
 633	if (qp->uobject) {
 
 
 
 
 
 
 634		mthca_unmap_user_db(to_mdev(qp->device),
 635				    &to_mucontext(qp->uobject->context)->uar,
 636				    to_mucontext(qp->uobject->context)->db_tab,
 637				    to_mqp(qp)->sq.db_index);
 638		mthca_unmap_user_db(to_mdev(qp->device),
 639				    &to_mucontext(qp->uobject->context)->uar,
 640				    to_mucontext(qp->uobject->context)->db_tab,
 641				    to_mqp(qp)->rq.db_index);
 642	}
 643	mthca_free_qp(to_mdev(qp->device), to_mqp(qp));
 644	kfree(qp);
 645	return 0;
 646}
 647
 648static struct ib_cq *mthca_create_cq(struct ib_device *ibdev,
 649				     const struct ib_cq_init_attr *attr,
 650				     struct ib_ucontext *context,
 651				     struct ib_udata *udata)
 652{
 
 653	int entries = attr->cqe;
 654	struct mthca_create_cq ucmd;
 655	struct mthca_cq *cq;
 656	int nent;
 657	int err;
 
 
 658
 659	if (attr->flags)
 660		return ERR_PTR(-EINVAL);
 661
 662	if (entries < 1 || entries > to_mdev(ibdev)->limits.max_cqes)
 663		return ERR_PTR(-EINVAL);
 664
 665	if (context) {
 666		if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd))
 667			return ERR_PTR(-EFAULT);
 668
 669		err = mthca_map_user_db(to_mdev(ibdev), &to_mucontext(context)->uar,
 670					to_mucontext(context)->db_tab,
 671					ucmd.set_db_index, ucmd.set_db_page);
 672		if (err)
 673			return ERR_PTR(err);
 674
 675		err = mthca_map_user_db(to_mdev(ibdev), &to_mucontext(context)->uar,
 676					to_mucontext(context)->db_tab,
 677					ucmd.arm_db_index, ucmd.arm_db_page);
 678		if (err)
 679			goto err_unmap_set;
 680	}
 681
 682	cq = kmalloc(sizeof *cq, GFP_KERNEL);
 683	if (!cq) {
 684		err = -ENOMEM;
 685		goto err_unmap_arm;
 686	}
 687
 688	if (context) {
 689		cq->buf.mr.ibmr.lkey = ucmd.lkey;
 690		cq->set_ci_db_index  = ucmd.set_db_index;
 691		cq->arm_db_index     = ucmd.arm_db_index;
 692	}
 693
 694	for (nent = 1; nent <= entries; nent <<= 1)
 695		; /* nothing */
 696
 697	err = mthca_init_cq(to_mdev(ibdev), nent,
 698			    context ? to_mucontext(context) : NULL,
 699			    context ? ucmd.pdn : to_mdev(ibdev)->driver_pd.pd_num,
 700			    cq);
 701	if (err)
 702		goto err_free;
 703
 704	if (context && ib_copy_to_udata(udata, &cq->cqn, sizeof (__u32))) {
 705		mthca_free_cq(to_mdev(ibdev), cq);
 706		err = -EFAULT;
 707		goto err_free;
 708	}
 709
 710	cq->resize_buf = NULL;
 711
 712	return &cq->ibcq;
 713
 714err_free:
 715	kfree(cq);
 716
 717err_unmap_arm:
 718	if (context)
 719		mthca_unmap_user_db(to_mdev(ibdev), &to_mucontext(context)->uar,
 720				    to_mucontext(context)->db_tab, ucmd.arm_db_index);
 721
 722err_unmap_set:
 723	if (context)
 724		mthca_unmap_user_db(to_mdev(ibdev), &to_mucontext(context)->uar,
 725				    to_mucontext(context)->db_tab, ucmd.set_db_index);
 726
 727	return ERR_PTR(err);
 728}
 729
 730static int mthca_alloc_resize_buf(struct mthca_dev *dev, struct mthca_cq *cq,
 731				  int entries)
 732{
 733	int ret;
 734
 735	spin_lock_irq(&cq->lock);
 736	if (cq->resize_buf) {
 737		ret = -EBUSY;
 738		goto unlock;
 739	}
 740
 741	cq->resize_buf = kmalloc(sizeof *cq->resize_buf, GFP_ATOMIC);
 742	if (!cq->resize_buf) {
 743		ret = -ENOMEM;
 744		goto unlock;
 745	}
 746
 747	cq->resize_buf->state = CQ_RESIZE_ALLOC;
 748
 749	ret = 0;
 750
 751unlock:
 752	spin_unlock_irq(&cq->lock);
 753
 754	if (ret)
 755		return ret;
 756
 757	ret = mthca_alloc_cq_buf(dev, &cq->resize_buf->buf, entries);
 758	if (ret) {
 759		spin_lock_irq(&cq->lock);
 760		kfree(cq->resize_buf);
 761		cq->resize_buf = NULL;
 762		spin_unlock_irq(&cq->lock);
 763		return ret;
 764	}
 765
 766	cq->resize_buf->cqe = entries - 1;
 767
 768	spin_lock_irq(&cq->lock);
 769	cq->resize_buf->state = CQ_RESIZE_READY;
 770	spin_unlock_irq(&cq->lock);
 771
 772	return 0;
 773}
 774
 775static int mthca_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata)
 776{
 777	struct mthca_dev *dev = to_mdev(ibcq->device);
 778	struct mthca_cq *cq = to_mcq(ibcq);
 779	struct mthca_resize_cq ucmd;
 780	u32 lkey;
 781	int ret;
 782
 783	if (entries < 1 || entries > dev->limits.max_cqes)
 784		return -EINVAL;
 785
 786	mutex_lock(&cq->mutex);
 787
 788	entries = roundup_pow_of_two(entries + 1);
 789	if (entries == ibcq->cqe + 1) {
 790		ret = 0;
 791		goto out;
 792	}
 793
 794	if (cq->is_kernel) {
 795		ret = mthca_alloc_resize_buf(dev, cq, entries);
 796		if (ret)
 797			goto out;
 798		lkey = cq->resize_buf->buf.mr.ibmr.lkey;
 799	} else {
 800		if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) {
 801			ret = -EFAULT;
 802			goto out;
 803		}
 804		lkey = ucmd.lkey;
 805	}
 806
 807	ret = mthca_RESIZE_CQ(dev, cq->cqn, lkey, ilog2(entries));
 808
 809	if (ret) {
 810		if (cq->resize_buf) {
 811			mthca_free_cq_buf(dev, &cq->resize_buf->buf,
 812					  cq->resize_buf->cqe);
 813			kfree(cq->resize_buf);
 814			spin_lock_irq(&cq->lock);
 815			cq->resize_buf = NULL;
 816			spin_unlock_irq(&cq->lock);
 817		}
 818		goto out;
 819	}
 820
 821	if (cq->is_kernel) {
 822		struct mthca_cq_buf tbuf;
 823		int tcqe;
 824
 825		spin_lock_irq(&cq->lock);
 826		if (cq->resize_buf->state == CQ_RESIZE_READY) {
 827			mthca_cq_resize_copy_cqes(cq);
 828			tbuf         = cq->buf;
 829			tcqe         = cq->ibcq.cqe;
 830			cq->buf      = cq->resize_buf->buf;
 831			cq->ibcq.cqe = cq->resize_buf->cqe;
 832		} else {
 833			tbuf = cq->resize_buf->buf;
 834			tcqe = cq->resize_buf->cqe;
 835		}
 836
 837		kfree(cq->resize_buf);
 838		cq->resize_buf = NULL;
 839		spin_unlock_irq(&cq->lock);
 840
 841		mthca_free_cq_buf(dev, &tbuf, tcqe);
 842	} else
 843		ibcq->cqe = entries - 1;
 844
 845out:
 846	mutex_unlock(&cq->mutex);
 847
 848	return ret;
 849}
 850
 851static int mthca_destroy_cq(struct ib_cq *cq)
 852{
 853	if (cq->uobject) {
 
 
 
 
 
 
 854		mthca_unmap_user_db(to_mdev(cq->device),
 855				    &to_mucontext(cq->uobject->context)->uar,
 856				    to_mucontext(cq->uobject->context)->db_tab,
 857				    to_mcq(cq)->arm_db_index);
 858		mthca_unmap_user_db(to_mdev(cq->device),
 859				    &to_mucontext(cq->uobject->context)->uar,
 860				    to_mucontext(cq->uobject->context)->db_tab,
 861				    to_mcq(cq)->set_ci_db_index);
 862	}
 863	mthca_free_cq(to_mdev(cq->device), to_mcq(cq));
 864	kfree(cq);
 865
 866	return 0;
 867}
 868
 869static inline u32 convert_access(int acc)
 870{
 871	return (acc & IB_ACCESS_REMOTE_ATOMIC ? MTHCA_MPT_FLAG_ATOMIC       : 0) |
 872	       (acc & IB_ACCESS_REMOTE_WRITE  ? MTHCA_MPT_FLAG_REMOTE_WRITE : 0) |
 873	       (acc & IB_ACCESS_REMOTE_READ   ? MTHCA_MPT_FLAG_REMOTE_READ  : 0) |
 874	       (acc & IB_ACCESS_LOCAL_WRITE   ? MTHCA_MPT_FLAG_LOCAL_WRITE  : 0) |
 875	       MTHCA_MPT_FLAG_LOCAL_READ;
 876}
 877
 878static struct ib_mr *mthca_get_dma_mr(struct ib_pd *pd, int acc)
 879{
 880	struct mthca_mr *mr;
 881	int err;
 882
 883	mr = kmalloc(sizeof *mr, GFP_KERNEL);
 884	if (!mr)
 885		return ERR_PTR(-ENOMEM);
 886
 887	err = mthca_mr_alloc_notrans(to_mdev(pd->device),
 888				     to_mpd(pd)->pd_num,
 889				     convert_access(acc), mr);
 890
 891	if (err) {
 892		kfree(mr);
 893		return ERR_PTR(err);
 894	}
 895
 896	mr->umem = NULL;
 897
 898	return &mr->ibmr;
 899}
 900
 901static struct ib_mr *mthca_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
 902				       u64 virt, int acc, struct ib_udata *udata)
 903{
 904	struct mthca_dev *dev = to_mdev(pd->device);
 905	struct scatterlist *sg;
 
 
 906	struct mthca_mr *mr;
 907	struct mthca_reg_mr ucmd;
 908	u64 *pages;
 909	int shift, n, len;
 910	int i, k, entry;
 911	int err = 0;
 912	int write_mtt_size;
 913
 914	if (udata->inlen - sizeof (struct ib_uverbs_cmd_hdr) < sizeof ucmd) {
 915		if (!to_mucontext(pd->uobject->context)->reg_mr_warned) {
 916			mthca_warn(dev, "Process '%s' did not pass in MR attrs.\n",
 917				   current->comm);
 918			mthca_warn(dev, "  Update libmthca to fix this.\n");
 919		}
 920		++to_mucontext(pd->uobject->context)->reg_mr_warned;
 921		ucmd.mr_attrs = 0;
 922	} else if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd))
 923		return ERR_PTR(-EFAULT);
 924
 925	mr = kmalloc(sizeof *mr, GFP_KERNEL);
 926	if (!mr)
 927		return ERR_PTR(-ENOMEM);
 928
 929	mr->umem = ib_umem_get(pd->uobject->context, start, length, acc,
 930			       ucmd.mr_attrs & MTHCA_MR_DMASYNC);
 931
 932	if (IS_ERR(mr->umem)) {
 933		err = PTR_ERR(mr->umem);
 934		goto err;
 935	}
 936
 937	shift = ffs(mr->umem->page_size) - 1;
 938	n = mr->umem->nmap;
 939
 940	mr->mtt = mthca_alloc_mtt(dev, n);
 941	if (IS_ERR(mr->mtt)) {
 942		err = PTR_ERR(mr->mtt);
 943		goto err_umem;
 944	}
 945
 946	pages = (u64 *) __get_free_page(GFP_KERNEL);
 947	if (!pages) {
 948		err = -ENOMEM;
 949		goto err_mtt;
 950	}
 951
 952	i = n = 0;
 953
 954	write_mtt_size = min(mthca_write_mtt_size(dev), (int) (PAGE_SIZE / sizeof *pages));
 955
 956	for_each_sg(mr->umem->sg_head.sgl, sg, mr->umem->nmap, entry) {
 957		len = sg_dma_len(sg) >> shift;
 958		for (k = 0; k < len; ++k) {
 959			pages[i++] = sg_dma_address(sg) +
 960				mr->umem->page_size * k;
 961			/*
 962			 * Be friendly to write_mtt and pass it chunks
 963			 * of appropriate size.
 964			 */
 965			if (i == write_mtt_size) {
 966				err = mthca_write_mtt(dev, mr->mtt, n, pages, i);
 967				if (err)
 968					goto mtt_done;
 969				n += i;
 970				i = 0;
 971			}
 972		}
 973	}
 974
 975	if (i)
 976		err = mthca_write_mtt(dev, mr->mtt, n, pages, i);
 977mtt_done:
 978	free_page((unsigned long) pages);
 979	if (err)
 980		goto err_mtt;
 981
 982	err = mthca_mr_alloc(dev, to_mpd(pd)->pd_num, shift, virt, length,
 983			     convert_access(acc), mr);
 984
 985	if (err)
 986		goto err_mtt;
 987
 988	return &mr->ibmr;
 989
 990err_mtt:
 991	mthca_free_mtt(dev, mr->mtt);
 992
 993err_umem:
 994	ib_umem_release(mr->umem);
 995
 996err:
 997	kfree(mr);
 998	return ERR_PTR(err);
 999}
1000
1001static int mthca_dereg_mr(struct ib_mr *mr)
1002{
1003	struct mthca_mr *mmr = to_mmr(mr);
1004
1005	mthca_free_mr(to_mdev(mr->device), mmr);
1006	if (mmr->umem)
1007		ib_umem_release(mmr->umem);
1008	kfree(mmr);
1009
1010	return 0;
1011}
1012
1013static struct ib_fmr *mthca_alloc_fmr(struct ib_pd *pd, int mr_access_flags,
1014				      struct ib_fmr_attr *fmr_attr)
1015{
1016	struct mthca_fmr *fmr;
1017	int err;
1018
1019	fmr = kmalloc(sizeof *fmr, GFP_KERNEL);
1020	if (!fmr)
1021		return ERR_PTR(-ENOMEM);
1022
1023	memcpy(&fmr->attr, fmr_attr, sizeof *fmr_attr);
1024	err = mthca_fmr_alloc(to_mdev(pd->device), to_mpd(pd)->pd_num,
1025			     convert_access(mr_access_flags), fmr);
1026
1027	if (err) {
1028		kfree(fmr);
1029		return ERR_PTR(err);
1030	}
1031
1032	return &fmr->ibmr;
1033}
 
1034
1035static int mthca_dealloc_fmr(struct ib_fmr *fmr)
1036{
1037	struct mthca_fmr *mfmr = to_mfmr(fmr);
1038	int err;
1039
1040	err = mthca_free_fmr(to_mdev(fmr->device), mfmr);
1041	if (err)
1042		return err;
1043
1044	kfree(mfmr);
1045	return 0;
1046}
1047
1048static int mthca_unmap_fmr(struct list_head *fmr_list)
1049{
1050	struct ib_fmr *fmr;
1051	int err;
1052	struct mthca_dev *mdev = NULL;
1053
1054	list_for_each_entry(fmr, fmr_list, list) {
1055		if (mdev && to_mdev(fmr->device) != mdev)
1056			return -EINVAL;
1057		mdev = to_mdev(fmr->device);
1058	}
1059
1060	if (!mdev)
1061		return 0;
1062
1063	if (mthca_is_memfree(mdev)) {
1064		list_for_each_entry(fmr, fmr_list, list)
1065			mthca_arbel_fmr_unmap(mdev, to_mfmr(fmr));
1066
1067		wmb();
1068	} else
1069		list_for_each_entry(fmr, fmr_list, list)
1070			mthca_tavor_fmr_unmap(mdev, to_mfmr(fmr));
1071
1072	err = mthca_SYNC_TPT(mdev);
1073	return err;
1074}
1075
1076static ssize_t show_rev(struct device *device, struct device_attribute *attr,
1077			char *buf)
1078{
1079	struct mthca_dev *dev =
1080		container_of(device, struct mthca_dev, ib_dev.dev);
1081	return sprintf(buf, "%x\n", dev->rev_id);
 
1082}
 
1083
1084static ssize_t show_fw_ver(struct device *device, struct device_attribute *attr,
1085			   char *buf)
1086{
1087	struct mthca_dev *dev =
1088		container_of(device, struct mthca_dev, ib_dev.dev);
1089	return sprintf(buf, "%d.%d.%d\n", (int) (dev->fw_ver >> 32),
1090		       (int) (dev->fw_ver >> 16) & 0xffff,
1091		       (int) dev->fw_ver & 0xffff);
1092}
1093
1094static ssize_t show_hca(struct device *device, struct device_attribute *attr,
1095			char *buf)
1096{
1097	struct mthca_dev *dev =
1098		container_of(device, struct mthca_dev, ib_dev.dev);
1099	switch (dev->pdev->device) {
1100	case PCI_DEVICE_ID_MELLANOX_TAVOR:
1101		return sprintf(buf, "MT23108\n");
1102	case PCI_DEVICE_ID_MELLANOX_ARBEL_COMPAT:
1103		return sprintf(buf, "MT25208 (MT23108 compat mode)\n");
1104	case PCI_DEVICE_ID_MELLANOX_ARBEL:
1105		return sprintf(buf, "MT25208\n");
1106	case PCI_DEVICE_ID_MELLANOX_SINAI:
1107	case PCI_DEVICE_ID_MELLANOX_SINAI_OLD:
1108		return sprintf(buf, "MT25204\n");
1109	default:
1110		return sprintf(buf, "unknown\n");
1111	}
1112}
 
1113
1114static ssize_t show_board(struct device *device, struct device_attribute *attr,
1115			  char *buf)
1116{
1117	struct mthca_dev *dev =
1118		container_of(device, struct mthca_dev, ib_dev.dev);
1119	return sprintf(buf, "%.*s\n", MTHCA_BOARD_ID_LEN, dev->board_id);
1120}
1121
1122static DEVICE_ATTR(hw_rev,   S_IRUGO, show_rev,    NULL);
1123static DEVICE_ATTR(fw_ver,   S_IRUGO, show_fw_ver, NULL);
1124static DEVICE_ATTR(hca_type, S_IRUGO, show_hca,    NULL);
1125static DEVICE_ATTR(board_id, S_IRUGO, show_board,  NULL);
1126
1127static struct device_attribute *mthca_dev_attributes[] = {
1128	&dev_attr_hw_rev,
1129	&dev_attr_fw_ver,
1130	&dev_attr_hca_type,
1131	&dev_attr_board_id
1132};
1133
1134static int mthca_init_node_data(struct mthca_dev *dev)
1135{
1136	struct ib_smp *in_mad  = NULL;
1137	struct ib_smp *out_mad = NULL;
1138	int err = -ENOMEM;
1139
1140	in_mad  = kzalloc(sizeof *in_mad, GFP_KERNEL);
1141	out_mad = kmalloc(sizeof *out_mad, GFP_KERNEL);
1142	if (!in_mad || !out_mad)
1143		goto out;
1144
1145	init_query_mad(in_mad);
1146	in_mad->attr_id = IB_SMP_ATTR_NODE_DESC;
1147
1148	err = mthca_MAD_IFC(dev, 1, 1,
1149			    1, NULL, NULL, in_mad, out_mad);
1150	if (err)
1151		goto out;
1152
1153	memcpy(dev->ib_dev.node_desc, out_mad->data, 64);
1154
1155	in_mad->attr_id = IB_SMP_ATTR_NODE_INFO;
1156
1157	err = mthca_MAD_IFC(dev, 1, 1,
1158			    1, NULL, NULL, in_mad, out_mad);
1159	if (err)
1160		goto out;
1161
1162	if (mthca_is_memfree(dev))
1163		dev->rev_id = be32_to_cpup((__be32 *) (out_mad->data + 32));
1164	memcpy(&dev->ib_dev.node_guid, out_mad->data + 12, 8);
1165
1166out:
1167	kfree(in_mad);
1168	kfree(out_mad);
1169	return err;
1170}
1171
1172static int mthca_port_immutable(struct ib_device *ibdev, u8 port_num,
1173			        struct ib_port_immutable *immutable)
1174{
1175	struct ib_port_attr attr;
1176	int err;
1177
1178	err = mthca_query_port(ibdev, port_num, &attr);
 
 
1179	if (err)
1180		return err;
1181
1182	immutable->pkey_tbl_len = attr.pkey_tbl_len;
1183	immutable->gid_tbl_len = attr.gid_tbl_len;
1184	immutable->core_cap_flags = RDMA_CORE_PORT_IBA_IB;
1185	immutable->max_mad_size = IB_MGMT_MAD_SIZE;
1186
1187	return 0;
1188}
1189
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1190int mthca_register_device(struct mthca_dev *dev)
1191{
1192	int ret;
1193	int i;
1194
1195	ret = mthca_init_node_data(dev);
1196	if (ret)
1197		return ret;
1198
1199	strlcpy(dev->ib_dev.name, "mthca%d", IB_DEVICE_NAME_MAX);
1200	dev->ib_dev.owner                = THIS_MODULE;
1201
1202	dev->ib_dev.uverbs_abi_ver	 = MTHCA_UVERBS_ABI_VERSION;
1203	dev->ib_dev.uverbs_cmd_mask	 =
1204		(1ull << IB_USER_VERBS_CMD_GET_CONTEXT)		|
1205		(1ull << IB_USER_VERBS_CMD_QUERY_DEVICE)	|
1206		(1ull << IB_USER_VERBS_CMD_QUERY_PORT)		|
1207		(1ull << IB_USER_VERBS_CMD_ALLOC_PD)		|
1208		(1ull << IB_USER_VERBS_CMD_DEALLOC_PD)		|
1209		(1ull << IB_USER_VERBS_CMD_REG_MR)		|
1210		(1ull << IB_USER_VERBS_CMD_DEREG_MR)		|
1211		(1ull << IB_USER_VERBS_CMD_CREATE_COMP_CHANNEL)	|
1212		(1ull << IB_USER_VERBS_CMD_CREATE_CQ)		|
1213		(1ull << IB_USER_VERBS_CMD_RESIZE_CQ)		|
1214		(1ull << IB_USER_VERBS_CMD_DESTROY_CQ)		|
1215		(1ull << IB_USER_VERBS_CMD_CREATE_QP)		|
1216		(1ull << IB_USER_VERBS_CMD_QUERY_QP)		|
1217		(1ull << IB_USER_VERBS_CMD_MODIFY_QP)		|
1218		(1ull << IB_USER_VERBS_CMD_DESTROY_QP)		|
1219		(1ull << IB_USER_VERBS_CMD_ATTACH_MCAST)	|
1220		(1ull << IB_USER_VERBS_CMD_DETACH_MCAST);
1221	dev->ib_dev.node_type            = RDMA_NODE_IB_CA;
1222	dev->ib_dev.phys_port_cnt        = dev->limits.num_ports;
1223	dev->ib_dev.num_comp_vectors     = 1;
1224	dev->ib_dev.dma_device           = &dev->pdev->dev;
1225	dev->ib_dev.query_device         = mthca_query_device;
1226	dev->ib_dev.query_port           = mthca_query_port;
1227	dev->ib_dev.modify_device        = mthca_modify_device;
1228	dev->ib_dev.modify_port          = mthca_modify_port;
1229	dev->ib_dev.query_pkey           = mthca_query_pkey;
1230	dev->ib_dev.query_gid            = mthca_query_gid;
1231	dev->ib_dev.alloc_ucontext       = mthca_alloc_ucontext;
1232	dev->ib_dev.dealloc_ucontext     = mthca_dealloc_ucontext;
1233	dev->ib_dev.mmap                 = mthca_mmap_uar;
1234	dev->ib_dev.alloc_pd             = mthca_alloc_pd;
1235	dev->ib_dev.dealloc_pd           = mthca_dealloc_pd;
1236	dev->ib_dev.create_ah            = mthca_ah_create;
1237	dev->ib_dev.query_ah             = mthca_ah_query;
1238	dev->ib_dev.destroy_ah           = mthca_ah_destroy;
1239
1240	if (dev->mthca_flags & MTHCA_FLAG_SRQ) {
1241		dev->ib_dev.create_srq           = mthca_create_srq;
1242		dev->ib_dev.modify_srq           = mthca_modify_srq;
1243		dev->ib_dev.query_srq            = mthca_query_srq;
1244		dev->ib_dev.destroy_srq          = mthca_destroy_srq;
1245		dev->ib_dev.uverbs_cmd_mask	|=
1246			(1ull << IB_USER_VERBS_CMD_CREATE_SRQ)		|
1247			(1ull << IB_USER_VERBS_CMD_MODIFY_SRQ)		|
1248			(1ull << IB_USER_VERBS_CMD_QUERY_SRQ)		|
1249			(1ull << IB_USER_VERBS_CMD_DESTROY_SRQ);
1250
1251		if (mthca_is_memfree(dev))
1252			dev->ib_dev.post_srq_recv = mthca_arbel_post_srq_recv;
 
1253		else
1254			dev->ib_dev.post_srq_recv = mthca_tavor_post_srq_recv;
 
1255	}
1256
1257	dev->ib_dev.create_qp            = mthca_create_qp;
1258	dev->ib_dev.modify_qp            = mthca_modify_qp;
1259	dev->ib_dev.query_qp             = mthca_query_qp;
1260	dev->ib_dev.destroy_qp           = mthca_destroy_qp;
1261	dev->ib_dev.create_cq            = mthca_create_cq;
1262	dev->ib_dev.resize_cq            = mthca_resize_cq;
1263	dev->ib_dev.destroy_cq           = mthca_destroy_cq;
1264	dev->ib_dev.poll_cq              = mthca_poll_cq;
1265	dev->ib_dev.get_dma_mr           = mthca_get_dma_mr;
1266	dev->ib_dev.reg_user_mr          = mthca_reg_user_mr;
1267	dev->ib_dev.dereg_mr             = mthca_dereg_mr;
1268	dev->ib_dev.get_port_immutable   = mthca_port_immutable;
1269
1270	if (dev->mthca_flags & MTHCA_FLAG_FMR) {
1271		dev->ib_dev.alloc_fmr            = mthca_alloc_fmr;
1272		dev->ib_dev.unmap_fmr            = mthca_unmap_fmr;
1273		dev->ib_dev.dealloc_fmr          = mthca_dealloc_fmr;
1274		if (mthca_is_memfree(dev))
1275			dev->ib_dev.map_phys_fmr = mthca_arbel_map_phys_fmr;
1276		else
1277			dev->ib_dev.map_phys_fmr = mthca_tavor_map_phys_fmr;
1278	}
1279
1280	dev->ib_dev.attach_mcast         = mthca_multicast_attach;
1281	dev->ib_dev.detach_mcast         = mthca_multicast_detach;
1282	dev->ib_dev.process_mad          = mthca_process_mad;
1283
1284	if (mthca_is_memfree(dev)) {
1285		dev->ib_dev.req_notify_cq = mthca_arbel_arm_cq;
1286		dev->ib_dev.post_send     = mthca_arbel_post_send;
1287		dev->ib_dev.post_recv     = mthca_arbel_post_receive;
1288	} else {
1289		dev->ib_dev.req_notify_cq = mthca_tavor_arm_cq;
1290		dev->ib_dev.post_send     = mthca_tavor_post_send;
1291		dev->ib_dev.post_recv     = mthca_tavor_post_receive;
1292	}
1293
1294	mutex_init(&dev->cap_mask_mutex);
1295
1296	ret = ib_register_device(&dev->ib_dev, NULL);
1297	if (ret)
1298		return ret;
1299
1300	for (i = 0; i < ARRAY_SIZE(mthca_dev_attributes); ++i) {
1301		ret = device_create_file(&dev->ib_dev.dev,
1302					 mthca_dev_attributes[i]);
1303		if (ret) {
1304			ib_unregister_device(&dev->ib_dev);
1305			return ret;
1306		}
1307	}
1308
1309	mthca_start_catas_poll(dev);
1310
1311	return 0;
1312}
1313
1314void mthca_unregister_device(struct mthca_dev *dev)
1315{
1316	mthca_stop_catas_poll(dev);
1317	ib_unregister_device(&dev->ib_dev);
1318}