Linux Audio

Check our new training course

Loading...
v6.8
   1/*
   2 * Copyright (c) 2004, 2005 Topspin Communications.  All rights reserved.
   3 * Copyright (c) 2005 Sun Microsystems, Inc. All rights reserved.
   4 * Copyright (c) 2005, 2006 Cisco Systems.  All rights reserved.
   5 * Copyright (c) 2005 Mellanox Technologies. All rights reserved.
   6 * Copyright (c) 2004 Voltaire, Inc. All rights reserved.
   7 *
   8 * This software is available to you under a choice of one of two
   9 * licenses.  You may choose to be licensed under the terms of the GNU
  10 * General Public License (GPL) Version 2, available from the file
  11 * COPYING in the main directory of this source tree, or the
  12 * OpenIB.org BSD license below:
  13 *
  14 *     Redistribution and use in source and binary forms, with or
  15 *     without modification, are permitted provided that the following
  16 *     conditions are met:
  17 *
  18 *      - Redistributions of source code must retain the above
  19 *        copyright notice, this list of conditions and the following
  20 *        disclaimer.
  21 *
  22 *      - Redistributions in binary form must reproduce the above
  23 *        copyright notice, this list of conditions and the following
  24 *        disclaimer in the documentation and/or other materials
  25 *        provided with the distribution.
  26 *
  27 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
  28 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
  29 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
  30 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
  31 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
  32 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
  33 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
  34 * SOFTWARE.
  35 */
  36
  37#include <rdma/ib_smi.h>
  38#include <rdma/ib_umem.h>
  39#include <rdma/ib_user_verbs.h>
  40#include <rdma/uverbs_ioctl.h>
  41
  42#include <linux/sched.h>
  43#include <linux/slab.h>
  44#include <linux/stat.h>
  45#include <linux/mm.h>
  46#include <linux/export.h>
  47
  48#include "mthca_dev.h"
  49#include "mthca_cmd.h"
  50#include <rdma/mthca-abi.h>
  51#include "mthca_memfree.h"
  52
 
 
 
 
 
 
 
 
  53static int mthca_query_device(struct ib_device *ibdev, struct ib_device_attr *props,
  54			      struct ib_udata *uhw)
  55{
  56	struct ib_smp *in_mad;
  57	struct ib_smp *out_mad;
  58	int err = -ENOMEM;
  59	struct mthca_dev *mdev = to_mdev(ibdev);
  60
  61	if (uhw->inlen || uhw->outlen)
  62		return -EINVAL;
  63
  64	in_mad  = kzalloc(sizeof *in_mad, GFP_KERNEL);
  65	out_mad = kmalloc(sizeof *out_mad, GFP_KERNEL);
  66	if (!in_mad || !out_mad)
  67		goto out;
  68
  69	memset(props, 0, sizeof *props);
  70
  71	props->fw_ver              = mdev->fw_ver;
  72
  73	ib_init_query_mad(in_mad);
  74	in_mad->attr_id = IB_SMP_ATTR_NODE_INFO;
  75
  76	err = mthca_MAD_IFC(mdev, 1, 1,
  77			    1, NULL, NULL, in_mad, out_mad);
  78	if (err)
  79		goto out;
  80
  81	props->device_cap_flags    = mdev->device_cap_flags;
  82	props->vendor_id           = be32_to_cpup((__be32 *) (out_mad->data + 36)) &
  83		0xffffff;
  84	props->vendor_part_id      = be16_to_cpup((__be16 *) (out_mad->data + 30));
  85	props->hw_ver              = be32_to_cpup((__be32 *) (out_mad->data + 32));
  86	memcpy(&props->sys_image_guid, out_mad->data +  4, 8);
  87
  88	props->max_mr_size         = ~0ull;
  89	props->page_size_cap       = mdev->limits.page_size_cap;
  90	props->max_qp              = mdev->limits.num_qps - mdev->limits.reserved_qps;
  91	props->max_qp_wr           = mdev->limits.max_wqes;
  92	props->max_send_sge        = mdev->limits.max_sg;
  93	props->max_recv_sge        = mdev->limits.max_sg;
  94	props->max_sge_rd          = mdev->limits.max_sg;
  95	props->max_cq              = mdev->limits.num_cqs - mdev->limits.reserved_cqs;
  96	props->max_cqe             = mdev->limits.max_cqes;
  97	props->max_mr              = mdev->limits.num_mpts - mdev->limits.reserved_mrws;
  98	props->max_pd              = mdev->limits.num_pds - mdev->limits.reserved_pds;
  99	props->max_qp_rd_atom      = 1 << mdev->qp_table.rdb_shift;
 100	props->max_qp_init_rd_atom = mdev->limits.max_qp_init_rdma;
 101	props->max_res_rd_atom     = props->max_qp_rd_atom * props->max_qp;
 102	props->max_srq             = mdev->limits.num_srqs - mdev->limits.reserved_srqs;
 103	props->max_srq_wr          = mdev->limits.max_srq_wqes;
 104	props->max_srq_sge         = mdev->limits.max_srq_sge;
 105	props->local_ca_ack_delay  = mdev->limits.local_ca_ack_delay;
 106	props->atomic_cap          = mdev->limits.flags & DEV_LIM_FLAG_ATOMIC ?
 107					IB_ATOMIC_HCA : IB_ATOMIC_NONE;
 108	props->max_pkeys           = mdev->limits.pkey_table_len;
 109	props->max_mcast_grp       = mdev->limits.num_mgms + mdev->limits.num_amgms;
 110	props->max_mcast_qp_attach = MTHCA_QP_PER_MGM;
 111	props->max_total_mcast_qp_attach = props->max_mcast_qp_attach *
 112					   props->max_mcast_grp;
 
 
 
 
 
 
 
 
 
 
 113
 114	err = 0;
 115 out:
 116	kfree(in_mad);
 117	kfree(out_mad);
 118	return err;
 119}
 120
 121static int mthca_query_port(struct ib_device *ibdev,
 122			    u32 port, struct ib_port_attr *props)
 123{
 124	struct ib_smp *in_mad;
 125	struct ib_smp *out_mad;
 126	int err = -ENOMEM;
 127
 128	in_mad  = kzalloc(sizeof *in_mad, GFP_KERNEL);
 129	out_mad = kmalloc(sizeof *out_mad, GFP_KERNEL);
 130	if (!in_mad || !out_mad)
 131		goto out;
 132
 133	/* props being zeroed by the caller, avoid zeroing it here */
 134
 135	ib_init_query_mad(in_mad);
 136	in_mad->attr_id  = IB_SMP_ATTR_PORT_INFO;
 137	in_mad->attr_mod = cpu_to_be32(port);
 138
 139	err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1,
 140			    port, NULL, NULL, in_mad, out_mad);
 141	if (err)
 142		goto out;
 143
 144	props->lid               = be16_to_cpup((__be16 *) (out_mad->data + 16));
 145	props->lmc               = out_mad->data[34] & 0x7;
 146	props->sm_lid            = be16_to_cpup((__be16 *) (out_mad->data + 18));
 147	props->sm_sl             = out_mad->data[36] & 0xf;
 148	props->state             = out_mad->data[32] & 0xf;
 149	props->phys_state        = out_mad->data[33] >> 4;
 150	props->port_cap_flags    = be32_to_cpup((__be32 *) (out_mad->data + 20));
 151	props->gid_tbl_len       = to_mdev(ibdev)->limits.gid_table_len;
 152	props->max_msg_sz        = 0x80000000;
 153	props->pkey_tbl_len      = to_mdev(ibdev)->limits.pkey_table_len;
 154	props->bad_pkey_cntr     = be16_to_cpup((__be16 *) (out_mad->data + 46));
 155	props->qkey_viol_cntr    = be16_to_cpup((__be16 *) (out_mad->data + 48));
 156	props->active_width      = out_mad->data[31] & 0xf;
 157	props->active_speed      = out_mad->data[35] >> 4;
 158	props->max_mtu           = out_mad->data[41] & 0xf;
 159	props->active_mtu        = out_mad->data[36] >> 4;
 160	props->subnet_timeout    = out_mad->data[51] & 0x1f;
 161	props->max_vl_num        = out_mad->data[37] >> 4;
 162	props->init_type_reply   = out_mad->data[41] >> 4;
 163
 164 out:
 165	kfree(in_mad);
 166	kfree(out_mad);
 167	return err;
 168}
 169
 170static int mthca_modify_device(struct ib_device *ibdev,
 171			       int mask,
 172			       struct ib_device_modify *props)
 173{
 174	if (mask & ~IB_DEVICE_MODIFY_NODE_DESC)
 175		return -EOPNOTSUPP;
 176
 177	if (mask & IB_DEVICE_MODIFY_NODE_DESC) {
 178		if (mutex_lock_interruptible(&to_mdev(ibdev)->cap_mask_mutex))
 179			return -ERESTARTSYS;
 180		memcpy(ibdev->node_desc, props->node_desc,
 181		       IB_DEVICE_NODE_DESC_MAX);
 182		mutex_unlock(&to_mdev(ibdev)->cap_mask_mutex);
 183	}
 184
 185	return 0;
 186}
 187
 188static int mthca_modify_port(struct ib_device *ibdev,
 189			     u32 port, int port_modify_mask,
 190			     struct ib_port_modify *props)
 191{
 192	struct mthca_set_ib_param set_ib;
 193	struct ib_port_attr attr;
 194	int err;
 195
 196	if (mutex_lock_interruptible(&to_mdev(ibdev)->cap_mask_mutex))
 197		return -ERESTARTSYS;
 198
 199	err = ib_query_port(ibdev, port, &attr);
 200	if (err)
 201		goto out;
 202
 203	set_ib.set_si_guid     = 0;
 204	set_ib.reset_qkey_viol = !!(port_modify_mask & IB_PORT_RESET_QKEY_CNTR);
 205
 206	set_ib.cap_mask = (attr.port_cap_flags | props->set_port_cap_mask) &
 207		~props->clr_port_cap_mask;
 208
 209	err = mthca_SET_IB(to_mdev(ibdev), &set_ib, port);
 210	if (err)
 211		goto out;
 212out:
 213	mutex_unlock(&to_mdev(ibdev)->cap_mask_mutex);
 214	return err;
 215}
 216
 217static int mthca_query_pkey(struct ib_device *ibdev,
 218			    u32 port, u16 index, u16 *pkey)
 219{
 220	struct ib_smp *in_mad;
 221	struct ib_smp *out_mad;
 222	int err = -ENOMEM;
 223
 224	in_mad  = kzalloc(sizeof *in_mad, GFP_KERNEL);
 225	out_mad = kmalloc(sizeof *out_mad, GFP_KERNEL);
 226	if (!in_mad || !out_mad)
 227		goto out;
 228
 229	ib_init_query_mad(in_mad);
 230	in_mad->attr_id  = IB_SMP_ATTR_PKEY_TABLE;
 231	in_mad->attr_mod = cpu_to_be32(index / 32);
 232
 233	err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1,
 234			    port, NULL, NULL, in_mad, out_mad);
 235	if (err)
 236		goto out;
 237
 238	*pkey = be16_to_cpu(((__be16 *) out_mad->data)[index % 32]);
 239
 240 out:
 241	kfree(in_mad);
 242	kfree(out_mad);
 243	return err;
 244}
 245
 246static int mthca_query_gid(struct ib_device *ibdev, u32 port,
 247			   int index, union ib_gid *gid)
 248{
 249	struct ib_smp *in_mad;
 250	struct ib_smp *out_mad;
 251	int err = -ENOMEM;
 252
 253	in_mad  = kzalloc(sizeof *in_mad, GFP_KERNEL);
 254	out_mad = kmalloc(sizeof *out_mad, GFP_KERNEL);
 255	if (!in_mad || !out_mad)
 256		goto out;
 257
 258	ib_init_query_mad(in_mad);
 259	in_mad->attr_id  = IB_SMP_ATTR_PORT_INFO;
 260	in_mad->attr_mod = cpu_to_be32(port);
 261
 262	err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1,
 263			    port, NULL, NULL, in_mad, out_mad);
 264	if (err)
 265		goto out;
 266
 267	memcpy(gid->raw, out_mad->data + 8, 8);
 268
 269	ib_init_query_mad(in_mad);
 270	in_mad->attr_id  = IB_SMP_ATTR_GUID_INFO;
 271	in_mad->attr_mod = cpu_to_be32(index / 8);
 272
 273	err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1,
 274			    port, NULL, NULL, in_mad, out_mad);
 275	if (err)
 276		goto out;
 277
 278	memcpy(gid->raw + 8, out_mad->data + (index % 8) * 8, 8);
 279
 280 out:
 281	kfree(in_mad);
 282	kfree(out_mad);
 283	return err;
 284}
 285
 286static int mthca_alloc_ucontext(struct ib_ucontext *uctx,
 287				struct ib_udata *udata)
 288{
 289	struct ib_device *ibdev = uctx->device;
 290	struct mthca_alloc_ucontext_resp uresp = {};
 291	struct mthca_ucontext *context = to_mucontext(uctx);
 292	int                              err;
 293
 294	if (!(to_mdev(ibdev)->active))
 295		return -EAGAIN;
 296
 297	uresp.qp_tab_size = to_mdev(ibdev)->limits.num_qps;
 298	if (mthca_is_memfree(to_mdev(ibdev)))
 299		uresp.uarc_size = to_mdev(ibdev)->uar_table.uarc_size;
 300	else
 301		uresp.uarc_size = 0;
 302
 303	err = mthca_uar_alloc(to_mdev(ibdev), &context->uar);
 304	if (err)
 305		return err;
 306
 307	context->db_tab = mthca_init_user_db_tab(to_mdev(ibdev));
 308	if (IS_ERR(context->db_tab)) {
 309		err = PTR_ERR(context->db_tab);
 310		mthca_uar_free(to_mdev(ibdev), &context->uar);
 311		return err;
 312	}
 313
 314	if (ib_copy_to_udata(udata, &uresp, sizeof(uresp))) {
 315		mthca_cleanup_user_db_tab(to_mdev(ibdev), &context->uar, context->db_tab);
 316		mthca_uar_free(to_mdev(ibdev), &context->uar);
 317		return -EFAULT;
 318	}
 319
 320	context->reg_mr_warned = 0;
 321
 322	return 0;
 323}
 324
 325static void mthca_dealloc_ucontext(struct ib_ucontext *context)
 326{
 327	mthca_cleanup_user_db_tab(to_mdev(context->device), &to_mucontext(context)->uar,
 328				  to_mucontext(context)->db_tab);
 329	mthca_uar_free(to_mdev(context->device), &to_mucontext(context)->uar);
 330}
 331
 332static int mthca_mmap_uar(struct ib_ucontext *context,
 333			  struct vm_area_struct *vma)
 334{
 335	if (vma->vm_end - vma->vm_start != PAGE_SIZE)
 336		return -EINVAL;
 337
 338	vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
 339
 340	if (io_remap_pfn_range(vma, vma->vm_start,
 341			       to_mucontext(context)->uar.pfn,
 342			       PAGE_SIZE, vma->vm_page_prot))
 343		return -EAGAIN;
 344
 345	return 0;
 346}
 347
 348static int mthca_alloc_pd(struct ib_pd *ibpd, struct ib_udata *udata)
 349{
 350	struct ib_device *ibdev = ibpd->device;
 351	struct mthca_pd *pd = to_mpd(ibpd);
 352	int err;
 353
 354	err = mthca_pd_alloc(to_mdev(ibdev), !udata, pd);
 355	if (err)
 356		return err;
 357
 358	if (udata) {
 359		if (ib_copy_to_udata(udata, &pd->pd_num, sizeof (__u32))) {
 360			mthca_pd_free(to_mdev(ibdev), pd);
 361			return -EFAULT;
 362		}
 363	}
 364
 365	return 0;
 366}
 367
 368static int mthca_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata)
 369{
 370	mthca_pd_free(to_mdev(pd->device), to_mpd(pd));
 371	return 0;
 372}
 373
 374static int mthca_ah_create(struct ib_ah *ibah,
 375			   struct rdma_ah_init_attr *init_attr,
 376			   struct ib_udata *udata)
 377
 378{
 379	struct mthca_ah *ah = to_mah(ibah);
 380
 381	return mthca_create_ah(to_mdev(ibah->device), to_mpd(ibah->pd),
 382			       init_attr->ah_attr, ah);
 383}
 384
 385static int mthca_ah_destroy(struct ib_ah *ah, u32 flags)
 386{
 387	mthca_destroy_ah(to_mdev(ah->device), to_mah(ah));
 388	return 0;
 389}
 390
 391static int mthca_create_srq(struct ib_srq *ibsrq,
 392			    struct ib_srq_init_attr *init_attr,
 393			    struct ib_udata *udata)
 394{
 395	struct mthca_create_srq ucmd;
 396	struct mthca_ucontext *context = rdma_udata_to_drv_context(
 397		udata, struct mthca_ucontext, ibucontext);
 398	struct mthca_srq *srq = to_msrq(ibsrq);
 399	int err;
 400
 401	if (init_attr->srq_type != IB_SRQT_BASIC)
 402		return -EOPNOTSUPP;
 403
 404	if (udata) {
 405		if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd)))
 406			return -EFAULT;
 407
 408		err = mthca_map_user_db(to_mdev(ibsrq->device), &context->uar,
 409					context->db_tab, ucmd.db_index,
 410					ucmd.db_page);
 411
 412		if (err)
 413			return err;
 414
 415		srq->mr.ibmr.lkey = ucmd.lkey;
 416		srq->db_index     = ucmd.db_index;
 417	}
 418
 419	err = mthca_alloc_srq(to_mdev(ibsrq->device), to_mpd(ibsrq->pd),
 420			      &init_attr->attr, srq, udata);
 421
 422	if (err && udata)
 423		mthca_unmap_user_db(to_mdev(ibsrq->device), &context->uar,
 424				    context->db_tab, ucmd.db_index);
 425
 426	if (err)
 427		return err;
 428
 429	if (context && ib_copy_to_udata(udata, &srq->srqn, sizeof(__u32))) {
 430		mthca_free_srq(to_mdev(ibsrq->device), srq);
 431		return -EFAULT;
 432	}
 433
 434	return 0;
 435}
 436
 437static int mthca_destroy_srq(struct ib_srq *srq, struct ib_udata *udata)
 438{
 439	if (udata) {
 440		struct mthca_ucontext *context =
 441			rdma_udata_to_drv_context(
 442				udata,
 443				struct mthca_ucontext,
 444				ibucontext);
 445
 446		mthca_unmap_user_db(to_mdev(srq->device), &context->uar,
 447				    context->db_tab, to_msrq(srq)->db_index);
 448	}
 449
 450	mthca_free_srq(to_mdev(srq->device), to_msrq(srq));
 451	return 0;
 452}
 453
 454static int mthca_create_qp(struct ib_qp *ibqp,
 455			   struct ib_qp_init_attr *init_attr,
 456			   struct ib_udata *udata)
 457{
 458	struct mthca_ucontext *context = rdma_udata_to_drv_context(
 459		udata, struct mthca_ucontext, ibucontext);
 460	struct mthca_create_qp ucmd;
 461	struct mthca_qp *qp = to_mqp(ibqp);
 462	struct mthca_dev *dev = to_mdev(ibqp->device);
 463	int err;
 464
 465	if (init_attr->create_flags)
 466		return -EOPNOTSUPP;
 467
 468	switch (init_attr->qp_type) {
 469	case IB_QPT_RC:
 470	case IB_QPT_UC:
 471	case IB_QPT_UD:
 472	{
 
 
 
 
 473		if (udata) {
 474			if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd)))
 475				return -EFAULT;
 
 
 476
 477			err = mthca_map_user_db(dev, &context->uar,
 478						context->db_tab,
 479						ucmd.sq_db_index,
 480						ucmd.sq_db_page);
 481			if (err)
 482				return err;
 
 483
 484			err = mthca_map_user_db(dev, &context->uar,
 485						context->db_tab,
 486						ucmd.rq_db_index,
 487						ucmd.rq_db_page);
 488			if (err) {
 489				mthca_unmap_user_db(dev, &context->uar,
 
 490						    context->db_tab,
 491						    ucmd.sq_db_index);
 492				return err;
 
 493			}
 494
 495			qp->mr.ibmr.lkey = ucmd.lkey;
 496			qp->sq.db_index  = ucmd.sq_db_index;
 497			qp->rq.db_index  = ucmd.rq_db_index;
 498		}
 499
 500		err = mthca_alloc_qp(dev, to_mpd(ibqp->pd),
 501				     to_mcq(init_attr->send_cq),
 502				     to_mcq(init_attr->recv_cq),
 503				     init_attr->qp_type, init_attr->sq_sig_type,
 504				     &init_attr->cap, qp, udata);
 505
 506		if (err && udata) {
 507			mthca_unmap_user_db(dev, &context->uar, context->db_tab,
 
 
 508					    ucmd.sq_db_index);
 509			mthca_unmap_user_db(dev, &context->uar, context->db_tab,
 
 
 510					    ucmd.rq_db_index);
 511		}
 512
 513		qp->ibqp.qp_num = qp->qpn;
 514		break;
 515	}
 516	case IB_QPT_SMI:
 517	case IB_QPT_GSI:
 518	{
 519		qp->sqp = kzalloc(sizeof(struct mthca_sqp), GFP_KERNEL);
 520		if (!qp->sqp)
 521			return -ENOMEM;
 
 
 
 
 522
 523		qp->ibqp.qp_num = init_attr->qp_type == IB_QPT_SMI ? 0 : 1;
 524
 525		err = mthca_alloc_sqp(dev, to_mpd(ibqp->pd),
 526				      to_mcq(init_attr->send_cq),
 527				      to_mcq(init_attr->recv_cq),
 528				      init_attr->sq_sig_type, &init_attr->cap,
 529				      qp->ibqp.qp_num, init_attr->port_num, qp,
 530				      udata);
 531		break;
 532	}
 533	default:
 534		/* Don't support raw QPs */
 535		return -EOPNOTSUPP;
 536	}
 537
 538	if (err) {
 539		kfree(qp->sqp);
 540		return err;
 541	}
 542
 543	init_attr->cap.max_send_wr     = qp->sq.max;
 544	init_attr->cap.max_recv_wr     = qp->rq.max;
 545	init_attr->cap.max_send_sge    = qp->sq.max_gs;
 546	init_attr->cap.max_recv_sge    = qp->rq.max_gs;
 547	init_attr->cap.max_inline_data = qp->max_inline_data;
 548
 549	return 0;
 550}
 551
 552static int mthca_destroy_qp(struct ib_qp *qp, struct ib_udata *udata)
 553{
 554	if (udata) {
 555		struct mthca_ucontext *context =
 556			rdma_udata_to_drv_context(
 557				udata,
 558				struct mthca_ucontext,
 559				ibucontext);
 560
 561		mthca_unmap_user_db(to_mdev(qp->device),
 562				    &context->uar,
 563				    context->db_tab,
 564				    to_mqp(qp)->sq.db_index);
 565		mthca_unmap_user_db(to_mdev(qp->device),
 566				    &context->uar,
 567				    context->db_tab,
 568				    to_mqp(qp)->rq.db_index);
 569	}
 570	mthca_free_qp(to_mdev(qp->device), to_mqp(qp));
 571	kfree(to_mqp(qp)->sqp);
 572	return 0;
 573}
 574
 575static int mthca_create_cq(struct ib_cq *ibcq,
 576			   const struct ib_cq_init_attr *attr,
 577			   struct ib_udata *udata)
 578{
 579	struct ib_device *ibdev = ibcq->device;
 580	int entries = attr->cqe;
 581	struct mthca_create_cq ucmd;
 582	struct mthca_cq *cq;
 583	int nent;
 584	int err;
 585	struct mthca_ucontext *context = rdma_udata_to_drv_context(
 586		udata, struct mthca_ucontext, ibucontext);
 587
 588	if (attr->flags)
 589		return -EOPNOTSUPP;
 590
 591	if (entries < 1 || entries > to_mdev(ibdev)->limits.max_cqes)
 592		return -EINVAL;
 593
 594	if (udata) {
 595		if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd)))
 596			return -EFAULT;
 597
 598		err = mthca_map_user_db(to_mdev(ibdev), &context->uar,
 599					context->db_tab, ucmd.set_db_index,
 600					ucmd.set_db_page);
 601		if (err)
 602			return err;
 603
 604		err = mthca_map_user_db(to_mdev(ibdev), &context->uar,
 605					context->db_tab, ucmd.arm_db_index,
 606					ucmd.arm_db_page);
 607		if (err)
 608			goto err_unmap_set;
 609	}
 610
 611	cq = to_mcq(ibcq);
 612
 613	if (udata) {
 614		cq->buf.mr.ibmr.lkey = ucmd.lkey;
 615		cq->set_ci_db_index  = ucmd.set_db_index;
 616		cq->arm_db_index     = ucmd.arm_db_index;
 617	}
 618
 619	for (nent = 1; nent <= entries; nent <<= 1)
 620		; /* nothing */
 621
 622	err = mthca_init_cq(to_mdev(ibdev), nent, context,
 623			    udata ? ucmd.pdn : to_mdev(ibdev)->driver_pd.pd_num,
 624			    cq);
 625	if (err)
 626		goto err_unmap_arm;
 627
 628	if (udata && ib_copy_to_udata(udata, &cq->cqn, sizeof(__u32))) {
 629		mthca_free_cq(to_mdev(ibdev), cq);
 630		err = -EFAULT;
 631		goto err_unmap_arm;
 632	}
 633
 634	cq->resize_buf = NULL;
 635
 636	return 0;
 637
 638err_unmap_arm:
 639	if (udata)
 640		mthca_unmap_user_db(to_mdev(ibdev), &context->uar,
 641				    context->db_tab, ucmd.arm_db_index);
 642
 643err_unmap_set:
 644	if (udata)
 645		mthca_unmap_user_db(to_mdev(ibdev), &context->uar,
 646				    context->db_tab, ucmd.set_db_index);
 647
 648	return err;
 649}
 650
 651static int mthca_alloc_resize_buf(struct mthca_dev *dev, struct mthca_cq *cq,
 652				  int entries)
 653{
 654	int ret;
 655
 656	spin_lock_irq(&cq->lock);
 657	if (cq->resize_buf) {
 658		ret = -EBUSY;
 659		goto unlock;
 660	}
 661
 662	cq->resize_buf = kmalloc(sizeof *cq->resize_buf, GFP_ATOMIC);
 663	if (!cq->resize_buf) {
 664		ret = -ENOMEM;
 665		goto unlock;
 666	}
 667
 668	cq->resize_buf->state = CQ_RESIZE_ALLOC;
 669
 670	ret = 0;
 671
 672unlock:
 673	spin_unlock_irq(&cq->lock);
 674
 675	if (ret)
 676		return ret;
 677
 678	ret = mthca_alloc_cq_buf(dev, &cq->resize_buf->buf, entries);
 679	if (ret) {
 680		spin_lock_irq(&cq->lock);
 681		kfree(cq->resize_buf);
 682		cq->resize_buf = NULL;
 683		spin_unlock_irq(&cq->lock);
 684		return ret;
 685	}
 686
 687	cq->resize_buf->cqe = entries - 1;
 688
 689	spin_lock_irq(&cq->lock);
 690	cq->resize_buf->state = CQ_RESIZE_READY;
 691	spin_unlock_irq(&cq->lock);
 692
 693	return 0;
 694}
 695
 696static int mthca_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata)
 697{
 698	struct mthca_dev *dev = to_mdev(ibcq->device);
 699	struct mthca_cq *cq = to_mcq(ibcq);
 700	struct mthca_resize_cq ucmd;
 701	u32 lkey;
 702	int ret;
 703
 704	if (entries < 1 || entries > dev->limits.max_cqes)
 705		return -EINVAL;
 706
 707	mutex_lock(&cq->mutex);
 708
 709	entries = roundup_pow_of_two(entries + 1);
 710	if (entries == ibcq->cqe + 1) {
 711		ret = 0;
 712		goto out;
 713	}
 714
 715	if (cq->is_kernel) {
 716		ret = mthca_alloc_resize_buf(dev, cq, entries);
 717		if (ret)
 718			goto out;
 719		lkey = cq->resize_buf->buf.mr.ibmr.lkey;
 720	} else {
 721		if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) {
 722			ret = -EFAULT;
 723			goto out;
 724		}
 725		lkey = ucmd.lkey;
 726	}
 727
 728	ret = mthca_RESIZE_CQ(dev, cq->cqn, lkey, ilog2(entries));
 729
 730	if (ret) {
 731		if (cq->resize_buf) {
 732			mthca_free_cq_buf(dev, &cq->resize_buf->buf,
 733					  cq->resize_buf->cqe);
 734			kfree(cq->resize_buf);
 735			spin_lock_irq(&cq->lock);
 736			cq->resize_buf = NULL;
 737			spin_unlock_irq(&cq->lock);
 738		}
 739		goto out;
 740	}
 741
 742	if (cq->is_kernel) {
 743		struct mthca_cq_buf tbuf;
 744		int tcqe;
 745
 746		spin_lock_irq(&cq->lock);
 747		if (cq->resize_buf->state == CQ_RESIZE_READY) {
 748			mthca_cq_resize_copy_cqes(cq);
 749			tbuf         = cq->buf;
 750			tcqe         = cq->ibcq.cqe;
 751			cq->buf      = cq->resize_buf->buf;
 752			cq->ibcq.cqe = cq->resize_buf->cqe;
 753		} else {
 754			tbuf = cq->resize_buf->buf;
 755			tcqe = cq->resize_buf->cqe;
 756		}
 757
 758		kfree(cq->resize_buf);
 759		cq->resize_buf = NULL;
 760		spin_unlock_irq(&cq->lock);
 761
 762		mthca_free_cq_buf(dev, &tbuf, tcqe);
 763	} else
 764		ibcq->cqe = entries - 1;
 765
 766out:
 767	mutex_unlock(&cq->mutex);
 768
 769	return ret;
 770}
 771
 772static int mthca_destroy_cq(struct ib_cq *cq, struct ib_udata *udata)
 773{
 774	if (udata) {
 775		struct mthca_ucontext *context =
 776			rdma_udata_to_drv_context(
 777				udata,
 778				struct mthca_ucontext,
 779				ibucontext);
 780
 781		mthca_unmap_user_db(to_mdev(cq->device),
 782				    &context->uar,
 783				    context->db_tab,
 784				    to_mcq(cq)->arm_db_index);
 785		mthca_unmap_user_db(to_mdev(cq->device),
 786				    &context->uar,
 787				    context->db_tab,
 788				    to_mcq(cq)->set_ci_db_index);
 789	}
 790	mthca_free_cq(to_mdev(cq->device), to_mcq(cq));
 791	return 0;
 792}
 793
 794static inline u32 convert_access(int acc)
 795{
 796	return (acc & IB_ACCESS_REMOTE_ATOMIC ? MTHCA_MPT_FLAG_ATOMIC       : 0) |
 797	       (acc & IB_ACCESS_REMOTE_WRITE  ? MTHCA_MPT_FLAG_REMOTE_WRITE : 0) |
 798	       (acc & IB_ACCESS_REMOTE_READ   ? MTHCA_MPT_FLAG_REMOTE_READ  : 0) |
 799	       (acc & IB_ACCESS_LOCAL_WRITE   ? MTHCA_MPT_FLAG_LOCAL_WRITE  : 0) |
 800	       MTHCA_MPT_FLAG_LOCAL_READ;
 801}
 802
 803static struct ib_mr *mthca_get_dma_mr(struct ib_pd *pd, int acc)
 804{
 805	struct mthca_mr *mr;
 806	int err;
 807
 808	mr = kmalloc(sizeof *mr, GFP_KERNEL);
 809	if (!mr)
 810		return ERR_PTR(-ENOMEM);
 811
 812	err = mthca_mr_alloc_notrans(to_mdev(pd->device),
 813				     to_mpd(pd)->pd_num,
 814				     convert_access(acc), mr);
 815
 816	if (err) {
 817		kfree(mr);
 818		return ERR_PTR(err);
 819	}
 820
 821	mr->umem = NULL;
 822
 823	return &mr->ibmr;
 824}
 825
 826static struct ib_mr *mthca_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
 827				       u64 virt, int acc, struct ib_udata *udata)
 828{
 829	struct mthca_dev *dev = to_mdev(pd->device);
 830	struct ib_block_iter biter;
 831	struct mthca_ucontext *context = rdma_udata_to_drv_context(
 832		udata, struct mthca_ucontext, ibucontext);
 833	struct mthca_mr *mr;
 834	struct mthca_reg_mr ucmd;
 835	u64 *pages;
 836	int n, i;
 837	int err = 0;
 838	int write_mtt_size;
 839
 840	if (udata->inlen < sizeof ucmd) {
 841		if (!context->reg_mr_warned) {
 842			mthca_warn(dev, "Process '%s' did not pass in MR attrs.\n",
 843				   current->comm);
 844			mthca_warn(dev, "  Update libmthca to fix this.\n");
 845		}
 846		++context->reg_mr_warned;
 847		ucmd.mr_attrs = 0;
 848	} else if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd))
 849		return ERR_PTR(-EFAULT);
 850
 851	mr = kmalloc(sizeof *mr, GFP_KERNEL);
 852	if (!mr)
 853		return ERR_PTR(-ENOMEM);
 854
 855	mr->umem = ib_umem_get(pd->device, start, length, acc);
 
 
 856	if (IS_ERR(mr->umem)) {
 857		err = PTR_ERR(mr->umem);
 858		goto err;
 859	}
 860
 861	n = ib_umem_num_dma_blocks(mr->umem, PAGE_SIZE);
 862
 863	mr->mtt = mthca_alloc_mtt(dev, n);
 864	if (IS_ERR(mr->mtt)) {
 865		err = PTR_ERR(mr->mtt);
 866		goto err_umem;
 867	}
 868
 869	pages = (u64 *) __get_free_page(GFP_KERNEL);
 870	if (!pages) {
 871		err = -ENOMEM;
 872		goto err_mtt;
 873	}
 874
 875	i = n = 0;
 876
 877	write_mtt_size = min(mthca_write_mtt_size(dev), (int) (PAGE_SIZE / sizeof *pages));
 878
 879	rdma_umem_for_each_dma_block(mr->umem, &biter, PAGE_SIZE) {
 880		pages[i++] = rdma_block_iter_dma_address(&biter);
 881
 882		/*
 883		 * Be friendly to write_mtt and pass it chunks
 884		 * of appropriate size.
 885		 */
 886		if (i == write_mtt_size) {
 887			err = mthca_write_mtt(dev, mr->mtt, n, pages, i);
 888			if (err)
 889				goto mtt_done;
 890			n += i;
 891			i = 0;
 892		}
 893	}
 894
 895	if (i)
 896		err = mthca_write_mtt(dev, mr->mtt, n, pages, i);
 897mtt_done:
 898	free_page((unsigned long) pages);
 899	if (err)
 900		goto err_mtt;
 901
 902	err = mthca_mr_alloc(dev, to_mpd(pd)->pd_num, PAGE_SHIFT, virt, length,
 903			     convert_access(acc), mr);
 904
 905	if (err)
 906		goto err_mtt;
 907
 908	return &mr->ibmr;
 909
 910err_mtt:
 911	mthca_free_mtt(dev, mr->mtt);
 912
 913err_umem:
 914	ib_umem_release(mr->umem);
 915
 916err:
 917	kfree(mr);
 918	return ERR_PTR(err);
 919}
 920
 921static int mthca_dereg_mr(struct ib_mr *mr, struct ib_udata *udata)
 922{
 923	struct mthca_mr *mmr = to_mmr(mr);
 924
 925	mthca_free_mr(to_mdev(mr->device), mmr);
 926	ib_umem_release(mmr->umem);
 927	kfree(mmr);
 928
 929	return 0;
 930}
 931
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 932static ssize_t hw_rev_show(struct device *device,
 933			   struct device_attribute *attr, char *buf)
 934{
 935	struct mthca_dev *dev =
 936		rdma_device_to_drv_device(device, struct mthca_dev, ib_dev);
 937
 938	return sysfs_emit(buf, "%x\n", dev->rev_id);
 939}
 940static DEVICE_ATTR_RO(hw_rev);
 941
 942static const char *hca_type_string(int hca_type)
 
 943{
 944	switch (hca_type) {
 
 
 
 945	case PCI_DEVICE_ID_MELLANOX_TAVOR:
 946		return "MT23108";
 947	case PCI_DEVICE_ID_MELLANOX_ARBEL_COMPAT:
 948		return "MT25208 (MT23108 compat mode)";
 949	case PCI_DEVICE_ID_MELLANOX_ARBEL:
 950		return "MT25208";
 951	case PCI_DEVICE_ID_MELLANOX_SINAI:
 952	case PCI_DEVICE_ID_MELLANOX_SINAI_OLD:
 953		return "MT25204";
 
 
 954	}
 955
 956	return "unknown";
 957}
 958
 959static ssize_t hca_type_show(struct device *device,
 960			     struct device_attribute *attr, char *buf)
 961{
 962	struct mthca_dev *dev =
 963		rdma_device_to_drv_device(device, struct mthca_dev, ib_dev);
 964
 965	return sysfs_emit(buf, "%s\n", hca_type_string(dev->pdev->device));
 966}
 967static DEVICE_ATTR_RO(hca_type);
 968
 969static ssize_t board_id_show(struct device *device,
 970			     struct device_attribute *attr, char *buf)
 971{
 972	struct mthca_dev *dev =
 973		rdma_device_to_drv_device(device, struct mthca_dev, ib_dev);
 974
 975	return sysfs_emit(buf, "%.*s\n", MTHCA_BOARD_ID_LEN, dev->board_id);
 976}
 977static DEVICE_ATTR_RO(board_id);
 978
 979static struct attribute *mthca_dev_attributes[] = {
 980	&dev_attr_hw_rev.attr,
 981	&dev_attr_hca_type.attr,
 982	&dev_attr_board_id.attr,
 983	NULL
 984};
 985
 986static const struct attribute_group mthca_attr_group = {
 987	.attrs = mthca_dev_attributes,
 988};
 989
 990static int mthca_init_node_data(struct mthca_dev *dev)
 991{
 992	struct ib_smp *in_mad;
 993	struct ib_smp *out_mad;
 994	int err = -ENOMEM;
 995
 996	in_mad  = kzalloc(sizeof *in_mad, GFP_KERNEL);
 997	out_mad = kmalloc(sizeof *out_mad, GFP_KERNEL);
 998	if (!in_mad || !out_mad)
 999		goto out;
1000
1001	ib_init_query_mad(in_mad);
1002	in_mad->attr_id = IB_SMP_ATTR_NODE_DESC;
1003
1004	err = mthca_MAD_IFC(dev, 1, 1,
1005			    1, NULL, NULL, in_mad, out_mad);
1006	if (err)
1007		goto out;
1008
1009	memcpy(dev->ib_dev.node_desc, out_mad->data, IB_DEVICE_NODE_DESC_MAX);
1010
1011	in_mad->attr_id = IB_SMP_ATTR_NODE_INFO;
1012
1013	err = mthca_MAD_IFC(dev, 1, 1,
1014			    1, NULL, NULL, in_mad, out_mad);
1015	if (err)
1016		goto out;
1017
1018	if (mthca_is_memfree(dev))
1019		dev->rev_id = be32_to_cpup((__be32 *) (out_mad->data + 32));
1020	memcpy(&dev->ib_dev.node_guid, out_mad->data + 12, 8);
1021
1022out:
1023	kfree(in_mad);
1024	kfree(out_mad);
1025	return err;
1026}
1027
1028static int mthca_port_immutable(struct ib_device *ibdev, u32 port_num,
1029			        struct ib_port_immutable *immutable)
1030{
1031	struct ib_port_attr attr;
1032	int err;
1033
1034	immutable->core_cap_flags = RDMA_CORE_PORT_IBA_IB;
1035
1036	err = ib_query_port(ibdev, port_num, &attr);
1037	if (err)
1038		return err;
1039
1040	immutable->pkey_tbl_len = attr.pkey_tbl_len;
1041	immutable->gid_tbl_len = attr.gid_tbl_len;
1042	immutable->max_mad_size = IB_MGMT_MAD_SIZE;
1043
1044	return 0;
1045}
1046
1047static void get_dev_fw_str(struct ib_device *device, char *str)
1048{
1049	struct mthca_dev *dev =
1050		container_of(device, struct mthca_dev, ib_dev);
1051	snprintf(str, IB_FW_VERSION_NAME_MAX, "%d.%d.%d",
1052		 (int) (dev->fw_ver >> 32),
1053		 (int) (dev->fw_ver >> 16) & 0xffff,
1054		 (int) dev->fw_ver & 0xffff);
1055}
1056
1057static const struct ib_device_ops mthca_dev_ops = {
1058	.owner = THIS_MODULE,
1059	.driver_id = RDMA_DRIVER_MTHCA,
1060	.uverbs_abi_ver = MTHCA_UVERBS_ABI_VERSION,
1061	.uverbs_no_driver_id_binding = 1,
1062
1063	.alloc_pd = mthca_alloc_pd,
1064	.alloc_ucontext = mthca_alloc_ucontext,
1065	.attach_mcast = mthca_multicast_attach,
1066	.create_ah = mthca_ah_create,
1067	.create_cq = mthca_create_cq,
1068	.create_qp = mthca_create_qp,
1069	.dealloc_pd = mthca_dealloc_pd,
1070	.dealloc_ucontext = mthca_dealloc_ucontext,
1071	.dereg_mr = mthca_dereg_mr,
1072	.destroy_ah = mthca_ah_destroy,
1073	.destroy_cq = mthca_destroy_cq,
1074	.destroy_qp = mthca_destroy_qp,
1075	.detach_mcast = mthca_multicast_detach,
1076	.device_group = &mthca_attr_group,
1077	.get_dev_fw_str = get_dev_fw_str,
1078	.get_dma_mr = mthca_get_dma_mr,
1079	.get_port_immutable = mthca_port_immutable,
1080	.mmap = mthca_mmap_uar,
1081	.modify_device = mthca_modify_device,
1082	.modify_port = mthca_modify_port,
1083	.modify_qp = mthca_modify_qp,
1084	.poll_cq = mthca_poll_cq,
1085	.process_mad = mthca_process_mad,
1086	.query_ah = mthca_ah_query,
1087	.query_device = mthca_query_device,
1088	.query_gid = mthca_query_gid,
1089	.query_pkey = mthca_query_pkey,
1090	.query_port = mthca_query_port,
1091	.query_qp = mthca_query_qp,
1092	.reg_user_mr = mthca_reg_user_mr,
1093	.resize_cq = mthca_resize_cq,
1094
1095	INIT_RDMA_OBJ_SIZE(ib_ah, mthca_ah, ibah),
1096	INIT_RDMA_OBJ_SIZE(ib_cq, mthca_cq, ibcq),
1097	INIT_RDMA_OBJ_SIZE(ib_pd, mthca_pd, ibpd),
1098	INIT_RDMA_OBJ_SIZE(ib_qp, mthca_qp, ibqp),
1099	INIT_RDMA_OBJ_SIZE(ib_ucontext, mthca_ucontext, ibucontext),
1100};
1101
1102static const struct ib_device_ops mthca_dev_arbel_srq_ops = {
1103	.create_srq = mthca_create_srq,
1104	.destroy_srq = mthca_destroy_srq,
1105	.modify_srq = mthca_modify_srq,
1106	.post_srq_recv = mthca_arbel_post_srq_recv,
1107	.query_srq = mthca_query_srq,
1108
1109	INIT_RDMA_OBJ_SIZE(ib_srq, mthca_srq, ibsrq),
1110};
1111
1112static const struct ib_device_ops mthca_dev_tavor_srq_ops = {
1113	.create_srq = mthca_create_srq,
1114	.destroy_srq = mthca_destroy_srq,
1115	.modify_srq = mthca_modify_srq,
1116	.post_srq_recv = mthca_tavor_post_srq_recv,
1117	.query_srq = mthca_query_srq,
1118
1119	INIT_RDMA_OBJ_SIZE(ib_srq, mthca_srq, ibsrq),
1120};
1121
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1122static const struct ib_device_ops mthca_dev_arbel_ops = {
1123	.post_recv = mthca_arbel_post_receive,
1124	.post_send = mthca_arbel_post_send,
1125	.req_notify_cq = mthca_arbel_arm_cq,
1126};
1127
1128static const struct ib_device_ops mthca_dev_tavor_ops = {
1129	.post_recv = mthca_tavor_post_receive,
1130	.post_send = mthca_tavor_post_send,
1131	.req_notify_cq = mthca_tavor_arm_cq,
1132};
1133
1134int mthca_register_device(struct mthca_dev *dev)
1135{
1136	int ret;
1137
1138	ret = mthca_init_node_data(dev);
1139	if (ret)
1140		return ret;
1141
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1142	dev->ib_dev.node_type            = RDMA_NODE_IB_CA;
1143	dev->ib_dev.phys_port_cnt        = dev->limits.num_ports;
1144	dev->ib_dev.num_comp_vectors     = 1;
1145	dev->ib_dev.dev.parent           = &dev->pdev->dev;
1146
1147	if (dev->mthca_flags & MTHCA_FLAG_SRQ) {
 
 
 
 
 
 
1148		if (mthca_is_memfree(dev))
1149			ib_set_device_ops(&dev->ib_dev,
1150					  &mthca_dev_arbel_srq_ops);
1151		else
1152			ib_set_device_ops(&dev->ib_dev,
1153					  &mthca_dev_tavor_srq_ops);
1154	}
1155
 
 
 
 
 
 
 
 
 
1156	ib_set_device_ops(&dev->ib_dev, &mthca_dev_ops);
1157
1158	if (mthca_is_memfree(dev))
1159		ib_set_device_ops(&dev->ib_dev, &mthca_dev_arbel_ops);
1160	else
1161		ib_set_device_ops(&dev->ib_dev, &mthca_dev_tavor_ops);
1162
1163	mutex_init(&dev->cap_mask_mutex);
1164
1165	ret = ib_register_device(&dev->ib_dev, "mthca%d", &dev->pdev->dev);
 
1166	if (ret)
1167		return ret;
1168
1169	mthca_start_catas_poll(dev);
1170
1171	return 0;
1172}
1173
1174void mthca_unregister_device(struct mthca_dev *dev)
1175{
1176	mthca_stop_catas_poll(dev);
1177	ib_unregister_device(&dev->ib_dev);
1178}
v5.4
   1/*
   2 * Copyright (c) 2004, 2005 Topspin Communications.  All rights reserved.
   3 * Copyright (c) 2005 Sun Microsystems, Inc. All rights reserved.
   4 * Copyright (c) 2005, 2006 Cisco Systems.  All rights reserved.
   5 * Copyright (c) 2005 Mellanox Technologies. All rights reserved.
   6 * Copyright (c) 2004 Voltaire, Inc. All rights reserved.
   7 *
   8 * This software is available to you under a choice of one of two
   9 * licenses.  You may choose to be licensed under the terms of the GNU
  10 * General Public License (GPL) Version 2, available from the file
  11 * COPYING in the main directory of this source tree, or the
  12 * OpenIB.org BSD license below:
  13 *
  14 *     Redistribution and use in source and binary forms, with or
  15 *     without modification, are permitted provided that the following
  16 *     conditions are met:
  17 *
  18 *      - Redistributions of source code must retain the above
  19 *        copyright notice, this list of conditions and the following
  20 *        disclaimer.
  21 *
  22 *      - Redistributions in binary form must reproduce the above
  23 *        copyright notice, this list of conditions and the following
  24 *        disclaimer in the documentation and/or other materials
  25 *        provided with the distribution.
  26 *
  27 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
  28 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
  29 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
  30 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
  31 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
  32 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
  33 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
  34 * SOFTWARE.
  35 */
  36
  37#include <rdma/ib_smi.h>
  38#include <rdma/ib_umem.h>
  39#include <rdma/ib_user_verbs.h>
  40#include <rdma/uverbs_ioctl.h>
  41
  42#include <linux/sched.h>
  43#include <linux/slab.h>
  44#include <linux/stat.h>
  45#include <linux/mm.h>
  46#include <linux/export.h>
  47
  48#include "mthca_dev.h"
  49#include "mthca_cmd.h"
  50#include <rdma/mthca-abi.h>
  51#include "mthca_memfree.h"
  52
  53static void init_query_mad(struct ib_smp *mad)
  54{
  55	mad->base_version  = 1;
  56	mad->mgmt_class    = IB_MGMT_CLASS_SUBN_LID_ROUTED;
  57	mad->class_version = 1;
  58	mad->method    	   = IB_MGMT_METHOD_GET;
  59}
  60
  61static int mthca_query_device(struct ib_device *ibdev, struct ib_device_attr *props,
  62			      struct ib_udata *uhw)
  63{
  64	struct ib_smp *in_mad  = NULL;
  65	struct ib_smp *out_mad = NULL;
  66	int err = -ENOMEM;
  67	struct mthca_dev *mdev = to_mdev(ibdev);
  68
  69	if (uhw->inlen || uhw->outlen)
  70		return -EINVAL;
  71
  72	in_mad  = kzalloc(sizeof *in_mad, GFP_KERNEL);
  73	out_mad = kmalloc(sizeof *out_mad, GFP_KERNEL);
  74	if (!in_mad || !out_mad)
  75		goto out;
  76
  77	memset(props, 0, sizeof *props);
  78
  79	props->fw_ver              = mdev->fw_ver;
  80
  81	init_query_mad(in_mad);
  82	in_mad->attr_id = IB_SMP_ATTR_NODE_INFO;
  83
  84	err = mthca_MAD_IFC(mdev, 1, 1,
  85			    1, NULL, NULL, in_mad, out_mad);
  86	if (err)
  87		goto out;
  88
  89	props->device_cap_flags    = mdev->device_cap_flags;
  90	props->vendor_id           = be32_to_cpup((__be32 *) (out_mad->data + 36)) &
  91		0xffffff;
  92	props->vendor_part_id      = be16_to_cpup((__be16 *) (out_mad->data + 30));
  93	props->hw_ver              = be32_to_cpup((__be32 *) (out_mad->data + 32));
  94	memcpy(&props->sys_image_guid, out_mad->data +  4, 8);
  95
  96	props->max_mr_size         = ~0ull;
  97	props->page_size_cap       = mdev->limits.page_size_cap;
  98	props->max_qp              = mdev->limits.num_qps - mdev->limits.reserved_qps;
  99	props->max_qp_wr           = mdev->limits.max_wqes;
 100	props->max_send_sge        = mdev->limits.max_sg;
 101	props->max_recv_sge        = mdev->limits.max_sg;
 102	props->max_sge_rd          = mdev->limits.max_sg;
 103	props->max_cq              = mdev->limits.num_cqs - mdev->limits.reserved_cqs;
 104	props->max_cqe             = mdev->limits.max_cqes;
 105	props->max_mr              = mdev->limits.num_mpts - mdev->limits.reserved_mrws;
 106	props->max_pd              = mdev->limits.num_pds - mdev->limits.reserved_pds;
 107	props->max_qp_rd_atom      = 1 << mdev->qp_table.rdb_shift;
 108	props->max_qp_init_rd_atom = mdev->limits.max_qp_init_rdma;
 109	props->max_res_rd_atom     = props->max_qp_rd_atom * props->max_qp;
 110	props->max_srq             = mdev->limits.num_srqs - mdev->limits.reserved_srqs;
 111	props->max_srq_wr          = mdev->limits.max_srq_wqes;
 112	props->max_srq_sge         = mdev->limits.max_srq_sge;
 113	props->local_ca_ack_delay  = mdev->limits.local_ca_ack_delay;
 114	props->atomic_cap          = mdev->limits.flags & DEV_LIM_FLAG_ATOMIC ?
 115					IB_ATOMIC_HCA : IB_ATOMIC_NONE;
 116	props->max_pkeys           = mdev->limits.pkey_table_len;
 117	props->max_mcast_grp       = mdev->limits.num_mgms + mdev->limits.num_amgms;
 118	props->max_mcast_qp_attach = MTHCA_QP_PER_MGM;
 119	props->max_total_mcast_qp_attach = props->max_mcast_qp_attach *
 120					   props->max_mcast_grp;
 121	/*
 122	 * If Sinai memory key optimization is being used, then only
 123	 * the 8-bit key portion will change.  For other HCAs, the
 124	 * unused index bits will also be used for FMR remapping.
 125	 */
 126	if (mdev->mthca_flags & MTHCA_FLAG_SINAI_OPT)
 127		props->max_map_per_fmr = 255;
 128	else
 129		props->max_map_per_fmr =
 130			(1 << (32 - ilog2(mdev->limits.num_mpts))) - 1;
 131
 132	err = 0;
 133 out:
 134	kfree(in_mad);
 135	kfree(out_mad);
 136	return err;
 137}
 138
 139static int mthca_query_port(struct ib_device *ibdev,
 140			    u8 port, struct ib_port_attr *props)
 141{
 142	struct ib_smp *in_mad  = NULL;
 143	struct ib_smp *out_mad = NULL;
 144	int err = -ENOMEM;
 145
 146	in_mad  = kzalloc(sizeof *in_mad, GFP_KERNEL);
 147	out_mad = kmalloc(sizeof *out_mad, GFP_KERNEL);
 148	if (!in_mad || !out_mad)
 149		goto out;
 150
 151	/* props being zeroed by the caller, avoid zeroing it here */
 152
 153	init_query_mad(in_mad);
 154	in_mad->attr_id  = IB_SMP_ATTR_PORT_INFO;
 155	in_mad->attr_mod = cpu_to_be32(port);
 156
 157	err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1,
 158			    port, NULL, NULL, in_mad, out_mad);
 159	if (err)
 160		goto out;
 161
 162	props->lid               = be16_to_cpup((__be16 *) (out_mad->data + 16));
 163	props->lmc               = out_mad->data[34] & 0x7;
 164	props->sm_lid            = be16_to_cpup((__be16 *) (out_mad->data + 18));
 165	props->sm_sl             = out_mad->data[36] & 0xf;
 166	props->state             = out_mad->data[32] & 0xf;
 167	props->phys_state        = out_mad->data[33] >> 4;
 168	props->port_cap_flags    = be32_to_cpup((__be32 *) (out_mad->data + 20));
 169	props->gid_tbl_len       = to_mdev(ibdev)->limits.gid_table_len;
 170	props->max_msg_sz        = 0x80000000;
 171	props->pkey_tbl_len      = to_mdev(ibdev)->limits.pkey_table_len;
 172	props->bad_pkey_cntr     = be16_to_cpup((__be16 *) (out_mad->data + 46));
 173	props->qkey_viol_cntr    = be16_to_cpup((__be16 *) (out_mad->data + 48));
 174	props->active_width      = out_mad->data[31] & 0xf;
 175	props->active_speed      = out_mad->data[35] >> 4;
 176	props->max_mtu           = out_mad->data[41] & 0xf;
 177	props->active_mtu        = out_mad->data[36] >> 4;
 178	props->subnet_timeout    = out_mad->data[51] & 0x1f;
 179	props->max_vl_num        = out_mad->data[37] >> 4;
 180	props->init_type_reply   = out_mad->data[41] >> 4;
 181
 182 out:
 183	kfree(in_mad);
 184	kfree(out_mad);
 185	return err;
 186}
 187
 188static int mthca_modify_device(struct ib_device *ibdev,
 189			       int mask,
 190			       struct ib_device_modify *props)
 191{
 192	if (mask & ~IB_DEVICE_MODIFY_NODE_DESC)
 193		return -EOPNOTSUPP;
 194
 195	if (mask & IB_DEVICE_MODIFY_NODE_DESC) {
 196		if (mutex_lock_interruptible(&to_mdev(ibdev)->cap_mask_mutex))
 197			return -ERESTARTSYS;
 198		memcpy(ibdev->node_desc, props->node_desc,
 199		       IB_DEVICE_NODE_DESC_MAX);
 200		mutex_unlock(&to_mdev(ibdev)->cap_mask_mutex);
 201	}
 202
 203	return 0;
 204}
 205
 206static int mthca_modify_port(struct ib_device *ibdev,
 207			     u8 port, int port_modify_mask,
 208			     struct ib_port_modify *props)
 209{
 210	struct mthca_set_ib_param set_ib;
 211	struct ib_port_attr attr;
 212	int err;
 213
 214	if (mutex_lock_interruptible(&to_mdev(ibdev)->cap_mask_mutex))
 215		return -ERESTARTSYS;
 216
 217	err = ib_query_port(ibdev, port, &attr);
 218	if (err)
 219		goto out;
 220
 221	set_ib.set_si_guid     = 0;
 222	set_ib.reset_qkey_viol = !!(port_modify_mask & IB_PORT_RESET_QKEY_CNTR);
 223
 224	set_ib.cap_mask = (attr.port_cap_flags | props->set_port_cap_mask) &
 225		~props->clr_port_cap_mask;
 226
 227	err = mthca_SET_IB(to_mdev(ibdev), &set_ib, port);
 228	if (err)
 229		goto out;
 230out:
 231	mutex_unlock(&to_mdev(ibdev)->cap_mask_mutex);
 232	return err;
 233}
 234
 235static int mthca_query_pkey(struct ib_device *ibdev,
 236			    u8 port, u16 index, u16 *pkey)
 237{
 238	struct ib_smp *in_mad  = NULL;
 239	struct ib_smp *out_mad = NULL;
 240	int err = -ENOMEM;
 241
 242	in_mad  = kzalloc(sizeof *in_mad, GFP_KERNEL);
 243	out_mad = kmalloc(sizeof *out_mad, GFP_KERNEL);
 244	if (!in_mad || !out_mad)
 245		goto out;
 246
 247	init_query_mad(in_mad);
 248	in_mad->attr_id  = IB_SMP_ATTR_PKEY_TABLE;
 249	in_mad->attr_mod = cpu_to_be32(index / 32);
 250
 251	err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1,
 252			    port, NULL, NULL, in_mad, out_mad);
 253	if (err)
 254		goto out;
 255
 256	*pkey = be16_to_cpu(((__be16 *) out_mad->data)[index % 32]);
 257
 258 out:
 259	kfree(in_mad);
 260	kfree(out_mad);
 261	return err;
 262}
 263
 264static int mthca_query_gid(struct ib_device *ibdev, u8 port,
 265			   int index, union ib_gid *gid)
 266{
 267	struct ib_smp *in_mad  = NULL;
 268	struct ib_smp *out_mad = NULL;
 269	int err = -ENOMEM;
 270
 271	in_mad  = kzalloc(sizeof *in_mad, GFP_KERNEL);
 272	out_mad = kmalloc(sizeof *out_mad, GFP_KERNEL);
 273	if (!in_mad || !out_mad)
 274		goto out;
 275
 276	init_query_mad(in_mad);
 277	in_mad->attr_id  = IB_SMP_ATTR_PORT_INFO;
 278	in_mad->attr_mod = cpu_to_be32(port);
 279
 280	err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1,
 281			    port, NULL, NULL, in_mad, out_mad);
 282	if (err)
 283		goto out;
 284
 285	memcpy(gid->raw, out_mad->data + 8, 8);
 286
 287	init_query_mad(in_mad);
 288	in_mad->attr_id  = IB_SMP_ATTR_GUID_INFO;
 289	in_mad->attr_mod = cpu_to_be32(index / 8);
 290
 291	err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1,
 292			    port, NULL, NULL, in_mad, out_mad);
 293	if (err)
 294		goto out;
 295
 296	memcpy(gid->raw + 8, out_mad->data + (index % 8) * 8, 8);
 297
 298 out:
 299	kfree(in_mad);
 300	kfree(out_mad);
 301	return err;
 302}
 303
 304static int mthca_alloc_ucontext(struct ib_ucontext *uctx,
 305				struct ib_udata *udata)
 306{
 307	struct ib_device *ibdev = uctx->device;
 308	struct mthca_alloc_ucontext_resp uresp = {};
 309	struct mthca_ucontext *context = to_mucontext(uctx);
 310	int                              err;
 311
 312	if (!(to_mdev(ibdev)->active))
 313		return -EAGAIN;
 314
 315	uresp.qp_tab_size = to_mdev(ibdev)->limits.num_qps;
 316	if (mthca_is_memfree(to_mdev(ibdev)))
 317		uresp.uarc_size = to_mdev(ibdev)->uar_table.uarc_size;
 318	else
 319		uresp.uarc_size = 0;
 320
 321	err = mthca_uar_alloc(to_mdev(ibdev), &context->uar);
 322	if (err)
 323		return err;
 324
 325	context->db_tab = mthca_init_user_db_tab(to_mdev(ibdev));
 326	if (IS_ERR(context->db_tab)) {
 327		err = PTR_ERR(context->db_tab);
 328		mthca_uar_free(to_mdev(ibdev), &context->uar);
 329		return err;
 330	}
 331
 332	if (ib_copy_to_udata(udata, &uresp, sizeof(uresp))) {
 333		mthca_cleanup_user_db_tab(to_mdev(ibdev), &context->uar, context->db_tab);
 334		mthca_uar_free(to_mdev(ibdev), &context->uar);
 335		return -EFAULT;
 336	}
 337
 338	context->reg_mr_warned = 0;
 339
 340	return 0;
 341}
 342
 343static void mthca_dealloc_ucontext(struct ib_ucontext *context)
 344{
 345	mthca_cleanup_user_db_tab(to_mdev(context->device), &to_mucontext(context)->uar,
 346				  to_mucontext(context)->db_tab);
 347	mthca_uar_free(to_mdev(context->device), &to_mucontext(context)->uar);
 348}
 349
 350static int mthca_mmap_uar(struct ib_ucontext *context,
 351			  struct vm_area_struct *vma)
 352{
 353	if (vma->vm_end - vma->vm_start != PAGE_SIZE)
 354		return -EINVAL;
 355
 356	vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
 357
 358	if (io_remap_pfn_range(vma, vma->vm_start,
 359			       to_mucontext(context)->uar.pfn,
 360			       PAGE_SIZE, vma->vm_page_prot))
 361		return -EAGAIN;
 362
 363	return 0;
 364}
 365
 366static int mthca_alloc_pd(struct ib_pd *ibpd, struct ib_udata *udata)
 367{
 368	struct ib_device *ibdev = ibpd->device;
 369	struct mthca_pd *pd = to_mpd(ibpd);
 370	int err;
 371
 372	err = mthca_pd_alloc(to_mdev(ibdev), !udata, pd);
 373	if (err)
 374		return err;
 375
 376	if (udata) {
 377		if (ib_copy_to_udata(udata, &pd->pd_num, sizeof (__u32))) {
 378			mthca_pd_free(to_mdev(ibdev), pd);
 379			return -EFAULT;
 380		}
 381	}
 382
 383	return 0;
 384}
 385
 386static void mthca_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata)
 387{
 388	mthca_pd_free(to_mdev(pd->device), to_mpd(pd));
 
 389}
 390
 391static int mthca_ah_create(struct ib_ah *ibah, struct rdma_ah_attr *ah_attr,
 392			   u32 flags, struct ib_udata *udata)
 
 393
 394{
 395	struct mthca_ah *ah = to_mah(ibah);
 396
 397	return mthca_create_ah(to_mdev(ibah->device), to_mpd(ibah->pd), ah_attr,
 398			       ah);
 399}
 400
 401static void mthca_ah_destroy(struct ib_ah *ah, u32 flags)
 402{
 403	mthca_destroy_ah(to_mdev(ah->device), to_mah(ah));
 
 404}
 405
 406static int mthca_create_srq(struct ib_srq *ibsrq,
 407			    struct ib_srq_init_attr *init_attr,
 408			    struct ib_udata *udata)
 409{
 410	struct mthca_create_srq ucmd;
 411	struct mthca_ucontext *context = rdma_udata_to_drv_context(
 412		udata, struct mthca_ucontext, ibucontext);
 413	struct mthca_srq *srq = to_msrq(ibsrq);
 414	int err;
 415
 416	if (init_attr->srq_type != IB_SRQT_BASIC)
 417		return -EOPNOTSUPP;
 418
 419	if (udata) {
 420		if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd)))
 421			return -EFAULT;
 422
 423		err = mthca_map_user_db(to_mdev(ibsrq->device), &context->uar,
 424					context->db_tab, ucmd.db_index,
 425					ucmd.db_page);
 426
 427		if (err)
 428			return err;
 429
 430		srq->mr.ibmr.lkey = ucmd.lkey;
 431		srq->db_index     = ucmd.db_index;
 432	}
 433
 434	err = mthca_alloc_srq(to_mdev(ibsrq->device), to_mpd(ibsrq->pd),
 435			      &init_attr->attr, srq, udata);
 436
 437	if (err && udata)
 438		mthca_unmap_user_db(to_mdev(ibsrq->device), &context->uar,
 439				    context->db_tab, ucmd.db_index);
 440
 441	if (err)
 442		return err;
 443
 444	if (context && ib_copy_to_udata(udata, &srq->srqn, sizeof(__u32))) {
 445		mthca_free_srq(to_mdev(ibsrq->device), srq);
 446		return -EFAULT;
 447	}
 448
 449	return 0;
 450}
 451
 452static void mthca_destroy_srq(struct ib_srq *srq, struct ib_udata *udata)
 453{
 454	if (udata) {
 455		struct mthca_ucontext *context =
 456			rdma_udata_to_drv_context(
 457				udata,
 458				struct mthca_ucontext,
 459				ibucontext);
 460
 461		mthca_unmap_user_db(to_mdev(srq->device), &context->uar,
 462				    context->db_tab, to_msrq(srq)->db_index);
 463	}
 464
 465	mthca_free_srq(to_mdev(srq->device), to_msrq(srq));
 
 466}
 467
 468static struct ib_qp *mthca_create_qp(struct ib_pd *pd,
 469				     struct ib_qp_init_attr *init_attr,
 470				     struct ib_udata *udata)
 471{
 472	struct mthca_ucontext *context = rdma_udata_to_drv_context(
 473		udata, struct mthca_ucontext, ibucontext);
 474	struct mthca_create_qp ucmd;
 475	struct mthca_qp *qp;
 
 476	int err;
 477
 478	if (init_attr->create_flags)
 479		return ERR_PTR(-EINVAL);
 480
 481	switch (init_attr->qp_type) {
 482	case IB_QPT_RC:
 483	case IB_QPT_UC:
 484	case IB_QPT_UD:
 485	{
 486		qp = kzalloc(sizeof(*qp), GFP_KERNEL);
 487		if (!qp)
 488			return ERR_PTR(-ENOMEM);
 489
 490		if (udata) {
 491			if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) {
 492				kfree(qp);
 493				return ERR_PTR(-EFAULT);
 494			}
 495
 496			err = mthca_map_user_db(to_mdev(pd->device), &context->uar,
 497						context->db_tab,
 498						ucmd.sq_db_index, ucmd.sq_db_page);
 499			if (err) {
 500				kfree(qp);
 501				return ERR_PTR(err);
 502			}
 503
 504			err = mthca_map_user_db(to_mdev(pd->device), &context->uar,
 505						context->db_tab,
 506						ucmd.rq_db_index, ucmd.rq_db_page);
 
 507			if (err) {
 508				mthca_unmap_user_db(to_mdev(pd->device),
 509						    &context->uar,
 510						    context->db_tab,
 511						    ucmd.sq_db_index);
 512				kfree(qp);
 513				return ERR_PTR(err);
 514			}
 515
 516			qp->mr.ibmr.lkey = ucmd.lkey;
 517			qp->sq.db_index  = ucmd.sq_db_index;
 518			qp->rq.db_index  = ucmd.rq_db_index;
 519		}
 520
 521		err = mthca_alloc_qp(to_mdev(pd->device), to_mpd(pd),
 522				     to_mcq(init_attr->send_cq),
 523				     to_mcq(init_attr->recv_cq),
 524				     init_attr->qp_type, init_attr->sq_sig_type,
 525				     &init_attr->cap, qp, udata);
 526
 527		if (err && udata) {
 528			mthca_unmap_user_db(to_mdev(pd->device),
 529					    &context->uar,
 530					    context->db_tab,
 531					    ucmd.sq_db_index);
 532			mthca_unmap_user_db(to_mdev(pd->device),
 533					    &context->uar,
 534					    context->db_tab,
 535					    ucmd.rq_db_index);
 536		}
 537
 538		qp->ibqp.qp_num = qp->qpn;
 539		break;
 540	}
 541	case IB_QPT_SMI:
 542	case IB_QPT_GSI:
 543	{
 544		/* Don't allow userspace to create special QPs */
 545		if (udata)
 546			return ERR_PTR(-EINVAL);
 547
 548		qp = kzalloc(sizeof(struct mthca_sqp), GFP_KERNEL);
 549		if (!qp)
 550			return ERR_PTR(-ENOMEM);
 551
 552		qp->ibqp.qp_num = init_attr->qp_type == IB_QPT_SMI ? 0 : 1;
 553
 554		err = mthca_alloc_sqp(to_mdev(pd->device), to_mpd(pd),
 555				      to_mcq(init_attr->send_cq),
 556				      to_mcq(init_attr->recv_cq),
 557				      init_attr->sq_sig_type, &init_attr->cap,
 558				      qp->ibqp.qp_num, init_attr->port_num,
 559				      to_msqp(qp), udata);
 560		break;
 561	}
 562	default:
 563		/* Don't support raw QPs */
 564		return ERR_PTR(-ENOSYS);
 565	}
 566
 567	if (err) {
 568		kfree(qp);
 569		return ERR_PTR(err);
 570	}
 571
 572	init_attr->cap.max_send_wr     = qp->sq.max;
 573	init_attr->cap.max_recv_wr     = qp->rq.max;
 574	init_attr->cap.max_send_sge    = qp->sq.max_gs;
 575	init_attr->cap.max_recv_sge    = qp->rq.max_gs;
 576	init_attr->cap.max_inline_data = qp->max_inline_data;
 577
 578	return &qp->ibqp;
 579}
 580
 581static int mthca_destroy_qp(struct ib_qp *qp, struct ib_udata *udata)
 582{
 583	if (udata) {
 584		struct mthca_ucontext *context =
 585			rdma_udata_to_drv_context(
 586				udata,
 587				struct mthca_ucontext,
 588				ibucontext);
 589
 590		mthca_unmap_user_db(to_mdev(qp->device),
 591				    &context->uar,
 592				    context->db_tab,
 593				    to_mqp(qp)->sq.db_index);
 594		mthca_unmap_user_db(to_mdev(qp->device),
 595				    &context->uar,
 596				    context->db_tab,
 597				    to_mqp(qp)->rq.db_index);
 598	}
 599	mthca_free_qp(to_mdev(qp->device), to_mqp(qp));
 600	kfree(qp);
 601	return 0;
 602}
 603
 604static int mthca_create_cq(struct ib_cq *ibcq,
 605			   const struct ib_cq_init_attr *attr,
 606			   struct ib_udata *udata)
 607{
 608	struct ib_device *ibdev = ibcq->device;
 609	int entries = attr->cqe;
 610	struct mthca_create_cq ucmd;
 611	struct mthca_cq *cq;
 612	int nent;
 613	int err;
 614	struct mthca_ucontext *context = rdma_udata_to_drv_context(
 615		udata, struct mthca_ucontext, ibucontext);
 616
 617	if (attr->flags)
 618		return -EINVAL;
 619
 620	if (entries < 1 || entries > to_mdev(ibdev)->limits.max_cqes)
 621		return -EINVAL;
 622
 623	if (udata) {
 624		if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd)))
 625			return -EFAULT;
 626
 627		err = mthca_map_user_db(to_mdev(ibdev), &context->uar,
 628					context->db_tab, ucmd.set_db_index,
 629					ucmd.set_db_page);
 630		if (err)
 631			return err;
 632
 633		err = mthca_map_user_db(to_mdev(ibdev), &context->uar,
 634					context->db_tab, ucmd.arm_db_index,
 635					ucmd.arm_db_page);
 636		if (err)
 637			goto err_unmap_set;
 638	}
 639
 640	cq = to_mcq(ibcq);
 641
 642	if (udata) {
 643		cq->buf.mr.ibmr.lkey = ucmd.lkey;
 644		cq->set_ci_db_index  = ucmd.set_db_index;
 645		cq->arm_db_index     = ucmd.arm_db_index;
 646	}
 647
 648	for (nent = 1; nent <= entries; nent <<= 1)
 649		; /* nothing */
 650
 651	err = mthca_init_cq(to_mdev(ibdev), nent, context,
 652			    udata ? ucmd.pdn : to_mdev(ibdev)->driver_pd.pd_num,
 653			    cq);
 654	if (err)
 655		goto err_unmap_arm;
 656
 657	if (udata && ib_copy_to_udata(udata, &cq->cqn, sizeof(__u32))) {
 658		mthca_free_cq(to_mdev(ibdev), cq);
 659		err = -EFAULT;
 660		goto err_unmap_arm;
 661	}
 662
 663	cq->resize_buf = NULL;
 664
 665	return 0;
 666
 667err_unmap_arm:
 668	if (udata)
 669		mthca_unmap_user_db(to_mdev(ibdev), &context->uar,
 670				    context->db_tab, ucmd.arm_db_index);
 671
 672err_unmap_set:
 673	if (udata)
 674		mthca_unmap_user_db(to_mdev(ibdev), &context->uar,
 675				    context->db_tab, ucmd.set_db_index);
 676
 677	return err;
 678}
 679
 680static int mthca_alloc_resize_buf(struct mthca_dev *dev, struct mthca_cq *cq,
 681				  int entries)
 682{
 683	int ret;
 684
 685	spin_lock_irq(&cq->lock);
 686	if (cq->resize_buf) {
 687		ret = -EBUSY;
 688		goto unlock;
 689	}
 690
 691	cq->resize_buf = kmalloc(sizeof *cq->resize_buf, GFP_ATOMIC);
 692	if (!cq->resize_buf) {
 693		ret = -ENOMEM;
 694		goto unlock;
 695	}
 696
 697	cq->resize_buf->state = CQ_RESIZE_ALLOC;
 698
 699	ret = 0;
 700
 701unlock:
 702	spin_unlock_irq(&cq->lock);
 703
 704	if (ret)
 705		return ret;
 706
 707	ret = mthca_alloc_cq_buf(dev, &cq->resize_buf->buf, entries);
 708	if (ret) {
 709		spin_lock_irq(&cq->lock);
 710		kfree(cq->resize_buf);
 711		cq->resize_buf = NULL;
 712		spin_unlock_irq(&cq->lock);
 713		return ret;
 714	}
 715
 716	cq->resize_buf->cqe = entries - 1;
 717
 718	spin_lock_irq(&cq->lock);
 719	cq->resize_buf->state = CQ_RESIZE_READY;
 720	spin_unlock_irq(&cq->lock);
 721
 722	return 0;
 723}
 724
 725static int mthca_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata)
 726{
 727	struct mthca_dev *dev = to_mdev(ibcq->device);
 728	struct mthca_cq *cq = to_mcq(ibcq);
 729	struct mthca_resize_cq ucmd;
 730	u32 lkey;
 731	int ret;
 732
 733	if (entries < 1 || entries > dev->limits.max_cqes)
 734		return -EINVAL;
 735
 736	mutex_lock(&cq->mutex);
 737
 738	entries = roundup_pow_of_two(entries + 1);
 739	if (entries == ibcq->cqe + 1) {
 740		ret = 0;
 741		goto out;
 742	}
 743
 744	if (cq->is_kernel) {
 745		ret = mthca_alloc_resize_buf(dev, cq, entries);
 746		if (ret)
 747			goto out;
 748		lkey = cq->resize_buf->buf.mr.ibmr.lkey;
 749	} else {
 750		if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) {
 751			ret = -EFAULT;
 752			goto out;
 753		}
 754		lkey = ucmd.lkey;
 755	}
 756
 757	ret = mthca_RESIZE_CQ(dev, cq->cqn, lkey, ilog2(entries));
 758
 759	if (ret) {
 760		if (cq->resize_buf) {
 761			mthca_free_cq_buf(dev, &cq->resize_buf->buf,
 762					  cq->resize_buf->cqe);
 763			kfree(cq->resize_buf);
 764			spin_lock_irq(&cq->lock);
 765			cq->resize_buf = NULL;
 766			spin_unlock_irq(&cq->lock);
 767		}
 768		goto out;
 769	}
 770
 771	if (cq->is_kernel) {
 772		struct mthca_cq_buf tbuf;
 773		int tcqe;
 774
 775		spin_lock_irq(&cq->lock);
 776		if (cq->resize_buf->state == CQ_RESIZE_READY) {
 777			mthca_cq_resize_copy_cqes(cq);
 778			tbuf         = cq->buf;
 779			tcqe         = cq->ibcq.cqe;
 780			cq->buf      = cq->resize_buf->buf;
 781			cq->ibcq.cqe = cq->resize_buf->cqe;
 782		} else {
 783			tbuf = cq->resize_buf->buf;
 784			tcqe = cq->resize_buf->cqe;
 785		}
 786
 787		kfree(cq->resize_buf);
 788		cq->resize_buf = NULL;
 789		spin_unlock_irq(&cq->lock);
 790
 791		mthca_free_cq_buf(dev, &tbuf, tcqe);
 792	} else
 793		ibcq->cqe = entries - 1;
 794
 795out:
 796	mutex_unlock(&cq->mutex);
 797
 798	return ret;
 799}
 800
 801static void mthca_destroy_cq(struct ib_cq *cq, struct ib_udata *udata)
 802{
 803	if (udata) {
 804		struct mthca_ucontext *context =
 805			rdma_udata_to_drv_context(
 806				udata,
 807				struct mthca_ucontext,
 808				ibucontext);
 809
 810		mthca_unmap_user_db(to_mdev(cq->device),
 811				    &context->uar,
 812				    context->db_tab,
 813				    to_mcq(cq)->arm_db_index);
 814		mthca_unmap_user_db(to_mdev(cq->device),
 815				    &context->uar,
 816				    context->db_tab,
 817				    to_mcq(cq)->set_ci_db_index);
 818	}
 819	mthca_free_cq(to_mdev(cq->device), to_mcq(cq));
 
 820}
 821
 822static inline u32 convert_access(int acc)
 823{
 824	return (acc & IB_ACCESS_REMOTE_ATOMIC ? MTHCA_MPT_FLAG_ATOMIC       : 0) |
 825	       (acc & IB_ACCESS_REMOTE_WRITE  ? MTHCA_MPT_FLAG_REMOTE_WRITE : 0) |
 826	       (acc & IB_ACCESS_REMOTE_READ   ? MTHCA_MPT_FLAG_REMOTE_READ  : 0) |
 827	       (acc & IB_ACCESS_LOCAL_WRITE   ? MTHCA_MPT_FLAG_LOCAL_WRITE  : 0) |
 828	       MTHCA_MPT_FLAG_LOCAL_READ;
 829}
 830
 831static struct ib_mr *mthca_get_dma_mr(struct ib_pd *pd, int acc)
 832{
 833	struct mthca_mr *mr;
 834	int err;
 835
 836	mr = kmalloc(sizeof *mr, GFP_KERNEL);
 837	if (!mr)
 838		return ERR_PTR(-ENOMEM);
 839
 840	err = mthca_mr_alloc_notrans(to_mdev(pd->device),
 841				     to_mpd(pd)->pd_num,
 842				     convert_access(acc), mr);
 843
 844	if (err) {
 845		kfree(mr);
 846		return ERR_PTR(err);
 847	}
 848
 849	mr->umem = NULL;
 850
 851	return &mr->ibmr;
 852}
 853
 854static struct ib_mr *mthca_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
 855				       u64 virt, int acc, struct ib_udata *udata)
 856{
 857	struct mthca_dev *dev = to_mdev(pd->device);
 858	struct sg_dma_page_iter sg_iter;
 859	struct mthca_ucontext *context = rdma_udata_to_drv_context(
 860		udata, struct mthca_ucontext, ibucontext);
 861	struct mthca_mr *mr;
 862	struct mthca_reg_mr ucmd;
 863	u64 *pages;
 864	int n, i;
 865	int err = 0;
 866	int write_mtt_size;
 867
 868	if (udata->inlen < sizeof ucmd) {
 869		if (!context->reg_mr_warned) {
 870			mthca_warn(dev, "Process '%s' did not pass in MR attrs.\n",
 871				   current->comm);
 872			mthca_warn(dev, "  Update libmthca to fix this.\n");
 873		}
 874		++context->reg_mr_warned;
 875		ucmd.mr_attrs = 0;
 876	} else if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd))
 877		return ERR_PTR(-EFAULT);
 878
 879	mr = kmalloc(sizeof *mr, GFP_KERNEL);
 880	if (!mr)
 881		return ERR_PTR(-ENOMEM);
 882
 883	mr->umem = ib_umem_get(udata, start, length, acc,
 884			       ucmd.mr_attrs & MTHCA_MR_DMASYNC);
 885
 886	if (IS_ERR(mr->umem)) {
 887		err = PTR_ERR(mr->umem);
 888		goto err;
 889	}
 890
 891	n = ib_umem_num_pages(mr->umem);
 892
 893	mr->mtt = mthca_alloc_mtt(dev, n);
 894	if (IS_ERR(mr->mtt)) {
 895		err = PTR_ERR(mr->mtt);
 896		goto err_umem;
 897	}
 898
 899	pages = (u64 *) __get_free_page(GFP_KERNEL);
 900	if (!pages) {
 901		err = -ENOMEM;
 902		goto err_mtt;
 903	}
 904
 905	i = n = 0;
 906
 907	write_mtt_size = min(mthca_write_mtt_size(dev), (int) (PAGE_SIZE / sizeof *pages));
 908
 909	for_each_sg_dma_page(mr->umem->sg_head.sgl, &sg_iter, mr->umem->nmap, 0) {
 910		pages[i++] = sg_page_iter_dma_address(&sg_iter);
 911
 912		/*
 913		 * Be friendly to write_mtt and pass it chunks
 914		 * of appropriate size.
 915		 */
 916		if (i == write_mtt_size) {
 917			err = mthca_write_mtt(dev, mr->mtt, n, pages, i);
 918			if (err)
 919				goto mtt_done;
 920			n += i;
 921			i = 0;
 922		}
 923	}
 924
 925	if (i)
 926		err = mthca_write_mtt(dev, mr->mtt, n, pages, i);
 927mtt_done:
 928	free_page((unsigned long) pages);
 929	if (err)
 930		goto err_mtt;
 931
 932	err = mthca_mr_alloc(dev, to_mpd(pd)->pd_num, PAGE_SHIFT, virt, length,
 933			     convert_access(acc), mr);
 934
 935	if (err)
 936		goto err_mtt;
 937
 938	return &mr->ibmr;
 939
 940err_mtt:
 941	mthca_free_mtt(dev, mr->mtt);
 942
 943err_umem:
 944	ib_umem_release(mr->umem);
 945
 946err:
 947	kfree(mr);
 948	return ERR_PTR(err);
 949}
 950
 951static int mthca_dereg_mr(struct ib_mr *mr, struct ib_udata *udata)
 952{
 953	struct mthca_mr *mmr = to_mmr(mr);
 954
 955	mthca_free_mr(to_mdev(mr->device), mmr);
 956	ib_umem_release(mmr->umem);
 957	kfree(mmr);
 958
 959	return 0;
 960}
 961
 962static struct ib_fmr *mthca_alloc_fmr(struct ib_pd *pd, int mr_access_flags,
 963				      struct ib_fmr_attr *fmr_attr)
 964{
 965	struct mthca_fmr *fmr;
 966	int err;
 967
 968	fmr = kmalloc(sizeof *fmr, GFP_KERNEL);
 969	if (!fmr)
 970		return ERR_PTR(-ENOMEM);
 971
 972	memcpy(&fmr->attr, fmr_attr, sizeof *fmr_attr);
 973	err = mthca_fmr_alloc(to_mdev(pd->device), to_mpd(pd)->pd_num,
 974			     convert_access(mr_access_flags), fmr);
 975
 976	if (err) {
 977		kfree(fmr);
 978		return ERR_PTR(err);
 979	}
 980
 981	return &fmr->ibmr;
 982}
 983
 984static int mthca_dealloc_fmr(struct ib_fmr *fmr)
 985{
 986	struct mthca_fmr *mfmr = to_mfmr(fmr);
 987	int err;
 988
 989	err = mthca_free_fmr(to_mdev(fmr->device), mfmr);
 990	if (err)
 991		return err;
 992
 993	kfree(mfmr);
 994	return 0;
 995}
 996
 997static int mthca_unmap_fmr(struct list_head *fmr_list)
 998{
 999	struct ib_fmr *fmr;
1000	int err;
1001	struct mthca_dev *mdev = NULL;
1002
1003	list_for_each_entry(fmr, fmr_list, list) {
1004		if (mdev && to_mdev(fmr->device) != mdev)
1005			return -EINVAL;
1006		mdev = to_mdev(fmr->device);
1007	}
1008
1009	if (!mdev)
1010		return 0;
1011
1012	if (mthca_is_memfree(mdev)) {
1013		list_for_each_entry(fmr, fmr_list, list)
1014			mthca_arbel_fmr_unmap(mdev, to_mfmr(fmr));
1015
1016		wmb();
1017	} else
1018		list_for_each_entry(fmr, fmr_list, list)
1019			mthca_tavor_fmr_unmap(mdev, to_mfmr(fmr));
1020
1021	err = mthca_SYNC_TPT(mdev);
1022	return err;
1023}
1024
1025static ssize_t hw_rev_show(struct device *device,
1026			   struct device_attribute *attr, char *buf)
1027{
1028	struct mthca_dev *dev =
1029		rdma_device_to_drv_device(device, struct mthca_dev, ib_dev);
1030
1031	return sprintf(buf, "%x\n", dev->rev_id);
1032}
1033static DEVICE_ATTR_RO(hw_rev);
1034
1035static ssize_t hca_type_show(struct device *device,
1036			     struct device_attribute *attr, char *buf)
1037{
1038	struct mthca_dev *dev =
1039		rdma_device_to_drv_device(device, struct mthca_dev, ib_dev);
1040
1041	switch (dev->pdev->device) {
1042	case PCI_DEVICE_ID_MELLANOX_TAVOR:
1043		return sprintf(buf, "MT23108\n");
1044	case PCI_DEVICE_ID_MELLANOX_ARBEL_COMPAT:
1045		return sprintf(buf, "MT25208 (MT23108 compat mode)\n");
1046	case PCI_DEVICE_ID_MELLANOX_ARBEL:
1047		return sprintf(buf, "MT25208\n");
1048	case PCI_DEVICE_ID_MELLANOX_SINAI:
1049	case PCI_DEVICE_ID_MELLANOX_SINAI_OLD:
1050		return sprintf(buf, "MT25204\n");
1051	default:
1052		return sprintf(buf, "unknown\n");
1053	}
 
 
 
 
 
 
 
 
 
 
 
1054}
1055static DEVICE_ATTR_RO(hca_type);
1056
1057static ssize_t board_id_show(struct device *device,
1058			     struct device_attribute *attr, char *buf)
1059{
1060	struct mthca_dev *dev =
1061		rdma_device_to_drv_device(device, struct mthca_dev, ib_dev);
1062
1063	return sprintf(buf, "%.*s\n", MTHCA_BOARD_ID_LEN, dev->board_id);
1064}
1065static DEVICE_ATTR_RO(board_id);
1066
1067static struct attribute *mthca_dev_attributes[] = {
1068	&dev_attr_hw_rev.attr,
1069	&dev_attr_hca_type.attr,
1070	&dev_attr_board_id.attr,
1071	NULL
1072};
1073
1074static const struct attribute_group mthca_attr_group = {
1075	.attrs = mthca_dev_attributes,
1076};
1077
1078static int mthca_init_node_data(struct mthca_dev *dev)
1079{
1080	struct ib_smp *in_mad  = NULL;
1081	struct ib_smp *out_mad = NULL;
1082	int err = -ENOMEM;
1083
1084	in_mad  = kzalloc(sizeof *in_mad, GFP_KERNEL);
1085	out_mad = kmalloc(sizeof *out_mad, GFP_KERNEL);
1086	if (!in_mad || !out_mad)
1087		goto out;
1088
1089	init_query_mad(in_mad);
1090	in_mad->attr_id = IB_SMP_ATTR_NODE_DESC;
1091
1092	err = mthca_MAD_IFC(dev, 1, 1,
1093			    1, NULL, NULL, in_mad, out_mad);
1094	if (err)
1095		goto out;
1096
1097	memcpy(dev->ib_dev.node_desc, out_mad->data, IB_DEVICE_NODE_DESC_MAX);
1098
1099	in_mad->attr_id = IB_SMP_ATTR_NODE_INFO;
1100
1101	err = mthca_MAD_IFC(dev, 1, 1,
1102			    1, NULL, NULL, in_mad, out_mad);
1103	if (err)
1104		goto out;
1105
1106	if (mthca_is_memfree(dev))
1107		dev->rev_id = be32_to_cpup((__be32 *) (out_mad->data + 32));
1108	memcpy(&dev->ib_dev.node_guid, out_mad->data + 12, 8);
1109
1110out:
1111	kfree(in_mad);
1112	kfree(out_mad);
1113	return err;
1114}
1115
1116static int mthca_port_immutable(struct ib_device *ibdev, u8 port_num,
1117			        struct ib_port_immutable *immutable)
1118{
1119	struct ib_port_attr attr;
1120	int err;
1121
1122	immutable->core_cap_flags = RDMA_CORE_PORT_IBA_IB;
1123
1124	err = ib_query_port(ibdev, port_num, &attr);
1125	if (err)
1126		return err;
1127
1128	immutable->pkey_tbl_len = attr.pkey_tbl_len;
1129	immutable->gid_tbl_len = attr.gid_tbl_len;
1130	immutable->max_mad_size = IB_MGMT_MAD_SIZE;
1131
1132	return 0;
1133}
1134
1135static void get_dev_fw_str(struct ib_device *device, char *str)
1136{
1137	struct mthca_dev *dev =
1138		container_of(device, struct mthca_dev, ib_dev);
1139	snprintf(str, IB_FW_VERSION_NAME_MAX, "%d.%d.%d",
1140		 (int) (dev->fw_ver >> 32),
1141		 (int) (dev->fw_ver >> 16) & 0xffff,
1142		 (int) dev->fw_ver & 0xffff);
1143}
1144
1145static const struct ib_device_ops mthca_dev_ops = {
1146	.owner = THIS_MODULE,
1147	.driver_id = RDMA_DRIVER_MTHCA,
1148	.uverbs_abi_ver = MTHCA_UVERBS_ABI_VERSION,
1149	.uverbs_no_driver_id_binding = 1,
1150
1151	.alloc_pd = mthca_alloc_pd,
1152	.alloc_ucontext = mthca_alloc_ucontext,
1153	.attach_mcast = mthca_multicast_attach,
1154	.create_ah = mthca_ah_create,
1155	.create_cq = mthca_create_cq,
1156	.create_qp = mthca_create_qp,
1157	.dealloc_pd = mthca_dealloc_pd,
1158	.dealloc_ucontext = mthca_dealloc_ucontext,
1159	.dereg_mr = mthca_dereg_mr,
1160	.destroy_ah = mthca_ah_destroy,
1161	.destroy_cq = mthca_destroy_cq,
1162	.destroy_qp = mthca_destroy_qp,
1163	.detach_mcast = mthca_multicast_detach,
 
1164	.get_dev_fw_str = get_dev_fw_str,
1165	.get_dma_mr = mthca_get_dma_mr,
1166	.get_port_immutable = mthca_port_immutable,
1167	.mmap = mthca_mmap_uar,
1168	.modify_device = mthca_modify_device,
1169	.modify_port = mthca_modify_port,
1170	.modify_qp = mthca_modify_qp,
1171	.poll_cq = mthca_poll_cq,
1172	.process_mad = mthca_process_mad,
1173	.query_ah = mthca_ah_query,
1174	.query_device = mthca_query_device,
1175	.query_gid = mthca_query_gid,
1176	.query_pkey = mthca_query_pkey,
1177	.query_port = mthca_query_port,
1178	.query_qp = mthca_query_qp,
1179	.reg_user_mr = mthca_reg_user_mr,
1180	.resize_cq = mthca_resize_cq,
1181
1182	INIT_RDMA_OBJ_SIZE(ib_ah, mthca_ah, ibah),
1183	INIT_RDMA_OBJ_SIZE(ib_cq, mthca_cq, ibcq),
1184	INIT_RDMA_OBJ_SIZE(ib_pd, mthca_pd, ibpd),
 
1185	INIT_RDMA_OBJ_SIZE(ib_ucontext, mthca_ucontext, ibucontext),
1186};
1187
1188static const struct ib_device_ops mthca_dev_arbel_srq_ops = {
1189	.create_srq = mthca_create_srq,
1190	.destroy_srq = mthca_destroy_srq,
1191	.modify_srq = mthca_modify_srq,
1192	.post_srq_recv = mthca_arbel_post_srq_recv,
1193	.query_srq = mthca_query_srq,
1194
1195	INIT_RDMA_OBJ_SIZE(ib_srq, mthca_srq, ibsrq),
1196};
1197
1198static const struct ib_device_ops mthca_dev_tavor_srq_ops = {
1199	.create_srq = mthca_create_srq,
1200	.destroy_srq = mthca_destroy_srq,
1201	.modify_srq = mthca_modify_srq,
1202	.post_srq_recv = mthca_tavor_post_srq_recv,
1203	.query_srq = mthca_query_srq,
1204
1205	INIT_RDMA_OBJ_SIZE(ib_srq, mthca_srq, ibsrq),
1206};
1207
1208static const struct ib_device_ops mthca_dev_arbel_fmr_ops = {
1209	.alloc_fmr = mthca_alloc_fmr,
1210	.dealloc_fmr = mthca_dealloc_fmr,
1211	.map_phys_fmr = mthca_arbel_map_phys_fmr,
1212	.unmap_fmr = mthca_unmap_fmr,
1213};
1214
1215static const struct ib_device_ops mthca_dev_tavor_fmr_ops = {
1216	.alloc_fmr = mthca_alloc_fmr,
1217	.dealloc_fmr = mthca_dealloc_fmr,
1218	.map_phys_fmr = mthca_tavor_map_phys_fmr,
1219	.unmap_fmr = mthca_unmap_fmr,
1220};
1221
1222static const struct ib_device_ops mthca_dev_arbel_ops = {
1223	.post_recv = mthca_arbel_post_receive,
1224	.post_send = mthca_arbel_post_send,
1225	.req_notify_cq = mthca_arbel_arm_cq,
1226};
1227
1228static const struct ib_device_ops mthca_dev_tavor_ops = {
1229	.post_recv = mthca_tavor_post_receive,
1230	.post_send = mthca_tavor_post_send,
1231	.req_notify_cq = mthca_tavor_arm_cq,
1232};
1233
1234int mthca_register_device(struct mthca_dev *dev)
1235{
1236	int ret;
1237
1238	ret = mthca_init_node_data(dev);
1239	if (ret)
1240		return ret;
1241
1242	dev->ib_dev.uverbs_cmd_mask	 =
1243		(1ull << IB_USER_VERBS_CMD_GET_CONTEXT)		|
1244		(1ull << IB_USER_VERBS_CMD_QUERY_DEVICE)	|
1245		(1ull << IB_USER_VERBS_CMD_QUERY_PORT)		|
1246		(1ull << IB_USER_VERBS_CMD_ALLOC_PD)		|
1247		(1ull << IB_USER_VERBS_CMD_DEALLOC_PD)		|
1248		(1ull << IB_USER_VERBS_CMD_REG_MR)		|
1249		(1ull << IB_USER_VERBS_CMD_DEREG_MR)		|
1250		(1ull << IB_USER_VERBS_CMD_CREATE_COMP_CHANNEL)	|
1251		(1ull << IB_USER_VERBS_CMD_CREATE_CQ)		|
1252		(1ull << IB_USER_VERBS_CMD_RESIZE_CQ)		|
1253		(1ull << IB_USER_VERBS_CMD_DESTROY_CQ)		|
1254		(1ull << IB_USER_VERBS_CMD_CREATE_QP)		|
1255		(1ull << IB_USER_VERBS_CMD_QUERY_QP)		|
1256		(1ull << IB_USER_VERBS_CMD_MODIFY_QP)		|
1257		(1ull << IB_USER_VERBS_CMD_DESTROY_QP)		|
1258		(1ull << IB_USER_VERBS_CMD_ATTACH_MCAST)	|
1259		(1ull << IB_USER_VERBS_CMD_DETACH_MCAST);
1260	dev->ib_dev.node_type            = RDMA_NODE_IB_CA;
1261	dev->ib_dev.phys_port_cnt        = dev->limits.num_ports;
1262	dev->ib_dev.num_comp_vectors     = 1;
1263	dev->ib_dev.dev.parent           = &dev->pdev->dev;
1264
1265	if (dev->mthca_flags & MTHCA_FLAG_SRQ) {
1266		dev->ib_dev.uverbs_cmd_mask	|=
1267			(1ull << IB_USER_VERBS_CMD_CREATE_SRQ)		|
1268			(1ull << IB_USER_VERBS_CMD_MODIFY_SRQ)		|
1269			(1ull << IB_USER_VERBS_CMD_QUERY_SRQ)		|
1270			(1ull << IB_USER_VERBS_CMD_DESTROY_SRQ);
1271
1272		if (mthca_is_memfree(dev))
1273			ib_set_device_ops(&dev->ib_dev,
1274					  &mthca_dev_arbel_srq_ops);
1275		else
1276			ib_set_device_ops(&dev->ib_dev,
1277					  &mthca_dev_tavor_srq_ops);
1278	}
1279
1280	if (dev->mthca_flags & MTHCA_FLAG_FMR) {
1281		if (mthca_is_memfree(dev))
1282			ib_set_device_ops(&dev->ib_dev,
1283					  &mthca_dev_arbel_fmr_ops);
1284		else
1285			ib_set_device_ops(&dev->ib_dev,
1286					  &mthca_dev_tavor_fmr_ops);
1287	}
1288
1289	ib_set_device_ops(&dev->ib_dev, &mthca_dev_ops);
1290
1291	if (mthca_is_memfree(dev))
1292		ib_set_device_ops(&dev->ib_dev, &mthca_dev_arbel_ops);
1293	else
1294		ib_set_device_ops(&dev->ib_dev, &mthca_dev_tavor_ops);
1295
1296	mutex_init(&dev->cap_mask_mutex);
1297
1298	rdma_set_device_sysfs_group(&dev->ib_dev, &mthca_attr_group);
1299	ret = ib_register_device(&dev->ib_dev, "mthca%d");
1300	if (ret)
1301		return ret;
1302
1303	mthca_start_catas_poll(dev);
1304
1305	return 0;
1306}
1307
1308void mthca_unregister_device(struct mthca_dev *dev)
1309{
1310	mthca_stop_catas_poll(dev);
1311	ib_unregister_device(&dev->ib_dev);
1312}