Linux Audio

Check our new training course

Loading...
Note: File does not exist in v3.1.
   1// SPDX-License-Identifier: GPL-2.0-only
   2/*
   3 * QLogic iSCSI Offload Driver
   4 * Copyright (c) 2016 Cavium Inc.
   5 */
   6
   7#include <linux/blkdev.h>
   8#include <scsi/scsi_tcq.h>
   9#include <linux/delay.h>
  10
  11#include "qedi.h"
  12#include "qedi_iscsi.h"
  13#include "qedi_gbl.h"
  14#include "qedi_fw_iscsi.h"
  15#include "qedi_fw_scsi.h"
  16
  17static int send_iscsi_tmf(struct qedi_conn *qedi_conn,
  18			  struct iscsi_task *mtask, struct iscsi_task *ctask);
  19
  20void qedi_iscsi_unmap_sg_list(struct qedi_cmd *cmd)
  21{
  22	struct scsi_cmnd *sc = cmd->scsi_cmd;
  23
  24	if (cmd->io_tbl.sge_valid && sc) {
  25		cmd->io_tbl.sge_valid = 0;
  26		scsi_dma_unmap(sc);
  27	}
  28}
  29
  30static void qedi_process_logout_resp(struct qedi_ctx *qedi,
  31				     union iscsi_cqe *cqe,
  32				     struct iscsi_task *task,
  33				     struct qedi_conn *qedi_conn)
  34{
  35	struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
  36	struct iscsi_logout_rsp *resp_hdr;
  37	struct iscsi_session *session = conn->session;
  38	struct iscsi_logout_response_hdr *cqe_logout_response;
  39	struct qedi_cmd *cmd;
  40
  41	cmd = (struct qedi_cmd *)task->dd_data;
  42	cqe_logout_response = &cqe->cqe_common.iscsi_hdr.logout_response;
  43	spin_lock(&session->back_lock);
  44	resp_hdr = (struct iscsi_logout_rsp *)&qedi_conn->gen_pdu.resp_hdr;
  45	memset(resp_hdr, 0, sizeof(struct iscsi_hdr));
  46	resp_hdr->opcode = cqe_logout_response->opcode;
  47	resp_hdr->flags = cqe_logout_response->flags;
  48	resp_hdr->hlength = 0;
  49
  50	resp_hdr->itt = build_itt(cqe->cqe_solicited.itid, conn->session->age);
  51	resp_hdr->statsn = cpu_to_be32(cqe_logout_response->stat_sn);
  52	resp_hdr->exp_cmdsn = cpu_to_be32(cqe_logout_response->exp_cmd_sn);
  53	resp_hdr->max_cmdsn = cpu_to_be32(cqe_logout_response->max_cmd_sn);
  54
  55	resp_hdr->t2wait = cpu_to_be32(cqe_logout_response->time_2_wait);
  56	resp_hdr->t2retain = cpu_to_be32(cqe_logout_response->time_2_retain);
  57
  58	QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_TID,
  59		  "Freeing tid=0x%x for cid=0x%x\n",
  60		  cmd->task_id, qedi_conn->iscsi_conn_id);
  61
  62	spin_lock(&qedi_conn->list_lock);
  63	if (likely(cmd->io_cmd_in_list)) {
  64		cmd->io_cmd_in_list = false;
  65		list_del_init(&cmd->io_cmd);
  66		qedi_conn->active_cmd_count--;
  67	} else {
  68		QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_INFO,
  69			  "Active cmd list node already deleted, tid=0x%x, cid=0x%x, io_cmd_node=%p\n",
  70			  cmd->task_id, qedi_conn->iscsi_conn_id,
  71			  &cmd->io_cmd);
  72	}
  73	spin_unlock(&qedi_conn->list_lock);
  74
  75	cmd->state = RESPONSE_RECEIVED;
  76	__iscsi_complete_pdu(conn, (struct iscsi_hdr *)resp_hdr, NULL, 0);
  77
  78	spin_unlock(&session->back_lock);
  79}
  80
  81static void qedi_process_text_resp(struct qedi_ctx *qedi,
  82				   union iscsi_cqe *cqe,
  83				   struct iscsi_task *task,
  84				   struct qedi_conn *qedi_conn)
  85{
  86	struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
  87	struct iscsi_session *session = conn->session;
  88	struct iscsi_task_context *task_ctx;
  89	struct iscsi_text_rsp *resp_hdr_ptr;
  90	struct iscsi_text_response_hdr *cqe_text_response;
  91	struct qedi_cmd *cmd;
  92	int pld_len;
  93
  94	cmd = (struct qedi_cmd *)task->dd_data;
  95	task_ctx = qedi_get_task_mem(&qedi->tasks, cmd->task_id);
  96
  97	cqe_text_response = &cqe->cqe_common.iscsi_hdr.text_response;
  98	spin_lock(&session->back_lock);
  99	resp_hdr_ptr =  (struct iscsi_text_rsp *)&qedi_conn->gen_pdu.resp_hdr;
 100	memset(resp_hdr_ptr, 0, sizeof(struct iscsi_hdr));
 101	resp_hdr_ptr->opcode = cqe_text_response->opcode;
 102	resp_hdr_ptr->flags = cqe_text_response->flags;
 103	resp_hdr_ptr->hlength = 0;
 104
 105	hton24(resp_hdr_ptr->dlength,
 106	       (cqe_text_response->hdr_second_dword &
 107		ISCSI_TEXT_RESPONSE_HDR_DATA_SEG_LEN_MASK));
 108
 109	resp_hdr_ptr->itt = build_itt(cqe->cqe_solicited.itid,
 110				      conn->session->age);
 111	resp_hdr_ptr->ttt = cqe_text_response->ttt;
 112	resp_hdr_ptr->statsn = cpu_to_be32(cqe_text_response->stat_sn);
 113	resp_hdr_ptr->exp_cmdsn = cpu_to_be32(cqe_text_response->exp_cmd_sn);
 114	resp_hdr_ptr->max_cmdsn = cpu_to_be32(cqe_text_response->max_cmd_sn);
 115
 116	pld_len = cqe_text_response->hdr_second_dword &
 117		  ISCSI_TEXT_RESPONSE_HDR_DATA_SEG_LEN_MASK;
 118	qedi_conn->gen_pdu.resp_wr_ptr = qedi_conn->gen_pdu.resp_buf + pld_len;
 119
 120	memset(task_ctx, '\0', sizeof(*task_ctx));
 121
 122	QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_TID,
 123		  "Freeing tid=0x%x for cid=0x%x\n",
 124		  cmd->task_id, qedi_conn->iscsi_conn_id);
 125
 126	spin_lock(&qedi_conn->list_lock);
 127	if (likely(cmd->io_cmd_in_list)) {
 128		cmd->io_cmd_in_list = false;
 129		list_del_init(&cmd->io_cmd);
 130		qedi_conn->active_cmd_count--;
 131	} else {
 132		QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_INFO,
 133			  "Active cmd list node already deleted, tid=0x%x, cid=0x%x, io_cmd_node=%p\n",
 134			  cmd->task_id, qedi_conn->iscsi_conn_id,
 135			  &cmd->io_cmd);
 136	}
 137	spin_unlock(&qedi_conn->list_lock);
 138
 139	cmd->state = RESPONSE_RECEIVED;
 140
 141	__iscsi_complete_pdu(conn, (struct iscsi_hdr *)resp_hdr_ptr,
 142			     qedi_conn->gen_pdu.resp_buf,
 143			     (qedi_conn->gen_pdu.resp_wr_ptr -
 144			      qedi_conn->gen_pdu.resp_buf));
 145	spin_unlock(&session->back_lock);
 146}
 147
 148static void qedi_tmf_resp_work(struct work_struct *work)
 149{
 150	struct qedi_cmd *qedi_cmd =
 151				container_of(work, struct qedi_cmd, tmf_work);
 152	struct qedi_conn *qedi_conn = qedi_cmd->conn;
 153	struct qedi_ctx *qedi = qedi_conn->qedi;
 154	struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
 155	struct iscsi_session *session = conn->session;
 156	struct iscsi_tm_rsp *resp_hdr_ptr;
 157	int rval = 0;
 158
 159	resp_hdr_ptr =  (struct iscsi_tm_rsp *)qedi_cmd->tmf_resp_buf;
 160
 161	rval = qedi_cleanup_all_io(qedi, qedi_conn, qedi_cmd->task, true);
 162	if (rval)
 163		goto exit_tmf_resp;
 164
 165	spin_lock(&session->back_lock);
 166	__iscsi_complete_pdu(conn, (struct iscsi_hdr *)resp_hdr_ptr, NULL, 0);
 167	spin_unlock(&session->back_lock);
 168
 169exit_tmf_resp:
 170	kfree(resp_hdr_ptr);
 171
 172	spin_lock(&qedi_conn->tmf_work_lock);
 173	qedi_conn->fw_cleanup_works--;
 174	spin_unlock(&qedi_conn->tmf_work_lock);
 175}
 176
 177static void qedi_process_tmf_resp(struct qedi_ctx *qedi,
 178				  union iscsi_cqe *cqe,
 179				  struct iscsi_task *task,
 180				  struct qedi_conn *qedi_conn)
 181
 182{
 183	struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
 184	struct iscsi_session *session = conn->session;
 185	struct iscsi_tmf_response_hdr *cqe_tmp_response;
 186	struct iscsi_tm_rsp *resp_hdr_ptr;
 187	struct iscsi_tm *tmf_hdr;
 188	struct qedi_cmd *qedi_cmd = NULL;
 189
 190	cqe_tmp_response = &cqe->cqe_common.iscsi_hdr.tmf_response;
 191
 192	qedi_cmd = task->dd_data;
 193	qedi_cmd->tmf_resp_buf = kzalloc(sizeof(*resp_hdr_ptr), GFP_ATOMIC);
 194	if (!qedi_cmd->tmf_resp_buf) {
 195		QEDI_ERR(&qedi->dbg_ctx,
 196			 "Failed to allocate resp buf, cid=0x%x\n",
 197			  qedi_conn->iscsi_conn_id);
 198		return;
 199	}
 200
 201	spin_lock(&session->back_lock);
 202	resp_hdr_ptr =  (struct iscsi_tm_rsp *)qedi_cmd->tmf_resp_buf;
 203	memset(resp_hdr_ptr, 0, sizeof(struct iscsi_tm_rsp));
 204
 205	/* Fill up the header */
 206	resp_hdr_ptr->opcode = cqe_tmp_response->opcode;
 207	resp_hdr_ptr->flags = cqe_tmp_response->hdr_flags;
 208	resp_hdr_ptr->response = cqe_tmp_response->hdr_response;
 209	resp_hdr_ptr->hlength = 0;
 210
 211	hton24(resp_hdr_ptr->dlength,
 212	       (cqe_tmp_response->hdr_second_dword &
 213		ISCSI_TMF_RESPONSE_HDR_DATA_SEG_LEN_MASK));
 214	resp_hdr_ptr->itt = build_itt(cqe->cqe_solicited.itid,
 215				      conn->session->age);
 216	resp_hdr_ptr->statsn = cpu_to_be32(cqe_tmp_response->stat_sn);
 217	resp_hdr_ptr->exp_cmdsn  = cpu_to_be32(cqe_tmp_response->exp_cmd_sn);
 218	resp_hdr_ptr->max_cmdsn = cpu_to_be32(cqe_tmp_response->max_cmd_sn);
 219
 220	tmf_hdr = (struct iscsi_tm *)qedi_cmd->task->hdr;
 221
 222	spin_lock(&qedi_conn->list_lock);
 223	if (likely(qedi_cmd->io_cmd_in_list)) {
 224		qedi_cmd->io_cmd_in_list = false;
 225		list_del_init(&qedi_cmd->io_cmd);
 226		qedi_conn->active_cmd_count--;
 227	}
 228	spin_unlock(&qedi_conn->list_lock);
 229
 230	spin_lock(&qedi_conn->tmf_work_lock);
 231	switch (tmf_hdr->flags & ISCSI_FLAG_TM_FUNC_MASK) {
 232	case ISCSI_TM_FUNC_LOGICAL_UNIT_RESET:
 233	case ISCSI_TM_FUNC_TARGET_WARM_RESET:
 234	case ISCSI_TM_FUNC_TARGET_COLD_RESET:
 235		if (qedi_conn->ep_disconnect_starting) {
 236			/* Session is down so ep_disconnect will clean up */
 237			spin_unlock(&qedi_conn->tmf_work_lock);
 238			goto unblock_sess;
 239		}
 240
 241		qedi_conn->fw_cleanup_works++;
 242		spin_unlock(&qedi_conn->tmf_work_lock);
 243
 244		INIT_WORK(&qedi_cmd->tmf_work, qedi_tmf_resp_work);
 245		queue_work(qedi->tmf_thread, &qedi_cmd->tmf_work);
 246		goto unblock_sess;
 247	}
 248	spin_unlock(&qedi_conn->tmf_work_lock);
 249
 250	__iscsi_complete_pdu(conn, (struct iscsi_hdr *)resp_hdr_ptr, NULL, 0);
 251	kfree(resp_hdr_ptr);
 252
 253unblock_sess:
 254	spin_unlock(&session->back_lock);
 255}
 256
 257static void qedi_process_login_resp(struct qedi_ctx *qedi,
 258				    union iscsi_cqe *cqe,
 259				    struct iscsi_task *task,
 260				    struct qedi_conn *qedi_conn)
 261{
 262	struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
 263	struct iscsi_session *session = conn->session;
 264	struct iscsi_task_context *task_ctx;
 265	struct iscsi_login_rsp *resp_hdr_ptr;
 266	struct iscsi_login_response_hdr *cqe_login_response;
 267	struct qedi_cmd *cmd;
 268	int pld_len;
 269
 270	cmd = (struct qedi_cmd *)task->dd_data;
 271
 272	cqe_login_response = &cqe->cqe_common.iscsi_hdr.login_response;
 273	task_ctx = qedi_get_task_mem(&qedi->tasks, cmd->task_id);
 274
 275	spin_lock(&session->back_lock);
 276	resp_hdr_ptr =  (struct iscsi_login_rsp *)&qedi_conn->gen_pdu.resp_hdr;
 277	memset(resp_hdr_ptr, 0, sizeof(struct iscsi_login_rsp));
 278	resp_hdr_ptr->opcode = cqe_login_response->opcode;
 279	resp_hdr_ptr->flags = cqe_login_response->flags_attr;
 280	resp_hdr_ptr->hlength = 0;
 281
 282	hton24(resp_hdr_ptr->dlength,
 283	       (cqe_login_response->hdr_second_dword &
 284		ISCSI_LOGIN_RESPONSE_HDR_DATA_SEG_LEN_MASK));
 285	resp_hdr_ptr->itt = build_itt(cqe->cqe_solicited.itid,
 286				      conn->session->age);
 287	resp_hdr_ptr->tsih = cqe_login_response->tsih;
 288	resp_hdr_ptr->statsn = cpu_to_be32(cqe_login_response->stat_sn);
 289	resp_hdr_ptr->exp_cmdsn = cpu_to_be32(cqe_login_response->exp_cmd_sn);
 290	resp_hdr_ptr->max_cmdsn = cpu_to_be32(cqe_login_response->max_cmd_sn);
 291	resp_hdr_ptr->status_class = cqe_login_response->status_class;
 292	resp_hdr_ptr->status_detail = cqe_login_response->status_detail;
 293	pld_len = cqe_login_response->hdr_second_dword &
 294		  ISCSI_LOGIN_RESPONSE_HDR_DATA_SEG_LEN_MASK;
 295	qedi_conn->gen_pdu.resp_wr_ptr = qedi_conn->gen_pdu.resp_buf + pld_len;
 296
 297	spin_lock(&qedi_conn->list_lock);
 298	if (likely(cmd->io_cmd_in_list)) {
 299		cmd->io_cmd_in_list = false;
 300		list_del_init(&cmd->io_cmd);
 301		qedi_conn->active_cmd_count--;
 302	}
 303	spin_unlock(&qedi_conn->list_lock);
 304
 305	memset(task_ctx, '\0', sizeof(*task_ctx));
 306
 307	__iscsi_complete_pdu(conn, (struct iscsi_hdr *)resp_hdr_ptr,
 308			     qedi_conn->gen_pdu.resp_buf,
 309			     (qedi_conn->gen_pdu.resp_wr_ptr -
 310			     qedi_conn->gen_pdu.resp_buf));
 311
 312	spin_unlock(&session->back_lock);
 313	QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_TID,
 314		  "Freeing tid=0x%x for cid=0x%x\n",
 315		  cmd->task_id, qedi_conn->iscsi_conn_id);
 316	cmd->state = RESPONSE_RECEIVED;
 317}
 318
 319static void qedi_get_rq_bdq_buf(struct qedi_ctx *qedi,
 320				struct iscsi_cqe_unsolicited *cqe,
 321				char *ptr, int len)
 322{
 323	u16 idx = 0;
 324
 325	QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
 326		  "pld_len [%d], bdq_prod_idx [%d], idx [%d]\n",
 327		  len, qedi->bdq_prod_idx,
 328		  (qedi->bdq_prod_idx % qedi->rq_num_entries));
 329
 330	/* Obtain buffer address from rqe_opaque */
 331	idx = cqe->rqe_opaque;
 332	if (idx > (QEDI_BDQ_NUM - 1)) {
 333		QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
 334			  "wrong idx %d returned by FW, dropping the unsolicited pkt\n",
 335			  idx);
 336		return;
 337	}
 338
 339	QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
 340		  "rqe_opaque [0x%p], idx [%d]\n", cqe->rqe_opaque, idx);
 341
 342	QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
 343		  "unsol_cqe_type = %d\n", cqe->unsol_cqe_type);
 344	switch (cqe->unsol_cqe_type) {
 345	case ISCSI_CQE_UNSOLICITED_SINGLE:
 346	case ISCSI_CQE_UNSOLICITED_FIRST:
 347		if (len)
 348			memcpy(ptr, (void *)qedi->bdq[idx].buf_addr, len);
 349		break;
 350	case ISCSI_CQE_UNSOLICITED_MIDDLE:
 351	case ISCSI_CQE_UNSOLICITED_LAST:
 352		break;
 353	default:
 354		break;
 355	}
 356}
 357
 358static void qedi_put_rq_bdq_buf(struct qedi_ctx *qedi,
 359				struct iscsi_cqe_unsolicited *cqe,
 360				int count)
 361{
 362	u16 idx = 0;
 363	struct scsi_bd *pbl;
 364
 365	/* Obtain buffer address from rqe_opaque */
 366	idx = cqe->rqe_opaque;
 367	if (idx > (QEDI_BDQ_NUM - 1)) {
 368		QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
 369			  "wrong idx %d returned by FW, dropping the unsolicited pkt\n",
 370			  idx);
 371		return;
 372	}
 373
 374	pbl = (struct scsi_bd *)qedi->bdq_pbl;
 375	pbl += (qedi->bdq_prod_idx % qedi->rq_num_entries);
 376	pbl->address.hi = cpu_to_le32(QEDI_U64_HI(qedi->bdq[idx].buf_dma));
 377	pbl->address.lo = cpu_to_le32(QEDI_U64_LO(qedi->bdq[idx].buf_dma));
 378	QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
 379		  "pbl [0x%p] pbl->address hi [0x%llx] lo [0x%llx] idx [%d]\n",
 380		  pbl, pbl->address.hi, pbl->address.lo, idx);
 381	pbl->opaque.iscsi_opaque.reserved_zero[0] = 0;
 382	pbl->opaque.iscsi_opaque.reserved_zero[1] = 0;
 383	pbl->opaque.iscsi_opaque.reserved_zero[2] = 0;
 384	pbl->opaque.iscsi_opaque.opaque = cpu_to_le32(idx);
 385
 386	/* Increment producer to let f/w know we've handled the frame */
 387	qedi->bdq_prod_idx += count;
 388
 389	writew(qedi->bdq_prod_idx, qedi->bdq_primary_prod);
 390	readw(qedi->bdq_primary_prod);
 391
 392	writew(qedi->bdq_prod_idx, qedi->bdq_secondary_prod);
 393	readw(qedi->bdq_secondary_prod);
 394}
 395
 396static void qedi_unsol_pdu_adjust_bdq(struct qedi_ctx *qedi,
 397				      struct iscsi_cqe_unsolicited *cqe,
 398				      u32 pdu_len, u32 num_bdqs,
 399				      char *bdq_data)
 400{
 401	QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
 402		  "num_bdqs [%d]\n", num_bdqs);
 403
 404	qedi_get_rq_bdq_buf(qedi, cqe, bdq_data, pdu_len);
 405	qedi_put_rq_bdq_buf(qedi, cqe, (num_bdqs + 1));
 406}
 407
 408static int qedi_process_nopin_mesg(struct qedi_ctx *qedi,
 409				   union iscsi_cqe *cqe,
 410				   struct iscsi_task *task,
 411				   struct qedi_conn *qedi_conn, u16 que_idx)
 412{
 413	struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
 414	struct iscsi_session *session = conn->session;
 415	struct iscsi_nop_in_hdr *cqe_nop_in;
 416	struct iscsi_nopin *hdr;
 417	struct qedi_cmd *cmd;
 418	int tgt_async_nop = 0;
 419	u32 lun[2];
 420	u32 pdu_len, num_bdqs;
 421	char bdq_data[QEDI_BDQ_BUF_SIZE];
 422	unsigned long flags;
 423
 424	spin_lock_bh(&session->back_lock);
 425	cqe_nop_in = &cqe->cqe_common.iscsi_hdr.nop_in;
 426
 427	pdu_len = cqe_nop_in->hdr_second_dword &
 428		  ISCSI_NOP_IN_HDR_DATA_SEG_LEN_MASK;
 429	num_bdqs = pdu_len / QEDI_BDQ_BUF_SIZE;
 430
 431	hdr = (struct iscsi_nopin *)&qedi_conn->gen_pdu.resp_hdr;
 432	memset(hdr, 0, sizeof(struct iscsi_hdr));
 433	hdr->opcode = cqe_nop_in->opcode;
 434	hdr->max_cmdsn = cpu_to_be32(cqe_nop_in->max_cmd_sn);
 435	hdr->exp_cmdsn = cpu_to_be32(cqe_nop_in->exp_cmd_sn);
 436	hdr->statsn = cpu_to_be32(cqe_nop_in->stat_sn);
 437	hdr->ttt = cpu_to_be32(cqe_nop_in->ttt);
 438
 439	if (cqe->cqe_common.cqe_type == ISCSI_CQE_TYPE_UNSOLICITED) {
 440		spin_lock_irqsave(&qedi->hba_lock, flags);
 441		qedi_unsol_pdu_adjust_bdq(qedi, &cqe->cqe_unsolicited,
 442					  pdu_len, num_bdqs, bdq_data);
 443		hdr->itt = RESERVED_ITT;
 444		tgt_async_nop = 1;
 445		spin_unlock_irqrestore(&qedi->hba_lock, flags);
 446		goto done;
 447	}
 448
 449	/* Response to one of our nop-outs */
 450	if (task) {
 451		cmd = task->dd_data;
 452		hdr->flags = ISCSI_FLAG_CMD_FINAL;
 453		hdr->itt = build_itt(cqe->cqe_solicited.itid,
 454				     conn->session->age);
 455		lun[0] = 0xffffffff;
 456		lun[1] = 0xffffffff;
 457		memcpy(&hdr->lun, lun, sizeof(struct scsi_lun));
 458		QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_TID,
 459			  "Freeing tid=0x%x for cid=0x%x\n",
 460			  cmd->task_id, qedi_conn->iscsi_conn_id);
 461		cmd->state = RESPONSE_RECEIVED;
 462		spin_lock(&qedi_conn->list_lock);
 463		if (likely(cmd->io_cmd_in_list)) {
 464			cmd->io_cmd_in_list = false;
 465			list_del_init(&cmd->io_cmd);
 466			qedi_conn->active_cmd_count--;
 467		}
 468
 469		spin_unlock(&qedi_conn->list_lock);
 470	}
 471
 472done:
 473	__iscsi_complete_pdu(conn, (struct iscsi_hdr *)hdr, bdq_data, pdu_len);
 474
 475	spin_unlock_bh(&session->back_lock);
 476	return tgt_async_nop;
 477}
 478
 479static void qedi_process_async_mesg(struct qedi_ctx *qedi,
 480				    union iscsi_cqe *cqe,
 481				    struct iscsi_task *task,
 482				    struct qedi_conn *qedi_conn,
 483				    u16 que_idx)
 484{
 485	struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
 486	struct iscsi_session *session = conn->session;
 487	struct iscsi_async_msg_hdr *cqe_async_msg;
 488	struct iscsi_async *resp_hdr;
 489	u32 lun[2];
 490	u32 pdu_len, num_bdqs;
 491	char bdq_data[QEDI_BDQ_BUF_SIZE];
 492	unsigned long flags;
 493
 494	spin_lock_bh(&session->back_lock);
 495
 496	cqe_async_msg = &cqe->cqe_common.iscsi_hdr.async_msg;
 497	pdu_len = cqe_async_msg->hdr_second_dword &
 498		ISCSI_ASYNC_MSG_HDR_DATA_SEG_LEN_MASK;
 499	num_bdqs = pdu_len / QEDI_BDQ_BUF_SIZE;
 500
 501	if (cqe->cqe_common.cqe_type == ISCSI_CQE_TYPE_UNSOLICITED) {
 502		spin_lock_irqsave(&qedi->hba_lock, flags);
 503		qedi_unsol_pdu_adjust_bdq(qedi, &cqe->cqe_unsolicited,
 504					  pdu_len, num_bdqs, bdq_data);
 505		spin_unlock_irqrestore(&qedi->hba_lock, flags);
 506	}
 507
 508	resp_hdr = (struct iscsi_async *)&qedi_conn->gen_pdu.resp_hdr;
 509	memset(resp_hdr, 0, sizeof(struct iscsi_hdr));
 510	resp_hdr->opcode = cqe_async_msg->opcode;
 511	resp_hdr->flags = 0x80;
 512
 513	lun[0] = cpu_to_be32(cqe_async_msg->lun.lo);
 514	lun[1] = cpu_to_be32(cqe_async_msg->lun.hi);
 515	memcpy(&resp_hdr->lun, lun, sizeof(struct scsi_lun));
 516	resp_hdr->exp_cmdsn = cpu_to_be32(cqe_async_msg->exp_cmd_sn);
 517	resp_hdr->max_cmdsn = cpu_to_be32(cqe_async_msg->max_cmd_sn);
 518	resp_hdr->statsn = cpu_to_be32(cqe_async_msg->stat_sn);
 519
 520	resp_hdr->async_event = cqe_async_msg->async_event;
 521	resp_hdr->async_vcode = cqe_async_msg->async_vcode;
 522
 523	resp_hdr->param1 = cpu_to_be16(cqe_async_msg->param1_rsrv);
 524	resp_hdr->param2 = cpu_to_be16(cqe_async_msg->param2_rsrv);
 525	resp_hdr->param3 = cpu_to_be16(cqe_async_msg->param3_rsrv);
 526
 527	__iscsi_complete_pdu(conn, (struct iscsi_hdr *)resp_hdr, bdq_data,
 528			     pdu_len);
 529
 530	spin_unlock_bh(&session->back_lock);
 531}
 532
 533static void qedi_process_reject_mesg(struct qedi_ctx *qedi,
 534				     union iscsi_cqe *cqe,
 535				     struct iscsi_task *task,
 536				     struct qedi_conn *qedi_conn,
 537				     uint16_t que_idx)
 538{
 539	struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
 540	struct iscsi_session *session = conn->session;
 541	struct iscsi_reject_hdr *cqe_reject;
 542	struct iscsi_reject *hdr;
 543	u32 pld_len, num_bdqs;
 544	unsigned long flags;
 545
 546	spin_lock_bh(&session->back_lock);
 547	cqe_reject = &cqe->cqe_common.iscsi_hdr.reject;
 548	pld_len = cqe_reject->hdr_second_dword &
 549		  ISCSI_REJECT_HDR_DATA_SEG_LEN_MASK;
 550	num_bdqs = pld_len / QEDI_BDQ_BUF_SIZE;
 551
 552	if (cqe->cqe_common.cqe_type == ISCSI_CQE_TYPE_UNSOLICITED) {
 553		spin_lock_irqsave(&qedi->hba_lock, flags);
 554		qedi_unsol_pdu_adjust_bdq(qedi, &cqe->cqe_unsolicited,
 555					  pld_len, num_bdqs, conn->data);
 556		spin_unlock_irqrestore(&qedi->hba_lock, flags);
 557	}
 558	hdr = (struct iscsi_reject *)&qedi_conn->gen_pdu.resp_hdr;
 559	memset(hdr, 0, sizeof(struct iscsi_hdr));
 560	hdr->opcode = cqe_reject->opcode;
 561	hdr->reason = cqe_reject->hdr_reason;
 562	hdr->flags = cqe_reject->hdr_flags;
 563	hton24(hdr->dlength, (cqe_reject->hdr_second_dword &
 564			      ISCSI_REJECT_HDR_DATA_SEG_LEN_MASK));
 565	hdr->max_cmdsn = cpu_to_be32(cqe_reject->max_cmd_sn);
 566	hdr->exp_cmdsn = cpu_to_be32(cqe_reject->exp_cmd_sn);
 567	hdr->statsn = cpu_to_be32(cqe_reject->stat_sn);
 568	hdr->ffffffff = cpu_to_be32(0xffffffff);
 569
 570	__iscsi_complete_pdu(conn, (struct iscsi_hdr *)hdr,
 571			     conn->data, pld_len);
 572	spin_unlock_bh(&session->back_lock);
 573}
 574
 575static void qedi_scsi_completion(struct qedi_ctx *qedi,
 576				 union iscsi_cqe *cqe,
 577				 struct iscsi_task *task,
 578				 struct iscsi_conn *conn)
 579{
 580	struct scsi_cmnd *sc_cmd;
 581	struct qedi_cmd *cmd = task->dd_data;
 582	struct iscsi_session *session = conn->session;
 583	struct iscsi_scsi_rsp *hdr;
 584	struct iscsi_data_in_hdr *cqe_data_in;
 585	int datalen = 0;
 586	struct qedi_conn *qedi_conn;
 587	u32 iscsi_cid;
 588	u8 cqe_err_bits = 0;
 589
 590	iscsi_cid  = cqe->cqe_common.conn_id;
 591	qedi_conn = qedi->cid_que.conn_cid_tbl[iscsi_cid];
 592
 593	cqe_data_in = &cqe->cqe_common.iscsi_hdr.data_in;
 594	cqe_err_bits =
 595		cqe->cqe_common.error_bitmap.error_bits.cqe_error_status_bits;
 596
 597	spin_lock_bh(&session->back_lock);
 598	/* get the scsi command */
 599	sc_cmd = cmd->scsi_cmd;
 600
 601	if (!sc_cmd) {
 602		QEDI_WARN(&qedi->dbg_ctx, "sc_cmd is NULL!\n");
 603		goto error;
 604	}
 605
 606	if (!iscsi_cmd(sc_cmd)->task) {
 607		QEDI_WARN(&qedi->dbg_ctx,
 608			  "NULL task pointer, returned in another context.\n");
 609		goto error;
 610	}
 611
 612	if (!scsi_cmd_to_rq(sc_cmd)->q) {
 613		QEDI_WARN(&qedi->dbg_ctx,
 614			  "request->q is NULL so request is not valid, sc_cmd=%p.\n",
 615			  sc_cmd);
 616		goto error;
 617	}
 618
 619	qedi_iscsi_unmap_sg_list(cmd);
 620
 621	hdr = (struct iscsi_scsi_rsp *)task->hdr;
 622	hdr->opcode = cqe_data_in->opcode;
 623	hdr->max_cmdsn = cpu_to_be32(cqe_data_in->max_cmd_sn);
 624	hdr->exp_cmdsn = cpu_to_be32(cqe_data_in->exp_cmd_sn);
 625	hdr->itt = build_itt(cqe->cqe_solicited.itid, conn->session->age);
 626	hdr->response = cqe_data_in->reserved1;
 627	hdr->cmd_status = cqe_data_in->status_rsvd;
 628	hdr->flags = cqe_data_in->flags;
 629	hdr->residual_count = cpu_to_be32(cqe_data_in->residual_count);
 630
 631	if (hdr->cmd_status == SAM_STAT_CHECK_CONDITION) {
 632		datalen = cqe_data_in->reserved2 &
 633			  ISCSI_COMMON_HDR_DATA_SEG_LEN_MASK;
 634		memcpy((char *)conn->data, (char *)cmd->sense_buffer, datalen);
 635	}
 636
 637	/* If f/w reports data underrun err then set residual to IO transfer
 638	 * length, set Underrun flag and clear Overrun flag explicitly
 639	 */
 640	if (unlikely(cqe_err_bits &&
 641		     GET_FIELD(cqe_err_bits, CQE_ERROR_BITMAP_UNDER_RUN_ERR))) {
 642		QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_INFO,
 643			  "Under flow itt=0x%x proto flags=0x%x tid=0x%x cid 0x%x fw resid 0x%x sc dlen 0x%x\n",
 644			  hdr->itt, cqe_data_in->flags, cmd->task_id,
 645			  qedi_conn->iscsi_conn_id, hdr->residual_count,
 646			  scsi_bufflen(sc_cmd));
 647		hdr->residual_count = cpu_to_be32(scsi_bufflen(sc_cmd));
 648		hdr->flags |= ISCSI_FLAG_CMD_UNDERFLOW;
 649		hdr->flags &= (~ISCSI_FLAG_CMD_OVERFLOW);
 650	}
 651
 652	spin_lock(&qedi_conn->list_lock);
 653	if (likely(cmd->io_cmd_in_list)) {
 654		cmd->io_cmd_in_list = false;
 655		list_del_init(&cmd->io_cmd);
 656		qedi_conn->active_cmd_count--;
 657	}
 658	spin_unlock(&qedi_conn->list_lock);
 659
 660	QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_TID,
 661		  "Freeing tid=0x%x for cid=0x%x\n",
 662		  cmd->task_id, qedi_conn->iscsi_conn_id);
 663	cmd->state = RESPONSE_RECEIVED;
 664	if (qedi_io_tracing)
 665		qedi_trace_io(qedi, task, cmd->task_id, QEDI_IO_TRACE_RSP);
 666
 667	__iscsi_complete_pdu(conn, (struct iscsi_hdr *)hdr,
 668			     conn->data, datalen);
 669error:
 670	spin_unlock_bh(&session->back_lock);
 671}
 672
 673static void qedi_mtask_completion(struct qedi_ctx *qedi,
 674				  union iscsi_cqe *cqe,
 675				  struct iscsi_task *task,
 676				  struct qedi_conn *conn, uint16_t que_idx)
 677{
 678	struct iscsi_conn *iscsi_conn;
 679	u32 hdr_opcode;
 680
 681	hdr_opcode = cqe->cqe_common.iscsi_hdr.common.hdr_first_byte;
 682	iscsi_conn = conn->cls_conn->dd_data;
 683
 684	switch (hdr_opcode) {
 685	case ISCSI_OPCODE_SCSI_RESPONSE:
 686	case ISCSI_OPCODE_DATA_IN:
 687		qedi_scsi_completion(qedi, cqe, task, iscsi_conn);
 688		break;
 689	case ISCSI_OPCODE_LOGIN_RESPONSE:
 690		qedi_process_login_resp(qedi, cqe, task, conn);
 691		break;
 692	case ISCSI_OPCODE_TMF_RESPONSE:
 693		qedi_process_tmf_resp(qedi, cqe, task, conn);
 694		break;
 695	case ISCSI_OPCODE_TEXT_RESPONSE:
 696		qedi_process_text_resp(qedi, cqe, task, conn);
 697		break;
 698	case ISCSI_OPCODE_LOGOUT_RESPONSE:
 699		qedi_process_logout_resp(qedi, cqe, task, conn);
 700		break;
 701	case ISCSI_OPCODE_NOP_IN:
 702		qedi_process_nopin_mesg(qedi, cqe, task, conn, que_idx);
 703		break;
 704	default:
 705		QEDI_ERR(&qedi->dbg_ctx, "unknown opcode\n");
 706	}
 707}
 708
 709static void qedi_process_nopin_local_cmpl(struct qedi_ctx *qedi,
 710					  struct iscsi_cqe_solicited *cqe,
 711					  struct iscsi_task *task,
 712					  struct qedi_conn *qedi_conn)
 713{
 714	struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
 715	struct iscsi_session *session = conn->session;
 716	struct qedi_cmd *cmd = task->dd_data;
 717
 718	QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_UNSOL,
 719		  "itid=0x%x, cmd task id=0x%x\n",
 720		  cqe->itid, cmd->task_id);
 721
 722	cmd->state = RESPONSE_RECEIVED;
 723
 724	spin_lock_bh(&session->back_lock);
 725	__iscsi_put_task(task);
 726	spin_unlock_bh(&session->back_lock);
 727}
 728
 729static void qedi_process_cmd_cleanup_resp(struct qedi_ctx *qedi,
 730					  struct iscsi_cqe_solicited *cqe,
 731					  struct iscsi_conn *conn)
 732{
 733	struct qedi_work_map *work, *work_tmp;
 734	u32 proto_itt = cqe->itid;
 735	int found = 0;
 736	struct qedi_cmd *qedi_cmd = NULL;
 737	u32 iscsi_cid;
 738	struct qedi_conn *qedi_conn;
 739	struct qedi_cmd *dbg_cmd;
 740	struct iscsi_task *mtask, *task;
 741	struct iscsi_tm *tmf_hdr = NULL;
 742
 743	iscsi_cid = cqe->conn_id;
 744	qedi_conn = qedi->cid_que.conn_cid_tbl[iscsi_cid];
 745	if (!qedi_conn) {
 746		QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_INFO,
 747			  "icid not found 0x%x\n", cqe->conn_id);
 748		return;
 749	}
 750
 751	/* Based on this itt get the corresponding qedi_cmd */
 752	spin_lock_bh(&qedi_conn->tmf_work_lock);
 753	list_for_each_entry_safe(work, work_tmp, &qedi_conn->tmf_work_list,
 754				 list) {
 755		if (work->rtid == proto_itt) {
 756			/* We found the command */
 757			qedi_cmd = work->qedi_cmd;
 758			if (!qedi_cmd->list_tmf_work) {
 759				QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
 760					  "TMF work not found, cqe->tid=0x%x, cid=0x%x\n",
 761					  proto_itt, qedi_conn->iscsi_conn_id);
 762				WARN_ON(1);
 763			}
 764			found = 1;
 765			mtask = qedi_cmd->task;
 766			task = work->ctask;
 767			tmf_hdr = (struct iscsi_tm *)mtask->hdr;
 768
 769			list_del_init(&work->list);
 770			kfree(work);
 771			qedi_cmd->list_tmf_work = NULL;
 772		}
 773	}
 774	spin_unlock_bh(&qedi_conn->tmf_work_lock);
 775
 776	if (!found)
 777		goto check_cleanup_reqs;
 778
 779	QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
 780		  "TMF work, cqe->tid=0x%x, tmf flags=0x%x, cid=0x%x\n",
 781		  proto_itt, tmf_hdr->flags, qedi_conn->iscsi_conn_id);
 782
 783	spin_lock_bh(&conn->session->back_lock);
 784	if (iscsi_task_is_completed(task)) {
 785		QEDI_NOTICE(&qedi->dbg_ctx,
 786			    "IO task completed, tmf rtt=0x%x, cid=0x%x\n",
 787			   get_itt(tmf_hdr->rtt), qedi_conn->iscsi_conn_id);
 788		goto unlock;
 789	}
 790
 791	dbg_cmd = task->dd_data;
 792
 793	QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
 794		  "Abort tmf rtt=0x%x, i/o itt=0x%x, i/o tid=0x%x, cid=0x%x\n",
 795		  get_itt(tmf_hdr->rtt), get_itt(task->itt), dbg_cmd->task_id,
 796		  qedi_conn->iscsi_conn_id);
 797
 798	spin_lock(&qedi_conn->list_lock);
 799	if (likely(dbg_cmd->io_cmd_in_list)) {
 800		dbg_cmd->io_cmd_in_list = false;
 801		list_del_init(&dbg_cmd->io_cmd);
 802		qedi_conn->active_cmd_count--;
 803	}
 804	spin_unlock(&qedi_conn->list_lock);
 805	qedi_cmd->state = CLEANUP_RECV;
 806unlock:
 807	spin_unlock_bh(&conn->session->back_lock);
 808	wake_up_interruptible(&qedi_conn->wait_queue);
 809	return;
 810
 811check_cleanup_reqs:
 812	if (atomic_inc_return(&qedi_conn->cmd_cleanup_cmpl) ==
 813	    qedi_conn->cmd_cleanup_req) {
 814		QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
 815			  "Freeing tid=0x%x for cid=0x%x\n",
 816			  cqe->itid, qedi_conn->iscsi_conn_id);
 817		wake_up(&qedi_conn->wait_queue);
 818	}
 819}
 820
 821void qedi_fp_process_cqes(struct qedi_work *work)
 822{
 823	struct qedi_ctx *qedi = work->qedi;
 824	union iscsi_cqe *cqe = &work->cqe;
 825	struct iscsi_task *task = NULL;
 826	struct iscsi_nopout *nopout_hdr;
 827	struct qedi_conn *q_conn;
 828	struct iscsi_conn *conn;
 829	struct qedi_cmd *qedi_cmd;
 830	u32 comp_type;
 831	u32 iscsi_cid;
 832	u32 hdr_opcode;
 833	u16 que_idx = work->que_idx;
 834	u8 cqe_err_bits = 0;
 835
 836	comp_type = cqe->cqe_common.cqe_type;
 837	hdr_opcode = cqe->cqe_common.iscsi_hdr.common.hdr_first_byte;
 838	cqe_err_bits =
 839		cqe->cqe_common.error_bitmap.error_bits.cqe_error_status_bits;
 840
 841	QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
 842		  "fw_cid=0x%x, cqe type=0x%x, opcode=0x%x\n",
 843		  cqe->cqe_common.conn_id, comp_type, hdr_opcode);
 844
 845	if (comp_type >= MAX_ISCSI_CQES_TYPE) {
 846		QEDI_WARN(&qedi->dbg_ctx, "Invalid CqE type\n");
 847		return;
 848	}
 849
 850	iscsi_cid  = cqe->cqe_common.conn_id;
 851	q_conn = qedi->cid_que.conn_cid_tbl[iscsi_cid];
 852	if (!q_conn) {
 853		QEDI_WARN(&qedi->dbg_ctx,
 854			  "Session no longer exists for cid=0x%x!!\n",
 855			  iscsi_cid);
 856		return;
 857	}
 858
 859	conn = q_conn->cls_conn->dd_data;
 860
 861	if (unlikely(cqe_err_bits &&
 862		     GET_FIELD(cqe_err_bits,
 863			       CQE_ERROR_BITMAP_DATA_DIGEST_ERR))) {
 864		iscsi_conn_failure(conn, ISCSI_ERR_DATA_DGST);
 865		return;
 866	}
 867
 868	switch (comp_type) {
 869	case ISCSI_CQE_TYPE_SOLICITED:
 870	case ISCSI_CQE_TYPE_SOLICITED_WITH_SENSE:
 871		qedi_cmd = container_of(work, struct qedi_cmd, cqe_work);
 872		task = qedi_cmd->task;
 873		if (!task) {
 874			QEDI_WARN(&qedi->dbg_ctx, "task is NULL\n");
 875			return;
 876		}
 877
 878		/* Process NOPIN local completion */
 879		nopout_hdr = (struct iscsi_nopout *)task->hdr;
 880		if ((nopout_hdr->itt == RESERVED_ITT) &&
 881		    (cqe->cqe_solicited.itid != (u16)RESERVED_ITT)) {
 882			qedi_process_nopin_local_cmpl(qedi, &cqe->cqe_solicited,
 883						      task, q_conn);
 884		} else {
 885			cqe->cqe_solicited.itid =
 886					       qedi_get_itt(cqe->cqe_solicited);
 887			/* Process other solicited responses */
 888			qedi_mtask_completion(qedi, cqe, task, q_conn, que_idx);
 889		}
 890		break;
 891	case ISCSI_CQE_TYPE_UNSOLICITED:
 892		switch (hdr_opcode) {
 893		case ISCSI_OPCODE_NOP_IN:
 894			qedi_process_nopin_mesg(qedi, cqe, task, q_conn,
 895						que_idx);
 896			break;
 897		case ISCSI_OPCODE_ASYNC_MSG:
 898			qedi_process_async_mesg(qedi, cqe, task, q_conn,
 899						que_idx);
 900			break;
 901		case ISCSI_OPCODE_REJECT:
 902			qedi_process_reject_mesg(qedi, cqe, task, q_conn,
 903						 que_idx);
 904			break;
 905		}
 906		goto exit_fp_process;
 907	case ISCSI_CQE_TYPE_DUMMY:
 908		QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM, "Dummy CqE\n");
 909		goto exit_fp_process;
 910	case ISCSI_CQE_TYPE_TASK_CLEANUP:
 911		QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM, "CleanUp CqE\n");
 912		qedi_process_cmd_cleanup_resp(qedi, &cqe->cqe_solicited, conn);
 913		goto exit_fp_process;
 914	default:
 915		QEDI_ERR(&qedi->dbg_ctx, "Error cqe.\n");
 916		break;
 917	}
 918
 919exit_fp_process:
 920	return;
 921}
 922
 923static void qedi_ring_doorbell(struct qedi_conn *qedi_conn)
 924{
 925	qedi_conn->ep->db_data.sq_prod = qedi_conn->ep->fw_sq_prod_idx;
 926
 927	/* wmb - Make sure fw idx is coherent */
 928	wmb();
 929	writel(*(u32 *)&qedi_conn->ep->db_data, qedi_conn->ep->p_doorbell);
 930
 931	/* Make sure fw write idx is coherent, and include both memory barriers
 932	 * as a failsafe as for some architectures the call is the same but on
 933	 * others they are two different assembly operations.
 934	 */
 935	wmb();
 936	QEDI_INFO(&qedi_conn->qedi->dbg_ctx, QEDI_LOG_MP_REQ,
 937		  "prod_idx=0x%x, fw_prod_idx=0x%x, cid=0x%x\n",
 938		  qedi_conn->ep->sq_prod_idx, qedi_conn->ep->fw_sq_prod_idx,
 939		  qedi_conn->iscsi_conn_id);
 940}
 941
 942static u16 qedi_get_wqe_idx(struct qedi_conn *qedi_conn)
 943{
 944	struct qedi_endpoint *ep;
 945	u16 rval;
 946
 947	ep = qedi_conn->ep;
 948	rval = ep->sq_prod_idx;
 949
 950	/* Increament SQ index */
 951	ep->sq_prod_idx++;
 952	ep->fw_sq_prod_idx++;
 953	if (ep->sq_prod_idx == QEDI_SQ_SIZE)
 954		ep->sq_prod_idx = 0;
 955
 956	return rval;
 957}
 958
 959int qedi_send_iscsi_login(struct qedi_conn *qedi_conn,
 960			  struct iscsi_task *task)
 961{
 962	struct iscsi_login_req_hdr login_req_pdu_header;
 963	struct scsi_sgl_task_params tx_sgl_task_params;
 964	struct scsi_sgl_task_params rx_sgl_task_params;
 965	struct iscsi_task_params task_params;
 966	struct iscsi_task_context *fw_task_ctx;
 967	struct qedi_ctx *qedi = qedi_conn->qedi;
 968	struct iscsi_login_req *login_hdr;
 969	struct scsi_sge *resp_sge = NULL;
 970	struct qedi_cmd *qedi_cmd;
 971	struct qedi_endpoint *ep;
 972	s16 tid = 0;
 973	u16 sq_idx = 0;
 974	int rval = 0;
 975
 976	resp_sge = (struct scsi_sge *)qedi_conn->gen_pdu.resp_bd_tbl;
 977	qedi_cmd = (struct qedi_cmd *)task->dd_data;
 978	ep = qedi_conn->ep;
 979	login_hdr = (struct iscsi_login_req *)task->hdr;
 980
 981	tid = qedi_get_task_idx(qedi);
 982	if (tid == -1)
 983		return -ENOMEM;
 984
 985	fw_task_ctx =
 986	     (struct iscsi_task_context *)qedi_get_task_mem(&qedi->tasks,
 987							       tid);
 988	memset(fw_task_ctx, 0, sizeof(struct iscsi_task_context));
 989
 990	qedi_cmd->task_id = tid;
 991
 992	memset(&task_params, 0, sizeof(task_params));
 993	memset(&login_req_pdu_header, 0, sizeof(login_req_pdu_header));
 994	memset(&tx_sgl_task_params, 0, sizeof(tx_sgl_task_params));
 995	memset(&rx_sgl_task_params, 0, sizeof(rx_sgl_task_params));
 996	/* Update header info */
 997	login_req_pdu_header.opcode = login_hdr->opcode;
 998	login_req_pdu_header.version_min = login_hdr->min_version;
 999	login_req_pdu_header.version_max = login_hdr->max_version;
1000	login_req_pdu_header.flags_attr = login_hdr->flags;
1001	login_req_pdu_header.isid_tabc = swab32p((u32 *)login_hdr->isid);
1002	login_req_pdu_header.isid_d = swab16p((u16 *)&login_hdr->isid[4]);
1003
1004	login_req_pdu_header.tsih = login_hdr->tsih;
1005	login_req_pdu_header.hdr_second_dword = ntoh24(login_hdr->dlength);
1006
1007	qedi_update_itt_map(qedi, tid, task->itt, qedi_cmd);
1008	login_req_pdu_header.itt = qedi_set_itt(tid, get_itt(task->itt));
1009	login_req_pdu_header.cid = qedi_conn->iscsi_conn_id;
1010	login_req_pdu_header.cmd_sn = be32_to_cpu(login_hdr->cmdsn);
1011	login_req_pdu_header.exp_stat_sn = be32_to_cpu(login_hdr->exp_statsn);
1012	login_req_pdu_header.exp_stat_sn = 0;
1013
1014	/* Fill tx AHS and rx buffer */
1015	tx_sgl_task_params.sgl =
1016			       (struct scsi_sge *)qedi_conn->gen_pdu.req_bd_tbl;
1017	tx_sgl_task_params.sgl_phys_addr.lo =
1018					 (u32)(qedi_conn->gen_pdu.req_dma_addr);
1019	tx_sgl_task_params.sgl_phys_addr.hi =
1020			      (u32)((u64)qedi_conn->gen_pdu.req_dma_addr >> 32);
1021	tx_sgl_task_params.total_buffer_size = ntoh24(login_hdr->dlength);
1022	tx_sgl_task_params.num_sges = 1;
1023
1024	rx_sgl_task_params.sgl =
1025			      (struct scsi_sge *)qedi_conn->gen_pdu.resp_bd_tbl;
1026	rx_sgl_task_params.sgl_phys_addr.lo =
1027					(u32)(qedi_conn->gen_pdu.resp_dma_addr);
1028	rx_sgl_task_params.sgl_phys_addr.hi =
1029			     (u32)((u64)qedi_conn->gen_pdu.resp_dma_addr >> 32);
1030	rx_sgl_task_params.total_buffer_size = resp_sge->sge_len;
1031	rx_sgl_task_params.num_sges = 1;
1032
1033	/* Fill fw input params */
1034	task_params.context = fw_task_ctx;
1035	task_params.conn_icid = (u16)qedi_conn->iscsi_conn_id;
1036	task_params.itid = tid;
1037	task_params.cq_rss_number = 0;
1038	task_params.tx_io_size = ntoh24(login_hdr->dlength);
1039	task_params.rx_io_size = resp_sge->sge_len;
1040
1041	sq_idx = qedi_get_wqe_idx(qedi_conn);
1042	task_params.sqe = &ep->sq[sq_idx];
1043
1044	memset(task_params.sqe, 0, sizeof(struct iscsi_wqe));
1045	rval = init_initiator_login_request_task(&task_params,
1046						 &login_req_pdu_header,
1047						 &tx_sgl_task_params,
1048						 &rx_sgl_task_params);
1049	if (rval)
1050		return -1;
1051
1052	spin_lock(&qedi_conn->list_lock);
1053	list_add_tail(&qedi_cmd->io_cmd, &qedi_conn->active_cmd_list);
1054	qedi_cmd->io_cmd_in_list = true;
1055	qedi_conn->active_cmd_count++;
1056	spin_unlock(&qedi_conn->list_lock);
1057
1058	qedi_ring_doorbell(qedi_conn);
1059	return 0;
1060}
1061
1062int qedi_send_iscsi_logout(struct qedi_conn *qedi_conn,
1063			   struct iscsi_task *task)
1064{
1065	struct iscsi_logout_req_hdr logout_pdu_header;
1066	struct scsi_sgl_task_params tx_sgl_task_params;
1067	struct scsi_sgl_task_params rx_sgl_task_params;
1068	struct iscsi_task_params task_params;
1069	struct iscsi_task_context *fw_task_ctx;
1070	struct iscsi_logout *logout_hdr = NULL;
1071	struct qedi_ctx *qedi = qedi_conn->qedi;
1072	struct qedi_cmd *qedi_cmd;
1073	struct qedi_endpoint *ep;
1074	s16 tid = 0;
1075	u16 sq_idx = 0;
1076	int rval = 0;
1077
1078	qedi_cmd = (struct qedi_cmd *)task->dd_data;
1079	logout_hdr = (struct iscsi_logout *)task->hdr;
1080	ep = qedi_conn->ep;
1081
1082	tid = qedi_get_task_idx(qedi);
1083	if (tid == -1)
1084		return -ENOMEM;
1085
1086	fw_task_ctx =
1087	     (struct iscsi_task_context *)qedi_get_task_mem(&qedi->tasks,
1088							       tid);
1089	memset(fw_task_ctx, 0, sizeof(struct iscsi_task_context));
1090
1091	qedi_cmd->task_id = tid;
1092
1093	memset(&task_params, 0, sizeof(task_params));
1094	memset(&logout_pdu_header, 0, sizeof(logout_pdu_header));
1095	memset(&tx_sgl_task_params, 0, sizeof(tx_sgl_task_params));
1096	memset(&rx_sgl_task_params, 0, sizeof(rx_sgl_task_params));
1097
1098	/* Update header info */
1099	logout_pdu_header.opcode = logout_hdr->opcode;
1100	logout_pdu_header.reason_code = 0x80 | logout_hdr->flags;
1101	qedi_update_itt_map(qedi, tid, task->itt, qedi_cmd);
1102	logout_pdu_header.itt = qedi_set_itt(tid, get_itt(task->itt));
1103	logout_pdu_header.exp_stat_sn = be32_to_cpu(logout_hdr->exp_statsn);
1104	logout_pdu_header.cmd_sn = be32_to_cpu(logout_hdr->cmdsn);
1105	logout_pdu_header.cid = qedi_conn->iscsi_conn_id;
1106
1107	/* Fill fw input params */
1108	task_params.context = fw_task_ctx;
1109	task_params.conn_icid = (u16)qedi_conn->iscsi_conn_id;
1110	task_params.itid = tid;
1111	task_params.cq_rss_number = 0;
1112	task_params.tx_io_size = 0;
1113	task_params.rx_io_size = 0;
1114
1115	sq_idx = qedi_get_wqe_idx(qedi_conn);
1116	task_params.sqe = &ep->sq[sq_idx];
1117	memset(task_params.sqe, 0, sizeof(struct iscsi_wqe));
1118
1119	rval = init_initiator_logout_request_task(&task_params,
1120						  &logout_pdu_header,
1121						  NULL, NULL);
1122	if (rval)
1123		return -1;
1124
1125	spin_lock(&qedi_conn->list_lock);
1126	list_add_tail(&qedi_cmd->io_cmd, &qedi_conn->active_cmd_list);
1127	qedi_cmd->io_cmd_in_list = true;
1128	qedi_conn->active_cmd_count++;
1129	spin_unlock(&qedi_conn->list_lock);
1130
1131	qedi_ring_doorbell(qedi_conn);
1132	return 0;
1133}
1134
1135int qedi_cleanup_all_io(struct qedi_ctx *qedi, struct qedi_conn *qedi_conn,
1136			struct iscsi_task *task, bool in_recovery)
1137{
1138	int rval;
1139	struct iscsi_task *ctask;
1140	struct qedi_cmd *cmd, *cmd_tmp;
1141	struct iscsi_tm *tmf_hdr;
1142	unsigned int lun = 0;
1143	bool lun_reset = false;
1144	struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
1145	struct iscsi_session *session = conn->session;
1146
1147	/* From recovery, task is NULL or from tmf resp valid task */
1148	if (task) {
1149		tmf_hdr = (struct iscsi_tm *)task->hdr;
1150
1151		if ((tmf_hdr->flags & ISCSI_FLAG_TM_FUNC_MASK) ==
1152			ISCSI_TM_FUNC_LOGICAL_UNIT_RESET) {
1153			lun_reset = true;
1154			lun = scsilun_to_int(&tmf_hdr->lun);
1155		}
1156	}
1157
1158	qedi_conn->cmd_cleanup_req = 0;
1159	atomic_set(&qedi_conn->cmd_cleanup_cmpl, 0);
1160
1161	QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
1162		  "active_cmd_count=%d, cid=0x%x, in_recovery=%d, lun_reset=%d\n",
1163		  qedi_conn->active_cmd_count, qedi_conn->iscsi_conn_id,
1164		  in_recovery, lun_reset);
1165
1166	if (lun_reset)
1167		spin_lock_bh(&session->back_lock);
1168
1169	spin_lock(&qedi_conn->list_lock);
1170
1171	list_for_each_entry_safe(cmd, cmd_tmp, &qedi_conn->active_cmd_list,
1172				 io_cmd) {
1173		ctask = cmd->task;
1174		if (ctask == task)
1175			continue;
1176
1177		if (lun_reset) {
1178			if (cmd->scsi_cmd && cmd->scsi_cmd->device) {
1179				QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
1180					  "tid=0x%x itt=0x%x scsi_cmd_ptr=%p device=%p task_state=%d cmd_state=0%x cid=0x%x\n",
1181					  cmd->task_id, get_itt(ctask->itt),
1182					  cmd->scsi_cmd, cmd->scsi_cmd->device,
1183					  ctask->state, cmd->state,
1184					  qedi_conn->iscsi_conn_id);
1185				if (cmd->scsi_cmd->device->lun != lun)
1186					continue;
1187			}
1188		}
1189		qedi_conn->cmd_cleanup_req++;
1190		qedi_iscsi_cleanup_task(ctask, true);
1191
1192		cmd->io_cmd_in_list = false;
1193		list_del_init(&cmd->io_cmd);
1194		qedi_conn->active_cmd_count--;
1195		QEDI_WARN(&qedi->dbg_ctx,
1196			  "Deleted active cmd list node io_cmd=%p, cid=0x%x\n",
1197			  &cmd->io_cmd, qedi_conn->iscsi_conn_id);
1198	}
1199
1200	spin_unlock(&qedi_conn->list_lock);
1201
1202	if (lun_reset)
1203		spin_unlock_bh(&session->back_lock);
1204
1205	QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
1206		  "cmd_cleanup_req=%d, cid=0x%x\n",
1207		  qedi_conn->cmd_cleanup_req,
1208		  qedi_conn->iscsi_conn_id);
1209
1210	rval  = wait_event_interruptible_timeout(qedi_conn->wait_queue,
1211				(qedi_conn->cmd_cleanup_req ==
1212				 atomic_read(&qedi_conn->cmd_cleanup_cmpl)) ||
1213				test_bit(QEDI_IN_RECOVERY, &qedi->flags),
1214				5 * HZ);
1215	if (rval) {
1216		QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
1217			  "i/o cmd_cleanup_req=%d, equal to cmd_cleanup_cmpl=%d, cid=0x%x\n",
1218			  qedi_conn->cmd_cleanup_req,
1219			  atomic_read(&qedi_conn->cmd_cleanup_cmpl),
1220			  qedi_conn->iscsi_conn_id);
1221
1222		return 0;
1223	}
1224
1225	QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
1226		  "i/o cmd_cleanup_req=%d, not equal to cmd_cleanup_cmpl=%d, cid=0x%x\n",
1227		  qedi_conn->cmd_cleanup_req,
1228		  atomic_read(&qedi_conn->cmd_cleanup_cmpl),
1229		  qedi_conn->iscsi_conn_id);
1230
1231	iscsi_host_for_each_session(qedi->shost,
1232				    qedi_mark_device_missing);
1233	qedi_ops->common->drain(qedi->cdev);
1234
1235	/* Enable IOs for all other sessions except current.*/
1236	if (!wait_event_interruptible_timeout(qedi_conn->wait_queue,
1237				(qedi_conn->cmd_cleanup_req ==
1238				 atomic_read(&qedi_conn->cmd_cleanup_cmpl)) ||
1239				test_bit(QEDI_IN_RECOVERY, &qedi->flags),
1240				5 * HZ)) {
1241		iscsi_host_for_each_session(qedi->shost,
1242					    qedi_mark_device_available);
1243		return -1;
1244	}
1245
1246	iscsi_host_for_each_session(qedi->shost,
1247				    qedi_mark_device_available);
1248
1249	return 0;
1250}
1251
1252void qedi_clearsq(struct qedi_ctx *qedi, struct qedi_conn *qedi_conn,
1253		  struct iscsi_task *task)
1254{
1255	struct qedi_endpoint *qedi_ep;
1256	int rval;
1257
1258	qedi_ep = qedi_conn->ep;
1259	qedi_conn->cmd_cleanup_req = 0;
1260	atomic_set(&qedi_conn->cmd_cleanup_cmpl, 0);
1261
1262	if (!qedi_ep) {
1263		QEDI_WARN(&qedi->dbg_ctx,
1264			  "Cannot proceed, ep already disconnected, cid=0x%x\n",
1265			  qedi_conn->iscsi_conn_id);
1266		return;
1267	}
1268
1269	QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_INFO,
1270		  "Clearing SQ for cid=0x%x, conn=%p, ep=%p\n",
1271		  qedi_conn->iscsi_conn_id, qedi_conn, qedi_ep);
1272
1273	qedi_ops->clear_sq(qedi->cdev, qedi_ep->handle);
1274
1275	rval = qedi_cleanup_all_io(qedi, qedi_conn, task, true);
1276	if (rval) {
1277		QEDI_ERR(&qedi->dbg_ctx,
1278			 "fatal error, need hard reset, cid=0x%x\n",
1279			 qedi_conn->iscsi_conn_id);
1280		WARN_ON(1);
1281	}
1282}
1283
1284static int qedi_wait_for_cleanup_request(struct qedi_ctx *qedi,
1285					 struct qedi_conn *qedi_conn,
1286					 struct iscsi_task *task,
1287					 struct qedi_cmd *qedi_cmd,
1288					 struct qedi_work_map *list_work)
1289{
1290	struct qedi_cmd *cmd = (struct qedi_cmd *)task->dd_data;
1291	int wait;
1292
1293	wait  = wait_event_interruptible_timeout(qedi_conn->wait_queue,
1294						 ((qedi_cmd->state ==
1295						   CLEANUP_RECV) ||
1296						 ((qedi_cmd->type == TYPEIO) &&
1297						  (cmd->state ==
1298						   RESPONSE_RECEIVED))),
1299						 5 * HZ);
1300	if (!wait) {
1301		qedi_cmd->state = CLEANUP_WAIT_FAILED;
1302
1303		QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
1304			  "Cleanup timedout tid=0x%x, issue connection recovery, cid=0x%x\n",
1305			  cmd->task_id, qedi_conn->iscsi_conn_id);
1306
1307		return -1;
1308	}
1309	return 0;
1310}
1311
1312static void qedi_abort_work(struct work_struct *work)
1313{
1314	struct qedi_cmd *qedi_cmd =
1315		container_of(work, struct qedi_cmd, tmf_work);
1316	struct qedi_conn *qedi_conn = qedi_cmd->conn;
1317	struct qedi_ctx *qedi = qedi_conn->qedi;
1318	struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
1319	struct qedi_work_map *list_work = NULL;
1320	struct iscsi_task *mtask;
1321	struct qedi_cmd *cmd;
1322	struct iscsi_task *ctask;
1323	struct iscsi_tm *tmf_hdr;
1324	s16 rval = 0;
1325
1326	mtask = qedi_cmd->task;
1327	tmf_hdr = (struct iscsi_tm *)mtask->hdr;
1328
1329	spin_lock_bh(&conn->session->back_lock);
1330	ctask = iscsi_itt_to_ctask(conn, tmf_hdr->rtt);
1331	if (!ctask) {
1332		spin_unlock_bh(&conn->session->back_lock);
1333		QEDI_ERR(&qedi->dbg_ctx, "Invalid RTT. Letting abort timeout.\n");
1334		goto clear_cleanup;
1335	}
1336
1337	if (iscsi_task_is_completed(ctask)) {
1338		spin_unlock_bh(&conn->session->back_lock);
1339		QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_INFO,
1340			  "Task already completed\n");
1341		/*
1342		 * We have to still send the TMF because libiscsi needs the
1343		 * response to avoid a timeout.
1344		 */
1345		goto send_tmf;
1346	}
1347	spin_unlock_bh(&conn->session->back_lock);
1348
1349	cmd = (struct qedi_cmd *)ctask->dd_data;
1350	QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_INFO,
1351		  "Abort tmf rtt=0x%x, cmd itt=0x%x, cmd tid=0x%x, cid=0x%x\n",
1352		  get_itt(tmf_hdr->rtt), get_itt(ctask->itt), cmd->task_id,
1353		  qedi_conn->iscsi_conn_id);
1354
1355	if (qedi_do_not_recover) {
1356		QEDI_ERR(&qedi->dbg_ctx, "DONT SEND CLEANUP/ABORT %d\n",
1357			 qedi_do_not_recover);
1358		goto clear_cleanup;
1359	}
1360
1361	list_work = kzalloc(sizeof(*list_work), GFP_NOIO);
1362	if (!list_work) {
1363		QEDI_ERR(&qedi->dbg_ctx, "Memory allocation failed\n");
1364		goto clear_cleanup;
1365	}
1366
1367	qedi_cmd->type = TYPEIO;
1368	qedi_cmd->state = CLEANUP_WAIT;
1369	list_work->qedi_cmd = qedi_cmd;
1370	list_work->rtid = cmd->task_id;
1371	list_work->state = QEDI_WORK_SCHEDULED;
1372	list_work->ctask = ctask;
1373	qedi_cmd->list_tmf_work = list_work;
1374
1375	QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
1376		  "Queue tmf work=%p, list node=%p, cid=0x%x, tmf flags=0x%x\n",
1377		  list_work->ptr_tmf_work, list_work, qedi_conn->iscsi_conn_id,
1378		  tmf_hdr->flags);
1379
1380	spin_lock_bh(&qedi_conn->tmf_work_lock);
1381	list_add_tail(&list_work->list, &qedi_conn->tmf_work_list);
1382	spin_unlock_bh(&qedi_conn->tmf_work_lock);
1383
1384	qedi_iscsi_cleanup_task(ctask, false);
1385
1386	rval = qedi_wait_for_cleanup_request(qedi, qedi_conn, ctask, qedi_cmd,
1387					     list_work);
1388	if (rval == -1) {
1389		QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_INFO,
1390			  "FW cleanup got escalated, cid=0x%x\n",
1391			  qedi_conn->iscsi_conn_id);
1392		goto ldel_exit;
1393	}
1394
1395send_tmf:
1396	send_iscsi_tmf(qedi_conn, qedi_cmd->task, ctask);
1397	goto clear_cleanup;
1398
1399ldel_exit:
1400	spin_lock_bh(&qedi_conn->tmf_work_lock);
1401	if (qedi_cmd->list_tmf_work) {
1402		list_del_init(&list_work->list);
1403		qedi_cmd->list_tmf_work = NULL;
1404		kfree(list_work);
1405	}
1406	spin_unlock_bh(&qedi_conn->tmf_work_lock);
1407
1408	spin_lock(&qedi_conn->list_lock);
1409	if (likely(cmd->io_cmd_in_list)) {
1410		cmd->io_cmd_in_list = false;
1411		list_del_init(&cmd->io_cmd);
1412		qedi_conn->active_cmd_count--;
1413	}
1414	spin_unlock(&qedi_conn->list_lock);
1415
1416clear_cleanup:
1417	spin_lock(&qedi_conn->tmf_work_lock);
1418	qedi_conn->fw_cleanup_works--;
1419	spin_unlock(&qedi_conn->tmf_work_lock);
1420}
1421
1422static int send_iscsi_tmf(struct qedi_conn *qedi_conn, struct iscsi_task *mtask,
1423			  struct iscsi_task *ctask)
1424{
1425	struct iscsi_tmf_request_hdr tmf_pdu_header;
1426	struct iscsi_task_params task_params;
1427	struct qedi_ctx *qedi = qedi_conn->qedi;
1428	struct iscsi_task_context *fw_task_ctx;
1429	struct iscsi_tm *tmf_hdr;
1430	struct qedi_cmd *qedi_cmd;
1431	struct qedi_cmd *cmd;
1432	struct qedi_endpoint *ep;
1433	u32 scsi_lun[2];
1434	s16 tid = 0;
1435	u16 sq_idx = 0;
1436
1437	tmf_hdr = (struct iscsi_tm *)mtask->hdr;
1438	qedi_cmd = (struct qedi_cmd *)mtask->dd_data;
1439	ep = qedi_conn->ep;
1440	if (!ep)
1441		return -ENODEV;
1442
1443	tid = qedi_get_task_idx(qedi);
1444	if (tid == -1)
1445		return -ENOMEM;
1446
1447	fw_task_ctx =
1448	     (struct iscsi_task_context *)qedi_get_task_mem(&qedi->tasks,
1449							       tid);
1450	memset(fw_task_ctx, 0, sizeof(struct iscsi_task_context));
1451
1452	qedi_cmd->task_id = tid;
1453
1454	memset(&task_params, 0, sizeof(task_params));
1455	memset(&tmf_pdu_header, 0, sizeof(tmf_pdu_header));
1456
1457	/* Update header info */
1458	qedi_update_itt_map(qedi, tid, mtask->itt, qedi_cmd);
1459	tmf_pdu_header.itt = qedi_set_itt(tid, get_itt(mtask->itt));
1460	tmf_pdu_header.cmd_sn = be32_to_cpu(tmf_hdr->cmdsn);
1461
1462	memcpy(scsi_lun, &tmf_hdr->lun, sizeof(struct scsi_lun));
1463	tmf_pdu_header.lun.lo = be32_to_cpu(scsi_lun[0]);
1464	tmf_pdu_header.lun.hi = be32_to_cpu(scsi_lun[1]);
1465
1466	if ((tmf_hdr->flags & ISCSI_FLAG_TM_FUNC_MASK) ==
1467	     ISCSI_TM_FUNC_ABORT_TASK) {
1468		cmd = (struct qedi_cmd *)ctask->dd_data;
1469		tmf_pdu_header.rtt =
1470				qedi_set_itt(cmd->task_id,
1471					     get_itt(tmf_hdr->rtt));
1472	} else {
1473		tmf_pdu_header.rtt = ISCSI_RESERVED_TAG;
1474	}
1475
1476	tmf_pdu_header.opcode = tmf_hdr->opcode;
1477	tmf_pdu_header.function = tmf_hdr->flags;
1478	tmf_pdu_header.hdr_second_dword = ntoh24(tmf_hdr->dlength);
1479	tmf_pdu_header.ref_cmd_sn = be32_to_cpu(tmf_hdr->refcmdsn);
1480
1481	/* Fill fw input params */
1482	task_params.context = fw_task_ctx;
1483	task_params.conn_icid = (u16)qedi_conn->iscsi_conn_id;
1484	task_params.itid = tid;
1485	task_params.cq_rss_number = 0;
1486	task_params.tx_io_size = 0;
1487	task_params.rx_io_size = 0;
1488
1489	sq_idx = qedi_get_wqe_idx(qedi_conn);
1490	task_params.sqe = &ep->sq[sq_idx];
1491
1492	memset(task_params.sqe, 0, sizeof(struct iscsi_wqe));
1493	init_initiator_tmf_request_task(&task_params, &tmf_pdu_header);
1494
1495	spin_lock(&qedi_conn->list_lock);
1496	list_add_tail(&qedi_cmd->io_cmd, &qedi_conn->active_cmd_list);
1497	qedi_cmd->io_cmd_in_list = true;
1498	qedi_conn->active_cmd_count++;
1499	spin_unlock(&qedi_conn->list_lock);
1500
1501	qedi_ring_doorbell(qedi_conn);
1502	return 0;
1503}
1504
1505int qedi_send_iscsi_tmf(struct qedi_conn *qedi_conn, struct iscsi_task *mtask)
1506{
1507	struct iscsi_tm *tmf_hdr = (struct iscsi_tm *)mtask->hdr;
1508	struct qedi_cmd *qedi_cmd = mtask->dd_data;
1509	struct qedi_ctx *qedi = qedi_conn->qedi;
1510	int rc = 0;
1511
1512	switch (tmf_hdr->flags & ISCSI_FLAG_TM_FUNC_MASK) {
1513	case ISCSI_TM_FUNC_ABORT_TASK:
1514		spin_lock(&qedi_conn->tmf_work_lock);
1515		qedi_conn->fw_cleanup_works++;
1516		spin_unlock(&qedi_conn->tmf_work_lock);
1517
1518		INIT_WORK(&qedi_cmd->tmf_work, qedi_abort_work);
1519		queue_work(qedi->tmf_thread, &qedi_cmd->tmf_work);
1520		break;
1521	case ISCSI_TM_FUNC_LOGICAL_UNIT_RESET:
1522	case ISCSI_TM_FUNC_TARGET_WARM_RESET:
1523	case ISCSI_TM_FUNC_TARGET_COLD_RESET:
1524		rc = send_iscsi_tmf(qedi_conn, mtask, NULL);
1525		break;
1526	default:
1527		QEDI_ERR(&qedi->dbg_ctx, "Invalid tmf, cid=0x%x\n",
1528			 qedi_conn->iscsi_conn_id);
1529		return -EINVAL;
1530	}
1531
1532	return rc;
1533}
1534
1535int qedi_send_iscsi_text(struct qedi_conn *qedi_conn,
1536			 struct iscsi_task *task)
1537{
1538	struct iscsi_text_request_hdr text_request_pdu_header;
1539	struct scsi_sgl_task_params tx_sgl_task_params;
1540	struct scsi_sgl_task_params rx_sgl_task_params;
1541	struct iscsi_task_params task_params;
1542	struct iscsi_task_context *fw_task_ctx;
1543	struct qedi_ctx *qedi = qedi_conn->qedi;
1544	struct iscsi_text *text_hdr;
1545	struct scsi_sge *req_sge = NULL;
1546	struct scsi_sge *resp_sge = NULL;
1547	struct qedi_cmd *qedi_cmd;
1548	struct qedi_endpoint *ep;
1549	s16 tid = 0;
1550	u16 sq_idx = 0;
1551	int rval = 0;
1552
1553	req_sge = (struct scsi_sge *)qedi_conn->gen_pdu.req_bd_tbl;
1554	resp_sge = (struct scsi_sge *)qedi_conn->gen_pdu.resp_bd_tbl;
1555	qedi_cmd = (struct qedi_cmd *)task->dd_data;
1556	text_hdr = (struct iscsi_text *)task->hdr;
1557	ep = qedi_conn->ep;
1558
1559	tid = qedi_get_task_idx(qedi);
1560	if (tid == -1)
1561		return -ENOMEM;
1562
1563	fw_task_ctx =
1564	     (struct iscsi_task_context *)qedi_get_task_mem(&qedi->tasks,
1565							       tid);
1566	memset(fw_task_ctx, 0, sizeof(struct iscsi_task_context));
1567
1568	qedi_cmd->task_id = tid;
1569
1570	memset(&task_params, 0, sizeof(task_params));
1571	memset(&text_request_pdu_header, 0, sizeof(text_request_pdu_header));
1572	memset(&tx_sgl_task_params, 0, sizeof(tx_sgl_task_params));
1573	memset(&rx_sgl_task_params, 0, sizeof(rx_sgl_task_params));
1574
1575	/* Update header info */
1576	text_request_pdu_header.opcode = text_hdr->opcode;
1577	text_request_pdu_header.flags_attr = text_hdr->flags;
1578
1579	qedi_update_itt_map(qedi, tid, task->itt, qedi_cmd);
1580	text_request_pdu_header.itt = qedi_set_itt(tid, get_itt(task->itt));
1581	text_request_pdu_header.ttt = text_hdr->ttt;
1582	text_request_pdu_header.cmd_sn = be32_to_cpu(text_hdr->cmdsn);
1583	text_request_pdu_header.exp_stat_sn = be32_to_cpu(text_hdr->exp_statsn);
1584	text_request_pdu_header.hdr_second_dword = ntoh24(text_hdr->dlength);
1585
1586	/* Fill tx AHS and rx buffer */
1587	tx_sgl_task_params.sgl =
1588			       (struct scsi_sge *)qedi_conn->gen_pdu.req_bd_tbl;
1589	tx_sgl_task_params.sgl_phys_addr.lo =
1590					 (u32)(qedi_conn->gen_pdu.req_dma_addr);
1591	tx_sgl_task_params.sgl_phys_addr.hi =
1592			      (u32)((u64)qedi_conn->gen_pdu.req_dma_addr >> 32);
1593	tx_sgl_task_params.total_buffer_size = req_sge->sge_len;
1594	tx_sgl_task_params.num_sges = 1;
1595
1596	rx_sgl_task_params.sgl =
1597			      (struct scsi_sge *)qedi_conn->gen_pdu.resp_bd_tbl;
1598	rx_sgl_task_params.sgl_phys_addr.lo =
1599					(u32)(qedi_conn->gen_pdu.resp_dma_addr);
1600	rx_sgl_task_params.sgl_phys_addr.hi =
1601			     (u32)((u64)qedi_conn->gen_pdu.resp_dma_addr >> 32);
1602	rx_sgl_task_params.total_buffer_size = resp_sge->sge_len;
1603	rx_sgl_task_params.num_sges = 1;
1604
1605	/* Fill fw input params */
1606	task_params.context = fw_task_ctx;
1607	task_params.conn_icid = (u16)qedi_conn->iscsi_conn_id;
1608	task_params.itid = tid;
1609	task_params.cq_rss_number = 0;
1610	task_params.tx_io_size = ntoh24(text_hdr->dlength);
1611	task_params.rx_io_size = resp_sge->sge_len;
1612
1613	sq_idx = qedi_get_wqe_idx(qedi_conn);
1614	task_params.sqe = &ep->sq[sq_idx];
1615
1616	memset(task_params.sqe, 0, sizeof(struct iscsi_wqe));
1617	rval = init_initiator_text_request_task(&task_params,
1618						&text_request_pdu_header,
1619						&tx_sgl_task_params,
1620						&rx_sgl_task_params);
1621	if (rval)
1622		return -1;
1623
1624	spin_lock(&qedi_conn->list_lock);
1625	list_add_tail(&qedi_cmd->io_cmd, &qedi_conn->active_cmd_list);
1626	qedi_cmd->io_cmd_in_list = true;
1627	qedi_conn->active_cmd_count++;
1628	spin_unlock(&qedi_conn->list_lock);
1629
1630	qedi_ring_doorbell(qedi_conn);
1631	return 0;
1632}
1633
1634int qedi_send_iscsi_nopout(struct qedi_conn *qedi_conn,
1635			   struct iscsi_task *task,
1636			   char *datap, int data_len, int unsol)
1637{
1638	struct iscsi_nop_out_hdr nop_out_pdu_header;
1639	struct scsi_sgl_task_params tx_sgl_task_params;
1640	struct scsi_sgl_task_params rx_sgl_task_params;
1641	struct iscsi_task_params task_params;
1642	struct qedi_ctx *qedi = qedi_conn->qedi;
1643	struct iscsi_task_context *fw_task_ctx;
1644	struct iscsi_nopout *nopout_hdr;
1645	struct scsi_sge *resp_sge = NULL;
1646	struct qedi_cmd *qedi_cmd;
1647	struct qedi_endpoint *ep;
1648	u32 scsi_lun[2];
1649	s16 tid = 0;
1650	u16 sq_idx = 0;
1651	int rval = 0;
1652
1653	resp_sge = (struct scsi_sge *)qedi_conn->gen_pdu.resp_bd_tbl;
1654	qedi_cmd = (struct qedi_cmd *)task->dd_data;
1655	nopout_hdr = (struct iscsi_nopout *)task->hdr;
1656	ep = qedi_conn->ep;
1657
1658	tid = qedi_get_task_idx(qedi);
1659	if (tid == -1)
1660		return -ENOMEM;
1661
1662	fw_task_ctx =
1663	     (struct iscsi_task_context *)qedi_get_task_mem(&qedi->tasks,
1664							       tid);
1665	memset(fw_task_ctx, 0, sizeof(struct iscsi_task_context));
1666
1667	qedi_cmd->task_id = tid;
1668
1669	memset(&task_params, 0, sizeof(task_params));
1670	memset(&nop_out_pdu_header, 0, sizeof(nop_out_pdu_header));
1671	memset(&tx_sgl_task_params, 0, sizeof(tx_sgl_task_params));
1672	memset(&rx_sgl_task_params, 0, sizeof(rx_sgl_task_params));
1673
1674	/* Update header info */
1675	nop_out_pdu_header.opcode = nopout_hdr->opcode;
1676	SET_FIELD(nop_out_pdu_header.flags_attr, ISCSI_NOP_OUT_HDR_CONST1, 1);
1677	SET_FIELD(nop_out_pdu_header.flags_attr, ISCSI_NOP_OUT_HDR_RSRV, 0);
1678
1679	memcpy(scsi_lun, &nopout_hdr->lun, sizeof(struct scsi_lun));
1680	nop_out_pdu_header.lun.lo = be32_to_cpu(scsi_lun[0]);
1681	nop_out_pdu_header.lun.hi = be32_to_cpu(scsi_lun[1]);
1682	nop_out_pdu_header.cmd_sn = be32_to_cpu(nopout_hdr->cmdsn);
1683	nop_out_pdu_header.exp_stat_sn = be32_to_cpu(nopout_hdr->exp_statsn);
1684
1685	qedi_update_itt_map(qedi, tid, task->itt, qedi_cmd);
1686
1687	if (nopout_hdr->ttt != ISCSI_TTT_ALL_ONES) {
1688		nop_out_pdu_header.itt = be32_to_cpu(nopout_hdr->itt);
1689		nop_out_pdu_header.ttt = be32_to_cpu(nopout_hdr->ttt);
1690	} else {
1691		nop_out_pdu_header.itt = qedi_set_itt(tid, get_itt(task->itt));
1692		nop_out_pdu_header.ttt = ISCSI_TTT_ALL_ONES;
1693
1694		spin_lock(&qedi_conn->list_lock);
1695		list_add_tail(&qedi_cmd->io_cmd, &qedi_conn->active_cmd_list);
1696		qedi_cmd->io_cmd_in_list = true;
1697		qedi_conn->active_cmd_count++;
1698		spin_unlock(&qedi_conn->list_lock);
1699	}
1700
1701	/* Fill tx AHS and rx buffer */
1702	if (data_len) {
1703		tx_sgl_task_params.sgl =
1704			       (struct scsi_sge *)qedi_conn->gen_pdu.req_bd_tbl;
1705		tx_sgl_task_params.sgl_phys_addr.lo =
1706					 (u32)(qedi_conn->gen_pdu.req_dma_addr);
1707		tx_sgl_task_params.sgl_phys_addr.hi =
1708			      (u32)((u64)qedi_conn->gen_pdu.req_dma_addr >> 32);
1709		tx_sgl_task_params.total_buffer_size = data_len;
1710		tx_sgl_task_params.num_sges = 1;
1711
1712		rx_sgl_task_params.sgl =
1713			      (struct scsi_sge *)qedi_conn->gen_pdu.resp_bd_tbl;
1714		rx_sgl_task_params.sgl_phys_addr.lo =
1715					(u32)(qedi_conn->gen_pdu.resp_dma_addr);
1716		rx_sgl_task_params.sgl_phys_addr.hi =
1717			     (u32)((u64)qedi_conn->gen_pdu.resp_dma_addr >> 32);
1718		rx_sgl_task_params.total_buffer_size = resp_sge->sge_len;
1719		rx_sgl_task_params.num_sges = 1;
1720	}
1721
1722	/* Fill fw input params */
1723	task_params.context = fw_task_ctx;
1724	task_params.conn_icid = (u16)qedi_conn->iscsi_conn_id;
1725	task_params.itid = tid;
1726	task_params.cq_rss_number = 0;
1727	task_params.tx_io_size = data_len;
1728	task_params.rx_io_size = resp_sge->sge_len;
1729
1730	sq_idx = qedi_get_wqe_idx(qedi_conn);
1731	task_params.sqe = &ep->sq[sq_idx];
1732
1733	memset(task_params.sqe, 0, sizeof(struct iscsi_wqe));
1734	rval = init_initiator_nop_out_task(&task_params,
1735					   &nop_out_pdu_header,
1736					   &tx_sgl_task_params,
1737					   &rx_sgl_task_params);
1738	if (rval)
1739		return -1;
1740
1741	qedi_ring_doorbell(qedi_conn);
1742	return 0;
1743}
1744
1745static int qedi_split_bd(struct qedi_cmd *cmd, u64 addr, int sg_len,
1746			 int bd_index)
1747{
1748	struct scsi_sge *bd = cmd->io_tbl.sge_tbl;
1749	int frag_size, sg_frags;
1750
1751	sg_frags = 0;
1752
1753	while (sg_len) {
1754		if (addr % QEDI_PAGE_SIZE)
1755			frag_size =
1756				   (QEDI_PAGE_SIZE - (addr % QEDI_PAGE_SIZE));
1757		else
1758			frag_size = (sg_len > QEDI_BD_SPLIT_SZ) ? 0 :
1759				    (sg_len % QEDI_BD_SPLIT_SZ);
1760
1761		if (frag_size == 0)
1762			frag_size = QEDI_BD_SPLIT_SZ;
1763
1764		bd[bd_index + sg_frags].sge_addr.lo = (addr & 0xffffffff);
1765		bd[bd_index + sg_frags].sge_addr.hi = (addr >> 32);
1766		bd[bd_index + sg_frags].sge_len = (u16)frag_size;
1767		QEDI_INFO(&cmd->conn->qedi->dbg_ctx, QEDI_LOG_IO,
1768			  "split sge %d: addr=%llx, len=%x",
1769			  (bd_index + sg_frags), addr, frag_size);
1770
1771		addr += (u64)frag_size;
1772		sg_frags++;
1773		sg_len -= frag_size;
1774	}
1775	return sg_frags;
1776}
1777
1778static int qedi_map_scsi_sg(struct qedi_ctx *qedi, struct qedi_cmd *cmd)
1779{
1780	struct scsi_cmnd *sc = cmd->scsi_cmd;
1781	struct scsi_sge *bd = cmd->io_tbl.sge_tbl;
1782	struct scatterlist *sg;
1783	int byte_count = 0;
1784	int bd_count = 0;
1785	int sg_count;
1786	int sg_len;
1787	int sg_frags;
1788	u64 addr, end_addr;
1789	int i;
1790
1791	WARN_ON(scsi_sg_count(sc) > QEDI_ISCSI_MAX_BDS_PER_CMD);
1792
1793	sg_count = dma_map_sg(&qedi->pdev->dev, scsi_sglist(sc),
1794			      scsi_sg_count(sc), sc->sc_data_direction);
1795
1796	/*
1797	 * New condition to send single SGE as cached-SGL.
1798	 * Single SGE with length less than 64K.
1799	 */
1800	sg = scsi_sglist(sc);
1801	if ((sg_count == 1) && (sg_dma_len(sg) <= MAX_SGLEN_FOR_CACHESGL)) {
1802		sg_len = sg_dma_len(sg);
1803		addr = (u64)sg_dma_address(sg);
1804
1805		bd[bd_count].sge_addr.lo = (addr & 0xffffffff);
1806		bd[bd_count].sge_addr.hi = (addr >> 32);
1807		bd[bd_count].sge_len = (u16)sg_len;
1808
1809		QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_IO,
1810			  "single-cached-sgl: bd_count:%d addr=%llx, len=%x",
1811			  sg_count, addr, sg_len);
1812
1813		return ++bd_count;
1814	}
1815
1816	scsi_for_each_sg(sc, sg, sg_count, i) {
1817		sg_len = sg_dma_len(sg);
1818		addr = (u64)sg_dma_address(sg);
1819		end_addr = (addr + sg_len);
1820
1821		/*
1822		 * first sg elem in the 'list',
1823		 * check if end addr is page-aligned.
1824		 */
1825		if ((i == 0) && (sg_count > 1) && (end_addr % QEDI_PAGE_SIZE))
1826			cmd->use_slowpath = true;
1827
1828		/*
1829		 * last sg elem in the 'list',
1830		 * check if start addr is page-aligned.
1831		 */
1832		else if ((i == (sg_count - 1)) &&
1833			 (sg_count > 1) && (addr % QEDI_PAGE_SIZE))
1834			cmd->use_slowpath = true;
1835
1836		/*
1837		 * middle sg elements in list,
1838		 * check if start and end addr is page-aligned
1839		 */
1840		else if ((i != 0) && (i != (sg_count - 1)) &&
1841			 ((addr % QEDI_PAGE_SIZE) ||
1842			 (end_addr % QEDI_PAGE_SIZE)))
1843			cmd->use_slowpath = true;
1844
1845		QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_IO, "sg[%d] size=0x%x",
1846			  i, sg_len);
1847
1848		if (sg_len > QEDI_BD_SPLIT_SZ) {
1849			sg_frags = qedi_split_bd(cmd, addr, sg_len, bd_count);
1850		} else {
1851			sg_frags = 1;
1852			bd[bd_count].sge_addr.lo = addr & 0xffffffff;
1853			bd[bd_count].sge_addr.hi = addr >> 32;
1854			bd[bd_count].sge_len = sg_len;
1855		}
1856		byte_count += sg_len;
1857		bd_count += sg_frags;
1858	}
1859
1860	if (byte_count != scsi_bufflen(sc))
1861		QEDI_ERR(&qedi->dbg_ctx,
1862			 "byte_count = %d != scsi_bufflen = %d\n", byte_count,
1863			 scsi_bufflen(sc));
1864	else
1865		QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_IO, "byte_count = %d\n",
1866			  byte_count);
1867
1868	WARN_ON(byte_count != scsi_bufflen(sc));
1869
1870	return bd_count;
1871}
1872
1873static void qedi_iscsi_map_sg_list(struct qedi_cmd *cmd)
1874{
1875	int bd_count;
1876	struct scsi_cmnd *sc = cmd->scsi_cmd;
1877
1878	if (scsi_sg_count(sc)) {
1879		bd_count  = qedi_map_scsi_sg(cmd->conn->qedi, cmd);
1880		if (bd_count == 0)
1881			return;
1882	} else {
1883		struct scsi_sge *bd = cmd->io_tbl.sge_tbl;
1884
1885		bd[0].sge_addr.lo = 0;
1886		bd[0].sge_addr.hi = 0;
1887		bd[0].sge_len = 0;
1888		bd_count = 0;
1889	}
1890	cmd->io_tbl.sge_valid = bd_count;
1891}
1892
1893static void qedi_cpy_scsi_cdb(struct scsi_cmnd *sc, u32 *dstp)
1894{
1895	u32 dword;
1896	int lpcnt;
1897	u8 *srcp;
1898
1899	lpcnt = sc->cmd_len / sizeof(dword);
1900	srcp = (u8 *)sc->cmnd;
1901	while (lpcnt--) {
1902		memcpy(&dword, (const void *)srcp, 4);
1903		*dstp = cpu_to_be32(dword);
1904		srcp += 4;
1905		dstp++;
1906	}
1907	if (sc->cmd_len & 0x3) {
1908		dword = (u32)srcp[0] | ((u32)srcp[1] << 8);
1909		*dstp = cpu_to_be32(dword);
1910	}
1911}
1912
1913void qedi_trace_io(struct qedi_ctx *qedi, struct iscsi_task *task,
1914		   u16 tid, int8_t direction)
1915{
1916	struct qedi_io_log *io_log;
1917	struct iscsi_conn *conn = task->conn;
1918	struct qedi_conn *qedi_conn = conn->dd_data;
1919	struct scsi_cmnd *sc_cmd = task->sc;
1920	unsigned long flags;
1921
1922	spin_lock_irqsave(&qedi->io_trace_lock, flags);
1923
1924	io_log = &qedi->io_trace_buf[qedi->io_trace_idx];
1925	io_log->direction = direction;
1926	io_log->task_id = tid;
1927	io_log->cid = qedi_conn->iscsi_conn_id;
1928	io_log->lun = sc_cmd->device->lun;
1929	io_log->op = sc_cmd->cmnd[0];
1930	io_log->lba[0] = sc_cmd->cmnd[2];
1931	io_log->lba[1] = sc_cmd->cmnd[3];
1932	io_log->lba[2] = sc_cmd->cmnd[4];
1933	io_log->lba[3] = sc_cmd->cmnd[5];
1934	io_log->bufflen = scsi_bufflen(sc_cmd);
1935	io_log->sg_count = scsi_sg_count(sc_cmd);
1936	io_log->fast_sgs = qedi->fast_sgls;
1937	io_log->cached_sgs = qedi->cached_sgls;
1938	io_log->slow_sgs = qedi->slow_sgls;
1939	io_log->cached_sge = qedi->use_cached_sge;
1940	io_log->slow_sge = qedi->use_slow_sge;
1941	io_log->fast_sge = qedi->use_fast_sge;
1942	io_log->result = sc_cmd->result;
1943	io_log->jiffies = jiffies;
1944	io_log->blk_req_cpu = smp_processor_id();
1945
1946	if (direction == QEDI_IO_TRACE_REQ) {
1947		/* For requests we only care about the submission CPU */
1948		io_log->req_cpu = smp_processor_id() % qedi->num_queues;
1949		io_log->intr_cpu = 0;
1950		io_log->blk_rsp_cpu = 0;
1951	} else if (direction == QEDI_IO_TRACE_RSP) {
1952		io_log->req_cpu = smp_processor_id() % qedi->num_queues;
1953		io_log->intr_cpu = qedi->intr_cpu;
1954		io_log->blk_rsp_cpu = smp_processor_id();
1955	}
1956
1957	qedi->io_trace_idx++;
1958	if (qedi->io_trace_idx == QEDI_IO_TRACE_SIZE)
1959		qedi->io_trace_idx = 0;
1960
1961	qedi->use_cached_sge = false;
1962	qedi->use_slow_sge = false;
1963	qedi->use_fast_sge = false;
1964
1965	spin_unlock_irqrestore(&qedi->io_trace_lock, flags);
1966}
1967
1968int qedi_iscsi_send_ioreq(struct iscsi_task *task)
1969{
1970	struct iscsi_conn *conn = task->conn;
1971	struct iscsi_session *session = conn->session;
1972	struct Scsi_Host *shost = iscsi_session_to_shost(session->cls_session);
1973	struct qedi_ctx *qedi = iscsi_host_priv(shost);
1974	struct qedi_conn *qedi_conn = conn->dd_data;
1975	struct qedi_cmd *cmd = task->dd_data;
1976	struct scsi_cmnd *sc = task->sc;
1977	struct iscsi_cmd_hdr cmd_pdu_header;
1978	struct scsi_sgl_task_params tx_sgl_task_params;
1979	struct scsi_sgl_task_params rx_sgl_task_params;
1980	struct scsi_sgl_task_params *prx_sgl = NULL;
1981	struct scsi_sgl_task_params *ptx_sgl = NULL;
1982	struct iscsi_task_params task_params;
1983	struct iscsi_conn_params conn_params;
1984	struct scsi_initiator_cmd_params cmd_params;
1985	struct iscsi_task_context *fw_task_ctx;
1986	struct iscsi_cls_conn *cls_conn;
1987	struct iscsi_scsi_req *hdr = (struct iscsi_scsi_req *)task->hdr;
1988	enum iscsi_task_type task_type = MAX_ISCSI_TASK_TYPE;
1989	struct qedi_endpoint *ep;
1990	u32 scsi_lun[2];
1991	s16 tid = 0;
1992	u16 sq_idx = 0;
1993	u16 cq_idx;
1994	int rval = 0;
1995
1996	ep = qedi_conn->ep;
1997	cls_conn = qedi_conn->cls_conn;
1998	conn = cls_conn->dd_data;
1999
2000	qedi_iscsi_map_sg_list(cmd);
2001	int_to_scsilun(sc->device->lun, (struct scsi_lun *)scsi_lun);
2002
2003	tid = qedi_get_task_idx(qedi);
2004	if (tid == -1)
2005		return -ENOMEM;
2006
2007	fw_task_ctx =
2008	     (struct iscsi_task_context *)qedi_get_task_mem(&qedi->tasks,
2009							       tid);
2010	memset(fw_task_ctx, 0, sizeof(struct iscsi_task_context));
2011
2012	cmd->task_id = tid;
2013
2014	memset(&task_params, 0, sizeof(task_params));
2015	memset(&cmd_pdu_header, 0, sizeof(cmd_pdu_header));
2016	memset(&tx_sgl_task_params, 0, sizeof(tx_sgl_task_params));
2017	memset(&rx_sgl_task_params, 0, sizeof(rx_sgl_task_params));
2018	memset(&conn_params, 0, sizeof(conn_params));
2019	memset(&cmd_params, 0, sizeof(cmd_params));
2020
2021	cq_idx = smp_processor_id() % qedi->num_queues;
2022	/* Update header info */
2023	SET_FIELD(cmd_pdu_header.flags_attr, ISCSI_CMD_HDR_ATTR,
2024		  ISCSI_ATTR_SIMPLE);
2025	if (hdr->cdb[0] != TEST_UNIT_READY) {
2026		if (sc->sc_data_direction == DMA_TO_DEVICE) {
2027			SET_FIELD(cmd_pdu_header.flags_attr,
2028				  ISCSI_CMD_HDR_WRITE, 1);
2029			task_type = ISCSI_TASK_TYPE_INITIATOR_WRITE;
2030		} else {
2031			SET_FIELD(cmd_pdu_header.flags_attr,
2032				  ISCSI_CMD_HDR_READ, 1);
2033			task_type = ISCSI_TASK_TYPE_INITIATOR_READ;
2034		}
2035	}
2036
2037	cmd_pdu_header.lun.lo = be32_to_cpu(scsi_lun[0]);
2038	cmd_pdu_header.lun.hi = be32_to_cpu(scsi_lun[1]);
2039
2040	qedi_update_itt_map(qedi, tid, task->itt, cmd);
2041	cmd_pdu_header.itt = qedi_set_itt(tid, get_itt(task->itt));
2042	cmd_pdu_header.expected_transfer_length = cpu_to_be32(hdr->data_length);
2043	cmd_pdu_header.hdr_second_dword = ntoh24(hdr->dlength);
2044	cmd_pdu_header.cmd_sn = be32_to_cpu(hdr->cmdsn);
2045	cmd_pdu_header.hdr_first_byte = hdr->opcode;
2046	qedi_cpy_scsi_cdb(sc, (u32 *)cmd_pdu_header.cdb);
2047
2048	/* Fill tx AHS and rx buffer */
2049	if (task_type == ISCSI_TASK_TYPE_INITIATOR_WRITE) {
2050		tx_sgl_task_params.sgl = cmd->io_tbl.sge_tbl;
2051		tx_sgl_task_params.sgl_phys_addr.lo =
2052						 (u32)(cmd->io_tbl.sge_tbl_dma);
2053		tx_sgl_task_params.sgl_phys_addr.hi =
2054				      (u32)((u64)cmd->io_tbl.sge_tbl_dma >> 32);
2055		tx_sgl_task_params.total_buffer_size = scsi_bufflen(sc);
2056		tx_sgl_task_params.num_sges = cmd->io_tbl.sge_valid;
2057		if (cmd->use_slowpath)
2058			tx_sgl_task_params.small_mid_sge = true;
2059	} else if (task_type == ISCSI_TASK_TYPE_INITIATOR_READ) {
2060		rx_sgl_task_params.sgl = cmd->io_tbl.sge_tbl;
2061		rx_sgl_task_params.sgl_phys_addr.lo =
2062						 (u32)(cmd->io_tbl.sge_tbl_dma);
2063		rx_sgl_task_params.sgl_phys_addr.hi =
2064				      (u32)((u64)cmd->io_tbl.sge_tbl_dma >> 32);
2065		rx_sgl_task_params.total_buffer_size = scsi_bufflen(sc);
2066		rx_sgl_task_params.num_sges = cmd->io_tbl.sge_valid;
2067	}
2068
2069	/* Add conn param */
2070	conn_params.first_burst_length = conn->session->first_burst;
2071	conn_params.max_send_pdu_length = conn->max_xmit_dlength;
2072	conn_params.max_burst_length = conn->session->max_burst;
2073	if (conn->session->initial_r2t_en)
2074		conn_params.initial_r2t = true;
2075	if (conn->session->imm_data_en)
2076		conn_params.immediate_data = true;
2077
2078	/* Add cmd params */
2079	cmd_params.sense_data_buffer_phys_addr.lo = (u32)cmd->sense_buffer_dma;
2080	cmd_params.sense_data_buffer_phys_addr.hi =
2081					(u32)((u64)cmd->sense_buffer_dma >> 32);
2082	/* Fill fw input params */
2083	task_params.context = fw_task_ctx;
2084	task_params.conn_icid = (u16)qedi_conn->iscsi_conn_id;
2085	task_params.itid = tid;
2086	task_params.cq_rss_number = cq_idx;
2087	if (task_type == ISCSI_TASK_TYPE_INITIATOR_WRITE)
2088		task_params.tx_io_size = scsi_bufflen(sc);
2089	else if (task_type == ISCSI_TASK_TYPE_INITIATOR_READ)
2090		task_params.rx_io_size = scsi_bufflen(sc);
2091
2092	sq_idx = qedi_get_wqe_idx(qedi_conn);
2093	task_params.sqe = &ep->sq[sq_idx];
2094
2095	QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_IO,
2096		  "%s: %s-SGL: sg_len=0x%x num_sges=0x%x first-sge-lo=0x%x first-sge-hi=0x%x\n",
2097		  (task_type == ISCSI_TASK_TYPE_INITIATOR_WRITE) ?
2098		  "Write " : "Read ", (cmd->io_tbl.sge_valid == 1) ?
2099		  "Single" : (cmd->use_slowpath ? "SLOW" : "FAST"),
2100		  (u16)cmd->io_tbl.sge_valid, scsi_bufflen(sc),
2101		  (u32)(cmd->io_tbl.sge_tbl_dma),
2102		  (u32)((u64)cmd->io_tbl.sge_tbl_dma >> 32));
2103
2104	memset(task_params.sqe, 0, sizeof(struct iscsi_wqe));
2105
2106	if (task_params.tx_io_size != 0)
2107		ptx_sgl = &tx_sgl_task_params;
2108	if (task_params.rx_io_size != 0)
2109		prx_sgl = &rx_sgl_task_params;
2110
2111	rval = init_initiator_rw_iscsi_task(&task_params, &conn_params,
2112					    &cmd_params, &cmd_pdu_header,
2113					    ptx_sgl, prx_sgl,
2114					    NULL);
2115	if (rval)
2116		return -1;
2117
2118	spin_lock(&qedi_conn->list_lock);
2119	list_add_tail(&cmd->io_cmd, &qedi_conn->active_cmd_list);
2120	cmd->io_cmd_in_list = true;
2121	qedi_conn->active_cmd_count++;
2122	spin_unlock(&qedi_conn->list_lock);
2123
2124	qedi_ring_doorbell(qedi_conn);
2125	return 0;
2126}
2127
2128int qedi_iscsi_cleanup_task(struct iscsi_task *task, bool mark_cmd_node_deleted)
2129{
2130	struct iscsi_task_params task_params;
2131	struct qedi_endpoint *ep;
2132	struct iscsi_conn *conn = task->conn;
2133	struct qedi_conn *qedi_conn = conn->dd_data;
2134	struct qedi_cmd *cmd = task->dd_data;
2135	u16 sq_idx = 0;
2136	int rval = 0;
2137
2138	QEDI_INFO(&qedi_conn->qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
2139		  "issue cleanup tid=0x%x itt=0x%x task_state=%d cmd_state=0%x cid=0x%x\n",
2140		  cmd->task_id, get_itt(task->itt), task->state,
2141		  cmd->state, qedi_conn->iscsi_conn_id);
2142
2143	memset(&task_params, 0, sizeof(task_params));
2144	ep = qedi_conn->ep;
2145
2146	sq_idx = qedi_get_wqe_idx(qedi_conn);
2147
2148	task_params.sqe = &ep->sq[sq_idx];
2149	memset(task_params.sqe, 0, sizeof(struct iscsi_wqe));
2150	task_params.itid = cmd->task_id;
2151
2152	rval = init_cleanup_task(&task_params);
2153	if (rval)
2154		return rval;
2155
2156	qedi_ring_doorbell(qedi_conn);
2157	return 0;
2158}