Linux Audio

Check our new training course

Loading...
Note: File does not exist in v3.1.
   1// SPDX-License-Identifier: GPL-2.0-only
   2/******************************************************************************
   3 *
   4 * Copyright(c) 2008 - 2014 Intel Corporation. All rights reserved.
   5 *
   6 * Contact Information:
   7 *  Intel Linux Wireless <linuxwifi@intel.com>
   8 * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
   9 *
  10 *****************************************************************************/
  11#include <linux/etherdevice.h>
  12#include <linux/kernel.h>
  13#include <linux/module.h>
  14#include <linux/sched.h>
  15#include <net/mac80211.h>
  16
  17#include "iwl-io.h"
  18#include "iwl-agn-hw.h"
  19#include "iwl-trans.h"
  20#include "iwl-modparams.h"
  21
  22#include "dev.h"
  23#include "agn.h"
  24
  25int iwlagn_hw_valid_rtc_data_addr(u32 addr)
  26{
  27	return (addr >= IWLAGN_RTC_DATA_LOWER_BOUND) &&
  28		(addr < IWLAGN_RTC_DATA_UPPER_BOUND);
  29}
  30
  31int iwlagn_send_tx_power(struct iwl_priv *priv)
  32{
  33	struct iwlagn_tx_power_dbm_cmd tx_power_cmd;
  34	u8 tx_ant_cfg_cmd;
  35
  36	if (WARN_ONCE(test_bit(STATUS_SCAN_HW, &priv->status),
  37		      "TX Power requested while scanning!\n"))
  38		return -EAGAIN;
  39
  40	/* half dBm need to multiply */
  41	tx_power_cmd.global_lmt = (s8)(2 * priv->tx_power_user_lmt);
  42
  43	if (tx_power_cmd.global_lmt > priv->nvm_data->max_tx_pwr_half_dbm) {
  44		/*
  45		 * For the newer devices which using enhanced/extend tx power
  46		 * table in EEPROM, the format is in half dBm. driver need to
  47		 * convert to dBm format before report to mac80211.
  48		 * By doing so, there is a possibility of 1/2 dBm resolution
  49		 * lost. driver will perform "round-up" operation before
  50		 * reporting, but it will cause 1/2 dBm tx power over the
  51		 * regulatory limit. Perform the checking here, if the
  52		 * "tx_power_user_lmt" is higher than EEPROM value (in
  53		 * half-dBm format), lower the tx power based on EEPROM
  54		 */
  55		tx_power_cmd.global_lmt =
  56			priv->nvm_data->max_tx_pwr_half_dbm;
  57	}
  58	tx_power_cmd.flags = IWLAGN_TX_POWER_NO_CLOSED;
  59	tx_power_cmd.srv_chan_lmt = IWLAGN_TX_POWER_AUTO;
  60
  61	if (IWL_UCODE_API(priv->fw->ucode_ver) == 1)
  62		tx_ant_cfg_cmd = REPLY_TX_POWER_DBM_CMD_V1;
  63	else
  64		tx_ant_cfg_cmd = REPLY_TX_POWER_DBM_CMD;
  65
  66	return iwl_dvm_send_cmd_pdu(priv, tx_ant_cfg_cmd, 0,
  67			sizeof(tx_power_cmd), &tx_power_cmd);
  68}
  69
  70void iwlagn_temperature(struct iwl_priv *priv)
  71{
  72	lockdep_assert_held(&priv->statistics.lock);
  73
  74	/* store temperature from correct statistics (in Celsius) */
  75	priv->temperature = le32_to_cpu(priv->statistics.common.temperature);
  76	iwl_tt_handler(priv);
  77}
  78
  79int iwlagn_hwrate_to_mac80211_idx(u32 rate_n_flags, enum nl80211_band band)
  80{
  81	int idx = 0;
  82	int band_offset = 0;
  83
  84	/* HT rate format: mac80211 wants an MCS number, which is just LSB */
  85	if (rate_n_flags & RATE_MCS_HT_MSK) {
  86		idx = (rate_n_flags & 0xff);
  87		return idx;
  88	/* Legacy rate format, search for match in table */
  89	} else {
  90		if (band == NL80211_BAND_5GHZ)
  91			band_offset = IWL_FIRST_OFDM_RATE;
  92		for (idx = band_offset; idx < IWL_RATE_COUNT_LEGACY; idx++)
  93			if (iwl_rates[idx].plcp == (rate_n_flags & 0xFF))
  94				return idx - band_offset;
  95	}
  96
  97	return -1;
  98}
  99
 100int iwlagn_manage_ibss_station(struct iwl_priv *priv,
 101			       struct ieee80211_vif *vif, bool add)
 102{
 103	struct iwl_vif_priv *vif_priv = (void *)vif->drv_priv;
 104
 105	if (add)
 106		return iwlagn_add_bssid_station(priv, vif_priv->ctx,
 107						vif->bss_conf.bssid,
 108						&vif_priv->ibss_bssid_sta_id);
 109	return iwl_remove_station(priv, vif_priv->ibss_bssid_sta_id,
 110				  vif->bss_conf.bssid);
 111}
 112
 113/**
 114 * iwlagn_txfifo_flush: send REPLY_TXFIFO_FLUSH command to uCode
 115 *
 116 * pre-requirements:
 117 *  1. acquire mutex before calling
 118 *  2. make sure rf is on and not in exit state
 119 */
 120int iwlagn_txfifo_flush(struct iwl_priv *priv, u32 scd_q_msk)
 121{
 122	struct iwl_txfifo_flush_cmd_v3 flush_cmd_v3 = {
 123		.flush_control = cpu_to_le16(IWL_DROP_ALL),
 124	};
 125	struct iwl_txfifo_flush_cmd_v2 flush_cmd_v2 = {
 126		.flush_control = cpu_to_le16(IWL_DROP_ALL),
 127	};
 128
 129	u32 queue_control = IWL_SCD_VO_MSK | IWL_SCD_VI_MSK |
 130			    IWL_SCD_BE_MSK | IWL_SCD_BK_MSK | IWL_SCD_MGMT_MSK;
 131
 132	if ((priv->valid_contexts != BIT(IWL_RXON_CTX_BSS)))
 133		queue_control |= IWL_PAN_SCD_VO_MSK | IWL_PAN_SCD_VI_MSK |
 134				 IWL_PAN_SCD_BE_MSK | IWL_PAN_SCD_BK_MSK |
 135				 IWL_PAN_SCD_MGMT_MSK |
 136				 IWL_PAN_SCD_MULTICAST_MSK;
 137
 138	if (priv->nvm_data->sku_cap_11n_enable)
 139		queue_control |= IWL_AGG_TX_QUEUE_MSK;
 140
 141	if (scd_q_msk)
 142		queue_control = scd_q_msk;
 143
 144	IWL_DEBUG_INFO(priv, "queue control: 0x%x\n", queue_control);
 145	flush_cmd_v3.queue_control = cpu_to_le32(queue_control);
 146	flush_cmd_v2.queue_control = cpu_to_le16((u16)queue_control);
 147
 148	if (IWL_UCODE_API(priv->fw->ucode_ver) > 2)
 149		return iwl_dvm_send_cmd_pdu(priv, REPLY_TXFIFO_FLUSH, 0,
 150					    sizeof(flush_cmd_v3),
 151					    &flush_cmd_v3);
 152	return iwl_dvm_send_cmd_pdu(priv, REPLY_TXFIFO_FLUSH, 0,
 153				    sizeof(flush_cmd_v2), &flush_cmd_v2);
 154}
 155
 156void iwlagn_dev_txfifo_flush(struct iwl_priv *priv)
 157{
 158	mutex_lock(&priv->mutex);
 159	ieee80211_stop_queues(priv->hw);
 160	if (iwlagn_txfifo_flush(priv, 0)) {
 161		IWL_ERR(priv, "flush request fail\n");
 162		goto done;
 163	}
 164	IWL_DEBUG_INFO(priv, "wait transmit/flush all frames\n");
 165	iwl_trans_wait_tx_queues_empty(priv->trans, 0xffffffff);
 166done:
 167	ieee80211_wake_queues(priv->hw);
 168	mutex_unlock(&priv->mutex);
 169}
 170
 171/*
 172 * BT coex
 173 */
 174/* Notmal TDM */
 175static const __le32 iwlagn_def_3w_lookup[IWLAGN_BT_DECISION_LUT_SIZE] = {
 176	cpu_to_le32(0xaaaaaaaa),
 177	cpu_to_le32(0xaaaaaaaa),
 178	cpu_to_le32(0xaeaaaaaa),
 179	cpu_to_le32(0xaaaaaaaa),
 180	cpu_to_le32(0xcc00ff28),
 181	cpu_to_le32(0x0000aaaa),
 182	cpu_to_le32(0xcc00aaaa),
 183	cpu_to_le32(0x0000aaaa),
 184	cpu_to_le32(0xc0004000),
 185	cpu_to_le32(0x00004000),
 186	cpu_to_le32(0xf0005000),
 187	cpu_to_le32(0xf0005000),
 188};
 189
 190/* Full concurrency */
 191static const __le32 iwlagn_concurrent_lookup[IWLAGN_BT_DECISION_LUT_SIZE] = {
 192	cpu_to_le32(0xaaaaaaaa),
 193	cpu_to_le32(0xaaaaaaaa),
 194	cpu_to_le32(0xaaaaaaaa),
 195	cpu_to_le32(0xaaaaaaaa),
 196	cpu_to_le32(0xaaaaaaaa),
 197	cpu_to_le32(0xaaaaaaaa),
 198	cpu_to_le32(0xaaaaaaaa),
 199	cpu_to_le32(0xaaaaaaaa),
 200	cpu_to_le32(0x00000000),
 201	cpu_to_le32(0x00000000),
 202	cpu_to_le32(0x00000000),
 203	cpu_to_le32(0x00000000),
 204};
 205
 206void iwlagn_send_advance_bt_config(struct iwl_priv *priv)
 207{
 208	struct iwl_basic_bt_cmd basic = {
 209		.max_kill = IWLAGN_BT_MAX_KILL_DEFAULT,
 210		.bt3_timer_t7_value = IWLAGN_BT3_T7_DEFAULT,
 211		.bt3_prio_sample_time = IWLAGN_BT3_PRIO_SAMPLE_DEFAULT,
 212		.bt3_timer_t2_value = IWLAGN_BT3_T2_DEFAULT,
 213	};
 214	struct iwl_bt_cmd_v1 bt_cmd_v1;
 215	struct iwl_bt_cmd_v2 bt_cmd_v2;
 216	int ret;
 217
 218	BUILD_BUG_ON(sizeof(iwlagn_def_3w_lookup) !=
 219			sizeof(basic.bt3_lookup_table));
 220
 221	if (priv->lib->bt_params) {
 222		/*
 223		 * newer generation of devices (2000 series and newer)
 224		 * use the version 2 of the bt command
 225		 * we need to make sure sending the host command
 226		 * with correct data structure to avoid uCode assert
 227		 */
 228		if (priv->lib->bt_params->bt_session_2) {
 229			bt_cmd_v2.prio_boost = cpu_to_le32(
 230				priv->lib->bt_params->bt_prio_boost);
 231			bt_cmd_v2.tx_prio_boost = 0;
 232			bt_cmd_v2.rx_prio_boost = 0;
 233		} else {
 234			/* older version only has 8 bits */
 235			WARN_ON(priv->lib->bt_params->bt_prio_boost & ~0xFF);
 236			bt_cmd_v1.prio_boost =
 237				priv->lib->bt_params->bt_prio_boost;
 238			bt_cmd_v1.tx_prio_boost = 0;
 239			bt_cmd_v1.rx_prio_boost = 0;
 240		}
 241	} else {
 242		IWL_ERR(priv, "failed to construct BT Coex Config\n");
 243		return;
 244	}
 245
 246	/*
 247	 * Possible situations when BT needs to take over for receive,
 248	 * at the same time where STA needs to response to AP's frame(s),
 249	 * reduce the tx power of the required response frames, by that,
 250	 * allow the concurrent BT receive & WiFi transmit
 251	 * (BT - ANT A, WiFi -ANT B), without interference to one another
 252	 *
 253	 * Reduced tx power apply to control frames only (ACK/Back/CTS)
 254	 * when indicated by the BT config command
 255	 */
 256	basic.kill_ack_mask = priv->kill_ack_mask;
 257	basic.kill_cts_mask = priv->kill_cts_mask;
 258	if (priv->reduced_txpower)
 259		basic.reduce_txpower = IWLAGN_BT_REDUCED_TX_PWR;
 260	basic.valid = priv->bt_valid;
 261
 262	/*
 263	 * Configure BT coex mode to "no coexistence" when the
 264	 * user disabled BT coexistence, we have no interface
 265	 * (might be in monitor mode), or the interface is in
 266	 * IBSS mode (no proper uCode support for coex then).
 267	 */
 268	if (!iwlwifi_mod_params.bt_coex_active ||
 269	    priv->iw_mode == NL80211_IFTYPE_ADHOC) {
 270		basic.flags = IWLAGN_BT_FLAG_COEX_MODE_DISABLED;
 271	} else {
 272		basic.flags = IWLAGN_BT_FLAG_COEX_MODE_3W <<
 273					IWLAGN_BT_FLAG_COEX_MODE_SHIFT;
 274
 275		if (!priv->bt_enable_pspoll)
 276			basic.flags |= IWLAGN_BT_FLAG_SYNC_2_BT_DISABLE;
 277		else
 278			basic.flags &= ~IWLAGN_BT_FLAG_SYNC_2_BT_DISABLE;
 279
 280		if (priv->bt_ch_announce)
 281			basic.flags |= IWLAGN_BT_FLAG_CHANNEL_INHIBITION;
 282		IWL_DEBUG_COEX(priv, "BT coex flag: 0X%x\n", basic.flags);
 283	}
 284	priv->bt_enable_flag = basic.flags;
 285	if (priv->bt_full_concurrent)
 286		memcpy(basic.bt3_lookup_table, iwlagn_concurrent_lookup,
 287			sizeof(iwlagn_concurrent_lookup));
 288	else
 289		memcpy(basic.bt3_lookup_table, iwlagn_def_3w_lookup,
 290			sizeof(iwlagn_def_3w_lookup));
 291
 292	IWL_DEBUG_COEX(priv, "BT coex %s in %s mode\n",
 293		       basic.flags ? "active" : "disabled",
 294		       priv->bt_full_concurrent ?
 295		       "full concurrency" : "3-wire");
 296
 297	if (priv->lib->bt_params->bt_session_2) {
 298		memcpy(&bt_cmd_v2.basic, &basic,
 299			sizeof(basic));
 300		ret = iwl_dvm_send_cmd_pdu(priv, REPLY_BT_CONFIG,
 301			0, sizeof(bt_cmd_v2), &bt_cmd_v2);
 302	} else {
 303		memcpy(&bt_cmd_v1.basic, &basic,
 304			sizeof(basic));
 305		ret = iwl_dvm_send_cmd_pdu(priv, REPLY_BT_CONFIG,
 306			0, sizeof(bt_cmd_v1), &bt_cmd_v1);
 307	}
 308	if (ret)
 309		IWL_ERR(priv, "failed to send BT Coex Config\n");
 310
 311}
 312
 313void iwlagn_bt_adjust_rssi_monitor(struct iwl_priv *priv, bool rssi_ena)
 314{
 315	struct iwl_rxon_context *ctx, *found_ctx = NULL;
 316	bool found_ap = false;
 317
 318	lockdep_assert_held(&priv->mutex);
 319
 320	/* Check whether AP or GO mode is active. */
 321	if (rssi_ena) {
 322		for_each_context(priv, ctx) {
 323			if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_AP &&
 324			    iwl_is_associated_ctx(ctx)) {
 325				found_ap = true;
 326				break;
 327			}
 328		}
 329	}
 330
 331	/*
 332	 * If disable was received or If GO/AP mode, disable RSSI
 333	 * measurements.
 334	 */
 335	if (!rssi_ena || found_ap) {
 336		if (priv->cur_rssi_ctx) {
 337			ctx = priv->cur_rssi_ctx;
 338			ieee80211_disable_rssi_reports(ctx->vif);
 339			priv->cur_rssi_ctx = NULL;
 340		}
 341		return;
 342	}
 343
 344	/*
 345	 * If rssi measurements need to be enabled, consider all cases now.
 346	 * Figure out how many contexts are active.
 347	 */
 348	for_each_context(priv, ctx) {
 349		if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_STATION &&
 350		    iwl_is_associated_ctx(ctx)) {
 351			found_ctx = ctx;
 352			break;
 353		}
 354	}
 355
 356	/*
 357	 * rssi monitor already enabled for the correct interface...nothing
 358	 * to do.
 359	 */
 360	if (found_ctx == priv->cur_rssi_ctx)
 361		return;
 362
 363	/*
 364	 * Figure out if rssi monitor is currently enabled, and needs
 365	 * to be changed. If rssi monitor is already enabled, disable
 366	 * it first else just enable rssi measurements on the
 367	 * interface found above.
 368	 */
 369	if (priv->cur_rssi_ctx) {
 370		ctx = priv->cur_rssi_ctx;
 371		if (ctx->vif)
 372			ieee80211_disable_rssi_reports(ctx->vif);
 373	}
 374
 375	priv->cur_rssi_ctx = found_ctx;
 376
 377	if (!found_ctx)
 378		return;
 379
 380	ieee80211_enable_rssi_reports(found_ctx->vif,
 381			IWLAGN_BT_PSP_MIN_RSSI_THRESHOLD,
 382			IWLAGN_BT_PSP_MAX_RSSI_THRESHOLD);
 383}
 384
 385static bool iwlagn_bt_traffic_is_sco(struct iwl_bt_uart_msg *uart_msg)
 386{
 387	return (BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3) >>
 388		BT_UART_MSG_FRAME3SCOESCO_POS;
 389}
 390
 391static void iwlagn_bt_traffic_change_work(struct work_struct *work)
 392{
 393	struct iwl_priv *priv =
 394		container_of(work, struct iwl_priv, bt_traffic_change_work);
 395	struct iwl_rxon_context *ctx;
 396	int smps_request = -1;
 397
 398	if (priv->bt_enable_flag == IWLAGN_BT_FLAG_COEX_MODE_DISABLED) {
 399		/* bt coex disabled */
 400		return;
 401	}
 402
 403	/*
 404	 * Note: bt_traffic_load can be overridden by scan complete and
 405	 * coex profile notifications. Ignore that since only bad consequence
 406	 * can be not matching debug print with actual state.
 407	 */
 408	IWL_DEBUG_COEX(priv, "BT traffic load changes: %d\n",
 409		       priv->bt_traffic_load);
 410
 411	switch (priv->bt_traffic_load) {
 412	case IWL_BT_COEX_TRAFFIC_LOAD_NONE:
 413		if (priv->bt_status)
 414			smps_request = IEEE80211_SMPS_DYNAMIC;
 415		else
 416			smps_request = IEEE80211_SMPS_AUTOMATIC;
 417		break;
 418	case IWL_BT_COEX_TRAFFIC_LOAD_LOW:
 419		smps_request = IEEE80211_SMPS_DYNAMIC;
 420		break;
 421	case IWL_BT_COEX_TRAFFIC_LOAD_HIGH:
 422	case IWL_BT_COEX_TRAFFIC_LOAD_CONTINUOUS:
 423		smps_request = IEEE80211_SMPS_STATIC;
 424		break;
 425	default:
 426		IWL_ERR(priv, "Invalid BT traffic load: %d\n",
 427			priv->bt_traffic_load);
 428		break;
 429	}
 430
 431	mutex_lock(&priv->mutex);
 432
 433	/*
 434	 * We can not send command to firmware while scanning. When the scan
 435	 * complete we will schedule this work again. We do check with mutex
 436	 * locked to prevent new scan request to arrive. We do not check
 437	 * STATUS_SCANNING to avoid race when queue_work two times from
 438	 * different notifications, but quit and not perform any work at all.
 439	 */
 440	if (test_bit(STATUS_SCAN_HW, &priv->status))
 441		goto out;
 442
 443	iwl_update_chain_flags(priv);
 444
 445	if (smps_request != -1) {
 446		priv->current_ht_config.smps = smps_request;
 447		for_each_context(priv, ctx) {
 448			if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_STATION)
 449				ieee80211_request_smps(ctx->vif, smps_request);
 450		}
 451	}
 452
 453	/*
 454	 * Dynamic PS poll related functionality. Adjust RSSI measurements if
 455	 * necessary.
 456	 */
 457	iwlagn_bt_coex_rssi_monitor(priv);
 458out:
 459	mutex_unlock(&priv->mutex);
 460}
 461
 462/*
 463 * If BT sco traffic, and RSSI monitor is enabled, move measurements to the
 464 * correct interface or disable it if this is the last interface to be
 465 * removed.
 466 */
 467void iwlagn_bt_coex_rssi_monitor(struct iwl_priv *priv)
 468{
 469	if (priv->bt_is_sco &&
 470	    priv->bt_traffic_load == IWL_BT_COEX_TRAFFIC_LOAD_CONTINUOUS)
 471		iwlagn_bt_adjust_rssi_monitor(priv, true);
 472	else
 473		iwlagn_bt_adjust_rssi_monitor(priv, false);
 474}
 475
 476static void iwlagn_print_uartmsg(struct iwl_priv *priv,
 477				struct iwl_bt_uart_msg *uart_msg)
 478{
 479	IWL_DEBUG_COEX(priv, "Message Type = 0x%X, SSN = 0x%X, "
 480			"Update Req = 0x%X\n",
 481		(BT_UART_MSG_FRAME1MSGTYPE_MSK & uart_msg->frame1) >>
 482			BT_UART_MSG_FRAME1MSGTYPE_POS,
 483		(BT_UART_MSG_FRAME1SSN_MSK & uart_msg->frame1) >>
 484			BT_UART_MSG_FRAME1SSN_POS,
 485		(BT_UART_MSG_FRAME1UPDATEREQ_MSK & uart_msg->frame1) >>
 486			BT_UART_MSG_FRAME1UPDATEREQ_POS);
 487
 488	IWL_DEBUG_COEX(priv, "Open connections = 0x%X, Traffic load = 0x%X, "
 489			"Chl_SeqN = 0x%X, In band = 0x%X\n",
 490		(BT_UART_MSG_FRAME2OPENCONNECTIONS_MSK & uart_msg->frame2) >>
 491			BT_UART_MSG_FRAME2OPENCONNECTIONS_POS,
 492		(BT_UART_MSG_FRAME2TRAFFICLOAD_MSK & uart_msg->frame2) >>
 493			BT_UART_MSG_FRAME2TRAFFICLOAD_POS,
 494		(BT_UART_MSG_FRAME2CHLSEQN_MSK & uart_msg->frame2) >>
 495			BT_UART_MSG_FRAME2CHLSEQN_POS,
 496		(BT_UART_MSG_FRAME2INBAND_MSK & uart_msg->frame2) >>
 497			BT_UART_MSG_FRAME2INBAND_POS);
 498
 499	IWL_DEBUG_COEX(priv, "SCO/eSCO = 0x%X, Sniff = 0x%X, A2DP = 0x%X, "
 500			"ACL = 0x%X, Master = 0x%X, OBEX = 0x%X\n",
 501		(BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3) >>
 502			BT_UART_MSG_FRAME3SCOESCO_POS,
 503		(BT_UART_MSG_FRAME3SNIFF_MSK & uart_msg->frame3) >>
 504			BT_UART_MSG_FRAME3SNIFF_POS,
 505		(BT_UART_MSG_FRAME3A2DP_MSK & uart_msg->frame3) >>
 506			BT_UART_MSG_FRAME3A2DP_POS,
 507		(BT_UART_MSG_FRAME3ACL_MSK & uart_msg->frame3) >>
 508			BT_UART_MSG_FRAME3ACL_POS,
 509		(BT_UART_MSG_FRAME3MASTER_MSK & uart_msg->frame3) >>
 510			BT_UART_MSG_FRAME3MASTER_POS,
 511		(BT_UART_MSG_FRAME3OBEX_MSK & uart_msg->frame3) >>
 512			BT_UART_MSG_FRAME3OBEX_POS);
 513
 514	IWL_DEBUG_COEX(priv, "Idle duration = 0x%X\n",
 515		(BT_UART_MSG_FRAME4IDLEDURATION_MSK & uart_msg->frame4) >>
 516			BT_UART_MSG_FRAME4IDLEDURATION_POS);
 517
 518	IWL_DEBUG_COEX(priv, "Tx Activity = 0x%X, Rx Activity = 0x%X, "
 519			"eSCO Retransmissions = 0x%X\n",
 520		(BT_UART_MSG_FRAME5TXACTIVITY_MSK & uart_msg->frame5) >>
 521			BT_UART_MSG_FRAME5TXACTIVITY_POS,
 522		(BT_UART_MSG_FRAME5RXACTIVITY_MSK & uart_msg->frame5) >>
 523			BT_UART_MSG_FRAME5RXACTIVITY_POS,
 524		(BT_UART_MSG_FRAME5ESCORETRANSMIT_MSK & uart_msg->frame5) >>
 525			BT_UART_MSG_FRAME5ESCORETRANSMIT_POS);
 526
 527	IWL_DEBUG_COEX(priv, "Sniff Interval = 0x%X, Discoverable = 0x%X\n",
 528		(BT_UART_MSG_FRAME6SNIFFINTERVAL_MSK & uart_msg->frame6) >>
 529			BT_UART_MSG_FRAME6SNIFFINTERVAL_POS,
 530		(BT_UART_MSG_FRAME6DISCOVERABLE_MSK & uart_msg->frame6) >>
 531			BT_UART_MSG_FRAME6DISCOVERABLE_POS);
 532
 533	IWL_DEBUG_COEX(priv, "Sniff Activity = 0x%X, Page = "
 534			"0x%X, Inquiry = 0x%X, Connectable = 0x%X\n",
 535		(BT_UART_MSG_FRAME7SNIFFACTIVITY_MSK & uart_msg->frame7) >>
 536			BT_UART_MSG_FRAME7SNIFFACTIVITY_POS,
 537		(BT_UART_MSG_FRAME7PAGE_MSK & uart_msg->frame7) >>
 538			BT_UART_MSG_FRAME7PAGE_POS,
 539		(BT_UART_MSG_FRAME7INQUIRY_MSK & uart_msg->frame7) >>
 540			BT_UART_MSG_FRAME7INQUIRY_POS,
 541		(BT_UART_MSG_FRAME7CONNECTABLE_MSK & uart_msg->frame7) >>
 542			BT_UART_MSG_FRAME7CONNECTABLE_POS);
 543}
 544
 545static bool iwlagn_set_kill_msk(struct iwl_priv *priv,
 546				struct iwl_bt_uart_msg *uart_msg)
 547{
 548	bool need_update = false;
 549	u8 kill_msk = IWL_BT_KILL_REDUCE;
 550	static const __le32 bt_kill_ack_msg[3] = {
 551		IWLAGN_BT_KILL_ACK_MASK_DEFAULT,
 552		IWLAGN_BT_KILL_ACK_CTS_MASK_SCO,
 553		IWLAGN_BT_KILL_ACK_CTS_MASK_REDUCE};
 554	static const __le32 bt_kill_cts_msg[3] = {
 555		IWLAGN_BT_KILL_CTS_MASK_DEFAULT,
 556		IWLAGN_BT_KILL_ACK_CTS_MASK_SCO,
 557		IWLAGN_BT_KILL_ACK_CTS_MASK_REDUCE};
 558
 559	if (!priv->reduced_txpower)
 560		kill_msk = (BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3)
 561			? IWL_BT_KILL_OVERRIDE : IWL_BT_KILL_DEFAULT;
 562	if (priv->kill_ack_mask != bt_kill_ack_msg[kill_msk] ||
 563	    priv->kill_cts_mask != bt_kill_cts_msg[kill_msk]) {
 564		priv->bt_valid |= IWLAGN_BT_VALID_KILL_ACK_MASK;
 565		priv->kill_ack_mask = bt_kill_ack_msg[kill_msk];
 566		priv->bt_valid |= IWLAGN_BT_VALID_KILL_CTS_MASK;
 567		priv->kill_cts_mask = bt_kill_cts_msg[kill_msk];
 568		need_update = true;
 569	}
 570	return need_update;
 571}
 572
 573/*
 574 * Upon RSSI changes, sends a bt config command with following changes
 575 *  1. enable/disable "reduced control frames tx power
 576 *  2. update the "kill)ack_mask" and "kill_cts_mask"
 577 *
 578 * If "reduced tx power" is enabled, uCode shall
 579 *  1. ACK/Back/CTS rate shall reduced to 6Mbps
 580 *  2. not use duplciate 20/40MHz mode
 581 */
 582static bool iwlagn_fill_txpower_mode(struct iwl_priv *priv,
 583				struct iwl_bt_uart_msg *uart_msg)
 584{
 585	bool need_update = false;
 586	struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS];
 587	int ave_rssi;
 588
 589	if (!ctx->vif || (ctx->vif->type != NL80211_IFTYPE_STATION)) {
 590		IWL_DEBUG_INFO(priv, "BSS ctx not active or not in sta mode\n");
 591		return false;
 592	}
 593
 594	ave_rssi = ieee80211_ave_rssi(ctx->vif);
 595	if (!ave_rssi) {
 596		/* no rssi data, no changes to reduce tx power */
 597		IWL_DEBUG_COEX(priv, "no rssi data available\n");
 598		return need_update;
 599	}
 600	if (!priv->reduced_txpower &&
 601	    !iwl_is_associated(priv, IWL_RXON_CTX_PAN) &&
 602	    (ave_rssi > BT_ENABLE_REDUCED_TXPOWER_THRESHOLD) &&
 603	    (uart_msg->frame3 & (BT_UART_MSG_FRAME3ACL_MSK |
 604	    BT_UART_MSG_FRAME3OBEX_MSK)) &&
 605	    !(uart_msg->frame3 & (BT_UART_MSG_FRAME3SCOESCO_MSK |
 606	    BT_UART_MSG_FRAME3SNIFF_MSK | BT_UART_MSG_FRAME3A2DP_MSK))) {
 607		/* enabling reduced tx power */
 608		priv->reduced_txpower = true;
 609		priv->bt_valid |= IWLAGN_BT_VALID_REDUCED_TX_PWR;
 610		need_update = true;
 611	} else if (priv->reduced_txpower &&
 612		   (iwl_is_associated(priv, IWL_RXON_CTX_PAN) ||
 613		   (ave_rssi < BT_DISABLE_REDUCED_TXPOWER_THRESHOLD) ||
 614		   (uart_msg->frame3 & (BT_UART_MSG_FRAME3SCOESCO_MSK |
 615		   BT_UART_MSG_FRAME3SNIFF_MSK | BT_UART_MSG_FRAME3A2DP_MSK)) ||
 616		   !(uart_msg->frame3 & (BT_UART_MSG_FRAME3ACL_MSK |
 617		   BT_UART_MSG_FRAME3OBEX_MSK)))) {
 618		/* disable reduced tx power */
 619		priv->reduced_txpower = false;
 620		priv->bt_valid |= IWLAGN_BT_VALID_REDUCED_TX_PWR;
 621		need_update = true;
 622	}
 623
 624	return need_update;
 625}
 626
 627static void iwlagn_bt_coex_profile_notif(struct iwl_priv *priv,
 628					 struct iwl_rx_cmd_buffer *rxb)
 629{
 630	struct iwl_rx_packet *pkt = rxb_addr(rxb);
 631	struct iwl_bt_coex_profile_notif *coex = (void *)pkt->data;
 632	struct iwl_bt_uart_msg *uart_msg = &coex->last_bt_uart_msg;
 633
 634	if (priv->bt_enable_flag == IWLAGN_BT_FLAG_COEX_MODE_DISABLED) {
 635		/* bt coex disabled */
 636		return;
 637	}
 638
 639	IWL_DEBUG_COEX(priv, "BT Coex notification:\n");
 640	IWL_DEBUG_COEX(priv, "    status: %d\n", coex->bt_status);
 641	IWL_DEBUG_COEX(priv, "    traffic load: %d\n", coex->bt_traffic_load);
 642	IWL_DEBUG_COEX(priv, "    CI compliance: %d\n",
 643			coex->bt_ci_compliance);
 644	iwlagn_print_uartmsg(priv, uart_msg);
 645
 646	priv->last_bt_traffic_load = priv->bt_traffic_load;
 647	priv->bt_is_sco = iwlagn_bt_traffic_is_sco(uart_msg);
 648
 649	if (priv->iw_mode != NL80211_IFTYPE_ADHOC) {
 650		if (priv->bt_status != coex->bt_status ||
 651		    priv->last_bt_traffic_load != coex->bt_traffic_load) {
 652			if (coex->bt_status) {
 653				/* BT on */
 654				if (!priv->bt_ch_announce)
 655					priv->bt_traffic_load =
 656						IWL_BT_COEX_TRAFFIC_LOAD_HIGH;
 657				else
 658					priv->bt_traffic_load =
 659						coex->bt_traffic_load;
 660			} else {
 661				/* BT off */
 662				priv->bt_traffic_load =
 663					IWL_BT_COEX_TRAFFIC_LOAD_NONE;
 664			}
 665			priv->bt_status = coex->bt_status;
 666			queue_work(priv->workqueue,
 667				   &priv->bt_traffic_change_work);
 668		}
 669	}
 670
 671	/* schedule to send runtime bt_config */
 672	/* check reduce power before change ack/cts kill mask */
 673	if (iwlagn_fill_txpower_mode(priv, uart_msg) ||
 674	    iwlagn_set_kill_msk(priv, uart_msg))
 675		queue_work(priv->workqueue, &priv->bt_runtime_config);
 676
 677
 678	/* FIXME: based on notification, adjust the prio_boost */
 679
 680	priv->bt_ci_compliance = coex->bt_ci_compliance;
 681}
 682
 683void iwlagn_bt_rx_handler_setup(struct iwl_priv *priv)
 684{
 685	priv->rx_handlers[REPLY_BT_COEX_PROFILE_NOTIF] =
 686		iwlagn_bt_coex_profile_notif;
 687}
 688
 689void iwlagn_bt_setup_deferred_work(struct iwl_priv *priv)
 690{
 691	INIT_WORK(&priv->bt_traffic_change_work,
 692		  iwlagn_bt_traffic_change_work);
 693}
 694
 695void iwlagn_bt_cancel_deferred_work(struct iwl_priv *priv)
 696{
 697	cancel_work_sync(&priv->bt_traffic_change_work);
 698}
 699
 700static bool is_single_rx_stream(struct iwl_priv *priv)
 701{
 702	return priv->current_ht_config.smps == IEEE80211_SMPS_STATIC ||
 703	       priv->current_ht_config.single_chain_sufficient;
 704}
 705
 706#define IWL_NUM_RX_CHAINS_MULTIPLE	3
 707#define IWL_NUM_RX_CHAINS_SINGLE	2
 708#define IWL_NUM_IDLE_CHAINS_DUAL	2
 709#define IWL_NUM_IDLE_CHAINS_SINGLE	1
 710
 711/*
 712 * Determine how many receiver/antenna chains to use.
 713 *
 714 * More provides better reception via diversity.  Fewer saves power
 715 * at the expense of throughput, but only when not in powersave to
 716 * start with.
 717 *
 718 * MIMO (dual stream) requires at least 2, but works better with 3.
 719 * This does not determine *which* chains to use, just how many.
 720 */
 721static int iwl_get_active_rx_chain_count(struct iwl_priv *priv)
 722{
 723	if (priv->lib->bt_params &&
 724	    priv->lib->bt_params->advanced_bt_coexist &&
 725	    (priv->bt_full_concurrent ||
 726	     priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH)) {
 727		/*
 728		 * only use chain 'A' in bt high traffic load or
 729		 * full concurrency mode
 730		 */
 731		return IWL_NUM_RX_CHAINS_SINGLE;
 732	}
 733	/* # of Rx chains to use when expecting MIMO. */
 734	if (is_single_rx_stream(priv))
 735		return IWL_NUM_RX_CHAINS_SINGLE;
 736	else
 737		return IWL_NUM_RX_CHAINS_MULTIPLE;
 738}
 739
 740/*
 741 * When we are in power saving mode, unless device support spatial
 742 * multiplexing power save, use the active count for rx chain count.
 743 */
 744static int iwl_get_idle_rx_chain_count(struct iwl_priv *priv, int active_cnt)
 745{
 746	/* # Rx chains when idling, depending on SMPS mode */
 747	switch (priv->current_ht_config.smps) {
 748	case IEEE80211_SMPS_STATIC:
 749	case IEEE80211_SMPS_DYNAMIC:
 750		return IWL_NUM_IDLE_CHAINS_SINGLE;
 751	case IEEE80211_SMPS_AUTOMATIC:
 752	case IEEE80211_SMPS_OFF:
 753		return active_cnt;
 754	default:
 755		WARN(1, "invalid SMPS mode %d",
 756		     priv->current_ht_config.smps);
 757		return active_cnt;
 758	}
 759}
 760
 761/* up to 4 chains */
 762static u8 iwl_count_chain_bitmap(u32 chain_bitmap)
 763{
 764	u8 res;
 765	res = (chain_bitmap & BIT(0)) >> 0;
 766	res += (chain_bitmap & BIT(1)) >> 1;
 767	res += (chain_bitmap & BIT(2)) >> 2;
 768	res += (chain_bitmap & BIT(3)) >> 3;
 769	return res;
 770}
 771
 772/**
 773 * iwlagn_set_rxon_chain - Set up Rx chain usage in "staging" RXON image
 774 *
 775 * Selects how many and which Rx receivers/antennas/chains to use.
 776 * This should not be used for scan command ... it puts data in wrong place.
 777 */
 778void iwlagn_set_rxon_chain(struct iwl_priv *priv, struct iwl_rxon_context *ctx)
 779{
 780	bool is_single = is_single_rx_stream(priv);
 781	bool is_cam = !test_bit(STATUS_POWER_PMI, &priv->status);
 782	u8 idle_rx_cnt, active_rx_cnt, valid_rx_cnt;
 783	u32 active_chains;
 784	u16 rx_chain;
 785
 786	/* Tell uCode which antennas are actually connected.
 787	 * Before first association, we assume all antennas are connected.
 788	 * Just after first association, iwl_chain_noise_calibration()
 789	 *    checks which antennas actually *are* connected. */
 790	if (priv->chain_noise_data.active_chains)
 791		active_chains = priv->chain_noise_data.active_chains;
 792	else
 793		active_chains = priv->nvm_data->valid_rx_ant;
 794
 795	if (priv->lib->bt_params &&
 796	    priv->lib->bt_params->advanced_bt_coexist &&
 797	    (priv->bt_full_concurrent ||
 798	     priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH)) {
 799		/*
 800		 * only use chain 'A' in bt high traffic load or
 801		 * full concurrency mode
 802		 */
 803		active_chains = first_antenna(active_chains);
 804	}
 805
 806	rx_chain = active_chains << RXON_RX_CHAIN_VALID_POS;
 807
 808	/* How many receivers should we use? */
 809	active_rx_cnt = iwl_get_active_rx_chain_count(priv);
 810	idle_rx_cnt = iwl_get_idle_rx_chain_count(priv, active_rx_cnt);
 811
 812
 813	/* correct rx chain count according hw settings
 814	 * and chain noise calibration
 815	 */
 816	valid_rx_cnt = iwl_count_chain_bitmap(active_chains);
 817	if (valid_rx_cnt < active_rx_cnt)
 818		active_rx_cnt = valid_rx_cnt;
 819
 820	if (valid_rx_cnt < idle_rx_cnt)
 821		idle_rx_cnt = valid_rx_cnt;
 822
 823	rx_chain |= active_rx_cnt << RXON_RX_CHAIN_MIMO_CNT_POS;
 824	rx_chain |= idle_rx_cnt  << RXON_RX_CHAIN_CNT_POS;
 825
 826	ctx->staging.rx_chain = cpu_to_le16(rx_chain);
 827
 828	if (!is_single && (active_rx_cnt >= IWL_NUM_RX_CHAINS_SINGLE) && is_cam)
 829		ctx->staging.rx_chain |= RXON_RX_CHAIN_MIMO_FORCE_MSK;
 830	else
 831		ctx->staging.rx_chain &= ~RXON_RX_CHAIN_MIMO_FORCE_MSK;
 832
 833	IWL_DEBUG_ASSOC(priv, "rx_chain=0x%X active=%d idle=%d\n",
 834			ctx->staging.rx_chain,
 835			active_rx_cnt, idle_rx_cnt);
 836
 837	WARN_ON(active_rx_cnt == 0 || idle_rx_cnt == 0 ||
 838		active_rx_cnt < idle_rx_cnt);
 839}
 840
 841u8 iwl_toggle_tx_ant(struct iwl_priv *priv, u8 ant, u8 valid)
 842{
 843	int i;
 844	u8 ind = ant;
 845
 846	if (priv->band == NL80211_BAND_2GHZ &&
 847	    priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH)
 848		return 0;
 849
 850	for (i = 0; i < RATE_ANT_NUM - 1; i++) {
 851		ind = (ind + 1) < RATE_ANT_NUM ?  ind + 1 : 0;
 852		if (valid & BIT(ind))
 853			return ind;
 854	}
 855	return ant;
 856}
 857
 858#ifdef CONFIG_PM_SLEEP
 859static void iwlagn_convert_p1k(u16 *p1k, __le16 *out)
 860{
 861	int i;
 862
 863	for (i = 0; i < IWLAGN_P1K_SIZE; i++)
 864		out[i] = cpu_to_le16(p1k[i]);
 865}
 866
 867struct wowlan_key_data {
 868	struct iwl_rxon_context *ctx;
 869	struct iwlagn_wowlan_rsc_tsc_params_cmd *rsc_tsc;
 870	struct iwlagn_wowlan_tkip_params_cmd *tkip;
 871	const u8 *bssid;
 872	bool error, use_rsc_tsc, use_tkip;
 873};
 874
 875
 876static void iwlagn_wowlan_program_keys(struct ieee80211_hw *hw,
 877			       struct ieee80211_vif *vif,
 878			       struct ieee80211_sta *sta,
 879			       struct ieee80211_key_conf *key,
 880			       void *_data)
 881{
 882	struct iwl_priv *priv = IWL_MAC80211_GET_DVM(hw);
 883	struct wowlan_key_data *data = _data;
 884	struct iwl_rxon_context *ctx = data->ctx;
 885	struct aes_sc *aes_sc, *aes_tx_sc = NULL;
 886	struct tkip_sc *tkip_sc, *tkip_tx_sc = NULL;
 887	struct iwlagn_p1k_cache *rx_p1ks;
 888	u8 *rx_mic_key;
 889	struct ieee80211_key_seq seq;
 890	u32 cur_rx_iv32 = 0;
 891	u16 p1k[IWLAGN_P1K_SIZE];
 892	int ret, i;
 893
 894	mutex_lock(&priv->mutex);
 895
 896	if ((key->cipher == WLAN_CIPHER_SUITE_WEP40 ||
 897	     key->cipher == WLAN_CIPHER_SUITE_WEP104) &&
 898	     !sta && !ctx->key_mapping_keys)
 899		ret = iwl_set_default_wep_key(priv, ctx, key);
 900	else
 901		ret = iwl_set_dynamic_key(priv, ctx, key, sta);
 902
 903	if (ret) {
 904		IWL_ERR(priv, "Error setting key during suspend!\n");
 905		data->error = true;
 906	}
 907
 908	switch (key->cipher) {
 909	case WLAN_CIPHER_SUITE_TKIP:
 910		if (sta) {
 911			u64 pn64;
 912
 913			tkip_sc = data->rsc_tsc->all_tsc_rsc.tkip.unicast_rsc;
 914			tkip_tx_sc = &data->rsc_tsc->all_tsc_rsc.tkip.tsc;
 915
 916			rx_p1ks = data->tkip->rx_uni;
 917
 918			pn64 = atomic64_read(&key->tx_pn);
 919			tkip_tx_sc->iv16 = cpu_to_le16(TKIP_PN_TO_IV16(pn64));
 920			tkip_tx_sc->iv32 = cpu_to_le32(TKIP_PN_TO_IV32(pn64));
 921
 922			ieee80211_get_tkip_p1k_iv(key, seq.tkip.iv32, p1k);
 923			iwlagn_convert_p1k(p1k, data->tkip->tx.p1k);
 924
 925			memcpy(data->tkip->mic_keys.tx,
 926			       &key->key[NL80211_TKIP_DATA_OFFSET_TX_MIC_KEY],
 927			       IWLAGN_MIC_KEY_SIZE);
 928
 929			rx_mic_key = data->tkip->mic_keys.rx_unicast;
 930		} else {
 931			tkip_sc =
 932				data->rsc_tsc->all_tsc_rsc.tkip.multicast_rsc;
 933			rx_p1ks = data->tkip->rx_multi;
 934			rx_mic_key = data->tkip->mic_keys.rx_mcast;
 935		}
 936
 937		/*
 938		 * For non-QoS this relies on the fact that both the uCode and
 939		 * mac80211 use TID 0 (as they need to to avoid replay attacks)
 940		 * for checking the IV in the frames.
 941		 */
 942		for (i = 0; i < IWLAGN_NUM_RSC; i++) {
 943			ieee80211_get_key_rx_seq(key, i, &seq);
 944			tkip_sc[i].iv16 = cpu_to_le16(seq.tkip.iv16);
 945			tkip_sc[i].iv32 = cpu_to_le32(seq.tkip.iv32);
 946			/* wrapping isn't allowed, AP must rekey */
 947			if (seq.tkip.iv32 > cur_rx_iv32)
 948				cur_rx_iv32 = seq.tkip.iv32;
 949		}
 950
 951		ieee80211_get_tkip_rx_p1k(key, data->bssid, cur_rx_iv32, p1k);
 952		iwlagn_convert_p1k(p1k, rx_p1ks[0].p1k);
 953		ieee80211_get_tkip_rx_p1k(key, data->bssid,
 954					  cur_rx_iv32 + 1, p1k);
 955		iwlagn_convert_p1k(p1k, rx_p1ks[1].p1k);
 956
 957		memcpy(rx_mic_key,
 958		       &key->key[NL80211_TKIP_DATA_OFFSET_RX_MIC_KEY],
 959		       IWLAGN_MIC_KEY_SIZE);
 960
 961		data->use_tkip = true;
 962		data->use_rsc_tsc = true;
 963		break;
 964	case WLAN_CIPHER_SUITE_CCMP:
 965		if (sta) {
 966			u64 pn64;
 967
 968			aes_sc = data->rsc_tsc->all_tsc_rsc.aes.unicast_rsc;
 969			aes_tx_sc = &data->rsc_tsc->all_tsc_rsc.aes.tsc;
 970
 971			pn64 = atomic64_read(&key->tx_pn);
 972			aes_tx_sc->pn = cpu_to_le64(pn64);
 973		} else
 974			aes_sc = data->rsc_tsc->all_tsc_rsc.aes.multicast_rsc;
 975
 976		/*
 977		 * For non-QoS this relies on the fact that both the uCode and
 978		 * mac80211 use TID 0 for checking the IV in the frames.
 979		 */
 980		for (i = 0; i < IWLAGN_NUM_RSC; i++) {
 981			u8 *pn = seq.ccmp.pn;
 982
 983			ieee80211_get_key_rx_seq(key, i, &seq);
 984			aes_sc[i].pn = cpu_to_le64(
 985					(u64)pn[5] |
 986					((u64)pn[4] << 8) |
 987					((u64)pn[3] << 16) |
 988					((u64)pn[2] << 24) |
 989					((u64)pn[1] << 32) |
 990					((u64)pn[0] << 40));
 991		}
 992		data->use_rsc_tsc = true;
 993		break;
 994	}
 995
 996	mutex_unlock(&priv->mutex);
 997}
 998
 999int iwlagn_send_patterns(struct iwl_priv *priv,
1000			struct cfg80211_wowlan *wowlan)
1001{
1002	struct iwlagn_wowlan_patterns_cmd *pattern_cmd;
1003	struct iwl_host_cmd cmd = {
1004		.id = REPLY_WOWLAN_PATTERNS,
1005		.dataflags[0] = IWL_HCMD_DFL_NOCOPY,
1006	};
1007	int i, err;
1008
1009	if (!wowlan->n_patterns)
1010		return 0;
1011
1012	cmd.len[0] = struct_size(pattern_cmd, patterns, wowlan->n_patterns);
1013
1014	pattern_cmd = kmalloc(cmd.len[0], GFP_KERNEL);
1015	if (!pattern_cmd)
1016		return -ENOMEM;
1017
1018	pattern_cmd->n_patterns = cpu_to_le32(wowlan->n_patterns);
1019
1020	for (i = 0; i < wowlan->n_patterns; i++) {
1021		int mask_len = DIV_ROUND_UP(wowlan->patterns[i].pattern_len, 8);
1022
1023		memcpy(&pattern_cmd->patterns[i].mask,
1024			wowlan->patterns[i].mask, mask_len);
1025		memcpy(&pattern_cmd->patterns[i].pattern,
1026			wowlan->patterns[i].pattern,
1027			wowlan->patterns[i].pattern_len);
1028		pattern_cmd->patterns[i].mask_size = mask_len;
1029		pattern_cmd->patterns[i].pattern_size =
1030			wowlan->patterns[i].pattern_len;
1031	}
1032
1033	cmd.data[0] = pattern_cmd;
1034	err = iwl_dvm_send_cmd(priv, &cmd);
1035	kfree(pattern_cmd);
1036	return err;
1037}
1038
1039int iwlagn_suspend(struct iwl_priv *priv, struct cfg80211_wowlan *wowlan)
1040{
1041	struct iwlagn_wowlan_wakeup_filter_cmd wakeup_filter_cmd;
1042	struct iwl_rxon_cmd rxon;
1043	struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS];
1044	struct iwlagn_wowlan_kek_kck_material_cmd kek_kck_cmd;
1045	struct iwlagn_wowlan_tkip_params_cmd tkip_cmd = {};
1046	struct iwlagn_d3_config_cmd d3_cfg_cmd = {
1047		/*
1048		 * Program the minimum sleep time to 10 seconds, as many
1049		 * platforms have issues processing a wakeup signal while
1050		 * still being in the process of suspending.
1051		 */
1052		.min_sleep_time = cpu_to_le32(10 * 1000 * 1000),
1053	};
1054	struct wowlan_key_data key_data = {
1055		.ctx = ctx,
1056		.bssid = ctx->active.bssid_addr,
1057		.use_rsc_tsc = false,
1058		.tkip = &tkip_cmd,
1059		.use_tkip = false,
1060	};
1061	int ret, i;
1062	u16 seq;
1063
1064	key_data.rsc_tsc = kzalloc(sizeof(*key_data.rsc_tsc), GFP_KERNEL);
1065	if (!key_data.rsc_tsc)
1066		return -ENOMEM;
1067
1068	memset(&wakeup_filter_cmd, 0, sizeof(wakeup_filter_cmd));
1069
1070	/*
1071	 * We know the last used seqno, and the uCode expects to know that
1072	 * one, it will increment before TX.
1073	 */
1074	seq = le16_to_cpu(priv->last_seq_ctl) & IEEE80211_SCTL_SEQ;
1075	wakeup_filter_cmd.non_qos_seq = cpu_to_le16(seq);
1076
1077	/*
1078	 * For QoS counters, we store the one to use next, so subtract 0x10
1079	 * since the uCode will add 0x10 before using the value.
1080	 */
1081	for (i = 0; i < IWL_MAX_TID_COUNT; i++) {
1082		seq = priv->tid_data[IWL_AP_ID][i].seq_number;
1083		seq -= 0x10;
1084		wakeup_filter_cmd.qos_seq[i] = cpu_to_le16(seq);
1085	}
1086
1087	if (wowlan->disconnect)
1088		wakeup_filter_cmd.enabled |=
1089			cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_BEACON_MISS |
1090				    IWLAGN_WOWLAN_WAKEUP_LINK_CHANGE);
1091	if (wowlan->magic_pkt)
1092		wakeup_filter_cmd.enabled |=
1093			cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_MAGIC_PACKET);
1094	if (wowlan->gtk_rekey_failure)
1095		wakeup_filter_cmd.enabled |=
1096			cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_GTK_REKEY_FAIL);
1097	if (wowlan->eap_identity_req)
1098		wakeup_filter_cmd.enabled |=
1099			cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_EAP_IDENT_REQ);
1100	if (wowlan->four_way_handshake)
1101		wakeup_filter_cmd.enabled |=
1102			cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_4WAY_HANDSHAKE);
1103	if (wowlan->n_patterns)
1104		wakeup_filter_cmd.enabled |=
1105			cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_PATTERN_MATCH);
1106
1107	if (wowlan->rfkill_release)
1108		d3_cfg_cmd.wakeup_flags |=
1109			cpu_to_le32(IWLAGN_D3_WAKEUP_RFKILL);
1110
1111	iwl_scan_cancel_timeout(priv, 200);
1112
1113	memcpy(&rxon, &ctx->active, sizeof(rxon));
1114
1115	priv->ucode_loaded = false;
1116	iwl_trans_stop_device(priv->trans);
1117	ret = iwl_trans_start_hw(priv->trans);
1118	if (ret)
1119		goto out;
1120
1121	priv->wowlan = true;
1122
1123	ret = iwl_load_ucode_wait_alive(priv, IWL_UCODE_WOWLAN);
1124	if (ret)
1125		goto out;
1126
1127	/* now configure WoWLAN ucode */
1128	ret = iwl_alive_start(priv);
1129	if (ret)
1130		goto out;
1131
1132	memcpy(&ctx->staging, &rxon, sizeof(rxon));
1133	ret = iwlagn_commit_rxon(priv, ctx);
1134	if (ret)
1135		goto out;
1136
1137	ret = iwl_power_update_mode(priv, true);
1138	if (ret)
1139		goto out;
1140
1141	if (!iwlwifi_mod_params.swcrypto) {
1142		/* mark all keys clear */
1143		priv->ucode_key_table = 0;
1144		ctx->key_mapping_keys = 0;
1145
1146		/*
1147		 * This needs to be unlocked due to lock ordering
1148		 * constraints. Since we're in the suspend path
1149		 * that isn't really a problem though.
1150		 */
1151		mutex_unlock(&priv->mutex);
1152		ieee80211_iter_keys(priv->hw, ctx->vif,
1153				    iwlagn_wowlan_program_keys,
1154				    &key_data);
1155		mutex_lock(&priv->mutex);
1156		if (key_data.error) {
1157			ret = -EIO;
1158			goto out;
1159		}
1160
1161		if (key_data.use_rsc_tsc) {
1162			struct iwl_host_cmd rsc_tsc_cmd = {
1163				.id = REPLY_WOWLAN_TSC_RSC_PARAMS,
1164				.data[0] = key_data.rsc_tsc,
1165				.dataflags[0] = IWL_HCMD_DFL_NOCOPY,
1166				.len[0] = sizeof(*key_data.rsc_tsc),
1167			};
1168
1169			ret = iwl_dvm_send_cmd(priv, &rsc_tsc_cmd);
1170			if (ret)
1171				goto out;
1172		}
1173
1174		if (key_data.use_tkip) {
1175			ret = iwl_dvm_send_cmd_pdu(priv,
1176						 REPLY_WOWLAN_TKIP_PARAMS,
1177						 0, sizeof(tkip_cmd),
1178						 &tkip_cmd);
1179			if (ret)
1180				goto out;
1181		}
1182
1183		if (priv->have_rekey_data) {
1184			memset(&kek_kck_cmd, 0, sizeof(kek_kck_cmd));
1185			memcpy(kek_kck_cmd.kck, priv->kck, NL80211_KCK_LEN);
1186			kek_kck_cmd.kck_len = cpu_to_le16(NL80211_KCK_LEN);
1187			memcpy(kek_kck_cmd.kek, priv->kek, NL80211_KEK_LEN);
1188			kek_kck_cmd.kek_len = cpu_to_le16(NL80211_KEK_LEN);
1189			kek_kck_cmd.replay_ctr = priv->replay_ctr;
1190
1191			ret = iwl_dvm_send_cmd_pdu(priv,
1192						 REPLY_WOWLAN_KEK_KCK_MATERIAL,
1193						 0, sizeof(kek_kck_cmd),
1194						 &kek_kck_cmd);
1195			if (ret)
1196				goto out;
1197		}
1198	}
1199
1200	ret = iwl_dvm_send_cmd_pdu(priv, REPLY_D3_CONFIG, 0,
1201				     sizeof(d3_cfg_cmd), &d3_cfg_cmd);
1202	if (ret)
1203		goto out;
1204
1205	ret = iwl_dvm_send_cmd_pdu(priv, REPLY_WOWLAN_WAKEUP_FILTER,
1206				 0, sizeof(wakeup_filter_cmd),
1207				 &wakeup_filter_cmd);
1208	if (ret)
1209		goto out;
1210
1211	ret = iwlagn_send_patterns(priv, wowlan);
1212 out:
1213	kfree(key_data.rsc_tsc);
1214	return ret;
1215}
1216#endif
1217
1218int iwl_dvm_send_cmd(struct iwl_priv *priv, struct iwl_host_cmd *cmd)
1219{
1220	if (iwl_is_rfkill(priv) || iwl_is_ctkill(priv)) {
1221		IWL_WARN(priv, "Not sending command - %s KILL\n",
1222			 iwl_is_rfkill(priv) ? "RF" : "CT");
1223		return -EIO;
1224	}
1225
1226	if (test_bit(STATUS_FW_ERROR, &priv->status)) {
1227		IWL_ERR(priv, "Command %s failed: FW Error\n",
1228			iwl_get_cmd_string(priv->trans, cmd->id));
1229		return -EIO;
1230	}
1231
1232	/*
1233	 * This can happen upon FW ASSERT: we clear the STATUS_FW_ERROR flag
1234	 * in iwl_down but cancel the workers only later.
1235	 */
1236	if (!priv->ucode_loaded) {
1237		IWL_ERR(priv, "Fw not loaded - dropping CMD: %x\n", cmd->id);
1238		return -EIO;
1239	}
1240
1241	/*
1242	 * Synchronous commands from this op-mode must hold
1243	 * the mutex, this ensures we don't try to send two
1244	 * (or more) synchronous commands at a time.
1245	 */
1246	if (!(cmd->flags & CMD_ASYNC))
1247		lockdep_assert_held(&priv->mutex);
1248
1249	return iwl_trans_send_cmd(priv->trans, cmd);
1250}
1251
1252int iwl_dvm_send_cmd_pdu(struct iwl_priv *priv, u8 id,
1253			 u32 flags, u16 len, const void *data)
1254{
1255	struct iwl_host_cmd cmd = {
1256		.id = id,
1257		.len = { len, },
1258		.data = { data, },
1259		.flags = flags,
1260	};
1261
1262	return iwl_dvm_send_cmd(priv, &cmd);
1263}