Loading...
1// SPDX-License-Identifier: GPL-2.0-only
2/******************************************************************************
3 *
4 * Copyright(c) 2008 - 2014, 2022 Intel Corporation. All rights reserved.
5 *****************************************************************************/
6#include <linux/etherdevice.h>
7#include <linux/kernel.h>
8#include <linux/module.h>
9#include <linux/sched.h>
10#include <net/mac80211.h>
11
12#include "iwl-io.h"
13#include "iwl-agn-hw.h"
14#include "iwl-trans.h"
15#include "iwl-modparams.h"
16
17#include "dev.h"
18#include "agn.h"
19
20int iwlagn_hw_valid_rtc_data_addr(u32 addr)
21{
22 return (addr >= IWLAGN_RTC_DATA_LOWER_BOUND) &&
23 (addr < IWLAGN_RTC_DATA_UPPER_BOUND);
24}
25
26int iwlagn_send_tx_power(struct iwl_priv *priv)
27{
28 struct iwlagn_tx_power_dbm_cmd tx_power_cmd;
29 u8 tx_ant_cfg_cmd;
30
31 if (WARN_ONCE(test_bit(STATUS_SCAN_HW, &priv->status),
32 "TX Power requested while scanning!\n"))
33 return -EAGAIN;
34
35 /* half dBm need to multiply */
36 tx_power_cmd.global_lmt = (s8)(2 * priv->tx_power_user_lmt);
37
38 if (tx_power_cmd.global_lmt > priv->nvm_data->max_tx_pwr_half_dbm) {
39 /*
40 * For the newer devices which using enhanced/extend tx power
41 * table in EEPROM, the format is in half dBm. driver need to
42 * convert to dBm format before report to mac80211.
43 * By doing so, there is a possibility of 1/2 dBm resolution
44 * lost. driver will perform "round-up" operation before
45 * reporting, but it will cause 1/2 dBm tx power over the
46 * regulatory limit. Perform the checking here, if the
47 * "tx_power_user_lmt" is higher than EEPROM value (in
48 * half-dBm format), lower the tx power based on EEPROM
49 */
50 tx_power_cmd.global_lmt =
51 priv->nvm_data->max_tx_pwr_half_dbm;
52 }
53 tx_power_cmd.flags = IWLAGN_TX_POWER_NO_CLOSED;
54 tx_power_cmd.srv_chan_lmt = IWLAGN_TX_POWER_AUTO;
55
56 if (IWL_UCODE_API(priv->fw->ucode_ver) == 1)
57 tx_ant_cfg_cmd = REPLY_TX_POWER_DBM_CMD_V1;
58 else
59 tx_ant_cfg_cmd = REPLY_TX_POWER_DBM_CMD;
60
61 return iwl_dvm_send_cmd_pdu(priv, tx_ant_cfg_cmd, 0,
62 sizeof(tx_power_cmd), &tx_power_cmd);
63}
64
65void iwlagn_temperature(struct iwl_priv *priv)
66{
67 lockdep_assert_held(&priv->statistics.lock);
68
69 /* store temperature from correct statistics (in Celsius) */
70 priv->temperature = le32_to_cpu(priv->statistics.common.temperature);
71 iwl_tt_handler(priv);
72}
73
74int iwlagn_hwrate_to_mac80211_idx(u32 rate_n_flags, enum nl80211_band band)
75{
76 int idx = 0;
77 int band_offset = 0;
78
79 /* HT rate format: mac80211 wants an MCS number, which is just LSB */
80 if (rate_n_flags & RATE_MCS_HT_MSK) {
81 idx = (rate_n_flags & 0xff);
82 return idx;
83 /* Legacy rate format, search for match in table */
84 } else {
85 if (band == NL80211_BAND_5GHZ)
86 band_offset = IWL_FIRST_OFDM_RATE;
87 for (idx = band_offset; idx < IWL_RATE_COUNT_LEGACY; idx++)
88 if (iwl_rates[idx].plcp == (rate_n_flags & 0xFF))
89 return idx - band_offset;
90 }
91
92 return -1;
93}
94
95int iwlagn_manage_ibss_station(struct iwl_priv *priv,
96 struct ieee80211_vif *vif, bool add)
97{
98 struct iwl_vif_priv *vif_priv = (void *)vif->drv_priv;
99
100 if (add)
101 return iwlagn_add_bssid_station(priv, vif_priv->ctx,
102 vif->bss_conf.bssid,
103 &vif_priv->ibss_bssid_sta_id);
104 return iwl_remove_station(priv, vif_priv->ibss_bssid_sta_id,
105 vif->bss_conf.bssid);
106}
107
108/*
109 * iwlagn_txfifo_flush: send REPLY_TXFIFO_FLUSH command to uCode
110 *
111 * pre-requirements:
112 * 1. acquire mutex before calling
113 * 2. make sure rf is on and not in exit state
114 */
115int iwlagn_txfifo_flush(struct iwl_priv *priv, u32 scd_q_msk)
116{
117 struct iwl_txfifo_flush_cmd_v3 flush_cmd_v3 = {
118 .flush_control = cpu_to_le16(IWL_DROP_ALL),
119 };
120 struct iwl_txfifo_flush_cmd_v2 flush_cmd_v2 = {
121 .flush_control = cpu_to_le16(IWL_DROP_ALL),
122 };
123
124 u32 queue_control = IWL_SCD_VO_MSK | IWL_SCD_VI_MSK |
125 IWL_SCD_BE_MSK | IWL_SCD_BK_MSK | IWL_SCD_MGMT_MSK;
126
127 if ((priv->valid_contexts != BIT(IWL_RXON_CTX_BSS)))
128 queue_control |= IWL_PAN_SCD_VO_MSK | IWL_PAN_SCD_VI_MSK |
129 IWL_PAN_SCD_BE_MSK | IWL_PAN_SCD_BK_MSK |
130 IWL_PAN_SCD_MGMT_MSK |
131 IWL_PAN_SCD_MULTICAST_MSK;
132
133 if (priv->nvm_data->sku_cap_11n_enable)
134 queue_control |= IWL_AGG_TX_QUEUE_MSK;
135
136 if (scd_q_msk)
137 queue_control = scd_q_msk;
138
139 IWL_DEBUG_INFO(priv, "queue control: 0x%x\n", queue_control);
140 flush_cmd_v3.queue_control = cpu_to_le32(queue_control);
141 flush_cmd_v2.queue_control = cpu_to_le16((u16)queue_control);
142
143 if (IWL_UCODE_API(priv->fw->ucode_ver) > 2)
144 return iwl_dvm_send_cmd_pdu(priv, REPLY_TXFIFO_FLUSH, 0,
145 sizeof(flush_cmd_v3),
146 &flush_cmd_v3);
147 return iwl_dvm_send_cmd_pdu(priv, REPLY_TXFIFO_FLUSH, 0,
148 sizeof(flush_cmd_v2), &flush_cmd_v2);
149}
150
151void iwlagn_dev_txfifo_flush(struct iwl_priv *priv)
152{
153 mutex_lock(&priv->mutex);
154 ieee80211_stop_queues(priv->hw);
155 if (iwlagn_txfifo_flush(priv, 0)) {
156 IWL_ERR(priv, "flush request fail\n");
157 goto done;
158 }
159 IWL_DEBUG_INFO(priv, "wait transmit/flush all frames\n");
160 iwl_trans_wait_tx_queues_empty(priv->trans, 0xffffffff);
161done:
162 ieee80211_wake_queues(priv->hw);
163 mutex_unlock(&priv->mutex);
164}
165
166/*
167 * BT coex
168 */
169/* Notmal TDM */
170static const __le32 iwlagn_def_3w_lookup[IWLAGN_BT_DECISION_LUT_SIZE] = {
171 cpu_to_le32(0xaaaaaaaa),
172 cpu_to_le32(0xaaaaaaaa),
173 cpu_to_le32(0xaeaaaaaa),
174 cpu_to_le32(0xaaaaaaaa),
175 cpu_to_le32(0xcc00ff28),
176 cpu_to_le32(0x0000aaaa),
177 cpu_to_le32(0xcc00aaaa),
178 cpu_to_le32(0x0000aaaa),
179 cpu_to_le32(0xc0004000),
180 cpu_to_le32(0x00004000),
181 cpu_to_le32(0xf0005000),
182 cpu_to_le32(0xf0005000),
183};
184
185/* Full concurrency */
186static const __le32 iwlagn_concurrent_lookup[IWLAGN_BT_DECISION_LUT_SIZE] = {
187 cpu_to_le32(0xaaaaaaaa),
188 cpu_to_le32(0xaaaaaaaa),
189 cpu_to_le32(0xaaaaaaaa),
190 cpu_to_le32(0xaaaaaaaa),
191 cpu_to_le32(0xaaaaaaaa),
192 cpu_to_le32(0xaaaaaaaa),
193 cpu_to_le32(0xaaaaaaaa),
194 cpu_to_le32(0xaaaaaaaa),
195 cpu_to_le32(0x00000000),
196 cpu_to_le32(0x00000000),
197 cpu_to_le32(0x00000000),
198 cpu_to_le32(0x00000000),
199};
200
201void iwlagn_send_advance_bt_config(struct iwl_priv *priv)
202{
203 struct iwl_basic_bt_cmd basic = {
204 .max_kill = IWLAGN_BT_MAX_KILL_DEFAULT,
205 .bt3_timer_t7_value = IWLAGN_BT3_T7_DEFAULT,
206 .bt3_prio_sample_time = IWLAGN_BT3_PRIO_SAMPLE_DEFAULT,
207 .bt3_timer_t2_value = IWLAGN_BT3_T2_DEFAULT,
208 };
209 struct iwl_bt_cmd_v1 bt_cmd_v1;
210 struct iwl_bt_cmd_v2 bt_cmd_v2;
211 int ret;
212
213 BUILD_BUG_ON(sizeof(iwlagn_def_3w_lookup) !=
214 sizeof(basic.bt3_lookup_table));
215
216 if (priv->lib->bt_params) {
217 /*
218 * newer generation of devices (2000 series and newer)
219 * use the version 2 of the bt command
220 * we need to make sure sending the host command
221 * with correct data structure to avoid uCode assert
222 */
223 if (priv->lib->bt_params->bt_session_2) {
224 bt_cmd_v2.prio_boost = cpu_to_le32(
225 priv->lib->bt_params->bt_prio_boost);
226 bt_cmd_v2.tx_prio_boost = 0;
227 bt_cmd_v2.rx_prio_boost = 0;
228 } else {
229 /* older version only has 8 bits */
230 WARN_ON(priv->lib->bt_params->bt_prio_boost & ~0xFF);
231 bt_cmd_v1.prio_boost =
232 priv->lib->bt_params->bt_prio_boost;
233 bt_cmd_v1.tx_prio_boost = 0;
234 bt_cmd_v1.rx_prio_boost = 0;
235 }
236 } else {
237 IWL_ERR(priv, "failed to construct BT Coex Config\n");
238 return;
239 }
240
241 /*
242 * Possible situations when BT needs to take over for receive,
243 * at the same time where STA needs to response to AP's frame(s),
244 * reduce the tx power of the required response frames, by that,
245 * allow the concurrent BT receive & WiFi transmit
246 * (BT - ANT A, WiFi -ANT B), without interference to one another
247 *
248 * Reduced tx power apply to control frames only (ACK/Back/CTS)
249 * when indicated by the BT config command
250 */
251 basic.kill_ack_mask = priv->kill_ack_mask;
252 basic.kill_cts_mask = priv->kill_cts_mask;
253 if (priv->reduced_txpower)
254 basic.reduce_txpower = IWLAGN_BT_REDUCED_TX_PWR;
255 basic.valid = priv->bt_valid;
256
257 /*
258 * Configure BT coex mode to "no coexistence" when the
259 * user disabled BT coexistence, we have no interface
260 * (might be in monitor mode), or the interface is in
261 * IBSS mode (no proper uCode support for coex then).
262 */
263 if (!iwlwifi_mod_params.bt_coex_active ||
264 priv->iw_mode == NL80211_IFTYPE_ADHOC) {
265 basic.flags = IWLAGN_BT_FLAG_COEX_MODE_DISABLED;
266 } else {
267 basic.flags = IWLAGN_BT_FLAG_COEX_MODE_3W <<
268 IWLAGN_BT_FLAG_COEX_MODE_SHIFT;
269
270 if (!priv->bt_enable_pspoll)
271 basic.flags |= IWLAGN_BT_FLAG_SYNC_2_BT_DISABLE;
272 else
273 basic.flags &= ~IWLAGN_BT_FLAG_SYNC_2_BT_DISABLE;
274
275 if (priv->bt_ch_announce)
276 basic.flags |= IWLAGN_BT_FLAG_CHANNEL_INHIBITION;
277 IWL_DEBUG_COEX(priv, "BT coex flag: 0X%x\n", basic.flags);
278 }
279 priv->bt_enable_flag = basic.flags;
280 if (priv->bt_full_concurrent)
281 memcpy(basic.bt3_lookup_table, iwlagn_concurrent_lookup,
282 sizeof(iwlagn_concurrent_lookup));
283 else
284 memcpy(basic.bt3_lookup_table, iwlagn_def_3w_lookup,
285 sizeof(iwlagn_def_3w_lookup));
286
287 IWL_DEBUG_COEX(priv, "BT coex %s in %s mode\n",
288 basic.flags ? "active" : "disabled",
289 priv->bt_full_concurrent ?
290 "full concurrency" : "3-wire");
291
292 if (priv->lib->bt_params->bt_session_2) {
293 memcpy(&bt_cmd_v2.basic, &basic,
294 sizeof(basic));
295 ret = iwl_dvm_send_cmd_pdu(priv, REPLY_BT_CONFIG,
296 0, sizeof(bt_cmd_v2), &bt_cmd_v2);
297 } else {
298 memcpy(&bt_cmd_v1.basic, &basic,
299 sizeof(basic));
300 ret = iwl_dvm_send_cmd_pdu(priv, REPLY_BT_CONFIG,
301 0, sizeof(bt_cmd_v1), &bt_cmd_v1);
302 }
303 if (ret)
304 IWL_ERR(priv, "failed to send BT Coex Config\n");
305
306}
307
308void iwlagn_bt_adjust_rssi_monitor(struct iwl_priv *priv, bool rssi_ena)
309{
310 struct iwl_rxon_context *ctx, *found_ctx = NULL;
311 bool found_ap = false;
312
313 lockdep_assert_held(&priv->mutex);
314
315 /* Check whether AP or GO mode is active. */
316 if (rssi_ena) {
317 for_each_context(priv, ctx) {
318 if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_AP &&
319 iwl_is_associated_ctx(ctx)) {
320 found_ap = true;
321 break;
322 }
323 }
324 }
325
326 /*
327 * If disable was received or If GO/AP mode, disable RSSI
328 * measurements.
329 */
330 if (!rssi_ena || found_ap) {
331 if (priv->cur_rssi_ctx) {
332 ctx = priv->cur_rssi_ctx;
333 ieee80211_disable_rssi_reports(ctx->vif);
334 priv->cur_rssi_ctx = NULL;
335 }
336 return;
337 }
338
339 /*
340 * If rssi measurements need to be enabled, consider all cases now.
341 * Figure out how many contexts are active.
342 */
343 for_each_context(priv, ctx) {
344 if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_STATION &&
345 iwl_is_associated_ctx(ctx)) {
346 found_ctx = ctx;
347 break;
348 }
349 }
350
351 /*
352 * rssi monitor already enabled for the correct interface...nothing
353 * to do.
354 */
355 if (found_ctx == priv->cur_rssi_ctx)
356 return;
357
358 /*
359 * Figure out if rssi monitor is currently enabled, and needs
360 * to be changed. If rssi monitor is already enabled, disable
361 * it first else just enable rssi measurements on the
362 * interface found above.
363 */
364 if (priv->cur_rssi_ctx) {
365 ctx = priv->cur_rssi_ctx;
366 if (ctx->vif)
367 ieee80211_disable_rssi_reports(ctx->vif);
368 }
369
370 priv->cur_rssi_ctx = found_ctx;
371
372 if (!found_ctx)
373 return;
374
375 ieee80211_enable_rssi_reports(found_ctx->vif,
376 IWLAGN_BT_PSP_MIN_RSSI_THRESHOLD,
377 IWLAGN_BT_PSP_MAX_RSSI_THRESHOLD);
378}
379
380static bool iwlagn_bt_traffic_is_sco(struct iwl_bt_uart_msg *uart_msg)
381{
382 return (BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3) >>
383 BT_UART_MSG_FRAME3SCOESCO_POS;
384}
385
386static void iwlagn_bt_traffic_change_work(struct work_struct *work)
387{
388 struct iwl_priv *priv =
389 container_of(work, struct iwl_priv, bt_traffic_change_work);
390 struct iwl_rxon_context *ctx;
391 int smps_request = -1;
392
393 if (priv->bt_enable_flag == IWLAGN_BT_FLAG_COEX_MODE_DISABLED) {
394 /* bt coex disabled */
395 return;
396 }
397
398 /*
399 * Note: bt_traffic_load can be overridden by scan complete and
400 * coex profile notifications. Ignore that since only bad consequence
401 * can be not matching debug print with actual state.
402 */
403 IWL_DEBUG_COEX(priv, "BT traffic load changes: %d\n",
404 priv->bt_traffic_load);
405
406 switch (priv->bt_traffic_load) {
407 case IWL_BT_COEX_TRAFFIC_LOAD_NONE:
408 if (priv->bt_status)
409 smps_request = IEEE80211_SMPS_DYNAMIC;
410 else
411 smps_request = IEEE80211_SMPS_AUTOMATIC;
412 break;
413 case IWL_BT_COEX_TRAFFIC_LOAD_LOW:
414 smps_request = IEEE80211_SMPS_DYNAMIC;
415 break;
416 case IWL_BT_COEX_TRAFFIC_LOAD_HIGH:
417 case IWL_BT_COEX_TRAFFIC_LOAD_CONTINUOUS:
418 smps_request = IEEE80211_SMPS_STATIC;
419 break;
420 default:
421 IWL_ERR(priv, "Invalid BT traffic load: %d\n",
422 priv->bt_traffic_load);
423 break;
424 }
425
426 mutex_lock(&priv->mutex);
427
428 /*
429 * We can not send command to firmware while scanning. When the scan
430 * complete we will schedule this work again. We do check with mutex
431 * locked to prevent new scan request to arrive. We do not check
432 * STATUS_SCANNING to avoid race when queue_work two times from
433 * different notifications, but quit and not perform any work at all.
434 */
435 if (test_bit(STATUS_SCAN_HW, &priv->status))
436 goto out;
437
438 iwl_update_chain_flags(priv);
439
440 if (smps_request != -1) {
441 priv->current_ht_config.smps = smps_request;
442 for_each_context(priv, ctx) {
443 if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_STATION)
444 ieee80211_request_smps(ctx->vif, 0, smps_request);
445 }
446 }
447
448 /*
449 * Dynamic PS poll related functionality. Adjust RSSI measurements if
450 * necessary.
451 */
452 iwlagn_bt_coex_rssi_monitor(priv);
453out:
454 mutex_unlock(&priv->mutex);
455}
456
457/*
458 * If BT sco traffic, and RSSI monitor is enabled, move measurements to the
459 * correct interface or disable it if this is the last interface to be
460 * removed.
461 */
462void iwlagn_bt_coex_rssi_monitor(struct iwl_priv *priv)
463{
464 if (priv->bt_is_sco &&
465 priv->bt_traffic_load == IWL_BT_COEX_TRAFFIC_LOAD_CONTINUOUS)
466 iwlagn_bt_adjust_rssi_monitor(priv, true);
467 else
468 iwlagn_bt_adjust_rssi_monitor(priv, false);
469}
470
471static void iwlagn_print_uartmsg(struct iwl_priv *priv,
472 struct iwl_bt_uart_msg *uart_msg)
473{
474 IWL_DEBUG_COEX(priv, "Message Type = 0x%X, SSN = 0x%X, "
475 "Update Req = 0x%X\n",
476 (BT_UART_MSG_FRAME1MSGTYPE_MSK & uart_msg->frame1) >>
477 BT_UART_MSG_FRAME1MSGTYPE_POS,
478 (BT_UART_MSG_FRAME1SSN_MSK & uart_msg->frame1) >>
479 BT_UART_MSG_FRAME1SSN_POS,
480 (BT_UART_MSG_FRAME1UPDATEREQ_MSK & uart_msg->frame1) >>
481 BT_UART_MSG_FRAME1UPDATEREQ_POS);
482
483 IWL_DEBUG_COEX(priv, "Open connections = 0x%X, Traffic load = 0x%X, "
484 "Chl_SeqN = 0x%X, In band = 0x%X\n",
485 (BT_UART_MSG_FRAME2OPENCONNECTIONS_MSK & uart_msg->frame2) >>
486 BT_UART_MSG_FRAME2OPENCONNECTIONS_POS,
487 (BT_UART_MSG_FRAME2TRAFFICLOAD_MSK & uart_msg->frame2) >>
488 BT_UART_MSG_FRAME2TRAFFICLOAD_POS,
489 (BT_UART_MSG_FRAME2CHLSEQN_MSK & uart_msg->frame2) >>
490 BT_UART_MSG_FRAME2CHLSEQN_POS,
491 (BT_UART_MSG_FRAME2INBAND_MSK & uart_msg->frame2) >>
492 BT_UART_MSG_FRAME2INBAND_POS);
493
494 IWL_DEBUG_COEX(priv, "SCO/eSCO = 0x%X, Sniff = 0x%X, A2DP = 0x%X, "
495 "ACL = 0x%X, Master = 0x%X, OBEX = 0x%X\n",
496 (BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3) >>
497 BT_UART_MSG_FRAME3SCOESCO_POS,
498 (BT_UART_MSG_FRAME3SNIFF_MSK & uart_msg->frame3) >>
499 BT_UART_MSG_FRAME3SNIFF_POS,
500 (BT_UART_MSG_FRAME3A2DP_MSK & uart_msg->frame3) >>
501 BT_UART_MSG_FRAME3A2DP_POS,
502 (BT_UART_MSG_FRAME3ACL_MSK & uart_msg->frame3) >>
503 BT_UART_MSG_FRAME3ACL_POS,
504 (BT_UART_MSG_FRAME3MASTER_MSK & uart_msg->frame3) >>
505 BT_UART_MSG_FRAME3MASTER_POS,
506 (BT_UART_MSG_FRAME3OBEX_MSK & uart_msg->frame3) >>
507 BT_UART_MSG_FRAME3OBEX_POS);
508
509 IWL_DEBUG_COEX(priv, "Idle duration = 0x%X\n",
510 (BT_UART_MSG_FRAME4IDLEDURATION_MSK & uart_msg->frame4) >>
511 BT_UART_MSG_FRAME4IDLEDURATION_POS);
512
513 IWL_DEBUG_COEX(priv, "Tx Activity = 0x%X, Rx Activity = 0x%X, "
514 "eSCO Retransmissions = 0x%X\n",
515 (BT_UART_MSG_FRAME5TXACTIVITY_MSK & uart_msg->frame5) >>
516 BT_UART_MSG_FRAME5TXACTIVITY_POS,
517 (BT_UART_MSG_FRAME5RXACTIVITY_MSK & uart_msg->frame5) >>
518 BT_UART_MSG_FRAME5RXACTIVITY_POS,
519 (BT_UART_MSG_FRAME5ESCORETRANSMIT_MSK & uart_msg->frame5) >>
520 BT_UART_MSG_FRAME5ESCORETRANSMIT_POS);
521
522 IWL_DEBUG_COEX(priv, "Sniff Interval = 0x%X, Discoverable = 0x%X\n",
523 (BT_UART_MSG_FRAME6SNIFFINTERVAL_MSK & uart_msg->frame6) >>
524 BT_UART_MSG_FRAME6SNIFFINTERVAL_POS,
525 (BT_UART_MSG_FRAME6DISCOVERABLE_MSK & uart_msg->frame6) >>
526 BT_UART_MSG_FRAME6DISCOVERABLE_POS);
527
528 IWL_DEBUG_COEX(priv, "Sniff Activity = 0x%X, Page = "
529 "0x%X, Inquiry = 0x%X, Connectable = 0x%X\n",
530 (BT_UART_MSG_FRAME7SNIFFACTIVITY_MSK & uart_msg->frame7) >>
531 BT_UART_MSG_FRAME7SNIFFACTIVITY_POS,
532 (BT_UART_MSG_FRAME7PAGE_MSK & uart_msg->frame7) >>
533 BT_UART_MSG_FRAME7PAGE_POS,
534 (BT_UART_MSG_FRAME7INQUIRY_MSK & uart_msg->frame7) >>
535 BT_UART_MSG_FRAME7INQUIRY_POS,
536 (BT_UART_MSG_FRAME7CONNECTABLE_MSK & uart_msg->frame7) >>
537 BT_UART_MSG_FRAME7CONNECTABLE_POS);
538}
539
540static bool iwlagn_set_kill_msk(struct iwl_priv *priv,
541 struct iwl_bt_uart_msg *uart_msg)
542{
543 bool need_update = false;
544 u8 kill_msk = IWL_BT_KILL_REDUCE;
545 static const __le32 bt_kill_ack_msg[3] = {
546 IWLAGN_BT_KILL_ACK_MASK_DEFAULT,
547 IWLAGN_BT_KILL_ACK_CTS_MASK_SCO,
548 IWLAGN_BT_KILL_ACK_CTS_MASK_REDUCE};
549 static const __le32 bt_kill_cts_msg[3] = {
550 IWLAGN_BT_KILL_CTS_MASK_DEFAULT,
551 IWLAGN_BT_KILL_ACK_CTS_MASK_SCO,
552 IWLAGN_BT_KILL_ACK_CTS_MASK_REDUCE};
553
554 if (!priv->reduced_txpower)
555 kill_msk = (BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3)
556 ? IWL_BT_KILL_OVERRIDE : IWL_BT_KILL_DEFAULT;
557 if (priv->kill_ack_mask != bt_kill_ack_msg[kill_msk] ||
558 priv->kill_cts_mask != bt_kill_cts_msg[kill_msk]) {
559 priv->bt_valid |= IWLAGN_BT_VALID_KILL_ACK_MASK;
560 priv->kill_ack_mask = bt_kill_ack_msg[kill_msk];
561 priv->bt_valid |= IWLAGN_BT_VALID_KILL_CTS_MASK;
562 priv->kill_cts_mask = bt_kill_cts_msg[kill_msk];
563 need_update = true;
564 }
565 return need_update;
566}
567
568/*
569 * Upon RSSI changes, sends a bt config command with following changes
570 * 1. enable/disable "reduced control frames tx power
571 * 2. update the "kill)ack_mask" and "kill_cts_mask"
572 *
573 * If "reduced tx power" is enabled, uCode shall
574 * 1. ACK/Back/CTS rate shall reduced to 6Mbps
575 * 2. not use duplciate 20/40MHz mode
576 */
577static bool iwlagn_fill_txpower_mode(struct iwl_priv *priv,
578 struct iwl_bt_uart_msg *uart_msg)
579{
580 bool need_update = false;
581 struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS];
582 int ave_rssi;
583
584 if (!ctx->vif || (ctx->vif->type != NL80211_IFTYPE_STATION)) {
585 IWL_DEBUG_INFO(priv, "BSS ctx not active or not in sta mode\n");
586 return false;
587 }
588
589 ave_rssi = ieee80211_ave_rssi(ctx->vif);
590 if (!ave_rssi) {
591 /* no rssi data, no changes to reduce tx power */
592 IWL_DEBUG_COEX(priv, "no rssi data available\n");
593 return need_update;
594 }
595 if (!priv->reduced_txpower &&
596 !iwl_is_associated(priv, IWL_RXON_CTX_PAN) &&
597 (ave_rssi > BT_ENABLE_REDUCED_TXPOWER_THRESHOLD) &&
598 (uart_msg->frame3 & (BT_UART_MSG_FRAME3ACL_MSK |
599 BT_UART_MSG_FRAME3OBEX_MSK)) &&
600 !(uart_msg->frame3 & (BT_UART_MSG_FRAME3SCOESCO_MSK |
601 BT_UART_MSG_FRAME3SNIFF_MSK | BT_UART_MSG_FRAME3A2DP_MSK))) {
602 /* enabling reduced tx power */
603 priv->reduced_txpower = true;
604 priv->bt_valid |= IWLAGN_BT_VALID_REDUCED_TX_PWR;
605 need_update = true;
606 } else if (priv->reduced_txpower &&
607 (iwl_is_associated(priv, IWL_RXON_CTX_PAN) ||
608 (ave_rssi < BT_DISABLE_REDUCED_TXPOWER_THRESHOLD) ||
609 (uart_msg->frame3 & (BT_UART_MSG_FRAME3SCOESCO_MSK |
610 BT_UART_MSG_FRAME3SNIFF_MSK | BT_UART_MSG_FRAME3A2DP_MSK)) ||
611 !(uart_msg->frame3 & (BT_UART_MSG_FRAME3ACL_MSK |
612 BT_UART_MSG_FRAME3OBEX_MSK)))) {
613 /* disable reduced tx power */
614 priv->reduced_txpower = false;
615 priv->bt_valid |= IWLAGN_BT_VALID_REDUCED_TX_PWR;
616 need_update = true;
617 }
618
619 return need_update;
620}
621
622static void iwlagn_bt_coex_profile_notif(struct iwl_priv *priv,
623 struct iwl_rx_cmd_buffer *rxb)
624{
625 struct iwl_rx_packet *pkt = rxb_addr(rxb);
626 struct iwl_bt_coex_profile_notif *coex = (void *)pkt->data;
627 struct iwl_bt_uart_msg *uart_msg = &coex->last_bt_uart_msg;
628
629 if (priv->bt_enable_flag == IWLAGN_BT_FLAG_COEX_MODE_DISABLED) {
630 /* bt coex disabled */
631 return;
632 }
633
634 IWL_DEBUG_COEX(priv, "BT Coex notification:\n");
635 IWL_DEBUG_COEX(priv, " status: %d\n", coex->bt_status);
636 IWL_DEBUG_COEX(priv, " traffic load: %d\n", coex->bt_traffic_load);
637 IWL_DEBUG_COEX(priv, " CI compliance: %d\n",
638 coex->bt_ci_compliance);
639 iwlagn_print_uartmsg(priv, uart_msg);
640
641 priv->last_bt_traffic_load = priv->bt_traffic_load;
642 priv->bt_is_sco = iwlagn_bt_traffic_is_sco(uart_msg);
643
644 if (priv->iw_mode != NL80211_IFTYPE_ADHOC) {
645 if (priv->bt_status != coex->bt_status ||
646 priv->last_bt_traffic_load != coex->bt_traffic_load) {
647 if (coex->bt_status) {
648 /* BT on */
649 if (!priv->bt_ch_announce)
650 priv->bt_traffic_load =
651 IWL_BT_COEX_TRAFFIC_LOAD_HIGH;
652 else
653 priv->bt_traffic_load =
654 coex->bt_traffic_load;
655 } else {
656 /* BT off */
657 priv->bt_traffic_load =
658 IWL_BT_COEX_TRAFFIC_LOAD_NONE;
659 }
660 priv->bt_status = coex->bt_status;
661 queue_work(priv->workqueue,
662 &priv->bt_traffic_change_work);
663 }
664 }
665
666 /* schedule to send runtime bt_config */
667 /* check reduce power before change ack/cts kill mask */
668 if (iwlagn_fill_txpower_mode(priv, uart_msg) ||
669 iwlagn_set_kill_msk(priv, uart_msg))
670 queue_work(priv->workqueue, &priv->bt_runtime_config);
671
672
673 /* FIXME: based on notification, adjust the prio_boost */
674
675 priv->bt_ci_compliance = coex->bt_ci_compliance;
676}
677
678void iwlagn_bt_rx_handler_setup(struct iwl_priv *priv)
679{
680 priv->rx_handlers[REPLY_BT_COEX_PROFILE_NOTIF] =
681 iwlagn_bt_coex_profile_notif;
682}
683
684void iwlagn_bt_setup_deferred_work(struct iwl_priv *priv)
685{
686 INIT_WORK(&priv->bt_traffic_change_work,
687 iwlagn_bt_traffic_change_work);
688}
689
690void iwlagn_bt_cancel_deferred_work(struct iwl_priv *priv)
691{
692 cancel_work_sync(&priv->bt_traffic_change_work);
693}
694
695static bool is_single_rx_stream(struct iwl_priv *priv)
696{
697 return priv->current_ht_config.smps == IEEE80211_SMPS_STATIC ||
698 priv->current_ht_config.single_chain_sufficient;
699}
700
701#define IWL_NUM_RX_CHAINS_MULTIPLE 3
702#define IWL_NUM_RX_CHAINS_SINGLE 2
703#define IWL_NUM_IDLE_CHAINS_DUAL 2
704#define IWL_NUM_IDLE_CHAINS_SINGLE 1
705
706/*
707 * Determine how many receiver/antenna chains to use.
708 *
709 * More provides better reception via diversity. Fewer saves power
710 * at the expense of throughput, but only when not in powersave to
711 * start with.
712 *
713 * MIMO (dual stream) requires at least 2, but works better with 3.
714 * This does not determine *which* chains to use, just how many.
715 */
716static int iwl_get_active_rx_chain_count(struct iwl_priv *priv)
717{
718 if (priv->lib->bt_params &&
719 priv->lib->bt_params->advanced_bt_coexist &&
720 (priv->bt_full_concurrent ||
721 priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH)) {
722 /*
723 * only use chain 'A' in bt high traffic load or
724 * full concurrency mode
725 */
726 return IWL_NUM_RX_CHAINS_SINGLE;
727 }
728 /* # of Rx chains to use when expecting MIMO. */
729 if (is_single_rx_stream(priv))
730 return IWL_NUM_RX_CHAINS_SINGLE;
731 else
732 return IWL_NUM_RX_CHAINS_MULTIPLE;
733}
734
735/*
736 * When we are in power saving mode, unless device support spatial
737 * multiplexing power save, use the active count for rx chain count.
738 */
739static int iwl_get_idle_rx_chain_count(struct iwl_priv *priv, int active_cnt)
740{
741 /* # Rx chains when idling, depending on SMPS mode */
742 switch (priv->current_ht_config.smps) {
743 case IEEE80211_SMPS_STATIC:
744 case IEEE80211_SMPS_DYNAMIC:
745 return IWL_NUM_IDLE_CHAINS_SINGLE;
746 case IEEE80211_SMPS_AUTOMATIC:
747 case IEEE80211_SMPS_OFF:
748 return active_cnt;
749 default:
750 WARN(1, "invalid SMPS mode %d",
751 priv->current_ht_config.smps);
752 return active_cnt;
753 }
754}
755
756/* up to 4 chains */
757static u8 iwl_count_chain_bitmap(u32 chain_bitmap)
758{
759 u8 res;
760 res = (chain_bitmap & BIT(0)) >> 0;
761 res += (chain_bitmap & BIT(1)) >> 1;
762 res += (chain_bitmap & BIT(2)) >> 2;
763 res += (chain_bitmap & BIT(3)) >> 3;
764 return res;
765}
766
767/*
768 * iwlagn_set_rxon_chain - Set up Rx chain usage in "staging" RXON image
769 *
770 * Selects how many and which Rx receivers/antennas/chains to use.
771 * This should not be used for scan command ... it puts data in wrong place.
772 */
773void iwlagn_set_rxon_chain(struct iwl_priv *priv, struct iwl_rxon_context *ctx)
774{
775 bool is_single = is_single_rx_stream(priv);
776 bool is_cam = !test_bit(STATUS_POWER_PMI, &priv->status);
777 u8 idle_rx_cnt, active_rx_cnt, valid_rx_cnt;
778 u32 active_chains;
779 u16 rx_chain;
780
781 /* Tell uCode which antennas are actually connected.
782 * Before first association, we assume all antennas are connected.
783 * Just after first association, iwl_chain_noise_calibration()
784 * checks which antennas actually *are* connected. */
785 if (priv->chain_noise_data.active_chains)
786 active_chains = priv->chain_noise_data.active_chains;
787 else
788 active_chains = priv->nvm_data->valid_rx_ant;
789
790 if (priv->lib->bt_params &&
791 priv->lib->bt_params->advanced_bt_coexist &&
792 (priv->bt_full_concurrent ||
793 priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH)) {
794 /*
795 * only use chain 'A' in bt high traffic load or
796 * full concurrency mode
797 */
798 active_chains = first_antenna(active_chains);
799 }
800
801 rx_chain = active_chains << RXON_RX_CHAIN_VALID_POS;
802
803 /* How many receivers should we use? */
804 active_rx_cnt = iwl_get_active_rx_chain_count(priv);
805 idle_rx_cnt = iwl_get_idle_rx_chain_count(priv, active_rx_cnt);
806
807
808 /* correct rx chain count according hw settings
809 * and chain noise calibration
810 */
811 valid_rx_cnt = iwl_count_chain_bitmap(active_chains);
812 if (valid_rx_cnt < active_rx_cnt)
813 active_rx_cnt = valid_rx_cnt;
814
815 if (valid_rx_cnt < idle_rx_cnt)
816 idle_rx_cnt = valid_rx_cnt;
817
818 rx_chain |= active_rx_cnt << RXON_RX_CHAIN_MIMO_CNT_POS;
819 rx_chain |= idle_rx_cnt << RXON_RX_CHAIN_CNT_POS;
820
821 ctx->staging.rx_chain = cpu_to_le16(rx_chain);
822
823 if (!is_single && (active_rx_cnt >= IWL_NUM_RX_CHAINS_SINGLE) && is_cam)
824 ctx->staging.rx_chain |= RXON_RX_CHAIN_MIMO_FORCE_MSK;
825 else
826 ctx->staging.rx_chain &= ~RXON_RX_CHAIN_MIMO_FORCE_MSK;
827
828 IWL_DEBUG_ASSOC(priv, "rx_chain=0x%X active=%d idle=%d\n",
829 ctx->staging.rx_chain,
830 active_rx_cnt, idle_rx_cnt);
831
832 WARN_ON(active_rx_cnt == 0 || idle_rx_cnt == 0 ||
833 active_rx_cnt < idle_rx_cnt);
834}
835
836u8 iwl_toggle_tx_ant(struct iwl_priv *priv, u8 ant, u8 valid)
837{
838 int i;
839 u8 ind = ant;
840
841 if (priv->band == NL80211_BAND_2GHZ &&
842 priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH)
843 return 0;
844
845 for (i = 0; i < RATE_ANT_NUM - 1; i++) {
846 ind = (ind + 1) < RATE_ANT_NUM ? ind + 1 : 0;
847 if (valid & BIT(ind))
848 return ind;
849 }
850 return ant;
851}
852
853#ifdef CONFIG_PM_SLEEP
854static void iwlagn_convert_p1k(u16 *p1k, __le16 *out)
855{
856 int i;
857
858 for (i = 0; i < IWLAGN_P1K_SIZE; i++)
859 out[i] = cpu_to_le16(p1k[i]);
860}
861
862struct wowlan_key_data {
863 struct iwl_rxon_context *ctx;
864 struct iwlagn_wowlan_rsc_tsc_params_cmd *rsc_tsc;
865 struct iwlagn_wowlan_tkip_params_cmd *tkip;
866 const u8 *bssid;
867 bool error, use_rsc_tsc, use_tkip;
868};
869
870
871static void iwlagn_wowlan_program_keys(struct ieee80211_hw *hw,
872 struct ieee80211_vif *vif,
873 struct ieee80211_sta *sta,
874 struct ieee80211_key_conf *key,
875 void *_data)
876{
877 struct iwl_priv *priv = IWL_MAC80211_GET_DVM(hw);
878 struct wowlan_key_data *data = _data;
879 struct iwl_rxon_context *ctx = data->ctx;
880 struct aes_sc *aes_sc, *aes_tx_sc = NULL;
881 struct tkip_sc *tkip_sc, *tkip_tx_sc = NULL;
882 struct iwlagn_p1k_cache *rx_p1ks;
883 u8 *rx_mic_key;
884 struct ieee80211_key_seq seq;
885 u32 cur_rx_iv32 = 0;
886 u16 p1k[IWLAGN_P1K_SIZE];
887 int ret, i;
888
889 mutex_lock(&priv->mutex);
890
891 if ((key->cipher == WLAN_CIPHER_SUITE_WEP40 ||
892 key->cipher == WLAN_CIPHER_SUITE_WEP104) &&
893 !sta && !ctx->key_mapping_keys)
894 ret = iwl_set_default_wep_key(priv, ctx, key);
895 else
896 ret = iwl_set_dynamic_key(priv, ctx, key, sta);
897
898 if (ret) {
899 IWL_ERR(priv, "Error setting key during suspend!\n");
900 data->error = true;
901 }
902
903 switch (key->cipher) {
904 case WLAN_CIPHER_SUITE_TKIP:
905 if (sta) {
906 u64 pn64;
907
908 tkip_sc = data->rsc_tsc->all_tsc_rsc.tkip.unicast_rsc;
909 tkip_tx_sc = &data->rsc_tsc->all_tsc_rsc.tkip.tsc;
910
911 rx_p1ks = data->tkip->rx_uni;
912
913 pn64 = atomic64_read(&key->tx_pn);
914 tkip_tx_sc->iv16 = cpu_to_le16(TKIP_PN_TO_IV16(pn64));
915 tkip_tx_sc->iv32 = cpu_to_le32(TKIP_PN_TO_IV32(pn64));
916
917 ieee80211_get_tkip_p1k_iv(key, seq.tkip.iv32, p1k);
918 iwlagn_convert_p1k(p1k, data->tkip->tx.p1k);
919
920 memcpy(data->tkip->mic_keys.tx,
921 &key->key[NL80211_TKIP_DATA_OFFSET_TX_MIC_KEY],
922 IWLAGN_MIC_KEY_SIZE);
923
924 rx_mic_key = data->tkip->mic_keys.rx_unicast;
925 } else {
926 tkip_sc =
927 data->rsc_tsc->all_tsc_rsc.tkip.multicast_rsc;
928 rx_p1ks = data->tkip->rx_multi;
929 rx_mic_key = data->tkip->mic_keys.rx_mcast;
930 }
931
932 /*
933 * For non-QoS this relies on the fact that both the uCode and
934 * mac80211 use TID 0 (as they need to to avoid replay attacks)
935 * for checking the IV in the frames.
936 */
937 for (i = 0; i < IWLAGN_NUM_RSC; i++) {
938 ieee80211_get_key_rx_seq(key, i, &seq);
939 tkip_sc[i].iv16 = cpu_to_le16(seq.tkip.iv16);
940 tkip_sc[i].iv32 = cpu_to_le32(seq.tkip.iv32);
941 /* wrapping isn't allowed, AP must rekey */
942 if (seq.tkip.iv32 > cur_rx_iv32)
943 cur_rx_iv32 = seq.tkip.iv32;
944 }
945
946 ieee80211_get_tkip_rx_p1k(key, data->bssid, cur_rx_iv32, p1k);
947 iwlagn_convert_p1k(p1k, rx_p1ks[0].p1k);
948 ieee80211_get_tkip_rx_p1k(key, data->bssid,
949 cur_rx_iv32 + 1, p1k);
950 iwlagn_convert_p1k(p1k, rx_p1ks[1].p1k);
951
952 memcpy(rx_mic_key,
953 &key->key[NL80211_TKIP_DATA_OFFSET_RX_MIC_KEY],
954 IWLAGN_MIC_KEY_SIZE);
955
956 data->use_tkip = true;
957 data->use_rsc_tsc = true;
958 break;
959 case WLAN_CIPHER_SUITE_CCMP:
960 if (sta) {
961 u64 pn64;
962
963 aes_sc = data->rsc_tsc->all_tsc_rsc.aes.unicast_rsc;
964 aes_tx_sc = &data->rsc_tsc->all_tsc_rsc.aes.tsc;
965
966 pn64 = atomic64_read(&key->tx_pn);
967 aes_tx_sc->pn = cpu_to_le64(pn64);
968 } else
969 aes_sc = data->rsc_tsc->all_tsc_rsc.aes.multicast_rsc;
970
971 /*
972 * For non-QoS this relies on the fact that both the uCode and
973 * mac80211 use TID 0 for checking the IV in the frames.
974 */
975 for (i = 0; i < IWLAGN_NUM_RSC; i++) {
976 u8 *pn = seq.ccmp.pn;
977
978 ieee80211_get_key_rx_seq(key, i, &seq);
979 aes_sc[i].pn = cpu_to_le64(
980 (u64)pn[5] |
981 ((u64)pn[4] << 8) |
982 ((u64)pn[3] << 16) |
983 ((u64)pn[2] << 24) |
984 ((u64)pn[1] << 32) |
985 ((u64)pn[0] << 40));
986 }
987 data->use_rsc_tsc = true;
988 break;
989 }
990
991 mutex_unlock(&priv->mutex);
992}
993
994int iwlagn_send_patterns(struct iwl_priv *priv,
995 struct cfg80211_wowlan *wowlan)
996{
997 struct iwlagn_wowlan_patterns_cmd *pattern_cmd;
998 struct iwl_host_cmd cmd = {
999 .id = REPLY_WOWLAN_PATTERNS,
1000 .dataflags[0] = IWL_HCMD_DFL_NOCOPY,
1001 };
1002 int i, err;
1003
1004 if (!wowlan->n_patterns)
1005 return 0;
1006
1007 cmd.len[0] = struct_size(pattern_cmd, patterns, wowlan->n_patterns);
1008
1009 pattern_cmd = kmalloc(cmd.len[0], GFP_KERNEL);
1010 if (!pattern_cmd)
1011 return -ENOMEM;
1012
1013 pattern_cmd->n_patterns = cpu_to_le32(wowlan->n_patterns);
1014
1015 for (i = 0; i < wowlan->n_patterns; i++) {
1016 int mask_len = DIV_ROUND_UP(wowlan->patterns[i].pattern_len, 8);
1017
1018 memcpy(&pattern_cmd->patterns[i].mask,
1019 wowlan->patterns[i].mask, mask_len);
1020 memcpy(&pattern_cmd->patterns[i].pattern,
1021 wowlan->patterns[i].pattern,
1022 wowlan->patterns[i].pattern_len);
1023 pattern_cmd->patterns[i].mask_size = mask_len;
1024 pattern_cmd->patterns[i].pattern_size =
1025 wowlan->patterns[i].pattern_len;
1026 }
1027
1028 cmd.data[0] = pattern_cmd;
1029 err = iwl_dvm_send_cmd(priv, &cmd);
1030 kfree(pattern_cmd);
1031 return err;
1032}
1033
1034int iwlagn_suspend(struct iwl_priv *priv, struct cfg80211_wowlan *wowlan)
1035{
1036 struct iwlagn_wowlan_wakeup_filter_cmd wakeup_filter_cmd;
1037 struct iwl_rxon_cmd rxon;
1038 struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS];
1039 struct iwlagn_wowlan_kek_kck_material_cmd kek_kck_cmd;
1040 struct iwlagn_wowlan_tkip_params_cmd tkip_cmd = {};
1041 struct iwlagn_d3_config_cmd d3_cfg_cmd = {
1042 /*
1043 * Program the minimum sleep time to 10 seconds, as many
1044 * platforms have issues processing a wakeup signal while
1045 * still being in the process of suspending.
1046 */
1047 .min_sleep_time = cpu_to_le32(10 * 1000 * 1000),
1048 };
1049 struct wowlan_key_data key_data = {
1050 .ctx = ctx,
1051 .bssid = ctx->active.bssid_addr,
1052 .use_rsc_tsc = false,
1053 .tkip = &tkip_cmd,
1054 .use_tkip = false,
1055 };
1056 int ret, i;
1057 u16 seq;
1058
1059 key_data.rsc_tsc = kzalloc(sizeof(*key_data.rsc_tsc), GFP_KERNEL);
1060 if (!key_data.rsc_tsc)
1061 return -ENOMEM;
1062
1063 memset(&wakeup_filter_cmd, 0, sizeof(wakeup_filter_cmd));
1064
1065 /*
1066 * We know the last used seqno, and the uCode expects to know that
1067 * one, it will increment before TX.
1068 */
1069 seq = le16_to_cpu(priv->last_seq_ctl) & IEEE80211_SCTL_SEQ;
1070 wakeup_filter_cmd.non_qos_seq = cpu_to_le16(seq);
1071
1072 /*
1073 * For QoS counters, we store the one to use next, so subtract 0x10
1074 * since the uCode will add 0x10 before using the value.
1075 */
1076 for (i = 0; i < IWL_MAX_TID_COUNT; i++) {
1077 seq = priv->tid_data[IWL_AP_ID][i].seq_number;
1078 seq -= 0x10;
1079 wakeup_filter_cmd.qos_seq[i] = cpu_to_le16(seq);
1080 }
1081
1082 if (wowlan->disconnect)
1083 wakeup_filter_cmd.enabled |=
1084 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_BEACON_MISS |
1085 IWLAGN_WOWLAN_WAKEUP_LINK_CHANGE);
1086 if (wowlan->magic_pkt)
1087 wakeup_filter_cmd.enabled |=
1088 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_MAGIC_PACKET);
1089 if (wowlan->gtk_rekey_failure)
1090 wakeup_filter_cmd.enabled |=
1091 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_GTK_REKEY_FAIL);
1092 if (wowlan->eap_identity_req)
1093 wakeup_filter_cmd.enabled |=
1094 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_EAP_IDENT_REQ);
1095 if (wowlan->four_way_handshake)
1096 wakeup_filter_cmd.enabled |=
1097 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_4WAY_HANDSHAKE);
1098 if (wowlan->n_patterns)
1099 wakeup_filter_cmd.enabled |=
1100 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_PATTERN_MATCH);
1101
1102 if (wowlan->rfkill_release)
1103 d3_cfg_cmd.wakeup_flags |=
1104 cpu_to_le32(IWLAGN_D3_WAKEUP_RFKILL);
1105
1106 iwl_scan_cancel_timeout(priv, 200);
1107
1108 memcpy(&rxon, &ctx->active, sizeof(rxon));
1109
1110 priv->ucode_loaded = false;
1111 iwl_trans_stop_device(priv->trans);
1112 ret = iwl_trans_start_hw(priv->trans);
1113 if (ret)
1114 goto out;
1115
1116 priv->wowlan = true;
1117
1118 ret = iwl_load_ucode_wait_alive(priv, IWL_UCODE_WOWLAN);
1119 if (ret)
1120 goto out;
1121
1122 /* now configure WoWLAN ucode */
1123 ret = iwl_alive_start(priv);
1124 if (ret)
1125 goto out;
1126
1127 memcpy(&ctx->staging, &rxon, sizeof(rxon));
1128 ret = iwlagn_commit_rxon(priv, ctx);
1129 if (ret)
1130 goto out;
1131
1132 ret = iwl_power_update_mode(priv, true);
1133 if (ret)
1134 goto out;
1135
1136 if (!iwlwifi_mod_params.swcrypto) {
1137 /* mark all keys clear */
1138 priv->ucode_key_table = 0;
1139 ctx->key_mapping_keys = 0;
1140
1141 /*
1142 * This needs to be unlocked due to lock ordering
1143 * constraints. Since we're in the suspend path
1144 * that isn't really a problem though.
1145 */
1146 mutex_unlock(&priv->mutex);
1147 ieee80211_iter_keys(priv->hw, ctx->vif,
1148 iwlagn_wowlan_program_keys,
1149 &key_data);
1150 mutex_lock(&priv->mutex);
1151 if (key_data.error) {
1152 ret = -EIO;
1153 goto out;
1154 }
1155
1156 if (key_data.use_rsc_tsc) {
1157 struct iwl_host_cmd rsc_tsc_cmd = {
1158 .id = REPLY_WOWLAN_TSC_RSC_PARAMS,
1159 .data[0] = key_data.rsc_tsc,
1160 .dataflags[0] = IWL_HCMD_DFL_NOCOPY,
1161 .len[0] = sizeof(*key_data.rsc_tsc),
1162 };
1163
1164 ret = iwl_dvm_send_cmd(priv, &rsc_tsc_cmd);
1165 if (ret)
1166 goto out;
1167 }
1168
1169 if (key_data.use_tkip) {
1170 ret = iwl_dvm_send_cmd_pdu(priv,
1171 REPLY_WOWLAN_TKIP_PARAMS,
1172 0, sizeof(tkip_cmd),
1173 &tkip_cmd);
1174 if (ret)
1175 goto out;
1176 }
1177
1178 if (priv->have_rekey_data) {
1179 memset(&kek_kck_cmd, 0, sizeof(kek_kck_cmd));
1180 memcpy(kek_kck_cmd.kck, priv->kck, NL80211_KCK_LEN);
1181 kek_kck_cmd.kck_len = cpu_to_le16(NL80211_KCK_LEN);
1182 memcpy(kek_kck_cmd.kek, priv->kek, NL80211_KEK_LEN);
1183 kek_kck_cmd.kek_len = cpu_to_le16(NL80211_KEK_LEN);
1184 kek_kck_cmd.replay_ctr = priv->replay_ctr;
1185
1186 ret = iwl_dvm_send_cmd_pdu(priv,
1187 REPLY_WOWLAN_KEK_KCK_MATERIAL,
1188 0, sizeof(kek_kck_cmd),
1189 &kek_kck_cmd);
1190 if (ret)
1191 goto out;
1192 }
1193 }
1194
1195 ret = iwl_dvm_send_cmd_pdu(priv, REPLY_D3_CONFIG, 0,
1196 sizeof(d3_cfg_cmd), &d3_cfg_cmd);
1197 if (ret)
1198 goto out;
1199
1200 ret = iwl_dvm_send_cmd_pdu(priv, REPLY_WOWLAN_WAKEUP_FILTER,
1201 0, sizeof(wakeup_filter_cmd),
1202 &wakeup_filter_cmd);
1203 if (ret)
1204 goto out;
1205
1206 ret = iwlagn_send_patterns(priv, wowlan);
1207 out:
1208 kfree(key_data.rsc_tsc);
1209 return ret;
1210}
1211#endif
1212
1213int iwl_dvm_send_cmd(struct iwl_priv *priv, struct iwl_host_cmd *cmd)
1214{
1215 if (iwl_is_rfkill(priv) || iwl_is_ctkill(priv)) {
1216 IWL_WARN(priv, "Not sending command - %s KILL\n",
1217 iwl_is_rfkill(priv) ? "RF" : "CT");
1218 return -EIO;
1219 }
1220
1221 if (test_bit(STATUS_FW_ERROR, &priv->status)) {
1222 IWL_ERR(priv, "Command %s failed: FW Error\n",
1223 iwl_get_cmd_string(priv->trans, cmd->id));
1224 return -EIO;
1225 }
1226
1227 /*
1228 * This can happen upon FW ASSERT: we clear the STATUS_FW_ERROR flag
1229 * in iwl_down but cancel the workers only later.
1230 */
1231 if (!priv->ucode_loaded) {
1232 IWL_ERR(priv, "Fw not loaded - dropping CMD: %x\n", cmd->id);
1233 return -EIO;
1234 }
1235
1236 /*
1237 * Synchronous commands from this op-mode must hold
1238 * the mutex, this ensures we don't try to send two
1239 * (or more) synchronous commands at a time.
1240 */
1241 if (!(cmd->flags & CMD_ASYNC))
1242 lockdep_assert_held(&priv->mutex);
1243
1244 return iwl_trans_send_cmd(priv->trans, cmd);
1245}
1246
1247int iwl_dvm_send_cmd_pdu(struct iwl_priv *priv, u8 id,
1248 u32 flags, u16 len, const void *data)
1249{
1250 struct iwl_host_cmd cmd = {
1251 .id = id,
1252 .len = { len, },
1253 .data = { data, },
1254 .flags = flags,
1255 };
1256
1257 return iwl_dvm_send_cmd(priv, &cmd);
1258}
1// SPDX-License-Identifier: GPL-2.0-only
2/******************************************************************************
3 *
4 * Copyright(c) 2008 - 2014 Intel Corporation. All rights reserved.
5 *
6 * Contact Information:
7 * Intel Linux Wireless <linuxwifi@intel.com>
8 * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
9 *
10 *****************************************************************************/
11#include <linux/etherdevice.h>
12#include <linux/kernel.h>
13#include <linux/module.h>
14#include <linux/sched.h>
15#include <net/mac80211.h>
16
17#include "iwl-io.h"
18#include "iwl-agn-hw.h"
19#include "iwl-trans.h"
20#include "iwl-modparams.h"
21
22#include "dev.h"
23#include "agn.h"
24
25int iwlagn_hw_valid_rtc_data_addr(u32 addr)
26{
27 return (addr >= IWLAGN_RTC_DATA_LOWER_BOUND) &&
28 (addr < IWLAGN_RTC_DATA_UPPER_BOUND);
29}
30
31int iwlagn_send_tx_power(struct iwl_priv *priv)
32{
33 struct iwlagn_tx_power_dbm_cmd tx_power_cmd;
34 u8 tx_ant_cfg_cmd;
35
36 if (WARN_ONCE(test_bit(STATUS_SCAN_HW, &priv->status),
37 "TX Power requested while scanning!\n"))
38 return -EAGAIN;
39
40 /* half dBm need to multiply */
41 tx_power_cmd.global_lmt = (s8)(2 * priv->tx_power_user_lmt);
42
43 if (tx_power_cmd.global_lmt > priv->nvm_data->max_tx_pwr_half_dbm) {
44 /*
45 * For the newer devices which using enhanced/extend tx power
46 * table in EEPROM, the format is in half dBm. driver need to
47 * convert to dBm format before report to mac80211.
48 * By doing so, there is a possibility of 1/2 dBm resolution
49 * lost. driver will perform "round-up" operation before
50 * reporting, but it will cause 1/2 dBm tx power over the
51 * regulatory limit. Perform the checking here, if the
52 * "tx_power_user_lmt" is higher than EEPROM value (in
53 * half-dBm format), lower the tx power based on EEPROM
54 */
55 tx_power_cmd.global_lmt =
56 priv->nvm_data->max_tx_pwr_half_dbm;
57 }
58 tx_power_cmd.flags = IWLAGN_TX_POWER_NO_CLOSED;
59 tx_power_cmd.srv_chan_lmt = IWLAGN_TX_POWER_AUTO;
60
61 if (IWL_UCODE_API(priv->fw->ucode_ver) == 1)
62 tx_ant_cfg_cmd = REPLY_TX_POWER_DBM_CMD_V1;
63 else
64 tx_ant_cfg_cmd = REPLY_TX_POWER_DBM_CMD;
65
66 return iwl_dvm_send_cmd_pdu(priv, tx_ant_cfg_cmd, 0,
67 sizeof(tx_power_cmd), &tx_power_cmd);
68}
69
70void iwlagn_temperature(struct iwl_priv *priv)
71{
72 lockdep_assert_held(&priv->statistics.lock);
73
74 /* store temperature from correct statistics (in Celsius) */
75 priv->temperature = le32_to_cpu(priv->statistics.common.temperature);
76 iwl_tt_handler(priv);
77}
78
79int iwlagn_hwrate_to_mac80211_idx(u32 rate_n_flags, enum nl80211_band band)
80{
81 int idx = 0;
82 int band_offset = 0;
83
84 /* HT rate format: mac80211 wants an MCS number, which is just LSB */
85 if (rate_n_flags & RATE_MCS_HT_MSK) {
86 idx = (rate_n_flags & 0xff);
87 return idx;
88 /* Legacy rate format, search for match in table */
89 } else {
90 if (band == NL80211_BAND_5GHZ)
91 band_offset = IWL_FIRST_OFDM_RATE;
92 for (idx = band_offset; idx < IWL_RATE_COUNT_LEGACY; idx++)
93 if (iwl_rates[idx].plcp == (rate_n_flags & 0xFF))
94 return idx - band_offset;
95 }
96
97 return -1;
98}
99
100int iwlagn_manage_ibss_station(struct iwl_priv *priv,
101 struct ieee80211_vif *vif, bool add)
102{
103 struct iwl_vif_priv *vif_priv = (void *)vif->drv_priv;
104
105 if (add)
106 return iwlagn_add_bssid_station(priv, vif_priv->ctx,
107 vif->bss_conf.bssid,
108 &vif_priv->ibss_bssid_sta_id);
109 return iwl_remove_station(priv, vif_priv->ibss_bssid_sta_id,
110 vif->bss_conf.bssid);
111}
112
113/**
114 * iwlagn_txfifo_flush: send REPLY_TXFIFO_FLUSH command to uCode
115 *
116 * pre-requirements:
117 * 1. acquire mutex before calling
118 * 2. make sure rf is on and not in exit state
119 */
120int iwlagn_txfifo_flush(struct iwl_priv *priv, u32 scd_q_msk)
121{
122 struct iwl_txfifo_flush_cmd_v3 flush_cmd_v3 = {
123 .flush_control = cpu_to_le16(IWL_DROP_ALL),
124 };
125 struct iwl_txfifo_flush_cmd_v2 flush_cmd_v2 = {
126 .flush_control = cpu_to_le16(IWL_DROP_ALL),
127 };
128
129 u32 queue_control = IWL_SCD_VO_MSK | IWL_SCD_VI_MSK |
130 IWL_SCD_BE_MSK | IWL_SCD_BK_MSK | IWL_SCD_MGMT_MSK;
131
132 if ((priv->valid_contexts != BIT(IWL_RXON_CTX_BSS)))
133 queue_control |= IWL_PAN_SCD_VO_MSK | IWL_PAN_SCD_VI_MSK |
134 IWL_PAN_SCD_BE_MSK | IWL_PAN_SCD_BK_MSK |
135 IWL_PAN_SCD_MGMT_MSK |
136 IWL_PAN_SCD_MULTICAST_MSK;
137
138 if (priv->nvm_data->sku_cap_11n_enable)
139 queue_control |= IWL_AGG_TX_QUEUE_MSK;
140
141 if (scd_q_msk)
142 queue_control = scd_q_msk;
143
144 IWL_DEBUG_INFO(priv, "queue control: 0x%x\n", queue_control);
145 flush_cmd_v3.queue_control = cpu_to_le32(queue_control);
146 flush_cmd_v2.queue_control = cpu_to_le16((u16)queue_control);
147
148 if (IWL_UCODE_API(priv->fw->ucode_ver) > 2)
149 return iwl_dvm_send_cmd_pdu(priv, REPLY_TXFIFO_FLUSH, 0,
150 sizeof(flush_cmd_v3),
151 &flush_cmd_v3);
152 return iwl_dvm_send_cmd_pdu(priv, REPLY_TXFIFO_FLUSH, 0,
153 sizeof(flush_cmd_v2), &flush_cmd_v2);
154}
155
156void iwlagn_dev_txfifo_flush(struct iwl_priv *priv)
157{
158 mutex_lock(&priv->mutex);
159 ieee80211_stop_queues(priv->hw);
160 if (iwlagn_txfifo_flush(priv, 0)) {
161 IWL_ERR(priv, "flush request fail\n");
162 goto done;
163 }
164 IWL_DEBUG_INFO(priv, "wait transmit/flush all frames\n");
165 iwl_trans_wait_tx_queues_empty(priv->trans, 0xffffffff);
166done:
167 ieee80211_wake_queues(priv->hw);
168 mutex_unlock(&priv->mutex);
169}
170
171/*
172 * BT coex
173 */
174/* Notmal TDM */
175static const __le32 iwlagn_def_3w_lookup[IWLAGN_BT_DECISION_LUT_SIZE] = {
176 cpu_to_le32(0xaaaaaaaa),
177 cpu_to_le32(0xaaaaaaaa),
178 cpu_to_le32(0xaeaaaaaa),
179 cpu_to_le32(0xaaaaaaaa),
180 cpu_to_le32(0xcc00ff28),
181 cpu_to_le32(0x0000aaaa),
182 cpu_to_le32(0xcc00aaaa),
183 cpu_to_le32(0x0000aaaa),
184 cpu_to_le32(0xc0004000),
185 cpu_to_le32(0x00004000),
186 cpu_to_le32(0xf0005000),
187 cpu_to_le32(0xf0005000),
188};
189
190/* Full concurrency */
191static const __le32 iwlagn_concurrent_lookup[IWLAGN_BT_DECISION_LUT_SIZE] = {
192 cpu_to_le32(0xaaaaaaaa),
193 cpu_to_le32(0xaaaaaaaa),
194 cpu_to_le32(0xaaaaaaaa),
195 cpu_to_le32(0xaaaaaaaa),
196 cpu_to_le32(0xaaaaaaaa),
197 cpu_to_le32(0xaaaaaaaa),
198 cpu_to_le32(0xaaaaaaaa),
199 cpu_to_le32(0xaaaaaaaa),
200 cpu_to_le32(0x00000000),
201 cpu_to_le32(0x00000000),
202 cpu_to_le32(0x00000000),
203 cpu_to_le32(0x00000000),
204};
205
206void iwlagn_send_advance_bt_config(struct iwl_priv *priv)
207{
208 struct iwl_basic_bt_cmd basic = {
209 .max_kill = IWLAGN_BT_MAX_KILL_DEFAULT,
210 .bt3_timer_t7_value = IWLAGN_BT3_T7_DEFAULT,
211 .bt3_prio_sample_time = IWLAGN_BT3_PRIO_SAMPLE_DEFAULT,
212 .bt3_timer_t2_value = IWLAGN_BT3_T2_DEFAULT,
213 };
214 struct iwl_bt_cmd_v1 bt_cmd_v1;
215 struct iwl_bt_cmd_v2 bt_cmd_v2;
216 int ret;
217
218 BUILD_BUG_ON(sizeof(iwlagn_def_3w_lookup) !=
219 sizeof(basic.bt3_lookup_table));
220
221 if (priv->lib->bt_params) {
222 /*
223 * newer generation of devices (2000 series and newer)
224 * use the version 2 of the bt command
225 * we need to make sure sending the host command
226 * with correct data structure to avoid uCode assert
227 */
228 if (priv->lib->bt_params->bt_session_2) {
229 bt_cmd_v2.prio_boost = cpu_to_le32(
230 priv->lib->bt_params->bt_prio_boost);
231 bt_cmd_v2.tx_prio_boost = 0;
232 bt_cmd_v2.rx_prio_boost = 0;
233 } else {
234 /* older version only has 8 bits */
235 WARN_ON(priv->lib->bt_params->bt_prio_boost & ~0xFF);
236 bt_cmd_v1.prio_boost =
237 priv->lib->bt_params->bt_prio_boost;
238 bt_cmd_v1.tx_prio_boost = 0;
239 bt_cmd_v1.rx_prio_boost = 0;
240 }
241 } else {
242 IWL_ERR(priv, "failed to construct BT Coex Config\n");
243 return;
244 }
245
246 /*
247 * Possible situations when BT needs to take over for receive,
248 * at the same time where STA needs to response to AP's frame(s),
249 * reduce the tx power of the required response frames, by that,
250 * allow the concurrent BT receive & WiFi transmit
251 * (BT - ANT A, WiFi -ANT B), without interference to one another
252 *
253 * Reduced tx power apply to control frames only (ACK/Back/CTS)
254 * when indicated by the BT config command
255 */
256 basic.kill_ack_mask = priv->kill_ack_mask;
257 basic.kill_cts_mask = priv->kill_cts_mask;
258 if (priv->reduced_txpower)
259 basic.reduce_txpower = IWLAGN_BT_REDUCED_TX_PWR;
260 basic.valid = priv->bt_valid;
261
262 /*
263 * Configure BT coex mode to "no coexistence" when the
264 * user disabled BT coexistence, we have no interface
265 * (might be in monitor mode), or the interface is in
266 * IBSS mode (no proper uCode support for coex then).
267 */
268 if (!iwlwifi_mod_params.bt_coex_active ||
269 priv->iw_mode == NL80211_IFTYPE_ADHOC) {
270 basic.flags = IWLAGN_BT_FLAG_COEX_MODE_DISABLED;
271 } else {
272 basic.flags = IWLAGN_BT_FLAG_COEX_MODE_3W <<
273 IWLAGN_BT_FLAG_COEX_MODE_SHIFT;
274
275 if (!priv->bt_enable_pspoll)
276 basic.flags |= IWLAGN_BT_FLAG_SYNC_2_BT_DISABLE;
277 else
278 basic.flags &= ~IWLAGN_BT_FLAG_SYNC_2_BT_DISABLE;
279
280 if (priv->bt_ch_announce)
281 basic.flags |= IWLAGN_BT_FLAG_CHANNEL_INHIBITION;
282 IWL_DEBUG_COEX(priv, "BT coex flag: 0X%x\n", basic.flags);
283 }
284 priv->bt_enable_flag = basic.flags;
285 if (priv->bt_full_concurrent)
286 memcpy(basic.bt3_lookup_table, iwlagn_concurrent_lookup,
287 sizeof(iwlagn_concurrent_lookup));
288 else
289 memcpy(basic.bt3_lookup_table, iwlagn_def_3w_lookup,
290 sizeof(iwlagn_def_3w_lookup));
291
292 IWL_DEBUG_COEX(priv, "BT coex %s in %s mode\n",
293 basic.flags ? "active" : "disabled",
294 priv->bt_full_concurrent ?
295 "full concurrency" : "3-wire");
296
297 if (priv->lib->bt_params->bt_session_2) {
298 memcpy(&bt_cmd_v2.basic, &basic,
299 sizeof(basic));
300 ret = iwl_dvm_send_cmd_pdu(priv, REPLY_BT_CONFIG,
301 0, sizeof(bt_cmd_v2), &bt_cmd_v2);
302 } else {
303 memcpy(&bt_cmd_v1.basic, &basic,
304 sizeof(basic));
305 ret = iwl_dvm_send_cmd_pdu(priv, REPLY_BT_CONFIG,
306 0, sizeof(bt_cmd_v1), &bt_cmd_v1);
307 }
308 if (ret)
309 IWL_ERR(priv, "failed to send BT Coex Config\n");
310
311}
312
313void iwlagn_bt_adjust_rssi_monitor(struct iwl_priv *priv, bool rssi_ena)
314{
315 struct iwl_rxon_context *ctx, *found_ctx = NULL;
316 bool found_ap = false;
317
318 lockdep_assert_held(&priv->mutex);
319
320 /* Check whether AP or GO mode is active. */
321 if (rssi_ena) {
322 for_each_context(priv, ctx) {
323 if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_AP &&
324 iwl_is_associated_ctx(ctx)) {
325 found_ap = true;
326 break;
327 }
328 }
329 }
330
331 /*
332 * If disable was received or If GO/AP mode, disable RSSI
333 * measurements.
334 */
335 if (!rssi_ena || found_ap) {
336 if (priv->cur_rssi_ctx) {
337 ctx = priv->cur_rssi_ctx;
338 ieee80211_disable_rssi_reports(ctx->vif);
339 priv->cur_rssi_ctx = NULL;
340 }
341 return;
342 }
343
344 /*
345 * If rssi measurements need to be enabled, consider all cases now.
346 * Figure out how many contexts are active.
347 */
348 for_each_context(priv, ctx) {
349 if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_STATION &&
350 iwl_is_associated_ctx(ctx)) {
351 found_ctx = ctx;
352 break;
353 }
354 }
355
356 /*
357 * rssi monitor already enabled for the correct interface...nothing
358 * to do.
359 */
360 if (found_ctx == priv->cur_rssi_ctx)
361 return;
362
363 /*
364 * Figure out if rssi monitor is currently enabled, and needs
365 * to be changed. If rssi monitor is already enabled, disable
366 * it first else just enable rssi measurements on the
367 * interface found above.
368 */
369 if (priv->cur_rssi_ctx) {
370 ctx = priv->cur_rssi_ctx;
371 if (ctx->vif)
372 ieee80211_disable_rssi_reports(ctx->vif);
373 }
374
375 priv->cur_rssi_ctx = found_ctx;
376
377 if (!found_ctx)
378 return;
379
380 ieee80211_enable_rssi_reports(found_ctx->vif,
381 IWLAGN_BT_PSP_MIN_RSSI_THRESHOLD,
382 IWLAGN_BT_PSP_MAX_RSSI_THRESHOLD);
383}
384
385static bool iwlagn_bt_traffic_is_sco(struct iwl_bt_uart_msg *uart_msg)
386{
387 return (BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3) >>
388 BT_UART_MSG_FRAME3SCOESCO_POS;
389}
390
391static void iwlagn_bt_traffic_change_work(struct work_struct *work)
392{
393 struct iwl_priv *priv =
394 container_of(work, struct iwl_priv, bt_traffic_change_work);
395 struct iwl_rxon_context *ctx;
396 int smps_request = -1;
397
398 if (priv->bt_enable_flag == IWLAGN_BT_FLAG_COEX_MODE_DISABLED) {
399 /* bt coex disabled */
400 return;
401 }
402
403 /*
404 * Note: bt_traffic_load can be overridden by scan complete and
405 * coex profile notifications. Ignore that since only bad consequence
406 * can be not matching debug print with actual state.
407 */
408 IWL_DEBUG_COEX(priv, "BT traffic load changes: %d\n",
409 priv->bt_traffic_load);
410
411 switch (priv->bt_traffic_load) {
412 case IWL_BT_COEX_TRAFFIC_LOAD_NONE:
413 if (priv->bt_status)
414 smps_request = IEEE80211_SMPS_DYNAMIC;
415 else
416 smps_request = IEEE80211_SMPS_AUTOMATIC;
417 break;
418 case IWL_BT_COEX_TRAFFIC_LOAD_LOW:
419 smps_request = IEEE80211_SMPS_DYNAMIC;
420 break;
421 case IWL_BT_COEX_TRAFFIC_LOAD_HIGH:
422 case IWL_BT_COEX_TRAFFIC_LOAD_CONTINUOUS:
423 smps_request = IEEE80211_SMPS_STATIC;
424 break;
425 default:
426 IWL_ERR(priv, "Invalid BT traffic load: %d\n",
427 priv->bt_traffic_load);
428 break;
429 }
430
431 mutex_lock(&priv->mutex);
432
433 /*
434 * We can not send command to firmware while scanning. When the scan
435 * complete we will schedule this work again. We do check with mutex
436 * locked to prevent new scan request to arrive. We do not check
437 * STATUS_SCANNING to avoid race when queue_work two times from
438 * different notifications, but quit and not perform any work at all.
439 */
440 if (test_bit(STATUS_SCAN_HW, &priv->status))
441 goto out;
442
443 iwl_update_chain_flags(priv);
444
445 if (smps_request != -1) {
446 priv->current_ht_config.smps = smps_request;
447 for_each_context(priv, ctx) {
448 if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_STATION)
449 ieee80211_request_smps(ctx->vif, smps_request);
450 }
451 }
452
453 /*
454 * Dynamic PS poll related functionality. Adjust RSSI measurements if
455 * necessary.
456 */
457 iwlagn_bt_coex_rssi_monitor(priv);
458out:
459 mutex_unlock(&priv->mutex);
460}
461
462/*
463 * If BT sco traffic, and RSSI monitor is enabled, move measurements to the
464 * correct interface or disable it if this is the last interface to be
465 * removed.
466 */
467void iwlagn_bt_coex_rssi_monitor(struct iwl_priv *priv)
468{
469 if (priv->bt_is_sco &&
470 priv->bt_traffic_load == IWL_BT_COEX_TRAFFIC_LOAD_CONTINUOUS)
471 iwlagn_bt_adjust_rssi_monitor(priv, true);
472 else
473 iwlagn_bt_adjust_rssi_monitor(priv, false);
474}
475
476static void iwlagn_print_uartmsg(struct iwl_priv *priv,
477 struct iwl_bt_uart_msg *uart_msg)
478{
479 IWL_DEBUG_COEX(priv, "Message Type = 0x%X, SSN = 0x%X, "
480 "Update Req = 0x%X\n",
481 (BT_UART_MSG_FRAME1MSGTYPE_MSK & uart_msg->frame1) >>
482 BT_UART_MSG_FRAME1MSGTYPE_POS,
483 (BT_UART_MSG_FRAME1SSN_MSK & uart_msg->frame1) >>
484 BT_UART_MSG_FRAME1SSN_POS,
485 (BT_UART_MSG_FRAME1UPDATEREQ_MSK & uart_msg->frame1) >>
486 BT_UART_MSG_FRAME1UPDATEREQ_POS);
487
488 IWL_DEBUG_COEX(priv, "Open connections = 0x%X, Traffic load = 0x%X, "
489 "Chl_SeqN = 0x%X, In band = 0x%X\n",
490 (BT_UART_MSG_FRAME2OPENCONNECTIONS_MSK & uart_msg->frame2) >>
491 BT_UART_MSG_FRAME2OPENCONNECTIONS_POS,
492 (BT_UART_MSG_FRAME2TRAFFICLOAD_MSK & uart_msg->frame2) >>
493 BT_UART_MSG_FRAME2TRAFFICLOAD_POS,
494 (BT_UART_MSG_FRAME2CHLSEQN_MSK & uart_msg->frame2) >>
495 BT_UART_MSG_FRAME2CHLSEQN_POS,
496 (BT_UART_MSG_FRAME2INBAND_MSK & uart_msg->frame2) >>
497 BT_UART_MSG_FRAME2INBAND_POS);
498
499 IWL_DEBUG_COEX(priv, "SCO/eSCO = 0x%X, Sniff = 0x%X, A2DP = 0x%X, "
500 "ACL = 0x%X, Master = 0x%X, OBEX = 0x%X\n",
501 (BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3) >>
502 BT_UART_MSG_FRAME3SCOESCO_POS,
503 (BT_UART_MSG_FRAME3SNIFF_MSK & uart_msg->frame3) >>
504 BT_UART_MSG_FRAME3SNIFF_POS,
505 (BT_UART_MSG_FRAME3A2DP_MSK & uart_msg->frame3) >>
506 BT_UART_MSG_FRAME3A2DP_POS,
507 (BT_UART_MSG_FRAME3ACL_MSK & uart_msg->frame3) >>
508 BT_UART_MSG_FRAME3ACL_POS,
509 (BT_UART_MSG_FRAME3MASTER_MSK & uart_msg->frame3) >>
510 BT_UART_MSG_FRAME3MASTER_POS,
511 (BT_UART_MSG_FRAME3OBEX_MSK & uart_msg->frame3) >>
512 BT_UART_MSG_FRAME3OBEX_POS);
513
514 IWL_DEBUG_COEX(priv, "Idle duration = 0x%X\n",
515 (BT_UART_MSG_FRAME4IDLEDURATION_MSK & uart_msg->frame4) >>
516 BT_UART_MSG_FRAME4IDLEDURATION_POS);
517
518 IWL_DEBUG_COEX(priv, "Tx Activity = 0x%X, Rx Activity = 0x%X, "
519 "eSCO Retransmissions = 0x%X\n",
520 (BT_UART_MSG_FRAME5TXACTIVITY_MSK & uart_msg->frame5) >>
521 BT_UART_MSG_FRAME5TXACTIVITY_POS,
522 (BT_UART_MSG_FRAME5RXACTIVITY_MSK & uart_msg->frame5) >>
523 BT_UART_MSG_FRAME5RXACTIVITY_POS,
524 (BT_UART_MSG_FRAME5ESCORETRANSMIT_MSK & uart_msg->frame5) >>
525 BT_UART_MSG_FRAME5ESCORETRANSMIT_POS);
526
527 IWL_DEBUG_COEX(priv, "Sniff Interval = 0x%X, Discoverable = 0x%X\n",
528 (BT_UART_MSG_FRAME6SNIFFINTERVAL_MSK & uart_msg->frame6) >>
529 BT_UART_MSG_FRAME6SNIFFINTERVAL_POS,
530 (BT_UART_MSG_FRAME6DISCOVERABLE_MSK & uart_msg->frame6) >>
531 BT_UART_MSG_FRAME6DISCOVERABLE_POS);
532
533 IWL_DEBUG_COEX(priv, "Sniff Activity = 0x%X, Page = "
534 "0x%X, Inquiry = 0x%X, Connectable = 0x%X\n",
535 (BT_UART_MSG_FRAME7SNIFFACTIVITY_MSK & uart_msg->frame7) >>
536 BT_UART_MSG_FRAME7SNIFFACTIVITY_POS,
537 (BT_UART_MSG_FRAME7PAGE_MSK & uart_msg->frame7) >>
538 BT_UART_MSG_FRAME7PAGE_POS,
539 (BT_UART_MSG_FRAME7INQUIRY_MSK & uart_msg->frame7) >>
540 BT_UART_MSG_FRAME7INQUIRY_POS,
541 (BT_UART_MSG_FRAME7CONNECTABLE_MSK & uart_msg->frame7) >>
542 BT_UART_MSG_FRAME7CONNECTABLE_POS);
543}
544
545static bool iwlagn_set_kill_msk(struct iwl_priv *priv,
546 struct iwl_bt_uart_msg *uart_msg)
547{
548 bool need_update = false;
549 u8 kill_msk = IWL_BT_KILL_REDUCE;
550 static const __le32 bt_kill_ack_msg[3] = {
551 IWLAGN_BT_KILL_ACK_MASK_DEFAULT,
552 IWLAGN_BT_KILL_ACK_CTS_MASK_SCO,
553 IWLAGN_BT_KILL_ACK_CTS_MASK_REDUCE};
554 static const __le32 bt_kill_cts_msg[3] = {
555 IWLAGN_BT_KILL_CTS_MASK_DEFAULT,
556 IWLAGN_BT_KILL_ACK_CTS_MASK_SCO,
557 IWLAGN_BT_KILL_ACK_CTS_MASK_REDUCE};
558
559 if (!priv->reduced_txpower)
560 kill_msk = (BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3)
561 ? IWL_BT_KILL_OVERRIDE : IWL_BT_KILL_DEFAULT;
562 if (priv->kill_ack_mask != bt_kill_ack_msg[kill_msk] ||
563 priv->kill_cts_mask != bt_kill_cts_msg[kill_msk]) {
564 priv->bt_valid |= IWLAGN_BT_VALID_KILL_ACK_MASK;
565 priv->kill_ack_mask = bt_kill_ack_msg[kill_msk];
566 priv->bt_valid |= IWLAGN_BT_VALID_KILL_CTS_MASK;
567 priv->kill_cts_mask = bt_kill_cts_msg[kill_msk];
568 need_update = true;
569 }
570 return need_update;
571}
572
573/*
574 * Upon RSSI changes, sends a bt config command with following changes
575 * 1. enable/disable "reduced control frames tx power
576 * 2. update the "kill)ack_mask" and "kill_cts_mask"
577 *
578 * If "reduced tx power" is enabled, uCode shall
579 * 1. ACK/Back/CTS rate shall reduced to 6Mbps
580 * 2. not use duplciate 20/40MHz mode
581 */
582static bool iwlagn_fill_txpower_mode(struct iwl_priv *priv,
583 struct iwl_bt_uart_msg *uart_msg)
584{
585 bool need_update = false;
586 struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS];
587 int ave_rssi;
588
589 if (!ctx->vif || (ctx->vif->type != NL80211_IFTYPE_STATION)) {
590 IWL_DEBUG_INFO(priv, "BSS ctx not active or not in sta mode\n");
591 return false;
592 }
593
594 ave_rssi = ieee80211_ave_rssi(ctx->vif);
595 if (!ave_rssi) {
596 /* no rssi data, no changes to reduce tx power */
597 IWL_DEBUG_COEX(priv, "no rssi data available\n");
598 return need_update;
599 }
600 if (!priv->reduced_txpower &&
601 !iwl_is_associated(priv, IWL_RXON_CTX_PAN) &&
602 (ave_rssi > BT_ENABLE_REDUCED_TXPOWER_THRESHOLD) &&
603 (uart_msg->frame3 & (BT_UART_MSG_FRAME3ACL_MSK |
604 BT_UART_MSG_FRAME3OBEX_MSK)) &&
605 !(uart_msg->frame3 & (BT_UART_MSG_FRAME3SCOESCO_MSK |
606 BT_UART_MSG_FRAME3SNIFF_MSK | BT_UART_MSG_FRAME3A2DP_MSK))) {
607 /* enabling reduced tx power */
608 priv->reduced_txpower = true;
609 priv->bt_valid |= IWLAGN_BT_VALID_REDUCED_TX_PWR;
610 need_update = true;
611 } else if (priv->reduced_txpower &&
612 (iwl_is_associated(priv, IWL_RXON_CTX_PAN) ||
613 (ave_rssi < BT_DISABLE_REDUCED_TXPOWER_THRESHOLD) ||
614 (uart_msg->frame3 & (BT_UART_MSG_FRAME3SCOESCO_MSK |
615 BT_UART_MSG_FRAME3SNIFF_MSK | BT_UART_MSG_FRAME3A2DP_MSK)) ||
616 !(uart_msg->frame3 & (BT_UART_MSG_FRAME3ACL_MSK |
617 BT_UART_MSG_FRAME3OBEX_MSK)))) {
618 /* disable reduced tx power */
619 priv->reduced_txpower = false;
620 priv->bt_valid |= IWLAGN_BT_VALID_REDUCED_TX_PWR;
621 need_update = true;
622 }
623
624 return need_update;
625}
626
627static void iwlagn_bt_coex_profile_notif(struct iwl_priv *priv,
628 struct iwl_rx_cmd_buffer *rxb)
629{
630 struct iwl_rx_packet *pkt = rxb_addr(rxb);
631 struct iwl_bt_coex_profile_notif *coex = (void *)pkt->data;
632 struct iwl_bt_uart_msg *uart_msg = &coex->last_bt_uart_msg;
633
634 if (priv->bt_enable_flag == IWLAGN_BT_FLAG_COEX_MODE_DISABLED) {
635 /* bt coex disabled */
636 return;
637 }
638
639 IWL_DEBUG_COEX(priv, "BT Coex notification:\n");
640 IWL_DEBUG_COEX(priv, " status: %d\n", coex->bt_status);
641 IWL_DEBUG_COEX(priv, " traffic load: %d\n", coex->bt_traffic_load);
642 IWL_DEBUG_COEX(priv, " CI compliance: %d\n",
643 coex->bt_ci_compliance);
644 iwlagn_print_uartmsg(priv, uart_msg);
645
646 priv->last_bt_traffic_load = priv->bt_traffic_load;
647 priv->bt_is_sco = iwlagn_bt_traffic_is_sco(uart_msg);
648
649 if (priv->iw_mode != NL80211_IFTYPE_ADHOC) {
650 if (priv->bt_status != coex->bt_status ||
651 priv->last_bt_traffic_load != coex->bt_traffic_load) {
652 if (coex->bt_status) {
653 /* BT on */
654 if (!priv->bt_ch_announce)
655 priv->bt_traffic_load =
656 IWL_BT_COEX_TRAFFIC_LOAD_HIGH;
657 else
658 priv->bt_traffic_load =
659 coex->bt_traffic_load;
660 } else {
661 /* BT off */
662 priv->bt_traffic_load =
663 IWL_BT_COEX_TRAFFIC_LOAD_NONE;
664 }
665 priv->bt_status = coex->bt_status;
666 queue_work(priv->workqueue,
667 &priv->bt_traffic_change_work);
668 }
669 }
670
671 /* schedule to send runtime bt_config */
672 /* check reduce power before change ack/cts kill mask */
673 if (iwlagn_fill_txpower_mode(priv, uart_msg) ||
674 iwlagn_set_kill_msk(priv, uart_msg))
675 queue_work(priv->workqueue, &priv->bt_runtime_config);
676
677
678 /* FIXME: based on notification, adjust the prio_boost */
679
680 priv->bt_ci_compliance = coex->bt_ci_compliance;
681}
682
683void iwlagn_bt_rx_handler_setup(struct iwl_priv *priv)
684{
685 priv->rx_handlers[REPLY_BT_COEX_PROFILE_NOTIF] =
686 iwlagn_bt_coex_profile_notif;
687}
688
689void iwlagn_bt_setup_deferred_work(struct iwl_priv *priv)
690{
691 INIT_WORK(&priv->bt_traffic_change_work,
692 iwlagn_bt_traffic_change_work);
693}
694
695void iwlagn_bt_cancel_deferred_work(struct iwl_priv *priv)
696{
697 cancel_work_sync(&priv->bt_traffic_change_work);
698}
699
700static bool is_single_rx_stream(struct iwl_priv *priv)
701{
702 return priv->current_ht_config.smps == IEEE80211_SMPS_STATIC ||
703 priv->current_ht_config.single_chain_sufficient;
704}
705
706#define IWL_NUM_RX_CHAINS_MULTIPLE 3
707#define IWL_NUM_RX_CHAINS_SINGLE 2
708#define IWL_NUM_IDLE_CHAINS_DUAL 2
709#define IWL_NUM_IDLE_CHAINS_SINGLE 1
710
711/*
712 * Determine how many receiver/antenna chains to use.
713 *
714 * More provides better reception via diversity. Fewer saves power
715 * at the expense of throughput, but only when not in powersave to
716 * start with.
717 *
718 * MIMO (dual stream) requires at least 2, but works better with 3.
719 * This does not determine *which* chains to use, just how many.
720 */
721static int iwl_get_active_rx_chain_count(struct iwl_priv *priv)
722{
723 if (priv->lib->bt_params &&
724 priv->lib->bt_params->advanced_bt_coexist &&
725 (priv->bt_full_concurrent ||
726 priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH)) {
727 /*
728 * only use chain 'A' in bt high traffic load or
729 * full concurrency mode
730 */
731 return IWL_NUM_RX_CHAINS_SINGLE;
732 }
733 /* # of Rx chains to use when expecting MIMO. */
734 if (is_single_rx_stream(priv))
735 return IWL_NUM_RX_CHAINS_SINGLE;
736 else
737 return IWL_NUM_RX_CHAINS_MULTIPLE;
738}
739
740/*
741 * When we are in power saving mode, unless device support spatial
742 * multiplexing power save, use the active count for rx chain count.
743 */
744static int iwl_get_idle_rx_chain_count(struct iwl_priv *priv, int active_cnt)
745{
746 /* # Rx chains when idling, depending on SMPS mode */
747 switch (priv->current_ht_config.smps) {
748 case IEEE80211_SMPS_STATIC:
749 case IEEE80211_SMPS_DYNAMIC:
750 return IWL_NUM_IDLE_CHAINS_SINGLE;
751 case IEEE80211_SMPS_AUTOMATIC:
752 case IEEE80211_SMPS_OFF:
753 return active_cnt;
754 default:
755 WARN(1, "invalid SMPS mode %d",
756 priv->current_ht_config.smps);
757 return active_cnt;
758 }
759}
760
761/* up to 4 chains */
762static u8 iwl_count_chain_bitmap(u32 chain_bitmap)
763{
764 u8 res;
765 res = (chain_bitmap & BIT(0)) >> 0;
766 res += (chain_bitmap & BIT(1)) >> 1;
767 res += (chain_bitmap & BIT(2)) >> 2;
768 res += (chain_bitmap & BIT(3)) >> 3;
769 return res;
770}
771
772/**
773 * iwlagn_set_rxon_chain - Set up Rx chain usage in "staging" RXON image
774 *
775 * Selects how many and which Rx receivers/antennas/chains to use.
776 * This should not be used for scan command ... it puts data in wrong place.
777 */
778void iwlagn_set_rxon_chain(struct iwl_priv *priv, struct iwl_rxon_context *ctx)
779{
780 bool is_single = is_single_rx_stream(priv);
781 bool is_cam = !test_bit(STATUS_POWER_PMI, &priv->status);
782 u8 idle_rx_cnt, active_rx_cnt, valid_rx_cnt;
783 u32 active_chains;
784 u16 rx_chain;
785
786 /* Tell uCode which antennas are actually connected.
787 * Before first association, we assume all antennas are connected.
788 * Just after first association, iwl_chain_noise_calibration()
789 * checks which antennas actually *are* connected. */
790 if (priv->chain_noise_data.active_chains)
791 active_chains = priv->chain_noise_data.active_chains;
792 else
793 active_chains = priv->nvm_data->valid_rx_ant;
794
795 if (priv->lib->bt_params &&
796 priv->lib->bt_params->advanced_bt_coexist &&
797 (priv->bt_full_concurrent ||
798 priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH)) {
799 /*
800 * only use chain 'A' in bt high traffic load or
801 * full concurrency mode
802 */
803 active_chains = first_antenna(active_chains);
804 }
805
806 rx_chain = active_chains << RXON_RX_CHAIN_VALID_POS;
807
808 /* How many receivers should we use? */
809 active_rx_cnt = iwl_get_active_rx_chain_count(priv);
810 idle_rx_cnt = iwl_get_idle_rx_chain_count(priv, active_rx_cnt);
811
812
813 /* correct rx chain count according hw settings
814 * and chain noise calibration
815 */
816 valid_rx_cnt = iwl_count_chain_bitmap(active_chains);
817 if (valid_rx_cnt < active_rx_cnt)
818 active_rx_cnt = valid_rx_cnt;
819
820 if (valid_rx_cnt < idle_rx_cnt)
821 idle_rx_cnt = valid_rx_cnt;
822
823 rx_chain |= active_rx_cnt << RXON_RX_CHAIN_MIMO_CNT_POS;
824 rx_chain |= idle_rx_cnt << RXON_RX_CHAIN_CNT_POS;
825
826 ctx->staging.rx_chain = cpu_to_le16(rx_chain);
827
828 if (!is_single && (active_rx_cnt >= IWL_NUM_RX_CHAINS_SINGLE) && is_cam)
829 ctx->staging.rx_chain |= RXON_RX_CHAIN_MIMO_FORCE_MSK;
830 else
831 ctx->staging.rx_chain &= ~RXON_RX_CHAIN_MIMO_FORCE_MSK;
832
833 IWL_DEBUG_ASSOC(priv, "rx_chain=0x%X active=%d idle=%d\n",
834 ctx->staging.rx_chain,
835 active_rx_cnt, idle_rx_cnt);
836
837 WARN_ON(active_rx_cnt == 0 || idle_rx_cnt == 0 ||
838 active_rx_cnt < idle_rx_cnt);
839}
840
841u8 iwl_toggle_tx_ant(struct iwl_priv *priv, u8 ant, u8 valid)
842{
843 int i;
844 u8 ind = ant;
845
846 if (priv->band == NL80211_BAND_2GHZ &&
847 priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH)
848 return 0;
849
850 for (i = 0; i < RATE_ANT_NUM - 1; i++) {
851 ind = (ind + 1) < RATE_ANT_NUM ? ind + 1 : 0;
852 if (valid & BIT(ind))
853 return ind;
854 }
855 return ant;
856}
857
858#ifdef CONFIG_PM_SLEEP
859static void iwlagn_convert_p1k(u16 *p1k, __le16 *out)
860{
861 int i;
862
863 for (i = 0; i < IWLAGN_P1K_SIZE; i++)
864 out[i] = cpu_to_le16(p1k[i]);
865}
866
867struct wowlan_key_data {
868 struct iwl_rxon_context *ctx;
869 struct iwlagn_wowlan_rsc_tsc_params_cmd *rsc_tsc;
870 struct iwlagn_wowlan_tkip_params_cmd *tkip;
871 const u8 *bssid;
872 bool error, use_rsc_tsc, use_tkip;
873};
874
875
876static void iwlagn_wowlan_program_keys(struct ieee80211_hw *hw,
877 struct ieee80211_vif *vif,
878 struct ieee80211_sta *sta,
879 struct ieee80211_key_conf *key,
880 void *_data)
881{
882 struct iwl_priv *priv = IWL_MAC80211_GET_DVM(hw);
883 struct wowlan_key_data *data = _data;
884 struct iwl_rxon_context *ctx = data->ctx;
885 struct aes_sc *aes_sc, *aes_tx_sc = NULL;
886 struct tkip_sc *tkip_sc, *tkip_tx_sc = NULL;
887 struct iwlagn_p1k_cache *rx_p1ks;
888 u8 *rx_mic_key;
889 struct ieee80211_key_seq seq;
890 u32 cur_rx_iv32 = 0;
891 u16 p1k[IWLAGN_P1K_SIZE];
892 int ret, i;
893
894 mutex_lock(&priv->mutex);
895
896 if ((key->cipher == WLAN_CIPHER_SUITE_WEP40 ||
897 key->cipher == WLAN_CIPHER_SUITE_WEP104) &&
898 !sta && !ctx->key_mapping_keys)
899 ret = iwl_set_default_wep_key(priv, ctx, key);
900 else
901 ret = iwl_set_dynamic_key(priv, ctx, key, sta);
902
903 if (ret) {
904 IWL_ERR(priv, "Error setting key during suspend!\n");
905 data->error = true;
906 }
907
908 switch (key->cipher) {
909 case WLAN_CIPHER_SUITE_TKIP:
910 if (sta) {
911 u64 pn64;
912
913 tkip_sc = data->rsc_tsc->all_tsc_rsc.tkip.unicast_rsc;
914 tkip_tx_sc = &data->rsc_tsc->all_tsc_rsc.tkip.tsc;
915
916 rx_p1ks = data->tkip->rx_uni;
917
918 pn64 = atomic64_read(&key->tx_pn);
919 tkip_tx_sc->iv16 = cpu_to_le16(TKIP_PN_TO_IV16(pn64));
920 tkip_tx_sc->iv32 = cpu_to_le32(TKIP_PN_TO_IV32(pn64));
921
922 ieee80211_get_tkip_p1k_iv(key, seq.tkip.iv32, p1k);
923 iwlagn_convert_p1k(p1k, data->tkip->tx.p1k);
924
925 memcpy(data->tkip->mic_keys.tx,
926 &key->key[NL80211_TKIP_DATA_OFFSET_TX_MIC_KEY],
927 IWLAGN_MIC_KEY_SIZE);
928
929 rx_mic_key = data->tkip->mic_keys.rx_unicast;
930 } else {
931 tkip_sc =
932 data->rsc_tsc->all_tsc_rsc.tkip.multicast_rsc;
933 rx_p1ks = data->tkip->rx_multi;
934 rx_mic_key = data->tkip->mic_keys.rx_mcast;
935 }
936
937 /*
938 * For non-QoS this relies on the fact that both the uCode and
939 * mac80211 use TID 0 (as they need to to avoid replay attacks)
940 * for checking the IV in the frames.
941 */
942 for (i = 0; i < IWLAGN_NUM_RSC; i++) {
943 ieee80211_get_key_rx_seq(key, i, &seq);
944 tkip_sc[i].iv16 = cpu_to_le16(seq.tkip.iv16);
945 tkip_sc[i].iv32 = cpu_to_le32(seq.tkip.iv32);
946 /* wrapping isn't allowed, AP must rekey */
947 if (seq.tkip.iv32 > cur_rx_iv32)
948 cur_rx_iv32 = seq.tkip.iv32;
949 }
950
951 ieee80211_get_tkip_rx_p1k(key, data->bssid, cur_rx_iv32, p1k);
952 iwlagn_convert_p1k(p1k, rx_p1ks[0].p1k);
953 ieee80211_get_tkip_rx_p1k(key, data->bssid,
954 cur_rx_iv32 + 1, p1k);
955 iwlagn_convert_p1k(p1k, rx_p1ks[1].p1k);
956
957 memcpy(rx_mic_key,
958 &key->key[NL80211_TKIP_DATA_OFFSET_RX_MIC_KEY],
959 IWLAGN_MIC_KEY_SIZE);
960
961 data->use_tkip = true;
962 data->use_rsc_tsc = true;
963 break;
964 case WLAN_CIPHER_SUITE_CCMP:
965 if (sta) {
966 u64 pn64;
967
968 aes_sc = data->rsc_tsc->all_tsc_rsc.aes.unicast_rsc;
969 aes_tx_sc = &data->rsc_tsc->all_tsc_rsc.aes.tsc;
970
971 pn64 = atomic64_read(&key->tx_pn);
972 aes_tx_sc->pn = cpu_to_le64(pn64);
973 } else
974 aes_sc = data->rsc_tsc->all_tsc_rsc.aes.multicast_rsc;
975
976 /*
977 * For non-QoS this relies on the fact that both the uCode and
978 * mac80211 use TID 0 for checking the IV in the frames.
979 */
980 for (i = 0; i < IWLAGN_NUM_RSC; i++) {
981 u8 *pn = seq.ccmp.pn;
982
983 ieee80211_get_key_rx_seq(key, i, &seq);
984 aes_sc[i].pn = cpu_to_le64(
985 (u64)pn[5] |
986 ((u64)pn[4] << 8) |
987 ((u64)pn[3] << 16) |
988 ((u64)pn[2] << 24) |
989 ((u64)pn[1] << 32) |
990 ((u64)pn[0] << 40));
991 }
992 data->use_rsc_tsc = true;
993 break;
994 }
995
996 mutex_unlock(&priv->mutex);
997}
998
999int iwlagn_send_patterns(struct iwl_priv *priv,
1000 struct cfg80211_wowlan *wowlan)
1001{
1002 struct iwlagn_wowlan_patterns_cmd *pattern_cmd;
1003 struct iwl_host_cmd cmd = {
1004 .id = REPLY_WOWLAN_PATTERNS,
1005 .dataflags[0] = IWL_HCMD_DFL_NOCOPY,
1006 };
1007 int i, err;
1008
1009 if (!wowlan->n_patterns)
1010 return 0;
1011
1012 cmd.len[0] = struct_size(pattern_cmd, patterns, wowlan->n_patterns);
1013
1014 pattern_cmd = kmalloc(cmd.len[0], GFP_KERNEL);
1015 if (!pattern_cmd)
1016 return -ENOMEM;
1017
1018 pattern_cmd->n_patterns = cpu_to_le32(wowlan->n_patterns);
1019
1020 for (i = 0; i < wowlan->n_patterns; i++) {
1021 int mask_len = DIV_ROUND_UP(wowlan->patterns[i].pattern_len, 8);
1022
1023 memcpy(&pattern_cmd->patterns[i].mask,
1024 wowlan->patterns[i].mask, mask_len);
1025 memcpy(&pattern_cmd->patterns[i].pattern,
1026 wowlan->patterns[i].pattern,
1027 wowlan->patterns[i].pattern_len);
1028 pattern_cmd->patterns[i].mask_size = mask_len;
1029 pattern_cmd->patterns[i].pattern_size =
1030 wowlan->patterns[i].pattern_len;
1031 }
1032
1033 cmd.data[0] = pattern_cmd;
1034 err = iwl_dvm_send_cmd(priv, &cmd);
1035 kfree(pattern_cmd);
1036 return err;
1037}
1038
1039int iwlagn_suspend(struct iwl_priv *priv, struct cfg80211_wowlan *wowlan)
1040{
1041 struct iwlagn_wowlan_wakeup_filter_cmd wakeup_filter_cmd;
1042 struct iwl_rxon_cmd rxon;
1043 struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS];
1044 struct iwlagn_wowlan_kek_kck_material_cmd kek_kck_cmd;
1045 struct iwlagn_wowlan_tkip_params_cmd tkip_cmd = {};
1046 struct iwlagn_d3_config_cmd d3_cfg_cmd = {
1047 /*
1048 * Program the minimum sleep time to 10 seconds, as many
1049 * platforms have issues processing a wakeup signal while
1050 * still being in the process of suspending.
1051 */
1052 .min_sleep_time = cpu_to_le32(10 * 1000 * 1000),
1053 };
1054 struct wowlan_key_data key_data = {
1055 .ctx = ctx,
1056 .bssid = ctx->active.bssid_addr,
1057 .use_rsc_tsc = false,
1058 .tkip = &tkip_cmd,
1059 .use_tkip = false,
1060 };
1061 int ret, i;
1062 u16 seq;
1063
1064 key_data.rsc_tsc = kzalloc(sizeof(*key_data.rsc_tsc), GFP_KERNEL);
1065 if (!key_data.rsc_tsc)
1066 return -ENOMEM;
1067
1068 memset(&wakeup_filter_cmd, 0, sizeof(wakeup_filter_cmd));
1069
1070 /*
1071 * We know the last used seqno, and the uCode expects to know that
1072 * one, it will increment before TX.
1073 */
1074 seq = le16_to_cpu(priv->last_seq_ctl) & IEEE80211_SCTL_SEQ;
1075 wakeup_filter_cmd.non_qos_seq = cpu_to_le16(seq);
1076
1077 /*
1078 * For QoS counters, we store the one to use next, so subtract 0x10
1079 * since the uCode will add 0x10 before using the value.
1080 */
1081 for (i = 0; i < IWL_MAX_TID_COUNT; i++) {
1082 seq = priv->tid_data[IWL_AP_ID][i].seq_number;
1083 seq -= 0x10;
1084 wakeup_filter_cmd.qos_seq[i] = cpu_to_le16(seq);
1085 }
1086
1087 if (wowlan->disconnect)
1088 wakeup_filter_cmd.enabled |=
1089 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_BEACON_MISS |
1090 IWLAGN_WOWLAN_WAKEUP_LINK_CHANGE);
1091 if (wowlan->magic_pkt)
1092 wakeup_filter_cmd.enabled |=
1093 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_MAGIC_PACKET);
1094 if (wowlan->gtk_rekey_failure)
1095 wakeup_filter_cmd.enabled |=
1096 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_GTK_REKEY_FAIL);
1097 if (wowlan->eap_identity_req)
1098 wakeup_filter_cmd.enabled |=
1099 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_EAP_IDENT_REQ);
1100 if (wowlan->four_way_handshake)
1101 wakeup_filter_cmd.enabled |=
1102 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_4WAY_HANDSHAKE);
1103 if (wowlan->n_patterns)
1104 wakeup_filter_cmd.enabled |=
1105 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_PATTERN_MATCH);
1106
1107 if (wowlan->rfkill_release)
1108 d3_cfg_cmd.wakeup_flags |=
1109 cpu_to_le32(IWLAGN_D3_WAKEUP_RFKILL);
1110
1111 iwl_scan_cancel_timeout(priv, 200);
1112
1113 memcpy(&rxon, &ctx->active, sizeof(rxon));
1114
1115 priv->ucode_loaded = false;
1116 iwl_trans_stop_device(priv->trans);
1117 ret = iwl_trans_start_hw(priv->trans);
1118 if (ret)
1119 goto out;
1120
1121 priv->wowlan = true;
1122
1123 ret = iwl_load_ucode_wait_alive(priv, IWL_UCODE_WOWLAN);
1124 if (ret)
1125 goto out;
1126
1127 /* now configure WoWLAN ucode */
1128 ret = iwl_alive_start(priv);
1129 if (ret)
1130 goto out;
1131
1132 memcpy(&ctx->staging, &rxon, sizeof(rxon));
1133 ret = iwlagn_commit_rxon(priv, ctx);
1134 if (ret)
1135 goto out;
1136
1137 ret = iwl_power_update_mode(priv, true);
1138 if (ret)
1139 goto out;
1140
1141 if (!iwlwifi_mod_params.swcrypto) {
1142 /* mark all keys clear */
1143 priv->ucode_key_table = 0;
1144 ctx->key_mapping_keys = 0;
1145
1146 /*
1147 * This needs to be unlocked due to lock ordering
1148 * constraints. Since we're in the suspend path
1149 * that isn't really a problem though.
1150 */
1151 mutex_unlock(&priv->mutex);
1152 ieee80211_iter_keys(priv->hw, ctx->vif,
1153 iwlagn_wowlan_program_keys,
1154 &key_data);
1155 mutex_lock(&priv->mutex);
1156 if (key_data.error) {
1157 ret = -EIO;
1158 goto out;
1159 }
1160
1161 if (key_data.use_rsc_tsc) {
1162 struct iwl_host_cmd rsc_tsc_cmd = {
1163 .id = REPLY_WOWLAN_TSC_RSC_PARAMS,
1164 .data[0] = key_data.rsc_tsc,
1165 .dataflags[0] = IWL_HCMD_DFL_NOCOPY,
1166 .len[0] = sizeof(*key_data.rsc_tsc),
1167 };
1168
1169 ret = iwl_dvm_send_cmd(priv, &rsc_tsc_cmd);
1170 if (ret)
1171 goto out;
1172 }
1173
1174 if (key_data.use_tkip) {
1175 ret = iwl_dvm_send_cmd_pdu(priv,
1176 REPLY_WOWLAN_TKIP_PARAMS,
1177 0, sizeof(tkip_cmd),
1178 &tkip_cmd);
1179 if (ret)
1180 goto out;
1181 }
1182
1183 if (priv->have_rekey_data) {
1184 memset(&kek_kck_cmd, 0, sizeof(kek_kck_cmd));
1185 memcpy(kek_kck_cmd.kck, priv->kck, NL80211_KCK_LEN);
1186 kek_kck_cmd.kck_len = cpu_to_le16(NL80211_KCK_LEN);
1187 memcpy(kek_kck_cmd.kek, priv->kek, NL80211_KEK_LEN);
1188 kek_kck_cmd.kek_len = cpu_to_le16(NL80211_KEK_LEN);
1189 kek_kck_cmd.replay_ctr = priv->replay_ctr;
1190
1191 ret = iwl_dvm_send_cmd_pdu(priv,
1192 REPLY_WOWLAN_KEK_KCK_MATERIAL,
1193 0, sizeof(kek_kck_cmd),
1194 &kek_kck_cmd);
1195 if (ret)
1196 goto out;
1197 }
1198 }
1199
1200 ret = iwl_dvm_send_cmd_pdu(priv, REPLY_D3_CONFIG, 0,
1201 sizeof(d3_cfg_cmd), &d3_cfg_cmd);
1202 if (ret)
1203 goto out;
1204
1205 ret = iwl_dvm_send_cmd_pdu(priv, REPLY_WOWLAN_WAKEUP_FILTER,
1206 0, sizeof(wakeup_filter_cmd),
1207 &wakeup_filter_cmd);
1208 if (ret)
1209 goto out;
1210
1211 ret = iwlagn_send_patterns(priv, wowlan);
1212 out:
1213 kfree(key_data.rsc_tsc);
1214 return ret;
1215}
1216#endif
1217
1218int iwl_dvm_send_cmd(struct iwl_priv *priv, struct iwl_host_cmd *cmd)
1219{
1220 if (iwl_is_rfkill(priv) || iwl_is_ctkill(priv)) {
1221 IWL_WARN(priv, "Not sending command - %s KILL\n",
1222 iwl_is_rfkill(priv) ? "RF" : "CT");
1223 return -EIO;
1224 }
1225
1226 if (test_bit(STATUS_FW_ERROR, &priv->status)) {
1227 IWL_ERR(priv, "Command %s failed: FW Error\n",
1228 iwl_get_cmd_string(priv->trans, cmd->id));
1229 return -EIO;
1230 }
1231
1232 /*
1233 * This can happen upon FW ASSERT: we clear the STATUS_FW_ERROR flag
1234 * in iwl_down but cancel the workers only later.
1235 */
1236 if (!priv->ucode_loaded) {
1237 IWL_ERR(priv, "Fw not loaded - dropping CMD: %x\n", cmd->id);
1238 return -EIO;
1239 }
1240
1241 /*
1242 * Synchronous commands from this op-mode must hold
1243 * the mutex, this ensures we don't try to send two
1244 * (or more) synchronous commands at a time.
1245 */
1246 if (!(cmd->flags & CMD_ASYNC))
1247 lockdep_assert_held(&priv->mutex);
1248
1249 return iwl_trans_send_cmd(priv->trans, cmd);
1250}
1251
1252int iwl_dvm_send_cmd_pdu(struct iwl_priv *priv, u8 id,
1253 u32 flags, u16 len, const void *data)
1254{
1255 struct iwl_host_cmd cmd = {
1256 .id = id,
1257 .len = { len, },
1258 .data = { data, },
1259 .flags = flags,
1260 };
1261
1262 return iwl_dvm_send_cmd(priv, &cmd);
1263}