1 // SPDX-License-Identifier: GPL-2.0-only
2 /******************************************************************************
3  *
4  * Copyright(c) 2008 - 2014 Intel Corporation. All rights reserved.
5  *
6  * Contact Information:
7  *  Intel Linux Wireless <linuxwifi@intel.com>
8  * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
9  *
10  *****************************************************************************/
11 #include <linux/etherdevice.h>
12 #include <linux/kernel.h>
13 #include <linux/module.h>
14 #include <linux/sched.h>
15 #include <net/mac80211.h>
16 
17 #include "iwl-io.h"
18 #include "iwl-agn-hw.h"
19 #include "iwl-trans.h"
20 #include "iwl-modparams.h"
21 
22 #include "dev.h"
23 #include "agn.h"
24 
iwlagn_hw_valid_rtc_data_addr(u32 addr)25 int iwlagn_hw_valid_rtc_data_addr(u32 addr)
26 {
27 	return (addr >= IWLAGN_RTC_DATA_LOWER_BOUND) &&
28 		(addr < IWLAGN_RTC_DATA_UPPER_BOUND);
29 }
30 
iwlagn_send_tx_power(struct iwl_priv * priv)31 int iwlagn_send_tx_power(struct iwl_priv *priv)
32 {
33 	struct iwlagn_tx_power_dbm_cmd tx_power_cmd;
34 	u8 tx_ant_cfg_cmd;
35 
36 	if (WARN_ONCE(test_bit(STATUS_SCAN_HW, &priv->status),
37 		      "TX Power requested while scanning!\n"))
38 		return -EAGAIN;
39 
40 	/* half dBm need to multiply */
41 	tx_power_cmd.global_lmt = (s8)(2 * priv->tx_power_user_lmt);
42 
43 	if (tx_power_cmd.global_lmt > priv->nvm_data->max_tx_pwr_half_dbm) {
44 		/*
45 		 * For the newer devices which using enhanced/extend tx power
46 		 * table in EEPROM, the format is in half dBm. driver need to
47 		 * convert to dBm format before report to mac80211.
48 		 * By doing so, there is a possibility of 1/2 dBm resolution
49 		 * lost. driver will perform "round-up" operation before
50 		 * reporting, but it will cause 1/2 dBm tx power over the
51 		 * regulatory limit. Perform the checking here, if the
52 		 * "tx_power_user_lmt" is higher than EEPROM value (in
53 		 * half-dBm format), lower the tx power based on EEPROM
54 		 */
55 		tx_power_cmd.global_lmt =
56 			priv->nvm_data->max_tx_pwr_half_dbm;
57 	}
58 	tx_power_cmd.flags = IWLAGN_TX_POWER_NO_CLOSED;
59 	tx_power_cmd.srv_chan_lmt = IWLAGN_TX_POWER_AUTO;
60 
61 	if (IWL_UCODE_API(priv->fw->ucode_ver) == 1)
62 		tx_ant_cfg_cmd = REPLY_TX_POWER_DBM_CMD_V1;
63 	else
64 		tx_ant_cfg_cmd = REPLY_TX_POWER_DBM_CMD;
65 
66 	return iwl_dvm_send_cmd_pdu(priv, tx_ant_cfg_cmd, 0,
67 			sizeof(tx_power_cmd), &tx_power_cmd);
68 }
69 
iwlagn_temperature(struct iwl_priv * priv)70 void iwlagn_temperature(struct iwl_priv *priv)
71 {
72 	lockdep_assert_held(&priv->statistics.lock);
73 
74 	/* store temperature from correct statistics (in Celsius) */
75 	priv->temperature = le32_to_cpu(priv->statistics.common.temperature);
76 	iwl_tt_handler(priv);
77 }
78 
iwlagn_hwrate_to_mac80211_idx(u32 rate_n_flags,enum nl80211_band band)79 int iwlagn_hwrate_to_mac80211_idx(u32 rate_n_flags, enum nl80211_band band)
80 {
81 	int idx = 0;
82 	int band_offset = 0;
83 
84 	/* HT rate format: mac80211 wants an MCS number, which is just LSB */
85 	if (rate_n_flags & RATE_MCS_HT_MSK) {
86 		idx = (rate_n_flags & 0xff);
87 		return idx;
88 	/* Legacy rate format, search for match in table */
89 	} else {
90 		if (band == NL80211_BAND_5GHZ)
91 			band_offset = IWL_FIRST_OFDM_RATE;
92 		for (idx = band_offset; idx < IWL_RATE_COUNT_LEGACY; idx++)
93 			if (iwl_rates[idx].plcp == (rate_n_flags & 0xFF))
94 				return idx - band_offset;
95 	}
96 
97 	return -1;
98 }
99 
iwlagn_manage_ibss_station(struct iwl_priv * priv,struct ieee80211_vif * vif,bool add)100 int iwlagn_manage_ibss_station(struct iwl_priv *priv,
101 			       struct ieee80211_vif *vif, bool add)
102 {
103 	struct iwl_vif_priv *vif_priv = (void *)vif->drv_priv;
104 
105 	if (add)
106 		return iwlagn_add_bssid_station(priv, vif_priv->ctx,
107 						vif->bss_conf.bssid,
108 						&vif_priv->ibss_bssid_sta_id);
109 	return iwl_remove_station(priv, vif_priv->ibss_bssid_sta_id,
110 				  vif->bss_conf.bssid);
111 }
112 
113 /*
114  * iwlagn_txfifo_flush: send REPLY_TXFIFO_FLUSH command to uCode
115  *
116  * pre-requirements:
117  *  1. acquire mutex before calling
118  *  2. make sure rf is on and not in exit state
119  */
iwlagn_txfifo_flush(struct iwl_priv * priv,u32 scd_q_msk)120 int iwlagn_txfifo_flush(struct iwl_priv *priv, u32 scd_q_msk)
121 {
122 	struct iwl_txfifo_flush_cmd_v3 flush_cmd_v3 = {
123 		.flush_control = cpu_to_le16(IWL_DROP_ALL),
124 	};
125 	struct iwl_txfifo_flush_cmd_v2 flush_cmd_v2 = {
126 		.flush_control = cpu_to_le16(IWL_DROP_ALL),
127 	};
128 
129 	u32 queue_control = IWL_SCD_VO_MSK | IWL_SCD_VI_MSK |
130 			    IWL_SCD_BE_MSK | IWL_SCD_BK_MSK | IWL_SCD_MGMT_MSK;
131 
132 	if ((priv->valid_contexts != BIT(IWL_RXON_CTX_BSS)))
133 		queue_control |= IWL_PAN_SCD_VO_MSK | IWL_PAN_SCD_VI_MSK |
134 				 IWL_PAN_SCD_BE_MSK | IWL_PAN_SCD_BK_MSK |
135 				 IWL_PAN_SCD_MGMT_MSK |
136 				 IWL_PAN_SCD_MULTICAST_MSK;
137 
138 	if (priv->nvm_data->sku_cap_11n_enable)
139 		queue_control |= IWL_AGG_TX_QUEUE_MSK;
140 
141 	if (scd_q_msk)
142 		queue_control = scd_q_msk;
143 
144 	IWL_DEBUG_INFO(priv, "queue control: 0x%x\n", queue_control);
145 	flush_cmd_v3.queue_control = cpu_to_le32(queue_control);
146 	flush_cmd_v2.queue_control = cpu_to_le16((u16)queue_control);
147 
148 	if (IWL_UCODE_API(priv->fw->ucode_ver) > 2)
149 		return iwl_dvm_send_cmd_pdu(priv, REPLY_TXFIFO_FLUSH, 0,
150 					    sizeof(flush_cmd_v3),
151 					    &flush_cmd_v3);
152 	return iwl_dvm_send_cmd_pdu(priv, REPLY_TXFIFO_FLUSH, 0,
153 				    sizeof(flush_cmd_v2), &flush_cmd_v2);
154 }
155 
iwlagn_dev_txfifo_flush(struct iwl_priv * priv)156 void iwlagn_dev_txfifo_flush(struct iwl_priv *priv)
157 {
158 	mutex_lock(&priv->mutex);
159 	ieee80211_stop_queues(priv->hw);
160 	if (iwlagn_txfifo_flush(priv, 0)) {
161 		IWL_ERR(priv, "flush request fail\n");
162 		goto done;
163 	}
164 	IWL_DEBUG_INFO(priv, "wait transmit/flush all frames\n");
165 	iwl_trans_wait_tx_queues_empty(priv->trans, 0xffffffff);
166 done:
167 	ieee80211_wake_queues(priv->hw);
168 	mutex_unlock(&priv->mutex);
169 }
170 
171 /*
172  * BT coex
173  */
174 /* Notmal TDM */
175 static const __le32 iwlagn_def_3w_lookup[IWLAGN_BT_DECISION_LUT_SIZE] = {
176 	cpu_to_le32(0xaaaaaaaa),
177 	cpu_to_le32(0xaaaaaaaa),
178 	cpu_to_le32(0xaeaaaaaa),
179 	cpu_to_le32(0xaaaaaaaa),
180 	cpu_to_le32(0xcc00ff28),
181 	cpu_to_le32(0x0000aaaa),
182 	cpu_to_le32(0xcc00aaaa),
183 	cpu_to_le32(0x0000aaaa),
184 	cpu_to_le32(0xc0004000),
185 	cpu_to_le32(0x00004000),
186 	cpu_to_le32(0xf0005000),
187 	cpu_to_le32(0xf0005000),
188 };
189 
190 /* Full concurrency */
191 static const __le32 iwlagn_concurrent_lookup[IWLAGN_BT_DECISION_LUT_SIZE] = {
192 	cpu_to_le32(0xaaaaaaaa),
193 	cpu_to_le32(0xaaaaaaaa),
194 	cpu_to_le32(0xaaaaaaaa),
195 	cpu_to_le32(0xaaaaaaaa),
196 	cpu_to_le32(0xaaaaaaaa),
197 	cpu_to_le32(0xaaaaaaaa),
198 	cpu_to_le32(0xaaaaaaaa),
199 	cpu_to_le32(0xaaaaaaaa),
200 	cpu_to_le32(0x00000000),
201 	cpu_to_le32(0x00000000),
202 	cpu_to_le32(0x00000000),
203 	cpu_to_le32(0x00000000),
204 };
205 
iwlagn_send_advance_bt_config(struct iwl_priv * priv)206 void iwlagn_send_advance_bt_config(struct iwl_priv *priv)
207 {
208 	struct iwl_basic_bt_cmd basic = {
209 		.max_kill = IWLAGN_BT_MAX_KILL_DEFAULT,
210 		.bt3_timer_t7_value = IWLAGN_BT3_T7_DEFAULT,
211 		.bt3_prio_sample_time = IWLAGN_BT3_PRIO_SAMPLE_DEFAULT,
212 		.bt3_timer_t2_value = IWLAGN_BT3_T2_DEFAULT,
213 	};
214 	struct iwl_bt_cmd_v1 bt_cmd_v1;
215 	struct iwl_bt_cmd_v2 bt_cmd_v2;
216 	int ret;
217 
218 	BUILD_BUG_ON(sizeof(iwlagn_def_3w_lookup) !=
219 			sizeof(basic.bt3_lookup_table));
220 
221 	if (priv->lib->bt_params) {
222 		/*
223 		 * newer generation of devices (2000 series and newer)
224 		 * use the version 2 of the bt command
225 		 * we need to make sure sending the host command
226 		 * with correct data structure to avoid uCode assert
227 		 */
228 		if (priv->lib->bt_params->bt_session_2) {
229 			bt_cmd_v2.prio_boost = cpu_to_le32(
230 				priv->lib->bt_params->bt_prio_boost);
231 			bt_cmd_v2.tx_prio_boost = 0;
232 			bt_cmd_v2.rx_prio_boost = 0;
233 		} else {
234 			/* older version only has 8 bits */
235 			WARN_ON(priv->lib->bt_params->bt_prio_boost & ~0xFF);
236 			bt_cmd_v1.prio_boost =
237 				priv->lib->bt_params->bt_prio_boost;
238 			bt_cmd_v1.tx_prio_boost = 0;
239 			bt_cmd_v1.rx_prio_boost = 0;
240 		}
241 	} else {
242 		IWL_ERR(priv, "failed to construct BT Coex Config\n");
243 		return;
244 	}
245 
246 	/*
247 	 * Possible situations when BT needs to take over for receive,
248 	 * at the same time where STA needs to response to AP's frame(s),
249 	 * reduce the tx power of the required response frames, by that,
250 	 * allow the concurrent BT receive & WiFi transmit
251 	 * (BT - ANT A, WiFi -ANT B), without interference to one another
252 	 *
253 	 * Reduced tx power apply to control frames only (ACK/Back/CTS)
254 	 * when indicated by the BT config command
255 	 */
256 	basic.kill_ack_mask = priv->kill_ack_mask;
257 	basic.kill_cts_mask = priv->kill_cts_mask;
258 	if (priv->reduced_txpower)
259 		basic.reduce_txpower = IWLAGN_BT_REDUCED_TX_PWR;
260 	basic.valid = priv->bt_valid;
261 
262 	/*
263 	 * Configure BT coex mode to "no coexistence" when the
264 	 * user disabled BT coexistence, we have no interface
265 	 * (might be in monitor mode), or the interface is in
266 	 * IBSS mode (no proper uCode support for coex then).
267 	 */
268 	if (!iwlwifi_mod_params.bt_coex_active ||
269 	    priv->iw_mode == NL80211_IFTYPE_ADHOC) {
270 		basic.flags = IWLAGN_BT_FLAG_COEX_MODE_DISABLED;
271 	} else {
272 		basic.flags = IWLAGN_BT_FLAG_COEX_MODE_3W <<
273 					IWLAGN_BT_FLAG_COEX_MODE_SHIFT;
274 
275 		if (!priv->bt_enable_pspoll)
276 			basic.flags |= IWLAGN_BT_FLAG_SYNC_2_BT_DISABLE;
277 		else
278 			basic.flags &= ~IWLAGN_BT_FLAG_SYNC_2_BT_DISABLE;
279 
280 		if (priv->bt_ch_announce)
281 			basic.flags |= IWLAGN_BT_FLAG_CHANNEL_INHIBITION;
282 		IWL_DEBUG_COEX(priv, "BT coex flag: 0X%x\n", basic.flags);
283 	}
284 	priv->bt_enable_flag = basic.flags;
285 	if (priv->bt_full_concurrent)
286 		memcpy(basic.bt3_lookup_table, iwlagn_concurrent_lookup,
287 			sizeof(iwlagn_concurrent_lookup));
288 	else
289 		memcpy(basic.bt3_lookup_table, iwlagn_def_3w_lookup,
290 			sizeof(iwlagn_def_3w_lookup));
291 
292 	IWL_DEBUG_COEX(priv, "BT coex %s in %s mode\n",
293 		       basic.flags ? "active" : "disabled",
294 		       priv->bt_full_concurrent ?
295 		       "full concurrency" : "3-wire");
296 
297 	if (priv->lib->bt_params->bt_session_2) {
298 		memcpy(&bt_cmd_v2.basic, &basic,
299 			sizeof(basic));
300 		ret = iwl_dvm_send_cmd_pdu(priv, REPLY_BT_CONFIG,
301 			0, sizeof(bt_cmd_v2), &bt_cmd_v2);
302 	} else {
303 		memcpy(&bt_cmd_v1.basic, &basic,
304 			sizeof(basic));
305 		ret = iwl_dvm_send_cmd_pdu(priv, REPLY_BT_CONFIG,
306 			0, sizeof(bt_cmd_v1), &bt_cmd_v1);
307 	}
308 	if (ret)
309 		IWL_ERR(priv, "failed to send BT Coex Config\n");
310 
311 }
312 
iwlagn_bt_adjust_rssi_monitor(struct iwl_priv * priv,bool rssi_ena)313 void iwlagn_bt_adjust_rssi_monitor(struct iwl_priv *priv, bool rssi_ena)
314 {
315 	struct iwl_rxon_context *ctx, *found_ctx = NULL;
316 	bool found_ap = false;
317 
318 	lockdep_assert_held(&priv->mutex);
319 
320 	/* Check whether AP or GO mode is active. */
321 	if (rssi_ena) {
322 		for_each_context(priv, ctx) {
323 			if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_AP &&
324 			    iwl_is_associated_ctx(ctx)) {
325 				found_ap = true;
326 				break;
327 			}
328 		}
329 	}
330 
331 	/*
332 	 * If disable was received or If GO/AP mode, disable RSSI
333 	 * measurements.
334 	 */
335 	if (!rssi_ena || found_ap) {
336 		if (priv->cur_rssi_ctx) {
337 			ctx = priv->cur_rssi_ctx;
338 			ieee80211_disable_rssi_reports(ctx->vif);
339 			priv->cur_rssi_ctx = NULL;
340 		}
341 		return;
342 	}
343 
344 	/*
345 	 * If rssi measurements need to be enabled, consider all cases now.
346 	 * Figure out how many contexts are active.
347 	 */
348 	for_each_context(priv, ctx) {
349 		if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_STATION &&
350 		    iwl_is_associated_ctx(ctx)) {
351 			found_ctx = ctx;
352 			break;
353 		}
354 	}
355 
356 	/*
357 	 * rssi monitor already enabled for the correct interface...nothing
358 	 * to do.
359 	 */
360 	if (found_ctx == priv->cur_rssi_ctx)
361 		return;
362 
363 	/*
364 	 * Figure out if rssi monitor is currently enabled, and needs
365 	 * to be changed. If rssi monitor is already enabled, disable
366 	 * it first else just enable rssi measurements on the
367 	 * interface found above.
368 	 */
369 	if (priv->cur_rssi_ctx) {
370 		ctx = priv->cur_rssi_ctx;
371 		if (ctx->vif)
372 			ieee80211_disable_rssi_reports(ctx->vif);
373 	}
374 
375 	priv->cur_rssi_ctx = found_ctx;
376 
377 	if (!found_ctx)
378 		return;
379 
380 	ieee80211_enable_rssi_reports(found_ctx->vif,
381 			IWLAGN_BT_PSP_MIN_RSSI_THRESHOLD,
382 			IWLAGN_BT_PSP_MAX_RSSI_THRESHOLD);
383 }
384 
iwlagn_bt_traffic_is_sco(struct iwl_bt_uart_msg * uart_msg)385 static bool iwlagn_bt_traffic_is_sco(struct iwl_bt_uart_msg *uart_msg)
386 {
387 	return (BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3) >>
388 		BT_UART_MSG_FRAME3SCOESCO_POS;
389 }
390 
iwlagn_bt_traffic_change_work(struct work_struct * work)391 static void iwlagn_bt_traffic_change_work(struct work_struct *work)
392 {
393 	struct iwl_priv *priv =
394 		container_of(work, struct iwl_priv, bt_traffic_change_work);
395 	struct iwl_rxon_context *ctx;
396 	int smps_request = -1;
397 
398 	if (priv->bt_enable_flag == IWLAGN_BT_FLAG_COEX_MODE_DISABLED) {
399 		/* bt coex disabled */
400 		return;
401 	}
402 
403 	/*
404 	 * Note: bt_traffic_load can be overridden by scan complete and
405 	 * coex profile notifications. Ignore that since only bad consequence
406 	 * can be not matching debug print with actual state.
407 	 */
408 	IWL_DEBUG_COEX(priv, "BT traffic load changes: %d\n",
409 		       priv->bt_traffic_load);
410 
411 	switch (priv->bt_traffic_load) {
412 	case IWL_BT_COEX_TRAFFIC_LOAD_NONE:
413 		if (priv->bt_status)
414 			smps_request = IEEE80211_SMPS_DYNAMIC;
415 		else
416 			smps_request = IEEE80211_SMPS_AUTOMATIC;
417 		break;
418 	case IWL_BT_COEX_TRAFFIC_LOAD_LOW:
419 		smps_request = IEEE80211_SMPS_DYNAMIC;
420 		break;
421 	case IWL_BT_COEX_TRAFFIC_LOAD_HIGH:
422 	case IWL_BT_COEX_TRAFFIC_LOAD_CONTINUOUS:
423 		smps_request = IEEE80211_SMPS_STATIC;
424 		break;
425 	default:
426 		IWL_ERR(priv, "Invalid BT traffic load: %d\n",
427 			priv->bt_traffic_load);
428 		break;
429 	}
430 
431 	mutex_lock(&priv->mutex);
432 
433 	/*
434 	 * We can not send command to firmware while scanning. When the scan
435 	 * complete we will schedule this work again. We do check with mutex
436 	 * locked to prevent new scan request to arrive. We do not check
437 	 * STATUS_SCANNING to avoid race when queue_work two times from
438 	 * different notifications, but quit and not perform any work at all.
439 	 */
440 	if (test_bit(STATUS_SCAN_HW, &priv->status))
441 		goto out;
442 
443 	iwl_update_chain_flags(priv);
444 
445 	if (smps_request != -1) {
446 		priv->current_ht_config.smps = smps_request;
447 		for_each_context(priv, ctx) {
448 			if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_STATION)
449 				ieee80211_request_smps(ctx->vif, smps_request);
450 		}
451 	}
452 
453 	/*
454 	 * Dynamic PS poll related functionality. Adjust RSSI measurements if
455 	 * necessary.
456 	 */
457 	iwlagn_bt_coex_rssi_monitor(priv);
458 out:
459 	mutex_unlock(&priv->mutex);
460 }
461 
462 /*
463  * If BT sco traffic, and RSSI monitor is enabled, move measurements to the
464  * correct interface or disable it if this is the last interface to be
465  * removed.
466  */
iwlagn_bt_coex_rssi_monitor(struct iwl_priv * priv)467 void iwlagn_bt_coex_rssi_monitor(struct iwl_priv *priv)
468 {
469 	if (priv->bt_is_sco &&
470 	    priv->bt_traffic_load == IWL_BT_COEX_TRAFFIC_LOAD_CONTINUOUS)
471 		iwlagn_bt_adjust_rssi_monitor(priv, true);
472 	else
473 		iwlagn_bt_adjust_rssi_monitor(priv, false);
474 }
475 
iwlagn_print_uartmsg(struct iwl_priv * priv,struct iwl_bt_uart_msg * uart_msg)476 static void iwlagn_print_uartmsg(struct iwl_priv *priv,
477 				struct iwl_bt_uart_msg *uart_msg)
478 {
479 	IWL_DEBUG_COEX(priv, "Message Type = 0x%X, SSN = 0x%X, "
480 			"Update Req = 0x%X\n",
481 		(BT_UART_MSG_FRAME1MSGTYPE_MSK & uart_msg->frame1) >>
482 			BT_UART_MSG_FRAME1MSGTYPE_POS,
483 		(BT_UART_MSG_FRAME1SSN_MSK & uart_msg->frame1) >>
484 			BT_UART_MSG_FRAME1SSN_POS,
485 		(BT_UART_MSG_FRAME1UPDATEREQ_MSK & uart_msg->frame1) >>
486 			BT_UART_MSG_FRAME1UPDATEREQ_POS);
487 
488 	IWL_DEBUG_COEX(priv, "Open connections = 0x%X, Traffic load = 0x%X, "
489 			"Chl_SeqN = 0x%X, In band = 0x%X\n",
490 		(BT_UART_MSG_FRAME2OPENCONNECTIONS_MSK & uart_msg->frame2) >>
491 			BT_UART_MSG_FRAME2OPENCONNECTIONS_POS,
492 		(BT_UART_MSG_FRAME2TRAFFICLOAD_MSK & uart_msg->frame2) >>
493 			BT_UART_MSG_FRAME2TRAFFICLOAD_POS,
494 		(BT_UART_MSG_FRAME2CHLSEQN_MSK & uart_msg->frame2) >>
495 			BT_UART_MSG_FRAME2CHLSEQN_POS,
496 		(BT_UART_MSG_FRAME2INBAND_MSK & uart_msg->frame2) >>
497 			BT_UART_MSG_FRAME2INBAND_POS);
498 
499 	IWL_DEBUG_COEX(priv, "SCO/eSCO = 0x%X, Sniff = 0x%X, A2DP = 0x%X, "
500 			"ACL = 0x%X, Master = 0x%X, OBEX = 0x%X\n",
501 		(BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3) >>
502 			BT_UART_MSG_FRAME3SCOESCO_POS,
503 		(BT_UART_MSG_FRAME3SNIFF_MSK & uart_msg->frame3) >>
504 			BT_UART_MSG_FRAME3SNIFF_POS,
505 		(BT_UART_MSG_FRAME3A2DP_MSK & uart_msg->frame3) >>
506 			BT_UART_MSG_FRAME3A2DP_POS,
507 		(BT_UART_MSG_FRAME3ACL_MSK & uart_msg->frame3) >>
508 			BT_UART_MSG_FRAME3ACL_POS,
509 		(BT_UART_MSG_FRAME3MASTER_MSK & uart_msg->frame3) >>
510 			BT_UART_MSG_FRAME3MASTER_POS,
511 		(BT_UART_MSG_FRAME3OBEX_MSK & uart_msg->frame3) >>
512 			BT_UART_MSG_FRAME3OBEX_POS);
513 
514 	IWL_DEBUG_COEX(priv, "Idle duration = 0x%X\n",
515 		(BT_UART_MSG_FRAME4IDLEDURATION_MSK & uart_msg->frame4) >>
516 			BT_UART_MSG_FRAME4IDLEDURATION_POS);
517 
518 	IWL_DEBUG_COEX(priv, "Tx Activity = 0x%X, Rx Activity = 0x%X, "
519 			"eSCO Retransmissions = 0x%X\n",
520 		(BT_UART_MSG_FRAME5TXACTIVITY_MSK & uart_msg->frame5) >>
521 			BT_UART_MSG_FRAME5TXACTIVITY_POS,
522 		(BT_UART_MSG_FRAME5RXACTIVITY_MSK & uart_msg->frame5) >>
523 			BT_UART_MSG_FRAME5RXACTIVITY_POS,
524 		(BT_UART_MSG_FRAME5ESCORETRANSMIT_MSK & uart_msg->frame5) >>
525 			BT_UART_MSG_FRAME5ESCORETRANSMIT_POS);
526 
527 	IWL_DEBUG_COEX(priv, "Sniff Interval = 0x%X, Discoverable = 0x%X\n",
528 		(BT_UART_MSG_FRAME6SNIFFINTERVAL_MSK & uart_msg->frame6) >>
529 			BT_UART_MSG_FRAME6SNIFFINTERVAL_POS,
530 		(BT_UART_MSG_FRAME6DISCOVERABLE_MSK & uart_msg->frame6) >>
531 			BT_UART_MSG_FRAME6DISCOVERABLE_POS);
532 
533 	IWL_DEBUG_COEX(priv, "Sniff Activity = 0x%X, Page = "
534 			"0x%X, Inquiry = 0x%X, Connectable = 0x%X\n",
535 		(BT_UART_MSG_FRAME7SNIFFACTIVITY_MSK & uart_msg->frame7) >>
536 			BT_UART_MSG_FRAME7SNIFFACTIVITY_POS,
537 		(BT_UART_MSG_FRAME7PAGE_MSK & uart_msg->frame7) >>
538 			BT_UART_MSG_FRAME7PAGE_POS,
539 		(BT_UART_MSG_FRAME7INQUIRY_MSK & uart_msg->frame7) >>
540 			BT_UART_MSG_FRAME7INQUIRY_POS,
541 		(BT_UART_MSG_FRAME7CONNECTABLE_MSK & uart_msg->frame7) >>
542 			BT_UART_MSG_FRAME7CONNECTABLE_POS);
543 }
544 
iwlagn_set_kill_msk(struct iwl_priv * priv,struct iwl_bt_uart_msg * uart_msg)545 static bool iwlagn_set_kill_msk(struct iwl_priv *priv,
546 				struct iwl_bt_uart_msg *uart_msg)
547 {
548 	bool need_update = false;
549 	u8 kill_msk = IWL_BT_KILL_REDUCE;
550 	static const __le32 bt_kill_ack_msg[3] = {
551 		IWLAGN_BT_KILL_ACK_MASK_DEFAULT,
552 		IWLAGN_BT_KILL_ACK_CTS_MASK_SCO,
553 		IWLAGN_BT_KILL_ACK_CTS_MASK_REDUCE};
554 	static const __le32 bt_kill_cts_msg[3] = {
555 		IWLAGN_BT_KILL_CTS_MASK_DEFAULT,
556 		IWLAGN_BT_KILL_ACK_CTS_MASK_SCO,
557 		IWLAGN_BT_KILL_ACK_CTS_MASK_REDUCE};
558 
559 	if (!priv->reduced_txpower)
560 		kill_msk = (BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3)
561 			? IWL_BT_KILL_OVERRIDE : IWL_BT_KILL_DEFAULT;
562 	if (priv->kill_ack_mask != bt_kill_ack_msg[kill_msk] ||
563 	    priv->kill_cts_mask != bt_kill_cts_msg[kill_msk]) {
564 		priv->bt_valid |= IWLAGN_BT_VALID_KILL_ACK_MASK;
565 		priv->kill_ack_mask = bt_kill_ack_msg[kill_msk];
566 		priv->bt_valid |= IWLAGN_BT_VALID_KILL_CTS_MASK;
567 		priv->kill_cts_mask = bt_kill_cts_msg[kill_msk];
568 		need_update = true;
569 	}
570 	return need_update;
571 }
572 
573 /*
574  * Upon RSSI changes, sends a bt config command with following changes
575  *  1. enable/disable "reduced control frames tx power
576  *  2. update the "kill)ack_mask" and "kill_cts_mask"
577  *
578  * If "reduced tx power" is enabled, uCode shall
579  *  1. ACK/Back/CTS rate shall reduced to 6Mbps
580  *  2. not use duplciate 20/40MHz mode
581  */
iwlagn_fill_txpower_mode(struct iwl_priv * priv,struct iwl_bt_uart_msg * uart_msg)582 static bool iwlagn_fill_txpower_mode(struct iwl_priv *priv,
583 				struct iwl_bt_uart_msg *uart_msg)
584 {
585 	bool need_update = false;
586 	struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS];
587 	int ave_rssi;
588 
589 	if (!ctx->vif || (ctx->vif->type != NL80211_IFTYPE_STATION)) {
590 		IWL_DEBUG_INFO(priv, "BSS ctx not active or not in sta mode\n");
591 		return false;
592 	}
593 
594 	ave_rssi = ieee80211_ave_rssi(ctx->vif);
595 	if (!ave_rssi) {
596 		/* no rssi data, no changes to reduce tx power */
597 		IWL_DEBUG_COEX(priv, "no rssi data available\n");
598 		return need_update;
599 	}
600 	if (!priv->reduced_txpower &&
601 	    !iwl_is_associated(priv, IWL_RXON_CTX_PAN) &&
602 	    (ave_rssi > BT_ENABLE_REDUCED_TXPOWER_THRESHOLD) &&
603 	    (uart_msg->frame3 & (BT_UART_MSG_FRAME3ACL_MSK |
604 	    BT_UART_MSG_FRAME3OBEX_MSK)) &&
605 	    !(uart_msg->frame3 & (BT_UART_MSG_FRAME3SCOESCO_MSK |
606 	    BT_UART_MSG_FRAME3SNIFF_MSK | BT_UART_MSG_FRAME3A2DP_MSK))) {
607 		/* enabling reduced tx power */
608 		priv->reduced_txpower = true;
609 		priv->bt_valid |= IWLAGN_BT_VALID_REDUCED_TX_PWR;
610 		need_update = true;
611 	} else if (priv->reduced_txpower &&
612 		   (iwl_is_associated(priv, IWL_RXON_CTX_PAN) ||
613 		   (ave_rssi < BT_DISABLE_REDUCED_TXPOWER_THRESHOLD) ||
614 		   (uart_msg->frame3 & (BT_UART_MSG_FRAME3SCOESCO_MSK |
615 		   BT_UART_MSG_FRAME3SNIFF_MSK | BT_UART_MSG_FRAME3A2DP_MSK)) ||
616 		   !(uart_msg->frame3 & (BT_UART_MSG_FRAME3ACL_MSK |
617 		   BT_UART_MSG_FRAME3OBEX_MSK)))) {
618 		/* disable reduced tx power */
619 		priv->reduced_txpower = false;
620 		priv->bt_valid |= IWLAGN_BT_VALID_REDUCED_TX_PWR;
621 		need_update = true;
622 	}
623 
624 	return need_update;
625 }
626 
iwlagn_bt_coex_profile_notif(struct iwl_priv * priv,struct iwl_rx_cmd_buffer * rxb)627 static void iwlagn_bt_coex_profile_notif(struct iwl_priv *priv,
628 					 struct iwl_rx_cmd_buffer *rxb)
629 {
630 	struct iwl_rx_packet *pkt = rxb_addr(rxb);
631 	struct iwl_bt_coex_profile_notif *coex = (void *)pkt->data;
632 	struct iwl_bt_uart_msg *uart_msg = &coex->last_bt_uart_msg;
633 
634 	if (priv->bt_enable_flag == IWLAGN_BT_FLAG_COEX_MODE_DISABLED) {
635 		/* bt coex disabled */
636 		return;
637 	}
638 
639 	IWL_DEBUG_COEX(priv, "BT Coex notification:\n");
640 	IWL_DEBUG_COEX(priv, "    status: %d\n", coex->bt_status);
641 	IWL_DEBUG_COEX(priv, "    traffic load: %d\n", coex->bt_traffic_load);
642 	IWL_DEBUG_COEX(priv, "    CI compliance: %d\n",
643 			coex->bt_ci_compliance);
644 	iwlagn_print_uartmsg(priv, uart_msg);
645 
646 	priv->last_bt_traffic_load = priv->bt_traffic_load;
647 	priv->bt_is_sco = iwlagn_bt_traffic_is_sco(uart_msg);
648 
649 	if (priv->iw_mode != NL80211_IFTYPE_ADHOC) {
650 		if (priv->bt_status != coex->bt_status ||
651 		    priv->last_bt_traffic_load != coex->bt_traffic_load) {
652 			if (coex->bt_status) {
653 				/* BT on */
654 				if (!priv->bt_ch_announce)
655 					priv->bt_traffic_load =
656 						IWL_BT_COEX_TRAFFIC_LOAD_HIGH;
657 				else
658 					priv->bt_traffic_load =
659 						coex->bt_traffic_load;
660 			} else {
661 				/* BT off */
662 				priv->bt_traffic_load =
663 					IWL_BT_COEX_TRAFFIC_LOAD_NONE;
664 			}
665 			priv->bt_status = coex->bt_status;
666 			queue_work(priv->workqueue,
667 				   &priv->bt_traffic_change_work);
668 		}
669 	}
670 
671 	/* schedule to send runtime bt_config */
672 	/* check reduce power before change ack/cts kill mask */
673 	if (iwlagn_fill_txpower_mode(priv, uart_msg) ||
674 	    iwlagn_set_kill_msk(priv, uart_msg))
675 		queue_work(priv->workqueue, &priv->bt_runtime_config);
676 
677 
678 	/* FIXME: based on notification, adjust the prio_boost */
679 
680 	priv->bt_ci_compliance = coex->bt_ci_compliance;
681 }
682 
iwlagn_bt_rx_handler_setup(struct iwl_priv * priv)683 void iwlagn_bt_rx_handler_setup(struct iwl_priv *priv)
684 {
685 	priv->rx_handlers[REPLY_BT_COEX_PROFILE_NOTIF] =
686 		iwlagn_bt_coex_profile_notif;
687 }
688 
iwlagn_bt_setup_deferred_work(struct iwl_priv * priv)689 void iwlagn_bt_setup_deferred_work(struct iwl_priv *priv)
690 {
691 	INIT_WORK(&priv->bt_traffic_change_work,
692 		  iwlagn_bt_traffic_change_work);
693 }
694 
iwlagn_bt_cancel_deferred_work(struct iwl_priv * priv)695 void iwlagn_bt_cancel_deferred_work(struct iwl_priv *priv)
696 {
697 	cancel_work_sync(&priv->bt_traffic_change_work);
698 }
699 
is_single_rx_stream(struct iwl_priv * priv)700 static bool is_single_rx_stream(struct iwl_priv *priv)
701 {
702 	return priv->current_ht_config.smps == IEEE80211_SMPS_STATIC ||
703 	       priv->current_ht_config.single_chain_sufficient;
704 }
705 
706 #define IWL_NUM_RX_CHAINS_MULTIPLE	3
707 #define IWL_NUM_RX_CHAINS_SINGLE	2
708 #define IWL_NUM_IDLE_CHAINS_DUAL	2
709 #define IWL_NUM_IDLE_CHAINS_SINGLE	1
710 
711 /*
712  * Determine how many receiver/antenna chains to use.
713  *
714  * More provides better reception via diversity.  Fewer saves power
715  * at the expense of throughput, but only when not in powersave to
716  * start with.
717  *
718  * MIMO (dual stream) requires at least 2, but works better with 3.
719  * This does not determine *which* chains to use, just how many.
720  */
iwl_get_active_rx_chain_count(struct iwl_priv * priv)721 static int iwl_get_active_rx_chain_count(struct iwl_priv *priv)
722 {
723 	if (priv->lib->bt_params &&
724 	    priv->lib->bt_params->advanced_bt_coexist &&
725 	    (priv->bt_full_concurrent ||
726 	     priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH)) {
727 		/*
728 		 * only use chain 'A' in bt high traffic load or
729 		 * full concurrency mode
730 		 */
731 		return IWL_NUM_RX_CHAINS_SINGLE;
732 	}
733 	/* # of Rx chains to use when expecting MIMO. */
734 	if (is_single_rx_stream(priv))
735 		return IWL_NUM_RX_CHAINS_SINGLE;
736 	else
737 		return IWL_NUM_RX_CHAINS_MULTIPLE;
738 }
739 
740 /*
741  * When we are in power saving mode, unless device support spatial
742  * multiplexing power save, use the active count for rx chain count.
743  */
iwl_get_idle_rx_chain_count(struct iwl_priv * priv,int active_cnt)744 static int iwl_get_idle_rx_chain_count(struct iwl_priv *priv, int active_cnt)
745 {
746 	/* # Rx chains when idling, depending on SMPS mode */
747 	switch (priv->current_ht_config.smps) {
748 	case IEEE80211_SMPS_STATIC:
749 	case IEEE80211_SMPS_DYNAMIC:
750 		return IWL_NUM_IDLE_CHAINS_SINGLE;
751 	case IEEE80211_SMPS_AUTOMATIC:
752 	case IEEE80211_SMPS_OFF:
753 		return active_cnt;
754 	default:
755 		WARN(1, "invalid SMPS mode %d",
756 		     priv->current_ht_config.smps);
757 		return active_cnt;
758 	}
759 }
760 
761 /* up to 4 chains */
iwl_count_chain_bitmap(u32 chain_bitmap)762 static u8 iwl_count_chain_bitmap(u32 chain_bitmap)
763 {
764 	u8 res;
765 	res = (chain_bitmap & BIT(0)) >> 0;
766 	res += (chain_bitmap & BIT(1)) >> 1;
767 	res += (chain_bitmap & BIT(2)) >> 2;
768 	res += (chain_bitmap & BIT(3)) >> 3;
769 	return res;
770 }
771 
772 /*
773  * iwlagn_set_rxon_chain - Set up Rx chain usage in "staging" RXON image
774  *
775  * Selects how many and which Rx receivers/antennas/chains to use.
776  * This should not be used for scan command ... it puts data in wrong place.
777  */
iwlagn_set_rxon_chain(struct iwl_priv * priv,struct iwl_rxon_context * ctx)778 void iwlagn_set_rxon_chain(struct iwl_priv *priv, struct iwl_rxon_context *ctx)
779 {
780 	bool is_single = is_single_rx_stream(priv);
781 	bool is_cam = !test_bit(STATUS_POWER_PMI, &priv->status);
782 	u8 idle_rx_cnt, active_rx_cnt, valid_rx_cnt;
783 	u32 active_chains;
784 	u16 rx_chain;
785 
786 	/* Tell uCode which antennas are actually connected.
787 	 * Before first association, we assume all antennas are connected.
788 	 * Just after first association, iwl_chain_noise_calibration()
789 	 *    checks which antennas actually *are* connected. */
790 	if (priv->chain_noise_data.active_chains)
791 		active_chains = priv->chain_noise_data.active_chains;
792 	else
793 		active_chains = priv->nvm_data->valid_rx_ant;
794 
795 	if (priv->lib->bt_params &&
796 	    priv->lib->bt_params->advanced_bt_coexist &&
797 	    (priv->bt_full_concurrent ||
798 	     priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH)) {
799 		/*
800 		 * only use chain 'A' in bt high traffic load or
801 		 * full concurrency mode
802 		 */
803 		active_chains = first_antenna(active_chains);
804 	}
805 
806 	rx_chain = active_chains << RXON_RX_CHAIN_VALID_POS;
807 
808 	/* How many receivers should we use? */
809 	active_rx_cnt = iwl_get_active_rx_chain_count(priv);
810 	idle_rx_cnt = iwl_get_idle_rx_chain_count(priv, active_rx_cnt);
811 
812 
813 	/* correct rx chain count according hw settings
814 	 * and chain noise calibration
815 	 */
816 	valid_rx_cnt = iwl_count_chain_bitmap(active_chains);
817 	if (valid_rx_cnt < active_rx_cnt)
818 		active_rx_cnt = valid_rx_cnt;
819 
820 	if (valid_rx_cnt < idle_rx_cnt)
821 		idle_rx_cnt = valid_rx_cnt;
822 
823 	rx_chain |= active_rx_cnt << RXON_RX_CHAIN_MIMO_CNT_POS;
824 	rx_chain |= idle_rx_cnt  << RXON_RX_CHAIN_CNT_POS;
825 
826 	ctx->staging.rx_chain = cpu_to_le16(rx_chain);
827 
828 	if (!is_single && (active_rx_cnt >= IWL_NUM_RX_CHAINS_SINGLE) && is_cam)
829 		ctx->staging.rx_chain |= RXON_RX_CHAIN_MIMO_FORCE_MSK;
830 	else
831 		ctx->staging.rx_chain &= ~RXON_RX_CHAIN_MIMO_FORCE_MSK;
832 
833 	IWL_DEBUG_ASSOC(priv, "rx_chain=0x%X active=%d idle=%d\n",
834 			ctx->staging.rx_chain,
835 			active_rx_cnt, idle_rx_cnt);
836 
837 	WARN_ON(active_rx_cnt == 0 || idle_rx_cnt == 0 ||
838 		active_rx_cnt < idle_rx_cnt);
839 }
840 
iwl_toggle_tx_ant(struct iwl_priv * priv,u8 ant,u8 valid)841 u8 iwl_toggle_tx_ant(struct iwl_priv *priv, u8 ant, u8 valid)
842 {
843 	int i;
844 	u8 ind = ant;
845 
846 	if (priv->band == NL80211_BAND_2GHZ &&
847 	    priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH)
848 		return 0;
849 
850 	for (i = 0; i < RATE_ANT_NUM - 1; i++) {
851 		ind = (ind + 1) < RATE_ANT_NUM ?  ind + 1 : 0;
852 		if (valid & BIT(ind))
853 			return ind;
854 	}
855 	return ant;
856 }
857 
858 #ifdef CONFIG_PM_SLEEP
iwlagn_convert_p1k(u16 * p1k,__le16 * out)859 static void iwlagn_convert_p1k(u16 *p1k, __le16 *out)
860 {
861 	int i;
862 
863 	for (i = 0; i < IWLAGN_P1K_SIZE; i++)
864 		out[i] = cpu_to_le16(p1k[i]);
865 }
866 
867 struct wowlan_key_data {
868 	struct iwl_rxon_context *ctx;
869 	struct iwlagn_wowlan_rsc_tsc_params_cmd *rsc_tsc;
870 	struct iwlagn_wowlan_tkip_params_cmd *tkip;
871 	const u8 *bssid;
872 	bool error, use_rsc_tsc, use_tkip;
873 };
874 
875 
iwlagn_wowlan_program_keys(struct ieee80211_hw * hw,struct ieee80211_vif * vif,struct ieee80211_sta * sta,struct ieee80211_key_conf * key,void * _data)876 static void iwlagn_wowlan_program_keys(struct ieee80211_hw *hw,
877 			       struct ieee80211_vif *vif,
878 			       struct ieee80211_sta *sta,
879 			       struct ieee80211_key_conf *key,
880 			       void *_data)
881 {
882 	struct iwl_priv *priv = IWL_MAC80211_GET_DVM(hw);
883 	struct wowlan_key_data *data = _data;
884 	struct iwl_rxon_context *ctx = data->ctx;
885 	struct aes_sc *aes_sc, *aes_tx_sc = NULL;
886 	struct tkip_sc *tkip_sc, *tkip_tx_sc = NULL;
887 	struct iwlagn_p1k_cache *rx_p1ks;
888 	u8 *rx_mic_key;
889 	struct ieee80211_key_seq seq;
890 	u32 cur_rx_iv32 = 0;
891 	u16 p1k[IWLAGN_P1K_SIZE];
892 	int ret, i;
893 
894 	mutex_lock(&priv->mutex);
895 
896 	if ((key->cipher == WLAN_CIPHER_SUITE_WEP40 ||
897 	     key->cipher == WLAN_CIPHER_SUITE_WEP104) &&
898 	     !sta && !ctx->key_mapping_keys)
899 		ret = iwl_set_default_wep_key(priv, ctx, key);
900 	else
901 		ret = iwl_set_dynamic_key(priv, ctx, key, sta);
902 
903 	if (ret) {
904 		IWL_ERR(priv, "Error setting key during suspend!\n");
905 		data->error = true;
906 	}
907 
908 	switch (key->cipher) {
909 	case WLAN_CIPHER_SUITE_TKIP:
910 		if (sta) {
911 			u64 pn64;
912 
913 			tkip_sc = data->rsc_tsc->all_tsc_rsc.tkip.unicast_rsc;
914 			tkip_tx_sc = &data->rsc_tsc->all_tsc_rsc.tkip.tsc;
915 
916 			rx_p1ks = data->tkip->rx_uni;
917 
918 			pn64 = atomic64_read(&key->tx_pn);
919 			tkip_tx_sc->iv16 = cpu_to_le16(TKIP_PN_TO_IV16(pn64));
920 			tkip_tx_sc->iv32 = cpu_to_le32(TKIP_PN_TO_IV32(pn64));
921 
922 			ieee80211_get_tkip_p1k_iv(key, seq.tkip.iv32, p1k);
923 			iwlagn_convert_p1k(p1k, data->tkip->tx.p1k);
924 
925 			memcpy(data->tkip->mic_keys.tx,
926 			       &key->key[NL80211_TKIP_DATA_OFFSET_TX_MIC_KEY],
927 			       IWLAGN_MIC_KEY_SIZE);
928 
929 			rx_mic_key = data->tkip->mic_keys.rx_unicast;
930 		} else {
931 			tkip_sc =
932 				data->rsc_tsc->all_tsc_rsc.tkip.multicast_rsc;
933 			rx_p1ks = data->tkip->rx_multi;
934 			rx_mic_key = data->tkip->mic_keys.rx_mcast;
935 		}
936 
937 		/*
938 		 * For non-QoS this relies on the fact that both the uCode and
939 		 * mac80211 use TID 0 (as they need to to avoid replay attacks)
940 		 * for checking the IV in the frames.
941 		 */
942 		for (i = 0; i < IWLAGN_NUM_RSC; i++) {
943 			ieee80211_get_key_rx_seq(key, i, &seq);
944 			tkip_sc[i].iv16 = cpu_to_le16(seq.tkip.iv16);
945 			tkip_sc[i].iv32 = cpu_to_le32(seq.tkip.iv32);
946 			/* wrapping isn't allowed, AP must rekey */
947 			if (seq.tkip.iv32 > cur_rx_iv32)
948 				cur_rx_iv32 = seq.tkip.iv32;
949 		}
950 
951 		ieee80211_get_tkip_rx_p1k(key, data->bssid, cur_rx_iv32, p1k);
952 		iwlagn_convert_p1k(p1k, rx_p1ks[0].p1k);
953 		ieee80211_get_tkip_rx_p1k(key, data->bssid,
954 					  cur_rx_iv32 + 1, p1k);
955 		iwlagn_convert_p1k(p1k, rx_p1ks[1].p1k);
956 
957 		memcpy(rx_mic_key,
958 		       &key->key[NL80211_TKIP_DATA_OFFSET_RX_MIC_KEY],
959 		       IWLAGN_MIC_KEY_SIZE);
960 
961 		data->use_tkip = true;
962 		data->use_rsc_tsc = true;
963 		break;
964 	case WLAN_CIPHER_SUITE_CCMP:
965 		if (sta) {
966 			u64 pn64;
967 
968 			aes_sc = data->rsc_tsc->all_tsc_rsc.aes.unicast_rsc;
969 			aes_tx_sc = &data->rsc_tsc->all_tsc_rsc.aes.tsc;
970 
971 			pn64 = atomic64_read(&key->tx_pn);
972 			aes_tx_sc->pn = cpu_to_le64(pn64);
973 		} else
974 			aes_sc = data->rsc_tsc->all_tsc_rsc.aes.multicast_rsc;
975 
976 		/*
977 		 * For non-QoS this relies on the fact that both the uCode and
978 		 * mac80211 use TID 0 for checking the IV in the frames.
979 		 */
980 		for (i = 0; i < IWLAGN_NUM_RSC; i++) {
981 			u8 *pn = seq.ccmp.pn;
982 
983 			ieee80211_get_key_rx_seq(key, i, &seq);
984 			aes_sc[i].pn = cpu_to_le64(
985 					(u64)pn[5] |
986 					((u64)pn[4] << 8) |
987 					((u64)pn[3] << 16) |
988 					((u64)pn[2] << 24) |
989 					((u64)pn[1] << 32) |
990 					((u64)pn[0] << 40));
991 		}
992 		data->use_rsc_tsc = true;
993 		break;
994 	}
995 
996 	mutex_unlock(&priv->mutex);
997 }
998 
iwlagn_send_patterns(struct iwl_priv * priv,struct cfg80211_wowlan * wowlan)999 int iwlagn_send_patterns(struct iwl_priv *priv,
1000 			struct cfg80211_wowlan *wowlan)
1001 {
1002 	struct iwlagn_wowlan_patterns_cmd *pattern_cmd;
1003 	struct iwl_host_cmd cmd = {
1004 		.id = REPLY_WOWLAN_PATTERNS,
1005 		.dataflags[0] = IWL_HCMD_DFL_NOCOPY,
1006 	};
1007 	int i, err;
1008 
1009 	if (!wowlan->n_patterns)
1010 		return 0;
1011 
1012 	cmd.len[0] = struct_size(pattern_cmd, patterns, wowlan->n_patterns);
1013 
1014 	pattern_cmd = kmalloc(cmd.len[0], GFP_KERNEL);
1015 	if (!pattern_cmd)
1016 		return -ENOMEM;
1017 
1018 	pattern_cmd->n_patterns = cpu_to_le32(wowlan->n_patterns);
1019 
1020 	for (i = 0; i < wowlan->n_patterns; i++) {
1021 		int mask_len = DIV_ROUND_UP(wowlan->patterns[i].pattern_len, 8);
1022 
1023 		memcpy(&pattern_cmd->patterns[i].mask,
1024 			wowlan->patterns[i].mask, mask_len);
1025 		memcpy(&pattern_cmd->patterns[i].pattern,
1026 			wowlan->patterns[i].pattern,
1027 			wowlan->patterns[i].pattern_len);
1028 		pattern_cmd->patterns[i].mask_size = mask_len;
1029 		pattern_cmd->patterns[i].pattern_size =
1030 			wowlan->patterns[i].pattern_len;
1031 	}
1032 
1033 	cmd.data[0] = pattern_cmd;
1034 	err = iwl_dvm_send_cmd(priv, &cmd);
1035 	kfree(pattern_cmd);
1036 	return err;
1037 }
1038 
iwlagn_suspend(struct iwl_priv * priv,struct cfg80211_wowlan * wowlan)1039 int iwlagn_suspend(struct iwl_priv *priv, struct cfg80211_wowlan *wowlan)
1040 {
1041 	struct iwlagn_wowlan_wakeup_filter_cmd wakeup_filter_cmd;
1042 	struct iwl_rxon_cmd rxon;
1043 	struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS];
1044 	struct iwlagn_wowlan_kek_kck_material_cmd kek_kck_cmd;
1045 	struct iwlagn_wowlan_tkip_params_cmd tkip_cmd = {};
1046 	struct iwlagn_d3_config_cmd d3_cfg_cmd = {
1047 		/*
1048 		 * Program the minimum sleep time to 10 seconds, as many
1049 		 * platforms have issues processing a wakeup signal while
1050 		 * still being in the process of suspending.
1051 		 */
1052 		.min_sleep_time = cpu_to_le32(10 * 1000 * 1000),
1053 	};
1054 	struct wowlan_key_data key_data = {
1055 		.ctx = ctx,
1056 		.bssid = ctx->active.bssid_addr,
1057 		.use_rsc_tsc = false,
1058 		.tkip = &tkip_cmd,
1059 		.use_tkip = false,
1060 	};
1061 	int ret, i;
1062 	u16 seq;
1063 
1064 	key_data.rsc_tsc = kzalloc(sizeof(*key_data.rsc_tsc), GFP_KERNEL);
1065 	if (!key_data.rsc_tsc)
1066 		return -ENOMEM;
1067 
1068 	memset(&wakeup_filter_cmd, 0, sizeof(wakeup_filter_cmd));
1069 
1070 	/*
1071 	 * We know the last used seqno, and the uCode expects to know that
1072 	 * one, it will increment before TX.
1073 	 */
1074 	seq = le16_to_cpu(priv->last_seq_ctl) & IEEE80211_SCTL_SEQ;
1075 	wakeup_filter_cmd.non_qos_seq = cpu_to_le16(seq);
1076 
1077 	/*
1078 	 * For QoS counters, we store the one to use next, so subtract 0x10
1079 	 * since the uCode will add 0x10 before using the value.
1080 	 */
1081 	for (i = 0; i < IWL_MAX_TID_COUNT; i++) {
1082 		seq = priv->tid_data[IWL_AP_ID][i].seq_number;
1083 		seq -= 0x10;
1084 		wakeup_filter_cmd.qos_seq[i] = cpu_to_le16(seq);
1085 	}
1086 
1087 	if (wowlan->disconnect)
1088 		wakeup_filter_cmd.enabled |=
1089 			cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_BEACON_MISS |
1090 				    IWLAGN_WOWLAN_WAKEUP_LINK_CHANGE);
1091 	if (wowlan->magic_pkt)
1092 		wakeup_filter_cmd.enabled |=
1093 			cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_MAGIC_PACKET);
1094 	if (wowlan->gtk_rekey_failure)
1095 		wakeup_filter_cmd.enabled |=
1096 			cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_GTK_REKEY_FAIL);
1097 	if (wowlan->eap_identity_req)
1098 		wakeup_filter_cmd.enabled |=
1099 			cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_EAP_IDENT_REQ);
1100 	if (wowlan->four_way_handshake)
1101 		wakeup_filter_cmd.enabled |=
1102 			cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_4WAY_HANDSHAKE);
1103 	if (wowlan->n_patterns)
1104 		wakeup_filter_cmd.enabled |=
1105 			cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_PATTERN_MATCH);
1106 
1107 	if (wowlan->rfkill_release)
1108 		d3_cfg_cmd.wakeup_flags |=
1109 			cpu_to_le32(IWLAGN_D3_WAKEUP_RFKILL);
1110 
1111 	iwl_scan_cancel_timeout(priv, 200);
1112 
1113 	memcpy(&rxon, &ctx->active, sizeof(rxon));
1114 
1115 	priv->ucode_loaded = false;
1116 	iwl_trans_stop_device(priv->trans);
1117 	ret = iwl_trans_start_hw(priv->trans);
1118 	if (ret)
1119 		goto out;
1120 
1121 	priv->wowlan = true;
1122 
1123 	ret = iwl_load_ucode_wait_alive(priv, IWL_UCODE_WOWLAN);
1124 	if (ret)
1125 		goto out;
1126 
1127 	/* now configure WoWLAN ucode */
1128 	ret = iwl_alive_start(priv);
1129 	if (ret)
1130 		goto out;
1131 
1132 	memcpy(&ctx->staging, &rxon, sizeof(rxon));
1133 	ret = iwlagn_commit_rxon(priv, ctx);
1134 	if (ret)
1135 		goto out;
1136 
1137 	ret = iwl_power_update_mode(priv, true);
1138 	if (ret)
1139 		goto out;
1140 
1141 	if (!iwlwifi_mod_params.swcrypto) {
1142 		/* mark all keys clear */
1143 		priv->ucode_key_table = 0;
1144 		ctx->key_mapping_keys = 0;
1145 
1146 		/*
1147 		 * This needs to be unlocked due to lock ordering
1148 		 * constraints. Since we're in the suspend path
1149 		 * that isn't really a problem though.
1150 		 */
1151 		mutex_unlock(&priv->mutex);
1152 		ieee80211_iter_keys(priv->hw, ctx->vif,
1153 				    iwlagn_wowlan_program_keys,
1154 				    &key_data);
1155 		mutex_lock(&priv->mutex);
1156 		if (key_data.error) {
1157 			ret = -EIO;
1158 			goto out;
1159 		}
1160 
1161 		if (key_data.use_rsc_tsc) {
1162 			struct iwl_host_cmd rsc_tsc_cmd = {
1163 				.id = REPLY_WOWLAN_TSC_RSC_PARAMS,
1164 				.data[0] = key_data.rsc_tsc,
1165 				.dataflags[0] = IWL_HCMD_DFL_NOCOPY,
1166 				.len[0] = sizeof(*key_data.rsc_tsc),
1167 			};
1168 
1169 			ret = iwl_dvm_send_cmd(priv, &rsc_tsc_cmd);
1170 			if (ret)
1171 				goto out;
1172 		}
1173 
1174 		if (key_data.use_tkip) {
1175 			ret = iwl_dvm_send_cmd_pdu(priv,
1176 						 REPLY_WOWLAN_TKIP_PARAMS,
1177 						 0, sizeof(tkip_cmd),
1178 						 &tkip_cmd);
1179 			if (ret)
1180 				goto out;
1181 		}
1182 
1183 		if (priv->have_rekey_data) {
1184 			memset(&kek_kck_cmd, 0, sizeof(kek_kck_cmd));
1185 			memcpy(kek_kck_cmd.kck, priv->kck, NL80211_KCK_LEN);
1186 			kek_kck_cmd.kck_len = cpu_to_le16(NL80211_KCK_LEN);
1187 			memcpy(kek_kck_cmd.kek, priv->kek, NL80211_KEK_LEN);
1188 			kek_kck_cmd.kek_len = cpu_to_le16(NL80211_KEK_LEN);
1189 			kek_kck_cmd.replay_ctr = priv->replay_ctr;
1190 
1191 			ret = iwl_dvm_send_cmd_pdu(priv,
1192 						 REPLY_WOWLAN_KEK_KCK_MATERIAL,
1193 						 0, sizeof(kek_kck_cmd),
1194 						 &kek_kck_cmd);
1195 			if (ret)
1196 				goto out;
1197 		}
1198 	}
1199 
1200 	ret = iwl_dvm_send_cmd_pdu(priv, REPLY_D3_CONFIG, 0,
1201 				     sizeof(d3_cfg_cmd), &d3_cfg_cmd);
1202 	if (ret)
1203 		goto out;
1204 
1205 	ret = iwl_dvm_send_cmd_pdu(priv, REPLY_WOWLAN_WAKEUP_FILTER,
1206 				 0, sizeof(wakeup_filter_cmd),
1207 				 &wakeup_filter_cmd);
1208 	if (ret)
1209 		goto out;
1210 
1211 	ret = iwlagn_send_patterns(priv, wowlan);
1212  out:
1213 	kfree(key_data.rsc_tsc);
1214 	return ret;
1215 }
1216 #endif
1217 
iwl_dvm_send_cmd(struct iwl_priv * priv,struct iwl_host_cmd * cmd)1218 int iwl_dvm_send_cmd(struct iwl_priv *priv, struct iwl_host_cmd *cmd)
1219 {
1220 	if (iwl_is_rfkill(priv) || iwl_is_ctkill(priv)) {
1221 		IWL_WARN(priv, "Not sending command - %s KILL\n",
1222 			 iwl_is_rfkill(priv) ? "RF" : "CT");
1223 		return -EIO;
1224 	}
1225 
1226 	if (test_bit(STATUS_FW_ERROR, &priv->status)) {
1227 		IWL_ERR(priv, "Command %s failed: FW Error\n",
1228 			iwl_get_cmd_string(priv->trans, cmd->id));
1229 		return -EIO;
1230 	}
1231 
1232 	/*
1233 	 * This can happen upon FW ASSERT: we clear the STATUS_FW_ERROR flag
1234 	 * in iwl_down but cancel the workers only later.
1235 	 */
1236 	if (!priv->ucode_loaded) {
1237 		IWL_ERR(priv, "Fw not loaded - dropping CMD: %x\n", cmd->id);
1238 		return -EIO;
1239 	}
1240 
1241 	/*
1242 	 * Synchronous commands from this op-mode must hold
1243 	 * the mutex, this ensures we don't try to send two
1244 	 * (or more) synchronous commands at a time.
1245 	 */
1246 	if (!(cmd->flags & CMD_ASYNC))
1247 		lockdep_assert_held(&priv->mutex);
1248 
1249 	return iwl_trans_send_cmd(priv->trans, cmd);
1250 }
1251 
iwl_dvm_send_cmd_pdu(struct iwl_priv * priv,u8 id,u32 flags,u16 len,const void * data)1252 int iwl_dvm_send_cmd_pdu(struct iwl_priv *priv, u8 id,
1253 			 u32 flags, u16 len, const void *data)
1254 {
1255 	struct iwl_host_cmd cmd = {
1256 		.id = id,
1257 		.len = { len, },
1258 		.data = { data, },
1259 		.flags = flags,
1260 	};
1261 
1262 	return iwl_dvm_send_cmd(priv, &cmd);
1263 }
1264