1 /******************************************************************************
3 * Copyright(c) 2003 - 2011 Intel Corporation. All rights reserved.
5 * Portions of this file are derived from the ipw3945 project, as well
6 * as portions of the ieee80211 subsystem header files.
8 * This program is free software; you can redistribute it and/or modify it
9 * under the terms of version 2 of the GNU General Public License as
10 * published by the Free Software Foundation.
12 * This program is distributed in the hope that it will be useful, but WITHOUT
13 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
14 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
17 * You should have received a copy of the GNU General Public License along with
18 * this program; if not, write to the Free Software Foundation, Inc.,
19 * 51 Franklin Street, Fifth Floor, Boston, MA 02110, USA
21 * The full GNU General Public License is included in this distribution in the
22 * file called LICENSE.
24 * Contact Information:
25 * Intel Linux Wireless <ilw@linux.intel.com>
26 * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
28 *****************************************************************************/
29 #include <linux/kernel.h>
30 #include <linux/module.h>
31 #include <linux/init.h>
32 #include <linux/slab.h>
33 #include <linux/dma-mapping.h>
34 #include <linux/delay.h>
35 #include <linux/sched.h>
36 #include <linux/skbuff.h>
37 #include <linux/netdevice.h>
38 #include <linux/firmware.h>
39 #include <linux/etherdevice.h>
40 #include <linux/if_arp.h>
42 #include <net/mac80211.h>
44 #include <asm/div64.h>
46 #include "iwl-eeprom.h"
50 #include "iwl-helpers.h"
51 #include "iwl-agn-calib.h"
53 #include "iwl-shared.h"
55 #include "iwl-trans.h"
57 /******************************************************************************
61 ******************************************************************************/
64 * module name, copyright, version, etc.
66 #define DRV_DESCRIPTION "Intel(R) Wireless WiFi Link AGN driver for Linux"
68 #ifdef CONFIG_IWLWIFI_DEBUG
74 #define DRV_VERSION IWLWIFI_VERSION VD
77 MODULE_DESCRIPTION(DRV_DESCRIPTION
);
78 MODULE_VERSION(DRV_VERSION
);
79 MODULE_AUTHOR(DRV_COPYRIGHT
" " DRV_AUTHOR
);
80 MODULE_LICENSE("GPL");
81 MODULE_ALIAS("iwlagn");
83 void iwl_update_chain_flags(struct iwl_priv
*priv
)
85 struct iwl_rxon_context
*ctx
;
87 for_each_context(priv
, ctx
) {
88 iwlagn_set_rxon_chain(priv
, ctx
);
89 if (ctx
->active
.rx_chain
!= ctx
->staging
.rx_chain
)
90 iwlagn_commit_rxon(priv
, ctx
);
94 /* Parse the beacon frame to find the TIM element and set tim_idx & tim_size */
95 static void iwl_set_beacon_tim(struct iwl_priv
*priv
,
96 struct iwl_tx_beacon_cmd
*tx_beacon_cmd
,
97 u8
*beacon
, u32 frame_size
)
100 struct ieee80211_mgmt
*mgmt
= (struct ieee80211_mgmt
*)beacon
;
103 * The index is relative to frame start but we start looking at the
104 * variable-length part of the beacon.
106 tim_idx
= mgmt
->u
.beacon
.variable
- beacon
;
108 /* Parse variable-length elements of beacon to find WLAN_EID_TIM */
109 while ((tim_idx
< (frame_size
- 2)) &&
110 (beacon
[tim_idx
] != WLAN_EID_TIM
))
111 tim_idx
+= beacon
[tim_idx
+1] + 2;
113 /* If TIM field was found, set variables */
114 if ((tim_idx
< (frame_size
- 1)) && (beacon
[tim_idx
] == WLAN_EID_TIM
)) {
115 tx_beacon_cmd
->tim_idx
= cpu_to_le16(tim_idx
);
116 tx_beacon_cmd
->tim_size
= beacon
[tim_idx
+1];
118 IWL_WARN(priv
, "Unable to find TIM Element in beacon\n");
121 int iwlagn_send_beacon_cmd(struct iwl_priv
*priv
)
123 struct iwl_tx_beacon_cmd
*tx_beacon_cmd
;
124 struct iwl_host_cmd cmd
= {
125 .id
= REPLY_TX_BEACON
,
128 struct ieee80211_tx_info
*info
;
134 * We have to set up the TX command, the TX Beacon command, and the
138 lockdep_assert_held(&priv
->shrd
->mutex
);
140 if (!priv
->beacon_ctx
) {
141 IWL_ERR(priv
, "trying to build beacon w/o beacon context!\n");
145 if (WARN_ON(!priv
->beacon_skb
))
148 /* Allocate beacon command */
149 if (!priv
->beacon_cmd
)
150 priv
->beacon_cmd
= kzalloc(sizeof(*tx_beacon_cmd
), GFP_KERNEL
);
151 tx_beacon_cmd
= priv
->beacon_cmd
;
155 frame_size
= priv
->beacon_skb
->len
;
157 /* Set up TX command fields */
158 tx_beacon_cmd
->tx
.len
= cpu_to_le16((u16
)frame_size
);
159 tx_beacon_cmd
->tx
.sta_id
= priv
->beacon_ctx
->bcast_sta_id
;
160 tx_beacon_cmd
->tx
.stop_time
.life_time
= TX_CMD_LIFE_TIME_INFINITE
;
161 tx_beacon_cmd
->tx
.tx_flags
= TX_CMD_FLG_SEQ_CTL_MSK
|
162 TX_CMD_FLG_TSF_MSK
| TX_CMD_FLG_STA_RATE_MSK
;
164 /* Set up TX beacon command fields */
165 iwl_set_beacon_tim(priv
, tx_beacon_cmd
, priv
->beacon_skb
->data
,
168 /* Set up packet rate and flags */
169 info
= IEEE80211_SKB_CB(priv
->beacon_skb
);
172 * Let's set up the rate at least somewhat correctly;
173 * it will currently not actually be used by the uCode,
174 * it uses the broadcast station's rate instead.
176 if (info
->control
.rates
[0].idx
< 0 ||
177 info
->control
.rates
[0].flags
& IEEE80211_TX_RC_MCS
)
180 rate
= info
->control
.rates
[0].idx
;
182 priv
->mgmt_tx_ant
= iwl_toggle_tx_ant(priv
, priv
->mgmt_tx_ant
,
183 hw_params(priv
).valid_tx_ant
);
184 rate_flags
= iwl_ant_idx_to_flags(priv
->mgmt_tx_ant
);
186 /* In mac80211, rates for 5 GHz start at 0 */
187 if (info
->band
== IEEE80211_BAND_5GHZ
)
188 rate
+= IWL_FIRST_OFDM_RATE
;
189 else if (rate
>= IWL_FIRST_CCK_RATE
&& rate
<= IWL_LAST_CCK_RATE
)
190 rate_flags
|= RATE_MCS_CCK_MSK
;
192 tx_beacon_cmd
->tx
.rate_n_flags
=
193 iwl_hw_set_rate_n_flags(rate
, rate_flags
);
196 cmd
.len
[0] = sizeof(*tx_beacon_cmd
);
197 cmd
.data
[0] = tx_beacon_cmd
;
198 cmd
.dataflags
[0] = IWL_HCMD_DFL_NOCOPY
;
199 cmd
.len
[1] = frame_size
;
200 cmd
.data
[1] = priv
->beacon_skb
->data
;
201 cmd
.dataflags
[1] = IWL_HCMD_DFL_NOCOPY
;
203 return iwl_trans_send_cmd(trans(priv
), &cmd
);
206 static void iwl_bg_beacon_update(struct work_struct
*work
)
208 struct iwl_priv
*priv
=
209 container_of(work
, struct iwl_priv
, beacon_update
);
210 struct sk_buff
*beacon
;
212 mutex_lock(&priv
->shrd
->mutex
);
213 if (!priv
->beacon_ctx
) {
214 IWL_ERR(priv
, "updating beacon w/o beacon context!\n");
218 if (priv
->beacon_ctx
->vif
->type
!= NL80211_IFTYPE_AP
) {
220 * The ucode will send beacon notifications even in
221 * IBSS mode, but we don't want to process them. But
222 * we need to defer the type check to here due to
223 * requiring locking around the beacon_ctx access.
228 /* Pull updated AP beacon from mac80211. will fail if not in AP mode */
229 beacon
= ieee80211_beacon_get(priv
->hw
, priv
->beacon_ctx
->vif
);
231 IWL_ERR(priv
, "update beacon failed -- keeping old\n");
235 /* new beacon skb is allocated every time; dispose previous.*/
236 dev_kfree_skb(priv
->beacon_skb
);
238 priv
->beacon_skb
= beacon
;
240 iwlagn_send_beacon_cmd(priv
);
242 mutex_unlock(&priv
->shrd
->mutex
);
245 static void iwl_bg_bt_runtime_config(struct work_struct
*work
)
247 struct iwl_priv
*priv
=
248 container_of(work
, struct iwl_priv
, bt_runtime_config
);
250 if (test_bit(STATUS_EXIT_PENDING
, &priv
->shrd
->status
))
253 /* dont send host command if rf-kill is on */
254 if (!iwl_is_ready_rf(priv
->shrd
))
256 iwlagn_send_advance_bt_config(priv
);
259 static void iwl_bg_bt_full_concurrency(struct work_struct
*work
)
261 struct iwl_priv
*priv
=
262 container_of(work
, struct iwl_priv
, bt_full_concurrency
);
263 struct iwl_rxon_context
*ctx
;
265 mutex_lock(&priv
->shrd
->mutex
);
267 if (test_bit(STATUS_EXIT_PENDING
, &priv
->shrd
->status
))
270 /* dont send host command if rf-kill is on */
271 if (!iwl_is_ready_rf(priv
->shrd
))
274 IWL_DEBUG_INFO(priv
, "BT coex in %s mode\n",
275 priv
->bt_full_concurrent
?
276 "full concurrency" : "3-wire");
279 * LQ & RXON updated cmds must be sent before BT Config cmd
280 * to avoid 3-wire collisions
282 for_each_context(priv
, ctx
) {
283 iwlagn_set_rxon_chain(priv
, ctx
);
284 iwlagn_commit_rxon(priv
, ctx
);
287 iwlagn_send_advance_bt_config(priv
);
289 mutex_unlock(&priv
->shrd
->mutex
);
293 * iwl_bg_statistics_periodic - Timer callback to queue statistics
295 * This callback is provided in order to send a statistics request.
297 * This timer function is continually reset to execute within
298 * REG_RECALIB_PERIOD seconds since the last STATISTICS_NOTIFICATION
299 * was received. We need to ensure we receive the statistics in order
300 * to update the temperature used for calibrating the TXPOWER.
302 static void iwl_bg_statistics_periodic(unsigned long data
)
304 struct iwl_priv
*priv
= (struct iwl_priv
*)data
;
306 if (test_bit(STATUS_EXIT_PENDING
, &priv
->shrd
->status
))
309 /* dont send host command if rf-kill is on */
310 if (!iwl_is_ready_rf(priv
->shrd
))
313 iwl_send_statistics_request(priv
, CMD_ASYNC
, false);
317 static void iwl_print_cont_event_trace(struct iwl_priv
*priv
, u32 base
,
318 u32 start_idx
, u32 num_events
,
322 u32 ptr
; /* SRAM byte address of log data */
323 u32 ev
, time
, data
; /* event log data */
324 unsigned long reg_flags
;
327 ptr
= base
+ (4 * sizeof(u32
)) + (start_idx
* 2 * sizeof(u32
));
329 ptr
= base
+ (4 * sizeof(u32
)) + (start_idx
* 3 * sizeof(u32
));
331 /* Make sure device is powered up for SRAM reads */
332 spin_lock_irqsave(&bus(priv
)->reg_lock
, reg_flags
);
333 if (iwl_grab_nic_access(bus(priv
))) {
334 spin_unlock_irqrestore(&bus(priv
)->reg_lock
, reg_flags
);
338 /* Set starting address; reads will auto-increment */
339 iwl_write32(bus(priv
), HBUS_TARG_MEM_RADDR
, ptr
);
343 * "time" is actually "data" for mode 0 (no timestamp).
344 * place event id # at far right for easier visual parsing.
346 for (i
= 0; i
< num_events
; i
++) {
347 ev
= iwl_read32(bus(priv
), HBUS_TARG_MEM_RDAT
);
348 time
= iwl_read32(bus(priv
), HBUS_TARG_MEM_RDAT
);
350 trace_iwlwifi_dev_ucode_cont_event(priv
,
353 data
= iwl_read32(bus(priv
), HBUS_TARG_MEM_RDAT
);
354 trace_iwlwifi_dev_ucode_cont_event(priv
,
358 /* Allow device to power down */
359 iwl_release_nic_access(bus(priv
));
360 spin_unlock_irqrestore(&bus(priv
)->reg_lock
, reg_flags
);
363 static void iwl_continuous_event_trace(struct iwl_priv
*priv
)
365 u32 capacity
; /* event log capacity in # entries */
366 u32 base
; /* SRAM byte address of event log header */
367 u32 mode
; /* 0 - no timestamp, 1 - timestamp recorded */
368 u32 num_wraps
; /* # times uCode wrapped to top of log */
369 u32 next_entry
; /* index of next entry to be written by uCode */
371 base
= priv
->device_pointers
.error_event_table
;
372 if (iwlagn_hw_valid_rtc_data_addr(base
)) {
373 capacity
= iwl_read_targ_mem(bus(priv
), base
);
374 num_wraps
= iwl_read_targ_mem(bus(priv
),
375 base
+ (2 * sizeof(u32
)));
376 mode
= iwl_read_targ_mem(bus(priv
), base
+ (1 * sizeof(u32
)));
377 next_entry
= iwl_read_targ_mem(bus(priv
),
378 base
+ (3 * sizeof(u32
)));
382 if (num_wraps
== priv
->event_log
.num_wraps
) {
383 iwl_print_cont_event_trace(priv
,
384 base
, priv
->event_log
.next_entry
,
385 next_entry
- priv
->event_log
.next_entry
,
387 priv
->event_log
.non_wraps_count
++;
389 if ((num_wraps
- priv
->event_log
.num_wraps
) > 1)
390 priv
->event_log
.wraps_more_count
++;
392 priv
->event_log
.wraps_once_count
++;
393 trace_iwlwifi_dev_ucode_wrap_event(priv
,
394 num_wraps
- priv
->event_log
.num_wraps
,
395 next_entry
, priv
->event_log
.next_entry
);
396 if (next_entry
< priv
->event_log
.next_entry
) {
397 iwl_print_cont_event_trace(priv
, base
,
398 priv
->event_log
.next_entry
,
399 capacity
- priv
->event_log
.next_entry
,
402 iwl_print_cont_event_trace(priv
, base
, 0,
405 iwl_print_cont_event_trace(priv
, base
,
406 next_entry
, capacity
- next_entry
,
409 iwl_print_cont_event_trace(priv
, base
, 0,
413 priv
->event_log
.num_wraps
= num_wraps
;
414 priv
->event_log
.next_entry
= next_entry
;
418 * iwl_bg_ucode_trace - Timer callback to log ucode event
420 * The timer is continually set to execute every
421 * UCODE_TRACE_PERIOD milliseconds after the last timer expired
422 * this function is to perform continuous uCode event logging operation
425 static void iwl_bg_ucode_trace(unsigned long data
)
427 struct iwl_priv
*priv
= (struct iwl_priv
*)data
;
429 if (test_bit(STATUS_EXIT_PENDING
, &priv
->shrd
->status
))
432 if (priv
->event_log
.ucode_trace
) {
433 iwl_continuous_event_trace(priv
);
434 /* Reschedule the timer to occur in UCODE_TRACE_PERIOD */
435 mod_timer(&priv
->ucode_trace
,
436 jiffies
+ msecs_to_jiffies(UCODE_TRACE_PERIOD
));
440 static void iwl_bg_tx_flush(struct work_struct
*work
)
442 struct iwl_priv
*priv
=
443 container_of(work
, struct iwl_priv
, tx_flush
);
445 if (test_bit(STATUS_EXIT_PENDING
, &priv
->shrd
->status
))
448 /* do nothing if rf-kill is on */
449 if (!iwl_is_ready_rf(priv
->shrd
))
452 IWL_DEBUG_INFO(priv
, "device request: flush all tx frames\n");
453 iwlagn_dev_txfifo_flush(priv
, IWL_DROP_ALL
);
456 /******************************************************************************
458 * uCode download functions
460 ******************************************************************************/
462 static void iwl_free_fw_desc(struct iwl_priv
*priv
, struct fw_desc
*desc
)
465 dma_free_coherent(bus(priv
)->dev
, desc
->len
,
466 desc
->v_addr
, desc
->p_addr
);
471 static void iwl_free_fw_img(struct iwl_priv
*priv
, struct fw_img
*img
)
473 iwl_free_fw_desc(priv
, &img
->code
);
474 iwl_free_fw_desc(priv
, &img
->data
);
477 static void iwl_dealloc_ucode(struct iwl_priv
*priv
)
479 iwl_free_fw_img(priv
, &priv
->ucode_rt
);
480 iwl_free_fw_img(priv
, &priv
->ucode_init
);
481 iwl_free_fw_img(priv
, &priv
->ucode_wowlan
);
484 static int iwl_alloc_fw_desc(struct iwl_priv
*priv
, struct fw_desc
*desc
,
485 const void *data
, size_t len
)
492 desc
->v_addr
= dma_alloc_coherent(bus(priv
)->dev
, len
,
493 &desc
->p_addr
, GFP_KERNEL
);
498 memcpy(desc
->v_addr
, data
, len
);
502 static void iwl_init_context(struct iwl_priv
*priv
, u32 ucode_flags
)
507 * The default context is always valid,
508 * the PAN context depends on uCode.
510 priv
->shrd
->valid_contexts
= BIT(IWL_RXON_CTX_BSS
);
511 if (ucode_flags
& IWL_UCODE_TLV_FLAGS_PAN
)
512 priv
->shrd
->valid_contexts
|= BIT(IWL_RXON_CTX_PAN
);
514 for (i
= 0; i
< NUM_IWL_RXON_CTX
; i
++)
515 priv
->contexts
[i
].ctxid
= i
;
517 priv
->contexts
[IWL_RXON_CTX_BSS
].always_active
= true;
518 priv
->contexts
[IWL_RXON_CTX_BSS
].is_active
= true;
519 priv
->contexts
[IWL_RXON_CTX_BSS
].rxon_cmd
= REPLY_RXON
;
520 priv
->contexts
[IWL_RXON_CTX_BSS
].rxon_timing_cmd
= REPLY_RXON_TIMING
;
521 priv
->contexts
[IWL_RXON_CTX_BSS
].rxon_assoc_cmd
= REPLY_RXON_ASSOC
;
522 priv
->contexts
[IWL_RXON_CTX_BSS
].qos_cmd
= REPLY_QOS_PARAM
;
523 priv
->contexts
[IWL_RXON_CTX_BSS
].ap_sta_id
= IWL_AP_ID
;
524 priv
->contexts
[IWL_RXON_CTX_BSS
].wep_key_cmd
= REPLY_WEPKEY
;
525 priv
->contexts
[IWL_RXON_CTX_BSS
].exclusive_interface_modes
=
526 BIT(NL80211_IFTYPE_ADHOC
);
527 priv
->contexts
[IWL_RXON_CTX_BSS
].interface_modes
=
528 BIT(NL80211_IFTYPE_STATION
);
529 priv
->contexts
[IWL_RXON_CTX_BSS
].ap_devtype
= RXON_DEV_TYPE_AP
;
530 priv
->contexts
[IWL_RXON_CTX_BSS
].ibss_devtype
= RXON_DEV_TYPE_IBSS
;
531 priv
->contexts
[IWL_RXON_CTX_BSS
].station_devtype
= RXON_DEV_TYPE_ESS
;
532 priv
->contexts
[IWL_RXON_CTX_BSS
].unused_devtype
= RXON_DEV_TYPE_ESS
;
534 priv
->contexts
[IWL_RXON_CTX_PAN
].rxon_cmd
= REPLY_WIPAN_RXON
;
535 priv
->contexts
[IWL_RXON_CTX_PAN
].rxon_timing_cmd
=
536 REPLY_WIPAN_RXON_TIMING
;
537 priv
->contexts
[IWL_RXON_CTX_PAN
].rxon_assoc_cmd
=
538 REPLY_WIPAN_RXON_ASSOC
;
539 priv
->contexts
[IWL_RXON_CTX_PAN
].qos_cmd
= REPLY_WIPAN_QOS_PARAM
;
540 priv
->contexts
[IWL_RXON_CTX_PAN
].ap_sta_id
= IWL_AP_ID_PAN
;
541 priv
->contexts
[IWL_RXON_CTX_PAN
].wep_key_cmd
= REPLY_WIPAN_WEPKEY
;
542 priv
->contexts
[IWL_RXON_CTX_PAN
].bcast_sta_id
= IWLAGN_PAN_BCAST_ID
;
543 priv
->contexts
[IWL_RXON_CTX_PAN
].station_flags
= STA_FLG_PAN_STATION
;
544 priv
->contexts
[IWL_RXON_CTX_PAN
].interface_modes
=
545 BIT(NL80211_IFTYPE_STATION
) | BIT(NL80211_IFTYPE_AP
);
547 if (ucode_flags
& IWL_UCODE_TLV_FLAGS_P2P
)
548 priv
->contexts
[IWL_RXON_CTX_PAN
].interface_modes
|=
549 BIT(NL80211_IFTYPE_P2P_CLIENT
) |
550 BIT(NL80211_IFTYPE_P2P_GO
);
552 priv
->contexts
[IWL_RXON_CTX_PAN
].ap_devtype
= RXON_DEV_TYPE_CP
;
553 priv
->contexts
[IWL_RXON_CTX_PAN
].station_devtype
= RXON_DEV_TYPE_2STA
;
554 priv
->contexts
[IWL_RXON_CTX_PAN
].unused_devtype
= RXON_DEV_TYPE_P2P
;
556 BUILD_BUG_ON(NUM_IWL_RXON_CTX
!= 2);
560 struct iwlagn_ucode_capabilities
{
561 u32 max_probe_length
;
562 u32 standard_phy_calibration_size
;
566 static void iwl_ucode_callback(const struct firmware
*ucode_raw
, void *context
);
567 static int iwlagn_mac_setup_register(struct iwl_priv
*priv
,
568 struct iwlagn_ucode_capabilities
*capa
);
570 #define UCODE_EXPERIMENTAL_INDEX 100
571 #define UCODE_EXPERIMENTAL_TAG "exp"
573 static int __must_check
iwl_request_firmware(struct iwl_priv
*priv
, bool first
)
575 const char *name_pre
= priv
->cfg
->fw_name_pre
;
579 #ifdef CONFIG_IWLWIFI_DEBUG_EXPERIMENTAL_UCODE
580 priv
->fw_index
= UCODE_EXPERIMENTAL_INDEX
;
581 strcpy(tag
, UCODE_EXPERIMENTAL_TAG
);
582 } else if (priv
->fw_index
== UCODE_EXPERIMENTAL_INDEX
) {
584 priv
->fw_index
= priv
->cfg
->ucode_api_max
;
585 sprintf(tag
, "%d", priv
->fw_index
);
588 sprintf(tag
, "%d", priv
->fw_index
);
591 if (priv
->fw_index
< priv
->cfg
->ucode_api_min
) {
592 IWL_ERR(priv
, "no suitable firmware found!\n");
596 sprintf(priv
->firmware_name
, "%s%s%s", name_pre
, tag
, ".ucode");
598 IWL_DEBUG_INFO(priv
, "attempting to load firmware %s'%s'\n",
599 (priv
->fw_index
== UCODE_EXPERIMENTAL_INDEX
)
600 ? "EXPERIMENTAL " : "",
601 priv
->firmware_name
);
603 return request_firmware_nowait(THIS_MODULE
, 1, priv
->firmware_name
,
605 GFP_KERNEL
, priv
, iwl_ucode_callback
);
608 struct iwlagn_firmware_pieces
{
609 const void *inst
, *data
, *init
, *init_data
, *wowlan_inst
, *wowlan_data
;
610 size_t inst_size
, data_size
, init_size
, init_data_size
,
611 wowlan_inst_size
, wowlan_data_size
;
615 u32 init_evtlog_ptr
, init_evtlog_size
, init_errlog_ptr
;
616 u32 inst_evtlog_ptr
, inst_evtlog_size
, inst_errlog_ptr
;
619 static int iwlagn_load_legacy_firmware(struct iwl_priv
*priv
,
620 const struct firmware
*ucode_raw
,
621 struct iwlagn_firmware_pieces
*pieces
)
623 struct iwl_ucode_header
*ucode
= (void *)ucode_raw
->data
;
624 u32 api_ver
, hdr_size
;
627 priv
->ucode_ver
= le32_to_cpu(ucode
->ver
);
628 api_ver
= IWL_UCODE_API(priv
->ucode_ver
);
633 if (ucode_raw
->size
< hdr_size
) {
634 IWL_ERR(priv
, "File size too small!\n");
637 pieces
->build
= le32_to_cpu(ucode
->u
.v2
.build
);
638 pieces
->inst_size
= le32_to_cpu(ucode
->u
.v2
.inst_size
);
639 pieces
->data_size
= le32_to_cpu(ucode
->u
.v2
.data_size
);
640 pieces
->init_size
= le32_to_cpu(ucode
->u
.v2
.init_size
);
641 pieces
->init_data_size
= le32_to_cpu(ucode
->u
.v2
.init_data_size
);
642 src
= ucode
->u
.v2
.data
;
648 if (ucode_raw
->size
< hdr_size
) {
649 IWL_ERR(priv
, "File size too small!\n");
653 pieces
->inst_size
= le32_to_cpu(ucode
->u
.v1
.inst_size
);
654 pieces
->data_size
= le32_to_cpu(ucode
->u
.v1
.data_size
);
655 pieces
->init_size
= le32_to_cpu(ucode
->u
.v1
.init_size
);
656 pieces
->init_data_size
= le32_to_cpu(ucode
->u
.v1
.init_data_size
);
657 src
= ucode
->u
.v1
.data
;
661 /* Verify size of file vs. image size info in file's header */
662 if (ucode_raw
->size
!= hdr_size
+ pieces
->inst_size
+
663 pieces
->data_size
+ pieces
->init_size
+
664 pieces
->init_data_size
) {
667 "uCode file size %d does not match expected size\n",
668 (int)ucode_raw
->size
);
673 src
+= pieces
->inst_size
;
675 src
+= pieces
->data_size
;
677 src
+= pieces
->init_size
;
678 pieces
->init_data
= src
;
679 src
+= pieces
->init_data_size
;
684 static int iwlagn_load_firmware(struct iwl_priv
*priv
,
685 const struct firmware
*ucode_raw
,
686 struct iwlagn_firmware_pieces
*pieces
,
687 struct iwlagn_ucode_capabilities
*capa
)
689 struct iwl_tlv_ucode_header
*ucode
= (void *)ucode_raw
->data
;
690 struct iwl_ucode_tlv
*tlv
;
691 size_t len
= ucode_raw
->size
;
693 int wanted_alternative
= iwlagn_mod_params
.wanted_ucode_alternative
;
697 enum iwl_ucode_tlv_type tlv_type
;
700 if (len
< sizeof(*ucode
)) {
701 IWL_ERR(priv
, "uCode has invalid length: %zd\n", len
);
705 if (ucode
->magic
!= cpu_to_le32(IWL_TLV_UCODE_MAGIC
)) {
706 IWL_ERR(priv
, "invalid uCode magic: 0X%x\n",
707 le32_to_cpu(ucode
->magic
));
712 * Check which alternatives are present, and "downgrade"
713 * when the chosen alternative is not present, warning
714 * the user when that happens. Some files may not have
715 * any alternatives, so don't warn in that case.
717 alternatives
= le64_to_cpu(ucode
->alternatives
);
718 tmp
= wanted_alternative
;
719 if (wanted_alternative
> 63)
720 wanted_alternative
= 63;
721 while (wanted_alternative
&& !(alternatives
& BIT(wanted_alternative
)))
722 wanted_alternative
--;
723 if (wanted_alternative
&& wanted_alternative
!= tmp
)
725 "uCode alternative %d not available, choosing %d\n",
726 tmp
, wanted_alternative
);
728 priv
->ucode_ver
= le32_to_cpu(ucode
->ver
);
729 pieces
->build
= le32_to_cpu(ucode
->build
);
732 len
-= sizeof(*ucode
);
734 while (len
>= sizeof(*tlv
)) {
740 tlv_len
= le32_to_cpu(tlv
->length
);
741 tlv_type
= le16_to_cpu(tlv
->type
);
742 tlv_alt
= le16_to_cpu(tlv
->alternative
);
743 tlv_data
= tlv
->data
;
746 IWL_ERR(priv
, "invalid TLV len: %zd/%u\n",
750 len
-= ALIGN(tlv_len
, 4);
751 data
+= sizeof(*tlv
) + ALIGN(tlv_len
, 4);
754 * Alternative 0 is always valid.
756 * Skip alternative TLVs that are not selected.
758 if (tlv_alt
!= 0 && tlv_alt
!= wanted_alternative
)
762 case IWL_UCODE_TLV_INST
:
763 pieces
->inst
= tlv_data
;
764 pieces
->inst_size
= tlv_len
;
766 case IWL_UCODE_TLV_DATA
:
767 pieces
->data
= tlv_data
;
768 pieces
->data_size
= tlv_len
;
770 case IWL_UCODE_TLV_INIT
:
771 pieces
->init
= tlv_data
;
772 pieces
->init_size
= tlv_len
;
774 case IWL_UCODE_TLV_INIT_DATA
:
775 pieces
->init_data
= tlv_data
;
776 pieces
->init_data_size
= tlv_len
;
778 case IWL_UCODE_TLV_BOOT
:
779 IWL_ERR(priv
, "Found unexpected BOOT ucode\n");
781 case IWL_UCODE_TLV_PROBE_MAX_LEN
:
782 if (tlv_len
!= sizeof(u32
))
783 goto invalid_tlv_len
;
784 capa
->max_probe_length
=
785 le32_to_cpup((__le32
*)tlv_data
);
787 case IWL_UCODE_TLV_PAN
:
789 goto invalid_tlv_len
;
790 capa
->flags
|= IWL_UCODE_TLV_FLAGS_PAN
;
792 case IWL_UCODE_TLV_FLAGS
:
793 /* must be at least one u32 */
794 if (tlv_len
< sizeof(u32
))
795 goto invalid_tlv_len
;
796 /* and a proper number of u32s */
797 if (tlv_len
% sizeof(u32
))
798 goto invalid_tlv_len
;
800 * This driver only reads the first u32 as
801 * right now no more features are defined,
802 * if that changes then either the driver
803 * will not work with the new firmware, or
804 * it'll not take advantage of new features.
806 capa
->flags
= le32_to_cpup((__le32
*)tlv_data
);
808 case IWL_UCODE_TLV_INIT_EVTLOG_PTR
:
809 if (tlv_len
!= sizeof(u32
))
810 goto invalid_tlv_len
;
811 pieces
->init_evtlog_ptr
=
812 le32_to_cpup((__le32
*)tlv_data
);
814 case IWL_UCODE_TLV_INIT_EVTLOG_SIZE
:
815 if (tlv_len
!= sizeof(u32
))
816 goto invalid_tlv_len
;
817 pieces
->init_evtlog_size
=
818 le32_to_cpup((__le32
*)tlv_data
);
820 case IWL_UCODE_TLV_INIT_ERRLOG_PTR
:
821 if (tlv_len
!= sizeof(u32
))
822 goto invalid_tlv_len
;
823 pieces
->init_errlog_ptr
=
824 le32_to_cpup((__le32
*)tlv_data
);
826 case IWL_UCODE_TLV_RUNT_EVTLOG_PTR
:
827 if (tlv_len
!= sizeof(u32
))
828 goto invalid_tlv_len
;
829 pieces
->inst_evtlog_ptr
=
830 le32_to_cpup((__le32
*)tlv_data
);
832 case IWL_UCODE_TLV_RUNT_EVTLOG_SIZE
:
833 if (tlv_len
!= sizeof(u32
))
834 goto invalid_tlv_len
;
835 pieces
->inst_evtlog_size
=
836 le32_to_cpup((__le32
*)tlv_data
);
838 case IWL_UCODE_TLV_RUNT_ERRLOG_PTR
:
839 if (tlv_len
!= sizeof(u32
))
840 goto invalid_tlv_len
;
841 pieces
->inst_errlog_ptr
=
842 le32_to_cpup((__le32
*)tlv_data
);
844 case IWL_UCODE_TLV_ENHANCE_SENS_TBL
:
846 goto invalid_tlv_len
;
847 priv
->enhance_sensitivity_table
= true;
849 case IWL_UCODE_TLV_WOWLAN_INST
:
850 pieces
->wowlan_inst
= tlv_data
;
851 pieces
->wowlan_inst_size
= tlv_len
;
853 case IWL_UCODE_TLV_WOWLAN_DATA
:
854 pieces
->wowlan_data
= tlv_data
;
855 pieces
->wowlan_data_size
= tlv_len
;
857 case IWL_UCODE_TLV_PHY_CALIBRATION_SIZE
:
858 if (tlv_len
!= sizeof(u32
))
859 goto invalid_tlv_len
;
860 capa
->standard_phy_calibration_size
=
861 le32_to_cpup((__le32
*)tlv_data
);
864 IWL_DEBUG_INFO(priv
, "unknown TLV: %d\n", tlv_type
);
870 IWL_ERR(priv
, "invalid TLV after parsing: %zd\n", len
);
871 iwl_print_hex_dump(priv
, IWL_DL_FW
, (u8
*)data
, len
);
878 IWL_ERR(priv
, "TLV %d has invalid size: %u\n", tlv_type
, tlv_len
);
879 iwl_print_hex_dump(priv
, IWL_DL_FW
, tlv_data
, tlv_len
);
885 * iwl_ucode_callback - callback when firmware was loaded
887 * If loaded successfully, copies the firmware into buffers
888 * for the card to fetch (via DMA).
890 static void iwl_ucode_callback(const struct firmware
*ucode_raw
, void *context
)
892 struct iwl_priv
*priv
= context
;
893 struct iwl_ucode_header
*ucode
;
895 struct iwlagn_firmware_pieces pieces
;
896 const unsigned int api_max
= priv
->cfg
->ucode_api_max
;
897 unsigned int api_ok
= priv
->cfg
->ucode_api_ok
;
898 const unsigned int api_min
= priv
->cfg
->ucode_api_min
;
902 struct iwlagn_ucode_capabilities ucode_capa
= {
903 .max_probe_length
= 200,
904 .standard_phy_calibration_size
=
905 IWL_DEFAULT_STANDARD_PHY_CALIBRATE_TBL_SIZE
,
911 memset(&pieces
, 0, sizeof(pieces
));
914 if (priv
->fw_index
<= api_ok
)
916 "request for firmware file '%s' failed.\n",
917 priv
->firmware_name
);
921 IWL_DEBUG_INFO(priv
, "Loaded firmware file '%s' (%zd bytes).\n",
922 priv
->firmware_name
, ucode_raw
->size
);
924 /* Make sure that we got at least the API version number */
925 if (ucode_raw
->size
< 4) {
926 IWL_ERR(priv
, "File size way too small!\n");
930 /* Data from ucode file: header followed by uCode images */
931 ucode
= (struct iwl_ucode_header
*)ucode_raw
->data
;
934 err
= iwlagn_load_legacy_firmware(priv
, ucode_raw
, &pieces
);
936 err
= iwlagn_load_firmware(priv
, ucode_raw
, &pieces
,
942 api_ver
= IWL_UCODE_API(priv
->ucode_ver
);
943 build
= pieces
.build
;
946 * api_ver should match the api version forming part of the
947 * firmware filename ... but we don't check for that and only rely
948 * on the API version read from firmware header from here on forward
950 /* no api version check required for experimental uCode */
951 if (priv
->fw_index
!= UCODE_EXPERIMENTAL_INDEX
) {
952 if (api_ver
< api_min
|| api_ver
> api_max
) {
954 "Driver unable to support your firmware API. "
955 "Driver supports v%u, firmware is v%u.\n",
960 if (api_ver
< api_ok
) {
961 if (api_ok
!= api_max
)
962 IWL_ERR(priv
, "Firmware has old API version, "
963 "expected v%u through v%u, got v%u.\n",
964 api_ok
, api_max
, api_ver
);
966 IWL_ERR(priv
, "Firmware has old API version, "
967 "expected v%u, got v%u.\n",
969 IWL_ERR(priv
, "New firmware can be obtained from "
970 "http://www.intellinuxwireless.org/.\n");
975 sprintf(buildstr
, " build %u%s", build
,
976 (priv
->fw_index
== UCODE_EXPERIMENTAL_INDEX
)
981 IWL_INFO(priv
, "loaded firmware version %u.%u.%u.%u%s\n",
982 IWL_UCODE_MAJOR(priv
->ucode_ver
),
983 IWL_UCODE_MINOR(priv
->ucode_ver
),
984 IWL_UCODE_API(priv
->ucode_ver
),
985 IWL_UCODE_SERIAL(priv
->ucode_ver
),
988 snprintf(priv
->hw
->wiphy
->fw_version
,
989 sizeof(priv
->hw
->wiphy
->fw_version
),
991 IWL_UCODE_MAJOR(priv
->ucode_ver
),
992 IWL_UCODE_MINOR(priv
->ucode_ver
),
993 IWL_UCODE_API(priv
->ucode_ver
),
994 IWL_UCODE_SERIAL(priv
->ucode_ver
),
998 * For any of the failures below (before allocating pci memory)
999 * we will try to load a version with a smaller API -- maybe the
1000 * user just got a corrupted version of the latest API.
1003 IWL_DEBUG_INFO(priv
, "f/w package hdr ucode version raw = 0x%x\n",
1005 IWL_DEBUG_INFO(priv
, "f/w package hdr runtime inst size = %Zd\n",
1007 IWL_DEBUG_INFO(priv
, "f/w package hdr runtime data size = %Zd\n",
1009 IWL_DEBUG_INFO(priv
, "f/w package hdr init inst size = %Zd\n",
1011 IWL_DEBUG_INFO(priv
, "f/w package hdr init data size = %Zd\n",
1012 pieces
.init_data_size
);
1014 /* Verify that uCode images will fit in card's SRAM */
1015 if (pieces
.inst_size
> hw_params(priv
).max_inst_size
) {
1016 IWL_ERR(priv
, "uCode instr len %Zd too large to fit in\n",
1021 if (pieces
.data_size
> hw_params(priv
).max_data_size
) {
1022 IWL_ERR(priv
, "uCode data len %Zd too large to fit in\n",
1027 if (pieces
.init_size
> hw_params(priv
).max_inst_size
) {
1028 IWL_ERR(priv
, "uCode init instr len %Zd too large to fit in\n",
1033 if (pieces
.init_data_size
> hw_params(priv
).max_data_size
) {
1034 IWL_ERR(priv
, "uCode init data len %Zd too large to fit in\n",
1035 pieces
.init_data_size
);
1039 /* Allocate ucode buffers for card's bus-master loading ... */
1041 /* Runtime instructions and 2 copies of data:
1042 * 1) unmodified from disk
1043 * 2) backup cache for save/restore during power-downs */
1044 if (iwl_alloc_fw_desc(priv
, &priv
->ucode_rt
.code
,
1045 pieces
.inst
, pieces
.inst_size
))
1047 if (iwl_alloc_fw_desc(priv
, &priv
->ucode_rt
.data
,
1048 pieces
.data
, pieces
.data_size
))
1051 /* Initialization instructions and data */
1052 if (pieces
.init_size
&& pieces
.init_data_size
) {
1053 if (iwl_alloc_fw_desc(priv
, &priv
->ucode_init
.code
,
1054 pieces
.init
, pieces
.init_size
))
1056 if (iwl_alloc_fw_desc(priv
, &priv
->ucode_init
.data
,
1057 pieces
.init_data
, pieces
.init_data_size
))
1061 /* WoWLAN instructions and data */
1062 if (pieces
.wowlan_inst_size
&& pieces
.wowlan_data_size
) {
1063 if (iwl_alloc_fw_desc(priv
, &priv
->ucode_wowlan
.code
,
1065 pieces
.wowlan_inst_size
))
1067 if (iwl_alloc_fw_desc(priv
, &priv
->ucode_wowlan
.data
,
1069 pieces
.wowlan_data_size
))
1073 /* Now that we can no longer fail, copy information */
1076 * The (size - 16) / 12 formula is based on the information recorded
1077 * for each event, which is of mode 1 (including timestamp) for all
1078 * new microcodes that include this information.
1080 priv
->init_evtlog_ptr
= pieces
.init_evtlog_ptr
;
1081 if (pieces
.init_evtlog_size
)
1082 priv
->init_evtlog_size
= (pieces
.init_evtlog_size
- 16)/12;
1084 priv
->init_evtlog_size
=
1085 priv
->cfg
->base_params
->max_event_log_size
;
1086 priv
->init_errlog_ptr
= pieces
.init_errlog_ptr
;
1087 priv
->inst_evtlog_ptr
= pieces
.inst_evtlog_ptr
;
1088 if (pieces
.inst_evtlog_size
)
1089 priv
->inst_evtlog_size
= (pieces
.inst_evtlog_size
- 16)/12;
1091 priv
->inst_evtlog_size
=
1092 priv
->cfg
->base_params
->max_event_log_size
;
1093 priv
->inst_errlog_ptr
= pieces
.inst_errlog_ptr
;
1095 priv
->new_scan_threshold_behaviour
=
1096 !!(ucode_capa
.flags
& IWL_UCODE_TLV_FLAGS_NEWSCAN
);
1098 if (!(priv
->cfg
->sku
& EEPROM_SKU_CAP_IPAN_ENABLE
))
1099 ucode_capa
.flags
&= ~IWL_UCODE_TLV_FLAGS_PAN
;
1102 * if not PAN, then don't support P2P -- might be a uCode
1103 * packaging bug or due to the eeprom check above
1105 if (!(ucode_capa
.flags
& IWL_UCODE_TLV_FLAGS_PAN
))
1106 ucode_capa
.flags
&= ~IWL_UCODE_TLV_FLAGS_P2P
;
1108 if (ucode_capa
.flags
& IWL_UCODE_TLV_FLAGS_PAN
) {
1109 priv
->sta_key_max_num
= STA_KEY_MAX_NUM_PAN
;
1110 priv
->shrd
->cmd_queue
= IWL_IPAN_CMD_QUEUE_NUM
;
1112 priv
->sta_key_max_num
= STA_KEY_MAX_NUM
;
1113 priv
->shrd
->cmd_queue
= IWL_DEFAULT_CMD_QUEUE_NUM
;
1117 * figure out the offset of chain noise reset and gain commands
1118 * base on the size of standard phy calibration commands table size
1120 if (ucode_capa
.standard_phy_calibration_size
>
1121 IWL_MAX_PHY_CALIBRATE_TBL_SIZE
)
1122 ucode_capa
.standard_phy_calibration_size
=
1123 IWL_MAX_STANDARD_PHY_CALIBRATE_TBL_SIZE
;
1125 priv
->phy_calib_chain_noise_reset_cmd
=
1126 ucode_capa
.standard_phy_calibration_size
;
1127 priv
->phy_calib_chain_noise_gain_cmd
=
1128 ucode_capa
.standard_phy_calibration_size
+ 1;
1130 /* initialize all valid contexts */
1131 iwl_init_context(priv
, ucode_capa
.flags
);
1133 /**************************************************
1134 * This is still part of probe() in a sense...
1136 * 9. Setup and register with mac80211 and debugfs
1137 **************************************************/
1138 err
= iwlagn_mac_setup_register(priv
, &ucode_capa
);
1142 err
= iwl_dbgfs_register(priv
, DRV_NAME
);
1144 IWL_ERR(priv
, "failed to create debugfs files. Ignoring error: %d\n", err
);
1146 /* We have our copies now, allow OS release its copies */
1147 release_firmware(ucode_raw
);
1148 complete(&priv
->firmware_loading_complete
);
1152 /* try next, if any */
1153 if (iwl_request_firmware(priv
, false))
1155 release_firmware(ucode_raw
);
1159 IWL_ERR(priv
, "failed to allocate pci memory\n");
1160 iwl_dealloc_ucode(priv
);
1162 complete(&priv
->firmware_loading_complete
);
1163 device_release_driver(bus(priv
)->dev
);
1164 release_firmware(ucode_raw
);
1167 static void iwl_rf_kill_ct_config(struct iwl_priv
*priv
)
1169 struct iwl_ct_kill_config cmd
;
1170 struct iwl_ct_kill_throttling_config adv_cmd
;
1171 unsigned long flags
;
1174 spin_lock_irqsave(&priv
->shrd
->lock
, flags
);
1175 iwl_write32(bus(priv
), CSR_UCODE_DRV_GP1_CLR
,
1176 CSR_UCODE_DRV_GP1_REG_BIT_CT_KILL_EXIT
);
1177 spin_unlock_irqrestore(&priv
->shrd
->lock
, flags
);
1178 priv
->thermal_throttle
.ct_kill_toggle
= false;
1180 if (priv
->cfg
->base_params
->support_ct_kill_exit
) {
1181 adv_cmd
.critical_temperature_enter
=
1182 cpu_to_le32(hw_params(priv
).ct_kill_threshold
);
1183 adv_cmd
.critical_temperature_exit
=
1184 cpu_to_le32(hw_params(priv
).ct_kill_exit_threshold
);
1186 ret
= iwl_trans_send_cmd_pdu(trans(priv
),
1187 REPLY_CT_KILL_CONFIG_CMD
,
1188 CMD_SYNC
, sizeof(adv_cmd
), &adv_cmd
);
1190 IWL_ERR(priv
, "REPLY_CT_KILL_CONFIG_CMD failed\n");
1192 IWL_DEBUG_INFO(priv
, "REPLY_CT_KILL_CONFIG_CMD "
1193 "succeeded, critical temperature enter is %d,"
1195 hw_params(priv
).ct_kill_threshold
,
1196 hw_params(priv
).ct_kill_exit_threshold
);
1198 cmd
.critical_temperature_R
=
1199 cpu_to_le32(hw_params(priv
).ct_kill_threshold
);
1201 ret
= iwl_trans_send_cmd_pdu(trans(priv
),
1202 REPLY_CT_KILL_CONFIG_CMD
,
1203 CMD_SYNC
, sizeof(cmd
), &cmd
);
1205 IWL_ERR(priv
, "REPLY_CT_KILL_CONFIG_CMD failed\n");
1207 IWL_DEBUG_INFO(priv
, "REPLY_CT_KILL_CONFIG_CMD "
1209 "critical temperature is %d\n",
1210 hw_params(priv
).ct_kill_threshold
);
1214 static int iwlagn_send_calib_cfg_rt(struct iwl_priv
*priv
, u32 cfg
)
1216 struct iwl_calib_cfg_cmd calib_cfg_cmd
;
1217 struct iwl_host_cmd cmd
= {
1218 .id
= CALIBRATION_CFG_CMD
,
1219 .len
= { sizeof(struct iwl_calib_cfg_cmd
), },
1220 .data
= { &calib_cfg_cmd
, },
1223 memset(&calib_cfg_cmd
, 0, sizeof(calib_cfg_cmd
));
1224 calib_cfg_cmd
.ucd_calib_cfg
.once
.is_enable
= IWL_CALIB_RT_CFG_ALL
;
1225 calib_cfg_cmd
.ucd_calib_cfg
.once
.start
= cpu_to_le32(cfg
);
1227 return iwl_trans_send_cmd(trans(priv
), &cmd
);
1231 static int iwlagn_send_tx_ant_config(struct iwl_priv
*priv
, u8 valid_tx_ant
)
1233 struct iwl_tx_ant_config_cmd tx_ant_cmd
= {
1234 .valid
= cpu_to_le32(valid_tx_ant
),
1237 if (IWL_UCODE_API(priv
->ucode_ver
) > 1) {
1238 IWL_DEBUG_HC(priv
, "select valid tx ant: %u\n", valid_tx_ant
);
1239 return iwl_trans_send_cmd_pdu(trans(priv
),
1240 TX_ANT_CONFIGURATION_CMD
,
1242 sizeof(struct iwl_tx_ant_config_cmd
),
1245 IWL_DEBUG_HC(priv
, "TX_ANT_CONFIGURATION_CMD not supported\n");
1251 * iwl_alive_start - called after REPLY_ALIVE notification received
1252 * from protocol/runtime uCode (initialization uCode's
1253 * Alive gets handled by iwl_init_alive_start()).
1255 int iwl_alive_start(struct iwl_priv
*priv
)
1258 struct iwl_rxon_context
*ctx
= &priv
->contexts
[IWL_RXON_CTX_BSS
];
1260 /*TODO: this should go to the transport layer */
1261 iwl_reset_ict(trans(priv
));
1263 IWL_DEBUG_INFO(priv
, "Runtime Alive received.\n");
1265 /* After the ALIVE response, we can send host commands to the uCode */
1266 set_bit(STATUS_ALIVE
, &priv
->shrd
->status
);
1268 /* Enable watchdog to monitor the driver tx queues */
1269 iwl_setup_watchdog(priv
);
1271 if (iwl_is_rfkill(priv
->shrd
))
1274 /* download priority table before any calibration request */
1275 if (priv
->cfg
->bt_params
&&
1276 priv
->cfg
->bt_params
->advanced_bt_coexist
) {
1277 /* Configure Bluetooth device coexistence support */
1278 if (priv
->cfg
->bt_params
->bt_sco_disable
)
1279 priv
->bt_enable_pspoll
= false;
1281 priv
->bt_enable_pspoll
= true;
1283 priv
->bt_valid
= IWLAGN_BT_ALL_VALID_MSK
;
1284 priv
->kill_ack_mask
= IWLAGN_BT_KILL_ACK_MASK_DEFAULT
;
1285 priv
->kill_cts_mask
= IWLAGN_BT_KILL_CTS_MASK_DEFAULT
;
1286 iwlagn_send_advance_bt_config(priv
);
1287 priv
->bt_valid
= IWLAGN_BT_VALID_ENABLE_FLAGS
;
1288 priv
->cur_rssi_ctx
= NULL
;
1290 iwlagn_send_prio_tbl(priv
);
1292 /* FIXME: w/a to force change uCode BT state machine */
1293 ret
= iwlagn_send_bt_env(priv
, IWL_BT_COEX_ENV_OPEN
,
1294 BT_COEX_PRIO_TBL_EVT_INIT_CALIB2
);
1297 ret
= iwlagn_send_bt_env(priv
, IWL_BT_COEX_ENV_CLOSE
,
1298 BT_COEX_PRIO_TBL_EVT_INIT_CALIB2
);
1303 * default is 2-wire BT coexexistence support
1305 iwl_send_bt_config(priv
);
1308 if (hw_params(priv
).calib_rt_cfg
)
1309 iwlagn_send_calib_cfg_rt(priv
,
1310 hw_params(priv
).calib_rt_cfg
);
1312 ieee80211_wake_queues(priv
->hw
);
1314 priv
->active_rate
= IWL_RATES_MASK
;
1316 /* Configure Tx antenna selection based on H/W config */
1317 iwlagn_send_tx_ant_config(priv
, priv
->cfg
->valid_tx_ant
);
1319 if (iwl_is_associated_ctx(ctx
) && !priv
->shrd
->wowlan
) {
1320 struct iwl_rxon_cmd
*active_rxon
=
1321 (struct iwl_rxon_cmd
*)&ctx
->active
;
1322 /* apply any changes in staging */
1323 ctx
->staging
.filter_flags
|= RXON_FILTER_ASSOC_MSK
;
1324 active_rxon
->filter_flags
&= ~RXON_FILTER_ASSOC_MSK
;
1326 struct iwl_rxon_context
*tmp
;
1327 /* Initialize our rx_config data */
1328 for_each_context(priv
, tmp
)
1329 iwl_connection_init_rx_config(priv
, tmp
);
1331 iwlagn_set_rxon_chain(priv
, ctx
);
1334 if (!priv
->shrd
->wowlan
) {
1335 /* WoWLAN ucode will not reply in the same way, skip it */
1336 iwl_reset_run_time_calib(priv
);
1339 set_bit(STATUS_READY
, &priv
->shrd
->status
);
1341 /* Configure the adapter for unassociated operation */
1342 ret
= iwlagn_commit_rxon(priv
, ctx
);
1346 /* At this point, the NIC is initialized and operational */
1347 iwl_rf_kill_ct_config(priv
);
1349 IWL_DEBUG_INFO(priv
, "ALIVE processing complete.\n");
1351 return iwl_power_update_mode(priv
, true);
1354 static void iwl_cancel_deferred_work(struct iwl_priv
*priv
);
1356 static void __iwl_down(struct iwl_priv
*priv
)
1360 IWL_DEBUG_INFO(priv
, DRV_NAME
" is going down\n");
1362 iwl_scan_cancel_timeout(priv
, 200);
1365 * If active, scanning won't cancel it, so say it expired.
1366 * No race since we hold the mutex here and a new one
1367 * can't come in at this time.
1369 ieee80211_remain_on_channel_expired(priv
->hw
);
1372 test_and_set_bit(STATUS_EXIT_PENDING
, &priv
->shrd
->status
);
1374 /* Stop TX queues watchdog. We need to have STATUS_EXIT_PENDING bit set
1375 * to prevent rearm timer */
1376 del_timer_sync(&priv
->watchdog
);
1378 iwl_clear_ucode_stations(priv
, NULL
);
1379 iwl_dealloc_bcast_stations(priv
);
1380 iwl_clear_driver_stations(priv
);
1382 /* reset BT coex data */
1383 priv
->bt_status
= 0;
1384 priv
->cur_rssi_ctx
= NULL
;
1385 priv
->bt_is_sco
= 0;
1386 if (priv
->cfg
->bt_params
)
1387 priv
->bt_traffic_load
=
1388 priv
->cfg
->bt_params
->bt_init_traffic_load
;
1390 priv
->bt_traffic_load
= 0;
1391 priv
->bt_full_concurrent
= false;
1392 priv
->bt_ci_compliance
= 0;
1394 /* Wipe out the EXIT_PENDING status bit if we are not actually
1395 * exiting the module */
1397 clear_bit(STATUS_EXIT_PENDING
, &priv
->shrd
->status
);
1399 if (priv
->mac80211_registered
)
1400 ieee80211_stop_queues(priv
->hw
);
1402 iwl_trans_stop_device(trans(priv
));
1404 /* Clear out all status bits but a few that are stable across reset */
1405 priv
->shrd
->status
&=
1406 test_bit(STATUS_RF_KILL_HW
, &priv
->shrd
->status
) <<
1408 test_bit(STATUS_GEO_CONFIGURED
, &priv
->shrd
->status
) <<
1409 STATUS_GEO_CONFIGURED
|
1410 test_bit(STATUS_FW_ERROR
, &priv
->shrd
->status
) <<
1412 test_bit(STATUS_EXIT_PENDING
, &priv
->shrd
->status
) <<
1413 STATUS_EXIT_PENDING
;
1415 dev_kfree_skb(priv
->beacon_skb
);
1416 priv
->beacon_skb
= NULL
;
1419 static void iwl_down(struct iwl_priv
*priv
)
1421 mutex_lock(&priv
->shrd
->mutex
);
1423 mutex_unlock(&priv
->shrd
->mutex
);
1425 iwl_cancel_deferred_work(priv
);
1428 #define MAX_HW_RESTARTS 5
1430 static int __iwl_up(struct iwl_priv
*priv
)
1432 struct iwl_rxon_context
*ctx
;
1435 lockdep_assert_held(&priv
->shrd
->mutex
);
1437 if (test_bit(STATUS_EXIT_PENDING
, &priv
->shrd
->status
)) {
1438 IWL_WARN(priv
, "Exit pending; will not bring the NIC up\n");
1442 for_each_context(priv
, ctx
) {
1443 ret
= iwlagn_alloc_bcast_station(priv
, ctx
);
1445 iwl_dealloc_bcast_stations(priv
);
1450 ret
= iwlagn_run_init_ucode(priv
);
1452 IWL_ERR(priv
, "Failed to run INIT ucode: %d\n", ret
);
1456 ret
= iwlagn_load_ucode_wait_alive(priv
,
1460 IWL_ERR(priv
, "Failed to start RT ucode: %d\n", ret
);
1464 ret
= iwl_alive_start(priv
);
1470 set_bit(STATUS_EXIT_PENDING
, &priv
->shrd
->status
);
1472 clear_bit(STATUS_EXIT_PENDING
, &priv
->shrd
->status
);
1474 IWL_ERR(priv
, "Unable to initialize device.\n");
1479 /*****************************************************************************
1481 * Workqueue callbacks
1483 *****************************************************************************/
1485 static void iwl_bg_run_time_calib_work(struct work_struct
*work
)
1487 struct iwl_priv
*priv
= container_of(work
, struct iwl_priv
,
1488 run_time_calib_work
);
1490 mutex_lock(&priv
->shrd
->mutex
);
1492 if (test_bit(STATUS_EXIT_PENDING
, &priv
->shrd
->status
) ||
1493 test_bit(STATUS_SCANNING
, &priv
->shrd
->status
)) {
1494 mutex_unlock(&priv
->shrd
->mutex
);
1498 if (priv
->start_calib
) {
1499 iwl_chain_noise_calibration(priv
);
1500 iwl_sensitivity_calibration(priv
);
1503 mutex_unlock(&priv
->shrd
->mutex
);
1506 static void iwlagn_prepare_restart(struct iwl_priv
*priv
)
1508 struct iwl_rxon_context
*ctx
;
1509 bool bt_full_concurrent
;
1510 u8 bt_ci_compliance
;
1515 lockdep_assert_held(&priv
->shrd
->mutex
);
1517 for_each_context(priv
, ctx
)
1522 * __iwl_down() will clear the BT status variables,
1523 * which is correct, but when we restart we really
1524 * want to keep them so restore them afterwards.
1526 * The restart process will later pick them up and
1527 * re-configure the hw when we reconfigure the BT
1530 bt_full_concurrent
= priv
->bt_full_concurrent
;
1531 bt_ci_compliance
= priv
->bt_ci_compliance
;
1532 bt_load
= priv
->bt_traffic_load
;
1533 bt_status
= priv
->bt_status
;
1534 bt_is_sco
= priv
->bt_is_sco
;
1538 priv
->bt_full_concurrent
= bt_full_concurrent
;
1539 priv
->bt_ci_compliance
= bt_ci_compliance
;
1540 priv
->bt_traffic_load
= bt_load
;
1541 priv
->bt_status
= bt_status
;
1542 priv
->bt_is_sco
= bt_is_sco
;
1545 static void iwl_bg_restart(struct work_struct
*data
)
1547 struct iwl_priv
*priv
= container_of(data
, struct iwl_priv
, restart
);
1549 if (test_bit(STATUS_EXIT_PENDING
, &priv
->shrd
->status
))
1552 if (test_and_clear_bit(STATUS_FW_ERROR
, &priv
->shrd
->status
)) {
1553 mutex_lock(&priv
->shrd
->mutex
);
1554 iwlagn_prepare_restart(priv
);
1555 mutex_unlock(&priv
->shrd
->mutex
);
1556 iwl_cancel_deferred_work(priv
);
1557 ieee80211_restart_hw(priv
->hw
);
1563 /*****************************************************************************
1565 * mac80211 entry point functions
1567 *****************************************************************************/
1569 static const struct ieee80211_iface_limit iwlagn_sta_ap_limits
[] = {
1572 .types
= BIT(NL80211_IFTYPE_STATION
),
1576 .types
= BIT(NL80211_IFTYPE_AP
),
1580 static const struct ieee80211_iface_limit iwlagn_2sta_limits
[] = {
1583 .types
= BIT(NL80211_IFTYPE_STATION
),
1587 static const struct ieee80211_iface_limit iwlagn_p2p_sta_go_limits
[] = {
1590 .types
= BIT(NL80211_IFTYPE_STATION
),
1594 .types
= BIT(NL80211_IFTYPE_P2P_GO
) |
1595 BIT(NL80211_IFTYPE_AP
),
1599 static const struct ieee80211_iface_limit iwlagn_p2p_2sta_limits
[] = {
1602 .types
= BIT(NL80211_IFTYPE_STATION
),
1606 .types
= BIT(NL80211_IFTYPE_P2P_CLIENT
),
1610 static const struct ieee80211_iface_combination
1611 iwlagn_iface_combinations_dualmode
[] = {
1612 { .num_different_channels
= 1,
1613 .max_interfaces
= 2,
1614 .beacon_int_infra_match
= true,
1615 .limits
= iwlagn_sta_ap_limits
,
1616 .n_limits
= ARRAY_SIZE(iwlagn_sta_ap_limits
),
1618 { .num_different_channels
= 1,
1619 .max_interfaces
= 2,
1620 .limits
= iwlagn_2sta_limits
,
1621 .n_limits
= ARRAY_SIZE(iwlagn_2sta_limits
),
1625 static const struct ieee80211_iface_combination
1626 iwlagn_iface_combinations_p2p
[] = {
1627 { .num_different_channels
= 1,
1628 .max_interfaces
= 2,
1629 .beacon_int_infra_match
= true,
1630 .limits
= iwlagn_p2p_sta_go_limits
,
1631 .n_limits
= ARRAY_SIZE(iwlagn_p2p_sta_go_limits
),
1633 { .num_different_channels
= 1,
1634 .max_interfaces
= 2,
1635 .limits
= iwlagn_p2p_2sta_limits
,
1636 .n_limits
= ARRAY_SIZE(iwlagn_p2p_2sta_limits
),
1641 * Not a mac80211 entry point function, but it fits in with all the
1642 * other mac80211 functions grouped here.
1644 static int iwlagn_mac_setup_register(struct iwl_priv
*priv
,
1645 struct iwlagn_ucode_capabilities
*capa
)
1648 struct ieee80211_hw
*hw
= priv
->hw
;
1649 struct iwl_rxon_context
*ctx
;
1651 hw
->rate_control_algorithm
= "iwl-agn-rs";
1653 /* Tell mac80211 our characteristics */
1654 hw
->flags
= IEEE80211_HW_SIGNAL_DBM
|
1655 IEEE80211_HW_AMPDU_AGGREGATION
|
1656 IEEE80211_HW_NEED_DTIM_PERIOD
|
1657 IEEE80211_HW_SPECTRUM_MGMT
|
1658 IEEE80211_HW_REPORTS_TX_ACK_STATUS
;
1661 * Including the following line will crash some AP's. This
1662 * workaround removes the stimulus which causes the crash until
1663 * the AP software can be fixed.
1664 hw->max_tx_aggregation_subframes = LINK_QUAL_AGG_FRAME_LIMIT_DEF;
1667 hw
->flags
|= IEEE80211_HW_SUPPORTS_PS
|
1668 IEEE80211_HW_SUPPORTS_DYNAMIC_PS
;
1670 if (priv
->cfg
->sku
& EEPROM_SKU_CAP_11N_ENABLE
)
1671 hw
->flags
|= IEEE80211_HW_SUPPORTS_DYNAMIC_SMPS
|
1672 IEEE80211_HW_SUPPORTS_STATIC_SMPS
;
1674 if (capa
->flags
& IWL_UCODE_TLV_FLAGS_MFP
)
1675 hw
->flags
|= IEEE80211_HW_MFP_CAPABLE
;
1677 hw
->sta_data_size
= sizeof(struct iwl_station_priv
);
1678 hw
->vif_data_size
= sizeof(struct iwl_vif_priv
);
1680 for_each_context(priv
, ctx
) {
1681 hw
->wiphy
->interface_modes
|= ctx
->interface_modes
;
1682 hw
->wiphy
->interface_modes
|= ctx
->exclusive_interface_modes
;
1685 BUILD_BUG_ON(NUM_IWL_RXON_CTX
!= 2);
1687 if (hw
->wiphy
->interface_modes
& BIT(NL80211_IFTYPE_P2P_CLIENT
)) {
1688 hw
->wiphy
->iface_combinations
= iwlagn_iface_combinations_p2p
;
1689 hw
->wiphy
->n_iface_combinations
=
1690 ARRAY_SIZE(iwlagn_iface_combinations_p2p
);
1691 } else if (hw
->wiphy
->interface_modes
& BIT(NL80211_IFTYPE_AP
)) {
1692 hw
->wiphy
->iface_combinations
= iwlagn_iface_combinations_dualmode
;
1693 hw
->wiphy
->n_iface_combinations
=
1694 ARRAY_SIZE(iwlagn_iface_combinations_dualmode
);
1697 hw
->wiphy
->max_remain_on_channel_duration
= 1000;
1699 hw
->wiphy
->flags
|= WIPHY_FLAG_CUSTOM_REGULATORY
|
1700 WIPHY_FLAG_DISABLE_BEACON_HINTS
|
1701 WIPHY_FLAG_IBSS_RSN
;
1703 if (priv
->ucode_wowlan
.code
.len
&& device_can_wakeup(bus(priv
)->dev
)) {
1704 hw
->wiphy
->wowlan
.flags
= WIPHY_WOWLAN_MAGIC_PKT
|
1705 WIPHY_WOWLAN_DISCONNECT
|
1706 WIPHY_WOWLAN_EAP_IDENTITY_REQ
|
1707 WIPHY_WOWLAN_RFKILL_RELEASE
;
1708 if (!iwlagn_mod_params
.sw_crypto
)
1709 hw
->wiphy
->wowlan
.flags
|=
1710 WIPHY_WOWLAN_SUPPORTS_GTK_REKEY
|
1711 WIPHY_WOWLAN_GTK_REKEY_FAILURE
;
1713 hw
->wiphy
->wowlan
.n_patterns
= IWLAGN_WOWLAN_MAX_PATTERNS
;
1714 hw
->wiphy
->wowlan
.pattern_min_len
=
1715 IWLAGN_WOWLAN_MIN_PATTERN_LEN
;
1716 hw
->wiphy
->wowlan
.pattern_max_len
=
1717 IWLAGN_WOWLAN_MAX_PATTERN_LEN
;
1720 if (iwlagn_mod_params
.power_save
)
1721 hw
->wiphy
->flags
|= WIPHY_FLAG_PS_ON_BY_DEFAULT
;
1723 hw
->wiphy
->flags
&= ~WIPHY_FLAG_PS_ON_BY_DEFAULT
;
1725 hw
->wiphy
->max_scan_ssids
= PROBE_OPTION_MAX
;
1726 /* we create the 802.11 header and a zero-length SSID element */
1727 hw
->wiphy
->max_scan_ie_len
= capa
->max_probe_length
- 24 - 2;
1729 /* Default value; 4 EDCA QOS priorities */
1732 hw
->max_listen_interval
= IWL_CONN_MAX_LISTEN_INTERVAL
;
1734 if (priv
->bands
[IEEE80211_BAND_2GHZ
].n_channels
)
1735 priv
->hw
->wiphy
->bands
[IEEE80211_BAND_2GHZ
] =
1736 &priv
->bands
[IEEE80211_BAND_2GHZ
];
1737 if (priv
->bands
[IEEE80211_BAND_5GHZ
].n_channels
)
1738 priv
->hw
->wiphy
->bands
[IEEE80211_BAND_5GHZ
] =
1739 &priv
->bands
[IEEE80211_BAND_5GHZ
];
1741 iwl_leds_init(priv
);
1743 ret
= ieee80211_register_hw(priv
->hw
);
1745 IWL_ERR(priv
, "Failed to register hw (error %d)\n", ret
);
1748 priv
->mac80211_registered
= 1;
1754 static int iwlagn_mac_start(struct ieee80211_hw
*hw
)
1756 struct iwl_priv
*priv
= hw
->priv
;
1759 IWL_DEBUG_MAC80211(priv
, "enter\n");
1761 /* we should be verifying the device is ready to be opened */
1762 mutex_lock(&priv
->shrd
->mutex
);
1763 ret
= __iwl_up(priv
);
1764 mutex_unlock(&priv
->shrd
->mutex
);
1768 IWL_DEBUG_INFO(priv
, "Start UP work done.\n");
1770 /* Now we should be done, and the READY bit should be set. */
1771 if (WARN_ON(!test_bit(STATUS_READY
, &priv
->shrd
->status
)))
1774 iwlagn_led_enable(priv
);
1777 IWL_DEBUG_MAC80211(priv
, "leave\n");
1781 static void iwlagn_mac_stop(struct ieee80211_hw
*hw
)
1783 struct iwl_priv
*priv
= hw
->priv
;
1785 IWL_DEBUG_MAC80211(priv
, "enter\n");
1794 flush_workqueue(priv
->shrd
->workqueue
);
1796 /* User space software may expect getting rfkill changes
1797 * even if interface is down */
1798 iwl_write32(bus(priv
), CSR_INT
, 0xFFFFFFFF);
1799 iwl_enable_rfkill_int(priv
);
1801 IWL_DEBUG_MAC80211(priv
, "leave\n");
1804 #ifdef CONFIG_PM_SLEEP
1805 static int iwlagn_send_patterns(struct iwl_priv
*priv
,
1806 struct cfg80211_wowlan
*wowlan
)
1808 struct iwlagn_wowlan_patterns_cmd
*pattern_cmd
;
1809 struct iwl_host_cmd cmd
= {
1810 .id
= REPLY_WOWLAN_PATTERNS
,
1811 .dataflags
[0] = IWL_HCMD_DFL_NOCOPY
,
1816 if (!wowlan
->n_patterns
)
1819 cmd
.len
[0] = sizeof(*pattern_cmd
) +
1820 wowlan
->n_patterns
* sizeof(struct iwlagn_wowlan_pattern
);
1822 pattern_cmd
= kmalloc(cmd
.len
[0], GFP_KERNEL
);
1826 pattern_cmd
->n_patterns
= cpu_to_le32(wowlan
->n_patterns
);
1828 for (i
= 0; i
< wowlan
->n_patterns
; i
++) {
1829 int mask_len
= DIV_ROUND_UP(wowlan
->patterns
[i
].pattern_len
, 8);
1831 memcpy(&pattern_cmd
->patterns
[i
].mask
,
1832 wowlan
->patterns
[i
].mask
, mask_len
);
1833 memcpy(&pattern_cmd
->patterns
[i
].pattern
,
1834 wowlan
->patterns
[i
].pattern
,
1835 wowlan
->patterns
[i
].pattern_len
);
1836 pattern_cmd
->patterns
[i
].mask_size
= mask_len
;
1837 pattern_cmd
->patterns
[i
].pattern_size
=
1838 wowlan
->patterns
[i
].pattern_len
;
1841 cmd
.data
[0] = pattern_cmd
;
1842 err
= iwl_trans_send_cmd(trans(priv
), &cmd
);
1848 static void iwlagn_mac_set_rekey_data(struct ieee80211_hw
*hw
,
1849 struct ieee80211_vif
*vif
,
1850 struct cfg80211_gtk_rekey_data
*data
)
1852 struct iwl_priv
*priv
= hw
->priv
;
1854 if (iwlagn_mod_params
.sw_crypto
)
1857 IWL_DEBUG_MAC80211(priv
, "enter\n");
1858 mutex_lock(&priv
->shrd
->mutex
);
1860 if (priv
->contexts
[IWL_RXON_CTX_BSS
].vif
!= vif
)
1863 memcpy(priv
->kek
, data
->kek
, NL80211_KEK_LEN
);
1864 memcpy(priv
->kck
, data
->kck
, NL80211_KCK_LEN
);
1865 priv
->replay_ctr
= cpu_to_le64(be64_to_cpup((__be64
*)&data
->replay_ctr
));
1866 priv
->have_rekey_data
= true;
1869 mutex_unlock(&priv
->shrd
->mutex
);
1870 IWL_DEBUG_MAC80211(priv
, "leave\n");
1873 struct wowlan_key_data
{
1874 struct iwl_rxon_context
*ctx
;
1875 struct iwlagn_wowlan_rsc_tsc_params_cmd
*rsc_tsc
;
1876 struct iwlagn_wowlan_tkip_params_cmd
*tkip
;
1878 bool error
, use_rsc_tsc
, use_tkip
;
1881 #ifdef CONFIG_PM_SLEEP
1882 static void iwlagn_convert_p1k(u16
*p1k
, __le16
*out
)
1886 for (i
= 0; i
< IWLAGN_P1K_SIZE
; i
++)
1887 out
[i
] = cpu_to_le16(p1k
[i
]);
1890 static void iwlagn_wowlan_program_keys(struct ieee80211_hw
*hw
,
1891 struct ieee80211_vif
*vif
,
1892 struct ieee80211_sta
*sta
,
1893 struct ieee80211_key_conf
*key
,
1896 struct iwl_priv
*priv
= hw
->priv
;
1897 struct wowlan_key_data
*data
= _data
;
1898 struct iwl_rxon_context
*ctx
= data
->ctx
;
1899 struct aes_sc
*aes_sc
, *aes_tx_sc
= NULL
;
1900 struct tkip_sc
*tkip_sc
, *tkip_tx_sc
= NULL
;
1901 struct iwlagn_p1k_cache
*rx_p1ks
;
1903 struct ieee80211_key_seq seq
;
1904 u32 cur_rx_iv32
= 0;
1905 u16 p1k
[IWLAGN_P1K_SIZE
];
1908 mutex_lock(&priv
->shrd
->mutex
);
1910 if ((key
->cipher
== WLAN_CIPHER_SUITE_WEP40
||
1911 key
->cipher
== WLAN_CIPHER_SUITE_WEP104
) &&
1912 !sta
&& !ctx
->key_mapping_keys
)
1913 ret
= iwl_set_default_wep_key(priv
, ctx
, key
);
1915 ret
= iwl_set_dynamic_key(priv
, ctx
, key
, sta
);
1918 IWL_ERR(priv
, "Error setting key during suspend!\n");
1922 switch (key
->cipher
) {
1923 case WLAN_CIPHER_SUITE_TKIP
:
1925 tkip_sc
= data
->rsc_tsc
->all_tsc_rsc
.tkip
.unicast_rsc
;
1926 tkip_tx_sc
= &data
->rsc_tsc
->all_tsc_rsc
.tkip
.tsc
;
1928 rx_p1ks
= data
->tkip
->rx_uni
;
1930 ieee80211_get_key_tx_seq(key
, &seq
);
1931 tkip_tx_sc
->iv16
= cpu_to_le16(seq
.tkip
.iv16
);
1932 tkip_tx_sc
->iv32
= cpu_to_le32(seq
.tkip
.iv32
);
1934 ieee80211_get_tkip_p1k_iv(key
, seq
.tkip
.iv32
, p1k
);
1935 iwlagn_convert_p1k(p1k
, data
->tkip
->tx
.p1k
);
1937 memcpy(data
->tkip
->mic_keys
.tx
,
1938 &key
->key
[NL80211_TKIP_DATA_OFFSET_TX_MIC_KEY
],
1939 IWLAGN_MIC_KEY_SIZE
);
1941 rx_mic_key
= data
->tkip
->mic_keys
.rx_unicast
;
1943 tkip_sc
= data
->rsc_tsc
->all_tsc_rsc
.tkip
.multicast_rsc
;
1944 rx_p1ks
= data
->tkip
->rx_multi
;
1945 rx_mic_key
= data
->tkip
->mic_keys
.rx_mcast
;
1949 * For non-QoS this relies on the fact that both the uCode and
1950 * mac80211 use TID 0 (as they need to to avoid replay attacks)
1951 * for checking the IV in the frames.
1953 for (i
= 0; i
< IWLAGN_NUM_RSC
; i
++) {
1954 ieee80211_get_key_rx_seq(key
, i
, &seq
);
1955 tkip_sc
[i
].iv16
= cpu_to_le16(seq
.tkip
.iv16
);
1956 tkip_sc
[i
].iv32
= cpu_to_le32(seq
.tkip
.iv32
);
1957 /* wrapping isn't allowed, AP must rekey */
1958 if (seq
.tkip
.iv32
> cur_rx_iv32
)
1959 cur_rx_iv32
= seq
.tkip
.iv32
;
1962 ieee80211_get_tkip_rx_p1k(key
, data
->bssid
, cur_rx_iv32
, p1k
);
1963 iwlagn_convert_p1k(p1k
, rx_p1ks
[0].p1k
);
1964 ieee80211_get_tkip_rx_p1k(key
, data
->bssid
,
1965 cur_rx_iv32
+ 1, p1k
);
1966 iwlagn_convert_p1k(p1k
, rx_p1ks
[1].p1k
);
1969 &key
->key
[NL80211_TKIP_DATA_OFFSET_RX_MIC_KEY
],
1970 IWLAGN_MIC_KEY_SIZE
);
1972 data
->use_tkip
= true;
1973 data
->use_rsc_tsc
= true;
1975 case WLAN_CIPHER_SUITE_CCMP
:
1977 u8
*pn
= seq
.ccmp
.pn
;
1979 aes_sc
= data
->rsc_tsc
->all_tsc_rsc
.aes
.unicast_rsc
;
1980 aes_tx_sc
= &data
->rsc_tsc
->all_tsc_rsc
.aes
.tsc
;
1982 ieee80211_get_key_tx_seq(key
, &seq
);
1983 aes_tx_sc
->pn
= cpu_to_le64(
1986 ((u64
)pn
[3] << 16) |
1987 ((u64
)pn
[2] << 24) |
1988 ((u64
)pn
[1] << 32) |
1989 ((u64
)pn
[0] << 40));
1991 aes_sc
= data
->rsc_tsc
->all_tsc_rsc
.aes
.multicast_rsc
;
1994 * For non-QoS this relies on the fact that both the uCode and
1995 * mac80211 use TID 0 for checking the IV in the frames.
1997 for (i
= 0; i
< IWLAGN_NUM_RSC
; i
++) {
1998 u8
*pn
= seq
.ccmp
.pn
;
2000 ieee80211_get_key_rx_seq(key
, i
, &seq
);
2001 aes_sc
->pn
= cpu_to_le64(
2004 ((u64
)pn
[3] << 16) |
2005 ((u64
)pn
[2] << 24) |
2006 ((u64
)pn
[1] << 32) |
2007 ((u64
)pn
[0] << 40));
2009 data
->use_rsc_tsc
= true;
2013 mutex_unlock(&priv
->shrd
->mutex
);
2016 static int iwlagn_mac_suspend(struct ieee80211_hw
*hw
,
2017 struct cfg80211_wowlan
*wowlan
)
2019 struct iwl_priv
*priv
= hw
->priv
;
2020 struct iwlagn_wowlan_wakeup_filter_cmd wakeup_filter_cmd
;
2021 struct iwl_rxon_cmd rxon
;
2022 struct iwl_rxon_context
*ctx
= &priv
->contexts
[IWL_RXON_CTX_BSS
];
2023 struct iwlagn_wowlan_kek_kck_material_cmd kek_kck_cmd
;
2024 struct iwlagn_wowlan_tkip_params_cmd tkip_cmd
= {};
2025 struct wowlan_key_data key_data
= {
2027 .bssid
= ctx
->active
.bssid_addr
,
2028 .use_rsc_tsc
= false,
2035 if (WARN_ON(!wowlan
))
2038 IWL_DEBUG_MAC80211(priv
, "enter\n");
2039 mutex_lock(&priv
->shrd
->mutex
);
2041 /* Don't attempt WoWLAN when not associated, tear down instead. */
2042 if (!ctx
->vif
|| ctx
->vif
->type
!= NL80211_IFTYPE_STATION
||
2043 !iwl_is_associated_ctx(ctx
)) {
2048 key_data
.rsc_tsc
= kzalloc(sizeof(*key_data
.rsc_tsc
), GFP_KERNEL
);
2049 if (!key_data
.rsc_tsc
) {
2054 memset(&wakeup_filter_cmd
, 0, sizeof(wakeup_filter_cmd
));
2057 * We know the last used seqno, and the uCode expects to know that
2058 * one, it will increment before TX.
2060 seq
= le16_to_cpu(priv
->last_seq_ctl
) & IEEE80211_SCTL_SEQ
;
2061 wakeup_filter_cmd
.non_qos_seq
= cpu_to_le16(seq
);
2064 * For QoS counters, we store the one to use next, so subtract 0x10
2065 * since the uCode will add 0x10 before using the value.
2067 for (i
= 0; i
< 8; i
++) {
2068 seq
= priv
->shrd
->tid_data
[IWL_AP_ID
][i
].seq_number
;
2070 wakeup_filter_cmd
.qos_seq
[i
] = cpu_to_le16(seq
);
2073 if (wowlan
->disconnect
)
2074 wakeup_filter_cmd
.enabled
|=
2075 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_BEACON_MISS
|
2076 IWLAGN_WOWLAN_WAKEUP_LINK_CHANGE
);
2077 if (wowlan
->magic_pkt
)
2078 wakeup_filter_cmd
.enabled
|=
2079 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_MAGIC_PACKET
);
2080 if (wowlan
->gtk_rekey_failure
)
2081 wakeup_filter_cmd
.enabled
|=
2082 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_GTK_REKEY_FAIL
);
2083 if (wowlan
->eap_identity_req
)
2084 wakeup_filter_cmd
.enabled
|=
2085 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_EAP_IDENT_REQ
);
2086 if (wowlan
->four_way_handshake
)
2087 wakeup_filter_cmd
.enabled
|=
2088 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_4WAY_HANDSHAKE
);
2089 if (wowlan
->rfkill_release
)
2090 wakeup_filter_cmd
.enabled
|=
2091 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_RFKILL
);
2092 if (wowlan
->n_patterns
)
2093 wakeup_filter_cmd
.enabled
|=
2094 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_PATTERN_MATCH
);
2096 iwl_scan_cancel_timeout(priv
, 200);
2098 memcpy(&rxon
, &ctx
->active
, sizeof(rxon
));
2100 iwl_trans_stop_device(trans(priv
));
2102 priv
->shrd
->wowlan
= true;
2104 ret
= iwlagn_load_ucode_wait_alive(priv
, &priv
->ucode_wowlan
,
2109 /* now configure WoWLAN ucode */
2110 ret
= iwl_alive_start(priv
);
2114 memcpy(&ctx
->staging
, &rxon
, sizeof(rxon
));
2115 ret
= iwlagn_commit_rxon(priv
, ctx
);
2119 ret
= iwl_power_update_mode(priv
, true);
2123 if (!iwlagn_mod_params
.sw_crypto
) {
2124 /* mark all keys clear */
2125 priv
->ucode_key_table
= 0;
2126 ctx
->key_mapping_keys
= 0;
2129 * This needs to be unlocked due to lock ordering
2130 * constraints. Since we're in the suspend path
2131 * that isn't really a problem though.
2133 mutex_unlock(&priv
->shrd
->mutex
);
2134 ieee80211_iter_keys(priv
->hw
, ctx
->vif
,
2135 iwlagn_wowlan_program_keys
,
2137 mutex_lock(&priv
->shrd
->mutex
);
2138 if (key_data
.error
) {
2143 if (key_data
.use_rsc_tsc
) {
2144 struct iwl_host_cmd rsc_tsc_cmd
= {
2145 .id
= REPLY_WOWLAN_TSC_RSC_PARAMS
,
2147 .data
[0] = key_data
.rsc_tsc
,
2148 .dataflags
[0] = IWL_HCMD_DFL_NOCOPY
,
2149 .len
[0] = sizeof(*key_data
.rsc_tsc
),
2152 ret
= iwl_trans_send_cmd(trans(priv
), &rsc_tsc_cmd
);
2157 if (key_data
.use_tkip
) {
2158 ret
= iwl_trans_send_cmd_pdu(trans(priv
),
2159 REPLY_WOWLAN_TKIP_PARAMS
,
2160 CMD_SYNC
, sizeof(tkip_cmd
),
2166 if (priv
->have_rekey_data
) {
2167 memset(&kek_kck_cmd
, 0, sizeof(kek_kck_cmd
));
2168 memcpy(kek_kck_cmd
.kck
, priv
->kck
, NL80211_KCK_LEN
);
2169 kek_kck_cmd
.kck_len
= cpu_to_le16(NL80211_KCK_LEN
);
2170 memcpy(kek_kck_cmd
.kek
, priv
->kek
, NL80211_KEK_LEN
);
2171 kek_kck_cmd
.kek_len
= cpu_to_le16(NL80211_KEK_LEN
);
2172 kek_kck_cmd
.replay_ctr
= priv
->replay_ctr
;
2174 ret
= iwl_trans_send_cmd_pdu(trans(priv
),
2175 REPLY_WOWLAN_KEK_KCK_MATERIAL
,
2176 CMD_SYNC
, sizeof(kek_kck_cmd
),
2183 ret
= iwl_trans_send_cmd_pdu(trans(priv
), REPLY_WOWLAN_WAKEUP_FILTER
,
2184 CMD_SYNC
, sizeof(wakeup_filter_cmd
),
2185 &wakeup_filter_cmd
);
2189 ret
= iwlagn_send_patterns(priv
, wowlan
);
2193 device_set_wakeup_enable(bus(priv
)->dev
, true);
2195 /* Now let the ucode operate on its own */
2196 iwl_write32(bus(priv
), CSR_UCODE_DRV_GP1_SET
,
2197 CSR_UCODE_DRV_GP1_BIT_D3_CFG_COMPLETE
);
2202 priv
->shrd
->wowlan
= false;
2203 iwlagn_prepare_restart(priv
);
2204 ieee80211_restart_hw(priv
->hw
);
2206 mutex_unlock(&priv
->shrd
->mutex
);
2207 kfree(key_data
.rsc_tsc
);
2208 IWL_DEBUG_MAC80211(priv
, "leave\n");
2213 static int iwlagn_mac_resume(struct ieee80211_hw
*hw
)
2215 struct iwl_priv
*priv
= hw
->priv
;
2216 struct iwl_rxon_context
*ctx
= &priv
->contexts
[IWL_RXON_CTX_BSS
];
2217 struct ieee80211_vif
*vif
;
2218 unsigned long flags
;
2219 u32 base
, status
= 0xffffffff;
2222 IWL_DEBUG_MAC80211(priv
, "enter\n");
2223 mutex_lock(&priv
->shrd
->mutex
);
2225 iwl_write32(bus(priv
), CSR_UCODE_DRV_GP1_CLR
,
2226 CSR_UCODE_DRV_GP1_BIT_D3_CFG_COMPLETE
);
2228 base
= priv
->device_pointers
.error_event_table
;
2229 if (iwlagn_hw_valid_rtc_data_addr(base
)) {
2230 spin_lock_irqsave(&bus(priv
)->reg_lock
, flags
);
2231 ret
= iwl_grab_nic_access_silent(bus(priv
));
2233 iwl_write32(bus(priv
), HBUS_TARG_MEM_RADDR
, base
);
2234 status
= iwl_read32(bus(priv
), HBUS_TARG_MEM_RDAT
);
2235 iwl_release_nic_access(bus(priv
));
2237 spin_unlock_irqrestore(&bus(priv
)->reg_lock
, flags
);
2239 #ifdef CONFIG_IWLWIFI_DEBUGFS
2241 if (!priv
->wowlan_sram
)
2243 kzalloc(priv
->ucode_wowlan
.data
.len
,
2246 if (priv
->wowlan_sram
)
2247 _iwl_read_targ_mem_words(
2248 bus(priv
), 0x800000, priv
->wowlan_sram
,
2249 priv
->ucode_wowlan
.data
.len
/ 4);
2254 /* we'll clear ctx->vif during iwlagn_prepare_restart() */
2257 priv
->shrd
->wowlan
= false;
2259 device_set_wakeup_enable(bus(priv
)->dev
, false);
2261 iwlagn_prepare_restart(priv
);
2263 memset((void *)&ctx
->active
, 0, sizeof(ctx
->active
));
2264 iwl_connection_init_rx_config(priv
, ctx
);
2265 iwlagn_set_rxon_chain(priv
, ctx
);
2267 mutex_unlock(&priv
->shrd
->mutex
);
2268 IWL_DEBUG_MAC80211(priv
, "leave\n");
2270 ieee80211_resume_disconnect(vif
);
2276 static void iwlagn_mac_tx(struct ieee80211_hw
*hw
, struct sk_buff
*skb
)
2278 struct iwl_priv
*priv
= hw
->priv
;
2280 IWL_DEBUG_MACDUMP(priv
, "enter\n");
2282 IWL_DEBUG_TX(priv
, "dev->xmit(%d bytes) at rate 0x%02x\n", skb
->len
,
2283 ieee80211_get_tx_rate(hw
, IEEE80211_SKB_CB(skb
))->bitrate
);
2285 if (iwlagn_tx_skb(priv
, skb
))
2286 dev_kfree_skb_any(skb
);
2288 IWL_DEBUG_MACDUMP(priv
, "leave\n");
2291 static void iwlagn_mac_update_tkip_key(struct ieee80211_hw
*hw
,
2292 struct ieee80211_vif
*vif
,
2293 struct ieee80211_key_conf
*keyconf
,
2294 struct ieee80211_sta
*sta
,
2295 u32 iv32
, u16
*phase1key
)
2297 struct iwl_priv
*priv
= hw
->priv
;
2299 iwl_update_tkip_key(priv
, vif
, keyconf
, sta
, iv32
, phase1key
);
2302 static int iwlagn_mac_set_key(struct ieee80211_hw
*hw
, enum set_key_cmd cmd
,
2303 struct ieee80211_vif
*vif
,
2304 struct ieee80211_sta
*sta
,
2305 struct ieee80211_key_conf
*key
)
2307 struct iwl_priv
*priv
= hw
->priv
;
2308 struct iwl_vif_priv
*vif_priv
= (void *)vif
->drv_priv
;
2309 struct iwl_rxon_context
*ctx
= vif_priv
->ctx
;
2311 bool is_default_wep_key
= false;
2313 IWL_DEBUG_MAC80211(priv
, "enter\n");
2315 if (iwlagn_mod_params
.sw_crypto
) {
2316 IWL_DEBUG_MAC80211(priv
, "leave - hwcrypto disabled\n");
2321 * We could program these keys into the hardware as well, but we
2322 * don't expect much multicast traffic in IBSS and having keys
2323 * for more stations is probably more useful.
2325 * Mark key TX-only and return 0.
2327 if (vif
->type
== NL80211_IFTYPE_ADHOC
&&
2328 !(key
->flags
& IEEE80211_KEY_FLAG_PAIRWISE
)) {
2329 key
->hw_key_idx
= WEP_INVALID_OFFSET
;
2333 /* If they key was TX-only, accept deletion */
2334 if (cmd
== DISABLE_KEY
&& key
->hw_key_idx
== WEP_INVALID_OFFSET
)
2337 mutex_lock(&priv
->shrd
->mutex
);
2338 iwl_scan_cancel_timeout(priv
, 100);
2340 BUILD_BUG_ON(WEP_INVALID_OFFSET
== IWLAGN_HW_KEY_DEFAULT
);
2343 * If we are getting WEP group key and we didn't receive any key mapping
2344 * so far, we are in legacy wep mode (group key only), otherwise we are
2346 * In legacy wep mode, we use another host command to the uCode.
2348 if ((key
->cipher
== WLAN_CIPHER_SUITE_WEP40
||
2349 key
->cipher
== WLAN_CIPHER_SUITE_WEP104
) && !sta
) {
2351 is_default_wep_key
= !ctx
->key_mapping_keys
;
2353 is_default_wep_key
=
2354 key
->hw_key_idx
== IWLAGN_HW_KEY_DEFAULT
;
2360 if (is_default_wep_key
) {
2361 ret
= iwl_set_default_wep_key(priv
, vif_priv
->ctx
, key
);
2364 ret
= iwl_set_dynamic_key(priv
, vif_priv
->ctx
, key
, sta
);
2367 * can't add key for RX, but we don't need it
2368 * in the device for TX so still return 0
2371 key
->hw_key_idx
= WEP_INVALID_OFFSET
;
2374 IWL_DEBUG_MAC80211(priv
, "enable hwcrypto key\n");
2377 if (is_default_wep_key
)
2378 ret
= iwl_remove_default_wep_key(priv
, ctx
, key
);
2380 ret
= iwl_remove_dynamic_key(priv
, ctx
, key
, sta
);
2382 IWL_DEBUG_MAC80211(priv
, "disable hwcrypto key\n");
2388 mutex_unlock(&priv
->shrd
->mutex
);
2389 IWL_DEBUG_MAC80211(priv
, "leave\n");
2394 static int iwlagn_mac_ampdu_action(struct ieee80211_hw
*hw
,
2395 struct ieee80211_vif
*vif
,
2396 enum ieee80211_ampdu_mlme_action action
,
2397 struct ieee80211_sta
*sta
, u16 tid
, u16
*ssn
,
2400 struct iwl_priv
*priv
= hw
->priv
;
2402 struct iwl_station_priv
*sta_priv
= (void *) sta
->drv_priv
;
2403 struct iwl_rxon_context
*ctx
= iwl_rxon_ctx_from_vif(vif
);
2405 IWL_DEBUG_HT(priv
, "A-MPDU action on addr %pM tid %d\n",
2408 if (!(priv
->cfg
->sku
& EEPROM_SKU_CAP_11N_ENABLE
))
2411 IWL_DEBUG_MAC80211(priv
, "enter\n");
2412 mutex_lock(&priv
->shrd
->mutex
);
2415 case IEEE80211_AMPDU_RX_START
:
2416 IWL_DEBUG_HT(priv
, "start Rx\n");
2417 ret
= iwl_sta_rx_agg_start(priv
, sta
, tid
, *ssn
);
2419 case IEEE80211_AMPDU_RX_STOP
:
2420 IWL_DEBUG_HT(priv
, "stop Rx\n");
2421 ret
= iwl_sta_rx_agg_stop(priv
, sta
, tid
);
2422 if (test_bit(STATUS_EXIT_PENDING
, &priv
->shrd
->status
))
2425 case IEEE80211_AMPDU_TX_START
:
2426 IWL_DEBUG_HT(priv
, "start Tx\n");
2427 ret
= iwlagn_tx_agg_start(priv
, vif
, sta
, tid
, ssn
);
2429 case IEEE80211_AMPDU_TX_STOP
:
2430 IWL_DEBUG_HT(priv
, "stop Tx\n");
2431 ret
= iwlagn_tx_agg_stop(priv
, vif
, sta
, tid
);
2432 if ((ret
== 0) && (priv
->agg_tids_count
> 0)) {
2433 priv
->agg_tids_count
--;
2434 IWL_DEBUG_HT(priv
, "priv->agg_tids_count = %u\n",
2435 priv
->agg_tids_count
);
2437 if (test_bit(STATUS_EXIT_PENDING
, &priv
->shrd
->status
))
2439 if (!priv
->agg_tids_count
&& priv
->cfg
->ht_params
&&
2440 priv
->cfg
->ht_params
->use_rts_for_aggregation
) {
2442 * switch off RTS/CTS if it was previously enabled
2444 sta_priv
->lq_sta
.lq
.general_params
.flags
&=
2445 ~LINK_QUAL_FLAGS_SET_STA_TLC_RTS_MSK
;
2446 iwl_send_lq_cmd(priv
, iwl_rxon_ctx_from_vif(vif
),
2447 &sta_priv
->lq_sta
.lq
, CMD_ASYNC
, false);
2450 case IEEE80211_AMPDU_TX_OPERATIONAL
:
2451 buf_size
= min_t(int, buf_size
, LINK_QUAL_AGG_FRAME_LIMIT_DEF
);
2453 iwl_trans_tx_agg_setup(trans(priv
), ctx
->ctxid
, iwl_sta_id(sta
),
2457 * If the limit is 0, then it wasn't initialised yet,
2458 * use the default. We can do that since we take the
2459 * minimum below, and we don't want to go above our
2460 * default due to hardware restrictions.
2462 if (sta_priv
->max_agg_bufsize
== 0)
2463 sta_priv
->max_agg_bufsize
=
2464 LINK_QUAL_AGG_FRAME_LIMIT_DEF
;
2467 * Even though in theory the peer could have different
2468 * aggregation reorder buffer sizes for different sessions,
2469 * our ucode doesn't allow for that and has a global limit
2470 * for each station. Therefore, use the minimum of all the
2471 * aggregation sessions and our default value.
2473 sta_priv
->max_agg_bufsize
=
2474 min(sta_priv
->max_agg_bufsize
, buf_size
);
2476 if (priv
->cfg
->ht_params
&&
2477 priv
->cfg
->ht_params
->use_rts_for_aggregation
) {
2479 * switch to RTS/CTS if it is the prefer protection
2480 * method for HT traffic
2483 sta_priv
->lq_sta
.lq
.general_params
.flags
|=
2484 LINK_QUAL_FLAGS_SET_STA_TLC_RTS_MSK
;
2486 priv
->agg_tids_count
++;
2487 IWL_DEBUG_HT(priv
, "priv->agg_tids_count = %u\n",
2488 priv
->agg_tids_count
);
2490 sta_priv
->lq_sta
.lq
.agg_params
.agg_frame_cnt_limit
=
2491 sta_priv
->max_agg_bufsize
;
2493 iwl_send_lq_cmd(priv
, iwl_rxon_ctx_from_vif(vif
),
2494 &sta_priv
->lq_sta
.lq
, CMD_ASYNC
, false);
2496 IWL_INFO(priv
, "Tx aggregation enabled on ra = %pM tid = %d\n",
2501 mutex_unlock(&priv
->shrd
->mutex
);
2502 IWL_DEBUG_MAC80211(priv
, "leave\n");
2506 static int iwlagn_mac_sta_add(struct ieee80211_hw
*hw
,
2507 struct ieee80211_vif
*vif
,
2508 struct ieee80211_sta
*sta
)
2510 struct iwl_priv
*priv
= hw
->priv
;
2511 struct iwl_station_priv
*sta_priv
= (void *)sta
->drv_priv
;
2512 struct iwl_vif_priv
*vif_priv
= (void *)vif
->drv_priv
;
2513 bool is_ap
= vif
->type
== NL80211_IFTYPE_STATION
;
2517 IWL_DEBUG_MAC80211(priv
, "received request to add station %pM\n",
2519 mutex_lock(&priv
->shrd
->mutex
);
2520 IWL_DEBUG_INFO(priv
, "proceeding to add station %pM\n",
2522 sta_priv
->sta_id
= IWL_INVALID_STATION
;
2524 atomic_set(&sta_priv
->pending_frames
, 0);
2525 if (vif
->type
== NL80211_IFTYPE_AP
)
2526 sta_priv
->client
= true;
2528 ret
= iwl_add_station_common(priv
, vif_priv
->ctx
, sta
->addr
,
2529 is_ap
, sta
, &sta_id
);
2531 IWL_ERR(priv
, "Unable to add station %pM (%d)\n",
2533 /* Should we return success if return code is EEXIST ? */
2537 sta_priv
->sta_id
= sta_id
;
2539 /* Initialize rate scaling */
2540 IWL_DEBUG_INFO(priv
, "Initializing rate scaling for station %pM\n",
2542 iwl_rs_rate_init(priv
, sta
, sta_id
);
2544 mutex_unlock(&priv
->shrd
->mutex
);
2545 IWL_DEBUG_MAC80211(priv
, "leave\n");
2550 static void iwlagn_mac_channel_switch(struct ieee80211_hw
*hw
,
2551 struct ieee80211_channel_switch
*ch_switch
)
2553 struct iwl_priv
*priv
= hw
->priv
;
2554 const struct iwl_channel_info
*ch_info
;
2555 struct ieee80211_conf
*conf
= &hw
->conf
;
2556 struct ieee80211_channel
*channel
= ch_switch
->channel
;
2557 struct iwl_ht_config
*ht_conf
= &priv
->current_ht_config
;
2560 * When we add support for multiple interfaces, we need to
2561 * revisit this. The channel switch command in the device
2562 * only affects the BSS context, but what does that really
2563 * mean? And what if we get a CSA on the second interface?
2564 * This needs a lot of work.
2566 struct iwl_rxon_context
*ctx
= &priv
->contexts
[IWL_RXON_CTX_BSS
];
2569 IWL_DEBUG_MAC80211(priv
, "enter\n");
2571 mutex_lock(&priv
->shrd
->mutex
);
2573 if (iwl_is_rfkill(priv
->shrd
))
2576 if (test_bit(STATUS_EXIT_PENDING
, &priv
->shrd
->status
) ||
2577 test_bit(STATUS_SCANNING
, &priv
->shrd
->status
) ||
2578 test_bit(STATUS_CHANNEL_SWITCH_PENDING
, &priv
->shrd
->status
))
2581 if (!iwl_is_associated_ctx(ctx
))
2584 if (!priv
->cfg
->lib
->set_channel_switch
)
2587 ch
= channel
->hw_value
;
2588 if (le16_to_cpu(ctx
->active
.channel
) == ch
)
2591 ch_info
= iwl_get_channel_info(priv
, channel
->band
, ch
);
2592 if (!is_channel_valid(ch_info
)) {
2593 IWL_DEBUG_MAC80211(priv
, "invalid channel\n");
2597 spin_lock_irq(&priv
->shrd
->lock
);
2599 priv
->current_ht_config
.smps
= conf
->smps_mode
;
2601 /* Configure HT40 channels */
2602 ctx
->ht
.enabled
= conf_is_ht(conf
);
2603 if (ctx
->ht
.enabled
) {
2604 if (conf_is_ht40_minus(conf
)) {
2605 ctx
->ht
.extension_chan_offset
=
2606 IEEE80211_HT_PARAM_CHA_SEC_BELOW
;
2607 ctx
->ht
.is_40mhz
= true;
2608 } else if (conf_is_ht40_plus(conf
)) {
2609 ctx
->ht
.extension_chan_offset
=
2610 IEEE80211_HT_PARAM_CHA_SEC_ABOVE
;
2611 ctx
->ht
.is_40mhz
= true;
2613 ctx
->ht
.extension_chan_offset
=
2614 IEEE80211_HT_PARAM_CHA_SEC_NONE
;
2615 ctx
->ht
.is_40mhz
= false;
2618 ctx
->ht
.is_40mhz
= false;
2620 if ((le16_to_cpu(ctx
->staging
.channel
) != ch
))
2621 ctx
->staging
.flags
= 0;
2623 iwl_set_rxon_channel(priv
, channel
, ctx
);
2624 iwl_set_rxon_ht(priv
, ht_conf
);
2625 iwl_set_flags_for_band(priv
, ctx
, channel
->band
, ctx
->vif
);
2627 spin_unlock_irq(&priv
->shrd
->lock
);
2631 * at this point, staging_rxon has the
2632 * configuration for channel switch
2634 set_bit(STATUS_CHANNEL_SWITCH_PENDING
, &priv
->shrd
->status
);
2635 priv
->switch_channel
= cpu_to_le16(ch
);
2636 if (priv
->cfg
->lib
->set_channel_switch(priv
, ch_switch
)) {
2637 clear_bit(STATUS_CHANNEL_SWITCH_PENDING
, &priv
->shrd
->status
);
2638 priv
->switch_channel
= 0;
2639 ieee80211_chswitch_done(ctx
->vif
, false);
2643 mutex_unlock(&priv
->shrd
->mutex
);
2644 IWL_DEBUG_MAC80211(priv
, "leave\n");
2647 static void iwlagn_configure_filter(struct ieee80211_hw
*hw
,
2648 unsigned int changed_flags
,
2649 unsigned int *total_flags
,
2652 struct iwl_priv
*priv
= hw
->priv
;
2653 __le32 filter_or
= 0, filter_nand
= 0;
2654 struct iwl_rxon_context
*ctx
;
2656 #define CHK(test, flag) do { \
2657 if (*total_flags & (test)) \
2658 filter_or |= (flag); \
2660 filter_nand |= (flag); \
2663 IWL_DEBUG_MAC80211(priv
, "Enter: changed: 0x%x, total: 0x%x\n",
2664 changed_flags
, *total_flags
);
2666 CHK(FIF_OTHER_BSS
| FIF_PROMISC_IN_BSS
, RXON_FILTER_PROMISC_MSK
);
2667 /* Setting _just_ RXON_FILTER_CTL2HOST_MSK causes FH errors */
2668 CHK(FIF_CONTROL
, RXON_FILTER_CTL2HOST_MSK
| RXON_FILTER_PROMISC_MSK
);
2669 CHK(FIF_BCN_PRBRESP_PROMISC
, RXON_FILTER_BCON_AWARE_MSK
);
2673 mutex_lock(&priv
->shrd
->mutex
);
2675 for_each_context(priv
, ctx
) {
2676 ctx
->staging
.filter_flags
&= ~filter_nand
;
2677 ctx
->staging
.filter_flags
|= filter_or
;
2680 * Not committing directly because hardware can perform a scan,
2681 * but we'll eventually commit the filter flags change anyway.
2685 mutex_unlock(&priv
->shrd
->mutex
);
2688 * Receiving all multicast frames is always enabled by the
2689 * default flags setup in iwl_connection_init_rx_config()
2690 * since we currently do not support programming multicast
2691 * filters into the device.
2693 *total_flags
&= FIF_OTHER_BSS
| FIF_ALLMULTI
| FIF_PROMISC_IN_BSS
|
2694 FIF_BCN_PRBRESP_PROMISC
| FIF_CONTROL
;
2697 static void iwlagn_mac_flush(struct ieee80211_hw
*hw
, bool drop
)
2699 struct iwl_priv
*priv
= hw
->priv
;
2701 mutex_lock(&priv
->shrd
->mutex
);
2702 IWL_DEBUG_MAC80211(priv
, "enter\n");
2704 if (test_bit(STATUS_EXIT_PENDING
, &priv
->shrd
->status
)) {
2705 IWL_DEBUG_TX(priv
, "Aborting flush due to device shutdown\n");
2708 if (iwl_is_rfkill(priv
->shrd
)) {
2709 IWL_DEBUG_TX(priv
, "Aborting flush due to RF Kill\n");
2714 * mac80211 will not push any more frames for transmit
2715 * until the flush is completed
2718 IWL_DEBUG_MAC80211(priv
, "send flush command\n");
2719 if (iwlagn_txfifo_flush(priv
, IWL_DROP_ALL
)) {
2720 IWL_ERR(priv
, "flush request fail\n");
2724 IWL_DEBUG_MAC80211(priv
, "wait transmit/flush all frames\n");
2725 iwl_trans_wait_tx_queue_empty(trans(priv
));
2727 mutex_unlock(&priv
->shrd
->mutex
);
2728 IWL_DEBUG_MAC80211(priv
, "leave\n");
2731 void iwlagn_disable_roc(struct iwl_priv
*priv
)
2733 struct iwl_rxon_context
*ctx
= &priv
->contexts
[IWL_RXON_CTX_PAN
];
2735 lockdep_assert_held(&priv
->shrd
->mutex
);
2737 if (!priv
->hw_roc_setup
)
2740 ctx
->staging
.dev_type
= RXON_DEV_TYPE_P2P
;
2741 ctx
->staging
.filter_flags
&= ~RXON_FILTER_ASSOC_MSK
;
2743 priv
->hw_roc_channel
= NULL
;
2745 memset(ctx
->staging
.node_addr
, 0, ETH_ALEN
);
2747 iwlagn_commit_rxon(priv
, ctx
);
2749 ctx
->is_active
= false;
2750 priv
->hw_roc_setup
= false;
2753 static void iwlagn_disable_roc_work(struct work_struct
*work
)
2755 struct iwl_priv
*priv
= container_of(work
, struct iwl_priv
,
2756 hw_roc_disable_work
.work
);
2758 mutex_lock(&priv
->shrd
->mutex
);
2759 iwlagn_disable_roc(priv
);
2760 mutex_unlock(&priv
->shrd
->mutex
);
2763 static int iwlagn_mac_remain_on_channel(struct ieee80211_hw
*hw
,
2764 struct ieee80211_channel
*channel
,
2765 enum nl80211_channel_type channel_type
,
2768 struct iwl_priv
*priv
= hw
->priv
;
2769 struct iwl_rxon_context
*ctx
= &priv
->contexts
[IWL_RXON_CTX_PAN
];
2772 if (!(priv
->shrd
->valid_contexts
& BIT(IWL_RXON_CTX_PAN
)))
2775 if (!(ctx
->interface_modes
& BIT(NL80211_IFTYPE_P2P_CLIENT
)))
2778 IWL_DEBUG_MAC80211(priv
, "enter\n");
2779 mutex_lock(&priv
->shrd
->mutex
);
2781 if (test_bit(STATUS_SCAN_HW
, &priv
->shrd
->status
)) {
2786 priv
->hw_roc_channel
= channel
;
2787 priv
->hw_roc_chantype
= channel_type
;
2788 priv
->hw_roc_duration
= duration
;
2789 priv
->hw_roc_start_notified
= false;
2790 cancel_delayed_work(&priv
->hw_roc_disable_work
);
2792 if (!ctx
->is_active
) {
2793 ctx
->is_active
= true;
2794 ctx
->staging
.dev_type
= RXON_DEV_TYPE_P2P
;
2795 memcpy(ctx
->staging
.node_addr
,
2796 priv
->contexts
[IWL_RXON_CTX_BSS
].staging
.node_addr
,
2798 memcpy(ctx
->staging
.bssid_addr
,
2799 priv
->contexts
[IWL_RXON_CTX_BSS
].staging
.node_addr
,
2801 err
= iwlagn_commit_rxon(priv
, ctx
);
2804 ctx
->staging
.filter_flags
|= RXON_FILTER_ASSOC_MSK
|
2805 RXON_FILTER_PROMISC_MSK
|
2806 RXON_FILTER_CTL2HOST_MSK
;
2808 err
= iwlagn_commit_rxon(priv
, ctx
);
2810 iwlagn_disable_roc(priv
);
2813 priv
->hw_roc_setup
= true;
2816 err
= iwl_scan_initiate(priv
, ctx
->vif
, IWL_SCAN_ROC
, channel
->band
);
2818 iwlagn_disable_roc(priv
);
2821 mutex_unlock(&priv
->shrd
->mutex
);
2822 IWL_DEBUG_MAC80211(priv
, "leave\n");
2827 static int iwlagn_mac_cancel_remain_on_channel(struct ieee80211_hw
*hw
)
2829 struct iwl_priv
*priv
= hw
->priv
;
2831 if (!(priv
->shrd
->valid_contexts
& BIT(IWL_RXON_CTX_PAN
)))
2834 IWL_DEBUG_MAC80211(priv
, "enter\n");
2835 mutex_lock(&priv
->shrd
->mutex
);
2836 iwl_scan_cancel_timeout(priv
, priv
->hw_roc_duration
);
2837 iwlagn_disable_roc(priv
);
2838 mutex_unlock(&priv
->shrd
->mutex
);
2839 IWL_DEBUG_MAC80211(priv
, "leave\n");
2844 static int iwlagn_mac_tx_sync(struct ieee80211_hw
*hw
,
2845 struct ieee80211_vif
*vif
,
2847 enum ieee80211_tx_sync_type type
)
2849 struct iwl_priv
*priv
= hw
->priv
;
2850 struct iwl_vif_priv
*vif_priv
= (void *)vif
->drv_priv
;
2851 struct iwl_rxon_context
*ctx
= vif_priv
->ctx
;
2855 IWL_DEBUG_MAC80211(priv
, "enter\n");
2856 mutex_lock(&priv
->shrd
->mutex
);
2858 if (iwl_is_associated_ctx(ctx
)) {
2863 if (ctx
->preauth_bssid
|| test_bit(STATUS_SCAN_HW
, &priv
->shrd
->status
)) {
2868 ret
= iwl_add_station_common(priv
, ctx
, bssid
, true, NULL
, &sta_id
);
2872 if (WARN_ON(sta_id
!= ctx
->ap_sta_id
)) {
2874 goto out_remove_sta
;
2877 memcpy(ctx
->bssid
, bssid
, ETH_ALEN
);
2878 ctx
->preauth_bssid
= true;
2880 ret
= iwlagn_commit_rxon(priv
, ctx
);
2886 iwl_remove_station(priv
, sta_id
, bssid
);
2888 mutex_unlock(&priv
->shrd
->mutex
);
2889 IWL_DEBUG_MAC80211(priv
, "leave\n");
2894 static void iwlagn_mac_finish_tx_sync(struct ieee80211_hw
*hw
,
2895 struct ieee80211_vif
*vif
,
2897 enum ieee80211_tx_sync_type type
)
2899 struct iwl_priv
*priv
= hw
->priv
;
2900 struct iwl_vif_priv
*vif_priv
= (void *)vif
->drv_priv
;
2901 struct iwl_rxon_context
*ctx
= vif_priv
->ctx
;
2903 IWL_DEBUG_MAC80211(priv
, "enter\n");
2904 mutex_lock(&priv
->shrd
->mutex
);
2906 if (iwl_is_associated_ctx(ctx
))
2909 iwl_remove_station(priv
, ctx
->ap_sta_id
, bssid
);
2910 ctx
->preauth_bssid
= false;
2911 /* no need to commit */
2913 mutex_unlock(&priv
->shrd
->mutex
);
2914 IWL_DEBUG_MAC80211(priv
, "leave\n");
2917 /*****************************************************************************
2919 * driver setup and teardown
2921 *****************************************************************************/
2923 static void iwl_setup_deferred_work(struct iwl_priv
*priv
)
2925 priv
->shrd
->workqueue
= create_singlethread_workqueue(DRV_NAME
);
2927 init_waitqueue_head(&priv
->shrd
->wait_command_queue
);
2929 INIT_WORK(&priv
->restart
, iwl_bg_restart
);
2930 INIT_WORK(&priv
->beacon_update
, iwl_bg_beacon_update
);
2931 INIT_WORK(&priv
->run_time_calib_work
, iwl_bg_run_time_calib_work
);
2932 INIT_WORK(&priv
->tx_flush
, iwl_bg_tx_flush
);
2933 INIT_WORK(&priv
->bt_full_concurrency
, iwl_bg_bt_full_concurrency
);
2934 INIT_WORK(&priv
->bt_runtime_config
, iwl_bg_bt_runtime_config
);
2935 INIT_DELAYED_WORK(&priv
->hw_roc_disable_work
,
2936 iwlagn_disable_roc_work
);
2938 iwl_setup_scan_deferred_work(priv
);
2940 if (priv
->cfg
->lib
->bt_setup_deferred_work
)
2941 priv
->cfg
->lib
->bt_setup_deferred_work(priv
);
2943 init_timer(&priv
->statistics_periodic
);
2944 priv
->statistics_periodic
.data
= (unsigned long)priv
;
2945 priv
->statistics_periodic
.function
= iwl_bg_statistics_periodic
;
2947 init_timer(&priv
->ucode_trace
);
2948 priv
->ucode_trace
.data
= (unsigned long)priv
;
2949 priv
->ucode_trace
.function
= iwl_bg_ucode_trace
;
2951 init_timer(&priv
->watchdog
);
2952 priv
->watchdog
.data
= (unsigned long)priv
;
2953 priv
->watchdog
.function
= iwl_bg_watchdog
;
2956 static void iwl_cancel_deferred_work(struct iwl_priv
*priv
)
2958 if (priv
->cfg
->lib
->cancel_deferred_work
)
2959 priv
->cfg
->lib
->cancel_deferred_work(priv
);
2961 cancel_work_sync(&priv
->run_time_calib_work
);
2962 cancel_work_sync(&priv
->beacon_update
);
2964 iwl_cancel_scan_deferred_work(priv
);
2966 cancel_work_sync(&priv
->bt_full_concurrency
);
2967 cancel_work_sync(&priv
->bt_runtime_config
);
2968 cancel_delayed_work_sync(&priv
->hw_roc_disable_work
);
2970 del_timer_sync(&priv
->statistics_periodic
);
2971 del_timer_sync(&priv
->ucode_trace
);
2974 static void iwl_init_hw_rates(struct iwl_priv
*priv
,
2975 struct ieee80211_rate
*rates
)
2979 for (i
= 0; i
< IWL_RATE_COUNT_LEGACY
; i
++) {
2980 rates
[i
].bitrate
= iwl_rates
[i
].ieee
* 5;
2981 rates
[i
].hw_value
= i
; /* Rate scaling will work on indexes */
2982 rates
[i
].hw_value_short
= i
;
2984 if ((i
>= IWL_FIRST_CCK_RATE
) && (i
<= IWL_LAST_CCK_RATE
)) {
2986 * If CCK != 1M then set short preamble rate flag.
2989 (iwl_rates
[i
].plcp
== IWL_RATE_1M_PLCP
) ?
2990 0 : IEEE80211_RATE_SHORT_PREAMBLE
;
2995 static int iwl_init_drv(struct iwl_priv
*priv
)
2999 spin_lock_init(&priv
->shrd
->sta_lock
);
3001 mutex_init(&priv
->shrd
->mutex
);
3003 priv
->ieee_channels
= NULL
;
3004 priv
->ieee_rates
= NULL
;
3005 priv
->band
= IEEE80211_BAND_2GHZ
;
3007 priv
->iw_mode
= NL80211_IFTYPE_STATION
;
3008 priv
->current_ht_config
.smps
= IEEE80211_SMPS_STATIC
;
3009 priv
->missed_beacon_threshold
= IWL_MISSED_BEACON_THRESHOLD_DEF
;
3010 priv
->agg_tids_count
= 0;
3012 /* initialize force reset */
3013 priv
->force_reset
[IWL_RF_RESET
].reset_duration
=
3014 IWL_DELAY_NEXT_FORCE_RF_RESET
;
3015 priv
->force_reset
[IWL_FW_RESET
].reset_duration
=
3016 IWL_DELAY_NEXT_FORCE_FW_RELOAD
;
3018 priv
->rx_statistics_jiffies
= jiffies
;
3020 /* Choose which receivers/antennas to use */
3021 iwlagn_set_rxon_chain(priv
, &priv
->contexts
[IWL_RXON_CTX_BSS
]);
3023 iwl_init_scan_params(priv
);
3026 if (priv
->cfg
->bt_params
&&
3027 priv
->cfg
->bt_params
->advanced_bt_coexist
) {
3028 priv
->kill_ack_mask
= IWLAGN_BT_KILL_ACK_MASK_DEFAULT
;
3029 priv
->kill_cts_mask
= IWLAGN_BT_KILL_CTS_MASK_DEFAULT
;
3030 priv
->bt_valid
= IWLAGN_BT_ALL_VALID_MSK
;
3031 priv
->bt_on_thresh
= BT_ON_THRESHOLD_DEF
;
3032 priv
->bt_duration
= BT_DURATION_LIMIT_DEF
;
3033 priv
->dynamic_frag_thresh
= BT_FRAG_THRESHOLD_DEF
;
3036 ret
= iwl_init_channel_map(priv
);
3038 IWL_ERR(priv
, "initializing regulatory failed: %d\n", ret
);
3042 ret
= iwl_init_geos(priv
);
3044 IWL_ERR(priv
, "initializing geos failed: %d\n", ret
);
3045 goto err_free_channel_map
;
3047 iwl_init_hw_rates(priv
, priv
->ieee_rates
);
3051 err_free_channel_map
:
3052 iwl_free_channel_map(priv
);
3057 static void iwl_uninit_drv(struct iwl_priv
*priv
)
3059 iwl_calib_free_results(priv
);
3060 iwl_free_geos(priv
);
3061 iwl_free_channel_map(priv
);
3062 if (priv
->tx_cmd_pool
)
3063 kmem_cache_destroy(priv
->tx_cmd_pool
);
3064 kfree(priv
->scan_cmd
);
3065 kfree(priv
->beacon_cmd
);
3066 #ifdef CONFIG_IWLWIFI_DEBUGFS
3067 kfree(priv
->wowlan_sram
);
3071 static void iwlagn_mac_rssi_callback(struct ieee80211_hw
*hw
,
3072 enum ieee80211_rssi_event rssi_event
)
3074 struct iwl_priv
*priv
= hw
->priv
;
3076 IWL_DEBUG_MAC80211(priv
, "enter\n");
3077 mutex_lock(&priv
->shrd
->mutex
);
3079 if (priv
->cfg
->bt_params
&&
3080 priv
->cfg
->bt_params
->advanced_bt_coexist
) {
3081 if (rssi_event
== RSSI_EVENT_LOW
)
3082 priv
->bt_enable_pspoll
= true;
3083 else if (rssi_event
== RSSI_EVENT_HIGH
)
3084 priv
->bt_enable_pspoll
= false;
3086 iwlagn_send_advance_bt_config(priv
);
3088 IWL_DEBUG_MAC80211(priv
, "Advanced BT coex disabled,"
3089 "ignoring RSSI callback\n");
3092 mutex_unlock(&priv
->shrd
->mutex
);
3093 IWL_DEBUG_MAC80211(priv
, "leave\n");
3096 static int iwlagn_mac_set_tim(struct ieee80211_hw
*hw
,
3097 struct ieee80211_sta
*sta
, bool set
)
3099 struct iwl_priv
*priv
= hw
->priv
;
3101 queue_work(priv
->shrd
->workqueue
, &priv
->beacon_update
);
3106 struct ieee80211_ops iwlagn_hw_ops
= {
3107 .tx
= iwlagn_mac_tx
,
3108 .start
= iwlagn_mac_start
,
3109 .stop
= iwlagn_mac_stop
,
3110 #ifdef CONFIG_PM_SLEEP
3111 .suspend
= iwlagn_mac_suspend
,
3112 .resume
= iwlagn_mac_resume
,
3114 .add_interface
= iwlagn_mac_add_interface
,
3115 .remove_interface
= iwlagn_mac_remove_interface
,
3116 .change_interface
= iwlagn_mac_change_interface
,
3117 .config
= iwlagn_mac_config
,
3118 .configure_filter
= iwlagn_configure_filter
,
3119 .set_key
= iwlagn_mac_set_key
,
3120 .update_tkip_key
= iwlagn_mac_update_tkip_key
,
3121 .set_rekey_data
= iwlagn_mac_set_rekey_data
,
3122 .conf_tx
= iwlagn_mac_conf_tx
,
3123 .bss_info_changed
= iwlagn_bss_info_changed
,
3124 .ampdu_action
= iwlagn_mac_ampdu_action
,
3125 .hw_scan
= iwlagn_mac_hw_scan
,
3126 .sta_notify
= iwlagn_mac_sta_notify
,
3127 .sta_add
= iwlagn_mac_sta_add
,
3128 .sta_remove
= iwlagn_mac_sta_remove
,
3129 .channel_switch
= iwlagn_mac_channel_switch
,
3130 .flush
= iwlagn_mac_flush
,
3131 .tx_last_beacon
= iwlagn_mac_tx_last_beacon
,
3132 .remain_on_channel
= iwlagn_mac_remain_on_channel
,
3133 .cancel_remain_on_channel
= iwlagn_mac_cancel_remain_on_channel
,
3134 .rssi_callback
= iwlagn_mac_rssi_callback
,
3135 CFG80211_TESTMODE_CMD(iwlagn_mac_testmode_cmd
)
3136 CFG80211_TESTMODE_DUMP(iwlagn_mac_testmode_dump
)
3137 .tx_sync
= iwlagn_mac_tx_sync
,
3138 .finish_tx_sync
= iwlagn_mac_finish_tx_sync
,
3139 .set_tim
= iwlagn_mac_set_tim
,
3142 static u32
iwl_hw_detect(struct iwl_priv
*priv
)
3144 return iwl_read32(bus(priv
), CSR_HW_REV
);
3147 /* Size of one Rx buffer in host DRAM */
3148 #define IWL_RX_BUF_SIZE_4K (4 * 1024)
3149 #define IWL_RX_BUF_SIZE_8K (8 * 1024)
3151 static int iwl_set_hw_params(struct iwl_priv
*priv
)
3153 if (iwlagn_mod_params
.amsdu_size_8K
)
3154 hw_params(priv
).rx_page_order
=
3155 get_order(IWL_RX_BUF_SIZE_8K
);
3157 hw_params(priv
).rx_page_order
=
3158 get_order(IWL_RX_BUF_SIZE_4K
);
3160 if (iwlagn_mod_params
.disable_11n
)
3161 priv
->cfg
->sku
&= ~EEPROM_SKU_CAP_11N_ENABLE
;
3163 hw_params(priv
).num_ampdu_queues
=
3164 priv
->cfg
->base_params
->num_of_ampdu_queues
;
3165 hw_params(priv
).shadow_reg_enable
=
3166 priv
->cfg
->base_params
->shadow_reg_enable
;
3167 hw_params(priv
).sku
= priv
->cfg
->sku
;
3168 hw_params(priv
).wd_timeout
= priv
->cfg
->base_params
->wd_timeout
;
3170 /* Device-specific setup */
3171 return priv
->cfg
->lib
->set_hw_params(priv
);
3174 /* This function both allocates and initializes hw and priv. */
3175 static struct ieee80211_hw
*iwl_alloc_all(struct iwl_cfg
*cfg
)
3177 struct iwl_priv
*priv
;
3178 /* mac80211 allocates memory for this device instance, including
3179 * space for this driver's private structure */
3180 struct ieee80211_hw
*hw
;
3182 hw
= ieee80211_alloc_hw(sizeof(struct iwl_priv
), &iwlagn_hw_ops
);
3184 pr_err("%s: Can not allocate network device\n",
3196 int iwl_probe(struct iwl_bus
*bus
, const struct iwl_trans_ops
*trans_ops
,
3197 struct iwl_cfg
*cfg
)
3200 struct iwl_priv
*priv
;
3201 struct ieee80211_hw
*hw
;
3205 /************************
3206 * 1. Allocating HW data
3207 ************************/
3208 hw
= iwl_alloc_all(cfg
);
3215 priv
->shrd
= &priv
->_shrd
;
3216 bus
->shrd
= priv
->shrd
;
3217 priv
->shrd
->bus
= bus
;
3218 priv
->shrd
->priv
= priv
;
3220 priv
->shrd
->trans
= trans_ops
->alloc(priv
->shrd
);
3221 if (priv
->shrd
->trans
== NULL
) {
3223 goto out_free_traffic_mem
;
3226 /* At this point both hw and priv are allocated. */
3228 SET_IEEE80211_DEV(hw
, bus(priv
)->dev
);
3230 IWL_DEBUG_INFO(priv
, "*** LOAD DRIVER ***\n");
3233 /* is antenna coupling more than 35dB ? */
3234 priv
->bt_ant_couple_ok
=
3235 (iwlagn_mod_params
.ant_coupling
>
3236 IWL_BT_ANTENNA_COUPLING_THRESHOLD
) ?
3239 /* enable/disable bt channel inhibition */
3240 priv
->bt_ch_announce
= iwlagn_mod_params
.bt_ch_announce
;
3241 IWL_DEBUG_INFO(priv
, "BT channel inhibition is %s\n",
3242 (priv
->bt_ch_announce
) ? "On" : "Off");
3244 if (iwl_alloc_traffic_mem(priv
))
3245 IWL_ERR(priv
, "Not enough memory to generate traffic log\n");
3247 /* these spin locks will be used in apm_ops.init and EEPROM access
3248 * we should init now
3250 spin_lock_init(&bus(priv
)->reg_lock
);
3251 spin_lock_init(&priv
->shrd
->lock
);
3254 * stop and reset the on-board processor just in case it is in a
3255 * strange state ... like being left stranded by a primary kernel
3256 * and this is now the kdump kernel trying to start up
3258 iwl_write32(bus(priv
), CSR_RESET
, CSR_RESET_REG_FLAG_NEVO_RESET
);
3260 /***********************
3261 * 3. Read REV register
3262 ***********************/
3263 hw_rev
= iwl_hw_detect(priv
);
3264 IWL_INFO(priv
, "Detected %s, REV=0x%X\n",
3265 priv
->cfg
->name
, hw_rev
);
3267 err
= iwl_trans_request_irq(trans(priv
));
3269 goto out_free_trans
;
3271 if (iwl_trans_prepare_card_hw(trans(priv
))) {
3273 IWL_WARN(priv
, "Failed, HW not ready\n");
3274 goto out_free_trans
;
3280 /* Read the EEPROM */
3281 err
= iwl_eeprom_init(priv
, hw_rev
);
3283 IWL_ERR(priv
, "Unable to init EEPROM\n");
3284 goto out_free_trans
;
3286 err
= iwl_eeprom_check_version(priv
);
3288 goto out_free_eeprom
;
3290 err
= iwl_eeprom_check_sku(priv
);
3292 goto out_free_eeprom
;
3294 /* extract MAC Address */
3295 iwl_eeprom_get_mac(priv
, priv
->addresses
[0].addr
);
3296 IWL_DEBUG_INFO(priv
, "MAC address: %pM\n", priv
->addresses
[0].addr
);
3297 priv
->hw
->wiphy
->addresses
= priv
->addresses
;
3298 priv
->hw
->wiphy
->n_addresses
= 1;
3299 num_mac
= iwl_eeprom_query16(priv
, EEPROM_NUM_MAC_ADDRESS
);
3301 memcpy(priv
->addresses
[1].addr
, priv
->addresses
[0].addr
,
3303 priv
->addresses
[1].addr
[5]++;
3304 priv
->hw
->wiphy
->n_addresses
++;
3307 /************************
3308 * 5. Setup HW constants
3309 ************************/
3310 if (iwl_set_hw_params(priv
)) {
3312 IWL_ERR(priv
, "failed to set hw parameters\n");
3313 goto out_free_eeprom
;
3316 /*******************
3318 *******************/
3320 err
= iwl_init_drv(priv
);
3322 goto out_free_eeprom
;
3323 /* At this point both hw and priv are initialized. */
3325 /********************
3327 ********************/
3328 iwl_setup_deferred_work(priv
);
3329 iwl_setup_rx_handlers(priv
);
3330 iwl_testmode_init(priv
);
3332 /*********************************************
3333 * 8. Enable interrupts
3334 *********************************************/
3336 iwl_enable_rfkill_int(priv
);
3338 /* If platform's RF_KILL switch is NOT set to KILL */
3339 if (iwl_read32(bus(priv
),
3340 CSR_GP_CNTRL
) & CSR_GP_CNTRL_REG_FLAG_HW_RF_KILL_SW
)
3341 clear_bit(STATUS_RF_KILL_HW
, &priv
->shrd
->status
);
3343 set_bit(STATUS_RF_KILL_HW
, &priv
->shrd
->status
);
3345 wiphy_rfkill_set_hw_state(priv
->hw
->wiphy
,
3346 test_bit(STATUS_RF_KILL_HW
, &priv
->shrd
->status
));
3348 iwl_power_initialize(priv
);
3349 iwl_tt_initialize(priv
);
3351 init_completion(&priv
->firmware_loading_complete
);
3353 err
= iwl_request_firmware(priv
, true);
3355 goto out_destroy_workqueue
;
3359 out_destroy_workqueue
:
3360 destroy_workqueue(priv
->shrd
->workqueue
);
3361 priv
->shrd
->workqueue
= NULL
;
3362 iwl_uninit_drv(priv
);
3364 iwl_eeprom_free(priv
);
3366 iwl_trans_free(trans(priv
));
3367 out_free_traffic_mem
:
3368 iwl_free_traffic_mem(priv
);
3369 ieee80211_free_hw(priv
->hw
);
3374 void __devexit
iwl_remove(struct iwl_priv
* priv
)
3376 wait_for_completion(&priv
->firmware_loading_complete
);
3378 IWL_DEBUG_INFO(priv
, "*** UNLOAD DRIVER ***\n");
3380 iwl_dbgfs_unregister(priv
);
3382 /* ieee80211_unregister_hw call wil cause iwlagn_mac_stop to
3383 * to be called and iwl_down since we are removing the device
3384 * we need to set STATUS_EXIT_PENDING bit.
3386 set_bit(STATUS_EXIT_PENDING
, &priv
->shrd
->status
);
3388 iwl_testmode_cleanup(priv
);
3389 iwl_leds_exit(priv
);
3391 if (priv
->mac80211_registered
) {
3392 ieee80211_unregister_hw(priv
->hw
);
3393 priv
->mac80211_registered
= 0;
3398 /*This will stop the queues, move the device to low power state */
3399 iwl_trans_stop_device(trans(priv
));
3401 iwl_dealloc_ucode(priv
);
3403 iwl_eeprom_free(priv
);
3405 /*netif_stop_queue(dev); */
3406 flush_workqueue(priv
->shrd
->workqueue
);
3408 /* ieee80211_unregister_hw calls iwlagn_mac_stop, which flushes
3409 * priv->shrd->workqueue... so we can't take down the workqueue
3411 destroy_workqueue(priv
->shrd
->workqueue
);
3412 priv
->shrd
->workqueue
= NULL
;
3413 iwl_free_traffic_mem(priv
);
3415 iwl_trans_free(trans(priv
));
3417 iwl_uninit_drv(priv
);
3419 dev_kfree_skb(priv
->beacon_skb
);
3421 ieee80211_free_hw(priv
->hw
);
3425 /*****************************************************************************
3427 * driver and module entry point
3429 *****************************************************************************/
3430 static int __init
iwl_init(void)
3434 pr_info(DRV_DESCRIPTION
", " DRV_VERSION
"\n");
3435 pr_info(DRV_COPYRIGHT
"\n");
3437 ret
= iwlagn_rate_control_register();
3439 pr_err("Unable to register rate control algorithm: %d\n", ret
);
3443 ret
= iwl_pci_register_driver();
3446 goto error_register
;
3450 iwlagn_rate_control_unregister();
3454 static void __exit
iwl_exit(void)
3456 iwl_pci_unregister_driver();
3457 iwlagn_rate_control_unregister();
3460 module_exit(iwl_exit
);
3461 module_init(iwl_init
);
3463 #ifdef CONFIG_IWLWIFI_DEBUG
3464 module_param_named(debug
, iwlagn_mod_params
.debug_level
, uint
,
3466 MODULE_PARM_DESC(debug
, "debug output mask");
3469 module_param_named(swcrypto
, iwlagn_mod_params
.sw_crypto
, int, S_IRUGO
);
3470 MODULE_PARM_DESC(swcrypto
, "using crypto in software (default 0 [hardware])");
3471 module_param_named(queues_num
, iwlagn_mod_params
.num_of_queues
, int, S_IRUGO
);
3472 MODULE_PARM_DESC(queues_num
, "number of hw queues.");
3473 module_param_named(11n_disable
, iwlagn_mod_params
.disable_11n
, int, S_IRUGO
);
3474 MODULE_PARM_DESC(11n_disable
, "disable 11n functionality");
3475 module_param_named(amsdu_size_8K
, iwlagn_mod_params
.amsdu_size_8K
,
3477 MODULE_PARM_DESC(amsdu_size_8K
, "enable 8K amsdu size");
3478 module_param_named(fw_restart
, iwlagn_mod_params
.restart_fw
, int, S_IRUGO
);
3479 MODULE_PARM_DESC(fw_restart
, "restart firmware in case of error");
3481 module_param_named(ucode_alternative
,
3482 iwlagn_mod_params
.wanted_ucode_alternative
,
3484 MODULE_PARM_DESC(ucode_alternative
,
3485 "specify ucode alternative to use from ucode file");
3487 module_param_named(antenna_coupling
, iwlagn_mod_params
.ant_coupling
,
3489 MODULE_PARM_DESC(antenna_coupling
,
3490 "specify antenna coupling in dB (defualt: 0 dB)");
3492 module_param_named(bt_ch_inhibition
, iwlagn_mod_params
.bt_ch_announce
,
3494 MODULE_PARM_DESC(bt_ch_inhibition
,
3495 "Enable BT channel inhibition (default: enable)");
3497 module_param_named(plcp_check
, iwlagn_mod_params
.plcp_check
, bool, S_IRUGO
);
3498 MODULE_PARM_DESC(plcp_check
, "Check plcp health (default: 1 [enabled])");
3500 module_param_named(ack_check
, iwlagn_mod_params
.ack_check
, bool, S_IRUGO
);
3501 MODULE_PARM_DESC(ack_check
, "Check ack health (default: 0 [disabled])");
3503 module_param_named(wd_disable
, iwlagn_mod_params
.wd_disable
, bool, S_IRUGO
);
3504 MODULE_PARM_DESC(wd_disable
,
3505 "Disable stuck queue watchdog timer (default: 0 [enabled])");
3508 * set bt_coex_active to true, uCode will do kill/defer
3509 * every time the priority line is asserted (BT is sending signals on the
3510 * priority line in the PCIx).
3511 * set bt_coex_active to false, uCode will ignore the BT activity and
3512 * perform the normal operation
3514 * User might experience transmit issue on some platform due to WiFi/BT
3515 * co-exist problem. The possible behaviors are:
3516 * Able to scan and finding all the available AP
3517 * Not able to associate with any AP
3518 * On those platforms, WiFi communication can be restored by set
3519 * "bt_coex_active" module parameter to "false"
3521 * default: bt_coex_active = true (BT_COEX_ENABLE)
3523 module_param_named(bt_coex_active
, iwlagn_mod_params
.bt_coex_active
,
3525 MODULE_PARM_DESC(bt_coex_active
, "enable wifi/bt co-exist (default: enable)");
3527 module_param_named(led_mode
, iwlagn_mod_params
.led_mode
, int, S_IRUGO
);
3528 MODULE_PARM_DESC(led_mode
, "0=system default, "
3529 "1=On(RF On)/Off(RF Off), 2=blinking (default: 0)");
3531 module_param_named(power_save
, iwlagn_mod_params
.power_save
,
3533 MODULE_PARM_DESC(power_save
,
3534 "enable WiFi power management (default: disable)");
3536 module_param_named(power_level
, iwlagn_mod_params
.power_level
,
3538 MODULE_PARM_DESC(power_level
,
3539 "default power save level (range from 1 - 5, default: 1)");
3541 module_param_named(auto_agg
, iwlagn_mod_params
.auto_agg
,
3543 MODULE_PARM_DESC(auto_agg
,
3544 "enable agg w/o check traffic load (default: enable)");
3547 * For now, keep using power level 1 instead of automatically
3550 module_param_named(no_sleep_autoadjust
, iwlagn_mod_params
.no_sleep_autoadjust
,
3552 MODULE_PARM_DESC(no_sleep_autoadjust
,
3553 "don't automatically adjust sleep level "
3554 "according to maximum network latency (default: true)");