2 * Atheros AR9170 driver
4 * mac80211 interaction code
6 * Copyright 2008, Johannes Berg <johannes@sipsolutions.net>
7 * Copyright 2009, Christian Lamparter <chunkeey@web.de>
9 * This program is free software; you can redistribute it and/or modify
10 * it under the terms of the GNU General Public License as published by
11 * the Free Software Foundation; either version 2 of the License, or
12 * (at your option) any later version.
14 * This program is distributed in the hope that it will be useful,
15 * but WITHOUT ANY WARRANTY; without even the implied warranty of
16 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
17 * GNU General Public License for more details.
19 * You should have received a copy of the GNU General Public License
20 * along with this program; see the file COPYING. If not, see
21 * http://www.gnu.org/licenses/.
23 * This file incorporates work covered by the following copyright and
25 * Copyright (c) 2007-2008 Atheros Communications, Inc.
27 * Permission to use, copy, modify, and/or distribute this software for any
28 * purpose with or without fee is hereby granted, provided that the above
29 * copyright notice and this permission notice appear in all copies.
31 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
32 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
33 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
34 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
35 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
36 * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
37 * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
40 #include <linux/init.h>
41 #include <linux/module.h>
42 #include <linux/etherdevice.h>
43 #include <net/mac80211.h>
48 static int modparam_nohwcrypt
;
49 module_param_named(nohwcrypt
, modparam_nohwcrypt
, bool, S_IRUGO
);
50 MODULE_PARM_DESC(nohwcrypt
, "Disable hardware encryption.");
52 static int modparam_ht
;
53 module_param_named(ht
, modparam_ht
, bool, S_IRUGO
);
54 MODULE_PARM_DESC(ht
, "enable MPDU aggregation.");
56 #define RATE(_bitrate, _hw_rate, _txpidx, _flags) { \
57 .bitrate = (_bitrate), \
59 .hw_value = (_hw_rate) | (_txpidx) << 4, \
62 static struct ieee80211_rate __ar9170_ratetable
[] = {
64 RATE(20, 1, 1, IEEE80211_RATE_SHORT_PREAMBLE
),
65 RATE(55, 2, 2, IEEE80211_RATE_SHORT_PREAMBLE
),
66 RATE(110, 3, 3, IEEE80211_RATE_SHORT_PREAMBLE
),
78 #define ar9170_g_ratetable (__ar9170_ratetable + 0)
79 #define ar9170_g_ratetable_size 12
80 #define ar9170_a_ratetable (__ar9170_ratetable + 4)
81 #define ar9170_a_ratetable_size 8
84 * NB: The hw_value is used as an index into the ar9170_phy_freq_params
85 * array in phy.c so that we don't have to do frequency lookups!
87 #define CHAN(_freq, _idx) { \
88 .center_freq = (_freq), \
90 .max_power = 18, /* XXX */ \
93 static struct ieee80211_channel ar9170_2ghz_chantable
[] = {
110 static struct ieee80211_channel ar9170_5ghz_chantable
[] = {
149 #define AR9170_HT_CAP \
151 .ht_supported = true, \
152 .cap = IEEE80211_HT_CAP_MAX_AMSDU | \
153 IEEE80211_HT_CAP_SUP_WIDTH_20_40 | \
154 IEEE80211_HT_CAP_SGI_40 | \
155 IEEE80211_HT_CAP_GRN_FLD | \
156 IEEE80211_HT_CAP_DSSSCCK40 | \
157 IEEE80211_HT_CAP_SM_PS, \
159 .ampdu_density = 6, \
161 .rx_mask = { 0xff, 0xff, 0, 0, 0x1, 0, 0, 0, 0, 0, }, \
162 .rx_highest = cpu_to_le16(300), \
163 .tx_params = IEEE80211_HT_MCS_TX_DEFINED, \
167 static struct ieee80211_supported_band ar9170_band_2GHz
= {
168 .channels
= ar9170_2ghz_chantable
,
169 .n_channels
= ARRAY_SIZE(ar9170_2ghz_chantable
),
170 .bitrates
= ar9170_g_ratetable
,
171 .n_bitrates
= ar9170_g_ratetable_size
,
172 .ht_cap
= AR9170_HT_CAP
,
175 static struct ieee80211_supported_band ar9170_band_5GHz
= {
176 .channels
= ar9170_5ghz_chantable
,
177 .n_channels
= ARRAY_SIZE(ar9170_5ghz_chantable
),
178 .bitrates
= ar9170_a_ratetable
,
179 .n_bitrates
= ar9170_a_ratetable_size
,
180 .ht_cap
= AR9170_HT_CAP
,
183 static void ar9170_tx(struct ar9170
*ar
);
184 static bool ar9170_tx_ampdu(struct ar9170
*ar
);
186 static inline u16
ar9170_get_seq_h(struct ieee80211_hdr
*hdr
)
188 return le16_to_cpu(hdr
->seq_ctrl
) >> 4;
191 static inline u16
ar9170_get_seq(struct sk_buff
*skb
)
193 struct ar9170_tx_control
*txc
= (void *) skb
->data
;
194 return ar9170_get_seq_h((void *) txc
->frame_data
);
197 static inline u16
ar9170_get_tid(struct sk_buff
*skb
)
199 struct ar9170_tx_control
*txc
= (void *) skb
->data
;
200 struct ieee80211_hdr
*hdr
= (void *) txc
->frame_data
;
202 return (ieee80211_get_qos_ctl(hdr
))[0] & IEEE80211_QOS_CTL_TID_MASK
;
205 #define GET_NEXT_SEQ(seq) ((seq + 1) & 0x0fff)
206 #define GET_NEXT_SEQ_FROM_SKB(skb) (GET_NEXT_SEQ(ar9170_get_seq(skb)))
208 #if (defined AR9170_QUEUE_DEBUG) || (defined AR9170_TXAGG_DEBUG)
209 static void ar9170_print_txheader(struct ar9170
*ar
, struct sk_buff
*skb
)
211 struct ar9170_tx_control
*txc
= (void *) skb
->data
;
212 struct ieee80211_tx_info
*txinfo
= IEEE80211_SKB_CB(skb
);
213 struct ar9170_tx_info
*arinfo
= (void *) txinfo
->rate_driver_data
;
214 struct ieee80211_hdr
*hdr
= (void *) txc
->frame_data
;
216 printk(KERN_DEBUG
"%s: => FRAME [skb:%p, q:%d, DA:[%pM] flags:%x s:%d "
217 "mac_ctrl:%04x, phy_ctrl:%08x, timeout:[%d ms]]\n",
218 wiphy_name(ar
->hw
->wiphy
), skb
, skb_get_queue_mapping(skb
),
219 ieee80211_get_DA(hdr
), arinfo
->flags
, ar9170_get_seq_h(hdr
),
220 le16_to_cpu(txc
->mac_control
), le32_to_cpu(txc
->phy_control
),
221 jiffies_to_msecs(arinfo
->timeout
- jiffies
));
224 static void __ar9170_dump_txqueue(struct ar9170
*ar
,
225 struct sk_buff_head
*queue
)
230 printk(KERN_DEBUG
"---[ cut here ]---\n");
231 printk(KERN_DEBUG
"%s: %d entries in queue.\n",
232 wiphy_name(ar
->hw
->wiphy
), skb_queue_len(queue
));
234 skb_queue_walk(queue
, skb
) {
235 printk(KERN_DEBUG
"index:%d => \n", i
++);
236 ar9170_print_txheader(ar
, skb
);
238 if (i
!= skb_queue_len(queue
))
239 printk(KERN_DEBUG
"WARNING: queue frame counter "
240 "mismatch %d != %d\n", skb_queue_len(queue
), i
);
241 printk(KERN_DEBUG
"---[ end ]---\n");
243 #endif /* AR9170_QUEUE_DEBUG || AR9170_TXAGG_DEBUG */
245 #ifdef AR9170_QUEUE_DEBUG
246 static void ar9170_dump_txqueue(struct ar9170
*ar
,
247 struct sk_buff_head
*queue
)
251 spin_lock_irqsave(&queue
->lock
, flags
);
252 __ar9170_dump_txqueue(ar
, queue
);
253 spin_unlock_irqrestore(&queue
->lock
, flags
);
255 #endif /* AR9170_QUEUE_DEBUG */
257 #ifdef AR9170_QUEUE_STOP_DEBUG
258 static void __ar9170_dump_txstats(struct ar9170
*ar
)
262 printk(KERN_DEBUG
"%s: QoS queue stats\n",
263 wiphy_name(ar
->hw
->wiphy
));
265 for (i
= 0; i
< __AR9170_NUM_TXQ
; i
++)
266 printk(KERN_DEBUG
"%s: queue:%d limit:%d len:%d waitack:%d "
267 " stopped:%d\n", wiphy_name(ar
->hw
->wiphy
), i
,
268 ar
->tx_stats
[i
].limit
, ar
->tx_stats
[i
].len
,
269 skb_queue_len(&ar
->tx_status
[i
]),
270 ieee80211_queue_stopped(ar
->hw
, i
));
272 #endif /* AR9170_QUEUE_STOP_DEBUG */
274 #ifdef AR9170_TXAGG_DEBUG
275 static void ar9170_dump_tx_status_ampdu(struct ar9170
*ar
)
279 spin_lock_irqsave(&ar
->tx_status_ampdu
.lock
, flags
);
280 printk(KERN_DEBUG
"%s: A-MPDU tx_status queue => \n",
281 wiphy_name(ar
->hw
->wiphy
));
282 __ar9170_dump_txqueue(ar
, &ar
->tx_status_ampdu
);
283 spin_unlock_irqrestore(&ar
->tx_status_ampdu
.lock
, flags
);
286 #endif /* AR9170_TXAGG_DEBUG */
288 /* caller must guarantee exclusive access for _bin_ queue. */
289 static void ar9170_recycle_expired(struct ar9170
*ar
,
290 struct sk_buff_head
*queue
,
291 struct sk_buff_head
*bin
)
293 struct sk_buff
*skb
, *old
= NULL
;
296 spin_lock_irqsave(&queue
->lock
, flags
);
297 while ((skb
= skb_peek(queue
))) {
298 struct ieee80211_tx_info
*txinfo
;
299 struct ar9170_tx_info
*arinfo
;
301 txinfo
= IEEE80211_SKB_CB(skb
);
302 arinfo
= (void *) txinfo
->rate_driver_data
;
304 if (time_is_before_jiffies(arinfo
->timeout
)) {
305 #ifdef AR9170_QUEUE_DEBUG
306 printk(KERN_DEBUG
"%s: [%ld > %ld] frame expired => "
307 "recycle \n", wiphy_name(ar
->hw
->wiphy
),
308 jiffies
, arinfo
->timeout
);
309 ar9170_print_txheader(ar
, skb
);
310 #endif /* AR9170_QUEUE_DEBUG */
311 __skb_unlink(skb
, queue
);
312 __skb_queue_tail(bin
, skb
);
317 if (unlikely(old
== skb
)) {
318 /* bail out - queue is shot. */
325 spin_unlock_irqrestore(&queue
->lock
, flags
);
328 static void ar9170_tx_status(struct ar9170
*ar
, struct sk_buff
*skb
,
331 struct ieee80211_tx_info
*txinfo
;
332 unsigned int retries
= 0;
334 txinfo
= IEEE80211_SKB_CB(skb
);
335 ieee80211_tx_info_clear_status(txinfo
);
338 case AR9170_TX_STATUS_RETRY
:
340 case AR9170_TX_STATUS_COMPLETE
:
341 txinfo
->flags
|= IEEE80211_TX_STAT_ACK
;
344 case AR9170_TX_STATUS_FAILED
:
345 retries
= ar
->hw
->conf
.long_frame_max_tx_count
;
349 printk(KERN_ERR
"%s: invalid tx_status response (%x).\n",
350 wiphy_name(ar
->hw
->wiphy
), tx_status
);
354 txinfo
->status
.rates
[0].count
= retries
+ 1;
355 skb_pull(skb
, sizeof(struct ar9170_tx_control
));
356 ieee80211_tx_status_irqsafe(ar
->hw
, skb
);
359 static void ar9170_tx_fake_ampdu_status(struct ar9170
*ar
)
361 struct sk_buff_head success
;
364 unsigned long queue_bitmap
= 0;
366 skb_queue_head_init(&success
);
368 while (skb_queue_len(&ar
->tx_status_ampdu
) > AR9170_NUM_TX_STATUS
)
369 __skb_queue_tail(&success
, skb_dequeue(&ar
->tx_status_ampdu
));
371 ar9170_recycle_expired(ar
, &ar
->tx_status_ampdu
, &success
);
373 #ifdef AR9170_TXAGG_DEBUG
374 printk(KERN_DEBUG
"%s: collected %d A-MPDU frames.\n",
375 wiphy_name(ar
->hw
->wiphy
), skb_queue_len(&success
));
376 __ar9170_dump_txqueue(ar
, &success
);
377 #endif /* AR9170_TXAGG_DEBUG */
379 while ((skb
= __skb_dequeue(&success
))) {
380 struct ieee80211_tx_info
*txinfo
;
382 queue_bitmap
|= BIT(skb_get_queue_mapping(skb
));
384 txinfo
= IEEE80211_SKB_CB(skb
);
385 ieee80211_tx_info_clear_status(txinfo
);
387 txinfo
->flags
|= IEEE80211_TX_STAT_ACK
;
388 txinfo
->status
.rates
[0].count
= 1;
390 skb_pull(skb
, sizeof(struct ar9170_tx_control
));
391 ieee80211_tx_status_irqsafe(ar
->hw
, skb
);
394 for_each_bit(i
, &queue_bitmap
, BITS_PER_BYTE
) {
395 #ifdef AR9170_QUEUE_STOP_DEBUG
396 printk(KERN_DEBUG
"%s: wake queue %d\n",
397 wiphy_name(ar
->hw
->wiphy
), i
);
398 __ar9170_dump_txstats(ar
);
399 #endif /* AR9170_QUEUE_STOP_DEBUG */
400 ieee80211_wake_queue(ar
->hw
, i
);
407 static void ar9170_tx_ampdu_callback(struct ar9170
*ar
, struct sk_buff
*skb
)
409 struct ieee80211_tx_info
*txinfo
= IEEE80211_SKB_CB(skb
);
410 struct ar9170_tx_info
*arinfo
= (void *) txinfo
->rate_driver_data
;
412 arinfo
->timeout
= jiffies
+
413 msecs_to_jiffies(AR9170_BA_TIMEOUT
);
415 skb_queue_tail(&ar
->tx_status_ampdu
, skb
);
416 ar9170_tx_fake_ampdu_status(ar
);
417 ar
->tx_ampdu_pending
--;
419 if (!list_empty(&ar
->tx_ampdu_list
) && !ar
->tx_ampdu_pending
)
423 void ar9170_tx_callback(struct ar9170
*ar
, struct sk_buff
*skb
)
425 struct ieee80211_tx_info
*info
= IEEE80211_SKB_CB(skb
);
426 struct ar9170_tx_info
*arinfo
= (void *) info
->rate_driver_data
;
427 unsigned int queue
= skb_get_queue_mapping(skb
);
430 spin_lock_irqsave(&ar
->tx_stats_lock
, flags
);
431 ar
->tx_stats
[queue
].len
--;
433 if (skb_queue_empty(&ar
->tx_pending
[queue
])) {
434 #ifdef AR9170_QUEUE_STOP_DEBUG
435 printk(KERN_DEBUG
"%s: wake queue %d\n",
436 wiphy_name(ar
->hw
->wiphy
), queue
);
437 __ar9170_dump_txstats(ar
);
438 #endif /* AR9170_QUEUE_STOP_DEBUG */
439 ieee80211_wake_queue(ar
->hw
, queue
);
441 spin_unlock_irqrestore(&ar
->tx_stats_lock
, flags
);
443 if (arinfo
->flags
& AR9170_TX_FLAG_BLOCK_ACK
) {
444 ar9170_tx_ampdu_callback(ar
, skb
);
445 } else if (arinfo
->flags
& AR9170_TX_FLAG_WAIT_FOR_ACK
) {
446 arinfo
->timeout
= jiffies
+
447 msecs_to_jiffies(AR9170_TX_TIMEOUT
);
449 skb_queue_tail(&ar
->tx_status
[queue
], skb
);
450 } else if (arinfo
->flags
& AR9170_TX_FLAG_NO_ACK
) {
451 ar9170_tx_status(ar
, skb
, AR9170_TX_STATUS_FAILED
);
453 #ifdef AR9170_QUEUE_DEBUG
454 printk(KERN_DEBUG
"%s: unsupported frame flags!\n",
455 wiphy_name(ar
->hw
->wiphy
));
456 ar9170_print_txheader(ar
, skb
);
457 #endif /* AR9170_QUEUE_DEBUG */
458 dev_kfree_skb_any(skb
);
461 if (!ar
->tx_stats
[queue
].len
&&
462 !skb_queue_empty(&ar
->tx_pending
[queue
])) {
467 static struct sk_buff
*ar9170_get_queued_skb(struct ar9170
*ar
,
469 struct sk_buff_head
*queue
,
476 * Unfortunately, the firmware does not tell to which (queued) frame
477 * this transmission status report belongs to.
479 * So we have to make risky guesses - with the scarce information
480 * the firmware provided (-> destination MAC, and phy_control) -
481 * and hope that we picked the right one...
484 spin_lock_irqsave(&queue
->lock
, flags
);
485 skb_queue_walk(queue
, skb
) {
486 struct ar9170_tx_control
*txc
= (void *) skb
->data
;
487 struct ieee80211_hdr
*hdr
= (void *) txc
->frame_data
;
490 if (mac
&& compare_ether_addr(ieee80211_get_DA(hdr
), mac
)) {
491 #ifdef AR9170_QUEUE_DEBUG
492 printk(KERN_DEBUG
"%s: skip frame => DA %pM != %pM\n",
493 wiphy_name(ar
->hw
->wiphy
), mac
,
494 ieee80211_get_DA(hdr
));
495 ar9170_print_txheader(ar
, skb
);
496 #endif /* AR9170_QUEUE_DEBUG */
500 r
= (le32_to_cpu(txc
->phy_control
) & AR9170_TX_PHY_MCS_MASK
) >>
501 AR9170_TX_PHY_MCS_SHIFT
;
503 if ((rate
!= AR9170_TX_INVALID_RATE
) && (r
!= rate
)) {
504 #ifdef AR9170_QUEUE_DEBUG
505 printk(KERN_DEBUG
"%s: skip frame => rate %d != %d\n",
506 wiphy_name(ar
->hw
->wiphy
), rate
, r
);
507 ar9170_print_txheader(ar
, skb
);
508 #endif /* AR9170_QUEUE_DEBUG */
512 __skb_unlink(skb
, queue
);
513 spin_unlock_irqrestore(&queue
->lock
, flags
);
517 #ifdef AR9170_QUEUE_DEBUG
518 printk(KERN_ERR
"%s: ESS:[%pM] does not have any "
519 "outstanding frames in queue.\n",
520 wiphy_name(ar
->hw
->wiphy
), mac
);
521 __ar9170_dump_txqueue(ar
, queue
);
522 #endif /* AR9170_QUEUE_DEBUG */
523 spin_unlock_irqrestore(&queue
->lock
, flags
);
528 static void ar9170_handle_block_ack(struct ar9170
*ar
, u16 count
, u16 r
)
531 struct ieee80211_tx_info
*txinfo
;
534 skb
= ar9170_get_queued_skb(ar
, NULL
, &ar
->tx_status_ampdu
, r
);
538 txinfo
= IEEE80211_SKB_CB(skb
);
539 ieee80211_tx_info_clear_status(txinfo
);
541 /* FIXME: maybe more ? */
542 txinfo
->status
.rates
[0].count
= 1;
544 skb_pull(skb
, sizeof(struct ar9170_tx_control
));
545 ieee80211_tx_status_irqsafe(ar
->hw
, skb
);
549 #ifdef AR9170_TXAGG_DEBUG
551 printk(KERN_DEBUG
"%s: got %d more failed mpdus, but no more "
552 "suitable frames left in tx_status queue.\n",
553 wiphy_name(ar
->hw
->wiphy
), count
);
555 ar9170_dump_tx_status_ampdu(ar
);
557 #endif /* AR9170_TXAGG_DEBUG */
561 * This worker tries to keeps an maintain tx_status queues.
562 * So we can guarantee that incoming tx_status reports are
563 * actually for a pending frame.
566 static void ar9170_tx_janitor(struct work_struct
*work
)
568 struct ar9170
*ar
= container_of(work
, struct ar9170
,
570 struct sk_buff_head waste
;
572 bool resched
= false;
574 if (unlikely(!IS_STARTED(ar
)))
577 skb_queue_head_init(&waste
);
579 for (i
= 0; i
< __AR9170_NUM_TXQ
; i
++) {
580 #ifdef AR9170_QUEUE_DEBUG
581 printk(KERN_DEBUG
"%s: garbage collector scans queue:%d\n",
582 wiphy_name(ar
->hw
->wiphy
), i
);
583 ar9170_dump_txqueue(ar
, &ar
->tx_pending
[i
]);
584 ar9170_dump_txqueue(ar
, &ar
->tx_status
[i
]);
585 #endif /* AR9170_QUEUE_DEBUG */
587 ar9170_recycle_expired(ar
, &ar
->tx_status
[i
], &waste
);
588 ar9170_recycle_expired(ar
, &ar
->tx_pending
[i
], &waste
);
589 skb_queue_purge(&waste
);
591 if (!skb_queue_empty(&ar
->tx_status
[i
]) ||
592 !skb_queue_empty(&ar
->tx_pending
[i
]))
596 ar9170_tx_fake_ampdu_status(ar
);
599 queue_delayed_work(ar
->hw
->workqueue
,
601 msecs_to_jiffies(AR9170_JANITOR_DELAY
));
604 void ar9170_handle_command_response(struct ar9170
*ar
, void *buf
, u32 len
)
606 struct ar9170_cmd_response
*cmd
= (void *) buf
;
608 if ((cmd
->type
& 0xc0) != 0xc0) {
609 ar
->callback_cmd(ar
, len
, buf
);
613 /* hardware event handlers */
617 * TX status notification:
618 * bytes: 0c c1 XX YY M1 M2 M3 M4 M5 M6 R4 R3 R2 R1 S2 S1
622 * M1-M6 is the MAC address
623 * R1-R4 is the transmit rate
624 * S1-S2 is the transmit status
628 u32 phy
= le32_to_cpu(cmd
->tx_status
.rate
);
629 u32 q
= (phy
& AR9170_TX_PHY_QOS_MASK
) >>
630 AR9170_TX_PHY_QOS_SHIFT
;
631 #ifdef AR9170_QUEUE_DEBUG
632 printk(KERN_DEBUG
"%s: recv tx_status for %pM, p:%08x, q:%d\n",
633 wiphy_name(ar
->hw
->wiphy
), cmd
->tx_status
.dst
, phy
, q
);
634 #endif /* AR9170_QUEUE_DEBUG */
636 skb
= ar9170_get_queued_skb(ar
, cmd
->tx_status
.dst
,
638 AR9170_TX_INVALID_RATE
);
642 ar9170_tx_status(ar
, skb
, le16_to_cpu(cmd
->tx_status
.status
));
650 if (ar
->vif
&& ar
->vif
->type
== NL80211_IFTYPE_AP
)
651 queue_work(ar
->hw
->workqueue
, &ar
->beacon_work
);
656 * (IBSS) beacon send notification
657 * bytes: 04 c2 XX YY B4 B3 B2 B1
661 * B1-B4 "should" be the number of send out beacons.
666 /* End of Atim Window */
670 /* BlockACK bitmap */
674 /* BlockACK events */
675 ar9170_handle_block_ack(ar
,
676 le16_to_cpu(cmd
->ba_fail_cnt
.failed
),
677 le16_to_cpu(cmd
->ba_fail_cnt
.rate
));
678 ar9170_tx_fake_ampdu_status(ar
);
682 /* Watchdog Interrupt */
686 /* retransmission issue / SIFS/EIFS collision ?! */
691 printk(KERN_DEBUG
"ar9170 FW: %.*s\n", len
- 4, (char *)buf
+ 4);
698 printk(KERN_DEBUG
"ar9170 FW: u8: %#.2x\n",
702 printk(KERN_DEBUG
"ar9170 FW: u8: %#.4x\n",
703 le16_to_cpup((__le16
*)((char *)buf
+ 4)));
706 printk(KERN_DEBUG
"ar9170 FW: u8: %#.8x\n",
707 le32_to_cpup((__le32
*)((char *)buf
+ 4)));
710 printk(KERN_DEBUG
"ar9170 FW: u8: %#.16lx\n",
711 (unsigned long)le64_to_cpup(
712 (__le64
*)((char *)buf
+ 4)));
717 print_hex_dump_bytes("ar9170 FW:", DUMP_PREFIX_NONE
,
718 (char *)buf
+ 4, len
- 4);
722 printk(KERN_INFO
"received unhandled event %x\n", cmd
->type
);
723 print_hex_dump_bytes("dump:", DUMP_PREFIX_NONE
, buf
, len
);
728 static void ar9170_rx_reset_rx_mpdu(struct ar9170
*ar
)
730 memset(&ar
->rx_mpdu
.plcp
, 0, sizeof(struct ar9170_rx_head
));
731 ar
->rx_mpdu
.has_plcp
= false;
734 int ar9170_nag_limiter(struct ar9170
*ar
)
739 * we expect all sorts of errors in promiscuous mode.
740 * don't bother with it, it's OK!
742 if (ar
->sniffer_enabled
)
746 * only go for frequent errors! The hardware tends to
747 * do some stupid thing once in a while under load, in
748 * noisy environments or just for fun!
750 if (time_before(jiffies
, ar
->bad_hw_nagger
) && net_ratelimit())
751 print_message
= true;
753 print_message
= false;
755 /* reset threshold for "once in a while" */
756 ar
->bad_hw_nagger
= jiffies
+ HZ
/ 4;
757 return print_message
;
760 static int ar9170_rx_mac_status(struct ar9170
*ar
,
761 struct ar9170_rx_head
*head
,
762 struct ar9170_rx_macstatus
*mac
,
763 struct ieee80211_rx_status
*status
)
767 BUILD_BUG_ON(sizeof(struct ar9170_rx_head
) != 12);
768 BUILD_BUG_ON(sizeof(struct ar9170_rx_macstatus
) != 4);
771 if (error
& AR9170_RX_ERROR_MMIC
) {
772 status
->flag
|= RX_FLAG_MMIC_ERROR
;
773 error
&= ~AR9170_RX_ERROR_MMIC
;
776 if (error
& AR9170_RX_ERROR_PLCP
) {
777 status
->flag
|= RX_FLAG_FAILED_PLCP_CRC
;
778 error
&= ~AR9170_RX_ERROR_PLCP
;
780 if (!(ar
->filter_state
& FIF_PLCPFAIL
))
784 if (error
& AR9170_RX_ERROR_FCS
) {
785 status
->flag
|= RX_FLAG_FAILED_FCS_CRC
;
786 error
&= ~AR9170_RX_ERROR_FCS
;
788 if (!(ar
->filter_state
& FIF_FCSFAIL
))
792 decrypt
= ar9170_get_decrypt_type(mac
);
793 if (!(decrypt
& AR9170_RX_ENC_SOFTWARE
) &&
794 decrypt
!= AR9170_ENC_ALG_NONE
)
795 status
->flag
|= RX_FLAG_DECRYPTED
;
797 /* ignore wrong RA errors */
798 error
&= ~AR9170_RX_ERROR_WRONG_RA
;
800 if (error
& AR9170_RX_ERROR_DECRYPT
) {
801 error
&= ~AR9170_RX_ERROR_DECRYPT
;
803 * Rx decryption is done in place,
804 * the original data is lost anyway.
810 /* drop any other error frames */
811 if (unlikely(error
)) {
812 /* TODO: update netdevice's RX dropped/errors statistics */
814 if (ar9170_nag_limiter(ar
))
815 printk(KERN_DEBUG
"%s: received frame with "
816 "suspicious error code (%#x).\n",
817 wiphy_name(ar
->hw
->wiphy
), error
);
822 status
->band
= ar
->channel
->band
;
823 status
->freq
= ar
->channel
->center_freq
;
825 switch (mac
->status
& AR9170_RX_STATUS_MODULATION_MASK
) {
826 case AR9170_RX_STATUS_MODULATION_CCK
:
827 if (mac
->status
& AR9170_RX_STATUS_SHORT_PREAMBLE
)
828 status
->flag
|= RX_FLAG_SHORTPRE
;
829 switch (head
->plcp
[0]) {
831 status
->rate_idx
= 0;
834 status
->rate_idx
= 1;
837 status
->rate_idx
= 2;
840 status
->rate_idx
= 3;
843 if (ar9170_nag_limiter(ar
))
844 printk(KERN_ERR
"%s: invalid plcp cck rate "
845 "(%x).\n", wiphy_name(ar
->hw
->wiphy
),
851 case AR9170_RX_STATUS_MODULATION_OFDM
:
852 switch (head
->plcp
[0] & 0xf) {
854 status
->rate_idx
= 0;
857 status
->rate_idx
= 1;
860 status
->rate_idx
= 2;
863 status
->rate_idx
= 3;
866 status
->rate_idx
= 4;
869 status
->rate_idx
= 5;
872 status
->rate_idx
= 6;
875 status
->rate_idx
= 7;
878 if (ar9170_nag_limiter(ar
))
879 printk(KERN_ERR
"%s: invalid plcp ofdm rate "
880 "(%x).\n", wiphy_name(ar
->hw
->wiphy
),
884 if (status
->band
== IEEE80211_BAND_2GHZ
)
885 status
->rate_idx
+= 4;
888 case AR9170_RX_STATUS_MODULATION_HT
:
889 if (head
->plcp
[3] & 0x80)
890 status
->flag
|= RX_FLAG_40MHZ
;
891 if (head
->plcp
[6] & 0x80)
892 status
->flag
|= RX_FLAG_SHORT_GI
;
894 status
->rate_idx
= clamp(0, 75, head
->plcp
[6] & 0x7f);
895 status
->flag
|= RX_FLAG_HT
;
898 case AR9170_RX_STATUS_MODULATION_DUPOFDM
:
900 if (ar9170_nag_limiter(ar
))
901 printk(KERN_ERR
"%s: invalid modulation\n",
902 wiphy_name(ar
->hw
->wiphy
));
909 static void ar9170_rx_phy_status(struct ar9170
*ar
,
910 struct ar9170_rx_phystatus
*phy
,
911 struct ieee80211_rx_status
*status
)
915 BUILD_BUG_ON(sizeof(struct ar9170_rx_phystatus
) != 20);
917 for (i
= 0; i
< 3; i
++)
918 if (phy
->rssi
[i
] != 0x80)
919 status
->antenna
|= BIT(i
);
921 /* post-process RSSI */
922 for (i
= 0; i
< 7; i
++)
923 if (phy
->rssi
[i
] & 0x80)
924 phy
->rssi
[i
] = ((phy
->rssi
[i
] & 0x7f) + 1) & 0x7f;
926 /* TODO: we could do something with phy_errors */
927 status
->signal
= ar
->noise
[0] + phy
->rssi_combined
;
928 status
->noise
= ar
->noise
[0];
931 static struct sk_buff
*ar9170_rx_copy_data(u8
*buf
, int len
)
935 struct ieee80211_hdr
*hdr
= (void *) buf
;
937 if (ieee80211_is_data_qos(hdr
->frame_control
)) {
938 u8
*qc
= ieee80211_get_qos_ctl(hdr
);
939 reserved
+= NET_IP_ALIGN
;
941 if (*qc
& IEEE80211_QOS_CONTROL_A_MSDU_PRESENT
)
942 reserved
+= NET_IP_ALIGN
;
945 if (ieee80211_has_a4(hdr
->frame_control
))
946 reserved
+= NET_IP_ALIGN
;
948 reserved
= 32 + (reserved
& NET_IP_ALIGN
);
950 skb
= dev_alloc_skb(len
+ reserved
);
952 skb_reserve(skb
, reserved
);
953 memcpy(skb_put(skb
, len
), buf
, len
);
960 * If the frame alignment is right (or the kernel has
961 * CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS), and there
962 * is only a single MPDU in the USB frame, then we could
963 * submit to mac80211 the SKB directly. However, since
964 * there may be multiple packets in one SKB in stream
965 * mode, and we need to observe the proper ordering,
966 * this is non-trivial.
969 static void ar9170_handle_mpdu(struct ar9170
*ar
, u8
*buf
, int len
)
971 struct ar9170_rx_head
*head
;
972 struct ar9170_rx_macstatus
*mac
;
973 struct ar9170_rx_phystatus
*phy
= NULL
;
974 struct ieee80211_rx_status status
;
978 if (unlikely(!IS_STARTED(ar
) || len
< (sizeof(*mac
))))
982 mpdu_len
= len
- sizeof(*mac
);
984 mac
= (void *)(buf
+ mpdu_len
);
985 if (unlikely(mac
->error
& AR9170_RX_ERROR_FATAL
)) {
986 /* this frame is too damaged and can't be used - drop it */
991 switch (mac
->status
& AR9170_RX_STATUS_MPDU_MASK
) {
992 case AR9170_RX_STATUS_MPDU_FIRST
:
993 /* first mpdu packet has the plcp header */
994 if (likely(mpdu_len
>= sizeof(struct ar9170_rx_head
))) {
996 memcpy(&ar
->rx_mpdu
.plcp
, (void *) buf
,
997 sizeof(struct ar9170_rx_head
));
999 mpdu_len
-= sizeof(struct ar9170_rx_head
);
1000 buf
+= sizeof(struct ar9170_rx_head
);
1001 ar
->rx_mpdu
.has_plcp
= true;
1003 if (ar9170_nag_limiter(ar
))
1004 printk(KERN_ERR
"%s: plcp info is clipped.\n",
1005 wiphy_name(ar
->hw
->wiphy
));
1010 case AR9170_RX_STATUS_MPDU_LAST
:
1011 /* last mpdu has a extra tail with phy status information */
1013 if (likely(mpdu_len
>= sizeof(struct ar9170_rx_phystatus
))) {
1014 mpdu_len
-= sizeof(struct ar9170_rx_phystatus
);
1015 phy
= (void *)(buf
+ mpdu_len
);
1017 if (ar9170_nag_limiter(ar
))
1018 printk(KERN_ERR
"%s: frame tail is clipped.\n",
1019 wiphy_name(ar
->hw
->wiphy
));
1023 case AR9170_RX_STATUS_MPDU_MIDDLE
:
1024 /* middle mpdus are just data */
1025 if (unlikely(!ar
->rx_mpdu
.has_plcp
)) {
1026 if (!ar9170_nag_limiter(ar
))
1029 printk(KERN_ERR
"%s: rx stream did not start "
1030 "with a first_mpdu frame tag.\n",
1031 wiphy_name(ar
->hw
->wiphy
));
1036 head
= &ar
->rx_mpdu
.plcp
;
1039 case AR9170_RX_STATUS_MPDU_SINGLE
:
1040 /* single mpdu - has plcp (head) and phy status (tail) */
1041 head
= (void *) buf
;
1043 mpdu_len
-= sizeof(struct ar9170_rx_head
);
1044 mpdu_len
-= sizeof(struct ar9170_rx_phystatus
);
1046 buf
+= sizeof(struct ar9170_rx_head
);
1047 phy
= (void *)(buf
+ mpdu_len
);
1055 if (unlikely(mpdu_len
< FCS_LEN
))
1058 memset(&status
, 0, sizeof(status
));
1059 if (unlikely(ar9170_rx_mac_status(ar
, head
, mac
, &status
)))
1063 ar9170_rx_phy_status(ar
, phy
, &status
);
1065 skb
= ar9170_rx_copy_data(buf
, mpdu_len
);
1067 memcpy(IEEE80211_SKB_RXCB(skb
), &status
, sizeof(status
));
1068 ieee80211_rx_irqsafe(ar
->hw
, skb
);
1072 void ar9170_rx(struct ar9170
*ar
, struct sk_buff
*skb
)
1074 unsigned int i
, tlen
, resplen
, wlen
= 0, clen
= 0;
1081 clen
= tbuf
[1] << 8 | tbuf
[0];
1082 wlen
= ALIGN(clen
, 4);
1084 /* check if this is stream has a valid tag.*/
1085 if (tbuf
[2] != 0 || tbuf
[3] != 0x4e) {
1087 * TODO: handle the highly unlikely event that the
1088 * corrupted stream has the TAG at the right position.
1091 /* check if the frame can be repaired. */
1092 if (!ar
->rx_failover_missing
) {
1093 /* this is no "short read". */
1094 if (ar9170_nag_limiter(ar
)) {
1095 printk(KERN_ERR
"%s: missing tag!\n",
1096 wiphy_name(ar
->hw
->wiphy
));
1102 if (ar
->rx_failover_missing
> tlen
) {
1103 if (ar9170_nag_limiter(ar
)) {
1104 printk(KERN_ERR
"%s: possible multi "
1105 "stream corruption!\n",
1106 wiphy_name(ar
->hw
->wiphy
));
1112 memcpy(skb_put(ar
->rx_failover
, tlen
), tbuf
, tlen
);
1113 ar
->rx_failover_missing
-= tlen
;
1115 if (ar
->rx_failover_missing
<= 0) {
1117 * nested ar9170_rx call!
1118 * termination is guranteed, even when the
1119 * combined frame also have a element with
1123 ar
->rx_failover_missing
= 0;
1124 ar9170_rx(ar
, ar
->rx_failover
);
1126 skb_reset_tail_pointer(ar
->rx_failover
);
1127 skb_trim(ar
->rx_failover
, 0);
1133 /* check if stream is clipped */
1134 if (wlen
> tlen
- 4) {
1135 if (ar
->rx_failover_missing
) {
1136 /* TODO: handle double stream corruption. */
1137 if (ar9170_nag_limiter(ar
)) {
1138 printk(KERN_ERR
"%s: double rx stream "
1140 wiphy_name(ar
->hw
->wiphy
));
1147 * save incomplete data set.
1148 * the firmware will resend the missing bits when
1149 * the rx - descriptor comes round again.
1152 memcpy(skb_put(ar
->rx_failover
, tlen
), tbuf
, tlen
);
1153 ar
->rx_failover_missing
= clen
- tlen
;
1163 /* weird thing, but this is the same in the original driver */
1164 while (resplen
> 2 && i
< 12 &&
1165 respbuf
[0] == 0xff && respbuf
[1] == 0xff) {
1174 /* found the 6 * 0xffff marker? */
1176 ar9170_handle_command_response(ar
, respbuf
, resplen
);
1178 ar9170_handle_mpdu(ar
, respbuf
, clen
);
1182 if (net_ratelimit())
1183 printk(KERN_ERR
"%s: %d bytes of unprocessed "
1184 "data left in rx stream!\n",
1185 wiphy_name(ar
->hw
->wiphy
), tlen
);
1193 printk(KERN_ERR
"%s: damaged RX stream data [want:%d, "
1194 "data:%d, rx:%d, pending:%d ]\n",
1195 wiphy_name(ar
->hw
->wiphy
), clen
, wlen
, tlen
,
1196 ar
->rx_failover_missing
);
1198 if (ar
->rx_failover_missing
)
1199 print_hex_dump_bytes("rxbuf:", DUMP_PREFIX_OFFSET
,
1200 ar
->rx_failover
->data
,
1201 ar
->rx_failover
->len
);
1203 print_hex_dump_bytes("stream:", DUMP_PREFIX_OFFSET
,
1204 skb
->data
, skb
->len
);
1206 printk(KERN_ERR
"%s: please check your hardware and cables, if "
1207 "you see this message frequently.\n",
1208 wiphy_name(ar
->hw
->wiphy
));
1211 if (ar
->rx_failover_missing
) {
1212 skb_reset_tail_pointer(ar
->rx_failover
);
1213 skb_trim(ar
->rx_failover
, 0);
1214 ar
->rx_failover_missing
= 0;
1218 #define AR9170_FILL_QUEUE(queue, ai_fs, cwmin, cwmax, _txop) \
1220 queue.aifs = ai_fs; \
1221 queue.cw_min = cwmin; \
1222 queue.cw_max = cwmax; \
1223 queue.txop = _txop; \
1226 static int ar9170_op_start(struct ieee80211_hw
*hw
)
1228 struct ar9170
*ar
= hw
->priv
;
1231 mutex_lock(&ar
->mutex
);
1233 ar
->filter_changed
= 0;
1235 /* reinitialize queues statistics */
1236 memset(&ar
->tx_stats
, 0, sizeof(ar
->tx_stats
));
1237 for (i
= 0; i
< __AR9170_NUM_TXQ
; i
++)
1238 ar
->tx_stats
[i
].limit
= AR9170_TXQ_DEPTH
;
1240 /* reset QoS defaults */
1241 AR9170_FILL_QUEUE(ar
->edcf
[0], 3, 15, 1023, 0); /* BEST EFFORT*/
1242 AR9170_FILL_QUEUE(ar
->edcf
[1], 7, 15, 1023, 0); /* BACKGROUND */
1243 AR9170_FILL_QUEUE(ar
->edcf
[2], 2, 7, 15, 94); /* VIDEO */
1244 AR9170_FILL_QUEUE(ar
->edcf
[3], 2, 3, 7, 47); /* VOICE */
1245 AR9170_FILL_QUEUE(ar
->edcf
[4], 2, 3, 7, 0); /* SPECIAL */
1247 /* set sane AMPDU defaults */
1248 ar
->global_ampdu_density
= 6;
1249 ar
->global_ampdu_factor
= 3;
1251 ar
->bad_hw_nagger
= jiffies
;
1257 err
= ar9170_init_mac(ar
);
1261 err
= ar9170_set_qos(ar
);
1265 err
= ar9170_init_phy(ar
, IEEE80211_BAND_2GHZ
);
1269 err
= ar9170_init_rf(ar
);
1274 err
= ar9170_write_reg(ar
, 0x1c3d30, 0x100);
1278 ar
->state
= AR9170_STARTED
;
1281 mutex_unlock(&ar
->mutex
);
1285 static void ar9170_op_stop(struct ieee80211_hw
*hw
)
1287 struct ar9170
*ar
= hw
->priv
;
1291 ar
->state
= AR9170_IDLE
;
1293 flush_workqueue(ar
->hw
->workqueue
);
1295 cancel_delayed_work_sync(&ar
->tx_janitor
);
1296 cancel_delayed_work_sync(&ar
->led_work
);
1297 cancel_work_sync(&ar
->filter_config_work
);
1298 cancel_work_sync(&ar
->beacon_work
);
1299 mutex_lock(&ar
->mutex
);
1301 if (IS_ACCEPTING_CMD(ar
)) {
1302 ar9170_set_leds_state(ar
, 0);
1305 ar9170_write_reg(ar
, 0x1c3d30, 0);
1309 for (i
= 0; i
< __AR9170_NUM_TXQ
; i
++) {
1310 skb_queue_purge(&ar
->tx_pending
[i
]);
1311 skb_queue_purge(&ar
->tx_status
[i
]);
1313 skb_queue_purge(&ar
->tx_status_ampdu
);
1315 mutex_unlock(&ar
->mutex
);
1318 static void ar9170_tx_indicate_immba(struct ar9170
*ar
, struct sk_buff
*skb
)
1320 struct ar9170_tx_control
*txc
= (void *) skb
->data
;
1322 txc
->mac_control
|= cpu_to_le16(AR9170_TX_MAC_IMM_AMPDU
);
1325 static void ar9170_tx_copy_phy(struct ar9170
*ar
, struct sk_buff
*dst
,
1326 struct sk_buff
*src
)
1328 struct ar9170_tx_control
*dst_txc
, *src_txc
;
1329 struct ieee80211_tx_info
*dst_info
, *src_info
;
1330 struct ar9170_tx_info
*dst_arinfo
, *src_arinfo
;
1332 src_txc
= (void *) src
->data
;
1333 src_info
= IEEE80211_SKB_CB(src
);
1334 src_arinfo
= (void *) src_info
->rate_driver_data
;
1336 dst_txc
= (void *) dst
->data
;
1337 dst_info
= IEEE80211_SKB_CB(dst
);
1338 dst_arinfo
= (void *) dst_info
->rate_driver_data
;
1340 dst_txc
->phy_control
= src_txc
->phy_control
;
1342 /* same MCS for the whole aggregate */
1343 memcpy(dst_info
->driver_rates
, src_info
->driver_rates
,
1344 sizeof(dst_info
->driver_rates
));
1347 static int ar9170_tx_prepare(struct ar9170
*ar
, struct sk_buff
*skb
)
1349 struct ieee80211_hdr
*hdr
;
1350 struct ar9170_tx_control
*txc
;
1351 struct ieee80211_tx_info
*info
;
1352 struct ieee80211_tx_rate
*txrate
;
1353 struct ar9170_tx_info
*arinfo
;
1354 unsigned int queue
= skb_get_queue_mapping(skb
);
1358 BUILD_BUG_ON(sizeof(*arinfo
) > sizeof(info
->rate_driver_data
));
1360 hdr
= (void *)skb
->data
;
1361 info
= IEEE80211_SKB_CB(skb
);
1364 txc
= (void *)skb_push(skb
, sizeof(*txc
));
1366 if (info
->control
.hw_key
) {
1367 icv
= info
->control
.hw_key
->icv_len
;
1369 switch (info
->control
.hw_key
->alg
) {
1371 keytype
= AR9170_TX_MAC_ENCR_RC4
;
1374 keytype
= AR9170_TX_MAC_ENCR_RC4
;
1377 keytype
= AR9170_TX_MAC_ENCR_AES
;
1386 txc
->length
= cpu_to_le16(len
+ icv
+ 4);
1388 txc
->mac_control
= cpu_to_le16(AR9170_TX_MAC_HW_DURATION
|
1389 AR9170_TX_MAC_BACKOFF
);
1390 txc
->mac_control
|= cpu_to_le16(ar9170_qos_hwmap
[queue
] <<
1391 AR9170_TX_MAC_QOS_SHIFT
);
1392 txc
->mac_control
|= cpu_to_le16(keytype
);
1393 txc
->phy_control
= cpu_to_le32(0);
1395 if (info
->flags
& IEEE80211_TX_CTL_NO_ACK
)
1396 txc
->mac_control
|= cpu_to_le16(AR9170_TX_MAC_NO_ACK
);
1398 txrate
= &info
->control
.rates
[0];
1399 if (txrate
->flags
& IEEE80211_TX_RC_USE_CTS_PROTECT
)
1400 txc
->mac_control
|= cpu_to_le16(AR9170_TX_MAC_PROT_CTS
);
1401 else if (txrate
->flags
& IEEE80211_TX_RC_USE_RTS_CTS
)
1402 txc
->mac_control
|= cpu_to_le16(AR9170_TX_MAC_PROT_RTS
);
1404 arinfo
= (void *)info
->rate_driver_data
;
1405 arinfo
->timeout
= jiffies
+ msecs_to_jiffies(AR9170_QUEUE_TIMEOUT
);
1407 if (!(info
->flags
& IEEE80211_TX_CTL_NO_ACK
) &&
1408 (is_valid_ether_addr(ieee80211_get_DA(hdr
)))) {
1409 if (info
->flags
& IEEE80211_TX_CTL_AMPDU
) {
1410 if (unlikely(!info
->control
.sta
))
1413 txc
->mac_control
|= cpu_to_le16(AR9170_TX_MAC_AGGR
);
1414 arinfo
->flags
= AR9170_TX_FLAG_BLOCK_ACK
;
1419 txc
->mac_control
|= cpu_to_le16(AR9170_TX_MAC_RATE_PROBE
);
1422 * Putting the QoS queue bits into an unexplored territory is
1423 * certainly not elegant.
1425 * In my defense: This idea provides a reasonable way to
1426 * smuggle valuable information to the tx_status callback.
1427 * Also, the idea behind this bit-abuse came straight from
1428 * the original driver code.
1432 cpu_to_le32(queue
<< AR9170_TX_PHY_QOS_SHIFT
);
1433 arinfo
->flags
= AR9170_TX_FLAG_WAIT_FOR_ACK
;
1435 arinfo
->flags
= AR9170_TX_FLAG_NO_ACK
;
1442 skb_pull(skb
, sizeof(*txc
));
1446 static void ar9170_tx_prepare_phy(struct ar9170
*ar
, struct sk_buff
*skb
)
1448 struct ar9170_tx_control
*txc
;
1449 struct ieee80211_tx_info
*info
;
1450 struct ieee80211_rate
*rate
= NULL
;
1451 struct ieee80211_tx_rate
*txrate
;
1454 txc
= (void *) skb
->data
;
1455 info
= IEEE80211_SKB_CB(skb
);
1456 txrate
= &info
->control
.rates
[0];
1458 if (txrate
->flags
& IEEE80211_TX_RC_GREEN_FIELD
)
1459 txc
->phy_control
|= cpu_to_le32(AR9170_TX_PHY_GREENFIELD
);
1461 if (txrate
->flags
& IEEE80211_TX_RC_USE_SHORT_PREAMBLE
)
1462 txc
->phy_control
|= cpu_to_le32(AR9170_TX_PHY_SHORT_PREAMBLE
);
1464 if (txrate
->flags
& IEEE80211_TX_RC_40_MHZ_WIDTH
)
1465 txc
->phy_control
|= cpu_to_le32(AR9170_TX_PHY_BW_40MHZ
);
1466 /* this works because 40 MHz is 2 and dup is 3 */
1467 if (txrate
->flags
& IEEE80211_TX_RC_DUP_DATA
)
1468 txc
->phy_control
|= cpu_to_le32(AR9170_TX_PHY_BW_40MHZ_DUP
);
1470 if (txrate
->flags
& IEEE80211_TX_RC_SHORT_GI
)
1471 txc
->phy_control
|= cpu_to_le32(AR9170_TX_PHY_SHORT_GI
);
1473 if (txrate
->flags
& IEEE80211_TX_RC_MCS
) {
1474 u32 r
= txrate
->idx
;
1477 /* heavy clip control */
1478 txc
->phy_control
|= cpu_to_le32((r
& 0x7) << 7);
1480 r
<<= AR9170_TX_PHY_MCS_SHIFT
;
1481 BUG_ON(r
& ~AR9170_TX_PHY_MCS_MASK
);
1483 txc
->phy_control
|= cpu_to_le32(r
& AR9170_TX_PHY_MCS_MASK
);
1484 txc
->phy_control
|= cpu_to_le32(AR9170_TX_PHY_MOD_HT
);
1486 if (txrate
->flags
& IEEE80211_TX_RC_40_MHZ_WIDTH
) {
1487 if (info
->band
== IEEE80211_BAND_5GHZ
)
1488 txpower
= ar
->power_5G_ht40
;
1490 txpower
= ar
->power_2G_ht40
;
1492 if (info
->band
== IEEE80211_BAND_5GHZ
)
1493 txpower
= ar
->power_5G_ht20
;
1495 txpower
= ar
->power_2G_ht20
;
1498 power
= txpower
[(txrate
->idx
) & 7];
1503 u8 idx
= txrate
->idx
;
1505 if (info
->band
!= IEEE80211_BAND_2GHZ
) {
1507 txpower
= ar
->power_5G_leg
;
1508 mod
= AR9170_TX_PHY_MOD_OFDM
;
1511 txpower
= ar
->power_2G_cck
;
1512 mod
= AR9170_TX_PHY_MOD_CCK
;
1514 mod
= AR9170_TX_PHY_MOD_OFDM
;
1515 txpower
= ar
->power_2G_ofdm
;
1519 rate
= &__ar9170_ratetable
[idx
];
1521 phyrate
= rate
->hw_value
& 0xF;
1522 power
= txpower
[(rate
->hw_value
& 0x30) >> 4];
1523 phyrate
<<= AR9170_TX_PHY_MCS_SHIFT
;
1525 txc
->phy_control
|= cpu_to_le32(mod
);
1526 txc
->phy_control
|= cpu_to_le32(phyrate
);
1529 power
<<= AR9170_TX_PHY_TX_PWR_SHIFT
;
1530 power
&= AR9170_TX_PHY_TX_PWR_MASK
;
1531 txc
->phy_control
|= cpu_to_le32(power
);
1534 if (ar
->eeprom
.tx_mask
== 1) {
1535 chains
= AR9170_TX_PHY_TXCHAIN_1
;
1537 chains
= AR9170_TX_PHY_TXCHAIN_2
;
1539 /* >= 36M legacy OFDM - use only one chain */
1540 if (rate
&& rate
->bitrate
>= 360)
1541 chains
= AR9170_TX_PHY_TXCHAIN_1
;
1543 txc
->phy_control
|= cpu_to_le32(chains
<< AR9170_TX_PHY_TXCHAIN_SHIFT
);
1546 static bool ar9170_tx_ampdu(struct ar9170
*ar
)
1548 struct sk_buff_head agg
;
1549 struct ar9170_sta_tid
*tid_info
= NULL
, *tmp
;
1550 struct sk_buff
*skb
, *first
= NULL
;
1551 unsigned long flags
, f2
;
1553 u16 seq
, queue
, tmpssn
;
1556 skb_queue_head_init(&agg
);
1558 spin_lock_irqsave(&ar
->tx_ampdu_list_lock
, flags
);
1559 if (list_empty(&ar
->tx_ampdu_list
)) {
1560 #ifdef AR9170_TXAGG_DEBUG
1561 printk(KERN_DEBUG
"%s: aggregation list is empty.\n",
1562 wiphy_name(ar
->hw
->wiphy
));
1563 #endif /* AR9170_TXAGG_DEBUG */
1567 list_for_each_entry_safe(tid_info
, tmp
, &ar
->tx_ampdu_list
, list
) {
1568 if (tid_info
->state
!= AR9170_TID_STATE_COMPLETE
) {
1569 #ifdef AR9170_TXAGG_DEBUG
1570 printk(KERN_DEBUG
"%s: dangling aggregation entry!\n",
1571 wiphy_name(ar
->hw
->wiphy
));
1572 #endif /* AR9170_TXAGG_DEBUG */
1577 #ifdef AR9170_TXAGG_DEBUG
1578 printk(KERN_DEBUG
"%s: enough frames aggregated.\n",
1579 wiphy_name(ar
->hw
->wiphy
));
1580 #endif /* AR9170_TXAGG_DEBUG */
1584 queue
= TID_TO_WME_AC(tid_info
->tid
);
1586 if (skb_queue_len(&ar
->tx_pending
[queue
]) >=
1587 AR9170_NUM_TX_AGG_MAX
) {
1588 #ifdef AR9170_TXAGG_DEBUG
1589 printk(KERN_DEBUG
"%s: queue %d full.\n",
1590 wiphy_name(ar
->hw
->wiphy
), queue
);
1591 #endif /* AR9170_TXAGG_DEBUG */
1595 list_del_init(&tid_info
->list
);
1597 spin_lock_irqsave(&tid_info
->queue
.lock
, f2
);
1598 tmpssn
= seq
= tid_info
->ssn
;
1599 first
= skb_peek(&tid_info
->queue
);
1602 tmpssn
= ar9170_get_seq(first
);
1604 if (unlikely(tmpssn
!= seq
)) {
1605 #ifdef AR9170_TXAGG_DEBUG
1606 printk(KERN_DEBUG
"%s: ssn mismatch [%d != %d]\n.",
1607 wiphy_name(ar
->hw
->wiphy
), seq
, tmpssn
);
1608 #endif /* AR9170_TXAGG_DEBUG */
1609 tid_info
->ssn
= tmpssn
;
1612 #ifdef AR9170_TXAGG_DEBUG
1613 printk(KERN_DEBUG
"%s: generate A-MPDU for tid:%d ssn:%d with "
1614 "%d queued frames.\n", wiphy_name(ar
->hw
->wiphy
),
1615 tid_info
->tid
, tid_info
->ssn
,
1616 skb_queue_len(&tid_info
->queue
));
1617 __ar9170_dump_txqueue(ar
, &tid_info
->queue
);
1618 #endif /* AR9170_TXAGG_DEBUG */
1620 while ((skb
= skb_peek(&tid_info
->queue
))) {
1621 if (unlikely(ar9170_get_seq(skb
) != seq
))
1624 __skb_unlink(skb
, &tid_info
->queue
);
1625 tid_info
->ssn
= seq
= GET_NEXT_SEQ(seq
);
1627 if (unlikely(skb_get_queue_mapping(skb
) != queue
)) {
1628 #ifdef AR9170_TXAGG_DEBUG
1629 printk(KERN_DEBUG
"%s: tid:%d(q:%d) queue:%d "
1630 "!match.\n", wiphy_name(ar
->hw
->wiphy
),
1632 TID_TO_WME_AC(tid_info
->tid
),
1633 skb_get_queue_mapping(skb
));
1634 #endif /* AR9170_TXAGG_DEBUG */
1635 dev_kfree_skb_any(skb
);
1639 if (unlikely(first
== skb
)) {
1640 ar9170_tx_prepare_phy(ar
, skb
);
1641 __skb_queue_tail(&agg
, skb
);
1644 ar9170_tx_copy_phy(ar
, skb
, first
);
1645 __skb_queue_tail(&agg
, skb
);
1648 if (unlikely(skb_queue_len(&agg
) ==
1649 AR9170_NUM_TX_AGG_MAX
))
1653 if (skb_queue_empty(&tid_info
->queue
))
1654 tid_info
->active
= false;
1656 list_add_tail(&tid_info
->list
,
1657 &ar
->tx_ampdu_list
);
1659 spin_unlock_irqrestore(&tid_info
->queue
.lock
, f2
);
1661 if (unlikely(skb_queue_empty(&agg
))) {
1662 #ifdef AR9170_TXAGG_DEBUG
1663 printk(KERN_DEBUG
"%s: queued empty list!\n",
1664 wiphy_name(ar
->hw
->wiphy
));
1665 #endif /* AR9170_TXAGG_DEBUG */
1670 * tell the FW/HW that this is the last frame,
1671 * that way it will wait for the immediate block ack.
1673 if (likely(skb_peek_tail(&agg
)))
1674 ar9170_tx_indicate_immba(ar
, skb_peek_tail(&agg
));
1676 #ifdef AR9170_TXAGG_DEBUG
1677 printk(KERN_DEBUG
"%s: generated A-MPDU looks like this:\n",
1678 wiphy_name(ar
->hw
->wiphy
));
1679 __ar9170_dump_txqueue(ar
, &agg
);
1680 #endif /* AR9170_TXAGG_DEBUG */
1682 spin_unlock_irqrestore(&ar
->tx_ampdu_list_lock
, flags
);
1684 spin_lock_irqsave(&ar
->tx_pending
[queue
].lock
, flags
);
1685 skb_queue_splice_tail_init(&agg
, &ar
->tx_pending
[queue
]);
1686 spin_unlock_irqrestore(&ar
->tx_pending
[queue
].lock
, flags
);
1689 spin_lock_irqsave(&ar
->tx_ampdu_list_lock
, flags
);
1693 spin_unlock_irqrestore(&ar
->tx_ampdu_list_lock
, flags
);
1694 __skb_queue_purge(&agg
);
1699 static void ar9170_tx(struct ar9170
*ar
)
1701 struct sk_buff
*skb
;
1702 unsigned long flags
;
1703 struct ieee80211_tx_info
*info
;
1704 struct ar9170_tx_info
*arinfo
;
1705 unsigned int i
, frames
, frames_failed
, remaining_space
;
1707 bool schedule_garbagecollector
= false;
1709 BUILD_BUG_ON(sizeof(*arinfo
) > sizeof(info
->rate_driver_data
));
1711 if (unlikely(!IS_STARTED(ar
)))
1714 remaining_space
= AR9170_TX_MAX_PENDING
;
1716 for (i
= 0; i
< __AR9170_NUM_TXQ
; i
++) {
1717 spin_lock_irqsave(&ar
->tx_stats_lock
, flags
);
1718 if (ar
->tx_stats
[i
].len
>= ar
->tx_stats
[i
].limit
) {
1719 #ifdef AR9170_QUEUE_DEBUG
1720 printk(KERN_DEBUG
"%s: queue %d full\n",
1721 wiphy_name(ar
->hw
->wiphy
), i
);
1723 printk(KERN_DEBUG
"%s: stuck frames: ===> \n",
1724 wiphy_name(ar
->hw
->wiphy
));
1725 ar9170_dump_txqueue(ar
, &ar
->tx_pending
[i
]);
1726 ar9170_dump_txqueue(ar
, &ar
->tx_status
[i
]);
1727 #endif /* AR9170_QUEUE_DEBUG */
1729 #ifdef AR9170_QUEUE_STOP_DEBUG
1730 printk(KERN_DEBUG
"%s: stop queue %d\n",
1731 wiphy_name(ar
->hw
->wiphy
), i
);
1732 __ar9170_dump_txstats(ar
);
1733 #endif /* AR9170_QUEUE_STOP_DEBUG */
1734 ieee80211_stop_queue(ar
->hw
, i
);
1735 spin_unlock_irqrestore(&ar
->tx_stats_lock
, flags
);
1739 frames
= min(ar
->tx_stats
[i
].limit
- ar
->tx_stats
[i
].len
,
1740 skb_queue_len(&ar
->tx_pending
[i
]));
1742 if (remaining_space
< frames
) {
1743 #ifdef AR9170_QUEUE_DEBUG
1744 printk(KERN_DEBUG
"%s: tx quota reached queue:%d, "
1745 "remaining slots:%d, needed:%d\n",
1746 wiphy_name(ar
->hw
->wiphy
), i
, remaining_space
,
1748 #endif /* AR9170_QUEUE_DEBUG */
1749 frames
= remaining_space
;
1752 ar
->tx_stats
[i
].len
+= frames
;
1753 ar
->tx_stats
[i
].count
+= frames
;
1754 spin_unlock_irqrestore(&ar
->tx_stats_lock
, flags
);
1761 skb
= skb_dequeue(&ar
->tx_pending
[i
]);
1762 if (unlikely(!skb
)) {
1763 frames_failed
+= frames
;
1768 info
= IEEE80211_SKB_CB(skb
);
1769 arinfo
= (void *) info
->rate_driver_data
;
1771 /* TODO: cancel stuck frames */
1772 arinfo
->timeout
= jiffies
+
1773 msecs_to_jiffies(AR9170_TX_TIMEOUT
);
1775 if (arinfo
->flags
== AR9170_TX_FLAG_BLOCK_ACK
)
1776 ar
->tx_ampdu_pending
++;
1778 #ifdef AR9170_QUEUE_DEBUG
1779 printk(KERN_DEBUG
"%s: send frame q:%d =>\n",
1780 wiphy_name(ar
->hw
->wiphy
), i
);
1781 ar9170_print_txheader(ar
, skb
);
1782 #endif /* AR9170_QUEUE_DEBUG */
1784 err
= ar
->tx(ar
, skb
);
1785 if (unlikely(err
)) {
1786 if (arinfo
->flags
== AR9170_TX_FLAG_BLOCK_ACK
)
1787 ar
->tx_ampdu_pending
--;
1790 dev_kfree_skb_any(skb
);
1793 schedule_garbagecollector
= true;
1799 #ifdef AR9170_QUEUE_DEBUG
1800 printk(KERN_DEBUG
"%s: ar9170_tx report for queue %d\n",
1801 wiphy_name(ar
->hw
->wiphy
), i
);
1803 printk(KERN_DEBUG
"%s: unprocessed pending frames left:\n",
1804 wiphy_name(ar
->hw
->wiphy
));
1805 ar9170_dump_txqueue(ar
, &ar
->tx_pending
[i
]);
1806 #endif /* AR9170_QUEUE_DEBUG */
1808 if (unlikely(frames_failed
)) {
1809 #ifdef AR9170_QUEUE_DEBUG
1810 printk(KERN_DEBUG
"%s: frames failed %d =>\n",
1811 wiphy_name(ar
->hw
->wiphy
), frames_failed
);
1812 #endif /* AR9170_QUEUE_DEBUG */
1814 spin_lock_irqsave(&ar
->tx_stats_lock
, flags
);
1815 ar
->tx_stats
[i
].len
-= frames_failed
;
1816 ar
->tx_stats
[i
].count
-= frames_failed
;
1817 #ifdef AR9170_QUEUE_STOP_DEBUG
1818 printk(KERN_DEBUG
"%s: wake queue %d\n",
1819 wiphy_name(ar
->hw
->wiphy
), i
);
1820 __ar9170_dump_txstats(ar
);
1821 #endif /* AR9170_QUEUE_STOP_DEBUG */
1822 ieee80211_wake_queue(ar
->hw
, i
);
1823 spin_unlock_irqrestore(&ar
->tx_stats_lock
, flags
);
1827 if (schedule_garbagecollector
)
1828 queue_delayed_work(ar
->hw
->workqueue
,
1830 msecs_to_jiffies(AR9170_JANITOR_DELAY
));
1833 static bool ar9170_tx_ampdu_queue(struct ar9170
*ar
, struct sk_buff
*skb
)
1835 struct ieee80211_tx_info
*txinfo
;
1836 struct ar9170_sta_info
*sta_info
;
1837 struct ar9170_sta_tid
*agg
;
1838 struct sk_buff
*iter
;
1839 unsigned long flags
, f2
;
1842 bool run
= false, queue
= false;
1844 tid
= ar9170_get_tid(skb
);
1845 seq
= ar9170_get_seq(skb
);
1846 txinfo
= IEEE80211_SKB_CB(skb
);
1847 sta_info
= (void *) txinfo
->control
.sta
->drv_priv
;
1848 agg
= &sta_info
->agg
[tid
];
1849 max
= sta_info
->ampdu_max_len
;
1851 spin_lock_irqsave(&ar
->tx_ampdu_list_lock
, flags
);
1853 if (unlikely(agg
->state
!= AR9170_TID_STATE_COMPLETE
)) {
1854 #ifdef AR9170_TXAGG_DEBUG
1855 printk(KERN_DEBUG
"%s: BlockACK session not fully initialized "
1856 "for ESS:%pM tid:%d state:%d.\n",
1857 wiphy_name(ar
->hw
->wiphy
), agg
->addr
, agg
->tid
,
1859 #endif /* AR9170_TXAGG_DEBUG */
1869 /* check if seq is within the BA window */
1870 if (unlikely(!BAW_WITHIN(agg
->ssn
, max
, seq
))) {
1871 #ifdef AR9170_TXAGG_DEBUG
1872 printk(KERN_DEBUG
"%s: frame with tid:%d seq:%d does not "
1873 "fit into BA window (%d - %d)\n",
1874 wiphy_name(ar
->hw
->wiphy
), tid
, seq
, agg
->ssn
,
1875 (agg
->ssn
+ max
) & 0xfff);
1876 #endif /* AR9170_TXAGG_DEBUG */
1880 spin_lock_irqsave(&agg
->queue
.lock
, f2
);
1882 skb_queue_reverse_walk(&agg
->queue
, iter
) {
1883 qseq
= ar9170_get_seq(iter
);
1885 if (GET_NEXT_SEQ(qseq
) == seq
) {
1886 __skb_queue_after(&agg
->queue
, iter
, skb
);
1891 __skb_queue_head(&agg
->queue
, skb
);
1894 spin_unlock_irqrestore(&agg
->queue
.lock
, f2
);
1896 #ifdef AR9170_TXAGG_DEBUG
1897 printk(KERN_DEBUG
"%s: new aggregate %p queued.\n",
1898 wiphy_name(ar
->hw
->wiphy
), skb
);
1899 __ar9170_dump_txqueue(ar
, &agg
->queue
);
1900 #endif /* AR9170_TXAGG_DEBUG */
1902 if (skb_queue_len(&agg
->queue
) >= AR9170_NUM_TX_AGG_MAX
)
1906 list_add_tail(&agg
->list
, &ar
->tx_ampdu_list
);
1908 spin_unlock_irqrestore(&ar
->tx_ampdu_list_lock
, flags
);
1912 spin_unlock_irqrestore(&ar
->tx_ampdu_list_lock
, flags
);
1913 dev_kfree_skb_irq(skb
);
1917 int ar9170_op_tx(struct ieee80211_hw
*hw
, struct sk_buff
*skb
)
1919 struct ar9170
*ar
= hw
->priv
;
1920 struct ieee80211_tx_info
*info
;
1922 if (unlikely(!IS_STARTED(ar
)))
1925 if (unlikely(ar9170_tx_prepare(ar
, skb
)))
1928 info
= IEEE80211_SKB_CB(skb
);
1929 if (info
->flags
& IEEE80211_TX_CTL_AMPDU
) {
1930 bool run
= ar9170_tx_ampdu_queue(ar
, skb
);
1932 if (run
|| !ar
->tx_ampdu_pending
)
1933 ar9170_tx_ampdu(ar
);
1935 unsigned int queue
= skb_get_queue_mapping(skb
);
1937 ar9170_tx_prepare_phy(ar
, skb
);
1938 skb_queue_tail(&ar
->tx_pending
[queue
], skb
);
1942 return NETDEV_TX_OK
;
1945 dev_kfree_skb_any(skb
);
1946 return NETDEV_TX_OK
;
1949 static int ar9170_op_add_interface(struct ieee80211_hw
*hw
,
1950 struct ieee80211_if_init_conf
*conf
)
1952 struct ar9170
*ar
= hw
->priv
;
1955 mutex_lock(&ar
->mutex
);
1962 ar
->vif
= conf
->vif
;
1963 memcpy(ar
->mac_addr
, conf
->mac_addr
, ETH_ALEN
);
1965 if (modparam_nohwcrypt
|| (ar
->vif
->type
!= NL80211_IFTYPE_STATION
)) {
1966 ar
->rx_software_decryption
= true;
1967 ar
->disable_offload
= true;
1971 ar
->want_filter
= AR9170_MAC_REG_FTF_DEFAULTS
;
1972 err
= ar9170_update_frame_filter(ar
);
1976 err
= ar9170_set_operating_mode(ar
);
1979 mutex_unlock(&ar
->mutex
);
1983 static void ar9170_op_remove_interface(struct ieee80211_hw
*hw
,
1984 struct ieee80211_if_init_conf
*conf
)
1986 struct ar9170
*ar
= hw
->priv
;
1988 mutex_lock(&ar
->mutex
);
1990 ar
->want_filter
= 0;
1991 ar9170_update_frame_filter(ar
);
1992 ar9170_set_beacon_timers(ar
);
1993 dev_kfree_skb(ar
->beacon
);
1995 ar
->sniffer_enabled
= false;
1996 ar
->rx_software_decryption
= false;
1997 ar9170_set_operating_mode(ar
);
1998 mutex_unlock(&ar
->mutex
);
2001 static int ar9170_op_config(struct ieee80211_hw
*hw
, u32 changed
)
2003 struct ar9170
*ar
= hw
->priv
;
2006 mutex_lock(&ar
->mutex
);
2008 if (changed
& IEEE80211_CONF_CHANGE_LISTEN_INTERVAL
) {
2013 if (changed
& IEEE80211_CONF_CHANGE_PS
) {
2018 if (changed
& IEEE80211_CONF_CHANGE_POWER
) {
2023 if (changed
& IEEE80211_CONF_CHANGE_RETRY_LIMITS
) {
2025 * is it long_frame_max_tx_count or short_frame_max_tx_count?
2028 err
= ar9170_set_hwretry_limit(ar
,
2029 ar
->hw
->conf
.long_frame_max_tx_count
);
2034 if (changed
& BSS_CHANGED_BEACON_INT
) {
2035 err
= ar9170_set_beacon_timers(ar
);
2040 if (changed
& IEEE80211_CONF_CHANGE_CHANNEL
) {
2042 /* adjust slot time for 5 GHz */
2043 err
= ar9170_set_slot_time(ar
);
2047 err
= ar9170_set_dyn_sifs_ack(ar
);
2051 err
= ar9170_set_channel(ar
, hw
->conf
.channel
,
2053 nl80211_to_ar9170(hw
->conf
.channel_type
));
2059 mutex_unlock(&ar
->mutex
);
2063 static void ar9170_set_filters(struct work_struct
*work
)
2065 struct ar9170
*ar
= container_of(work
, struct ar9170
,
2066 filter_config_work
);
2069 if (unlikely(!IS_STARTED(ar
)))
2072 mutex_lock(&ar
->mutex
);
2073 if (test_and_clear_bit(AR9170_FILTER_CHANGED_MODE
,
2074 &ar
->filter_changed
)) {
2075 err
= ar9170_set_operating_mode(ar
);
2080 if (test_and_clear_bit(AR9170_FILTER_CHANGED_MULTICAST
,
2081 &ar
->filter_changed
)) {
2082 err
= ar9170_update_multicast(ar
);
2087 if (test_and_clear_bit(AR9170_FILTER_CHANGED_FRAMEFILTER
,
2088 &ar
->filter_changed
)) {
2089 err
= ar9170_update_frame_filter(ar
);
2095 mutex_unlock(&ar
->mutex
);
2098 static void ar9170_op_configure_filter(struct ieee80211_hw
*hw
,
2099 unsigned int changed_flags
,
2100 unsigned int *new_flags
,
2101 int mc_count
, struct dev_mc_list
*mclist
)
2103 struct ar9170
*ar
= hw
->priv
;
2105 /* mask supported flags */
2106 *new_flags
&= FIF_ALLMULTI
| FIF_CONTROL
| FIF_BCN_PRBRESP_PROMISC
|
2107 FIF_PROMISC_IN_BSS
| FIF_FCSFAIL
| FIF_PLCPFAIL
;
2108 ar
->filter_state
= *new_flags
;
2110 * We can support more by setting the sniffer bit and
2111 * then checking the error flags, later.
2114 if (changed_flags
& FIF_ALLMULTI
) {
2115 if (*new_flags
& FIF_ALLMULTI
) {
2116 ar
->want_mc_hash
= ~0ULL;
2121 /* always get broadcast frames */
2122 mchash
= 1ULL << (0xff >> 2);
2124 for (i
= 0; i
< mc_count
; i
++) {
2125 if (WARN_ON(!mclist
))
2127 mchash
|= 1ULL << (mclist
->dmi_addr
[5] >> 2);
2128 mclist
= mclist
->next
;
2130 ar
->want_mc_hash
= mchash
;
2132 set_bit(AR9170_FILTER_CHANGED_MULTICAST
, &ar
->filter_changed
);
2135 if (changed_flags
& FIF_CONTROL
) {
2136 u32 filter
= AR9170_MAC_REG_FTF_PSPOLL
|
2137 AR9170_MAC_REG_FTF_RTS
|
2138 AR9170_MAC_REG_FTF_CTS
|
2139 AR9170_MAC_REG_FTF_ACK
|
2140 AR9170_MAC_REG_FTF_CFE
|
2141 AR9170_MAC_REG_FTF_CFE_ACK
;
2143 if (*new_flags
& FIF_CONTROL
)
2144 ar
->want_filter
= ar
->cur_filter
| filter
;
2146 ar
->want_filter
= ar
->cur_filter
& ~filter
;
2148 set_bit(AR9170_FILTER_CHANGED_FRAMEFILTER
,
2149 &ar
->filter_changed
);
2152 if (changed_flags
& FIF_PROMISC_IN_BSS
) {
2153 ar
->sniffer_enabled
= ((*new_flags
) & FIF_PROMISC_IN_BSS
) != 0;
2154 set_bit(AR9170_FILTER_CHANGED_MODE
,
2155 &ar
->filter_changed
);
2158 if (likely(IS_STARTED(ar
)))
2159 queue_work(ar
->hw
->workqueue
, &ar
->filter_config_work
);
2162 static void ar9170_op_bss_info_changed(struct ieee80211_hw
*hw
,
2163 struct ieee80211_vif
*vif
,
2164 struct ieee80211_bss_conf
*bss_conf
,
2167 struct ar9170
*ar
= hw
->priv
;
2170 mutex_lock(&ar
->mutex
);
2172 if (changed
& BSS_CHANGED_BSSID
) {
2173 memcpy(ar
->bssid
, bss_conf
->bssid
, ETH_ALEN
);
2174 err
= ar9170_set_operating_mode(ar
);
2179 if (changed
& (BSS_CHANGED_BEACON
| BSS_CHANGED_BEACON_ENABLED
)) {
2180 err
= ar9170_update_beacon(ar
);
2184 err
= ar9170_set_beacon_timers(ar
);
2189 if (changed
& BSS_CHANGED_ASSOC
) {
2190 #ifndef CONFIG_AR9170_LEDS
2191 /* enable assoc LED. */
2192 err
= ar9170_set_leds_state(ar
, bss_conf
->assoc
? 2 : 0);
2193 #endif /* CONFIG_AR9170_LEDS */
2196 if (changed
& BSS_CHANGED_BEACON_INT
) {
2197 err
= ar9170_set_beacon_timers(ar
);
2202 if (changed
& BSS_CHANGED_HT
) {
2207 if (changed
& BSS_CHANGED_ERP_SLOT
) {
2208 err
= ar9170_set_slot_time(ar
);
2213 if (changed
& BSS_CHANGED_BASIC_RATES
) {
2214 err
= ar9170_set_basic_rates(ar
);
2220 mutex_unlock(&ar
->mutex
);
2223 static u64
ar9170_op_get_tsf(struct ieee80211_hw
*hw
)
2225 struct ar9170
*ar
= hw
->priv
;
2231 mutex_lock(&ar
->mutex
);
2232 err
= ar9170_read_reg(ar
, AR9170_MAC_REG_TSF_L
, &tsf_low
);
2234 err
= ar9170_read_reg(ar
, AR9170_MAC_REG_TSF_H
, &tsf_high
);
2235 mutex_unlock(&ar
->mutex
);
2241 tsf
= (tsf
<< 32) | tsf_low
;
2245 static int ar9170_set_key(struct ieee80211_hw
*hw
, enum set_key_cmd cmd
,
2246 struct ieee80211_vif
*vif
, struct ieee80211_sta
*sta
,
2247 struct ieee80211_key_conf
*key
)
2249 struct ar9170
*ar
= hw
->priv
;
2253 if ((!ar
->vif
) || (ar
->disable_offload
))
2258 if (key
->keylen
== WLAN_KEY_LEN_WEP40
)
2259 ktype
= AR9170_ENC_ALG_WEP64
;
2261 ktype
= AR9170_ENC_ALG_WEP128
;
2264 ktype
= AR9170_ENC_ALG_TKIP
;
2267 ktype
= AR9170_ENC_ALG_AESCCMP
;
2273 mutex_lock(&ar
->mutex
);
2274 if (cmd
== SET_KEY
) {
2275 if (unlikely(!IS_STARTED(ar
))) {
2280 /* group keys need all-zeroes address */
2281 if (!(key
->flags
& IEEE80211_KEY_FLAG_PAIRWISE
))
2284 if (key
->flags
& IEEE80211_KEY_FLAG_PAIRWISE
) {
2285 for (i
= 0; i
< 64; i
++)
2286 if (!(ar
->usedkeys
& BIT(i
)))
2289 ar
->rx_software_decryption
= true;
2290 ar9170_set_operating_mode(ar
);
2295 i
= 64 + key
->keyidx
;
2298 key
->hw_key_idx
= i
;
2300 err
= ar9170_upload_key(ar
, i
, sta
? sta
->addr
: NULL
, ktype
, 0,
2301 key
->key
, min_t(u8
, 16, key
->keylen
));
2305 if (key
->alg
== ALG_TKIP
) {
2306 err
= ar9170_upload_key(ar
, i
, sta
? sta
->addr
: NULL
,
2307 ktype
, 1, key
->key
+ 16, 16);
2312 * hardware is not capable generating the MMIC
2313 * for fragmented frames!
2315 key
->flags
|= IEEE80211_KEY_FLAG_GENERATE_MMIC
;
2319 ar
->usedkeys
|= BIT(i
);
2321 key
->flags
|= IEEE80211_KEY_FLAG_GENERATE_IV
;
2323 if (unlikely(!IS_STARTED(ar
))) {
2324 /* The device is gone... together with the key ;-) */
2329 err
= ar9170_disable_key(ar
, key
->hw_key_idx
);
2333 if (key
->hw_key_idx
< 64) {
2334 ar
->usedkeys
&= ~BIT(key
->hw_key_idx
);
2336 err
= ar9170_upload_key(ar
, key
->hw_key_idx
, NULL
,
2337 AR9170_ENC_ALG_NONE
, 0,
2342 if (key
->alg
== ALG_TKIP
) {
2343 err
= ar9170_upload_key(ar
, key
->hw_key_idx
,
2345 AR9170_ENC_ALG_NONE
, 1,
2354 ar9170_regwrite_begin(ar
);
2355 ar9170_regwrite(AR9170_MAC_REG_ROLL_CALL_TBL_L
, ar
->usedkeys
);
2356 ar9170_regwrite(AR9170_MAC_REG_ROLL_CALL_TBL_H
, ar
->usedkeys
>> 32);
2357 ar9170_regwrite_finish();
2358 err
= ar9170_regwrite_result();
2361 mutex_unlock(&ar
->mutex
);
2366 static void ar9170_sta_notify(struct ieee80211_hw
*hw
,
2367 struct ieee80211_vif
*vif
,
2368 enum sta_notify_cmd cmd
,
2369 struct ieee80211_sta
*sta
)
2371 struct ar9170
*ar
= hw
->priv
;
2372 struct ar9170_sta_info
*sta_info
= (void *) sta
->drv_priv
;
2376 case STA_NOTIFY_ADD
:
2377 memset(sta_info
, 0, sizeof(*sta_info
));
2379 if (!sta
->ht_cap
.ht_supported
)
2382 if (sta
->ht_cap
.ampdu_density
> ar
->global_ampdu_density
)
2383 ar
->global_ampdu_density
= sta
->ht_cap
.ampdu_density
;
2385 if (sta
->ht_cap
.ampdu_factor
< ar
->global_ampdu_factor
)
2386 ar
->global_ampdu_factor
= sta
->ht_cap
.ampdu_factor
;
2388 for (i
= 0; i
< AR9170_NUM_TID
; i
++) {
2389 sta_info
->agg
[i
].state
= AR9170_TID_STATE_SHUTDOWN
;
2390 sta_info
->agg
[i
].active
= false;
2391 sta_info
->agg
[i
].ssn
= 0;
2392 sta_info
->agg
[i
].retry
= 0;
2393 sta_info
->agg
[i
].tid
= i
;
2394 INIT_LIST_HEAD(&sta_info
->agg
[i
].list
);
2395 skb_queue_head_init(&sta_info
->agg
[i
].queue
);
2398 sta_info
->ampdu_max_len
= 1 << (3 + sta
->ht_cap
.ampdu_factor
);
2401 case STA_NOTIFY_REMOVE
:
2402 if (!sta
->ht_cap
.ht_supported
)
2405 for (i
= 0; i
< AR9170_NUM_TID
; i
++) {
2406 sta_info
->agg
[i
].state
= AR9170_TID_STATE_INVALID
;
2407 skb_queue_purge(&sta_info
->agg
[i
].queue
);
2416 if (IS_STARTED(ar
) && ar
->filter_changed
)
2417 queue_work(ar
->hw
->workqueue
, &ar
->filter_config_work
);
2420 static int ar9170_get_stats(struct ieee80211_hw
*hw
,
2421 struct ieee80211_low_level_stats
*stats
)
2423 struct ar9170
*ar
= hw
->priv
;
2427 mutex_lock(&ar
->mutex
);
2428 err
= ar9170_read_reg(ar
, AR9170_MAC_REG_TX_RETRY
, &val
);
2429 ar
->stats
.dot11ACKFailureCount
+= val
;
2431 memcpy(stats
, &ar
->stats
, sizeof(*stats
));
2432 mutex_unlock(&ar
->mutex
);
2437 static int ar9170_get_tx_stats(struct ieee80211_hw
*hw
,
2438 struct ieee80211_tx_queue_stats
*tx_stats
)
2440 struct ar9170
*ar
= hw
->priv
;
2442 spin_lock_bh(&ar
->tx_stats_lock
);
2443 memcpy(tx_stats
, ar
->tx_stats
, sizeof(tx_stats
[0]) * hw
->queues
);
2444 spin_unlock_bh(&ar
->tx_stats_lock
);
2449 static int ar9170_conf_tx(struct ieee80211_hw
*hw
, u16 queue
,
2450 const struct ieee80211_tx_queue_params
*param
)
2452 struct ar9170
*ar
= hw
->priv
;
2455 mutex_lock(&ar
->mutex
);
2456 if ((param
) && !(queue
> __AR9170_NUM_TXQ
)) {
2457 memcpy(&ar
->edcf
[ar9170_qos_hwmap
[queue
]],
2458 param
, sizeof(*param
));
2460 ret
= ar9170_set_qos(ar
);
2464 mutex_unlock(&ar
->mutex
);
2468 static int ar9170_ampdu_action(struct ieee80211_hw
*hw
,
2469 enum ieee80211_ampdu_mlme_action action
,
2470 struct ieee80211_sta
*sta
, u16 tid
, u16
*ssn
)
2472 struct ar9170
*ar
= hw
->priv
;
2473 struct ar9170_sta_info
*sta_info
= (void *) sta
->drv_priv
;
2474 struct ar9170_sta_tid
*tid_info
= &sta_info
->agg
[tid
];
2475 unsigned long flags
;
2481 case IEEE80211_AMPDU_TX_START
:
2482 spin_lock_irqsave(&ar
->tx_ampdu_list_lock
, flags
);
2483 if (tid_info
->state
!= AR9170_TID_STATE_SHUTDOWN
||
2484 !list_empty(&tid_info
->list
)) {
2485 spin_unlock_irqrestore(&ar
->tx_ampdu_list_lock
, flags
);
2486 #ifdef AR9170_TXAGG_DEBUG
2487 printk(KERN_INFO
"%s: A-MPDU [ESS:[%pM] tid:[%d]] "
2488 "is in a very bad state!\n",
2489 wiphy_name(hw
->wiphy
), sta
->addr
, tid
);
2490 #endif /* AR9170_TXAGG_DEBUG */
2494 *ssn
= tid_info
->ssn
;
2495 tid_info
->state
= AR9170_TID_STATE_PROGRESS
;
2496 tid_info
->active
= false;
2497 spin_unlock_irqrestore(&ar
->tx_ampdu_list_lock
, flags
);
2498 ieee80211_start_tx_ba_cb_irqsafe(hw
, sta
->addr
, tid
);
2501 case IEEE80211_AMPDU_TX_STOP
:
2502 spin_lock_irqsave(&ar
->tx_ampdu_list_lock
, flags
);
2503 tid_info
->state
= AR9170_TID_STATE_SHUTDOWN
;
2504 list_del_init(&tid_info
->list
);
2505 tid_info
->active
= false;
2506 skb_queue_purge(&tid_info
->queue
);
2507 spin_unlock_irqrestore(&ar
->tx_ampdu_list_lock
, flags
);
2508 ieee80211_stop_tx_ba_cb_irqsafe(hw
, sta
->addr
, tid
);
2511 case IEEE80211_AMPDU_TX_OPERATIONAL
:
2512 #ifdef AR9170_TXAGG_DEBUG
2513 printk(KERN_INFO
"%s: A-MPDU for %pM [tid:%d] Operational.\n",
2514 wiphy_name(hw
->wiphy
), sta
->addr
, tid
);
2515 #endif /* AR9170_TXAGG_DEBUG */
2516 spin_lock_irqsave(&ar
->tx_ampdu_list_lock
, flags
);
2517 sta_info
->agg
[tid
].state
= AR9170_TID_STATE_COMPLETE
;
2518 spin_unlock_irqrestore(&ar
->tx_ampdu_list_lock
, flags
);
2521 case IEEE80211_AMPDU_RX_START
:
2522 case IEEE80211_AMPDU_RX_STOP
:
2523 /* Handled by firmware */
2533 static const struct ieee80211_ops ar9170_ops
= {
2534 .start
= ar9170_op_start
,
2535 .stop
= ar9170_op_stop
,
2537 .add_interface
= ar9170_op_add_interface
,
2538 .remove_interface
= ar9170_op_remove_interface
,
2539 .config
= ar9170_op_config
,
2540 .configure_filter
= ar9170_op_configure_filter
,
2541 .conf_tx
= ar9170_conf_tx
,
2542 .bss_info_changed
= ar9170_op_bss_info_changed
,
2543 .get_tsf
= ar9170_op_get_tsf
,
2544 .set_key
= ar9170_set_key
,
2545 .sta_notify
= ar9170_sta_notify
,
2546 .get_stats
= ar9170_get_stats
,
2547 .get_tx_stats
= ar9170_get_tx_stats
,
2548 .ampdu_action
= ar9170_ampdu_action
,
2551 void *ar9170_alloc(size_t priv_size
)
2553 struct ieee80211_hw
*hw
;
2555 struct sk_buff
*skb
;
2559 * this buffer is used for rx stream reconstruction.
2560 * Under heavy load this device (or the transport layer?)
2561 * tends to split the streams into seperate rx descriptors.
2564 skb
= __dev_alloc_skb(AR9170_MAX_RX_BUFFER_SIZE
, GFP_KERNEL
);
2568 hw
= ieee80211_alloc_hw(priv_size
, &ar9170_ops
);
2574 ar
->rx_failover
= skb
;
2576 mutex_init(&ar
->mutex
);
2577 spin_lock_init(&ar
->cmdlock
);
2578 spin_lock_init(&ar
->tx_stats_lock
);
2579 spin_lock_init(&ar
->tx_ampdu_list_lock
);
2580 skb_queue_head_init(&ar
->tx_status_ampdu
);
2581 for (i
= 0; i
< __AR9170_NUM_TXQ
; i
++) {
2582 skb_queue_head_init(&ar
->tx_status
[i
]);
2583 skb_queue_head_init(&ar
->tx_pending
[i
]);
2585 ar9170_rx_reset_rx_mpdu(ar
);
2586 INIT_WORK(&ar
->filter_config_work
, ar9170_set_filters
);
2587 INIT_WORK(&ar
->beacon_work
, ar9170_new_beacon
);
2588 INIT_DELAYED_WORK(&ar
->tx_janitor
, ar9170_tx_janitor
);
2589 INIT_LIST_HEAD(&ar
->tx_ampdu_list
);
2591 /* all hw supports 2.4 GHz, so set channel to 1 by default */
2592 ar
->channel
= &ar9170_2ghz_chantable
[0];
2594 /* first part of wiphy init */
2595 ar
->hw
->wiphy
->interface_modes
= BIT(NL80211_IFTYPE_STATION
) |
2596 BIT(NL80211_IFTYPE_WDS
) |
2597 BIT(NL80211_IFTYPE_ADHOC
);
2598 ar
->hw
->flags
|= IEEE80211_HW_RX_INCLUDES_FCS
|
2599 IEEE80211_HW_HOST_BROADCAST_PS_BUFFERING
|
2600 IEEE80211_HW_SIGNAL_DBM
|
2601 IEEE80211_HW_NOISE_DBM
;
2604 ar
->hw
->flags
|= IEEE80211_HW_AMPDU_AGGREGATION
;
2606 ar9170_band_2GHz
.ht_cap
.ht_supported
= false;
2607 ar9170_band_5GHz
.ht_cap
.ht_supported
= false;
2610 ar
->hw
->queues
= __AR9170_NUM_TXQ
;
2611 ar
->hw
->extra_tx_headroom
= 8;
2612 ar
->hw
->sta_data_size
= sizeof(struct ar9170_sta_info
);
2614 ar
->hw
->max_rates
= 1;
2615 ar
->hw
->max_rate_tries
= 3;
2617 for (i
= 0; i
< ARRAY_SIZE(ar
->noise
); i
++)
2618 ar
->noise
[i
] = -95; /* ATH_DEFAULT_NOISE_FLOOR */
2624 return ERR_PTR(-ENOMEM
);
2627 static int ar9170_read_eeprom(struct ar9170
*ar
)
2629 #define RW 8 /* number of words to read at once */
2630 #define RB (sizeof(u32) * RW)
2631 DECLARE_MAC_BUF(mbuf
);
2632 u8
*eeprom
= (void *)&ar
->eeprom
;
2633 u8
*addr
= ar
->eeprom
.mac_address
;
2635 unsigned int rx_streams
, tx_streams
, tx_params
= 0;
2636 int i
, j
, err
, bands
= 0;
2638 BUILD_BUG_ON(sizeof(ar
->eeprom
) & 3);
2640 BUILD_BUG_ON(RB
> AR9170_MAX_CMD_LEN
- 4);
2642 /* don't want to handle trailing remains */
2643 BUILD_BUG_ON(sizeof(ar
->eeprom
) % RB
);
2646 for (i
= 0; i
< sizeof(ar
->eeprom
)/RB
; i
++) {
2647 for (j
= 0; j
< RW
; j
++)
2648 offsets
[j
] = cpu_to_le32(AR9170_EEPROM_START
+
2651 err
= ar
->exec_cmd(ar
, AR9170_CMD_RREG
,
2652 RB
, (u8
*) &offsets
,
2653 RB
, eeprom
+ RB
* i
);
2661 if (ar
->eeprom
.length
== cpu_to_le16(0xFFFF))
2664 if (ar
->eeprom
.operating_flags
& AR9170_OPFLAG_2GHZ
) {
2665 ar
->hw
->wiphy
->bands
[IEEE80211_BAND_2GHZ
] = &ar9170_band_2GHz
;
2668 if (ar
->eeprom
.operating_flags
& AR9170_OPFLAG_5GHZ
) {
2669 ar
->hw
->wiphy
->bands
[IEEE80211_BAND_5GHZ
] = &ar9170_band_5GHz
;
2673 rx_streams
= hweight8(ar
->eeprom
.rx_mask
);
2674 tx_streams
= hweight8(ar
->eeprom
.tx_mask
);
2676 if (rx_streams
!= tx_streams
)
2677 tx_params
= IEEE80211_HT_MCS_TX_RX_DIFF
;
2679 if (tx_streams
>= 1 && tx_streams
<= IEEE80211_HT_MCS_TX_MAX_STREAMS
)
2680 tx_params
= (tx_streams
- 1) <<
2681 IEEE80211_HT_MCS_TX_MAX_STREAMS_SHIFT
;
2683 ar9170_band_2GHz
.ht_cap
.mcs
.tx_params
|= tx_params
;
2684 ar9170_band_5GHz
.ht_cap
.mcs
.tx_params
|= tx_params
;
2687 * I measured this, a bandswitch takes roughly
2688 * 135 ms and a frequency switch about 80.
2690 * FIXME: measure these values again once EEPROM settings
2691 * are used, that will influence them!
2694 ar
->hw
->channel_change_time
= 135 * 1000;
2696 ar
->hw
->channel_change_time
= 80 * 1000;
2698 ar
->regulatory
.current_rd
= le16_to_cpu(ar
->eeprom
.reg_domain
[0]);
2699 ar
->regulatory
.current_rd_ext
= le16_to_cpu(ar
->eeprom
.reg_domain
[1]);
2701 /* second part of wiphy init */
2702 SET_IEEE80211_PERM_ADDR(ar
->hw
, addr
);
2704 return bands
? 0 : -EINVAL
;
2707 static int ar9170_reg_notifier(struct wiphy
*wiphy
,
2708 struct regulatory_request
*request
)
2710 struct ieee80211_hw
*hw
= wiphy_to_ieee80211_hw(wiphy
);
2711 struct ar9170
*ar
= hw
->priv
;
2713 return ath_reg_notifier_apply(wiphy
, request
, &ar
->regulatory
);
2716 int ar9170_register(struct ar9170
*ar
, struct device
*pdev
)
2720 /* try to read EEPROM, init MAC addr */
2721 err
= ar9170_read_eeprom(ar
);
2725 err
= ath_regd_init(&ar
->regulatory
, ar
->hw
->wiphy
,
2726 ar9170_reg_notifier
);
2730 err
= ieee80211_register_hw(ar
->hw
);
2734 if (!ath_is_world_regd(&ar
->regulatory
))
2735 regulatory_hint(ar
->hw
->wiphy
, ar
->regulatory
.alpha2
);
2737 err
= ar9170_init_leds(ar
);
2741 #ifdef CONFIG_AR9170_LEDS
2742 err
= ar9170_register_leds(ar
);
2745 #endif /* CONFIG_AR9170_LEDS */
2747 dev_info(pdev
, "Atheros AR9170 is registered as '%s'\n",
2748 wiphy_name(ar
->hw
->wiphy
));
2753 ieee80211_unregister_hw(ar
->hw
);
2759 void ar9170_unregister(struct ar9170
*ar
)
2761 #ifdef CONFIG_AR9170_LEDS
2762 ar9170_unregister_leds(ar
);
2763 #endif /* CONFIG_AR9170_LEDS */
2765 kfree_skb(ar
->rx_failover
);
2766 ieee80211_unregister_hw(ar
->hw
);
2767 mutex_destroy(&ar
->mutex
);