2 * Copyright (c) 2004-2011 Atheros Communications Inc.
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
11 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
13 * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
14 * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
21 static int ath6kl_wmi_sync_point(struct wmi
*wmi
);
23 static const s32 wmi_rate_tbl
[][2] = {
24 /* {W/O SGI, with SGI} */
56 /* 802.1d to AC mapping. Refer pg 57 of WMM-test-plan-v1.2 */
57 static const u8 up_to_ac
[] = {
68 void ath6kl_wmi_set_control_ep(struct wmi
*wmi
, enum htc_endpoint_id ep_id
)
70 if (WARN_ON(ep_id
== ENDPOINT_UNUSED
|| ep_id
>= ENDPOINT_MAX
))
76 enum htc_endpoint_id
ath6kl_wmi_get_control_ep(struct wmi
*wmi
)
81 /* Performs DIX to 802.3 encapsulation for transmit packets.
82 * Assumes the entire DIX header is contigous and that there is
83 * enough room in the buffer for a 802.3 mac header and LLC+SNAP headers.
85 int ath6kl_wmi_dix_2_dot3(struct wmi
*wmi
, struct sk_buff
*skb
)
87 struct ath6kl_llc_snap_hdr
*llc_hdr
;
88 struct ethhdr
*eth_hdr
;
94 if (WARN_ON(skb
== NULL
))
97 size
= sizeof(struct ath6kl_llc_snap_hdr
) + sizeof(struct wmi_data_hdr
);
98 if (skb_headroom(skb
) < size
)
101 eth_hdr
= (struct ethhdr
*) skb
->data
;
102 type
= eth_hdr
->h_proto
;
104 if (!is_ethertype(be16_to_cpu(type
))) {
105 ath6kl_dbg(ATH6KL_DBG_WMI
,
106 "%s: pkt is already in 802.3 format\n", __func__
);
110 new_len
= skb
->len
- sizeof(*eth_hdr
) + sizeof(*llc_hdr
);
112 skb_push(skb
, sizeof(struct ath6kl_llc_snap_hdr
));
115 eth_hdr
->h_proto
= cpu_to_be16(new_len
);
117 memcpy(datap
, eth_hdr
, sizeof(*eth_hdr
));
119 llc_hdr
= (struct ath6kl_llc_snap_hdr
*)(datap
+ sizeof(*eth_hdr
));
120 llc_hdr
->dsap
= 0xAA;
121 llc_hdr
->ssap
= 0xAA;
122 llc_hdr
->cntl
= 0x03;
123 llc_hdr
->org_code
[0] = 0x0;
124 llc_hdr
->org_code
[1] = 0x0;
125 llc_hdr
->org_code
[2] = 0x0;
126 llc_hdr
->eth_type
= type
;
131 static int ath6kl_wmi_meta_add(struct wmi
*wmi
, struct sk_buff
*skb
,
132 u8
*version
, void *tx_meta_info
)
134 struct wmi_tx_meta_v1
*v1
;
135 struct wmi_tx_meta_v2
*v2
;
137 if (WARN_ON(skb
== NULL
|| version
== NULL
))
141 case WMI_META_VERSION_1
:
142 skb_push(skb
, WMI_MAX_TX_META_SZ
);
143 v1
= (struct wmi_tx_meta_v1
*) skb
->data
;
145 v1
->rate_plcy_id
= 0;
146 *version
= WMI_META_VERSION_1
;
148 case WMI_META_VERSION_2
:
149 skb_push(skb
, WMI_MAX_TX_META_SZ
);
150 v2
= (struct wmi_tx_meta_v2
*) skb
->data
;
151 memcpy(v2
, (struct wmi_tx_meta_v2
*) tx_meta_info
,
152 sizeof(struct wmi_tx_meta_v2
));
159 int ath6kl_wmi_data_hdr_add(struct wmi
*wmi
, struct sk_buff
*skb
,
160 u8 msg_type
, bool more_data
,
161 enum wmi_data_hdr_data_type data_type
,
162 u8 meta_ver
, void *tx_meta_info
)
164 struct wmi_data_hdr
*data_hdr
;
167 if (WARN_ON(skb
== NULL
))
171 ret
= ath6kl_wmi_meta_add(wmi
, skb
, &meta_ver
, tx_meta_info
);
176 skb_push(skb
, sizeof(struct wmi_data_hdr
));
178 data_hdr
= (struct wmi_data_hdr
*)skb
->data
;
179 memset(data_hdr
, 0, sizeof(struct wmi_data_hdr
));
181 data_hdr
->info
= msg_type
<< WMI_DATA_HDR_MSG_TYPE_SHIFT
;
182 data_hdr
->info
|= data_type
<< WMI_DATA_HDR_DATA_TYPE_SHIFT
;
186 WMI_DATA_HDR_MORE_MASK
<< WMI_DATA_HDR_MORE_SHIFT
;
188 data_hdr
->info2
= cpu_to_le16(meta_ver
<< WMI_DATA_HDR_META_SHIFT
);
194 static u8
ath6kl_wmi_determine_user_priority(u8
*pkt
, u32 layer2_pri
)
196 struct iphdr
*ip_hdr
= (struct iphdr
*) pkt
;
200 * Determine IPTOS priority
203 * : DSCP(6-bits) ECN(2-bits)
204 * : DSCP - P2 P1 P0 X X X
205 * where (P2 P1 P0) form 802.1D
207 ip_pri
= ip_hdr
->tos
>> 5;
210 if ((layer2_pri
& 0x7) > ip_pri
)
211 return (u8
) layer2_pri
& 0x7;
216 int ath6kl_wmi_implicit_create_pstream(struct wmi
*wmi
, struct sk_buff
*skb
,
217 u32 layer2_priority
, bool wmm_enabled
,
220 struct wmi_data_hdr
*data_hdr
;
221 struct ath6kl_llc_snap_hdr
*llc_hdr
;
222 struct wmi_create_pstream_cmd cmd
;
223 u32 meta_size
, hdr_size
;
224 u16 ip_type
= IP_ETHERTYPE
;
225 u8 stream_exist
, usr_pri
;
226 u8 traffic_class
= WMM_AC_BE
;
229 if (WARN_ON(skb
== NULL
))
233 data_hdr
= (struct wmi_data_hdr
*) datap
;
235 meta_size
= ((le16_to_cpu(data_hdr
->info2
) >> WMI_DATA_HDR_META_SHIFT
) &
236 WMI_DATA_HDR_META_MASK
) ? WMI_MAX_TX_META_SZ
: 0;
239 /* If WMM is disabled all traffic goes as BE traffic */
242 hdr_size
= sizeof(struct ethhdr
);
244 llc_hdr
= (struct ath6kl_llc_snap_hdr
*)(datap
+
247 meta_size
+ hdr_size
);
249 if (llc_hdr
->eth_type
== htons(ip_type
)) {
251 * Extract the endpoint info from the TOS field
255 ath6kl_wmi_determine_user_priority(((u8
*) llc_hdr
) +
256 sizeof(struct ath6kl_llc_snap_hdr
),
259 usr_pri
= layer2_priority
& 0x7;
262 /* workaround for WMM S5 */
263 if ((wmi
->traffic_class
== WMM_AC_VI
) &&
264 ((usr_pri
== 5) || (usr_pri
== 4)))
267 /* Convert user priority to traffic class */
268 traffic_class
= up_to_ac
[usr_pri
& 0x7];
270 wmi_data_hdr_set_up(data_hdr
, usr_pri
);
272 spin_lock_bh(&wmi
->lock
);
273 stream_exist
= wmi
->fat_pipe_exist
;
274 spin_unlock_bh(&wmi
->lock
);
276 if (!(stream_exist
& (1 << traffic_class
))) {
277 memset(&cmd
, 0, sizeof(cmd
));
278 cmd
.traffic_class
= traffic_class
;
279 cmd
.user_pri
= usr_pri
;
281 cpu_to_le32(WMI_IMPLICIT_PSTREAM_INACTIVITY_INT
);
282 /* Implicit streams are created with TSID 0xFF */
283 cmd
.tsid
= WMI_IMPLICIT_PSTREAM
;
284 ath6kl_wmi_create_pstream_cmd(wmi
, &cmd
);
292 int ath6kl_wmi_dot11_hdr_remove(struct wmi
*wmi
, struct sk_buff
*skb
)
294 struct ieee80211_hdr_3addr
*pwh
, wh
;
295 struct ath6kl_llc_snap_hdr
*llc_hdr
;
296 struct ethhdr eth_hdr
;
301 if (WARN_ON(skb
== NULL
))
305 pwh
= (struct ieee80211_hdr_3addr
*) datap
;
307 sub_type
= pwh
->frame_control
& cpu_to_le16(IEEE80211_FCTL_STYPE
);
309 memcpy((u8
*) &wh
, datap
, sizeof(struct ieee80211_hdr_3addr
));
311 /* Strip off the 802.11 header */
312 if (sub_type
== cpu_to_le16(IEEE80211_STYPE_QOS_DATA
)) {
313 hdr_size
= roundup(sizeof(struct ieee80211_qos_hdr
),
315 skb_pull(skb
, hdr_size
);
316 } else if (sub_type
== cpu_to_le16(IEEE80211_STYPE_DATA
))
317 skb_pull(skb
, sizeof(struct ieee80211_hdr_3addr
));
320 llc_hdr
= (struct ath6kl_llc_snap_hdr
*)(datap
);
322 memset(ð_hdr
, 0, sizeof(eth_hdr
));
323 eth_hdr
.h_proto
= llc_hdr
->eth_type
;
325 switch ((le16_to_cpu(wh
.frame_control
)) &
326 (IEEE80211_FCTL_FROMDS
| IEEE80211_FCTL_TODS
)) {
328 memcpy(eth_hdr
.h_dest
, wh
.addr1
, ETH_ALEN
);
329 memcpy(eth_hdr
.h_source
, wh
.addr2
, ETH_ALEN
);
331 case IEEE80211_FCTL_TODS
:
332 memcpy(eth_hdr
.h_dest
, wh
.addr3
, ETH_ALEN
);
333 memcpy(eth_hdr
.h_source
, wh
.addr2
, ETH_ALEN
);
335 case IEEE80211_FCTL_FROMDS
:
336 memcpy(eth_hdr
.h_dest
, wh
.addr1
, ETH_ALEN
);
337 memcpy(eth_hdr
.h_source
, wh
.addr3
, ETH_ALEN
);
339 case IEEE80211_FCTL_FROMDS
| IEEE80211_FCTL_TODS
:
343 skb_pull(skb
, sizeof(struct ath6kl_llc_snap_hdr
));
344 skb_push(skb
, sizeof(eth_hdr
));
348 memcpy(datap
, ð_hdr
, sizeof(eth_hdr
));
354 * Performs 802.3 to DIX encapsulation for received packets.
355 * Assumes the entire 802.3 header is contigous.
357 int ath6kl_wmi_dot3_2_dix(struct sk_buff
*skb
)
359 struct ath6kl_llc_snap_hdr
*llc_hdr
;
360 struct ethhdr eth_hdr
;
363 if (WARN_ON(skb
== NULL
))
368 memcpy(ð_hdr
, datap
, sizeof(eth_hdr
));
370 llc_hdr
= (struct ath6kl_llc_snap_hdr
*) (datap
+ sizeof(eth_hdr
));
371 eth_hdr
.h_proto
= llc_hdr
->eth_type
;
373 skb_pull(skb
, sizeof(struct ath6kl_llc_snap_hdr
));
376 memcpy(datap
, ð_hdr
, sizeof(eth_hdr
));
381 static void ath6kl_wmi_convert_bssinfo_hdr2_to_hdr(struct sk_buff
*skb
,
384 struct wmi_bss_info_hdr2 bih2
;
385 struct wmi_bss_info_hdr
*bih
;
387 memcpy(&bih2
, datap
, sizeof(struct wmi_bss_info_hdr2
));
390 bih
= (struct wmi_bss_info_hdr
*) skb
->data
;
393 bih
->frame_type
= bih2
.frame_type
;
395 bih
->rssi
= a_cpu_to_sle16(bih2
.snr
- 95);
396 bih
->ie_mask
= cpu_to_le32(le16_to_cpu(bih2
.ie_mask
));
397 memcpy(bih
->bssid
, bih2
.bssid
, ETH_ALEN
);
400 static int ath6kl_wmi_tx_complete_event_rx(u8
*datap
, int len
)
402 struct tx_complete_msg_v1
*msg_v1
;
403 struct wmi_tx_complete_event
*evt
;
407 evt
= (struct wmi_tx_complete_event
*) datap
;
409 ath6kl_dbg(ATH6KL_DBG_WMI
, "comp: %d %d %d\n",
410 evt
->num_msg
, evt
->msg_len
, evt
->msg_type
);
412 if (!AR_DBG_LVL_CHECK(ATH6KL_DBG_WMI
))
415 for (index
= 0; index
< evt
->num_msg
; index
++) {
416 size
= sizeof(struct wmi_tx_complete_event
) +
417 (index
* sizeof(struct tx_complete_msg_v1
));
418 msg_v1
= (struct tx_complete_msg_v1
*)(datap
+ size
);
420 ath6kl_dbg(ATH6KL_DBG_WMI
, "msg: %d %d %d %d\n",
421 msg_v1
->status
, msg_v1
->pkt_id
,
422 msg_v1
->rate_idx
, msg_v1
->ack_failures
);
428 static int ath6kl_wmi_remain_on_chnl_event_rx(struct wmi
*wmi
, u8
*datap
,
431 struct wmi_remain_on_chnl_event
*ev
;
434 struct ieee80211_channel
*chan
;
435 struct ath6kl
*ar
= wmi
->parent_dev
;
437 if (len
< sizeof(*ev
))
440 ev
= (struct wmi_remain_on_chnl_event
*) datap
;
441 freq
= le32_to_cpu(ev
->freq
);
442 dur
= le32_to_cpu(ev
->duration
);
443 ath6kl_dbg(ATH6KL_DBG_WMI
, "remain_on_chnl: freq=%u dur=%u\n",
445 chan
= ieee80211_get_channel(ar
->wdev
->wiphy
, freq
);
447 ath6kl_dbg(ATH6KL_DBG_WMI
, "remain_on_chnl: Unknown channel "
448 "(freq=%u)\n", freq
);
451 cfg80211_ready_on_channel(ar
->net_dev
, 1, chan
, NL80211_CHAN_NO_HT
,
457 static int ath6kl_wmi_cancel_remain_on_chnl_event_rx(struct wmi
*wmi
,
460 struct wmi_cancel_remain_on_chnl_event
*ev
;
463 struct ieee80211_channel
*chan
;
464 struct ath6kl
*ar
= wmi
->parent_dev
;
466 if (len
< sizeof(*ev
))
469 ev
= (struct wmi_cancel_remain_on_chnl_event
*) datap
;
470 freq
= le32_to_cpu(ev
->freq
);
471 dur
= le32_to_cpu(ev
->duration
);
472 ath6kl_dbg(ATH6KL_DBG_WMI
, "cancel_remain_on_chnl: freq=%u dur=%u "
473 "status=%u\n", freq
, dur
, ev
->status
);
474 chan
= ieee80211_get_channel(ar
->wdev
->wiphy
, freq
);
476 ath6kl_dbg(ATH6KL_DBG_WMI
, "cancel_remain_on_chnl: Unknown "
477 "channel (freq=%u)\n", freq
);
480 cfg80211_remain_on_channel_expired(ar
->net_dev
, 1, chan
,
481 NL80211_CHAN_NO_HT
, GFP_ATOMIC
);
486 static int ath6kl_wmi_tx_status_event_rx(struct wmi
*wmi
, u8
*datap
, int len
)
488 struct wmi_tx_status_event
*ev
;
490 struct ath6kl
*ar
= wmi
->parent_dev
;
492 if (len
< sizeof(*ev
))
495 ev
= (struct wmi_tx_status_event
*) datap
;
496 id
= le32_to_cpu(ev
->id
);
497 ath6kl_dbg(ATH6KL_DBG_WMI
, "tx_status: id=%x ack_status=%u\n",
499 if (wmi
->last_mgmt_tx_frame
) {
500 cfg80211_mgmt_tx_status(ar
->net_dev
, id
,
501 wmi
->last_mgmt_tx_frame
,
502 wmi
->last_mgmt_tx_frame_len
,
503 !!ev
->ack_status
, GFP_ATOMIC
);
504 kfree(wmi
->last_mgmt_tx_frame
);
505 wmi
->last_mgmt_tx_frame
= NULL
;
506 wmi
->last_mgmt_tx_frame_len
= 0;
512 static int ath6kl_wmi_rx_probe_req_event_rx(struct wmi
*wmi
, u8
*datap
, int len
)
514 struct wmi_p2p_rx_probe_req_event
*ev
;
517 struct ath6kl
*ar
= wmi
->parent_dev
;
519 if (len
< sizeof(*ev
))
522 ev
= (struct wmi_p2p_rx_probe_req_event
*) datap
;
523 freq
= le32_to_cpu(ev
->freq
);
524 dlen
= le16_to_cpu(ev
->len
);
525 if (datap
+ len
< ev
->data
+ dlen
) {
526 ath6kl_err("invalid wmi_p2p_rx_probe_req_event: "
527 "len=%d dlen=%u\n", len
, dlen
);
530 ath6kl_dbg(ATH6KL_DBG_WMI
, "rx_probe_req: len=%u freq=%u "
531 "probe_req_report=%d\n",
532 dlen
, freq
, ar
->probe_req_report
);
534 if (ar
->probe_req_report
|| ar
->nw_type
== AP_NETWORK
)
535 cfg80211_rx_mgmt(ar
->net_dev
, freq
, ev
->data
, dlen
, GFP_ATOMIC
);
540 static int ath6kl_wmi_p2p_capabilities_event_rx(u8
*datap
, int len
)
542 struct wmi_p2p_capabilities_event
*ev
;
545 if (len
< sizeof(*ev
))
548 ev
= (struct wmi_p2p_capabilities_event
*) datap
;
549 dlen
= le16_to_cpu(ev
->len
);
550 ath6kl_dbg(ATH6KL_DBG_WMI
, "p2p_capab: len=%u\n", dlen
);
555 static int ath6kl_wmi_rx_action_event_rx(u8
*datap
, int len
)
557 struct wmi_rx_action_event
*ev
;
560 if (len
< sizeof(*ev
))
563 ev
= (struct wmi_rx_action_event
*) datap
;
564 dlen
= le16_to_cpu(ev
->len
);
565 ath6kl_dbg(ATH6KL_DBG_WMI
, "rx_action: len=%u\n", dlen
);
570 static int ath6kl_wmi_p2p_info_event_rx(u8
*datap
, int len
)
572 struct wmi_p2p_info_event
*ev
;
576 if (len
< sizeof(*ev
))
579 ev
= (struct wmi_p2p_info_event
*) datap
;
580 flags
= le32_to_cpu(ev
->info_req_flags
);
581 dlen
= le16_to_cpu(ev
->len
);
582 ath6kl_dbg(ATH6KL_DBG_WMI
, "p2p_info: flags=%x len=%d\n", flags
, dlen
);
584 if (flags
& P2P_FLAG_CAPABILITIES_REQ
) {
585 struct wmi_p2p_capabilities
*cap
;
586 if (dlen
< sizeof(*cap
))
588 cap
= (struct wmi_p2p_capabilities
*) ev
->data
;
589 ath6kl_dbg(ATH6KL_DBG_WMI
, "p2p_info: GO Power Save = %d\n",
593 if (flags
& P2P_FLAG_MACADDR_REQ
) {
594 struct wmi_p2p_macaddr
*mac
;
595 if (dlen
< sizeof(*mac
))
597 mac
= (struct wmi_p2p_macaddr
*) ev
->data
;
598 ath6kl_dbg(ATH6KL_DBG_WMI
, "p2p_info: MAC Address = %pM\n",
602 if (flags
& P2P_FLAG_HMODEL_REQ
) {
603 struct wmi_p2p_hmodel
*mod
;
604 if (dlen
< sizeof(*mod
))
606 mod
= (struct wmi_p2p_hmodel
*) ev
->data
;
607 ath6kl_dbg(ATH6KL_DBG_WMI
, "p2p_info: P2P Model = %d (%s)\n",
609 mod
->p2p_model
? "host" : "firmware");
614 static inline struct sk_buff
*ath6kl_wmi_get_new_buf(u32 size
)
618 skb
= ath6kl_buf_alloc(size
);
624 memset(skb
->data
, 0, size
);
629 /* Send a "simple" wmi command -- one with no arguments */
630 static int ath6kl_wmi_simple_cmd(struct wmi
*wmi
, enum wmi_cmd_id cmd_id
)
635 skb
= ath6kl_wmi_get_new_buf(0);
639 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, cmd_id
, NO_SYNC_WMIFLAG
);
644 static int ath6kl_wmi_ready_event_rx(struct wmi
*wmi
, u8
*datap
, int len
)
646 struct wmi_ready_event_2
*ev
= (struct wmi_ready_event_2
*) datap
;
648 if (len
< sizeof(struct wmi_ready_event_2
))
652 ath6kl_ready_event(wmi
->parent_dev
, ev
->mac_addr
,
653 le32_to_cpu(ev
->sw_version
),
654 le32_to_cpu(ev
->abi_version
));
659 static int ath6kl_wmi_connect_event_rx(struct wmi
*wmi
, u8
*datap
, int len
)
661 struct wmi_connect_event
*ev
;
664 if (len
< sizeof(struct wmi_connect_event
))
667 ev
= (struct wmi_connect_event
*) datap
;
669 ath6kl_dbg(ATH6KL_DBG_WMI
, "%s: freq %d bssid %pM\n",
670 __func__
, ev
->ch
, ev
->bssid
);
672 /* Start of assoc rsp IEs */
673 pie
= ev
->assoc_info
+ ev
->beacon_ie_len
+
674 ev
->assoc_req_len
+ (sizeof(u16
) * 3); /* capinfo, status, aid */
676 /* End of assoc rsp IEs */
677 peie
= ev
->assoc_info
+ ev
->beacon_ie_len
+ ev
->assoc_req_len
+
682 case WLAN_EID_VENDOR_SPECIFIC
:
683 if (pie
[1] > 3 && pie
[2] == 0x00 && pie
[3] == 0x50 &&
684 pie
[4] == 0xf2 && pie
[5] == WMM_OUI_TYPE
) {
685 /* WMM OUT (00:50:F2) */
687 && pie
[6] == WMM_PARAM_OUI_SUBTYPE
)
688 wmi
->is_wmm_enabled
= true;
693 if (wmi
->is_wmm_enabled
)
699 ath6kl_connect_event(wmi
->parent_dev
, le16_to_cpu(ev
->ch
), ev
->bssid
,
700 le16_to_cpu(ev
->listen_intvl
),
701 le16_to_cpu(ev
->beacon_intvl
),
702 le32_to_cpu(ev
->nw_type
),
703 ev
->beacon_ie_len
, ev
->assoc_req_len
,
704 ev
->assoc_resp_len
, ev
->assoc_info
);
709 static int ath6kl_wmi_disconnect_event_rx(struct wmi
*wmi
, u8
*datap
, int len
)
711 struct wmi_disconnect_event
*ev
;
712 wmi
->traffic_class
= 100;
714 if (len
< sizeof(struct wmi_disconnect_event
))
717 ev
= (struct wmi_disconnect_event
*) datap
;
719 wmi
->is_wmm_enabled
= false;
720 wmi
->pair_crypto_type
= NONE_CRYPT
;
721 wmi
->grp_crypto_type
= NONE_CRYPT
;
723 ath6kl_disconnect_event(wmi
->parent_dev
, ev
->disconn_reason
,
724 ev
->bssid
, ev
->assoc_resp_len
, ev
->assoc_info
,
725 le16_to_cpu(ev
->proto_reason_status
));
730 static int ath6kl_wmi_peer_node_event_rx(struct wmi
*wmi
, u8
*datap
, int len
)
732 struct wmi_peer_node_event
*ev
;
734 if (len
< sizeof(struct wmi_peer_node_event
))
737 ev
= (struct wmi_peer_node_event
*) datap
;
739 if (ev
->event_code
== PEER_NODE_JOIN_EVENT
)
740 ath6kl_dbg(ATH6KL_DBG_WMI
, "joined node with mac addr: %pM\n",
742 else if (ev
->event_code
== PEER_NODE_LEAVE_EVENT
)
743 ath6kl_dbg(ATH6KL_DBG_WMI
, "left node with mac addr: %pM\n",
749 static int ath6kl_wmi_tkip_micerr_event_rx(struct wmi
*wmi
, u8
*datap
, int len
)
751 struct wmi_tkip_micerr_event
*ev
;
753 if (len
< sizeof(struct wmi_tkip_micerr_event
))
756 ev
= (struct wmi_tkip_micerr_event
*) datap
;
758 ath6kl_tkip_micerr_event(wmi
->parent_dev
, ev
->key_id
, ev
->is_mcast
);
763 static int ath6kl_wlan_parse_beacon(u8
*buf
, int frame_len
,
764 struct ath6kl_common_ie
*cie
)
767 u8 elemid_ssid
= false;
770 efrm
= (u8
*) (frm
+ frame_len
);
773 * beacon/probe response frame format
775 * [2] beacon interval
776 * [2] capability information
778 * [tlv] supported rates
779 * [tlv] country information
780 * [tlv] parameter set (FH/DS)
781 * [tlv] erp information
782 * [tlv] extended supported rates
785 * [tlv] Atheros Advanced Capabilities
787 if ((efrm
- frm
) < 12)
790 memset(cie
, 0, sizeof(*cie
));
792 cie
->ie_tstamp
= frm
;
794 cie
->ie_beaconInt
= *(u16
*) frm
;
796 cie
->ie_capInfo
= *(u16
*) frm
;
808 case WLAN_EID_SUPP_RATES
:
811 case WLAN_EID_COUNTRY
:
812 cie
->ie_country
= frm
;
814 case WLAN_EID_FH_PARAMS
:
816 case WLAN_EID_DS_PARAMS
:
817 cie
->ie_chan
= frm
[2];
822 case WLAN_EID_IBSS_PARAMS
:
824 case WLAN_EID_EXT_SUPP_RATES
:
825 cie
->ie_xrates
= frm
;
827 case WLAN_EID_ERP_INFO
:
831 cie
->ie_erp
= frm
[2];
836 case WLAN_EID_HT_CAPABILITY
:
839 case WLAN_EID_HT_INFORMATION
:
842 case WLAN_EID_VENDOR_SPECIFIC
:
843 if (frm
[1] > 3 && frm
[2] == 0x00 && frm
[3] == 0x50 &&
845 /* OUT Type (00:50:F2) */
847 if (frm
[5] == WPA_OUI_TYPE
) {
850 } else if (frm
[5] == WMM_OUI_TYPE
) {
853 } else if (frm
[5] == WSC_OUT_TYPE
) {
858 } else if (frm
[1] > 3 && frm
[2] == 0x00
859 && frm
[3] == 0x03 && frm
[4] == 0x7f
860 && frm
[5] == ATH_OUI_TYPE
) {
861 /* Atheros OUI (00:03:7f) */
871 if ((cie
->ie_rates
== NULL
)
872 || (cie
->ie_rates
[1] > ATH6KL_RATE_MAXSIZE
))
875 if ((cie
->ie_ssid
== NULL
)
876 || (cie
->ie_ssid
[1] > IEEE80211_MAX_SSID_LEN
))
882 static int ath6kl_wmi_bssinfo_event_rx(struct wmi
*wmi
, u8
*datap
, int len
)
884 struct bss
*bss
= NULL
;
885 struct wmi_bss_info_hdr
*bih
;
886 u8 cached_ssid_len
= 0;
887 u8 cached_ssid
[IEEE80211_MAX_SSID_LEN
] = { 0 };
888 u8 beacon_ssid_len
= 0;
895 if (len
<= sizeof(struct wmi_bss_info_hdr
))
898 bih
= (struct wmi_bss_info_hdr
*) datap
;
899 bss
= wlan_find_node(&wmi
->parent_dev
->scan_table
, bih
->bssid
);
901 if (a_sle16_to_cpu(bih
->rssi
) > 0) {
905 bih
->rssi
= a_cpu_to_sle16(bss
->ni_rssi
);
908 buf
= datap
+ sizeof(struct wmi_bss_info_hdr
);
909 len
-= sizeof(struct wmi_bss_info_hdr
);
911 ath6kl_dbg(ATH6KL_DBG_WMI
,
912 "bss info evt - ch %u, rssi %02x, bssid \"%pM\"\n",
913 bih
->ch
, a_sle16_to_cpu(bih
->rssi
), bih
->bssid
);
917 * Free up the node. We are about to allocate a new node.
918 * In case of hidden AP, beacon will not have ssid,
919 * but a directed probe response will have it,
920 * so cache the probe-resp-ssid if already present.
922 if (wmi
->is_probe_ssid
&& (bih
->frame_type
== BEACON_FTYPE
)) {
923 ie_ssid
= bss
->ni_cie
.ie_ssid
;
924 if (ie_ssid
&& (ie_ssid
[1] <= IEEE80211_MAX_SSID_LEN
) &&
926 cached_ssid_len
= ie_ssid
[1];
927 memcpy(cached_ssid
, ie_ssid
+ 2,
933 * Use the current average rssi of associated AP base on
935 * 1. Most os with GUI will update RSSI by
936 * ath6kl_wmi_get_stats_cmd() periodically.
937 * 2. ath6kl_wmi_get_stats_cmd(..) will be called when calling
938 * ath6kl_wmi_startscan_cmd(...)
939 * The average value of RSSI give end-user better feeling for
940 * instance value of scan result. It also sync up RSSI info
941 * in GUI between scan result and RSSI signal icon.
943 if (memcmp(wmi
->parent_dev
->bssid
, bih
->bssid
, ETH_ALEN
) == 0) {
944 bih
->rssi
= a_cpu_to_sle16(bss
->ni_rssi
);
945 bih
->snr
= bss
->ni_snr
;
948 wlan_node_reclaim(&wmi
->parent_dev
->scan_table
, bss
);
952 * beacon/probe response frame format
954 * [2] beacon interval
955 * [2] capability information
958 beacon_ssid_len
= buf
[SSID_IE_LEN_INDEX
];
961 * If ssid is cached for this hidden AP, then change
962 * buffer len accordingly.
964 if (wmi
->is_probe_ssid
&& (bih
->frame_type
== BEACON_FTYPE
) &&
965 (cached_ssid_len
!= 0) &&
966 (beacon_ssid_len
== 0 || (cached_ssid_len
> beacon_ssid_len
&&
967 buf
[SSID_IE_LEN_INDEX
+ 1] == 0))) {
969 len
+= (cached_ssid_len
- beacon_ssid_len
);
972 bss
= wlan_node_alloc(len
);
976 bss
->ni_snr
= bih
->snr
;
977 bss
->ni_rssi
= a_sle16_to_cpu(bih
->rssi
);
979 if (WARN_ON(!bss
->ni_buf
))
983 * In case of hidden AP, beacon will not have ssid,
984 * but a directed probe response will have it,
985 * so place the cached-ssid(probe-resp) in the bss info.
987 if (wmi
->is_probe_ssid
&& (bih
->frame_type
== BEACON_FTYPE
) &&
988 (cached_ssid_len
!= 0) &&
989 (beacon_ssid_len
== 0 || (beacon_ssid_len
&&
990 buf
[SSID_IE_LEN_INDEX
+ 1] == 0))) {
991 ni_buf
= bss
->ni_buf
;
995 * Copy the first 14 bytes:
996 * time-stamp(8), beacon-interval(2),
997 * cap-info(2), ssid-id(1), ssid-len(1).
999 memcpy(ni_buf
, buf
, SSID_IE_LEN_INDEX
+ 1);
1001 ni_buf
[SSID_IE_LEN_INDEX
] = cached_ssid_len
;
1002 ni_buf
+= (SSID_IE_LEN_INDEX
+ 1);
1004 buf
+= (SSID_IE_LEN_INDEX
+ 1);
1005 buf_len
-= (SSID_IE_LEN_INDEX
+ 1);
1007 memcpy(ni_buf
, cached_ssid
, cached_ssid_len
);
1008 ni_buf
+= cached_ssid_len
;
1010 buf
+= beacon_ssid_len
;
1011 buf_len
-= beacon_ssid_len
;
1013 if (cached_ssid_len
> beacon_ssid_len
)
1014 buf_len
-= (cached_ssid_len
- beacon_ssid_len
);
1016 memcpy(ni_buf
, buf
, buf_len
);
1018 memcpy(bss
->ni_buf
, buf
, len
);
1020 bss
->ni_framelen
= len
;
1022 ret
= ath6kl_wlan_parse_beacon(bss
->ni_buf
, len
, &bss
->ni_cie
);
1024 wlan_node_free(bss
);
1029 * Update the frequency in ie_chan, overwriting of channel number
1030 * which is done in ath6kl_wlan_parse_beacon
1032 bss
->ni_cie
.ie_chan
= le16_to_cpu(bih
->ch
);
1033 wlan_setup_node(&wmi
->parent_dev
->scan_table
, bss
, bih
->bssid
);
1038 static int ath6kl_wmi_opt_frame_event_rx(struct wmi
*wmi
, u8
*datap
, int len
)
1041 struct wmi_opt_rx_info_hdr
*bih
;
1044 if (len
<= sizeof(struct wmi_opt_rx_info_hdr
))
1047 bih
= (struct wmi_opt_rx_info_hdr
*) datap
;
1048 buf
= datap
+ sizeof(struct wmi_opt_rx_info_hdr
);
1049 len
-= sizeof(struct wmi_opt_rx_info_hdr
);
1051 ath6kl_dbg(ATH6KL_DBG_WMI
, "opt frame event %2.2x:%2.2x\n",
1052 bih
->bssid
[4], bih
->bssid
[5]);
1054 bss
= wlan_find_node(&wmi
->parent_dev
->scan_table
, bih
->bssid
);
1056 /* Free up the node. We are about to allocate a new node. */
1057 wlan_node_reclaim(&wmi
->parent_dev
->scan_table
, bss
);
1060 bss
= wlan_node_alloc(len
);
1064 bss
->ni_snr
= bih
->snr
;
1065 bss
->ni_cie
.ie_chan
= le16_to_cpu(bih
->ch
);
1067 if (WARN_ON(!bss
->ni_buf
))
1070 memcpy(bss
->ni_buf
, buf
, len
);
1071 wlan_setup_node(&wmi
->parent_dev
->scan_table
, bss
, bih
->bssid
);
1076 /* Inactivity timeout of a fatpipe(pstream) at the target */
1077 static int ath6kl_wmi_pstream_timeout_event_rx(struct wmi
*wmi
, u8
*datap
,
1080 struct wmi_pstream_timeout_event
*ev
;
1082 if (len
< sizeof(struct wmi_pstream_timeout_event
))
1085 ev
= (struct wmi_pstream_timeout_event
*) datap
;
1088 * When the pstream (fat pipe == AC) timesout, it means there were
1089 * no thinStreams within this pstream & it got implicitly created
1090 * due to data flow on this AC. We start the inactivity timer only
1091 * for implicitly created pstream. Just reset the host state.
1093 spin_lock_bh(&wmi
->lock
);
1094 wmi
->stream_exist_for_ac
[ev
->traffic_class
] = 0;
1095 wmi
->fat_pipe_exist
&= ~(1 << ev
->traffic_class
);
1096 spin_unlock_bh(&wmi
->lock
);
1098 /* Indicate inactivity to driver layer for this fatpipe (pstream) */
1099 ath6kl_indicate_tx_activity(wmi
->parent_dev
, ev
->traffic_class
, false);
1104 static int ath6kl_wmi_bitrate_reply_rx(struct wmi
*wmi
, u8
*datap
, int len
)
1106 struct wmi_bit_rate_reply
*reply
;
1110 if (len
< sizeof(struct wmi_bit_rate_reply
))
1113 reply
= (struct wmi_bit_rate_reply
*) datap
;
1115 ath6kl_dbg(ATH6KL_DBG_WMI
, "rateindex %d\n", reply
->rate_index
);
1117 if (reply
->rate_index
== (s8
) RATE_AUTO
) {
1120 index
= reply
->rate_index
& 0x7f;
1121 sgi
= (reply
->rate_index
& 0x80) ? 1 : 0;
1122 rate
= wmi_rate_tbl
[index
][sgi
];
1125 ath6kl_wakeup_event(wmi
->parent_dev
);
1130 static int ath6kl_wmi_ratemask_reply_rx(struct wmi
*wmi
, u8
*datap
, int len
)
1132 if (len
< sizeof(struct wmi_fix_rates_reply
))
1135 ath6kl_wakeup_event(wmi
->parent_dev
);
1140 static int ath6kl_wmi_ch_list_reply_rx(struct wmi
*wmi
, u8
*datap
, int len
)
1142 if (len
< sizeof(struct wmi_channel_list_reply
))
1145 ath6kl_wakeup_event(wmi
->parent_dev
);
1150 static int ath6kl_wmi_tx_pwr_reply_rx(struct wmi
*wmi
, u8
*datap
, int len
)
1152 struct wmi_tx_pwr_reply
*reply
;
1154 if (len
< sizeof(struct wmi_tx_pwr_reply
))
1157 reply
= (struct wmi_tx_pwr_reply
*) datap
;
1158 ath6kl_txpwr_rx_evt(wmi
->parent_dev
, reply
->dbM
);
1163 static int ath6kl_wmi_keepalive_reply_rx(struct wmi
*wmi
, u8
*datap
, int len
)
1165 if (len
< sizeof(struct wmi_get_keepalive_cmd
))
1168 ath6kl_wakeup_event(wmi
->parent_dev
);
1173 static int ath6kl_wmi_scan_complete_rx(struct wmi
*wmi
, u8
*datap
, int len
)
1175 struct wmi_scan_complete_event
*ev
;
1177 ev
= (struct wmi_scan_complete_event
*) datap
;
1179 if (a_sle32_to_cpu(ev
->status
) == 0)
1180 wlan_refresh_inactive_nodes(wmi
->parent_dev
);
1182 ath6kl_scan_complete_evt(wmi
->parent_dev
, a_sle32_to_cpu(ev
->status
));
1183 wmi
->is_probe_ssid
= false;
1189 * Target is reporting a programming error. This is for
1190 * developer aid only. Target only checks a few common violations
1191 * and it is responsibility of host to do all error checking.
1192 * Behavior of target after wmi error event is undefined.
1193 * A reset is recommended.
1195 static int ath6kl_wmi_error_event_rx(struct wmi
*wmi
, u8
*datap
, int len
)
1197 const char *type
= "unknown error";
1198 struct wmi_cmd_error_event
*ev
;
1199 ev
= (struct wmi_cmd_error_event
*) datap
;
1201 switch (ev
->err_code
) {
1203 type
= "invalid parameter";
1206 type
= "invalid state";
1208 case INTERNAL_ERROR
:
1209 type
= "internal error";
1213 ath6kl_dbg(ATH6KL_DBG_WMI
, "programming error, cmd=%d %s\n",
1219 static int ath6kl_wmi_stats_event_rx(struct wmi
*wmi
, u8
*datap
, int len
)
1221 ath6kl_tgt_stats_event(wmi
->parent_dev
, datap
, len
);
1226 static u8
ath6kl_wmi_get_upper_threshold(s16 rssi
,
1227 struct sq_threshold_params
*sq_thresh
,
1231 u8 threshold
= (u8
) sq_thresh
->upper_threshold
[size
- 1];
1233 /* The list is already in sorted order. Get the next lower value */
1234 for (index
= 0; index
< size
; index
++) {
1235 if (rssi
< sq_thresh
->upper_threshold
[index
]) {
1236 threshold
= (u8
) sq_thresh
->upper_threshold
[index
];
1244 static u8
ath6kl_wmi_get_lower_threshold(s16 rssi
,
1245 struct sq_threshold_params
*sq_thresh
,
1249 u8 threshold
= (u8
) sq_thresh
->lower_threshold
[size
- 1];
1251 /* The list is already in sorted order. Get the next lower value */
1252 for (index
= 0; index
< size
; index
++) {
1253 if (rssi
> sq_thresh
->lower_threshold
[index
]) {
1254 threshold
= (u8
) sq_thresh
->lower_threshold
[index
];
1262 static int ath6kl_wmi_send_rssi_threshold_params(struct wmi
*wmi
,
1263 struct wmi_rssi_threshold_params_cmd
*rssi_cmd
)
1265 struct sk_buff
*skb
;
1266 struct wmi_rssi_threshold_params_cmd
*cmd
;
1268 skb
= ath6kl_wmi_get_new_buf(sizeof(*cmd
));
1272 cmd
= (struct wmi_rssi_threshold_params_cmd
*) skb
->data
;
1273 memcpy(cmd
, rssi_cmd
, sizeof(struct wmi_rssi_threshold_params_cmd
));
1275 return ath6kl_wmi_cmd_send(wmi
, skb
, WMI_RSSI_THRESHOLD_PARAMS_CMDID
,
1279 static int ath6kl_wmi_rssi_threshold_event_rx(struct wmi
*wmi
, u8
*datap
,
1282 struct wmi_rssi_threshold_event
*reply
;
1283 struct wmi_rssi_threshold_params_cmd cmd
;
1284 struct sq_threshold_params
*sq_thresh
;
1285 enum wmi_rssi_threshold_val new_threshold
;
1286 u8 upper_rssi_threshold
, lower_rssi_threshold
;
1290 if (len
< sizeof(struct wmi_rssi_threshold_event
))
1293 reply
= (struct wmi_rssi_threshold_event
*) datap
;
1294 new_threshold
= (enum wmi_rssi_threshold_val
) reply
->range
;
1295 rssi
= a_sle16_to_cpu(reply
->rssi
);
1297 sq_thresh
= &wmi
->sq_threshld
[SIGNAL_QUALITY_METRICS_RSSI
];
1300 * Identify the threshold breached and communicate that to the app.
1301 * After that install a new set of thresholds based on the signal
1302 * quality reported by the target
1304 if (new_threshold
) {
1305 /* Upper threshold breached */
1306 if (rssi
< sq_thresh
->upper_threshold
[0]) {
1307 ath6kl_dbg(ATH6KL_DBG_WMI
,
1308 "spurious upper rssi threshold event: %d\n",
1310 } else if ((rssi
< sq_thresh
->upper_threshold
[1]) &&
1311 (rssi
>= sq_thresh
->upper_threshold
[0])) {
1312 new_threshold
= WMI_RSSI_THRESHOLD1_ABOVE
;
1313 } else if ((rssi
< sq_thresh
->upper_threshold
[2]) &&
1314 (rssi
>= sq_thresh
->upper_threshold
[1])) {
1315 new_threshold
= WMI_RSSI_THRESHOLD2_ABOVE
;
1316 } else if ((rssi
< sq_thresh
->upper_threshold
[3]) &&
1317 (rssi
>= sq_thresh
->upper_threshold
[2])) {
1318 new_threshold
= WMI_RSSI_THRESHOLD3_ABOVE
;
1319 } else if ((rssi
< sq_thresh
->upper_threshold
[4]) &&
1320 (rssi
>= sq_thresh
->upper_threshold
[3])) {
1321 new_threshold
= WMI_RSSI_THRESHOLD4_ABOVE
;
1322 } else if ((rssi
< sq_thresh
->upper_threshold
[5]) &&
1323 (rssi
>= sq_thresh
->upper_threshold
[4])) {
1324 new_threshold
= WMI_RSSI_THRESHOLD5_ABOVE
;
1325 } else if (rssi
>= sq_thresh
->upper_threshold
[5]) {
1326 new_threshold
= WMI_RSSI_THRESHOLD6_ABOVE
;
1329 /* Lower threshold breached */
1330 if (rssi
> sq_thresh
->lower_threshold
[0]) {
1331 ath6kl_dbg(ATH6KL_DBG_WMI
,
1332 "spurious lower rssi threshold event: %d %d\n",
1333 rssi
, sq_thresh
->lower_threshold
[0]);
1334 } else if ((rssi
> sq_thresh
->lower_threshold
[1]) &&
1335 (rssi
<= sq_thresh
->lower_threshold
[0])) {
1336 new_threshold
= WMI_RSSI_THRESHOLD6_BELOW
;
1337 } else if ((rssi
> sq_thresh
->lower_threshold
[2]) &&
1338 (rssi
<= sq_thresh
->lower_threshold
[1])) {
1339 new_threshold
= WMI_RSSI_THRESHOLD5_BELOW
;
1340 } else if ((rssi
> sq_thresh
->lower_threshold
[3]) &&
1341 (rssi
<= sq_thresh
->lower_threshold
[2])) {
1342 new_threshold
= WMI_RSSI_THRESHOLD4_BELOW
;
1343 } else if ((rssi
> sq_thresh
->lower_threshold
[4]) &&
1344 (rssi
<= sq_thresh
->lower_threshold
[3])) {
1345 new_threshold
= WMI_RSSI_THRESHOLD3_BELOW
;
1346 } else if ((rssi
> sq_thresh
->lower_threshold
[5]) &&
1347 (rssi
<= sq_thresh
->lower_threshold
[4])) {
1348 new_threshold
= WMI_RSSI_THRESHOLD2_BELOW
;
1349 } else if (rssi
<= sq_thresh
->lower_threshold
[5]) {
1350 new_threshold
= WMI_RSSI_THRESHOLD1_BELOW
;
1354 /* Calculate and install the next set of thresholds */
1355 lower_rssi_threshold
= ath6kl_wmi_get_lower_threshold(rssi
, sq_thresh
,
1356 sq_thresh
->lower_threshold_valid_count
);
1357 upper_rssi_threshold
= ath6kl_wmi_get_upper_threshold(rssi
, sq_thresh
,
1358 sq_thresh
->upper_threshold_valid_count
);
1360 /* Issue a wmi command to install the thresholds */
1361 cmd
.thresh_above1_val
= a_cpu_to_sle16(upper_rssi_threshold
);
1362 cmd
.thresh_below1_val
= a_cpu_to_sle16(lower_rssi_threshold
);
1363 cmd
.weight
= sq_thresh
->weight
;
1364 cmd
.poll_time
= cpu_to_le32(sq_thresh
->polling_interval
);
1366 ret
= ath6kl_wmi_send_rssi_threshold_params(wmi
, &cmd
);
1368 ath6kl_err("unable to configure rssi thresholds\n");
1375 static int ath6kl_wmi_cac_event_rx(struct wmi
*wmi
, u8
*datap
, int len
)
1377 struct wmi_cac_event
*reply
;
1378 struct ieee80211_tspec_ie
*ts
;
1379 u16 active_tsids
, tsinfo
;
1383 if (len
< sizeof(struct wmi_cac_event
))
1386 reply
= (struct wmi_cac_event
*) datap
;
1388 if ((reply
->cac_indication
== CAC_INDICATION_ADMISSION_RESP
) &&
1389 (reply
->status_code
!= IEEE80211_TSPEC_STATUS_ADMISS_ACCEPTED
)) {
1391 ts
= (struct ieee80211_tspec_ie
*) &(reply
->tspec_suggestion
);
1392 tsinfo
= le16_to_cpu(ts
->tsinfo
);
1393 tsid
= (tsinfo
>> IEEE80211_WMM_IE_TSPEC_TID_SHIFT
) &
1394 IEEE80211_WMM_IE_TSPEC_TID_MASK
;
1396 ath6kl_wmi_delete_pstream_cmd(wmi
, reply
->ac
, tsid
);
1397 } else if (reply
->cac_indication
== CAC_INDICATION_NO_RESP
) {
1399 * Following assumes that there is only one outstanding
1400 * ADDTS request when this event is received
1402 spin_lock_bh(&wmi
->lock
);
1403 active_tsids
= wmi
->stream_exist_for_ac
[reply
->ac
];
1404 spin_unlock_bh(&wmi
->lock
);
1406 for (index
= 0; index
< sizeof(active_tsids
) * 8; index
++) {
1407 if ((active_tsids
>> index
) & 1)
1410 if (index
< (sizeof(active_tsids
) * 8))
1411 ath6kl_wmi_delete_pstream_cmd(wmi
, reply
->ac
, index
);
1415 * Clear active tsids and Add missing handling
1416 * for delete qos stream from AP
1418 else if (reply
->cac_indication
== CAC_INDICATION_DELETE
) {
1420 ts
= (struct ieee80211_tspec_ie
*) &(reply
->tspec_suggestion
);
1421 tsinfo
= le16_to_cpu(ts
->tsinfo
);
1422 ts_id
= ((tsinfo
>> IEEE80211_WMM_IE_TSPEC_TID_SHIFT
) &
1423 IEEE80211_WMM_IE_TSPEC_TID_MASK
);
1425 spin_lock_bh(&wmi
->lock
);
1426 wmi
->stream_exist_for_ac
[reply
->ac
] &= ~(1 << ts_id
);
1427 active_tsids
= wmi
->stream_exist_for_ac
[reply
->ac
];
1428 spin_unlock_bh(&wmi
->lock
);
1430 /* Indicate stream inactivity to driver layer only if all tsids
1431 * within this AC are deleted.
1433 if (!active_tsids
) {
1434 ath6kl_indicate_tx_activity(wmi
->parent_dev
, reply
->ac
,
1436 wmi
->fat_pipe_exist
&= ~(1 << reply
->ac
);
1443 static int ath6kl_wmi_send_snr_threshold_params(struct wmi
*wmi
,
1444 struct wmi_snr_threshold_params_cmd
*snr_cmd
)
1446 struct sk_buff
*skb
;
1447 struct wmi_snr_threshold_params_cmd
*cmd
;
1449 skb
= ath6kl_wmi_get_new_buf(sizeof(*cmd
));
1453 cmd
= (struct wmi_snr_threshold_params_cmd
*) skb
->data
;
1454 memcpy(cmd
, snr_cmd
, sizeof(struct wmi_snr_threshold_params_cmd
));
1456 return ath6kl_wmi_cmd_send(wmi
, skb
, WMI_SNR_THRESHOLD_PARAMS_CMDID
,
1460 static int ath6kl_wmi_snr_threshold_event_rx(struct wmi
*wmi
, u8
*datap
,
1463 struct wmi_snr_threshold_event
*reply
;
1464 struct sq_threshold_params
*sq_thresh
;
1465 struct wmi_snr_threshold_params_cmd cmd
;
1466 enum wmi_snr_threshold_val new_threshold
;
1467 u8 upper_snr_threshold
, lower_snr_threshold
;
1471 if (len
< sizeof(struct wmi_snr_threshold_event
))
1474 reply
= (struct wmi_snr_threshold_event
*) datap
;
1476 new_threshold
= (enum wmi_snr_threshold_val
) reply
->range
;
1479 sq_thresh
= &wmi
->sq_threshld
[SIGNAL_QUALITY_METRICS_SNR
];
1482 * Identify the threshold breached and communicate that to the app.
1483 * After that install a new set of thresholds based on the signal
1484 * quality reported by the target.
1486 if (new_threshold
) {
1487 /* Upper threshold breached */
1488 if (snr
< sq_thresh
->upper_threshold
[0]) {
1489 ath6kl_dbg(ATH6KL_DBG_WMI
,
1490 "spurious upper snr threshold event: %d\n",
1492 } else if ((snr
< sq_thresh
->upper_threshold
[1]) &&
1493 (snr
>= sq_thresh
->upper_threshold
[0])) {
1494 new_threshold
= WMI_SNR_THRESHOLD1_ABOVE
;
1495 } else if ((snr
< sq_thresh
->upper_threshold
[2]) &&
1496 (snr
>= sq_thresh
->upper_threshold
[1])) {
1497 new_threshold
= WMI_SNR_THRESHOLD2_ABOVE
;
1498 } else if ((snr
< sq_thresh
->upper_threshold
[3]) &&
1499 (snr
>= sq_thresh
->upper_threshold
[2])) {
1500 new_threshold
= WMI_SNR_THRESHOLD3_ABOVE
;
1501 } else if (snr
>= sq_thresh
->upper_threshold
[3]) {
1502 new_threshold
= WMI_SNR_THRESHOLD4_ABOVE
;
1505 /* Lower threshold breached */
1506 if (snr
> sq_thresh
->lower_threshold
[0]) {
1507 ath6kl_dbg(ATH6KL_DBG_WMI
,
1508 "spurious lower snr threshold event: %d\n",
1509 sq_thresh
->lower_threshold
[0]);
1510 } else if ((snr
> sq_thresh
->lower_threshold
[1]) &&
1511 (snr
<= sq_thresh
->lower_threshold
[0])) {
1512 new_threshold
= WMI_SNR_THRESHOLD4_BELOW
;
1513 } else if ((snr
> sq_thresh
->lower_threshold
[2]) &&
1514 (snr
<= sq_thresh
->lower_threshold
[1])) {
1515 new_threshold
= WMI_SNR_THRESHOLD3_BELOW
;
1516 } else if ((snr
> sq_thresh
->lower_threshold
[3]) &&
1517 (snr
<= sq_thresh
->lower_threshold
[2])) {
1518 new_threshold
= WMI_SNR_THRESHOLD2_BELOW
;
1519 } else if (snr
<= sq_thresh
->lower_threshold
[3]) {
1520 new_threshold
= WMI_SNR_THRESHOLD1_BELOW
;
1524 /* Calculate and install the next set of thresholds */
1525 lower_snr_threshold
= ath6kl_wmi_get_lower_threshold(snr
, sq_thresh
,
1526 sq_thresh
->lower_threshold_valid_count
);
1527 upper_snr_threshold
= ath6kl_wmi_get_upper_threshold(snr
, sq_thresh
,
1528 sq_thresh
->upper_threshold_valid_count
);
1530 /* Issue a wmi command to install the thresholds */
1531 cmd
.thresh_above1_val
= upper_snr_threshold
;
1532 cmd
.thresh_below1_val
= lower_snr_threshold
;
1533 cmd
.weight
= sq_thresh
->weight
;
1534 cmd
.poll_time
= cpu_to_le32(sq_thresh
->polling_interval
);
1536 ath6kl_dbg(ATH6KL_DBG_WMI
,
1537 "snr: %d, threshold: %d, lower: %d, upper: %d\n",
1539 lower_snr_threshold
, upper_snr_threshold
);
1541 ret
= ath6kl_wmi_send_snr_threshold_params(wmi
, &cmd
);
1543 ath6kl_err("unable to configure snr threshold\n");
1550 static int ath6kl_wmi_aplist_event_rx(struct wmi
*wmi
, u8
*datap
, int len
)
1552 u16 ap_info_entry_size
;
1553 struct wmi_aplist_event
*ev
= (struct wmi_aplist_event
*) datap
;
1554 struct wmi_ap_info_v1
*ap_info_v1
;
1557 if (len
< sizeof(struct wmi_aplist_event
) ||
1558 ev
->ap_list_ver
!= APLIST_VER1
)
1561 ap_info_entry_size
= sizeof(struct wmi_ap_info_v1
);
1562 ap_info_v1
= (struct wmi_ap_info_v1
*) ev
->ap_list
;
1564 ath6kl_dbg(ATH6KL_DBG_WMI
,
1565 "number of APs in aplist event: %d\n", ev
->num_ap
);
1567 if (len
< (int) (sizeof(struct wmi_aplist_event
) +
1568 (ev
->num_ap
- 1) * ap_info_entry_size
))
1571 /* AP list version 1 contents */
1572 for (index
= 0; index
< ev
->num_ap
; index
++) {
1573 ath6kl_dbg(ATH6KL_DBG_WMI
, "AP#%d BSSID %pM Channel %d\n",
1574 index
, ap_info_v1
->bssid
, ap_info_v1
->channel
);
1581 int ath6kl_wmi_cmd_send(struct wmi
*wmi
, struct sk_buff
*skb
,
1582 enum wmi_cmd_id cmd_id
, enum wmi_sync_flag sync_flag
)
1584 struct wmi_cmd_hdr
*cmd_hdr
;
1585 enum htc_endpoint_id ep_id
= wmi
->ep_id
;
1588 ath6kl_dbg(ATH6KL_DBG_WMI
, "%s: cmd_id=%d\n", __func__
, cmd_id
);
1590 if (WARN_ON(skb
== NULL
))
1593 if (sync_flag
>= END_WMIFLAG
) {
1598 if ((sync_flag
== SYNC_BEFORE_WMIFLAG
) ||
1599 (sync_flag
== SYNC_BOTH_WMIFLAG
)) {
1601 * Make sure all data currently queued is transmitted before
1602 * the cmd execution. Establish a new sync point.
1604 ath6kl_wmi_sync_point(wmi
);
1607 skb_push(skb
, sizeof(struct wmi_cmd_hdr
));
1609 cmd_hdr
= (struct wmi_cmd_hdr
*) skb
->data
;
1610 cmd_hdr
->cmd_id
= cpu_to_le16(cmd_id
);
1611 cmd_hdr
->info1
= 0; /* added for virtual interface */
1613 /* Only for OPT_TX_CMD, use BE endpoint. */
1614 if (cmd_id
== WMI_OPT_TX_FRAME_CMDID
) {
1615 ret
= ath6kl_wmi_data_hdr_add(wmi
, skb
, OPT_MSGTYPE
,
1616 false, false, 0, NULL
);
1621 ep_id
= ath6kl_ac2_endpoint_id(wmi
->parent_dev
, WMM_AC_BE
);
1624 ath6kl_control_tx(wmi
->parent_dev
, skb
, ep_id
);
1626 if ((sync_flag
== SYNC_AFTER_WMIFLAG
) ||
1627 (sync_flag
== SYNC_BOTH_WMIFLAG
)) {
1629 * Make sure all new data queued waits for the command to
1630 * execute. Establish a new sync point.
1632 ath6kl_wmi_sync_point(wmi
);
1638 int ath6kl_wmi_connect_cmd(struct wmi
*wmi
, enum network_type nw_type
,
1639 enum dot11_auth_mode dot11_auth_mode
,
1640 enum auth_mode auth_mode
,
1641 enum crypto_type pairwise_crypto
,
1642 u8 pairwise_crypto_len
,
1643 enum crypto_type group_crypto
,
1644 u8 group_crypto_len
, int ssid_len
, u8
*ssid
,
1645 u8
*bssid
, u16 channel
, u32 ctrl_flags
)
1647 struct sk_buff
*skb
;
1648 struct wmi_connect_cmd
*cc
;
1651 wmi
->traffic_class
= 100;
1653 if ((pairwise_crypto
== NONE_CRYPT
) && (group_crypto
!= NONE_CRYPT
))
1656 if ((pairwise_crypto
!= NONE_CRYPT
) && (group_crypto
== NONE_CRYPT
))
1659 skb
= ath6kl_wmi_get_new_buf(sizeof(struct wmi_connect_cmd
));
1663 cc
= (struct wmi_connect_cmd
*) skb
->data
;
1666 memcpy(cc
->ssid
, ssid
, ssid_len
);
1668 cc
->ssid_len
= ssid_len
;
1669 cc
->nw_type
= nw_type
;
1670 cc
->dot11_auth_mode
= dot11_auth_mode
;
1671 cc
->auth_mode
= auth_mode
;
1672 cc
->prwise_crypto_type
= pairwise_crypto
;
1673 cc
->prwise_crypto_len
= pairwise_crypto_len
;
1674 cc
->grp_crypto_type
= group_crypto
;
1675 cc
->grp_crypto_len
= group_crypto_len
;
1676 cc
->ch
= cpu_to_le16(channel
);
1677 cc
->ctrl_flags
= cpu_to_le32(ctrl_flags
);
1680 memcpy(cc
->bssid
, bssid
, ETH_ALEN
);
1682 wmi
->pair_crypto_type
= pairwise_crypto
;
1683 wmi
->grp_crypto_type
= group_crypto
;
1685 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_CONNECT_CMDID
, NO_SYNC_WMIFLAG
);
1690 int ath6kl_wmi_reconnect_cmd(struct wmi
*wmi
, u8
*bssid
, u16 channel
)
1692 struct sk_buff
*skb
;
1693 struct wmi_reconnect_cmd
*cc
;
1696 wmi
->traffic_class
= 100;
1698 skb
= ath6kl_wmi_get_new_buf(sizeof(struct wmi_reconnect_cmd
));
1702 cc
= (struct wmi_reconnect_cmd
*) skb
->data
;
1703 cc
->channel
= cpu_to_le16(channel
);
1706 memcpy(cc
->bssid
, bssid
, ETH_ALEN
);
1708 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_RECONNECT_CMDID
,
1714 int ath6kl_wmi_disconnect_cmd(struct wmi
*wmi
)
1718 wmi
->traffic_class
= 100;
1720 /* Disconnect command does not need to do a SYNC before. */
1721 ret
= ath6kl_wmi_simple_cmd(wmi
, WMI_DISCONNECT_CMDID
);
1726 int ath6kl_wmi_startscan_cmd(struct wmi
*wmi
, enum wmi_scan_type scan_type
,
1727 u32 force_fgscan
, u32 is_legacy
,
1728 u32 home_dwell_time
, u32 force_scan_interval
,
1729 s8 num_chan
, u16
*ch_list
)
1731 struct sk_buff
*skb
;
1732 struct wmi_start_scan_cmd
*sc
;
1736 size
= sizeof(struct wmi_start_scan_cmd
);
1738 if ((scan_type
!= WMI_LONG_SCAN
) && (scan_type
!= WMI_SHORT_SCAN
))
1741 if (num_chan
> WMI_MAX_CHANNELS
)
1745 size
+= sizeof(u16
) * (num_chan
- 1);
1747 skb
= ath6kl_wmi_get_new_buf(size
);
1751 sc
= (struct wmi_start_scan_cmd
*) skb
->data
;
1752 sc
->scan_type
= scan_type
;
1753 sc
->force_fg_scan
= cpu_to_le32(force_fgscan
);
1754 sc
->is_legacy
= cpu_to_le32(is_legacy
);
1755 sc
->home_dwell_time
= cpu_to_le32(home_dwell_time
);
1756 sc
->force_scan_intvl
= cpu_to_le32(force_scan_interval
);
1757 sc
->num_ch
= num_chan
;
1759 for (i
= 0; i
< num_chan
; i
++)
1760 sc
->ch_list
[i
] = cpu_to_le16(ch_list
[i
]);
1762 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_START_SCAN_CMDID
,
1768 int ath6kl_wmi_scanparams_cmd(struct wmi
*wmi
, u16 fg_start_sec
,
1769 u16 fg_end_sec
, u16 bg_sec
,
1770 u16 minact_chdw_msec
, u16 maxact_chdw_msec
,
1771 u16 pas_chdw_msec
, u8 short_scan_ratio
,
1772 u8 scan_ctrl_flag
, u32 max_dfsch_act_time
,
1773 u16 maxact_scan_per_ssid
)
1775 struct sk_buff
*skb
;
1776 struct wmi_scan_params_cmd
*sc
;
1779 skb
= ath6kl_wmi_get_new_buf(sizeof(*sc
));
1783 sc
= (struct wmi_scan_params_cmd
*) skb
->data
;
1784 sc
->fg_start_period
= cpu_to_le16(fg_start_sec
);
1785 sc
->fg_end_period
= cpu_to_le16(fg_end_sec
);
1786 sc
->bg_period
= cpu_to_le16(bg_sec
);
1787 sc
->minact_chdwell_time
= cpu_to_le16(minact_chdw_msec
);
1788 sc
->maxact_chdwell_time
= cpu_to_le16(maxact_chdw_msec
);
1789 sc
->pas_chdwell_time
= cpu_to_le16(pas_chdw_msec
);
1790 sc
->short_scan_ratio
= short_scan_ratio
;
1791 sc
->scan_ctrl_flags
= scan_ctrl_flag
;
1792 sc
->max_dfsch_act_time
= cpu_to_le32(max_dfsch_act_time
);
1793 sc
->maxact_scan_per_ssid
= cpu_to_le16(maxact_scan_per_ssid
);
1795 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_SET_SCAN_PARAMS_CMDID
,
1800 int ath6kl_wmi_bssfilter_cmd(struct wmi
*wmi
, u8 filter
, u32 ie_mask
)
1802 struct sk_buff
*skb
;
1803 struct wmi_bss_filter_cmd
*cmd
;
1806 if (filter
>= LAST_BSS_FILTER
)
1809 skb
= ath6kl_wmi_get_new_buf(sizeof(*cmd
));
1813 cmd
= (struct wmi_bss_filter_cmd
*) skb
->data
;
1814 cmd
->bss_filter
= filter
;
1815 cmd
->ie_mask
= cpu_to_le32(ie_mask
);
1817 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_SET_BSS_FILTER_CMDID
,
1822 int ath6kl_wmi_probedssid_cmd(struct wmi
*wmi
, u8 index
, u8 flag
,
1823 u8 ssid_len
, u8
*ssid
)
1825 struct sk_buff
*skb
;
1826 struct wmi_probed_ssid_cmd
*cmd
;
1829 if (index
> MAX_PROBED_SSID_INDEX
)
1832 if (ssid_len
> sizeof(cmd
->ssid
))
1835 if ((flag
& (DISABLE_SSID_FLAG
| ANY_SSID_FLAG
)) && (ssid_len
> 0))
1838 if ((flag
& SPECIFIC_SSID_FLAG
) && !ssid_len
)
1841 if (flag
& SPECIFIC_SSID_FLAG
)
1842 wmi
->is_probe_ssid
= true;
1844 skb
= ath6kl_wmi_get_new_buf(sizeof(*cmd
));
1848 cmd
= (struct wmi_probed_ssid_cmd
*) skb
->data
;
1849 cmd
->entry_index
= index
;
1851 cmd
->ssid_len
= ssid_len
;
1852 memcpy(cmd
->ssid
, ssid
, ssid_len
);
1854 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_SET_PROBED_SSID_CMDID
,
1859 int ath6kl_wmi_listeninterval_cmd(struct wmi
*wmi
, u16 listen_interval
,
1862 struct sk_buff
*skb
;
1863 struct wmi_listen_int_cmd
*cmd
;
1866 skb
= ath6kl_wmi_get_new_buf(sizeof(*cmd
));
1870 cmd
= (struct wmi_listen_int_cmd
*) skb
->data
;
1871 cmd
->listen_intvl
= cpu_to_le16(listen_interval
);
1872 cmd
->num_beacons
= cpu_to_le16(listen_beacons
);
1874 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_SET_LISTEN_INT_CMDID
,
1879 int ath6kl_wmi_powermode_cmd(struct wmi
*wmi
, u8 pwr_mode
)
1881 struct sk_buff
*skb
;
1882 struct wmi_power_mode_cmd
*cmd
;
1885 skb
= ath6kl_wmi_get_new_buf(sizeof(*cmd
));
1889 cmd
= (struct wmi_power_mode_cmd
*) skb
->data
;
1890 cmd
->pwr_mode
= pwr_mode
;
1891 wmi
->pwr_mode
= pwr_mode
;
1893 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_SET_POWER_MODE_CMDID
,
1898 int ath6kl_wmi_pmparams_cmd(struct wmi
*wmi
, u16 idle_period
,
1899 u16 ps_poll_num
, u16 dtim_policy
,
1900 u16 tx_wakeup_policy
, u16 num_tx_to_wakeup
,
1901 u16 ps_fail_event_policy
)
1903 struct sk_buff
*skb
;
1904 struct wmi_power_params_cmd
*pm
;
1907 skb
= ath6kl_wmi_get_new_buf(sizeof(*pm
));
1911 pm
= (struct wmi_power_params_cmd
*)skb
->data
;
1912 pm
->idle_period
= cpu_to_le16(idle_period
);
1913 pm
->pspoll_number
= cpu_to_le16(ps_poll_num
);
1914 pm
->dtim_policy
= cpu_to_le16(dtim_policy
);
1915 pm
->tx_wakeup_policy
= cpu_to_le16(tx_wakeup_policy
);
1916 pm
->num_tx_to_wakeup
= cpu_to_le16(num_tx_to_wakeup
);
1917 pm
->ps_fail_event_policy
= cpu_to_le16(ps_fail_event_policy
);
1919 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_SET_POWER_PARAMS_CMDID
,
1924 int ath6kl_wmi_disctimeout_cmd(struct wmi
*wmi
, u8 timeout
)
1926 struct sk_buff
*skb
;
1927 struct wmi_disc_timeout_cmd
*cmd
;
1930 skb
= ath6kl_wmi_get_new_buf(sizeof(*cmd
));
1934 cmd
= (struct wmi_disc_timeout_cmd
*) skb
->data
;
1935 cmd
->discon_timeout
= timeout
;
1937 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_SET_DISC_TIMEOUT_CMDID
,
1942 int ath6kl_wmi_addkey_cmd(struct wmi
*wmi
, u8 key_index
,
1943 enum crypto_type key_type
,
1944 u8 key_usage
, u8 key_len
,
1945 u8
*key_rsc
, u8
*key_material
,
1946 u8 key_op_ctrl
, u8
*mac_addr
,
1947 enum wmi_sync_flag sync_flag
)
1949 struct sk_buff
*skb
;
1950 struct wmi_add_cipher_key_cmd
*cmd
;
1953 ath6kl_dbg(ATH6KL_DBG_WMI
, "addkey cmd: key_index=%u key_type=%d "
1954 "key_usage=%d key_len=%d key_op_ctrl=%d\n",
1955 key_index
, key_type
, key_usage
, key_len
, key_op_ctrl
);
1957 if ((key_index
> WMI_MAX_KEY_INDEX
) || (key_len
> WMI_MAX_KEY_LEN
) ||
1958 (key_material
== NULL
))
1961 if ((WEP_CRYPT
!= key_type
) && (NULL
== key_rsc
))
1964 skb
= ath6kl_wmi_get_new_buf(sizeof(*cmd
));
1968 cmd
= (struct wmi_add_cipher_key_cmd
*) skb
->data
;
1969 cmd
->key_index
= key_index
;
1970 cmd
->key_type
= key_type
;
1971 cmd
->key_usage
= key_usage
;
1972 cmd
->key_len
= key_len
;
1973 memcpy(cmd
->key
, key_material
, key_len
);
1975 if (key_rsc
!= NULL
)
1976 memcpy(cmd
->key_rsc
, key_rsc
, sizeof(cmd
->key_rsc
));
1978 cmd
->key_op_ctrl
= key_op_ctrl
;
1981 memcpy(cmd
->key_mac_addr
, mac_addr
, ETH_ALEN
);
1983 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_ADD_CIPHER_KEY_CMDID
,
1989 int ath6kl_wmi_add_krk_cmd(struct wmi
*wmi
, u8
*krk
)
1991 struct sk_buff
*skb
;
1992 struct wmi_add_krk_cmd
*cmd
;
1995 skb
= ath6kl_wmi_get_new_buf(sizeof(*cmd
));
1999 cmd
= (struct wmi_add_krk_cmd
*) skb
->data
;
2000 memcpy(cmd
->krk
, krk
, WMI_KRK_LEN
);
2002 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_ADD_KRK_CMDID
, NO_SYNC_WMIFLAG
);
2007 int ath6kl_wmi_deletekey_cmd(struct wmi
*wmi
, u8 key_index
)
2009 struct sk_buff
*skb
;
2010 struct wmi_delete_cipher_key_cmd
*cmd
;
2013 if (key_index
> WMI_MAX_KEY_INDEX
)
2016 skb
= ath6kl_wmi_get_new_buf(sizeof(*cmd
));
2020 cmd
= (struct wmi_delete_cipher_key_cmd
*) skb
->data
;
2021 cmd
->key_index
= key_index
;
2023 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_DELETE_CIPHER_KEY_CMDID
,
2029 int ath6kl_wmi_setpmkid_cmd(struct wmi
*wmi
, const u8
*bssid
,
2030 const u8
*pmkid
, bool set
)
2032 struct sk_buff
*skb
;
2033 struct wmi_setpmkid_cmd
*cmd
;
2039 if (set
&& pmkid
== NULL
)
2042 skb
= ath6kl_wmi_get_new_buf(sizeof(*cmd
));
2046 cmd
= (struct wmi_setpmkid_cmd
*) skb
->data
;
2047 memcpy(cmd
->bssid
, bssid
, ETH_ALEN
);
2049 memcpy(cmd
->pmkid
, pmkid
, sizeof(cmd
->pmkid
));
2050 cmd
->enable
= PMKID_ENABLE
;
2052 memset(cmd
->pmkid
, 0, sizeof(cmd
->pmkid
));
2053 cmd
->enable
= PMKID_DISABLE
;
2056 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_SET_PMKID_CMDID
,
2062 static int ath6kl_wmi_data_sync_send(struct wmi
*wmi
, struct sk_buff
*skb
,
2063 enum htc_endpoint_id ep_id
)
2065 struct wmi_data_hdr
*data_hdr
;
2068 if (WARN_ON(skb
== NULL
|| ep_id
== wmi
->ep_id
))
2071 skb_push(skb
, sizeof(struct wmi_data_hdr
));
2073 data_hdr
= (struct wmi_data_hdr
*) skb
->data
;
2074 data_hdr
->info
= SYNC_MSGTYPE
<< WMI_DATA_HDR_MSG_TYPE_SHIFT
;
2075 data_hdr
->info3
= 0;
2077 ret
= ath6kl_control_tx(wmi
->parent_dev
, skb
, ep_id
);
2082 static int ath6kl_wmi_sync_point(struct wmi
*wmi
)
2084 struct sk_buff
*skb
;
2085 struct wmi_sync_cmd
*cmd
;
2086 struct wmi_data_sync_bufs data_sync_bufs
[WMM_NUM_AC
];
2087 enum htc_endpoint_id ep_id
;
2088 u8 index
, num_pri_streams
= 0;
2091 memset(data_sync_bufs
, 0, sizeof(data_sync_bufs
));
2093 spin_lock_bh(&wmi
->lock
);
2095 for (index
= 0; index
< WMM_NUM_AC
; index
++) {
2096 if (wmi
->fat_pipe_exist
& (1 << index
)) {
2098 data_sync_bufs
[num_pri_streams
- 1].traffic_class
=
2103 spin_unlock_bh(&wmi
->lock
);
2105 skb
= ath6kl_wmi_get_new_buf(sizeof(*cmd
));
2111 cmd
= (struct wmi_sync_cmd
*) skb
->data
;
2114 * In the SYNC cmd sent on the control Ep, send a bitmap
2115 * of the data eps on which the Data Sync will be sent
2117 cmd
->data_sync_map
= wmi
->fat_pipe_exist
;
2119 for (index
= 0; index
< num_pri_streams
; index
++) {
2120 data_sync_bufs
[index
].skb
= ath6kl_buf_alloc(0);
2121 if (data_sync_bufs
[index
].skb
== NULL
) {
2128 * If buffer allocation for any of the dataSync fails,
2129 * then do not send the Synchronize cmd on the control ep
2135 * Send sync cmd followed by sync data messages on all
2136 * endpoints being used
2138 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_SYNCHRONIZE_CMDID
,
2144 /* cmd buffer sent, we no longer own it */
2147 for (index
= 0; index
< num_pri_streams
; index
++) {
2149 if (WARN_ON(!data_sync_bufs
[index
].skb
))
2152 ep_id
= ath6kl_ac2_endpoint_id(wmi
->parent_dev
,
2153 data_sync_bufs
[index
].
2156 ath6kl_wmi_data_sync_send(wmi
, data_sync_bufs
[index
].skb
,
2162 data_sync_bufs
[index
].skb
= NULL
;
2166 /* free up any resources left over (possibly due to an error) */
2170 for (index
= 0; index
< num_pri_streams
; index
++) {
2171 if (data_sync_bufs
[index
].skb
!= NULL
) {
2172 dev_kfree_skb((struct sk_buff
*)data_sync_bufs
[index
].
2180 int ath6kl_wmi_create_pstream_cmd(struct wmi
*wmi
,
2181 struct wmi_create_pstream_cmd
*params
)
2183 struct sk_buff
*skb
;
2184 struct wmi_create_pstream_cmd
*cmd
;
2185 u8 fatpipe_exist_for_ac
= 0;
2187 s32 nominal_phy
= 0;
2190 if (!((params
->user_pri
< 8) &&
2191 (params
->user_pri
<= 0x7) &&
2192 (up_to_ac
[params
->user_pri
& 0x7] == params
->traffic_class
) &&
2193 (params
->traffic_direc
== UPLINK_TRAFFIC
||
2194 params
->traffic_direc
== DNLINK_TRAFFIC
||
2195 params
->traffic_direc
== BIDIR_TRAFFIC
) &&
2196 (params
->traffic_type
== TRAFFIC_TYPE_APERIODIC
||
2197 params
->traffic_type
== TRAFFIC_TYPE_PERIODIC
) &&
2198 (params
->voice_psc_cap
== DISABLE_FOR_THIS_AC
||
2199 params
->voice_psc_cap
== ENABLE_FOR_THIS_AC
||
2200 params
->voice_psc_cap
== ENABLE_FOR_ALL_AC
) &&
2201 (params
->tsid
== WMI_IMPLICIT_PSTREAM
||
2202 params
->tsid
<= WMI_MAX_THINSTREAM
))) {
2207 * Check nominal PHY rate is >= minimalPHY,
2208 * so that DUT can allow TSRS IE
2211 /* Get the physical rate (units of bps) */
2212 min_phy
= ((le32_to_cpu(params
->min_phy_rate
) / 1000) / 1000);
2214 /* Check minimal phy < nominal phy rate */
2215 if (params
->nominal_phy
>= min_phy
) {
2216 /* unit of 500 kbps */
2217 nominal_phy
= (params
->nominal_phy
* 1000) / 500;
2218 ath6kl_dbg(ATH6KL_DBG_WMI
,
2219 "TSRS IE enabled::MinPhy %x->NominalPhy ===> %x\n",
2220 min_phy
, nominal_phy
);
2222 params
->nominal_phy
= nominal_phy
;
2224 params
->nominal_phy
= 0;
2227 skb
= ath6kl_wmi_get_new_buf(sizeof(*cmd
));
2231 ath6kl_dbg(ATH6KL_DBG_WMI
,
2232 "sending create_pstream_cmd: ac=%d tsid:%d\n",
2233 params
->traffic_class
, params
->tsid
);
2235 cmd
= (struct wmi_create_pstream_cmd
*) skb
->data
;
2236 memcpy(cmd
, params
, sizeof(*cmd
));
2238 /* This is an implicitly created Fat pipe */
2239 if ((u32
) params
->tsid
== (u32
) WMI_IMPLICIT_PSTREAM
) {
2240 spin_lock_bh(&wmi
->lock
);
2241 fatpipe_exist_for_ac
= (wmi
->fat_pipe_exist
&
2242 (1 << params
->traffic_class
));
2243 wmi
->fat_pipe_exist
|= (1 << params
->traffic_class
);
2244 spin_unlock_bh(&wmi
->lock
);
2246 /* explicitly created thin stream within a fat pipe */
2247 spin_lock_bh(&wmi
->lock
);
2248 fatpipe_exist_for_ac
= (wmi
->fat_pipe_exist
&
2249 (1 << params
->traffic_class
));
2250 wmi
->stream_exist_for_ac
[params
->traffic_class
] |=
2251 (1 << params
->tsid
);
2253 * If a thinstream becomes active, the fat pipe automatically
2256 wmi
->fat_pipe_exist
|= (1 << params
->traffic_class
);
2257 spin_unlock_bh(&wmi
->lock
);
2261 * Indicate activty change to driver layer only if this is the
2262 * first TSID to get created in this AC explicitly or an implicit
2263 * fat pipe is getting created.
2265 if (!fatpipe_exist_for_ac
)
2266 ath6kl_indicate_tx_activity(wmi
->parent_dev
,
2267 params
->traffic_class
, true);
2269 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_CREATE_PSTREAM_CMDID
,
2274 int ath6kl_wmi_delete_pstream_cmd(struct wmi
*wmi
, u8 traffic_class
, u8 tsid
)
2276 struct sk_buff
*skb
;
2277 struct wmi_delete_pstream_cmd
*cmd
;
2278 u16 active_tsids
= 0;
2281 if (traffic_class
> 3) {
2282 ath6kl_err("invalid traffic class: %d\n", traffic_class
);
2286 skb
= ath6kl_wmi_get_new_buf(sizeof(*cmd
));
2290 cmd
= (struct wmi_delete_pstream_cmd
*) skb
->data
;
2291 cmd
->traffic_class
= traffic_class
;
2294 spin_lock_bh(&wmi
->lock
);
2295 active_tsids
= wmi
->stream_exist_for_ac
[traffic_class
];
2296 spin_unlock_bh(&wmi
->lock
);
2298 if (!(active_tsids
& (1 << tsid
))) {
2300 ath6kl_dbg(ATH6KL_DBG_WMI
,
2301 "TSID %d doesn't exist for traffic class: %d\n",
2302 tsid
, traffic_class
);
2306 ath6kl_dbg(ATH6KL_DBG_WMI
,
2307 "sending delete_pstream_cmd: traffic class: %d tsid=%d\n",
2308 traffic_class
, tsid
);
2310 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_DELETE_PSTREAM_CMDID
,
2311 SYNC_BEFORE_WMIFLAG
);
2313 spin_lock_bh(&wmi
->lock
);
2314 wmi
->stream_exist_for_ac
[traffic_class
] &= ~(1 << tsid
);
2315 active_tsids
= wmi
->stream_exist_for_ac
[traffic_class
];
2316 spin_unlock_bh(&wmi
->lock
);
2319 * Indicate stream inactivity to driver layer only if all tsids
2320 * within this AC are deleted.
2322 if (!active_tsids
) {
2323 ath6kl_indicate_tx_activity(wmi
->parent_dev
,
2324 traffic_class
, false);
2325 wmi
->fat_pipe_exist
&= ~(1 << traffic_class
);
2331 int ath6kl_wmi_set_ip_cmd(struct wmi
*wmi
, struct wmi_set_ip_cmd
*ip_cmd
)
2333 struct sk_buff
*skb
;
2334 struct wmi_set_ip_cmd
*cmd
;
2337 /* Multicast address are not valid */
2338 if ((*((u8
*) &ip_cmd
->ips
[0]) >= 0xE0) ||
2339 (*((u8
*) &ip_cmd
->ips
[1]) >= 0xE0))
2342 skb
= ath6kl_wmi_get_new_buf(sizeof(struct wmi_set_ip_cmd
));
2346 cmd
= (struct wmi_set_ip_cmd
*) skb
->data
;
2347 memcpy(cmd
, ip_cmd
, sizeof(struct wmi_set_ip_cmd
));
2349 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_SET_IP_CMDID
, NO_SYNC_WMIFLAG
);
2353 static int ath6kl_wmi_get_wow_list_event_rx(struct wmi
*wmi
, u8
* datap
,
2356 if (len
< sizeof(struct wmi_get_wow_list_reply
))
2362 static int ath6kl_wmi_cmd_send_xtnd(struct wmi
*wmi
, struct sk_buff
*skb
,
2363 enum wmix_command_id cmd_id
,
2364 enum wmi_sync_flag sync_flag
)
2366 struct wmix_cmd_hdr
*cmd_hdr
;
2369 skb_push(skb
, sizeof(struct wmix_cmd_hdr
));
2371 cmd_hdr
= (struct wmix_cmd_hdr
*) skb
->data
;
2372 cmd_hdr
->cmd_id
= cpu_to_le32(cmd_id
);
2374 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_EXTENSION_CMDID
, sync_flag
);
2379 int ath6kl_wmi_get_challenge_resp_cmd(struct wmi
*wmi
, u32 cookie
, u32 source
)
2381 struct sk_buff
*skb
;
2382 struct wmix_hb_challenge_resp_cmd
*cmd
;
2385 skb
= ath6kl_wmi_get_new_buf(sizeof(*cmd
));
2389 cmd
= (struct wmix_hb_challenge_resp_cmd
*) skb
->data
;
2390 cmd
->cookie
= cpu_to_le32(cookie
);
2391 cmd
->source
= cpu_to_le32(source
);
2393 ret
= ath6kl_wmi_cmd_send_xtnd(wmi
, skb
, WMIX_HB_CHALLENGE_RESP_CMDID
,
2398 int ath6kl_wmi_get_stats_cmd(struct wmi
*wmi
)
2400 return ath6kl_wmi_simple_cmd(wmi
, WMI_GET_STATISTICS_CMDID
);
2403 int ath6kl_wmi_set_tx_pwr_cmd(struct wmi
*wmi
, u8 dbM
)
2405 struct sk_buff
*skb
;
2406 struct wmi_set_tx_pwr_cmd
*cmd
;
2409 skb
= ath6kl_wmi_get_new_buf(sizeof(struct wmi_set_tx_pwr_cmd
));
2413 cmd
= (struct wmi_set_tx_pwr_cmd
*) skb
->data
;
2416 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_SET_TX_PWR_CMDID
,
2422 int ath6kl_wmi_get_tx_pwr_cmd(struct wmi
*wmi
)
2424 return ath6kl_wmi_simple_cmd(wmi
, WMI_GET_TX_PWR_CMDID
);
2427 int ath6kl_wmi_set_lpreamble_cmd(struct wmi
*wmi
, u8 status
, u8 preamble_policy
)
2429 struct sk_buff
*skb
;
2430 struct wmi_set_lpreamble_cmd
*cmd
;
2433 skb
= ath6kl_wmi_get_new_buf(sizeof(struct wmi_set_lpreamble_cmd
));
2437 cmd
= (struct wmi_set_lpreamble_cmd
*) skb
->data
;
2438 cmd
->status
= status
;
2439 cmd
->preamble_policy
= preamble_policy
;
2441 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_SET_LPREAMBLE_CMDID
,
2446 int ath6kl_wmi_set_rts_cmd(struct wmi
*wmi
, u16 threshold
)
2448 struct sk_buff
*skb
;
2449 struct wmi_set_rts_cmd
*cmd
;
2452 skb
= ath6kl_wmi_get_new_buf(sizeof(struct wmi_set_rts_cmd
));
2456 cmd
= (struct wmi_set_rts_cmd
*) skb
->data
;
2457 cmd
->threshold
= cpu_to_le16(threshold
);
2459 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_SET_RTS_CMDID
, NO_SYNC_WMIFLAG
);
2463 int ath6kl_wmi_set_wmm_txop(struct wmi
*wmi
, enum wmi_txop_cfg cfg
)
2465 struct sk_buff
*skb
;
2466 struct wmi_set_wmm_txop_cmd
*cmd
;
2469 if (!((cfg
== WMI_TXOP_DISABLED
) || (cfg
== WMI_TXOP_ENABLED
)))
2472 skb
= ath6kl_wmi_get_new_buf(sizeof(struct wmi_set_wmm_txop_cmd
));
2476 cmd
= (struct wmi_set_wmm_txop_cmd
*) skb
->data
;
2477 cmd
->txop_enable
= cfg
;
2479 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_SET_WMM_TXOP_CMDID
,
2484 int ath6kl_wmi_set_keepalive_cmd(struct wmi
*wmi
, u8 keep_alive_intvl
)
2486 struct sk_buff
*skb
;
2487 struct wmi_set_keepalive_cmd
*cmd
;
2490 skb
= ath6kl_wmi_get_new_buf(sizeof(*cmd
));
2494 cmd
= (struct wmi_set_keepalive_cmd
*) skb
->data
;
2495 cmd
->keep_alive_intvl
= keep_alive_intvl
;
2496 wmi
->keep_alive_intvl
= keep_alive_intvl
;
2498 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_SET_KEEPALIVE_CMDID
,
2503 s32
ath6kl_wmi_get_rate(s8 rate_index
)
2505 if (rate_index
== RATE_AUTO
)
2508 return wmi_rate_tbl
[(u32
) rate_index
][0];
2511 void ath6kl_wmi_node_return(struct wmi
*wmi
, struct bss
*bss
)
2514 wlan_node_return(&wmi
->parent_dev
->scan_table
, bss
);
2517 struct bss
*ath6kl_wmi_find_ssid_node(struct wmi
*wmi
, u8
* ssid
,
2518 u32 ssid_len
, bool is_wpa2
,
2521 struct bss
*node
= NULL
;
2523 node
= wlan_find_ssid_node(&wmi
->parent_dev
->scan_table
, ssid
,
2524 ssid_len
, is_wpa2
, match_ssid
);
2528 struct bss
*ath6kl_wmi_find_node(struct wmi
*wmi
, const u8
* mac_addr
)
2530 struct bss
*ni
= NULL
;
2532 ni
= wlan_find_node(&wmi
->parent_dev
->scan_table
, mac_addr
);
2537 void ath6kl_wmi_node_free(struct wmi
*wmi
, const u8
* mac_addr
)
2539 struct bss
*ni
= NULL
;
2541 ni
= wlan_find_node(&wmi
->parent_dev
->scan_table
, mac_addr
);
2543 wlan_node_reclaim(&wmi
->parent_dev
->scan_table
, ni
);
2548 static int ath6kl_wmi_get_pmkid_list_event_rx(struct wmi
*wmi
, u8
*datap
,
2551 struct wmi_pmkid_list_reply
*reply
;
2554 if (len
< sizeof(struct wmi_pmkid_list_reply
))
2557 reply
= (struct wmi_pmkid_list_reply
*)datap
;
2558 expected_len
= sizeof(reply
->num_pmkid
) +
2559 le32_to_cpu(reply
->num_pmkid
) * WMI_PMKID_LEN
;
2561 if (len
< expected_len
)
2567 static int ath6kl_wmi_addba_req_event_rx(struct wmi
*wmi
, u8
*datap
, int len
)
2569 struct wmi_addba_req_event
*cmd
= (struct wmi_addba_req_event
*) datap
;
2571 aggr_recv_addba_req_evt(wmi
->parent_dev
, cmd
->tid
,
2572 le16_to_cpu(cmd
->st_seq_no
), cmd
->win_sz
);
2577 static int ath6kl_wmi_delba_req_event_rx(struct wmi
*wmi
, u8
*datap
, int len
)
2579 struct wmi_delba_event
*cmd
= (struct wmi_delba_event
*) datap
;
2581 aggr_recv_delba_req_evt(wmi
->parent_dev
, cmd
->tid
);
2586 /* AP mode functions */
2588 int ath6kl_wmi_ap_profile_commit(struct wmi
*wmip
, struct wmi_connect_cmd
*p
)
2590 struct sk_buff
*skb
;
2591 struct wmi_connect_cmd
*cm
;
2594 skb
= ath6kl_wmi_get_new_buf(sizeof(*cm
));
2598 cm
= (struct wmi_connect_cmd
*) skb
->data
;
2599 memcpy(cm
, p
, sizeof(*cm
));
2601 res
= ath6kl_wmi_cmd_send(wmip
, skb
, WMI_AP_CONFIG_COMMIT_CMDID
,
2603 ath6kl_dbg(ATH6KL_DBG_WMI
, "%s: nw_type=%u auth_mode=%u ch=%u "
2604 "ctrl_flags=0x%x-> res=%d\n",
2605 __func__
, p
->nw_type
, p
->auth_mode
, le16_to_cpu(p
->ch
),
2606 le32_to_cpu(p
->ctrl_flags
), res
);
2610 int ath6kl_wmi_ap_set_mlme(struct wmi
*wmip
, u8 cmd
, const u8
*mac
, u16 reason
)
2612 struct sk_buff
*skb
;
2613 struct wmi_ap_set_mlme_cmd
*cm
;
2615 skb
= ath6kl_wmi_get_new_buf(sizeof(*cm
));
2619 cm
= (struct wmi_ap_set_mlme_cmd
*) skb
->data
;
2620 memcpy(cm
->mac
, mac
, ETH_ALEN
);
2621 cm
->reason
= cpu_to_le16(reason
);
2624 return ath6kl_wmi_cmd_send(wmip
, skb
, WMI_AP_SET_MLME_CMDID
,
2628 static int ath6kl_wmi_pspoll_event_rx(struct wmi
*wmi
, u8
*datap
, int len
)
2630 struct wmi_pspoll_event
*ev
;
2632 if (len
< sizeof(struct wmi_pspoll_event
))
2635 ev
= (struct wmi_pspoll_event
*) datap
;
2637 ath6kl_pspoll_event(wmi
->parent_dev
, le16_to_cpu(ev
->aid
));
2642 static int ath6kl_wmi_dtimexpiry_event_rx(struct wmi
*wmi
, u8
*datap
, int len
)
2644 ath6kl_dtimexpiry_event(wmi
->parent_dev
);
2649 int ath6kl_wmi_set_pvb_cmd(struct wmi
*wmi
, u16 aid
, bool flag
)
2651 struct sk_buff
*skb
;
2652 struct wmi_ap_set_pvb_cmd
*cmd
;
2655 skb
= ath6kl_wmi_get_new_buf(sizeof(struct wmi_ap_set_pvb_cmd
));
2659 cmd
= (struct wmi_ap_set_pvb_cmd
*) skb
->data
;
2660 cmd
->aid
= cpu_to_le16(aid
);
2661 cmd
->flag
= cpu_to_le32(flag
);
2663 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_AP_SET_PVB_CMDID
,
2669 int ath6kl_wmi_set_rx_frame_format_cmd(struct wmi
*wmi
, u8 rx_meta_ver
,
2670 bool rx_dot11_hdr
, bool defrag_on_host
)
2672 struct sk_buff
*skb
;
2673 struct wmi_rx_frame_format_cmd
*cmd
;
2676 skb
= ath6kl_wmi_get_new_buf(sizeof(*cmd
));
2680 cmd
= (struct wmi_rx_frame_format_cmd
*) skb
->data
;
2681 cmd
->dot11_hdr
= rx_dot11_hdr
? 1 : 0;
2682 cmd
->defrag_on_host
= defrag_on_host
? 1 : 0;
2683 cmd
->meta_ver
= rx_meta_ver
;
2685 /* Delete the local aggr state, on host */
2686 ret
= ath6kl_wmi_cmd_send(wmi
, skb
, WMI_RX_FRAME_FORMAT_CMDID
,
2692 int ath6kl_wmi_set_appie_cmd(struct wmi
*wmi
, u8 mgmt_frm_type
, const u8
*ie
,
2695 struct sk_buff
*skb
;
2696 struct wmi_set_appie_cmd
*p
;
2698 skb
= ath6kl_wmi_get_new_buf(sizeof(*p
) + ie_len
);
2702 ath6kl_dbg(ATH6KL_DBG_WMI
, "set_appie_cmd: mgmt_frm_type=%u "
2703 "ie_len=%u\n", mgmt_frm_type
, ie_len
);
2704 p
= (struct wmi_set_appie_cmd
*) skb
->data
;
2705 p
->mgmt_frm_type
= mgmt_frm_type
;
2707 memcpy(p
->ie_info
, ie
, ie_len
);
2708 return ath6kl_wmi_cmd_send(wmi
, skb
, WMI_SET_APPIE_CMDID
,
2712 int ath6kl_wmi_disable_11b_rates_cmd(struct wmi
*wmi
, bool disable
)
2714 struct sk_buff
*skb
;
2715 struct wmi_disable_11b_rates_cmd
*cmd
;
2717 skb
= ath6kl_wmi_get_new_buf(sizeof(*cmd
));
2721 ath6kl_dbg(ATH6KL_DBG_WMI
, "disable_11b_rates_cmd: disable=%u\n",
2723 cmd
= (struct wmi_disable_11b_rates_cmd
*) skb
->data
;
2724 cmd
->disable
= disable
? 1 : 0;
2726 return ath6kl_wmi_cmd_send(wmi
, skb
, WMI_DISABLE_11B_RATES_CMDID
,
2730 int ath6kl_wmi_remain_on_chnl_cmd(struct wmi
*wmi
, u32 freq
, u32 dur
)
2732 struct sk_buff
*skb
;
2733 struct wmi_remain_on_chnl_cmd
*p
;
2735 skb
= ath6kl_wmi_get_new_buf(sizeof(*p
));
2739 ath6kl_dbg(ATH6KL_DBG_WMI
, "remain_on_chnl_cmd: freq=%u dur=%u\n",
2741 p
= (struct wmi_remain_on_chnl_cmd
*) skb
->data
;
2742 p
->freq
= cpu_to_le32(freq
);
2743 p
->duration
= cpu_to_le32(dur
);
2744 return ath6kl_wmi_cmd_send(wmi
, skb
, WMI_REMAIN_ON_CHNL_CMDID
,
2748 int ath6kl_wmi_send_action_cmd(struct wmi
*wmi
, u32 id
, u32 freq
, u32 wait
,
2749 const u8
*data
, u16 data_len
)
2751 struct sk_buff
*skb
;
2752 struct wmi_send_action_cmd
*p
;
2756 return -EINVAL
; /* Offload for wait not supported */
2758 buf
= kmalloc(data_len
, GFP_KERNEL
);
2762 skb
= ath6kl_wmi_get_new_buf(sizeof(*p
) + data_len
);
2768 kfree(wmi
->last_mgmt_tx_frame
);
2769 wmi
->last_mgmt_tx_frame
= buf
;
2770 wmi
->last_mgmt_tx_frame_len
= data_len
;
2772 ath6kl_dbg(ATH6KL_DBG_WMI
, "send_action_cmd: id=%u freq=%u wait=%u "
2773 "len=%u\n", id
, freq
, wait
, data_len
);
2774 p
= (struct wmi_send_action_cmd
*) skb
->data
;
2775 p
->id
= cpu_to_le32(id
);
2776 p
->freq
= cpu_to_le32(freq
);
2777 p
->wait
= cpu_to_le32(wait
);
2778 p
->len
= cpu_to_le16(data_len
);
2779 memcpy(p
->data
, data
, data_len
);
2780 return ath6kl_wmi_cmd_send(wmi
, skb
, WMI_SEND_ACTION_CMDID
,
2784 int ath6kl_wmi_send_probe_response_cmd(struct wmi
*wmi
, u32 freq
,
2786 const u8
*data
, u16 data_len
)
2788 struct sk_buff
*skb
;
2789 struct wmi_p2p_probe_response_cmd
*p
;
2791 skb
= ath6kl_wmi_get_new_buf(sizeof(*p
) + data_len
);
2795 ath6kl_dbg(ATH6KL_DBG_WMI
, "send_probe_response_cmd: freq=%u dst=%pM "
2796 "len=%u\n", freq
, dst
, data_len
);
2797 p
= (struct wmi_p2p_probe_response_cmd
*) skb
->data
;
2798 p
->freq
= cpu_to_le32(freq
);
2799 memcpy(p
->destination_addr
, dst
, ETH_ALEN
);
2800 p
->len
= cpu_to_le16(data_len
);
2801 memcpy(p
->data
, data
, data_len
);
2802 return ath6kl_wmi_cmd_send(wmi
, skb
, WMI_SEND_PROBE_RESPONSE_CMDID
,
2806 int ath6kl_wmi_probe_report_req_cmd(struct wmi
*wmi
, bool enable
)
2808 struct sk_buff
*skb
;
2809 struct wmi_probe_req_report_cmd
*p
;
2811 skb
= ath6kl_wmi_get_new_buf(sizeof(*p
));
2815 ath6kl_dbg(ATH6KL_DBG_WMI
, "probe_report_req_cmd: enable=%u\n",
2817 p
= (struct wmi_probe_req_report_cmd
*) skb
->data
;
2818 p
->enable
= enable
? 1 : 0;
2819 return ath6kl_wmi_cmd_send(wmi
, skb
, WMI_PROBE_REQ_REPORT_CMDID
,
2823 int ath6kl_wmi_info_req_cmd(struct wmi
*wmi
, u32 info_req_flags
)
2825 struct sk_buff
*skb
;
2826 struct wmi_get_p2p_info
*p
;
2828 skb
= ath6kl_wmi_get_new_buf(sizeof(*p
));
2832 ath6kl_dbg(ATH6KL_DBG_WMI
, "info_req_cmd: flags=%x\n",
2834 p
= (struct wmi_get_p2p_info
*) skb
->data
;
2835 p
->info_req_flags
= cpu_to_le32(info_req_flags
);
2836 return ath6kl_wmi_cmd_send(wmi
, skb
, WMI_GET_P2P_INFO_CMDID
,
2840 int ath6kl_wmi_cancel_remain_on_chnl_cmd(struct wmi
*wmi
)
2842 ath6kl_dbg(ATH6KL_DBG_WMI
, "cancel_remain_on_chnl_cmd\n");
2843 return ath6kl_wmi_simple_cmd(wmi
, WMI_CANCEL_REMAIN_ON_CHNL_CMDID
);
2846 static int ath6kl_wmi_control_rx_xtnd(struct wmi
*wmi
, struct sk_buff
*skb
)
2848 struct wmix_cmd_hdr
*cmd
;
2854 if (skb
->len
< sizeof(struct wmix_cmd_hdr
)) {
2855 ath6kl_err("bad packet 1\n");
2856 wmi
->stat
.cmd_len_err
++;
2860 cmd
= (struct wmix_cmd_hdr
*) skb
->data
;
2861 id
= le32_to_cpu(cmd
->cmd_id
);
2863 skb_pull(skb
, sizeof(struct wmix_cmd_hdr
));
2869 case WMIX_HB_CHALLENGE_RESP_EVENTID
:
2871 case WMIX_DBGLOG_EVENTID
:
2874 ath6kl_err("unknown cmd id 0x%x\n", id
);
2875 wmi
->stat
.cmd_id_err
++;
2884 int ath6kl_wmi_control_rx(struct wmi
*wmi
, struct sk_buff
*skb
)
2886 struct wmi_cmd_hdr
*cmd
;
2892 if (WARN_ON(skb
== NULL
))
2895 if (skb
->len
< sizeof(struct wmi_cmd_hdr
)) {
2896 ath6kl_err("bad packet 1\n");
2898 wmi
->stat
.cmd_len_err
++;
2902 cmd
= (struct wmi_cmd_hdr
*) skb
->data
;
2903 id
= le16_to_cpu(cmd
->cmd_id
);
2905 skb_pull(skb
, sizeof(struct wmi_cmd_hdr
));
2910 ath6kl_dbg(ATH6KL_DBG_WMI
, "%s: wmi id: %d\n", __func__
, id
);
2911 ath6kl_dbg_dump(ATH6KL_DBG_RAW_BYTES
, "msg payload ", datap
, len
);
2914 case WMI_GET_BITRATE_CMDID
:
2915 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_GET_BITRATE_CMDID\n");
2916 ret
= ath6kl_wmi_bitrate_reply_rx(wmi
, datap
, len
);
2918 case WMI_GET_CHANNEL_LIST_CMDID
:
2919 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_GET_CHANNEL_LIST_CMDID\n");
2920 ret
= ath6kl_wmi_ch_list_reply_rx(wmi
, datap
, len
);
2922 case WMI_GET_TX_PWR_CMDID
:
2923 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_GET_TX_PWR_CMDID\n");
2924 ret
= ath6kl_wmi_tx_pwr_reply_rx(wmi
, datap
, len
);
2926 case WMI_READY_EVENTID
:
2927 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_READY_EVENTID\n");
2928 ret
= ath6kl_wmi_ready_event_rx(wmi
, datap
, len
);
2930 case WMI_CONNECT_EVENTID
:
2931 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_CONNECT_EVENTID\n");
2932 ret
= ath6kl_wmi_connect_event_rx(wmi
, datap
, len
);
2934 case WMI_DISCONNECT_EVENTID
:
2935 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_DISCONNECT_EVENTID\n");
2936 ret
= ath6kl_wmi_disconnect_event_rx(wmi
, datap
, len
);
2938 case WMI_PEER_NODE_EVENTID
:
2939 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_PEER_NODE_EVENTID\n");
2940 ret
= ath6kl_wmi_peer_node_event_rx(wmi
, datap
, len
);
2942 case WMI_TKIP_MICERR_EVENTID
:
2943 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_TKIP_MICERR_EVENTID\n");
2944 ret
= ath6kl_wmi_tkip_micerr_event_rx(wmi
, datap
, len
);
2946 case WMI_BSSINFO_EVENTID
:
2947 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_BSSINFO_EVENTID\n");
2948 ath6kl_wmi_convert_bssinfo_hdr2_to_hdr(skb
, datap
);
2949 ret
= ath6kl_wmi_bssinfo_event_rx(wmi
, skb
->data
, skb
->len
);
2951 case WMI_REGDOMAIN_EVENTID
:
2952 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_REGDOMAIN_EVENTID\n");
2954 case WMI_PSTREAM_TIMEOUT_EVENTID
:
2955 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_PSTREAM_TIMEOUT_EVENTID\n");
2956 ret
= ath6kl_wmi_pstream_timeout_event_rx(wmi
, datap
, len
);
2958 case WMI_NEIGHBOR_REPORT_EVENTID
:
2959 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_NEIGHBOR_REPORT_EVENTID\n");
2961 case WMI_SCAN_COMPLETE_EVENTID
:
2962 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_SCAN_COMPLETE_EVENTID\n");
2963 ret
= ath6kl_wmi_scan_complete_rx(wmi
, datap
, len
);
2965 case WMI_CMDERROR_EVENTID
:
2966 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_CMDERROR_EVENTID\n");
2967 ret
= ath6kl_wmi_error_event_rx(wmi
, datap
, len
);
2969 case WMI_REPORT_STATISTICS_EVENTID
:
2970 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_REPORT_STATISTICS_EVENTID\n");
2971 ret
= ath6kl_wmi_stats_event_rx(wmi
, datap
, len
);
2973 case WMI_RSSI_THRESHOLD_EVENTID
:
2974 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_RSSI_THRESHOLD_EVENTID\n");
2975 ret
= ath6kl_wmi_rssi_threshold_event_rx(wmi
, datap
, len
);
2977 case WMI_ERROR_REPORT_EVENTID
:
2978 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_ERROR_REPORT_EVENTID\n");
2980 case WMI_OPT_RX_FRAME_EVENTID
:
2981 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_OPT_RX_FRAME_EVENTID\n");
2982 ret
= ath6kl_wmi_opt_frame_event_rx(wmi
, datap
, len
);
2984 case WMI_REPORT_ROAM_TBL_EVENTID
:
2985 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_REPORT_ROAM_TBL_EVENTID\n");
2987 case WMI_EXTENSION_EVENTID
:
2988 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_EXTENSION_EVENTID\n");
2989 ret
= ath6kl_wmi_control_rx_xtnd(wmi
, skb
);
2991 case WMI_CAC_EVENTID
:
2992 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_CAC_EVENTID\n");
2993 ret
= ath6kl_wmi_cac_event_rx(wmi
, datap
, len
);
2995 case WMI_CHANNEL_CHANGE_EVENTID
:
2996 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_CHANNEL_CHANGE_EVENTID\n");
2998 case WMI_REPORT_ROAM_DATA_EVENTID
:
2999 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_REPORT_ROAM_DATA_EVENTID\n");
3001 case WMI_GET_FIXRATES_CMDID
:
3002 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_GET_FIXRATES_CMDID\n");
3003 ret
= ath6kl_wmi_ratemask_reply_rx(wmi
, datap
, len
);
3005 case WMI_TX_RETRY_ERR_EVENTID
:
3006 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_TX_RETRY_ERR_EVENTID\n");
3008 case WMI_SNR_THRESHOLD_EVENTID
:
3009 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_SNR_THRESHOLD_EVENTID\n");
3010 ret
= ath6kl_wmi_snr_threshold_event_rx(wmi
, datap
, len
);
3012 case WMI_LQ_THRESHOLD_EVENTID
:
3013 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_LQ_THRESHOLD_EVENTID\n");
3015 case WMI_APLIST_EVENTID
:
3016 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_APLIST_EVENTID\n");
3017 ret
= ath6kl_wmi_aplist_event_rx(wmi
, datap
, len
);
3019 case WMI_GET_KEEPALIVE_CMDID
:
3020 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_GET_KEEPALIVE_CMDID\n");
3021 ret
= ath6kl_wmi_keepalive_reply_rx(wmi
, datap
, len
);
3023 case WMI_GET_WOW_LIST_EVENTID
:
3024 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_GET_WOW_LIST_EVENTID\n");
3025 ret
= ath6kl_wmi_get_wow_list_event_rx(wmi
, datap
, len
);
3027 case WMI_GET_PMKID_LIST_EVENTID
:
3028 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_GET_PMKID_LIST_EVENTID\n");
3029 ret
= ath6kl_wmi_get_pmkid_list_event_rx(wmi
, datap
, len
);
3031 case WMI_PSPOLL_EVENTID
:
3032 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_PSPOLL_EVENTID\n");
3033 ret
= ath6kl_wmi_pspoll_event_rx(wmi
, datap
, len
);
3035 case WMI_DTIMEXPIRY_EVENTID
:
3036 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_DTIMEXPIRY_EVENTID\n");
3037 ret
= ath6kl_wmi_dtimexpiry_event_rx(wmi
, datap
, len
);
3039 case WMI_SET_PARAMS_REPLY_EVENTID
:
3040 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_SET_PARAMS_REPLY_EVENTID\n");
3042 case WMI_ADDBA_REQ_EVENTID
:
3043 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_ADDBA_REQ_EVENTID\n");
3044 ret
= ath6kl_wmi_addba_req_event_rx(wmi
, datap
, len
);
3046 case WMI_ADDBA_RESP_EVENTID
:
3047 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_ADDBA_RESP_EVENTID\n");
3049 case WMI_DELBA_REQ_EVENTID
:
3050 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_DELBA_REQ_EVENTID\n");
3051 ret
= ath6kl_wmi_delba_req_event_rx(wmi
, datap
, len
);
3053 case WMI_REPORT_BTCOEX_CONFIG_EVENTID
:
3054 ath6kl_dbg(ATH6KL_DBG_WMI
,
3055 "WMI_REPORT_BTCOEX_CONFIG_EVENTID\n");
3057 case WMI_REPORT_BTCOEX_STATS_EVENTID
:
3058 ath6kl_dbg(ATH6KL_DBG_WMI
,
3059 "WMI_REPORT_BTCOEX_STATS_EVENTID\n");
3061 case WMI_TX_COMPLETE_EVENTID
:
3062 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_TX_COMPLETE_EVENTID\n");
3063 ret
= ath6kl_wmi_tx_complete_event_rx(datap
, len
);
3065 case WMI_REMAIN_ON_CHNL_EVENTID
:
3066 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_REMAIN_ON_CHNL_EVENTID\n");
3067 ret
= ath6kl_wmi_remain_on_chnl_event_rx(wmi
, datap
, len
);
3069 case WMI_CANCEL_REMAIN_ON_CHNL_EVENTID
:
3070 ath6kl_dbg(ATH6KL_DBG_WMI
,
3071 "WMI_CANCEL_REMAIN_ON_CHNL_EVENTID\n");
3072 ret
= ath6kl_wmi_cancel_remain_on_chnl_event_rx(wmi
, datap
,
3075 case WMI_TX_STATUS_EVENTID
:
3076 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_TX_STATUS_EVENTID\n");
3077 ret
= ath6kl_wmi_tx_status_event_rx(wmi
, datap
, len
);
3079 case WMI_RX_PROBE_REQ_EVENTID
:
3080 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_RX_PROBE_REQ_EVENTID\n");
3081 ret
= ath6kl_wmi_rx_probe_req_event_rx(wmi
, datap
, len
);
3083 case WMI_P2P_CAPABILITIES_EVENTID
:
3084 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_P2P_CAPABILITIES_EVENTID\n");
3085 ret
= ath6kl_wmi_p2p_capabilities_event_rx(datap
, len
);
3087 case WMI_RX_ACTION_EVENTID
:
3088 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_RX_ACTION_EVENTID\n");
3089 ret
= ath6kl_wmi_rx_action_event_rx(datap
, len
);
3091 case WMI_P2P_INFO_EVENTID
:
3092 ath6kl_dbg(ATH6KL_DBG_WMI
, "WMI_P2P_INFO_EVENTID\n");
3093 ret
= ath6kl_wmi_p2p_info_event_rx(datap
, len
);
3096 ath6kl_dbg(ATH6KL_DBG_WMI
, "unknown cmd id 0x%x\n", id
);
3097 wmi
->stat
.cmd_id_err
++;
3107 static void ath6kl_wmi_qos_state_init(struct wmi
*wmi
)
3112 spin_lock_bh(&wmi
->lock
);
3114 wmi
->fat_pipe_exist
= 0;
3115 memset(wmi
->stream_exist_for_ac
, 0, sizeof(wmi
->stream_exist_for_ac
));
3117 spin_unlock_bh(&wmi
->lock
);
3120 void *ath6kl_wmi_init(struct ath6kl
*dev
)
3124 wmi
= kzalloc(sizeof(struct wmi
), GFP_KERNEL
);
3128 spin_lock_init(&wmi
->lock
);
3130 wmi
->parent_dev
= dev
;
3132 ath6kl_wmi_qos_state_init(wmi
);
3134 wmi
->pwr_mode
= REC_POWER
;
3135 wmi
->phy_mode
= WMI_11G_MODE
;
3137 wmi
->pair_crypto_type
= NONE_CRYPT
;
3138 wmi
->grp_crypto_type
= NONE_CRYPT
;
3140 wmi
->ht_allowed
[A_BAND_24GHZ
] = 1;
3141 wmi
->ht_allowed
[A_BAND_5GHZ
] = 1;
3146 void ath6kl_wmi_shutdown(struct wmi
*wmi
)
3151 kfree(wmi
->last_mgmt_tx_frame
);