M7350/external/compat-wireless/drivers/net/wireless/ath/ath6kl-3.5/main.c

3173 lines
74 KiB
C
Raw Permalink Normal View History

2024-09-09 08:57:42 +00:00
/*
* Copyright (c) 2004-2011 Atheros Communications Inc.
*
* Permission to use, copy, modify, and/or distribute this software for any
* purpose with or without fee is hereby granted, provided that the above
* copyright notice and this permission notice appear in all copies.
*
* THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
* WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
* MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
* ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
* WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
* ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
* OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
*/
#include "core.h"
#include "hif-ops.h"
#include "htc-ops.h"
#include "cfg80211.h"
#include "target.h"
#include "debug.h"
#include <linux/vmalloc.h>
#ifdef ATHTST_SUPPORT
#include "ieee80211_ioctl.h"
#include "ce_athtst.h"
#endif
int _string_to_mac(char *string, int len, u8 *macaddr)
{
int i, k, ret;
char temp[3] = {0};
unsigned long value;
/* assume string is "00:11:22:33:44:55". */
if (!string || (len < 17))
return -1;
i = k = 0;
while (i < len) {
memcpy(temp, string + i, 2);
ret = kstrtoul(temp, 16, &value);
if (ret < 0)
return -1;
macaddr[k++] = (u8)value;
i += 3;
}
return 0;
}
struct ath6kl_sta *ath6kl_find_sta(struct ath6kl_vif *vif, u8 *node_addr)
{
struct ath6kl_sta *conn = NULL;
u8 i, max_conn;
if (vif->nw_type != AP_NETWORK)
return &vif->sta_list[0];
if (is_zero_ether_addr(node_addr))
return NULL;
max_conn = (vif->nw_type == AP_NETWORK) ? AP_MAX_NUM_STA : 0;
for (i = 0; i < max_conn; i++) {
if (memcmp(node_addr, vif->sta_list[i].mac, ETH_ALEN) == 0) {
conn = &vif->sta_list[i];
break;
}
}
return conn;
}
struct ath6kl_sta *ath6kl_find_sta_by_aid(struct ath6kl_vif *vif, u8 aid)
{
struct ath6kl_sta *conn = NULL;
u8 ctr;
if (vif->nw_type != AP_NETWORK) {
conn = &vif->sta_list[0];
} else {
for (ctr = 0; ctr < AP_MAX_NUM_STA; ctr++) {
if (vif->sta_list[ctr].aid == aid) {
conn = &vif->sta_list[ctr];
break;
}
}
}
return conn;
}
static void ath6kl_add_new_sta(struct ath6kl_vif *vif, u8 *mac, u8 aid,
u8 *wpaie, u8 ielen, u8 keymgmt, u8 ucipher,
u8 auth, u8 apsd_info, bool ht_support,
u8 phymode)
{
struct ath6kl_sta *sta;
u8 free_slot;
BUG_ON(aid > AP_MAX_NUM_STA);
free_slot = aid - 1;
sta = &vif->sta_list[free_slot];
spin_lock_bh(&sta->lock);
memcpy(sta->mac, mac, ETH_ALEN);
memcpy(sta->wpa_ie, wpaie, ielen);
sta->aid = aid;
sta->keymgmt = keymgmt;
sta->ucipher = ucipher;
sta->auth = auth;
sta->apsd_info = apsd_info;
sta->phymode = phymode;
sta->vif = vif;
if (!sta->psq_age_active) {
init_timer(&sta->psq_age_timer);
sta->psq_age_timer.function = ath6kl_ps_queue_age_handler;
sta->psq_age_timer.data = (unsigned long)sta;
sta->psq_age_active = 1;
}
aggr_reset_state(sta->aggr_conn_cntxt);
sta->last_txrx_time_tgt = 0;
sta->last_txrx_time = 0;
vif->sta_list_index = vif->sta_list_index | (1 << free_slot);
vif->ap_stats.sta[free_slot].aid = aid;
if (ht_support)
sta->sta_flags |= STA_HT_SUPPORT;
else
vif->sta_no_ht_num++;
spin_unlock_bh(&sta->lock);
}
static void ath6kl_sta_cleanup(struct ath6kl_vif *vif, u8 i)
{
struct ath6kl_sta *sta = &vif->sta_list[i];
struct ath6kl *ar = vif->ar;
struct ath6kl_p2p_flowctrl *p2p_flowctrl = ar->p2p_flowctrl_ctx;
struct ath6kl_fw_conn_list *fw_conn;
struct list_head container;
int reclaim = 0, j;
if (sta->psq_age_active) {
del_timer_sync(&sta->psq_age_timer);
sta->psq_age_active = 0;
}
if (p2p_flowctrl &&
p2p_flowctrl->sche_type == P2P_FLOWCTRL_SCHE_TYPE_CONNECTION) {
INIT_LIST_HEAD(&container);
spin_lock_bh(&p2p_flowctrl->p2p_flowctrl_lock);
fw_conn = &p2p_flowctrl->fw_conn_list[0];
for (j = 0; j < NUM_CONN; j++, fw_conn++) {
if (fw_conn->connId == ATH6KL_P2P_FLOWCTRL_NULL_CONNID)
continue;
if (memcmp(fw_conn->mac_addr,
sta->mac,
ETH_ALEN) == 0) {
ath6kl_p2p_flowctrl_conn_collect_by_conn(
fw_conn, &container, &reclaim);
break;
}
}
spin_unlock_bh(&p2p_flowctrl->p2p_flowctrl_lock);
ath6kl_tx_complete(ar->htc_target, &container);
}
spin_lock_bh(&sta->lock);
sta->vif = NULL;
/* empty the queued pkts in the PS queue if any */
ath6kl_ps_queue_purge(&sta->psq_data);
ath6kl_ps_queue_purge(&sta->psq_mgmt);
memset(&vif->ap_stats.sta[sta->aid - 1], 0,
sizeof(struct wmi_per_sta_stat));
memset(sta->mac, 0, ETH_ALEN);
memset(sta->wpa_ie, 0, ATH6KL_MAX_IE);
if (!(sta->sta_flags & STA_HT_SUPPORT))
vif->sta_no_ht_num--;
sta->aid = 0;
sta->sta_flags = 0;
aggr_reset_state(sta->aggr_conn_cntxt);
vif->sta_list_index = vif->sta_list_index & ~(1 << i);
spin_unlock_bh(&sta->lock);
}
static u8 ath6kl_remove_sta(struct ath6kl_vif *vif, u8 *mac, u16 reason)
{
u8 i, removed = 0;
if (is_zero_ether_addr(mac))
return removed;
if (is_broadcast_ether_addr(mac)) {
ath6kl_dbg(ATH6KL_DBG_TRC, "deleting all station\n");
for (i = 0; i < AP_MAX_NUM_STA; i++) {
if (!is_zero_ether_addr(vif->sta_list[i].mac)) {
ath6kl_sta_cleanup(vif, i);
removed = 1;
}
}
} else {
for (i = 0; i < AP_MAX_NUM_STA; i++) {
if (memcmp(vif->sta_list[i].mac, mac, ETH_ALEN) == 0) {
ath6kl_dbg(ATH6KL_DBG_TRC,
"deleting station %pM aid=%d reason=%d\n",
mac, vif->sta_list[i].aid, reason);
ath6kl_sta_cleanup(vif, i);
removed = 1;
break;
}
}
}
return removed;
}
void ath6kl_ps_queue_init(struct ath6kl_ps_buf_head *psq,
enum ps_queue_type queue_type,
u32 age_cycle,
u32 max_depth)
{
INIT_LIST_HEAD(&psq->list);
psq->queue_type = queue_type;
psq->age_cycle = age_cycle;
psq->max_depth = max_depth;
psq->depth = 0;
psq->enqueued = 0;
psq->enqueued_err = 0;
psq->dequeued = 0;
psq->aged = 0;
spin_lock_init(&psq->lock);
ath6kl_dbg(ATH6KL_DBG_POWERSAVE,
"ps: init psq %p age_cycle %d\n",
psq,
psq->age_cycle);
return;
}
void ath6kl_ps_queue_purge(struct ath6kl_ps_buf_head *psq)
{
unsigned long flags;
struct ath6kl_ps_buf_desc *ps_buf, *ps_buf_n;
ath6kl_dbg(ATH6KL_DBG_POWERSAVE,
"ps: purge psq %p depth %d\n",
psq,
psq->depth);
WARN_ON(psq->depth < 0);
spin_lock_irqsave(&psq->lock, flags);
if (psq->depth == 0) {
spin_unlock_irqrestore(&psq->lock, flags);
return;
}
list_for_each_entry_safe(ps_buf, ps_buf_n, &psq->list, list) {
list_del(&ps_buf->list);
if (ps_buf->skb)
dev_kfree_skb(ps_buf->skb);
kfree(ps_buf);
psq->depth--;
psq->dequeued++;
}
WARN_ON(psq->depth != 0);
WARN_ON(psq->enqueued != psq->dequeued);
/* call ath6kl_ps_queue_init() instead? */
psq->depth = 0;
psq->enqueued = 0;
psq->enqueued_err = 0;
psq->dequeued = 0;
psq->aged = 0;
spin_unlock_irqrestore(&psq->lock, flags);
return;
}
int ath6kl_ps_queue_empty(struct ath6kl_ps_buf_head *psq)
{
unsigned long flags;
int empty;
WARN_ON(psq->depth < 0);
spin_lock_irqsave(&psq->lock, flags);
empty = (psq->depth == 0);
spin_unlock_irqrestore(&psq->lock, flags);
return empty;
}
int ath6kl_ps_queue_depth(struct ath6kl_ps_buf_head *psq)
{
unsigned long flags;
int depth;
WARN_ON(psq->depth < 0);
spin_lock_irqsave(&psq->lock, flags);
depth = psq->depth;
spin_unlock_irqrestore(&psq->lock, flags);
return depth;
}
void ath6kl_ps_queue_stat(struct ath6kl_ps_buf_head *psq,
int *depth,
u32 *enqueued,
u32 *enqueued_err,
u32 *dequeued,
u32 *aged)
{
unsigned long flags;
spin_lock_irqsave(&psq->lock, flags);
*depth = psq->depth;
*enqueued = psq->enqueued;
*enqueued_err = psq->enqueued_err;
*dequeued = psq->dequeued;
*aged = psq->aged;
spin_unlock_irqrestore(&psq->lock, flags);
ath6kl_dbg(ATH6KL_DBG_POWERSAVE,
"ps: stat psq %p depth %d enqueued %d enqueued_err %d dequeued %d aged %d\n",
psq,
*depth,
*enqueued,
*enqueued_err,
*dequeued,
*aged);
return;
}
struct ath6kl_ps_buf_desc *ath6kl_ps_queue_dequeue(
struct ath6kl_ps_buf_head *psq)
{
unsigned long flags;
struct ath6kl_ps_buf_desc *ps_buf;
ath6kl_dbg(ATH6KL_DBG_POWERSAVE,
"ps: dequeue psq %p depth %d\n",
psq,
psq->depth);
WARN_ON(psq->depth < 0);
spin_lock_irqsave(&psq->lock, flags);
if (psq->depth == 0) {
spin_unlock_irqrestore(&psq->lock, flags);
return NULL;
}
ps_buf = list_first_entry(&psq->list, struct ath6kl_ps_buf_desc, list);
list_del(&ps_buf->list);
psq->depth--;
psq->dequeued++;
spin_unlock_irqrestore(&psq->lock, flags);
return ps_buf;
}
int ath6kl_ps_queue_enqueue_mgmt(struct ath6kl_ps_buf_head *psq,
const u8 *buf,
u16 len,
u32 id,
u32 freq,
u32 wait,
bool no_cck,
bool dont_wait_for_ack)
{
unsigned long flags;
struct ath6kl_ps_buf_desc *ps_buf;
int mgmt_buf_size;
ath6kl_dbg(ATH6KL_DBG_POWERSAVE,
"ps: enqueue_mgmt psq %p depth %d buf %p\n",
psq,
psq->depth,
buf);
if ((psq->max_depth != ATH6KL_PS_QUEUE_NO_DEPTH) &&
(psq->depth > psq->max_depth)) {
psq->enqueued_err++;
return -EBUSY;
}
mgmt_buf_size = len + sizeof(struct ath6kl_ps_buf_desc) - 1;
ps_buf = kmalloc(mgmt_buf_size, GFP_ATOMIC);
if (!ps_buf) {
psq->enqueued_err++;
return -ENOMEM;
}
INIT_LIST_HEAD(&ps_buf->list);
ps_buf->id = id;
ps_buf->freq = freq;
ps_buf->wait = wait;
ps_buf->no_cck = no_cck;
ps_buf->dont_wait_for_ack = dont_wait_for_ack;
ps_buf->len = len;
ps_buf->age = 0;
ps_buf->skb = NULL;
memcpy(ps_buf->buf, buf, len);
spin_lock_irqsave(&psq->lock, flags);
list_add_tail(&ps_buf->list, &psq->list);
psq->depth++;
psq->enqueued++;
spin_unlock_irqrestore(&psq->lock, flags);
return 0;
}
int ath6kl_ps_queue_enqueue_data(struct ath6kl_ps_buf_head *psq,
struct sk_buff *skb)
{
unsigned long flags;
struct ath6kl_ps_buf_desc *ps_buf;
int data_buf_size;
ath6kl_dbg(ATH6KL_DBG_POWERSAVE,
"ps: enqueue_data psq %p depth %d skb %p\n",
psq,
psq->depth,
skb);
if ((psq->max_depth != ATH6KL_PS_QUEUE_NO_DEPTH) &&
(psq->depth > psq->max_depth)) {
psq->enqueued_err++;
return -EBUSY;
}
data_buf_size = sizeof(struct ath6kl_ps_buf_desc);
ps_buf = kmalloc(data_buf_size, GFP_ATOMIC);
if (!ps_buf) {
psq->enqueued_err++;
return -ENOMEM;
}
INIT_LIST_HEAD(&ps_buf->list);
ps_buf->age = 0;
ps_buf->skb = skb;
ps_buf->len = skb->len;
spin_lock_irqsave(&psq->lock, flags);
list_add_tail(&ps_buf->list, &psq->list);
psq->depth++;
psq->enqueued++;
spin_unlock_irqrestore(&psq->lock, flags);
return 0;
}
static int ath6kl_ps_queue_aging(struct ath6kl_ps_buf_head *psq)
{
struct ath6kl_ps_buf_desc *ps_buf, *ps_buf_n;
unsigned long flags;
u32 age;
int is_empty = 1;
spin_lock_irqsave(&psq->lock, flags);
if (psq->depth == 0) {
spin_unlock_irqrestore(&psq->lock, flags);
return is_empty;
}
list_for_each_entry_safe(ps_buf, ps_buf_n, &psq->list, list) {
age = ath6kl_ps_queue_get_age(ps_buf);
age++;
if ((psq->age_cycle != ATH6KL_PS_QUEUE_NO_AGE) &&
(age >= psq->age_cycle)) {
list_del(&ps_buf->list);
if (ps_buf->skb)
dev_kfree_skb(ps_buf->skb);
kfree(ps_buf);
psq->aged++;
psq->dequeued++;
psq->depth--;
} else
ath6kl_ps_queue_set_age(ps_buf, age);
}
is_empty = (psq->depth == 0);
spin_unlock_irqrestore(&psq->lock, flags);
ath6kl_dbg(ATH6KL_DBG_POWERSAVE,
"ps: aging psq %p depth %d aged %d\n",
psq,
psq->depth,
psq->aged);
return is_empty;
}
void ath6kl_ps_queue_age_handler(unsigned long ptr)
{
struct ath6kl_sta *conn = (struct ath6kl_sta *)ptr;
struct ath6kl_vif *vif = conn->vif;
spin_lock_bh(&conn->lock);
if (ath6kl_ps_queue_aging(&conn->psq_data) &&
ath6kl_ps_queue_aging(&conn->psq_mgmt))
ath6kl_wmi_set_pvb_cmd(vif->ar->wmi,
vif->fw_vif_idx, conn->aid, 0);
spin_unlock_bh(&conn->lock);
if (conn->psq_age_active)
mod_timer(&conn->psq_age_timer, jiffies +
msecs_to_jiffies(ATH6KL_PS_QUEUE_CHECK_AGE));
return;
}
void ath6kl_ps_queue_age_start(struct ath6kl_sta *conn)
{
ath6kl_dbg(ATH6KL_DBG_POWERSAVE,
"ps: aging_timer_start conn %p aid %d\n",
conn,
conn->aid);
if (conn->psq_age_active)
mod_timer(&conn->psq_age_timer, jiffies +
msecs_to_jiffies(ATH6KL_PS_QUEUE_CHECK_AGE));
return;
}
void ath6kl_ps_queue_age_stop(struct ath6kl_sta *conn)
{
ath6kl_dbg(ATH6KL_DBG_POWERSAVE,
"ps: aging_timer_stop conn %p aid %d\n",
conn,
conn->aid);
if (conn->psq_age_active)
del_timer_sync(&conn->psq_age_timer);
return;
}
struct bss_post_proc *ath6kl_bss_post_proc_init(struct ath6kl_vif *vif)
{
struct bss_post_proc *post_proc;
post_proc = kzalloc(sizeof(struct bss_post_proc), GFP_KERNEL);
if (!post_proc) {
ath6kl_err("failed to alloc memory for post_proc\n");
return NULL;
}
post_proc->vif = vif;
post_proc->flags = ATH6KL_BSS_POST_PROC_CACHED_BSS;
post_proc->aging_time = ATH6KL_BSS_POST_PROC_AGING_TIME;
spin_lock_init(&post_proc->bss_info_lock);
INIT_LIST_HEAD(&post_proc->bss_info_list);
ath6kl_dbg(ATH6KL_DBG_EXT_BSS_PROC,
"bss_proc init (vif %d)\n",
vif->fw_vif_idx);
return post_proc;
}
static void bss_proc_post_flush(struct bss_post_proc *post_proc, bool force)
{
struct bss_info_entry *bss_info, *tmp;
bool aging;
int aging_cnt, bss_cnt;
spin_lock_bh(&post_proc->bss_info_lock);
aging_cnt = bss_cnt = 0;
list_for_each_entry_safe(bss_info,
tmp,
&post_proc->bss_info_list,
list) {
bss_cnt++;
if (time_after(jiffies,
bss_info->shoot_time + post_proc->aging_time))
aging = true;
else
aging = false;
if ((force) || (aging)) {
aging_cnt++;
list_del(&bss_info->list);
kfree(bss_info->mgmt);
kfree(bss_info);
}
}
spin_unlock_bh(&post_proc->bss_info_lock);
ath6kl_dbg(ATH6KL_DBG_EXT_BSS_PROC,
"bss_proc flush %s, %d/%d\n",
(force ? "force" : "aging"),
aging_cnt,
bss_cnt);
return;
}
void ath6kl_bss_post_proc_deinit(struct ath6kl_vif *vif)
{
struct bss_post_proc *post_proc = vif->bss_post_proc_ctx;
if (post_proc) {
bss_proc_post_flush(post_proc, true);
kfree(post_proc);
}
vif->bss_post_proc_ctx = NULL;
ath6kl_dbg(ATH6KL_DBG_EXT_BSS_PROC,
"bss_proc deinit (vif %d)\n",
vif->fw_vif_idx);
return;
}
void ath6kl_bss_post_proc_bss_scan_start(struct ath6kl_vif *vif)
{
struct bss_post_proc *post_proc = vif->bss_post_proc_ctx;
if (!post_proc)
return;
/* Always flush old bss_info before start a new scan. */
if (post_proc->flags & ATH6KL_BSS_POST_PROC_CACHED_BSS)
bss_proc_post_flush(post_proc, false);
else
bss_proc_post_flush(post_proc, true);
set_bit(BSS_POST_PROC_SCAN_ONGOING, &post_proc->stat);
ath6kl_dbg(ATH6KL_DBG_EXT_BSS_PROC,
"bss_proc scan_start\n");
return;
}
int ath6kl_bss_post_proc_bss_complete_event(struct ath6kl_vif *vif)
{
struct bss_post_proc *post_proc = vif->bss_post_proc_ctx;
struct bss_info_entry *bss_info, *tmp;
struct cfg80211_bss *bss;
int cnt = 0;
if (!post_proc)
return 0;
if (!test_and_clear_bit(BSS_POST_PROC_SCAN_ONGOING, &post_proc->stat))
return 0;
spin_lock_bh(&post_proc->bss_info_lock);
list_for_each_entry_safe(bss_info,
tmp,
&post_proc->bss_info_list,
list) {
cnt++;
BUG_ON(!bss_info->mgmt);
bss = cfg80211_inform_bss_frame(vif->ar->wiphy,
bss_info->channel,
bss_info->mgmt,
bss_info->len,
bss_info->signal,
GFP_ATOMIC);
if (bss)
ath6kl_bss_put(vif->ar, bss);
}
spin_unlock_bh(&post_proc->bss_info_lock);
if (post_proc->flags & ATH6KL_BSS_POST_PROC_CACHED_BSS)
bss_proc_post_flush(post_proc, false);
else
bss_proc_post_flush(post_proc, true);
ath6kl_dbg(ATH6KL_DBG_EXT_BSS_PROC,
"bss_proc scan_comp, cnt %d\n", cnt);
return 0;
}
static struct bss_info_entry *bss_post_proc_bss_info(struct ath6kl_vif *vif,
struct ieee80211_mgmt *mgmt,
int len,
s32 snr,
struct ieee80211_channel *channel)
{
struct bss_post_proc *post_proc = vif->bss_post_proc_ctx;
struct bss_info_entry *bss_info, *tmp;
bool updated = false;
int hits = 0;
spin_lock_bh(&post_proc->bss_info_lock);
/*
* Expect reverse travel should hit in high possibility.
* And check some hints first to speed up. Also assume
* Duration & Sequence are always be zero and ignore the
* timestamp field.
*/
list_for_each_entry_safe_reverse(bss_info,
tmp,
&post_proc->bss_info_list,
list) {
BUG_ON(!bss_info->mgmt);
hits++;
if ((bss_info->channel == channel) &&
(bss_info->len == len) &&
(memcmp(bss_info->mgmt, mgmt, 24) == 0)) {
if (memcmp((u8 *)(bss_info->mgmt) + 24 + 8,
(u8 *)mgmt + 24 + 8,
len - 24 - 8) == 0) {
list_del(&bss_info->list);
/* Update */
memcpy((u8 *)(bss_info->mgmt), (u8 *)mgmt, len);
bss_info->shoot_time = jiffies;
bss_info->signal = snr; /* TODO: average SNR */
updated = true;
break;
}
}
}
/*
* Remove and insert back to the list to let the list is
* always keep shoot-time ordered.
*/
if (updated)
list_add_tail(&bss_info->list, &post_proc->bss_info_list);
else
bss_info = NULL;
spin_unlock_bh(&post_proc->bss_info_lock);
ath6kl_dbg(ATH6KL_DBG_EXT_BSS_PROC,
"bss_proc bssinfo %s, hits %d\n",
(updated ? "updated" : "add"),
hits);
return bss_info;
}
void ath6kl_bss_post_proc_bss_info(struct ath6kl_vif *vif,
struct ieee80211_mgmt *mgmt,
int len,
s32 snr,
struct ieee80211_channel *channel)
{
struct bss_post_proc *post_proc = vif->bss_post_proc_ctx;
struct bss_info_entry *bss_info;
if (!post_proc)
return;
if (!test_bit(BSS_POST_PROC_SCAN_ONGOING, &post_proc->stat))
return;
ath6kl_dbg(ATH6KL_DBG_EXT_BSS_PROC,
"bss_proc bssinfo (vif %d) BSSID "
"%02x:%02x:%02x:%02x:%02x:%02x FC %04x len %d\n",
vif->fw_vif_idx,
mgmt->bssid[0], mgmt->bssid[1], mgmt->bssid[2],
mgmt->bssid[3], mgmt->bssid[4], mgmt->bssid[5],
mgmt->frame_control,
len);
if (bss_post_proc_bss_info(vif, mgmt, len, snr, channel))
return;
bss_info = kzalloc(sizeof(struct bss_info_entry), GFP_ATOMIC);
if (!bss_info) {
ath6kl_err("failed to alloc memory for bss_info\n");
return;
}
bss_info->mgmt = kmalloc(len, GFP_ATOMIC);
if (!bss_info->mgmt) {
kfree(bss_info);
ath6kl_err("failed to alloc memory for bss_info->mgmt\n");
return;
}
/* Add BSS info. */
bss_info->channel = channel;
bss_info->signal = snr;
memcpy((u8 *)(bss_info->mgmt), (u8 *)mgmt, len);
bss_info->len = len;
bss_info->shoot_time = jiffies;
spin_lock_bh(&post_proc->bss_info_lock);
list_add_tail(&bss_info->list, &post_proc->bss_info_list);
spin_unlock_bh(&post_proc->bss_info_lock);
return;
}
static int bss_post_proc_candidate_chan(struct bss_info_entry *bss,
u16 *chan_list,
int chan_num)
{
u16 chan = bss->channel->center_freq;
int i, j, slot = -1;
bool add = true;
for (i = 0; i < chan_num; i++) {
if (chan == chan_list[i]) {
add = false;
break;
} else if (chan < chan_list[i])
break;
}
if (add) {
/* Order from lower to higher channel. */
if (i == chan_num)
slot = chan_num;
else {
slot = i;
for (j = 0; j < (chan_num - i); j++)
chan_list[chan_num - j] =
chan_list[chan_num - j - 1];
}
chan_list[slot] = chan;
chan_num++;
}
BUG_ON(slot >= chan_num);
ath6kl_dbg(ATH6KL_DBG_EXT_BSS_PROC,
"bss_proc candidate_chan %s chan %d slot %d chan_num %d\n",
(add ? "add" : "ignore"),
chan,
slot,
chan_num);
return chan_num;
}
int ath6kl_bss_post_proc_candidate_bss(struct ath6kl_vif *vif,
char *ssid,
int ssid_len,
u16 *chan_list)
{
struct bss_post_proc *post_proc = vif->bss_post_proc_ctx;
struct bss_info_entry *bss_info, *tmp;
u8 *ssid_ie;
int chan_num = 0;
if ((!post_proc) || (ssid_len == 0))
return 0;
spin_lock_bh(&post_proc->bss_info_lock);
list_for_each_entry_safe(bss_info,
tmp,
&post_proc->bss_info_list,
list) {
if (time_after(jiffies,
bss_info->shoot_time + post_proc->aging_time))
continue;
if (bss_info->len < 24 + 8 + 2 + 2 + 2 + 1)
continue;
BUG_ON(!bss_info->mgmt);
/* Always assume 1st IE is SSID IE. */
ssid_ie = &(bss_info->mgmt->u.beacon.variable[0]);
if ((ssid_ie[0] == WLAN_EID_SSID) &&
(ssid_ie[1] == ssid_len) &&
(memcmp(ssid_ie + 2, ssid, ssid_len) == 0))
chan_num = bss_post_proc_candidate_chan(bss_info,
chan_list,
chan_num);
}
spin_unlock_bh(&post_proc->bss_info_lock);
ath6kl_dbg(ATH6KL_DBG_EXT_BSS_PROC,
"bss_proc candidate_bss ssid %s chan_num %d\n",
ssid,
chan_num);
return chan_num;
}
void ath6kl_bss_post_proc_bss_config(struct ath6kl_vif *vif,
bool cache_bss,
int aging_time)
{
struct bss_post_proc *post_proc = vif->bss_post_proc_ctx;
if (!post_proc)
return;
if (cache_bss) {
post_proc->flags |= ATH6KL_BSS_POST_PROC_CACHED_BSS;
aging_time = msecs_to_jiffies(aging_time);
if (aging_time < ATH6KL_BSS_POST_PROC_AGING_TIME_MIN)
post_proc->aging_time =
ATH6KL_BSS_POST_PROC_AGING_TIME_MIN;
else
post_proc->aging_time = aging_time;
} else {
post_proc->flags &= ~ATH6KL_BSS_POST_PROC_CACHED_BSS;
post_proc->aging_time = ATH6KL_BSS_POST_PROC_AGING_TIME;
}
ath6kl_dbg(ATH6KL_DBG_EXT_BSS_PROC,
"bss_proc config, cache BSS %s aging time %d ms.",
(cache_bss ? "ON" : "OFF"),
jiffies_to_msecs(post_proc->aging_time));
return;
}
enum htc_endpoint_id ath6kl_ac2_endpoint_id(void *devt, u8 ac)
{
struct ath6kl *ar = devt;
return ar->ac2ep_map[ac];
}
static inline u8 *_cookie_malloc(int size, bool *alloc_from_vmalloc)
{
bool from_vmalloc = false;
u8 *ptr = NULL;
/* use kmalloc() first then vmalloc() if fail. */
ptr = kzalloc(size, GFP_ATOMIC);
if (!ptr) {
from_vmalloc = true;
ptr = vmalloc(size);
if (ptr)
memset(ptr, 0 , size);
}
*alloc_from_vmalloc = from_vmalloc;
return ptr;
}
static inline void _cookie_free(u8 *buf, bool alloc_from_vmalloc)
{
if (buf == NULL)
return;
if (alloc_from_vmalloc)
vfree(buf);
else
kfree(buf);
return;
}
static int ath6kl_cookie_pool_init(struct ath6kl *ar,
struct ath6kl_cookie_pool *cookie_pool,
enum cookie_type cookie_type,
u32 cookie_num)
{
u32 i, j, mem_size;
bool alloc_from_vmalloc = false;
WARN_ON(!cookie_num);
cookie_pool->cookie_type = cookie_type;
cookie_pool->cookie_list = NULL;
cookie_pool->cookie_count = 0;
mem_size = sizeof(struct ath6kl_cookie) * cookie_num;
cookie_pool->cookie_mem =
(struct ath6kl_cookie *)_cookie_malloc(mem_size,
&alloc_from_vmalloc);
if (!cookie_pool->cookie_mem) {
cookie_pool->cookie_num = 0;
ath6kl_err("unable to allocate cookie, type %d num %d\n",
cookie_type,
cookie_num);
return -ENOMEM;
} else {
cookie_pool->cookie_num = cookie_num;
cookie_pool->cookie_mem->alloc_from_vmalloc =
alloc_from_vmalloc;
}
for (i = 0; i < cookie_num; i++) {
/* Assign the parent then insert to free queue */
cookie_pool->cookie_mem[i].cookie_pool = cookie_pool;
cookie_pool->cookie_mem[i].htc_pkt =
(struct htc_packet *)_cookie_malloc(
sizeof(struct htc_packet),
&alloc_from_vmalloc);
if (cookie_pool->cookie_mem[i].htc_pkt) {
cookie_pool->cookie_mem[i].htc_pkt->alloc_from_vmalloc =
alloc_from_vmalloc;
ath6kl_free_cookie(ar, &cookie_pool->cookie_mem[i]);
} else {
struct htc_packet *htc_pkt;
ath6kl_err("unable to allocate htc_pkt\n");
for (j = 0 ; j < i ; j++) {
htc_pkt = cookie_pool->cookie_mem[j].htc_pkt;
if (htc_pkt)
_cookie_free((u8 *)htc_pkt,
htc_pkt->alloc_from_vmalloc);
cookie_pool->cookie_mem[j].htc_pkt = NULL;
}
_cookie_free((u8 *)(cookie_pool->cookie_mem),
cookie_pool->cookie_mem->alloc_from_vmalloc);
cookie_pool->cookie_mem = NULL;
cookie_pool->cookie_count = 0;
cookie_pool->cookie_num = 0;
return -ENOMEM;
}
}
/* Reset stats */
cookie_pool->cookie_alloc_cnt = 0;
cookie_pool->cookie_alloc_fail_cnt = 0;
cookie_pool->cookie_free_cnt = 0;
cookie_pool->cookie_peak_cnt = 0;
cookie_pool->cookie_fail_in_row = 0;
ath6kl_info("Create HTC cookie, type %d num %d\n"
"pool from vmalloc %d htc from vmalloc %d",
cookie_type,
cookie_num,
cookie_pool->cookie_mem->alloc_from_vmalloc,
alloc_from_vmalloc);
return 0;
}
static void ath6kl_cookie_pool_cleanup(struct ath6kl *ar,
struct ath6kl_cookie_pool *cookie_pool)
{
int i;
if (cookie_pool->cookie_num != cookie_pool->cookie_count)
ath6kl_err("Cookie unmber unsync, type %d num %d, %d\n",
cookie_pool->cookie_type,
cookie_pool->cookie_num,
cookie_pool->cookie_count);
ath6kl_info("Free HTC cookie, type %d curr_num %d\n",
cookie_pool->cookie_type,
cookie_pool->cookie_count);
cookie_pool->cookie_list = NULL;
cookie_pool->cookie_count = 0;
if (cookie_pool->cookie_mem) {
for (i = 0; i < cookie_pool->cookie_num; i++) {
struct htc_packet *htc_pkt;
htc_pkt = cookie_pool->cookie_mem[i].htc_pkt;
if (htc_pkt)
_cookie_free((u8 *)htc_pkt,
htc_pkt->alloc_from_vmalloc);
cookie_pool->cookie_mem[i].htc_pkt = NULL;
}
_cookie_free((u8 *)(cookie_pool->cookie_mem),
cookie_pool->cookie_mem->alloc_from_vmalloc);
cookie_pool->cookie_mem = NULL;
}
cookie_pool->cookie_alloc_cnt = 0;
cookie_pool->cookie_alloc_fail_cnt = 0;
cookie_pool->cookie_free_cnt = 0;
cookie_pool->cookie_peak_cnt = 0;
cookie_pool->cookie_fail_in_row = 0;
return;
}
struct ath6kl_cookie *ath6kl_alloc_cookie(struct ath6kl *ar,
enum cookie_type cookie_type)
{
struct ath6kl_cookie_pool *cookie_pool;
struct ath6kl_cookie *cookie;
u32 alloced;
if (cookie_type == COOKIE_TYPE_DATA)
cookie_pool = &ar->cookie_data;
else if (cookie_type == COOKIE_TYPE_CTRL)
cookie_pool = &ar->cookie_ctrl;
else
BUG_ON(1);
cookie = cookie_pool->cookie_list;
if (cookie != NULL) {
cookie_pool->cookie_list = cookie->arc_list_next;
cookie_pool->cookie_count--;
alloced = cookie_pool->cookie_num - cookie_pool->cookie_count;
cookie_pool->cookie_alloc_cnt++;
cookie_pool->cookie_fail_in_row = 0;
if (alloced > cookie_pool->cookie_peak_cnt)
cookie_pool->cookie_peak_cnt = alloced;
} else {
cookie_pool->cookie_alloc_fail_cnt++;
cookie_pool->cookie_fail_in_row++;
}
return cookie;
}
int ath6kl_cookie_init(struct ath6kl *ar)
{
/* Create HTC cookies pool for DATA frame */
if (ath6kl_cookie_pool_init(ar,
&ar->cookie_data,
COOKIE_TYPE_DATA,
MAX_COOKIE_DATA_NUM))
goto fail;
/* Create HTC cookies pool for CTRL command */
if (ath6kl_cookie_pool_init(ar,
&ar->cookie_ctrl,
COOKIE_TYPE_CTRL,
MAX_COOKIE_CTRL_NUM))
goto fail;
return 0;
fail:
ath6kl_cookie_cleanup(ar);
return -ENOMEM;
}
void ath6kl_cookie_cleanup(struct ath6kl *ar)
{
ath6kl_cookie_pool_cleanup(ar,
&ar->cookie_data);
ath6kl_cookie_pool_cleanup(ar,
&ar->cookie_ctrl);
return;
}
void ath6kl_free_cookie(struct ath6kl *ar, struct ath6kl_cookie *cookie)
{
struct ath6kl_cookie_pool *cookie_pool;
if (!ar || !cookie)
return;
cookie_pool = cookie->cookie_pool;
BUG_ON(!cookie_pool);
cookie->arc_list_next = cookie_pool->cookie_list;
cookie_pool->cookie_list = cookie;
cookie_pool->cookie_count++;
cookie_pool->cookie_free_cnt++;
return;
}
bool ath6kl_cookie_is_almost_full(struct ath6kl *ar,
enum cookie_type cookie_type)
{
struct ath6kl_cookie_pool *cookie_pool;
bool almost_full;
if (cookie_type == COOKIE_TYPE_DATA)
cookie_pool = &ar->cookie_data;
else if (cookie_type == COOKIE_TYPE_CTRL)
cookie_pool = &ar->cookie_ctrl;
else
BUG_ON(1);
if (cookie_pool->cookie_count < MAX_RESV_COOKIE_NUM)
almost_full = true;
else
almost_full = false;
return almost_full;
}
int ath6kl_diag_warm_reset(struct ath6kl *ar)
{
int ret;
ret = ath6kl_hif_diag_warm_reset(ar);
if (ret) {
ath6kl_err("failed to issue warm reset command\n");
return ret;
}
return 0;
}
/*
* Read from the hardware through its diagnostic window. No cooperation
* from the firmware is required for this.
*/
int ath6kl_diag_read32(struct ath6kl *ar, u32 address, u32 *value)
{
int ret;
ret = ath6kl_hif_diag_read32(ar, address, value);
if (ret) {
ath6kl_warn("failed to read32 through diagnose window: %d\n",
ret);
return ret;
}
return 0;
}
/*
* Write to the ATH6KL through its diagnostic window. No cooperation from
* the Target is required for this.
*/
int ath6kl_diag_write32(struct ath6kl *ar, u32 address, __le32 value)
{
int ret;
ret = ath6kl_hif_diag_write32(ar, address, value);
if (ret) {
ath6kl_err("failed to write 0x%x during diagnose window to 0x%d\n",
address, value);
return ret;
}
return 0;
}
int ath6kl_diag_read(struct ath6kl *ar, u32 address, void *data, u32 length)
{
u32 count, *buf = data;
int ret;
if (WARN_ON(length % 4))
return -EINVAL;
for (count = 0; count < length / 4; count++, address += 4) {
ret = ath6kl_diag_read32(ar, address, &buf[count]);
if (ret)
return ret;
}
return 0;
}
int ath6kl_diag_write(struct ath6kl *ar, u32 address, void *data, u32 length)
{
u32 count;
__le32 *buf = data;
int ret;
if (WARN_ON(length % 4))
return -EINVAL;
for (count = 0; count < length / 4; count++, address += 4) {
ret = ath6kl_diag_write32(ar, address, buf[count]);
if (ret)
return ret;
}
return 0;
}
int ath6kl_read_fwlogs(struct ath6kl *ar)
{
struct ath6kl_dbglog_hdr debug_hdr;
struct ath6kl_dbglog_buf debug_buf;
u32 address, length, dropped, firstbuf, debug_hdr_addr;
int ret = 0, loop;
u8 *buf;
buf = kmalloc(ATH6KL_FWLOG_PAYLOAD_SIZE, GFP_KERNEL);
if (!buf)
return -ENOMEM;
address = TARG_VTOP(ar->target_type,
ath6kl_get_hi_item_addr(ar,
HI_ITEM(hi_dbglog_hdr)));
ret = ath6kl_diag_read32(ar, address, &debug_hdr_addr);
if (ret)
goto out;
/* Get the contents of the ring buffer */
if (debug_hdr_addr == 0) {
ath6kl_warn("Invalid address for debug_hdr_addr\n");
ret = -EINVAL;
goto out;
}
address = TARG_VTOP(ar->target_type, debug_hdr_addr);
ath6kl_diag_read(ar, address, &debug_hdr, sizeof(debug_hdr));
address = TARG_VTOP(ar->target_type,
le32_to_cpu(debug_hdr.dbuf_addr));
firstbuf = address;
dropped = le32_to_cpu(debug_hdr.dropped);
ath6kl_diag_read(ar, address, &debug_buf, sizeof(debug_buf));
loop = 100;
do {
address = TARG_VTOP(ar->target_type,
le32_to_cpu(debug_buf.buffer_addr));
length = le32_to_cpu(debug_buf.length);
if (length != 0 && (le32_to_cpu(debug_buf.length) <=
le32_to_cpu(debug_buf.bufsize))) {
length = ALIGN(length, 4);
ret = ath6kl_diag_read(ar, address,
buf, length);
if (ret)
goto out;
ath6kl_debug_fwlog_event(ar, buf, length);
}
address = TARG_VTOP(ar->target_type,
le32_to_cpu(debug_buf.next));
ret = ath6kl_diag_read(ar, address, &debug_buf,
sizeof(debug_buf));
if (ret)
goto out;
loop--;
if (WARN_ON(loop == 0)) {
ret = -ETIMEDOUT;
goto out;
}
} while (address != firstbuf);
out:
kfree(buf);
return ret;
}
/* FIXME: move to a better place, target.h? */
#define AR6003_RESET_CONTROL_ADDRESS 0x00004000
#define AR6004_RESET_CONTROL_ADDRESS 0x00004000
#define AR6006_RESET_CONTROL_ADDRESS 0x00004000
void ath6kl_reset_device(struct ath6kl *ar, u32 target_type,
bool wait_fot_compltn, bool cold_reset)
{
int status = 0;
u32 address;
__le32 data;
#ifdef USB_AUTO_SUSPEND
struct ath6kl_vif *vif;
#endif
if (target_type != TARGET_TYPE_AR6003 &&
target_type != TARGET_TYPE_AR6004 &&
target_type != TARGET_TYPE_AR6006)
return;
data = cold_reset ? cpu_to_le32(RESET_CONTROL_COLD_RST) :
cpu_to_le32(RESET_CONTROL_MBOX_RST);
switch (target_type) {
case TARGET_TYPE_AR6004:
address = AR6004_RESET_CONTROL_ADDRESS;
break;
case TARGET_TYPE_AR6006:
address = AR6006_RESET_CONTROL_ADDRESS;
break;
default:
address = AR6003_RESET_CONTROL_ADDRESS;
break;
}
#ifdef USB_AUTO_SUSPEND
vif = ath6kl_vif_first(ar);
if (vif != NULL) {
if (ar->state == ATH6KL_STATE_WOW ||
ar->state == ATH6KL_STATE_DEEPSLEEP) {
ath6kl_hif_auto_pm_turnoff(ar);
msleep(20);
} else {
if (ar->autopm_turn_on) {
ar->autopm_defer_delay_change_cnt =
USB_SUSPEND_DEFER_DELAY_FOR_RECOVER;
ath6kl_hif_auto_pm_set_delay(ar,
USB_SUSPEND_DELAY_MAX);
}
}
}
#endif
/* If the bootstrap mode is HSIC, do warm reset */
if (BOOTSTRAP_IS_HSIC(ar->bootstrap_mode)) {
status = ath6kl_diag_warm_reset(ar);
ath6kl_info("%s: warm reset\n", __func__);
} else {
status = ath6kl_diag_write32(ar, address, data);
}
if (status)
ath6kl_err("failed to reset target\n");
}
void ath6kl_fw_crash_notify(struct ath6kl *ar)
{
struct ath6kl_vif *vif = ath6kl_vif_first(ar);
if (vif == NULL)
return;
ath6kl_info("notify firmware crash to user %p\n", ar);
/* TODO */
#ifdef ATH6KL_HSIC_RECOVER
if (BOOTSTRAP_IS_HSIC(ar->bootstrap_mode))
ath6kl_hif_sw_recover(ar);
#endif
return;
}
int ath6kl_fw_watchdog_enable(struct ath6kl *ar)
{
u32 param;
int ret;
ret = ath6kl_diag_read32(ar,
ath6kl_get_hi_item_addr(ar, HI_ITEM(hi_option_flag2)),
(u32 *)&param);
if (ret != 0)
return ret;
param |= HI_OPTION_FW_WATCHDOG_ENABLE;
ret = ath6kl_diag_write32(ar,
ath6kl_get_hi_item_addr(ar, HI_ITEM(hi_option_flag2)),
param);
return ret;
}
int ath6kl_fw_crash_cold_reset_enable(struct ath6kl *ar)
{
u32 param;
int ret;
ret = ath6kl_diag_read32(ar,
ath6kl_get_hi_item_addr(ar, HI_ITEM(hi_option_flag2)),
(u32 *)&param);
if (ret != 0)
return ret;
param |= HI_OPTION_FW_CRASH_COLD_RESET;
ret = ath6kl_diag_write32(ar,
ath6kl_get_hi_item_addr(ar, HI_ITEM(hi_option_flag2)),
param);
return ret;
}
static void ath6kl_install_static_wep_keys(struct ath6kl_vif *vif)
{
u8 index;
u8 keyusage;
for (index = WMI_MIN_KEY_INDEX; index <= WMI_MAX_KEY_INDEX; index++) {
if (vif->wep_key_list[index].key_len) {
keyusage = GROUP_USAGE;
if (index == vif->def_txkey_index)
keyusage |= TX_USAGE;
ath6kl_wmi_addkey_cmd(vif->ar->wmi, vif->fw_vif_idx,
index,
WEP_CRYPT,
keyusage,
vif->wep_key_list[index].key_len,
NULL, 0,
vif->wep_key_list[index].key,
KEY_OP_INIT_VAL, NULL,
NO_SYNC_WMIFLAG);
}
}
}
void ath6kl_connect_ap_mode_bss(struct ath6kl_vif *vif, u16 channel,
u8 *beacon, u8 beacon_len)
{
struct ath6kl *ar = vif->ar;
struct ath6kl_req_key *ik;
int res;
u8 key_rsc[ATH6KL_KEY_SEQ_LEN];
ik = &vif->ap_mode_bkey;
ath6kl_dbg(ATH6KL_DBG_WLAN_CFG, "AP mode started on %u MHz\n", channel);
vif->bss_ch = channel;
ath6kl_ap_beacon_info(vif, beacon, beacon_len);
switch (vif->auth_mode) {
case NONE_AUTH:
if (vif->prwise_crypto == WEP_CRYPT)
ath6kl_install_static_wep_keys(vif);
break;
case WPA_PSK_AUTH:
case WPA2_PSK_AUTH:
case (WPA_PSK_AUTH | WPA2_PSK_AUTH):
if (!ik->valid)
break;
ath6kl_dbg(ATH6KL_DBG_WLAN_CFG, "Delayed addkey for "
"the initial group key for AP mode\n");
memset(key_rsc, 0, sizeof(key_rsc));
res = ath6kl_wmi_addkey_cmd(
ar->wmi, vif->fw_vif_idx, ik->key_index, ik->key_type,
GROUP_USAGE, ik->key_len, key_rsc, ATH6KL_KEY_SEQ_LEN,
ik->key,
KEY_OP_INIT_VAL, NULL, SYNC_BOTH_WMIFLAG);
if (res) {
ath6kl_dbg(ATH6KL_DBG_WLAN_CFG, "Delayed "
"addkey failed: %d\n", res);
}
break;
}
ath6kl_wmi_bssfilter_cmd(ar->wmi, vif->fw_vif_idx, NONE_BSS_FILTER, 0);
set_bit(CONNECTED, &vif->flags);
ath6kl_judge_roam_parameter(vif, false);
if (ath6kl_wmi_set_rate_ctrl_cmd(ar->wmi,
vif->fw_vif_idx, RATECTRL_MODE_PERONLY))
ath6kl_err("set rate_ctrl failed\n");
ath6kl_switch_parameter_based_on_connection(vif, false);
netif_carrier_on(vif->ndev);
}
void ath6kl_connect_ap_mode_sta(struct ath6kl_vif *vif, u8 aid, u8 *mac_addr,
u8 keymgmt, u8 ucipher, u8 auth,
u16 assoc_req_len, u8 *assoc_info,
u8 apsd_info, u8 phymode)
{
u8 *ies = NULL, *wpa_ie = NULL, *pos;
size_t ies_len = 0;
struct station_info sinfo;
bool is_ht_sta = false;
ath6kl_dbg(ATH6KL_DBG_TRC, "new station %pM aid=%d\n", mac_addr, aid);
if (assoc_req_len > sizeof(struct ieee80211_hdr_3addr)) {
struct ieee80211_mgmt *mgmt =
(struct ieee80211_mgmt *) assoc_info;
if (ieee80211_is_assoc_req(mgmt->frame_control) &&
assoc_req_len >= sizeof(struct ieee80211_hdr_3addr) +
sizeof(mgmt->u.assoc_req)) {
ies = mgmt->u.assoc_req.variable;
ies_len = assoc_info + assoc_req_len - ies;
} else if (ieee80211_is_reassoc_req(mgmt->frame_control) &&
assoc_req_len >= sizeof(struct ieee80211_hdr_3addr)
+ sizeof(mgmt->u.reassoc_req)) {
ies = mgmt->u.reassoc_req.variable;
ies_len = assoc_info + assoc_req_len - ies;
}
}
pos = ies;
while (pos && pos + 1 < ies + ies_len) {
if (pos + 2 + pos[1] > ies + ies_len)
break;
if (pos[0] == WLAN_EID_RSN)
wpa_ie = pos; /* RSN IE */
else if (pos[0] == WLAN_EID_VENDOR_SPECIFIC &&
pos[1] >= 4 &&
pos[2] == 0x00 && pos[3] == 0x50 && pos[4] == 0xf2) {
if (pos[5] == 0x01)
wpa_ie = pos; /* WPA IE */
else if (pos[5] == 0x04) {
wpa_ie = pos; /* WPS IE */
break; /* overrides WPA/RSN IE */
}
} else if (pos[0] == WLAN_EID_HT_CAPABILITY &&
pos[1] >= 22)
is_ht_sta = true;
else if (pos[0] == 0x44 && wpa_ie == NULL) {
/*
* Note: WAPI Parameter Set IE re-uses Element ID that
* was officially allocated for BSS AC Access Delay. As
* such, we need to be a bit more careful on when
* parsing the frame. However, BSS AC Access Delay
* element is not supposed to be included in
* (Re)Association Request frames, so this should not
* cause problems.
*/
wpa_ie = pos; /* WAPI IE */
break;
}
pos += 2 + pos[1];
}
ath6kl_add_new_sta(vif, mac_addr, aid, wpa_ie,
wpa_ie ? 2 + wpa_ie[1] : 0,
keymgmt, ucipher, auth, apsd_info,
is_ht_sta, phymode);
/* send event to application */
memset(&sinfo, 0, sizeof(sinfo));
/* TODO: sinfo.generation */
sinfo.assoc_req_ies = ies;
sinfo.assoc_req_ies_len = ies_len;
sinfo.filled |= STATION_INFO_ASSOC_REQ_IES;
cfg80211_new_sta(vif->ndev, mac_addr, &sinfo, GFP_KERNEL);
netif_wake_queue(vif->ndev);
}
void disconnect_timer_handler(unsigned long ptr)
{
struct net_device *dev = (struct net_device *)ptr;
struct ath6kl_vif *vif = netdev_priv(dev);
ath6kl_init_profile_info(vif);
ath6kl_disconnect(vif);
}
int ath6kl_disconnect(struct ath6kl_vif *vif)
{
int ret = 0;
if (test_bit(CONNECTED, &vif->flags) ||
test_bit(CONNECT_PEND, &vif->flags)) {
ret = ath6kl_wmi_disconnect_cmd(vif->ar->wmi, vif->fw_vif_idx);
/*
* Disconnect command is issued, clear the connect pending
* flag. The connected flag will be cleared in
* disconnect event notification.
*/
clear_bit(CONNECT_PEND, &vif->flags);
}
return ret;
}
/* WMI Event handlers */
void ath6kl_ready_event(void *devt, u8 *datap, u32 sw_ver, u32 abi_ver)
{
struct ath6kl *ar = devt;
memcpy(ar->mac_addr, datap, ETH_ALEN);
ath6kl_dbg(ATH6KL_DBG_TRC, "%s: mac addr = %pM\n",
__func__, ar->mac_addr);
ar->version.wlan_ver = sw_ver;
ar->version.abi_ver = abi_ver;
snprintf(ar->wiphy->fw_version,
sizeof(ar->wiphy->fw_version),
"%u.%u.%u.%u",
(ar->version.wlan_ver & 0xf0000000) >> 28,
(ar->version.wlan_ver & 0x0f000000) >> 24,
(ar->version.wlan_ver & 0x00ff0000) >> 16,
(ar->version.wlan_ver & 0x0000ffff));
#ifdef ATH6KL_HSIC_RECOVER
memcpy(cached_mac, ar->mac_addr, ETH_ALEN);
cached_mac_valid = true;
#endif
/* indicate to the waiting thread that the ready event was received */
set_bit(WMI_READY, &ar->flag);
wake_up(&ar->event_wq);
}
void ath6kl_scan_complete_evt(struct ath6kl_vif *vif, int status)
{
struct ath6kl *ar = vif->ar;
bool aborted = false;
if (status != WMI_SCAN_STATUS_SUCCESS)
aborted = true;
#ifdef ACS_SUPPORT
/* FIXME : bad, may use call-back instead. */
ath6kl_acs_scan_complete_event(vif, aborted);
#endif
ath6kl_bss_post_proc_bss_complete_event(vif);
ath6kl_p2p_rc_scan_complete_event(vif, aborted);
if (ath6kl_htcoex_scan_complete_event(vif, aborted) ==
HTCOEX_PASS_SCAN_DONE)
ath6kl_cfg80211_scan_complete_event(vif, aborted);
if (!vif->usr_bss_filter) {
clear_bit(CLEAR_BSSFILTER_ON_BEACON, &vif->flags);
ath6kl_wmi_bssfilter_cmd(ar->wmi, vif->fw_vif_idx,
NONE_BSS_FILTER, 0);
}
ath6kl_dbg(ATH6KL_DBG_WLAN_CFG |
ATH6KL_DBG_EXT_INFO1 |
ATH6KL_DBG_EXT_SCAN,
"scan complete: %d\n",
status);
}
void ath6kl_connect_event(struct ath6kl_vif *vif, u16 channel, u8 *bssid,
u16 listen_int, u16 beacon_int,
enum network_type net_type, u8 beacon_ie_len,
u8 assoc_req_len, u8 assoc_resp_len,
u8 *assoc_info)
{
struct ath6kl *ar = vif->ar;
ath6kl_cfg80211_connect_event(vif, channel, bssid,
listen_int, beacon_int,
net_type, beacon_ie_len,
assoc_req_len, assoc_resp_len,
assoc_info);
memcpy(vif->bssid, bssid, sizeof(vif->bssid));
vif->bss_ch = channel;
if ((vif->nw_type == INFRA_NETWORK)) {
if (ar->wiphy->flags & WIPHY_FLAG_SUPPORTS_FW_ROAM &&
vif->wdev.iftype == NL80211_IFTYPE_STATION) {
ath6kl_wmi_disctimeout_cmd(ar->wmi, vif->fw_vif_idx,
ATH6KL_SEAMLESS_ROAMING_DISCONNECT_TIMEOUT);
} else {
#ifdef CE_SUPPORT
ath6kl_wmi_disctimeout_cmd(ar->wmi, vif->fw_vif_idx,
0);
#else
ath6kl_wmi_disctimeout_cmd(ar->wmi, vif->fw_vif_idx,
ATH6KL_DISCONNECT_TIMEOUT);
#endif
}
ath6kl_wmi_listeninterval_cmd(ar->wmi, vif->fw_vif_idx,
ar->listen_intvl_t,
ar->listen_intvl_b);
}
netif_wake_queue(vif->ndev);
/* Update connect & link status atomically */
spin_lock_bh(&vif->if_lock);
set_bit(CONNECTED, &vif->flags);
clear_bit(CONNECT_PEND, &vif->flags);
netif_carrier_on(vif->ndev);
spin_unlock_bh(&vif->if_lock);
aggr_reset_state(vif->sta_list[0].aggr_conn_cntxt);
vif->reconnect_flag = 0;
if ((vif->nw_type == ADHOC_NETWORK) && ar->ibss_ps_enable) {
memset(ar->node_map, 0, sizeof(ar->node_map));
ar->node_num = 0;
ar->next_ep_id = ENDPOINT_2;
}
if (!vif->usr_bss_filter) {
set_bit(CLEAR_BSSFILTER_ON_BEACON, &vif->flags);
ath6kl_wmi_bssfilter_cmd(ar->wmi, vif->fw_vif_idx,
CURRENT_BSS_FILTER, 0);
}
/* Hook connection event */
ath6kl_htcoex_connect_event(vif);
ath6kl_p2p_connect_event(vif,
beacon_ie_len,
assoc_req_len,
assoc_resp_len,
assoc_info);
aggr_tx_connect_event(vif,
beacon_ie_len,
assoc_req_len,
assoc_resp_len,
assoc_info);
ath6kl_switch_parameter_based_on_connection(vif, false);
}
void ath6kl_tkip_micerr_event(struct ath6kl_vif *vif, u8 keyid, bool ismcast)
{
struct ath6kl_sta *sta;
u8 tsc[6];
/*
* For AP case, keyid will have aid of STA which sent pkt with
* MIC error. Use this aid to get MAC & send it to hostapd.
*/
if (vif->nw_type == AP_NETWORK) {
sta = ath6kl_find_sta_by_aid(vif, (keyid >> 2));
if (!sta)
return;
ath6kl_dbg(ATH6KL_DBG_TRC,
"ap tkip mic error received from aid=%d\n", keyid);
memset(tsc, 0, sizeof(tsc)); /* FIX: get correct TSC */
cfg80211_michael_mic_failure(vif->ndev, sta->mac,
NL80211_KEYTYPE_PAIRWISE, keyid,
tsc, GFP_KERNEL);
} else
ath6kl_cfg80211_tkip_micerr_event(vif, keyid, ismcast);
}
static void ath6kl_update_target_stats(struct ath6kl_vif *vif, u8 *ptr, u32 len)
{
struct wmi_target_stats *tgt_stats =
(struct wmi_target_stats *) ptr;
struct ath6kl *ar = vif->ar;
struct target_stats *stats = &vif->target_stats;
struct tkip_ccmp_stats *ccmp_stats;
u8 ac;
if (len < sizeof(*tgt_stats))
return;
ath6kl_dbg(ATH6KL_DBG_TRC, "updating target stats\n");
stats->tx_pkt += le32_to_cpu(tgt_stats->stats.tx.pkt);
stats->tx_byte += le32_to_cpu(tgt_stats->stats.tx.byte);
stats->tx_ucast_pkt += le32_to_cpu(tgt_stats->stats.tx.ucast_pkt);
stats->tx_ucast_byte += le32_to_cpu(tgt_stats->stats.tx.ucast_byte);
stats->tx_mcast_pkt += le32_to_cpu(tgt_stats->stats.tx.mcast_pkt);
stats->tx_mcast_byte += le32_to_cpu(tgt_stats->stats.tx.mcast_byte);
stats->tx_bcast_pkt += le32_to_cpu(tgt_stats->stats.tx.bcast_pkt);
stats->tx_bcast_byte += le32_to_cpu(tgt_stats->stats.tx.bcast_byte);
stats->tx_rts_success_cnt +=
le32_to_cpu(tgt_stats->stats.tx.rts_success_cnt);
for (ac = 0; ac < WMM_NUM_AC; ac++)
stats->tx_pkt_per_ac[ac] +=
le32_to_cpu(tgt_stats->stats.tx.pkt_per_ac[ac]);
stats->tx_err += le32_to_cpu(tgt_stats->stats.tx.err);
stats->tx_fail_cnt += le32_to_cpu(tgt_stats->stats.tx.fail_cnt);
stats->tx_retry_cnt += le32_to_cpu(tgt_stats->stats.tx.retry_cnt);
stats->tx_mult_retry_cnt +=
le32_to_cpu(tgt_stats->stats.tx.mult_retry_cnt);
stats->tx_rts_fail_cnt +=
le32_to_cpu(tgt_stats->stats.tx.rts_fail_cnt);
if (ar->target_type == TARGET_TYPE_AR6004) {
stats->tx_ucast_rate =
ath6kl_wmi_get_rate_ar6004(
a_sle32_to_cpu(tgt_stats->stats.tx.ucast_rate));
stats->tx_rate_index = tgt_stats->stats.tx.ucast_rate;
} else {
stats->tx_ucast_rate =
ath6kl_wmi_get_rate(
a_sle32_to_cpu(tgt_stats->stats.tx.ucast_rate));
}
stats->rx_pkt += le32_to_cpu(tgt_stats->stats.rx.pkt);
stats->rx_byte += le32_to_cpu(tgt_stats->stats.rx.byte);
stats->rx_ucast_pkt += le32_to_cpu(tgt_stats->stats.rx.ucast_pkt);
stats->rx_ucast_byte += le32_to_cpu(tgt_stats->stats.rx.ucast_byte);
stats->rx_mcast_pkt += le32_to_cpu(tgt_stats->stats.rx.mcast_pkt);
stats->rx_mcast_byte += le32_to_cpu(tgt_stats->stats.rx.mcast_byte);
stats->rx_bcast_pkt += le32_to_cpu(tgt_stats->stats.rx.bcast_pkt);
stats->rx_bcast_byte += le32_to_cpu(tgt_stats->stats.rx.bcast_byte);
stats->rx_frgment_pkt += le32_to_cpu(tgt_stats->stats.rx.frgment_pkt);
stats->rx_err += le32_to_cpu(tgt_stats->stats.rx.err);
stats->rx_crc_err += le32_to_cpu(tgt_stats->stats.rx.crc_err);
stats->rx_key_cache_miss +=
le32_to_cpu(tgt_stats->stats.rx.key_cache_miss);
stats->rx_decrypt_err += le32_to_cpu(tgt_stats->stats.rx.decrypt_err);
stats->rx_dupl_frame += le32_to_cpu(tgt_stats->stats.rx.dupl_frame);
stats->rx_ucast_rate =
ath6kl_wmi_get_rate(a_sle32_to_cpu(tgt_stats->stats.rx.ucast_rate));
ccmp_stats = &tgt_stats->stats.tkip_ccmp_stats;
stats->tkip_local_mic_fail +=
le32_to_cpu(ccmp_stats->tkip_local_mic_fail);
stats->tkip_cnter_measures_invoked +=
le32_to_cpu(ccmp_stats->tkip_cnter_measures_invoked);
stats->tkip_fmt_err += le32_to_cpu(ccmp_stats->tkip_fmt_err);
stats->ccmp_fmt_err += le32_to_cpu(ccmp_stats->ccmp_fmt_err);
stats->ccmp_replays += le32_to_cpu(ccmp_stats->ccmp_replays);
stats->pwr_save_fail_cnt +=
le32_to_cpu(tgt_stats->pm_stats.pwr_save_failure_cnt);
stats->noise_floor_calib =
a_sle32_to_cpu(tgt_stats->noise_floor_calib);
stats->cs_bmiss_cnt +=
le32_to_cpu(tgt_stats->cserv_stats.cs_bmiss_cnt);
stats->cs_low_rssi_cnt +=
le32_to_cpu(tgt_stats->cserv_stats.cs_low_rssi_cnt);
stats->cs_connect_cnt +=
le16_to_cpu(tgt_stats->cserv_stats.cs_connect_cnt);
stats->cs_discon_cnt +=
le16_to_cpu(tgt_stats->cserv_stats.cs_discon_cnt);
stats->cs_roam_cnt +=
le16_to_cpu(tgt_stats->cserv_stats.cs_roam_count);
stats->cs_ave_beacon_rssi =
a_sle16_to_cpu(tgt_stats->cserv_stats.cs_ave_beacon_rssi);
stats->cs_last_roam_msec =
tgt_stats->cserv_stats.cs_last_roam_msec;
stats->cs_snr = tgt_stats->cserv_stats.cs_snr;
stats->cs_rssi = a_sle16_to_cpu(tgt_stats->cserv_stats.cs_rssi);
stats->lq_val = le32_to_cpu(tgt_stats->lq_val);
stats->wow_pkt_dropped +=
le32_to_cpu(tgt_stats->wow_stats.wow_pkt_dropped);
stats->wow_host_pkt_wakeups +=
tgt_stats->wow_stats.wow_host_pkt_wakeups;
stats->wow_host_evt_wakeups +=
tgt_stats->wow_stats.wow_host_evt_wakeups;
stats->wow_evt_discarded +=
le16_to_cpu(tgt_stats->wow_stats.wow_evt_discarded);
do_gettimeofday(&stats->update_time);
}
static void ath6kl_add_le32(__le32 *var, __le32 val)
{
*var = cpu_to_le32(le32_to_cpu(*var) + le32_to_cpu(val));
}
void ath6kl_tgt_stats_event(struct ath6kl_vif *vif, u8 *ptr, u32 len)
{
struct wmi_ap_mode_stat *p = (struct wmi_ap_mode_stat *) ptr;
struct wmi_ap_mode_stat *ap = &vif->ap_stats;
struct wmi_per_sta_stat *st_ap, *st_p;
u8 ac, slot;
u16 updated = 0;
if (vif->nw_type == AP_NETWORK) {
if ((len + 4) >= sizeof(*p)) {
for (ac = 0; ac < AP_MAX_NUM_STA; ac++) {
st_p = &p->sta[ac];
/*
* Target may insert garbage data and only
* update the associated stations.
*/
if ((st_p->aid == 0) ||
(st_p->aid > AP_MAX_NUM_STA))
continue;
st_ap = &ap->sta[st_p->aid - 1];
slot = (1 << (st_p->aid - 1));
if ((vif->sta_list_index & slot) &&
(!(updated & slot))) {
WARN_ON(st_ap->aid != st_p->aid);
updated |= slot;
ath6kl_add_le32(&st_ap->tx_bytes,
st_p->tx_bytes);
ath6kl_add_le32(&st_ap->tx_pkts,
st_p->tx_pkts);
ath6kl_add_le32(&st_ap->tx_error,
st_p->tx_error);
ath6kl_add_le32(&st_ap->tx_discard,
st_p->tx_discard);
ath6kl_add_le32(&st_ap->rx_bytes,
st_p->rx_bytes);
ath6kl_add_le32(&st_ap->rx_pkts,
st_p->rx_pkts);
ath6kl_add_le32(&st_ap->rx_error,
st_p->rx_error);
ath6kl_add_le32(&st_ap->rx_discard,
st_p->rx_discard);
st_ap->tx_ucast_rate =
st_p->tx_ucast_rate;
st_ap->last_txrx_time =
le16_to_cpu(st_p->last_txrx_time);
}
}
}
} else {
ath6kl_update_target_stats(vif, ptr, len);
}
if (test_bit(STATS_UPDATE_PEND, &vif->flags)) {
clear_bit(STATS_UPDATE_PEND, &vif->flags);
wake_up(&vif->ar->event_wq);
}
}
void ath6kl_wakeup_event(void *dev)
{
struct ath6kl *ar = (struct ath6kl *) dev;
wake_up(&ar->event_wq);
}
void ath6kl_txpwr_rx_evt(void *devt, u8 tx_pwr)
{
struct ath6kl *ar = (struct ath6kl *) devt;
ar->tx_pwr = tx_pwr;
wake_up(&ar->event_wq);
}
void ath6kl_pspoll_event(struct ath6kl_vif *vif, u8 aid)
{
struct ath6kl_sta *conn;
bool psq_empty = false;
bool is_data_psq_empty, is_mgmt_psq_empty;
struct ath6kl *ar = vif->ar;
struct ath6kl_ps_buf_desc *ps_buf;
conn = ath6kl_find_sta_by_aid(vif, aid);
if (!conn)
return;
/*
* Send out a packet queued on ps queue. When the ps queue
* becomes empty update the PVB for this station.
*/
spin_lock_bh(&conn->lock);
is_data_psq_empty = ath6kl_ps_queue_empty(&conn->psq_data);
is_mgmt_psq_empty = ath6kl_ps_queue_empty(&conn->psq_mgmt);
psq_empty = is_data_psq_empty && is_mgmt_psq_empty;
ath6kl_dbg(ATH6KL_DBG_POWERSAVE,
"%s: aid %d sta_flags %x psq_data %d psq_mgmt %d\n",
__func__, conn->aid, conn->sta_flags, !is_data_psq_empty,
!is_mgmt_psq_empty);
if (psq_empty) {
spin_unlock_bh(&conn->lock);
return;
}
if (!is_mgmt_psq_empty) {
struct ieee80211_mgmt *mgmt;
ps_buf = ath6kl_ps_queue_dequeue(&conn->psq_mgmt);
spin_unlock_bh(&conn->lock);
WARN_ON(!ps_buf);
conn->sta_flags |= STA_PS_POLLED;
mgmt = (struct ieee80211_mgmt *) ps_buf->buf;
if (ps_buf->buf + ps_buf->len >= mgmt->u.probe_resp.variable &&
ieee80211_is_probe_resp(mgmt->frame_control))
ath6kl_wmi_send_go_probe_response_cmd(ar->wmi,
vif,
ps_buf->buf,
ps_buf->len,
ps_buf->freq);
else
ath6kl_wmi_send_action_cmd(ar->wmi,
vif->fw_vif_idx,
ps_buf->id,
ps_buf->freq,
ps_buf->wait,
ps_buf->buf,
ps_buf->len);
conn->sta_flags &= ~STA_PS_POLLED;
kfree(ps_buf);
} else {
ps_buf = ath6kl_ps_queue_dequeue(&conn->psq_data);
spin_unlock_bh(&conn->lock);
if (ps_buf) {
WARN_ON(!ps_buf->skb);
if (ps_buf->skb) {
conn->sta_flags |= STA_PS_POLLED;
ath6kl_data_tx(ps_buf->skb, vif->ndev, true);
conn->sta_flags &= ~STA_PS_POLLED;
}
kfree(ps_buf);
}
}
spin_lock_bh(&conn->lock);
psq_empty = ath6kl_ps_queue_empty(&conn->psq_data) &&
ath6kl_ps_queue_empty(&conn->psq_mgmt);
spin_unlock_bh(&conn->lock);
if (psq_empty)
ath6kl_wmi_set_pvb_cmd(ar->wmi, vif->fw_vif_idx, conn->aid, 0);
}
void ath6kl_dtimexpiry_event(struct ath6kl_vif *vif)
{
struct ath6kl *ar = vif->ar;
struct ath6kl_ps_buf_desc *ps_buf;
/*
* If there are no associated STAs, ignore the DTIM expiry event.
* There can be potential race conditions where the last associated
* STA may disconnect & before the host could clear the 'Indicate
* DTIM' request to the firmware, the firmware would have just
* indicated a DTIM expiry event. The race is between 'clear DTIM
* expiry cmd' going from the host to the firmware & the DTIM
* expiry event happening from the firmware to the host.
*/
if (!vif->sta_list_index)
return;
spin_lock_bh(&vif->psq_mcast_lock);
if (ath6kl_ps_queue_empty(&vif->psq_mcast)) {
spin_unlock_bh(&vif->psq_mcast_lock);
return;
}
spin_unlock_bh(&vif->psq_mcast_lock);
/* set the STA flag to dtim_expired for the frame to go out */
set_bit(DTIM_EXPIRED, &vif->flags);
spin_lock_bh(&vif->psq_mcast_lock);
while ((ps_buf = ath6kl_ps_queue_dequeue(&vif->psq_mcast)) != NULL) {
spin_unlock_bh(&vif->psq_mcast_lock);
ath6kl_data_tx(ps_buf->skb, vif->ndev, true);
kfree(ps_buf);
spin_lock_bh(&vif->psq_mcast_lock);
}
spin_unlock_bh(&vif->psq_mcast_lock);
clear_bit(DTIM_EXPIRED, &vif->flags);
/* clear the LSB of the BitMapCtl field of the TIM IE */
ath6kl_wmi_set_pvb_cmd(ar->wmi, vif->fw_vif_idx, MCAST_AID, 0);
}
void ath6kl_disconnect_event(struct ath6kl_vif *vif, u8 reason, u8 *bssid,
u8 assoc_resp_len, u8 *assoc_info,
u16 prot_reason_status)
{
struct ath6kl *ar = vif->ar;
if (vif->nw_type == AP_NETWORK) {
if (!ath6kl_remove_sta(vif, bssid, prot_reason_status))
return;
ath6kl_ap_ht_update_ies(vif);
/* if no more associated STAs, empty the mcast PS q */
if (vif->sta_list_index == 0) {
ath6kl_ps_queue_purge(&vif->psq_mcast);
/* clear the LSB of the TIM IE's BitMapCtl field */
if (test_bit(WMI_READY, &ar->flag))
ath6kl_wmi_set_pvb_cmd(ar->wmi, vif->fw_vif_idx,
MCAST_AID, 0);
}
if (!is_broadcast_ether_addr(bssid)) {
/* send event to application */
cfg80211_del_sta(vif->ndev, bssid, GFP_KERNEL);
}
if (memcmp(vif->ndev->dev_addr, bssid, ETH_ALEN) == 0) {
vif->bss_ch = 0;
vif->phymode = ATH6KL_PHY_MODE_UNKNOWN;
vif->chan_type = ATH6KL_CHAN_TYPE_NONE;
memset(vif->wep_key_list, 0, sizeof(vif->wep_key_list));
clear_bit(CONNECTED, &vif->flags);
netif_carrier_off(vif->ndev);
netif_stop_queue(vif->ndev);
ath6kl_tx_data_cleanup_by_if(vif);
ath6kl_judge_roam_parameter(vif, true);
ath6kl_switch_parameter_based_on_connection(vif, true);
}
return;
} else if (vif->nw_type == INFRA_NETWORK) {
ath6kl_wmi_disctimeout_cmd(ar->wmi, vif->fw_vif_idx,
ATH6KL_DISCONNECT_TIMEOUT);
/* Support to triger supplicant to have another try. */
if (!is_valid_ether_addr(bssid) &&
is_valid_ether_addr(vif->req_bssid))
bssid = vif->req_bssid;
}
ath6kl_cfg80211_disconnect_event(vif, reason, bssid,
assoc_resp_len, assoc_info,
prot_reason_status);
aggr_reset_state(vif->sta_list[0].aggr_conn_cntxt);
del_timer(&vif->disconnect_timer);
ath6kl_dbg(ATH6KL_DBG_WLAN_CFG, "disconnect reason is %d\n", reason);
/*
* If the event is due to disconnect cmd from the host, only they
* the target would stop trying to connect. Under any other
* condition, target would keep trying to connect.
*/
if (reason == DISCONNECT_CMD) {
if (!vif->usr_bss_filter && test_bit(WMI_READY, &ar->flag))
ath6kl_wmi_bssfilter_cmd(ar->wmi, vif->fw_vif_idx,
NONE_BSS_FILTER, 0);
} else {
set_bit(CONNECT_PEND, &vif->flags);
if (((reason == ASSOC_FAILED) &&
(prot_reason_status == 0x11)) ||
((reason == ASSOC_FAILED) && (prot_reason_status == 0x0)
&& (vif->reconnect_flag == 1))) {
set_bit(CONNECTED, &vif->flags);
ath6kl_judge_roam_parameter(vif, false);
ath6kl_switch_parameter_based_on_connection(vif, false);
return;
}
}
/* update connect & link status atomically */
spin_lock_bh(&vif->if_lock);
clear_bit(CONNECTED, &vif->flags);
clear_bit(CONNECT_HANDSHAKE_PROTECT, &vif->flags);
clear_bit(PS_STICK, &vif->flags);
del_timer(&vif->shprotect_timer);
netif_carrier_off(vif->ndev);
spin_unlock_bh(&vif->if_lock);
if ((reason != CSERV_DISCONNECT) || (vif->reconnect_flag != 1))
vif->reconnect_flag = 0;
if (reason != CSERV_DISCONNECT)
ar->user_key_ctrl = 0;
netif_stop_queue(vif->ndev);
memset(vif->bssid, 0, sizeof(vif->bssid));
vif->bss_ch = 0;
vif->phymode = ATH6KL_PHY_MODE_UNKNOWN;
vif->chan_type = ATH6KL_CHAN_TYPE_NONE;
ath6kl_tx_data_cleanup_by_if(vif);
/* Hook disconnection event */
ath6kl_htcoex_disconnect_event(vif);
ath6kl_judge_roam_parameter(vif, true);
ath6kl_switch_parameter_based_on_connection(vif, true);
}
struct ath6kl_vif *ath6kl_vif_first(struct ath6kl *ar)
{
struct ath6kl_vif *vif;
spin_lock_bh(&ar->list_lock);
if (list_empty(&ar->vif_list)) {
spin_unlock_bh(&ar->list_lock);
return NULL;
}
vif = list_first_entry(&ar->vif_list, struct ath6kl_vif, list);
spin_unlock_bh(&ar->list_lock);
return vif;
}
static int ath6kl_open(struct net_device *dev)
{
struct ath6kl_vif *vif = netdev_priv(dev);
struct ath6kl *ar = ath6kl_priv(dev);
set_bit(WLAN_ENABLED, &vif->flags);
if (test_bit(CONNECTED, &vif->flags) ||
test_bit(TESTMODE_EPPING, &ar->flag)) {
netif_carrier_on(dev);
netif_wake_queue(dev);
} else
netif_carrier_off(dev);
return 0;
}
static int ath6kl_close(struct net_device *dev)
{
struct ath6kl *ar = ath6kl_priv(dev);
struct ath6kl_vif *vif = netdev_priv(dev);
netif_stop_queue(dev);
switch (vif->sme_state) {
case SME_CONNECTING:
ath6kl_cfg80211_connect_result(vif, vif->bssid, NULL, 0,
NULL, 0,
WLAN_STATUS_UNSPECIFIED_FAILURE,
GFP_KERNEL);
break;
case SME_CONNECTED:
ath6kl_cfg80211_disconnected(vif, 0, NULL, 0, GFP_KERNEL);
break;
case SME_DISCONNECTED:
default:
break;
}
/* Stop keep-alive. */
ath6kl_ap_keepalive_stop(vif);
/* Stop ACL. */
ath6kl_ap_acl_stop(vif);
/* Stop Admission-Control */
ath6kl_ap_admc_stop(vif);
ath6kl_disconnect(vif);
vif->sme_state = SME_DISCONNECTED;
clear_bit(CONNECTED, &vif->flags);
clear_bit(CONNECT_PEND, &vif->flags);
clear_bit(CONNECT_HANDSHAKE_PROTECT, &vif->flags);
clear_bit(PS_STICK, &vif->flags);
del_timer(&vif->shprotect_timer);
if (test_bit(WMI_READY, &ar->flag)) {
if (ath6kl_wmi_scanparams_cmd(ar->wmi, vif->fw_vif_idx, 0xFFFF,
0, 0, 0, 0, 0, 0, 0, 0, 0))
return -EIO;
}
ath6kl_cfg80211_scan_complete_event(vif, true);
clear_bit(WLAN_ENABLED, &vif->flags);
return 0;
}
static struct net_device_stats *ath6kl_get_stats(struct net_device *dev)
{
struct ath6kl_vif *vif = netdev_priv(dev);
return &vif->net_stats;
}
static int ath6kl_ioctl_p2p_set_ps(struct ath6kl_vif *vif,
char *user_cmd,
int len)
{
int ret = 0;
u8 pwr_mode;
/* SET::P2P_SET_PS {legacy_ps} {opp_ps} {ctwindow} */
if (len > 1) {
if (down_interruptible(&vif->ar->sem)) {
ath6kl_err("busy, couldn't get access\n");
return -EIO;
}
ret = 0;
pwr_mode = (user_cmd[0] != '0' ?
REC_POWER : MAX_PERF_POWER);
if (ath6kl_wmi_powermode_cmd(vif->ar->wmi,
vif->fw_vif_idx,
pwr_mode))
ret = -EIO;
up(&vif->ar->sem);
} else
ret = -EFAULT;
return ret;
}
static int ath6kl_ioctl_setband(struct ath6kl_vif *vif,
char *user_cmd,
int len)
{
int ret = 0, scanband_type = 0;
int i, f = 0;
u8 not_allow_ch;
/* SET::SETBAND {band} */
if (len > 1) {
ret = 0;
sscanf(user_cmd, "%d %d", &scanband_type, &f);
if (scanband_type == ANDROID_SETBAND_ALL)
vif->scanband_type = SCANBAND_TYPE_ALL;
else if (scanband_type == ANDROID_SETBAND_5G)
vif->scanband_type = SCANBAND_TYPE_5G;
else if (scanband_type == ANDROID_SETBAND_2G)
vif->scanband_type = SCANBAND_TYPE_2G;
else if (scanband_type == ANDROID_SETBAND_NO_DFS)
vif->scanband_type = SCANBAND_TYPE_IGNORE_DFS;
else if (scanband_type == ANDROID_SETBAND_NO_CH) {
vif->scanband_type = SCANBAND_TYPE_IGNORE_CH;
if (f == 1) { /* reset */
memset(vif->scanband_ignore_chan,
0,
sizeof(u16) * 64);
} else {
for (i = 0; i < 64; i++) {
if (vif->scanband_ignore_chan[i] == f)
break;
else if (!vif->scanband_ignore_chan[i])
break;
}
if (i < 64)
vif->scanband_ignore_chan[i] = f;
}
} else if ((scanband_type >= 2412) && (scanband_type <= 5825)) {
vif->scanband_type = SCANBAND_TYPE_CHAN_ONLY;
vif->scanband_chan = scanband_type;
} else
ret = -ENOTSUPP;
/* Disconnect if AP is in not allowed band. */
not_allow_ch = 0;
if ((vif->bss_ch) &&
(vif->scanband_type != SCANBAND_TYPE_ALL) &&
(((vif->scanband_type == SCANBAND_TYPE_5G) &&
(vif->bss_ch < 2484)) ||
((vif->scanband_type == SCANBAND_TYPE_2G) &&
(vif->bss_ch > 2484))))
not_allow_ch = 1;
if ((!ret) &&
(vif->nw_type == INFRA_NETWORK) &&
(not_allow_ch) &&
(test_bit(CONNECTED, &vif->flags))) {
ath6kl_dbg(ATH6KL_DBG_INFO,
"Disconnect because of band changed!");
vif->reconnect_flag = 0;
ret = ath6kl_disconnect(vif);
memset(vif->ssid, 0, sizeof(vif->ssid));
vif->ssid_len = 0;
}
} else
ret = -EFAULT;
return ret;
}
static int ath6kl_ioctl_setrts(struct ath6kl_vif *vif,
char *user_cmd,
int len)
{
int ret = 0, rts_threshold = 0;
/* SET::SETRTS {threshold, 0~2347} */
if (len > 1) {
ret = 0;
sscanf(user_cmd, "%d", &rts_threshold);
if (rts_threshold < 0 || rts_threshold > 2347)
return -EINVAL;
ath6kl_wmi_set_rts_cmd(vif->ar->wmi,
vif->fw_vif_idx,
rts_threshold);
} else
ret = -EFAULT;
return ret;
}
static int ath6kl_ioctl_p2p_dev_addr(struct ath6kl_vif *vif,
char *user_cmd,
u8 *buf)
{
int ret = 0;
struct ath6kl_vif *p2p_vif;
/* GET::P2P_DEV_ADDR */
/* In current design, the last vif is always be P2P-device. */
p2p_vif = ath6kl_get_vif_by_index(vif->ar, (vif->ar->vif_max - 1));
if (p2p_vif) {
if (copy_to_user(buf, p2p_vif->ndev->dev_addr, ETH_ALEN))
ret = -EFAULT;
else
ret = 0;
} else
ret = -EFAULT;
return ret;
}
static int ath6kl_ioctl_p2p_best_chan(struct ath6kl_vif *vif,
char *user_cmd,
u8 *buf,
int len)
{
char result[20];
struct ath6kl_rc_report rc_report;
int ret = 0;
/* GET::P2P_BEST_CHANNEL */
if ((strlen(buf) > 16) &&
strstr(buf, "0"))
goto done;
/*
* Current wpa_supplicant only uses best channel for P2P purpose.
* Hence, here just get P2P channels.
*/
memset(&rc_report, 0, sizeof(struct ath6kl_rc_report));
ret = ath6kl_p2p_rc_get(vif->ar, &rc_report);
done:
if (ret == 0) {
memset(result, 0, 20);
snprintf(result, 20,
"%d %d %d",
rc_report.rc_p2p_2g,
rc_report.rc_p2p_5g,
rc_report.rc_p2p_all);
result[strlen(result)] = '\0';
if (copy_to_user(buf, result, strlen(result)))
ret = -EFAULT;
else
ret = 0;
}
return ret;
}
static int ath6kl_ioctl_ap_acl(struct ath6kl_vif *vif,
char *user_cmd,
u8 *buf, /* reserved for GET op */
int len)
{
int ret = 0;
#ifdef NL80211_CMD_SET_AP_MAC_ACL
/* Configurate from NL80211. */
if (vif->ar->wiphy->max_acl_mac_addrs)
return -EINVAL;
#endif
/* SET::ACL {MACCMD|ADDMAC|DELMAC} {{[0|1|2]}|{MAC ADDRESS}} */
if (len > 1) {
int i, policy, addr[ETH_ALEN];
u8 mac_addr[ETH_ALEN];
if (strstr(user_cmd, "MACCMD ")) {
user_cmd += 7;
sscanf(user_cmd, "%d", &policy);
ret = ath6kl_ap_acl_config_policy(vif, policy);
} else if (strstr(user_cmd, "ADDMAC ")) {
user_cmd += 7;
if (sscanf(user_cmd, "%02x:%02x:%02x:%02x:%02x:%02x",
&addr[0], &addr[1], &addr[2],
&addr[3], &addr[4], &addr[5]) != ETH_ALEN)
return -EFAULT;
for (i = 0; i < ETH_ALEN; i++)
mac_addr[i] = (u8)addr[i];
ret = ath6kl_ap_acl_config_mac_list(vif,
mac_addr,
false);
} else if (strstr(user_cmd, "DELMAC ")) {
user_cmd += 7;
if (sscanf(user_cmd, "%02x:%02x:%02x:%02x:%02x:%02x",
&addr[0], &addr[1], &addr[2],
&addr[3], &addr[4], &addr[5]) != ETH_ALEN)
return -EFAULT;
for (i = 0; i < ETH_ALEN; i++)
mac_addr[i] = (u8)addr[i];
ret = ath6kl_ap_acl_config_mac_list(vif,
mac_addr,
true);
} else
ret = -EFAULT;
} else
ret = -EFAULT;
return ret;
}
bool ath6kl_ioctl_ready(struct ath6kl_vif *vif)
{
struct ath6kl *ar = vif->ar;
if (!test_bit(WMI_READY, &ar->flag)) {
ath6kl_err("wmi is not ready\n");
return false;
}
#if defined(USB_AUTO_SUSPEND)
if ((vif->ar->state == ATH6KL_STATE_WOW) ||
(vif->ar->state == ATH6KL_STATE_DEEPSLEEP) ||
(vif->ar->state == ATH6KL_STATE_PRE_SUSPEND)) {
ath6kl_dbg(ATH6KL_DBG_WLAN_CFG,
"ignore wlan disabled in AUTO suspend mode!\n");
} else {
if (!test_bit(WLAN_ENABLED, &vif->flags))
return false;
}
#else
if (!test_bit(WLAN_ENABLED, &vif->flags)) {
ath6kl_err("wlan disabled\n");
return false;
}
#endif
return true;
}
static int ath6kl_ioctl_standard(struct net_device *dev,
struct ifreq *rq, int cmd)
{
struct ath6kl_vif *vif = netdev_priv(dev);
void *data = (void *)(rq->ifr_data);
int ret = 0;
switch (cmd) {
case ATH6KL_IOCTL_STANDARD01:
{
struct ath6kl_android_wifi_priv_cmd android_cmd;
char *user_cmd;
if (copy_from_user(&android_cmd,
data,
sizeof(struct ath6kl_android_wifi_priv_cmd)))
ret = -EIO;
else {
if (android_cmd.used_len > android_cmd.total_len) {
ret = -EINVAL;
break;
}
user_cmd = kzalloc(android_cmd.total_len, GFP_KERNEL);
if (!user_cmd) {
ret = -ENOMEM;
break;
}
if (copy_from_user(user_cmd,
android_cmd.buf,
android_cmd.used_len))
ret = -EIO;
else {
if (strstr(user_cmd, "P2P_SET_PS "))
ret = ath6kl_ioctl_p2p_set_ps(vif,
(user_cmd + 11),
(android_cmd.used_len - 11));
else if (strstr(user_cmd, "SETBAND "))
ret = ath6kl_ioctl_setband(vif,
(user_cmd + 8),
(android_cmd.used_len - 8));
else if (strstr(user_cmd, "SETRTS "))
ret = ath6kl_ioctl_setrts(vif,
(user_cmd + 7),
(android_cmd.used_len - 7));
else if (strstr(user_cmd, "P2P_DEV_ADDR"))
ret = ath6kl_ioctl_p2p_dev_addr(vif,
user_cmd,
android_cmd.buf);
else if (strstr(user_cmd, "P2P_BEST_CHANNEL"))
ret = ath6kl_ioctl_p2p_best_chan(vif,
user_cmd,
android_cmd.buf,
android_cmd.used_len);
else if (strstr(user_cmd, "ACL "))
ret = ath6kl_ioctl_ap_acl(vif,
(user_cmd + 4),
NULL,
(android_cmd.used_len - 4));
else if (strstr(user_cmd, "SET_AP_WPS_P2P_IE"))
ret = 0; /* To avoid AP/GO up stuck. */
#ifdef CONFIG_ANDROID
else if (strstr(user_cmd, "SET_BT_ON ")) {
ath6kl_bt_on =
(user_cmd[10] == '1') ? 1 : 0;
ret = 0;
}
#endif
else {
ath6kl_dbg(ATH6KL_DBG_TRC,
"not yet support \"%s\"\n",
user_cmd);
ret = -EOPNOTSUPP;
}
}
kfree(user_cmd);
}
break;
}
case ATH6KL_IOCTL_STANDARD02:
{
struct ath6kl_ioctl_cmd ioctl_cmd;
if (copy_from_user(&ioctl_cmd,
data,
sizeof(struct ath6kl_ioctl_cmd)))
ret = -EIO;
else {
switch (ioctl_cmd.subcmd) {
case ATH6KL_IOCTL_AP_APSD:
ath6kl_wmi_ap_set_apsd(vif->ar->wmi,
vif->fw_vif_idx,
ioctl_cmd.options);
break;
case ATH6KL_IOCTL_AP_INTRABSS:
vif->intra_bss = ioctl_cmd.options;
break;
default:
ret = -EOPNOTSUPP;
break;
}
}
break;
}
case ATH6KL_IOCTL_STANDARD03:
{
struct btcoex_ioctl btcoex_cmd;
char *user_cmd;
if (copy_from_user(&btcoex_cmd,
data,
sizeof(struct btcoex_ioctl)))
ret = -EIO;
else {
user_cmd = kzalloc(btcoex_cmd.cmd_len, GFP_KERNEL);
if (!user_cmd) {
ret = -ENOMEM;
break;
}
if (copy_from_user(user_cmd,
btcoex_cmd.cmd,
btcoex_cmd.cmd_len))
ret = -EIO;
else {
if (!ath6kl_ioctl_ready(vif)) {
kfree(user_cmd);
return -EIO;
}
if (down_interruptible(&vif->ar->sem)) {
ath6kl_err("busy, couldn't get access\n");
kfree(user_cmd);
return -ERESTARTSYS;
}
ret = ath6kl_wmi_send_btcoex_cmd(vif->ar,
(u8 *)user_cmd, btcoex_cmd.cmd_len);
up(&vif->ar->sem);
}
kfree(user_cmd);
}
break;
}
default:
#if defined(ATHTST_SUPPORT) || defined(ACL_SUPPORT)
return ath6kl_ce_ioctl(dev, rq, cmd);
#else
ret = -EOPNOTSUPP;
break;
#endif
}
return ret;
}
static int ath6kl_ioctl_linkspeed(struct net_device *dev,
struct ifreq *rq,
int cmd)
{
struct ath6kl *ar = ath6kl_priv(dev);
struct ath6kl_vif *vif = netdev_priv(dev);
struct iwreq *req = (struct iwreq *)(rq);
char user_cmd[32];
u8 macaddr[6];
long left;
s32 rate = 0;
/* Only AR6004 now */
if (ar->target_type != TARGET_TYPE_AR6004)
return -EOPNOTSUPP;
if ((!req->u.data.pointer) || (!req->u.data.length))
return -EFAULT;
memset(user_cmd, 0, 32);
if (copy_from_user(user_cmd,
req->u.data.pointer,
req->u.data.length))
return -EFAULT;
if (_string_to_mac(user_cmd, req->u.data.length, macaddr))
return -EFAULT;
if (down_interruptible(&ar->sem))
return -EBUSY;
#ifdef CONFIG_ANDROID
/*
* WAR : Framework always use p2p0 to query linkspeed and here transfer
* to correct P2P-GO/P2P-Client interface.
*/
if ((ar->p2p) &&
(!ar->p2p_compat) &&
(ar->p2p_concurrent) &&
(ar->p2p_dedicate)) {
vif = ath6kl_get_vif_by_index(ar, ar->vif_max - 2);
if (!vif) {
up(&ar->sem);
return -EFAULT;
}
}
#endif
set_bit(STATS_UPDATE_PEND, &vif->flags);
if (ath6kl_wmi_get_stats_cmd(ar->wmi, vif->fw_vif_idx) != 0) {
up(&ar->sem);
return -EIO;
}
left = wait_event_interruptible_timeout(ar->event_wq,
!test_bit(STATS_UPDATE_PEND,
&vif->flags),
WMI_TIMEOUT);
up(&ar->sem);
if (left == 0)
return -ETIMEDOUT;
else if (left < 0)
return left;
memset(user_cmd, 0, 32);
if (vif->nw_type == AP_NETWORK) {
struct wmi_ap_mode_stat *ap = &vif->ap_stats;
struct ath6kl_sta *conn;
conn = ath6kl_find_sta(vif, macaddr);
if (conn) {
for (left = 0; left < AP_MAX_NUM_STA; left++) {
if (conn->aid == ap->sta[left].aid) {
rate = ath6kl_wmi_get_rate_ar6004(
ap->sta[left].tx_ucast_rate);
break;
}
}
WARN_ON(left == AP_MAX_NUM_STA);
}
} else
rate = vif->target_stats.tx_ucast_rate;
snprintf(user_cmd, 32, "%u", rate / 1000);
req->u.data.length = strlen(user_cmd);
user_cmd[req->u.data.length] = '\0';
return copy_to_user(req->u.data.pointer,
user_cmd,
req->u.data.length + 1) ? -EFAULT : 0;
}
static int ath6kl_ioctl_get_if_freq(struct net_device *dev,
struct ifreq *rq,
int cmd)
{
struct ath6kl_vif *vif = netdev_priv(dev);
struct iwreq *req = (struct iwreq *)(rq);
u16 freq;
if (vif->bss_ch == 0) {
return -EINVAL;
} else {
freq = vif->bss_ch;
return copy_to_user(req->u.data.pointer, &freq, sizeof(freq));
}
}
int ath6kl_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
{
struct ath6kl *ar = ath6kl_priv(dev);
int ret = true;
s8 *userdata;
/*
* ioctl operations may have to wait for the Target, so we cannot
* hold rtnl. Prevent the device from disappearing under us and
* release the lock during the ioctl operation.
*/
dev_hold(dev);
rtnl_unlock();
switch (cmd) {
case ATH6KL_IOCTL_STANDARD01: /* Android privacy command */
case ATH6KL_IOCTL_STANDARD02: /* supplicant escape purpose to
support WiFi-Direct Cert. */
case ATH6KL_IOCTL_STANDARD03: /* BTC command */
case ATH6KL_IOCTL_STANDARD12: /* hole, please reserved */
#ifdef ATHTST_SUPPORT
case ATHCFG_WCMD_IOCTL: /* athtst */
#else
case ATH6KL_IOCTL_STANDARD15: /* hole, please reserved */
#endif
#ifdef CE_SUPPORT
case IEEE80211_IOCTL_KICKMAC:
#endif
#ifdef ACL_SUPPORT
case IEEE80211_IOCTL_SETPARAM:
case IEEE80211_IOCTL_GETPARAM:
case IEEE80211_IOCTL_SETMLME:
case IEEE80211_IOCTL_ADDMAC:
case IEEE80211_IOCTL_DELMAC:
case IEEE80211_IOCTL_GET_MACADDR:
#endif
#ifdef TX99_SUPPORT
case SIOCIOCTLTX99:/* TX99 */
#else
case ATH6KL_IOCTL_STANDARD13: /* TX99 */
#endif
ret = ath6kl_ioctl_standard(dev, rq, cmd);
break;
#ifndef CE_SUPPORT
case ATH6KL_IOCTL_WEXT_PRIV6:
ret = ath6kl_ioctl_get_if_freq(dev, rq, cmd);
break;
#endif
case ATH6KL_IOCTL_WEXT_PRIV26: /* endpoint loopback purpose */
get_user(cmd, (int *)rq->ifr_data);
userdata = (char *)(((unsigned int *)rq->ifr_data)+1);
switch (cmd) {
case ATH6KL_XIOCTL_TRAFFIC_ACTIVITY_CHANGE:
if (ar->htc_target != NULL) {
struct ath6kl_traffic_activity_change data;
if (copy_from_user(&data,
userdata,
sizeof(data))) {
ret = -EFAULT;
goto ioctl_done;
}
ath6kl_htc_indicate_activity_change(
ar->htc_target,
(u8)data.stream_id,
data.active ?
true : false);
}
break;
}
break;
case ATH6KL_IOCTL_WEXT_PRIV27: /* QCSAP (old) */
case ATH6KL_IOCTL_WEXT_PRIV31: /* QCSAP */
ret = ath6kl_ioctl_linkspeed(dev, rq, cmd);
break;
default:
ret = -EOPNOTSUPP;
goto ioctl_done;
}
ioctl_done:
rtnl_lock(); /* restore rtnl state */
dev_put(dev);
return ret;
}
static struct net_device_ops ath6kl_netdev_ops = {
.ndo_open = ath6kl_open,
.ndo_stop = ath6kl_close,
.ndo_start_xmit = ath6kl_start_tx,
.ndo_get_stats = ath6kl_get_stats,
.ndo_do_ioctl = ath6kl_ioctl,
};
void init_netdev(struct net_device *dev)
{
struct ath6kl_vif *vif = netdev_priv(dev);
vif->needed_headroom = ETH_HLEN +
sizeof(struct ath6kl_llc_snap_hdr) +
sizeof(struct wmi_data_hdr) +
HTC_HDR_LENGTH +
WMI_MAX_TX_META_SZ +
ATH6KL_HTC_ALIGN_BYTES;
#ifdef CE_OLD_KERNEL_SUPPORT_2_6_23
dev->open = ath6kl_open;
dev->stop = ath6kl_close;
dev->hard_start_xmit = ath6kl_start_tx;
dev->get_stats = ath6kl_get_stats;
dev->do_ioctl = ath6kl_ioctl;
#else
dev->netdev_ops = &ath6kl_netdev_ops;
dev->needed_headroom = vif->needed_headroom;
#endif
dev->destructor = free_netdev;
dev->watchdog_timeo = ATH6KL_TX_TIMEOUT;
return;
}
#if defined(CONFIG_CRASH_DUMP) || defined(ATH6KL_HSIC_RECOVER)
int _readwrite_file(const char *filename, char *rbuf,
const char *wbuf, size_t length, int mode)
{
int ret = 0;
struct file *filp = (struct file *)-ENOENT;
mm_segment_t oldfs;
oldfs = get_fs();
set_fs(KERNEL_DS);
do {
filp = filp_open(filename, mode, S_IRUSR);
if (IS_ERR(filp) || !filp->f_op) {
ret = -ENOENT;
break;
}
if (!filp->f_op->write || !filp->f_op->read) {
filp_close(filp, NULL);
ret = -ENOENT;
break;
}
if (length == 0) {
/* Read the length of the file only */
struct inode *inode;
inode = GET_INODE_FROM_FILEP(filp);
if (!inode) {
printk(KERN_ERR
"_readwrite_file: Error 2\n");
ret = -ENOENT;
break;
}
ret = i_size_read(inode->i_mapping->host);
break;
}
if (wbuf) {
ret = filp->f_op->write(
filp, wbuf, length, &filp->f_pos);
if (ret < 0) {
printk(KERN_ERR
"_readwrite_file: Error 3\n");
break;
}
} else {
ret = filp->f_op->read(
filp, rbuf, length, &filp->f_pos);
if (ret < 0) {
printk(KERN_ERR
"_readwrite_file: Error 4\n");
break;
}
}
} while (0);
if (!IS_ERR(filp))
filp_close(filp, NULL);
set_fs(oldfs);
return ret;
}
#endif
#ifdef CONFIG_CRASH_DUMP
int check_dump_file_size(void)
{
int status = 0, size = 0;
size = _readwrite_file(CRASH_DUMP_FILE, NULL, NULL, 0, O_RDONLY);
if (size > (MAX_DUMP_FW_SIZE - DUMP_BUF_SIZE)) {
ath6kl_info("clean big log 0x%x\n", size);
status = _readwrite_file(CRASH_DUMP_FILE, NULL, NULL,
0, (O_WRONLY | O_TRUNC));
}
return status;
}
int print_to_file(const char *fmt, ...)
{
char *buf;
unsigned int len = 0, buf_len = MAX_STRDUMP_LEN;
int status = 0;
struct va_format vaf;
va_list args;
status = check_dump_file_size();
if (status)
ath6kl_info("log file check status code 0x%x\n", status);
buf = kmalloc(buf_len, GFP_ATOMIC);
if (buf == NULL)
return -ENOMEM;
memset(buf, 0, buf_len);
len = snprintf(buf + len, buf_len - len, "Drv ver %s\n",
DRV_VERSION);
va_start(args, fmt);
vaf.fmt = fmt;
vaf.va = &args;
len += scnprintf(buf + len, buf_len - len, "%pV", &vaf);
va_end(args);
status = _readwrite_file(CRASH_DUMP_FILE, NULL,
buf, len, (O_WRONLY | O_APPEND | O_CREAT));
if (status < 0)
ath6kl_info("write failed with status code 0x%x\n", status);
kfree(buf);
return status;
}
#else
int print_to_file(const char *fmt, ...)
{
return 0;
}
#endif