diff options
Diffstat (limited to 'drivers/net/wireless/eagle/esp_sip.c')
-rwxr-xr-x | drivers/net/wireless/eagle/esp_sip.c | 2377 |
1 files changed, 2377 insertions, 0 deletions
diff --git a/drivers/net/wireless/eagle/esp_sip.c b/drivers/net/wireless/eagle/esp_sip.c new file mode 100755 index 00000000..ef11ff57 --- /dev/null +++ b/drivers/net/wireless/eagle/esp_sip.c @@ -0,0 +1,2377 @@ +/* + * Copyright (c) 2009 - 2013 Espressif System. + */ + +#include <linux/ieee80211.h> +#include <net/mac80211.h> +#include <net/cfg80211.h> +#include <linux/skbuff.h> +#include <linux/bitops.h> +#include <linux/version.h> +#include <linux/mmc/card.h> +#include <linux/mmc/mmc.h> +#include <linux/mmc/host.h> +#include <linux/mmc/sdio_func.h> +#include <linux/mmc/sdio_ids.h> +#include <linux/mmc/sdio.h> +#include <linux/mmc/sd.h> +#include <linux/completion.h> + +#include "esp_pub.h" +#include "esp_sip.h" +#include "esp_ctrl.h" +#include "esp_sif.h" +#include "esp_debug.h" +#include "slc_host_register.h" +#include "esp_wmac.h" +#include "esp_utils.h" +#ifdef TEST_MODE +#include "testmode.h" +#endif + +#ifdef USE_EXT_GPIO +#include "esp_ext.h" +#endif /* USE_EXT_GPIO */ + +extern struct completion *gl_bootup_cplx; +static u32 bcn_counter = 0; +static u32 probe_rsp_counter = 0; + +static int old_signal = -35; +static int avg_signal = 0; +static int signal_loop = 0; + +#define SIGNAL_COUNT 300 + +#define TID_TO_AC(_tid) ((_tid)== 0||((_tid)==3)?WME_AC_BE:((_tid)<3)?WME_AC_BK:((_tid)<6)?WME_AC_VI:WME_AC_VO) + +#ifdef SIP_DEBUG +#define esp_sip_dbg esp_dbg +struct sip_trace { + u32 tx_data; + u32 tx_cmd; + u32 rx_data; + u32 rx_evt; + u32 rx_tx_status; + u32 tx_out_of_credit; + u32 tx_one_shot_overflow; +}; +static struct sip_trace str; +#define STRACE_TX_DATA_INC() (str.tx_data++) +#define STRACE_TX_CMD_INC() (str.tx_cmd++) +#define STRACE_RX_DATA_INC() (str.rx_data++) +#define STRACE_RX_EVENT_INC() (str.rx_evt++) +#define STRACE_RX_TXSTATUS_INC() (str.rx_tx_status++) +#define STRACE_TX_OUT_OF_CREDIT_INC() (str.tx_out_of_credit++) +#define STRACE_TX_ONE_SHOT_INC() (str.tx_one_shot_overflow++) + +#if 0 +static void sip_show_trace(struct esp_sip *sip); +#endif //0000 + +#define STRACE_SHOW(sip) sip_show_trace(sip) +#else +#define esp_sip_dbg(...) +#define STRACE_TX_DATA_INC() +#define STRACE_TX_CMD_INC() +#define STRACE_RX_DATA_INC() +#define STRACE_RX_EVENT_INC() +#define STRACE_RX_TXSTATUS_INC() +#define STRACE_TX_OUT_OF_CREDIT_INC() +#define STRACE_TX_ONE_SHOT_INC() +#define STRACE_SHOW(sip) +#endif /* SIP_DEBUG */ + +#define SIP_STOP_QUEUE_THRESHOLD 48 +#define SIP_RESUME_QUEUE_THRESHOLD 12 +#ifndef FAST_TX_STATUS +#define SIP_PENDING_STOP_TX_THRESHOLD 6 +#define SIP_PENDING_RESUME_TX_THRESHOLD 6 +#endif /* !FAST_TX_STATUS */ + +#define SIP_MIN_DATA_PKT_LEN (sizeof(struct esp_mac_rx_ctrl) + 24) //24 is min 80211hdr +#define TARGET_RX_SIZE 524 + +static struct sip_pkt *sip_get_ctrl_buf(struct esp_sip *sip, SIP_BUF_TYPE bftype); + +static void sip_reclaim_ctrl_buf(struct esp_sip *sip, struct sip_pkt *pkt, SIP_BUF_TYPE bftype); + +static void sip_free_init_ctrl_buf(struct esp_sip *sip); + +static void sip_dec_credit(struct esp_sip *sip); + + +static int sip_pack_pkt(struct esp_sip *sip, struct sk_buff *skb, int *pm_state); + +static struct esp_mac_rx_ctrl *sip_parse_normal_mac_ctrl(struct sk_buff *skb, int * pkt_len_enc, int *buf_len, int *pulled_len); + +static struct sk_buff * sip_parse_data_rx_info(struct esp_sip *sip, struct sk_buff *skb, int pkt_len_enc, int buf_len, struct esp_mac_rx_ctrl *mac_ctrl, int *pulled_len); + +#ifndef RX_SYNC +static inline void sip_rx_pkt_enqueue(struct esp_sip *sip, struct sk_buff *skb); + +static inline struct sk_buff * sip_rx_pkt_dequeue(struct esp_sip *sip); +#endif /* RX_SYNC */ +#ifndef FAST_TX_STATUS +static void sip_after_tx_status_update(struct esp_sip *sip); +#endif /* !FAST_TX_STATUS */ + +static void sip_after_write_pkts(struct esp_sip *sip); + +//static void show_data_seq(u8 *pkt); + +static void sip_update_tx_credits(struct esp_sip *sip, u16 recycled_credits); + +//static void sip_trigger_txq_process(struct esp_sip *sip); + +#ifndef RX_SYNC +static bool sip_rx_pkt_process(struct esp_sip * sip, struct sk_buff *skb); +#else +static void sip_rx_pkt_process_sync(struct esp_sip *sip, struct sk_buff *skb); +#endif /* RX_SYNC */ + +#ifdef FAST_TX_STATUS +static void sip_tx_status_report(struct esp_sip *sip, struct sk_buff *skb, struct ieee80211_tx_info* tx_info, bool success); +#endif /* FAST_TX_STATUS */ + +#if (LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 35)) +static void sip_check_skb_alignment(struct sk_buff *skb); +#endif /* NEW_KERNEL */ + +#ifdef ESP_RX_COPYBACK_TEST +/* only for rx test */ +static u8 *copyback_buf; +static u32 copyback_offset = 0; +#endif /* ESP_RX_COPYBACK_TEST */ + +#ifdef FPGA_TXDATA +int sip_send_tx_data(struct esp_sip *sip); +#endif/* FPGA_TXDATA */ + +#ifdef FPGA_LOOPBACK +int sip_send_loopback_cmd_mblk(struct esp_sip *sip); +#endif /* FPGA_LOOPBACK */ + +#ifdef SIP_DEBUG +#if 0 +static void sip_show_trace(struct esp_sip *sip) +{ + esp_sip_dbg(ESP_DBG_TRACE, "\n \t tx_data %u \t tx_cmd %u \t rx_data %u \t rx_evt %u \t rx_tx_status %u \n\n", \ + str.tx_data, str.tx_cmd, str.rx_data, str.rx_evt, str.rx_tx_status); +} +#endif //0000 +#endif //SIP_DEBUG + +#if 0 +static void show_data_seq(u8 *pkt) +{ + struct ieee80211_hdr * wh = (struct ieee80211_hdr *)pkt; + u16 seq = 0; + + if (ieee80211_is_data(wh->frame_control)) { + seq = (le16_to_cpu(wh->seq_ctrl) >> 4); + esp_sip_dbg(ESP_DBG_TRACE, " ieee80211 seq %u addr1 %pM\n", seq, wh->addr1); + } else if (ieee80211_is_beacon(wh->frame_control) || ieee80211_is_probe_resp(wh->frame_control)) + esp_sip_dbg(ESP_DBG_TRACE, " ieee80211 probe resp or beacon 0x%04x\n", wh->frame_control); + else + esp_sip_dbg(ESP_DBG_TRACE, " ieee80211 other mgmt pkt 0x%04x\n", wh->frame_control); +} +#endif //0000 + +//#if (LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 35)) +static bool check_ac_tid(u8 *pkt, u8 ac, u8 tid) +{ + struct ieee80211_hdr * wh = (struct ieee80211_hdr *)pkt; +#ifdef TID_DEBUG + u16 real_tid = 0; +#endif //TID_DEBUG + + if (ieee80211_is_data_qos(wh->frame_control)) { +#ifdef TID_DEBUG + real_tid = *ieee80211_get_qos_ctl(wh) & IEEE80211_QOS_CTL_TID_MASK; + + esp_sip_dbg(ESP_SHOW, "ac:%u, tid:%u, tid in pkt:%u\n", ac, tid, real_tid); + if (tid != real_tid) { + esp_sip_dbg(ESP_DBG_ERROR, "111 ac:%u, tid:%u, tid in pkt:%u\n", ac, tid, real_tid); + } + if (TID_TO_AC(tid) != ac) { + esp_sip_dbg(ESP_DBG_ERROR, "222 ac:%u, tid:%u, tid in pkt:%u\n", ac, tid, real_tid); + } + + //ASSERT(tid == real_tid); + //ASSERT(TID_TO_AC(tid) == ac); +#endif /* TID_DEBUG*/ + } else if (ieee80211_is_mgmt(wh->frame_control)) { +#ifdef TID_DEBUG + esp_sip_dbg(ESP_SHOW, "ac:%u, tid:%u\n", ac, tid); + if (tid != 7 || ac != WME_AC_VO) { + esp_sip_dbg(ESP_DBG_ERROR, "333 ac:%u, tid:%u\n", ac, tid); + } + //ASSERT(tid == 7); + //ASSERT(ac == WME_AC_VO); +#endif /* TID_DEBUG*/ + } else { + if (ieee80211_is_ctl(wh->frame_control)) { +#ifdef TID_DEBUG + esp_sip_dbg(ESP_SHOW, "%s is ctrl pkt fc 0x%04x ac:%u, tid:%u, tid in pkt:%u\n", __func__, wh->frame_control, ac, tid, real_tid); +#endif /* TID_DEBUG*/ + } else { + if (tid != 0 || ac != WME_AC_BE) { + //show_buf(pkt, 24); + esp_sip_dbg(ESP_DBG_LOG, "444 ac:%u, tid:%u \n", ac, tid); + if (tid == 7 && ac == WME_AC_VO) + return false; + } + return true; //hack to modify non-qos null data. + + } + //ASSERT(tid == 0); + //ASSERT(ac = WME_AC_BE); + } + + return false; +} +//#endif /* NEW_KERNEL || KERNEL_35 */ + +static void sip_update_tx_credits(struct esp_sip *sip, u16 recycled_credits) +{ + esp_sip_dbg(ESP_DBG_TRACE, "%s:before add, credits is %d\n", __func__, atomic_read(&sip->tx_credits)); + atomic_add(recycled_credits, &sip->tx_credits); + esp_sip_dbg(ESP_DBG_TRACE, "%s:after add %d, credits is %d\n", __func__, recycled_credits, atomic_read(&sip->tx_credits)); +} + +void sip_trigger_txq_process(struct esp_sip *sip) +{ + if (atomic_read(&sip->tx_credits) <= sip->credit_to_reserve) //no credits, do nothing + return; + + if (sip_queue_may_resume(sip)) { + /* wakeup upper queue only if we have sufficient credits */ + esp_sip_dbg(ESP_DBG_TRACE, "%s wakeup ieee80211 txq \n", __func__); + atomic_set(&sip->epub->txq_stopped, false); + ieee80211_wake_queues(sip->epub->hw); + } else if (atomic_read(&sip->epub->txq_stopped) ) { + esp_sip_dbg(ESP_DBG_TRACE, "%s can't wake txq, credits: %d \n", __func__, atomic_read(&sip->tx_credits) ); + } + + if (!skb_queue_empty(&sip->epub->txq)) { + /* try to send out pkt already in sip queue once we have credits */ + esp_sip_dbg(ESP_DBG_TRACE, "%s resume sip txq \n", __func__); +#if !defined(FPGA_LOOPBACK) && !defined(FPGA_TXDATA) && (LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 32)) + ieee80211_queue_work(sip->epub->hw, &sip->epub->tx_work); +#else + queue_work(sip->epub->esp_wkq, &sip->epub->tx_work); +#endif + } +} + +static bool sip_ampdu_occupy_buf(struct esp_sip *sip, struct esp_rx_ampdu_len * ampdu_len) +{ + return (ampdu_len->substate == 0 || esp_wmac_rxsec_error(ampdu_len->substate) || (sip->dump_rpbm_err && ampdu_len->substate == RX_RPBM_ERR)); +} + +#ifdef RX_SYNC +static void sip_rx_pkt_process_sync(struct esp_sip *sip, struct sk_buff *skb) +#else +static bool sip_rx_pkt_process(struct esp_sip * sip, struct sk_buff *skb) +#endif /* RX_SYNC */ +{ +#define DO_NOT_COPY false +#define DO_COPY true + + struct sip_hdr * hdr = NULL; + struct sk_buff * rskb = NULL; + int remains_len = 0; + int first_pkt_len = 0; + u8 *bufptr = NULL; + int ret = 0; + bool trigger_rxq = false; +#ifdef RX_SYNC + bool trigger_txq = false; +#endif/* RX_SYNC */ + + if (skb == NULL) { + esp_sip_dbg(ESP_DBG_ERROR, "%s NULL SKB!!!!!!!! \n", __func__); +#ifdef RX_SYNC + return; +#else + return trigger_rxq; +#endif /* RX_SYNC */ + } + + hdr = (struct sip_hdr *)skb->data; + bufptr = skb->data; + + + esp_sip_dbg(ESP_DBG_TRACE, "%s Hcredits 0x%08x, realCredits %d\n", __func__, hdr->h_credits, hdr->h_credits & SIP_CREDITS_MASK); + if (hdr->h_credits & SIP_CREDITS_MASK) { + sip_update_tx_credits(sip, hdr->h_credits & SIP_CREDITS_MASK); +#ifdef RX_SYNC + trigger_txq = true; +#endif/* RX_SYNC */ + } + + hdr->h_credits &= ~SIP_CREDITS_MASK; /* clean credits in sip_hdr, prevent over-add */ + + esp_sip_dbg(ESP_DBG_TRACE, "%s credits %d\n", __func__, hdr->h_credits); + + /* + * first pkt's length is stored in recycled_credits first 20 bits + * config w3 [31:12] + * repair hdr->len of first pkt + */ + remains_len = hdr->len; + first_pkt_len = hdr->h_credits >> 12; + hdr->len = first_pkt_len; + + esp_dbg(ESP_DBG_TRACE, "%s first_pkt_len %d, whole pkt len %d \n", __func__, first_pkt_len, remains_len); + if (first_pkt_len > remains_len) { + printk("first_pkt_len %d, whole pkt len %d\n", first_pkt_len, remains_len); + show_buf((u8 *)hdr, first_pkt_len); + ASSERT(0); + } + + /* + * pkts handling, including the first pkt, should alloc new skb for each data pkt. + * free the original whole skb after parsing is done. + */ + while (remains_len) { + ASSERT(remains_len >= sizeof(struct sip_hdr)); + hdr = (struct sip_hdr *)bufptr; + ASSERT(hdr->len > 0); + if((hdr->len & 3) != 0) { + show_buf((u8 *)hdr, 512); + } + ASSERT((hdr->len & 3) == 0); + if (unlikely(hdr->seq != sip->rxseq++)) { + esp_dbg(ESP_DBG_ERROR, "%s seq mismatch! got %u, expect %u\n", __func__, hdr->seq, sip->rxseq-1); + show_buf(bufptr, 32); + ASSERT(0); + } + + if (SIP_HDR_IS_CTRL(hdr)) { + STRACE_RX_EVENT_INC(); + esp_sip_dbg(ESP_DBG_TRACE, "seq %u \n", hdr->seq); + + ret = sip_parse_events(sip, bufptr); + + skb_pull(skb, hdr->len); + + } else if (SIP_HDR_IS_DATA(hdr)) { + struct esp_mac_rx_ctrl * mac_ctrl = NULL; + int pkt_len_enc = 0, buf_len = 0, pulled_len = 0; + + STRACE_RX_DATA_INC(); + esp_sip_dbg(ESP_DBG_TRACE, "seq %u \n", hdr->seq); + mac_ctrl = sip_parse_normal_mac_ctrl(skb, &pkt_len_enc, &buf_len, &pulled_len); + rskb = sip_parse_data_rx_info(sip, skb, pkt_len_enc, buf_len, mac_ctrl, &pulled_len); + + if(rskb == NULL) + goto _move_on; + +#if (LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 35)) + sip_check_skb_alignment(rskb); +#endif /* !NEW_KERNEL */ + if (likely(atomic_read(&sip->epub->wl.off) == 0)) { +#ifndef RX_SENDUP_SYNC + skb_queue_tail(&sip->epub->rxq, rskb); + trigger_rxq = true; +#else +#ifdef RX_CHECKSUM_TEST + esp_rx_checksum_test(rskb); +#endif + local_bh_disable(); +#if (LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 32)) + ieee80211_rx(sip->epub->hw, rskb); +#else + //simulate IEEE80211_SKB_RXCB in 2.6.32 + ieee80211_rx(sip->epub->hw, rskb ,(struct ieee80211_rx_status *)rskb->cb); +#endif + local_bh_enable(); +#endif /* RX_SENDUP_SYNC */ + } else { + /* still need go thro parsing as skb_pull should invoke */ + kfree_skb(rskb); + } + } else if (SIP_HDR_IS_AMPDU(hdr)) { + struct esp_mac_rx_ctrl * mac_ctrl = NULL; + struct esp_mac_rx_ctrl new_mac_ctrl; + struct esp_rx_ampdu_len *ampdu_len; + int pkt_num; + int pulled_len = 0; + static int pkt_dropped = 0; + static int pkt_total = 0; + bool have_rxabort = false; + bool have_goodpkt = false; + static u8 frame_head[16]; + static u8 frame_buf_ttl = 0; + + ampdu_len = (struct esp_rx_ampdu_len *)(skb->data + hdr->len/sip->rx_blksz * sip->rx_blksz); + esp_sip_dbg(ESP_DBG_TRACE, "%s rx ampdu total len %u\n", __func__, hdr->len); + if(skb->data != (u8 *)hdr) { + printk("%p %p\n", skb->data, hdr); + show_buf(skb->data, 512); + show_buf((u8 *)hdr, 512); + } + ASSERT(skb->data == (u8 *)hdr); + mac_ctrl = sip_parse_normal_mac_ctrl(skb, NULL, NULL, &pulled_len); + memcpy(&new_mac_ctrl, mac_ctrl, sizeof(struct esp_mac_rx_ctrl)); + mac_ctrl = &new_mac_ctrl; + pkt_num = mac_ctrl->ampdu_cnt; + esp_sip_dbg(ESP_DBG_TRACE, "%s %d rx ampdu %u pkts, %d pkts dumped, first len %u\n", + __func__, __LINE__, (unsigned int)((hdr->len % sip->rx_blksz) / sizeof(struct esp_rx_ampdu_len)), pkt_num, (unsigned int)ampdu_len->sublen); + + pkt_total += mac_ctrl->ampdu_cnt; + //esp_sip_dbg(ESP_DBG_ERROR, "%s ampdu dropped %d/%d\n", __func__, pkt_dropped, pkt_total); + while (pkt_num > 0) { + esp_sip_dbg(ESP_DBG_TRACE, "%s %d ampdu sub state %02x,\n", __func__, __LINE__, ampdu_len->substate); + + if (sip_ampdu_occupy_buf(sip, ampdu_len)) { //pkt is dumped + + rskb = sip_parse_data_rx_info(sip, skb, ampdu_len->sublen - FCS_LEN, 0, mac_ctrl, &pulled_len); + assert(rskb != NULL); + + if (likely(atomic_read(&sip->epub->wl.off) == 0) && + (ampdu_len->substate == 0 || ampdu_len->substate == RX_TKIPMIC_ERR || + (sip->sendup_rpbm_pkt && ampdu_len->substate == RX_RPBM_ERR)) && + (sip->rxabort_fixed || !have_rxabort) ) + { + if(!have_goodpkt) { + have_goodpkt = true; + memcpy(frame_head, rskb->data, 16); + frame_head[1] &= ~0x80; + frame_buf_ttl = 3; + } +#ifndef RX_SENDUP_SYNC + skb_queue_tail(&sip->epub->rxq, rskb); + trigger_rxq = true; +#else +#ifdef RX_CHECKSUM_TEST + esp_rx_checksum_test(rskb); +#endif + local_bh_disable(); +#if (LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 32)) + ieee80211_rx(sip->epub->hw, rskb); +#else + //simulate IEEE80211_SKB_RXCB in 2.6.32 + ieee80211_rx(sip->epub->hw, rskb ,(struct ieee80211_rx_status *)rskb->cb); +#endif + local_bh_enable(); +#endif /* RX_SENDUP_SYNC */ + + } else { + kfree_skb(rskb); + } + } else { + if (ampdu_len->substate == RX_ABORT) { + u8 * a; + have_rxabort = true; + esp_sip_dbg(ESP_DBG_TRACE, "rx abort %d %d\n", frame_buf_ttl, pkt_num); + if(frame_buf_ttl && !sip->rxabort_fixed) { + struct esp_rx_ampdu_len * next_good_ampdu_len = ampdu_len + 1; + a = frame_head; + esp_sip_dbg(ESP_DBG_TRACE, "frame:%02x %02x %02x %02x %02x %02x %02x %02x %02x %02x %02x %02x %02x %02x %02x %02x\n", + a[0], a[1], a[2], a[3], a[4], a[5], a[6], a[7], a[8], a[9], a[10], a[11], a[12], a[13], a[14], a[15]); + while(!sip_ampdu_occupy_buf(sip, next_good_ampdu_len)) { + if(next_good_ampdu_len > ampdu_len + pkt_num - 1) + break; + next_good_ampdu_len++; + + } + if(next_good_ampdu_len <= ampdu_len + pkt_num -1) { + bool b0, b10, b11; + a = skb->data; + esp_sip_dbg(ESP_DBG_TRACE, "buf:%02x %02x %02x %02x %02x %02x %02x %02x %02x %02x %02x %02x %02x %02x %02x %02x\n", + a[0], a[1], a[2], a[3], a[4], a[5], a[6], a[7], a[8], a[9], a[10], a[11], a[12], a[13], a[14], a[15]); + b0 = memcmp(frame_head + 4, skb->data + 4, 12) == 0; + b10 = memcmp(frame_head + 10, skb->data, 6) == 0; + b11 = memcpy(frame_head + 11, skb->data, 5) == 0; + esp_sip_dbg(ESP_DBG_TRACE, "com %d %d %d\n", b0, b10, b11); + if(b0 && !b10 && !b11) { + have_rxabort = false; + esp_sip_dbg(ESP_DBG_TRACE, "repair 0\n"); + } else if(!b0 && b10 && !b11) { + skb_push(skb, 10); + memcpy(skb->data, frame_head, 10); + have_rxabort = false; + pulled_len -= 10; + esp_sip_dbg(ESP_DBG_TRACE, "repair 10\n"); + } else if(!b0 && !b10 && b11) { + skb_push(skb, 11); + memcpy(skb->data, frame_head, 11); + have_rxabort = false; + pulled_len -= 11; + esp_sip_dbg(ESP_DBG_TRACE, "repair 11\n"); + } + } + } + } + pkt_dropped++; + esp_sip_dbg(ESP_DBG_LOG, "%s ampdu dropped %d/%d\n", __func__, pkt_dropped, pkt_total); + } + pkt_num--; + ampdu_len++; + } + if(frame_buf_ttl) + frame_buf_ttl--; + skb_pull(skb, hdr->len - pulled_len); + } else { + esp_sip_dbg(ESP_DBG_ERROR, "%s %d unknown type\n", __func__, __LINE__); + } + +_move_on: + if (hdr->len < remains_len) { + remains_len -= hdr->len; + } else { + break; + } + bufptr += hdr->len; + } + + do { + ASSERT(skb != rskb); + kfree_skb(skb); + } while (0); + +#ifdef RX_SYNC + if (trigger_rxq) { + queue_work(sip->epub->esp_wkq, &sip->epub->sendup_work); + } + if (trigger_txq) { + sip_trigger_txq_process(sip); + } +#else + return trigger_rxq; +#endif /* RX_SYNC */ + +#undef DO_NOT_COPY +#undef DO_COPY +} + +#ifndef RX_SYNC +static void _sip_rxq_process(struct esp_sip *sip) +{ + struct sk_buff *skb = NULL; + bool sendup = false; + + while ((skb = skb_dequeue(&sip->rxq))) { + if (sip_rx_pkt_process(sip, skb)) + sendup = true; + } +#ifndef RX_SENDUP_SYNC + if (sendup) { + queue_work(sip->epub->esp_wkq, &sip->epub->sendup_work); + } +#endif /* !RX_SENDUP_SYNC */ + + /* probably tx_credit is updated, try txq */ + sip_trigger_txq_process(sip); +} + +void sip_rxq_process(struct work_struct *work) +{ + struct esp_sip *sip = container_of(work, struct esp_sip, rx_process_work); + ASSERT(sip != NULL); + + if (unlikely(atomic_read(&sip->state) == SIP_SEND_INIT)) { + sip_send_chip_init(sip); + atomic_set(&sip->state, SIP_WAIT_BOOTUP); + return; + } + + spin_lock(&sip->rx_lock); + _sip_rxq_process(sip); + spin_unlock(&sip->rx_lock); +} + +static inline void sip_rx_pkt_enqueue(struct esp_sip *sip, struct sk_buff *skb) +{ + skb_queue_tail(&sip->rxq, skb); +} + +static inline struct sk_buff * sip_rx_pkt_dequeue(struct esp_sip *sip) { + return skb_dequeue(&sip->rxq); +} +#endif /* RX_SYNC */ + +static u32 sip_rx_count = 0; +void sip_debug_show(struct esp_sip *sip) +{ + esp_sip_dbg(ESP_DBG_ERROR, "txq left %d %d\n", skb_queue_len(&sip->epub->txq), atomic_read(&sip->tx_data_pkt_queued)); + esp_sip_dbg(ESP_DBG_ERROR, "tx queues stop ? %d\n", atomic_read(&sip->epub->txq_stopped)); + esp_sip_dbg(ESP_DBG_ERROR, "txq stop? %d\n", test_bit(ESP_WL_FLAG_STOP_TXQ, &sip->epub->wl.flags)); + esp_sip_dbg(ESP_DBG_ERROR, "tx credit %d\n", atomic_read(&sip->tx_credits)); + esp_sip_dbg(ESP_DBG_ERROR, "rx collect %d\n", sip_rx_count); + sip_rx_count = 0; +} + +#ifndef LOOKAHEAD +int sip_rx(struct esp_pub *epub) +{ + struct sip_hdr *shdr = NULL; + struct esp_sip *sip = epub->sip; + int err = 0; + struct sk_buff *first_skb = NULL; + struct sk_buff *next_skb = NULL; + u8 *rx_buf = NULL; + u32 rx_blksz; + u32 first_sz = 4; + struct sk_buff *rx_skb = NULL; + + esp_sip_dbg(ESP_DBG_LOG, "%s enter\n", __func__); + + + /* first read one block out, if we luck enough, that's it + * + * To make design as simple as possible, we allocate skb(s) + * separately for each sif read operation to avoid global + * read_buf_pointe access. It coule be optimized late. + */ + rx_blksz = sif_get_blksz(epub); + first_skb = __dev_alloc_skb(first_sz, GFP_KERNEL); + if (first_skb == NULL) { + esp_sip_dbg(ESP_DBG_ERROR, "%s first no memory \n", __func__); + goto _err; + } + + rx_buf = skb_put(first_skb, first_sz); + esp_sip_dbg(ESP_DBG_LOG, "%s rx_buf ptr %p, first_sz %d\n", __func__, rx_buf, first_sz); + + sif_lock_bus(epub); + +#ifdef USE_EXT_GPIO + do{ + int err2; + u16 value = 0; + u16 intr_mask = ext_gpio_get_int_mask_reg(); + if(!intr_mask) + break; + err2 = sif_get_gpio_intr(epub, intr_mask, &value); + if(!err2 && value) { + esp_sip_dbg(ESP_DBG_TRACE, "%s intr_mask[0x%04x] value[0x%04x]\n", __func__, intr_mask, value); + ext_gpio_int_process(value); + } + }while(0); +#endif +#ifdef ESP_ACK_INTERRUPT +#ifdef ESP_ACK_LATER + err = esp_common_read(epub, rx_buf, first_sz, ESP_SIF_NOSYNC, true); + sif_platform_ack_interrupt(epub); +#else + sif_platform_ack_interrupt(epub); + err = esp_common_read(epub, rx_buf, first_sz, ESP_SIF_NOSYNC, true); +#endif /* ESP_ACK_LATER */ +#else + err = esp_common_read(epub, rx_buf, first_sz, ESP_SIF_NOSYNC, true); +#endif //ESP_ACK_INTERRUPT + + sip_rx_count++; + if (unlikely(err)) { + esp_dbg(ESP_DBG_ERROR, " %s first read err %d %d\n", __func__, err, sif_get_regs(epub)->config_w0); + kfree_skb(first_skb); + sif_unlock_bus(epub); + goto _err; + } + + shdr = (struct sip_hdr *)rx_buf; + if(SIP_HDR_IS_CTRL(shdr) && (shdr->c_evtid == SIP_EVT_SLEEP)) { + atomic_set(&sip->epub->ps.state, ESP_PM_ON); + esp_dbg(ESP_DBG_TRACE, "s\n"); + } + + ASSERT((shdr->len & 3) == 0); + //ASSERT(shdr->len >= SIP_HDR_LEN && shdr->len <= SIP_PKT_MAX_LEN); + if (shdr->len > first_sz) { + /* larger than one blk, fetch the rest */ + next_skb = __dev_alloc_skb(roundup(shdr->len, rx_blksz) + first_sz, GFP_KERNEL); + if (unlikely(next_skb == NULL)) { + sif_unlock_bus(epub); + esp_sip_dbg(ESP_DBG_ERROR, "%s next no memory \n", __func__); + kfree_skb(first_skb); + goto _err; + } + rx_buf = skb_put(next_skb, shdr->len); + rx_buf += first_sz; /* skip the first block */ + + err = esp_common_read(epub, rx_buf, (shdr->len - first_sz), ESP_SIF_NOSYNC, false); + sif_unlock_bus(epub); + + if (unlikely(err)) { + esp_sip_dbg(ESP_DBG_ERROR, "%s next read err %d \n", __func__, err); + kfree(first_skb); + kfree(next_skb); + goto _err; + } + /* merge two skbs, TBD: could be optimized by skb_linearize*/ + memcpy(next_skb->data, first_skb->data, first_sz); + esp_dbg(ESP_DBG_TRACE, " %s next skb\n", __func__); + + rx_skb = next_skb; + kfree_skb(first_skb); + } else { + sif_unlock_bus(epub); + skb_trim(first_skb, shdr->len); + esp_dbg(ESP_DBG_TRACE, " %s first_skb only\n", __func__); + + rx_skb = first_skb; + } + if (atomic_read(&sip->state) == SIP_STOP) { + kfree_skb(rx_skb); + esp_sip_dbg(ESP_DBG_ERROR, "%s when sip stopped\n", __func__); + return 0; + } +#ifndef RX_SYNC + sip_rx_pkt_enqueue(sip, rx_skb); + queue_work(sip->epub->esp_wkq, &sip->rx_process_work); +#else + sip_rx_pkt_process_sync(sip, rx_skb); +#endif /* RX_SYNC */ + +_err: + + return err; +} + +#else + +int sip_rx(struct esp_pub *epub) +{ + struct sip_hdr *shdr = NULL; + struct esp_sip *sip = epub->sip; + int err = 0; + struct sk_buff *rx_skb = NULL; + u8 *rx_buf = NULL; + u32 lookahead = 0; + struct slc_host_regs *regs = sif_get_regs(epub); + bool free_skb = false; + + esp_sip_dbg(ESP_DBG_LOG, "%s enter\n", __func__); + + lookahead = regs->config_w0; + + esp_sip_dbg(ESP_DBG_TRACE, "%s lookahead %u \n", __func__, lookahead); + rx_skb = __dev_alloc_skb(lookahead, GFP_KERNEL); + if (rx_skb == NULL) { + esp_sip_dbg(ESP_DBG_ERROR, "%s no memory \n", __func__); + goto __out; + } + rx_buf = skb_put(rx_skb, lookahead); + err = sif_lldesc_read_raw(epub, rx_buf, lookahead); + if (unlikely(err)) { + esp_dbg(ESP_DBG_ERROR, " %s read err %d \n", __func__, err); + goto __out; + } + + shdr = (struct sip_hdr *)rx_buf; + esp_sip_dbg(ESP_DBG_TRACE, "%s len %d totlen %d\n", __func__, shdr->len, (shdr->h_credits >> 12)); + show_buf(rx_buf, 32); + //ASSERT(shdr->len >= SIP_HDR_LEN && shdr->len <= SIP_PKT_MAX_LEN); + + esp_sip_dbg(ESP_DBG_TRACE, "%s credits %d\n", __func__, shdr->h_credits); + if (shdr->h_credits & SIP_CREDITS_MASK) { + sip_update_tx_credits(sip, shdr->h_credits & SIP_CREDITS_MASK); + } +#if 0 + if (SIP_HDR_IS_CREDIT_RPT(shdr)) { + esp_sip_dbg(ESP_DBG_TRACE, "%s credit update %d\n", __func__, shdr->h_credits & SIP_CREDITS_MASK); + if (unlikely(shdr->seq != sip->rxseq++)) { + esp_sip_dbg(ESP_DBG_ERROR, "%s seq mismatch! got %u, expect %u\n", __func__, shdr->seq, sip->rxseq-1); + } + + free_skb = true; + goto __out; + } +#endif + shdr->h_credits &= ~SIP_CREDITS_MASK; /* clean credits in sip_hdr, prevent over-add */ + + /* we need some place to store the total length of the pkt chain, the simple way is make it consistent with + * non-lookahead case + */ + shdr->h_credits = (shdr->len << 12); + shdr->len = lookahead; + + sip_rx_pkt_enqueue(sip, rx_skb); + queue_work(sip->epub->esp_wkq, &sip->rx_process_work); + +__out: + if (free_skb) + kfree_skb(rx_skb); + + sip_trigger_txq_process(sip); + + return err; +} +#endif /* LOOKAHEAD */ + +int sip_get_raw_credits(struct esp_sip *sip) +{ +#if 1 + unsigned long timeout; + int err = 0; + + esp_dbg(ESP_DBG_TRACE, "%s entern \n", __func__); + + /* 1s timeout */ + timeout = jiffies + msecs_to_jiffies(1000); + + while (time_before(jiffies, timeout) && !sip->boot_credits) { + + err = esp_common_read_with_addr(sip->epub, SLC_HOST_TOKEN_RDATA, (u8 *)&sip->boot_credits, 4, ESP_SIF_SYNC); + + if (err) { + esp_dbg(ESP_DBG_ERROR, "Can't read credits\n"); + return err; + } + sip->boot_credits &= SLC_HOST_TOKEN0_MASK; +#ifdef SIP_DEBUG + if (sip->boot_credits == 0) { + esp_dbg(ESP_DBG_ERROR, "no credit, try again\n"); + mdelay(50); + } +#endif /* SIP_DEBUG */ + } + + if (!sip->boot_credits) { + esp_dbg(ESP_DBG_ERROR, "read credits timeout\n"); + return -ETIMEDOUT; + } + + esp_dbg(ESP_DBG_TRACE, "%s got credits: %d\n", __func__, sip->boot_credits); +#endif //0000 + + return 0; +} + + +/* Only cooperate with get_raw_credits */ +static void +sip_dec_credit(struct esp_sip *sip) +{ +#if 0 + u32 reg = 0; + int err = 0; + + reg = SLC_HOST_TOKEN0_WR | SLC_HOST_TOKEN0_DEC; + memcpy(sip->rawbuf, ®, sizeof(u32)); + err = sif_io_sync(sip->epub, SLC_HOST_INT_CLR, sip->rawbuf, sizeof(u32), SIF_TO_DEVICE | SIF_SYNC | SIF_BYTE_BASIS | SIF_INC_ADDR); + + if (err) + esp_dbg(ESP_DBG_ERROR, "%s can't clear target token0 \n", __func__); + +#endif //0000 + /* SLC 2.0, token reg is read-to-clean, thus no need to access target */ + sip->boot_credits--; +} + +int sip_post_init(struct esp_sip *sip, struct sip_evt_bootup2 *bevt) +{ + struct esp_pub *epub; + int po = 0; + + ASSERT(sip != NULL); + + epub = sip->epub; + + po = get_order(SIP_TX_AGGR_BUF_SIZE); + sip->tx_aggr_buf = (u8 *)__get_free_pages(GFP_ATOMIC, po); + if (sip->tx_aggr_buf == NULL) { + esp_dbg(ESP_DBG_ERROR, "no mem for tx_aggr_buf! \n"); + return -ENOMEM; + } +#if 0 + po = get_order(SIP_RX_AGGR_BUF_SIZE); + sip->rx_aggr_buf = (u8 *)__get_free_pages(GFP_ATOMIC, po); + + if (sip->rx_aggr_buf == NULL) { + po = get_order(SIP_TX_AGGR_BUF_SIZE); + free_pages((unsigned long)sip->tx_aggr_buf, po); + esp_dbg(ESP_DBG_ERROR, "no mem for rx_aggr_buf! \n"); + return -ENOMEM; + } +#endif + + sip->tx_aggr_write_ptr = sip->tx_aggr_buf; + + sip->tx_blksz = bevt->tx_blksz; + sip->rx_blksz = bevt->rx_blksz; + sip->credit_to_reserve = bevt->credit_to_reserve; + + sip->dump_rpbm_err = (bevt->options & SIP_DUMP_RPBM_ERR); + sip->rxabort_fixed = (bevt->options & SIP_RXABORT_FIXED); + sip->support_bgscan = (bevt->options & SIP_SUPPORT_BGSCAN); + + sip->sendup_rpbm_pkt = sip->dump_rpbm_err && false; + + /* print out MAC addr... */ + memcpy(epub->mac_addr, bevt->mac_addr, ETH_ALEN); + atomic_set(&sip->noise_floor, bevt->noise_floor); + + esp_sip_dbg(ESP_DBG_TRACE, "%s tx_blksz %d rx_blksz %d mac addr %pM\n", __func__, sip->tx_blksz, sip->rx_blksz, epub->mac_addr); + + return 0; +} + +/* write pkts in aggr buf to target memory */ +static void sip_write_pkts(struct esp_sip *sip, int pm_state) +{ + int tx_aggr_len = 0; + struct sip_hdr *first_shdr = NULL; + int err = 0; + + tx_aggr_len = sip->tx_aggr_write_ptr - sip->tx_aggr_buf; + if (tx_aggr_len < sizeof(struct sip_hdr)) { + printk("%s tx_aggr_len %d \n", __func__, tx_aggr_len); + ASSERT(0); + } + //ASSERT(tx_aggr_len >= sizeof(struct sip_hdr)); + ASSERT((tx_aggr_len & 0x3) == 0); + + first_shdr = (struct sip_hdr *)sip->tx_aggr_buf; + + if (atomic_read(&sip->tx_credits) <= SIP_CREDITS_LOW_THRESHOLD) { + first_shdr->fc[1] |= SIP_HDR_F_NEED_CRDT_RPT; + } + + ASSERT(tx_aggr_len == sip->tx_tot_len); + + /* still use lock bus instead of sif_lldesc_write_sync since we want to protect several global varibles assignments */ + sif_lock_bus(sip->epub); + + /* add a dummy read in power saving mode, although target may not really in + * sleep + */ + //if (atomic_read(&sip->epub->ps.state) == ESP_PM_ON) { + // atomic_set(&sip->epub->ps.state, ESP_PM_OFF); + sif_raw_dummy_read(sip->epub); + //} + //esp_dbg(ESP_DBG_ERROR, "%s write len %u\n", __func__, tx_aggr_len); + + err = esp_common_write(sip->epub, sip->tx_aggr_buf, tx_aggr_len, ESP_SIF_NOSYNC); + + sip->tx_aggr_write_ptr = sip->tx_aggr_buf; + sip->tx_tot_len = 0; + +#if 0 + if (pm_state == ESP_PM_TURNING_ON) { + ASSERT(atomic_read(&sip->epub->ps.state) == ESP_PM_TURNING_ON); + atomic_set(&sip->epub->ps.state, ESP_PM_ON); + } else if (pm_state == ESP_PM_TURNING_OFF) { + ASSERT(atomic_read(&sip->epub->ps.state) == ESP_PM_TURNING_OFF); + atomic_set(&sip->epub->ps.state, ESP_PM_OFF); + esp_dbg(ESP_DBG_PS, "----beacon %d, probe rsp %d -------\n", bcn_counter, probe_rsp_counter); + bcn_counter = 0; probe_rsp_counter = 0; + } else if (pm_state && atomic_read(&sip->epub->ps.state)) { + esp_dbg(ESP_DBG_ERROR, "%s wrong pm_state %d ps.state %d \n", __func__, pm_state, atomic_read(&sip->epub->ps.state)); + } +#endif + sif_unlock_bus(sip->epub); + + if (err) + esp_sip_dbg(ESP_DBG_ERROR, "func %s err!!!!!!!!!: %d\n", __func__, err); + +} + +/* setup sip header and tx info, copy pkt into aggr buf */ +static int sip_pack_pkt(struct esp_sip *sip, struct sk_buff *skb, int *pm_state) +{ + struct ieee80211_tx_info *itx_info; + struct sip_hdr *shdr; + u32 tx_len = 0, offset = 0; + bool is_data = true; + + itx_info = IEEE80211_SKB_CB(skb); + + if (itx_info->flags == 0xffffffff) { + shdr = (struct sip_hdr *)skb->data; + is_data = false; + tx_len = skb->len; + } else { + struct ieee80211_hdr * wh = (struct ieee80211_hdr *)skb->data; + struct esp_vif *evif = (struct esp_vif *)itx_info->control.vif->drv_priv; + u8 sta_index; + struct esp_node *node; + /* update sip header */ + shdr = (struct sip_hdr *)sip->tx_aggr_write_ptr; + + shdr->fc[0] = 0; + shdr->fc[1] = 0; + + if ((itx_info->flags & IEEE80211_TX_CTL_AMPDU) && (true || esp_is_ip_pkt(skb))) + SIP_HDR_SET_TYPE(shdr->fc[0], SIP_DATA_AMPDU); + else + SIP_HDR_SET_TYPE(shdr->fc[0], SIP_DATA); + + if(evif->epub == NULL){ +#ifndef FAST_TX_STATUS + /* TBD */ +#else + sip_tx_status_report(sip, skb, itx_info, false); + atomic_dec(&sip->tx_data_pkt_queued); + return -EINVAL; +#endif /* FAST_TX_STATUS */ + } + + /* make room for encrypted pkt */ + if (itx_info->control.hw_key) { +#if (LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 39)) + shdr->d_enc_flag= itx_info->control.hw_key->alg+1; +#else + int alg = esp_cipher2alg(itx_info->control.hw_key->cipher); + if (unlikely(alg == -1)) { +#ifndef FAST_TX_STATUS + /* TBD */ +#else + sip_tx_status_report(sip, skb, itx_info, false); + atomic_dec(&sip->tx_data_pkt_queued); + return -1; +#endif /* FAST_TX_STATUS */ + } else { + shdr->d_enc_flag = alg + 1; + } + +#endif /* NEW_KERNEL */ + shdr->d_hw_kid = itx_info->control.hw_key->hw_key_idx | (evif->index<<7); + } else { + shdr->d_enc_flag=0; + shdr->d_hw_kid = (evif->index << 7 | evif->index); + } + + /* update sip tx info */ +#if (LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 28)) + if(itx_info->control.sta == NULL){ + node = NULL; + } else { + node = esp_get_node_by_addr(sip->epub, itx_info->control.sta->addr); + } +#else + + node = esp_get_node_by_addr(sip->epub, wh->addr1); +#endif + if(node != NULL) + sta_index = node->index; + else + sta_index = ESP_PUB_MAX_STA + 1; + SIP_HDR_SET_IFIDX(shdr->fc[0], evif->index << 3 | sta_index); +#if (LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 37)) + shdr->d_p2p = itx_info->control.vif->p2p; + if(evif->index == 1) + shdr->d_p2p = 1; +#endif + shdr->d_ac = skb_get_queue_mapping(skb); + shdr->d_tid = skb->priority & IEEE80211_QOS_CTL_TAG1D_MASK; + wh = (struct ieee80211_hdr *)skb->data; + if (ieee80211_is_mgmt(wh->frame_control)) { + /* addba/delba/bar may use different tid/ac */ + if (ieee80211_is_beacon(wh->frame_control)||shdr->d_ac == WME_AC_VO) { + shdr->d_tid = 7; + } + } +//#if (LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 35)) + if (check_ac_tid(skb->data, shdr->d_ac, shdr->d_tid)) { + shdr->d_ac = WME_AC_BE; + shdr->d_tid = 0; + } +//#endif /* NEW_KERNEL || KERNEL_35 */ + + + /* make sure data is start at 4 bytes aligned addr. */ + offset = roundup(sizeof(struct sip_hdr), 4); + +#ifdef HOST_RC + esp_sip_dbg(ESP_DBG_TRACE, "%s offset0 %d \n", __func__, offset); + memcpy(sip->tx_aggr_write_ptr + offset, (void *)&itx_info->control, + sizeof(struct sip_tx_rc)); + + offset += roundup(sizeof(struct sip_tx_rc), 4); + esp_show_tx_rates(&itx_info->control.rates[0]); + +#endif /* HOST_RC */ + + if (SIP_HDR_IS_AMPDU(shdr)) { + memset(sip->tx_aggr_write_ptr + offset, 0, sizeof(struct esp_tx_ampdu_entry)); + offset += roundup(sizeof(struct esp_tx_ampdu_entry), 4); + } + + tx_len = offset + skb->len; + shdr->len = tx_len; /* actual len */ + + esp_sip_dbg(ESP_DBG_TRACE, "%s offset %d skblen %d txlen %d\n", __func__, offset, skb->len, tx_len); + + } + + shdr->seq = sip->txseq++; + //esp_sip_dbg(ESP_DBG_ERROR, "%s seq %u, %u %u\n", __func__, shdr->seq, SIP_HDR_GET_TYPE(shdr->fc[0]),shdr->c_cmdid); + + /* copy skb to aggr buf */ + memcpy(sip->tx_aggr_write_ptr + offset, skb->data, skb->len); + + if (is_data) { + spin_lock_bh(&sip->epub->tx_lock); + sip->txdataseq = shdr->seq; + spin_unlock_bh(&sip->epub->tx_lock); +#ifndef FAST_TX_STATUS + /* store seq in driver data, need seq to pick pkt during tx status report */ + *(u32 *)itx_info->driver_data = shdr->seq; + atomic_inc(&sip->pending_tx_status); +#else + /* fake a tx_status and report to mac80211 stack to speed up tx, may affect + * 1) rate control (now it's all in target, so should be OK) + * 2) ps mode, mac80211 want to check ACK of ps/nulldata to see if AP is awake + * 3) BAR, mac80211 do BAR by checking ACK + */ + /* + * XXX: need to adjust for 11n, e.g. report tx_status according to BA received in target + * + */ + sip_tx_status_report(sip, skb, itx_info, true); + atomic_dec(&sip->tx_data_pkt_queued); + +#endif /* FAST_TX_STATUS */ + STRACE_TX_DATA_INC(); + } else { + /* check pm state here */ +#if 0 + if (*pm_state == 0) { + if (SIP_HDR_IS_PM_TURNING_ON(shdr)) + *pm_state = ESP_PM_TURNING_ON; + else if (SIP_HDR_IS_PM_TURNING_OFF(shdr)) + *pm_state = ESP_PM_TURNING_OFF; + } else { + if (SIP_HDR_IS_PM_TURNING_ON(shdr) || SIP_HDR_IS_PM_TURNING_OFF(shdr)) { + esp_dbg(ESP_DBG_ERROR, "%s two pm states in one bunch - prev: %d; curr %d\n", __func__, + *pm_state, SIP_HDR_IS_PM_TURNING_ON(shdr)?ESP_PM_TURNING_ON:ESP_PM_TURNING_OFF); + if (*pm_state == ESP_PM_TURNING_ON && SIP_HDR_IS_PM_TURNING_OFF(shdr)) { + *pm_state = 0; + atomic_set(&sip->epub->ps.state, ESP_PM_OFF); + } else if (*pm_state == ESP_PM_TURNING_OFF && SIP_HDR_IS_PM_TURNING_ON(shdr)) { + *pm_state = 0; + atomic_set(&sip->epub->ps.state, ESP_PM_ON); + } else { + ASSERT(0); + } + } + } +#endif + /* no need to hold ctrl skb */ + sip_free_ctrl_skbuff(sip, skb); + STRACE_TX_CMD_INC(); + } + + /* TBD: roundup here or whole aggr-buf */ + tx_len = roundup(tx_len, sip->tx_blksz); + + sip->tx_aggr_write_ptr += tx_len; + sip->tx_tot_len += tx_len; + + return 0; +} + +#ifndef FAST_TX_STATUS +static void +sip_after_tx_status_update(struct esp_sip *sip) +{ + if (atomic_read(&sip->data_tx_stopped) == true && sip_tx_data_may_resume(sip)) { + atomic_set(&sip->data_tx_stopped, false); + if (sip_is_tx_mblk_avail(sip) == false) { + esp_sip_dbg(ESP_DBG_ERROR, "%s mblk still unavail \n", __func__); + } else { + esp_sip_dbg(ESP_DBG_TRACE, "%s trigger txq \n", __func__); + sip_trigger_txq_process(sip); + } + } else if (!sip_tx_data_may_resume(sip)) { //JLU: this is redundant + STRACE_SHOW(sip); + } +} +#endif /* !FAST_TX_STATUS */ + +#ifdef HOST_RC +static void sip_set_tx_rate_status(struct sip_rc_status *rcstatus, struct ieee80211_tx_rate *irates) +{ + int i; + u8 shift = 0; + u32 cnt = 0; + + for (i = 0; i < IEEE80211_TX_MAX_RATES; i++) { + if (rcstatus->rc_map & BIT(i)) { + shift = i << 2; + cnt = (rcstatus->rc_cnt_store >> shift) & RC_CNT_MASK; + irates[i].idx = i; + irates[i].count = (u8)cnt; + } else { + irates[i].idx = -1; + irates[i].count = 0; + } + } + + esp_show_rcstatus(rcstatus); + esp_show_tx_rates(irates); +} +#endif /* HOST_RC */ + +#ifndef FAST_TX_STATUS +static void +sip_txdoneq_process(struct esp_sip *sip, struct sip_evt_tx_report *tx_report) +{ + struct sk_buff *skb, *tmp; + struct esp_pub *epub = sip->epub; + int matchs = 0; + struct ieee80211_tx_info *tx_info; + struct sip_tx_status *tx_status; + int i; + + esp_sip_dbg(ESP_DBG_LOG, "%s enter, report->pkts %d, pending tx_status %d\n", __func__, tx_report->pkts, atomic_read(&sip->pending_tx_status)); + + /* traversal the txdone queue, find out matched skb by seq, hand over + * to up layer stack + */ + for (i = 0; i < tx_report->pkts; i++) { + //esp_sip_dbg(ESP_DBG_TRACE, "%s status %d seq %u\n", __func__, i, tx_report->status[i].sip_seq); + skb_queue_walk_safe(&epub->txdoneq, skb, tmp) { + tx_info = IEEE80211_SKB_CB(skb); + + //esp_sip_dbg(ESP_DBG_TRACE, "%s skb seq %u\n", __func__, *(u32 *)tx_info->driver_data); + if (tx_report->status[i].sip_seq == *(u32 *)tx_info->driver_data) { + tx_status = &tx_report->status[i]; + __skb_unlink(skb, &epub->txdoneq); + + //fill up ieee80211_tx_info + //TBD: lock ?? + if (tx_status->errno == SIP_TX_ST_OK && + !(tx_info->flags & IEEE80211_TX_CTL_NO_ACK)) { + tx_info->flags |= IEEE80211_TX_STAT_ACK; + } +#ifdef HOST_RC + sip_set_tx_rate_status(&tx_report->status[i].rcstatus, &tx_info->status.rates[0]); + esp_sip_dbg(ESP_DBG_TRACE, "%s idx0 %d, cnt0 %d, flags0 0x%02x\n", __func__, tx_info->status.rates[0].idx,tx_info->status.rates[0].count, tx_info->status.rates[0].flags); + +#else + /* manipulate rate status... */ + tx_info->status.rates[0].idx = 0; + tx_info->status.rates[0].count = 1; + tx_info->status.rates[0].flags = 0; + tx_info->status.rates[1].idx = -1; +#endif /* HOST_RC */ + + ieee80211_tx_status(epub->hw, skb); + matchs++; + atomic_dec(&sip->pending_tx_status); + STRACE_RX_TXSTATUS_INC(); + } + } + } + + if (matchs < tx_report->pkts) { + esp_sip_dbg(ESP_DBG_ERROR, "%s tx report mismatch! \n", __func__); + } else { + //esp_sip_dbg(ESP_DBG_TRACE, "%s tx report %d pkts! \n", __func__, matchs); + } + + sip_after_tx_status_update(sip); +} +#else +#ifndef FAST_TX_NOWAIT + +static void +sip_txdoneq_process(struct esp_sip *sip) +{ + struct esp_pub *epub = sip->epub; + struct sk_buff *skb; + while ((skb = skb_dequeue(&epub->txdoneq))) { + ieee80211_tx_status(epub->hw, skb); + } +} +#endif +#endif /* !FAST_TX_STATUS */ + +#ifdef FAST_TX_STATUS +static void sip_tx_status_report(struct esp_sip *sip, struct sk_buff *skb, struct ieee80211_tx_info *tx_info, bool success) +{ + if(!(tx_info->flags & IEEE80211_TX_CTL_AMPDU)) { + if (likely(success)) + tx_info->flags |= IEEE80211_TX_STAT_ACK; + else + tx_info->flags &= ~IEEE80211_TX_STAT_ACK; + +#if (LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 29)) + /* manipulate rate status... */ + tx_info->status.rates[0].idx = 11; + tx_info->status.rates[0].count = 1; + tx_info->status.rates[0].flags = 0; + tx_info->status.rates[1].idx = -1; +#else + tx_info->status.retry_count = 1; + tx_info->status.excessive_retries = false; +#endif + + } else { + tx_info->flags |= IEEE80211_TX_STAT_AMPDU | IEEE80211_TX_STAT_ACK; +#if (LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 39)) + tx_info->status.ampdu_ack_map = 1; +#else + tx_info->status.ampdu_len = 1; +#endif + tx_info->status.ampdu_ack_len = 1; + +#if (LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 29)) + /* manipulate rate status... */ + tx_info->status.rates[0].idx = 7; + tx_info->status.rates[0].count = 1; + tx_info->status.rates[0].flags = IEEE80211_TX_RC_MCS | IEEE80211_TX_RC_SHORT_GI; + tx_info->status.rates[1].idx = -1; +#else + tx_info->status.retry_count = 1; + tx_info->status.excessive_retries = false; +#endif + + } + + if(tx_info->flags & IEEE80211_TX_STAT_AMPDU) + esp_sip_dbg(ESP_DBG_TRACE, "%s ampdu status! \n", __func__); + + if (!mod_support_no_txampdu() && +#if (LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 29)) + sip->epub->hw->conf.channel_type != NL80211_CHAN_NO_HT +#else + !(sip->epub->hw->conf.flags&IEEE80211_CONF_SUPPORT_HT_MODE) +#endif + ) { + struct ieee80211_tx_info * tx_info = IEEE80211_SKB_CB(skb); + struct ieee80211_hdr * wh = (struct ieee80211_hdr *)skb->data; + if(ieee80211_is_data_qos(wh->frame_control)) { + if(!(tx_info->flags & IEEE80211_TX_CTL_AMPDU)) { + u8 tidno = ieee80211_get_qos_ctl(wh)[0] & IEEE80211_QOS_CTL_TID_MASK; + struct esp_node * node; + struct esp_tx_tid *tid; +#if (LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 28)) + struct ieee80211_sta *sta; + sta = tx_info->control.sta; + if(sta == NULL) + goto _exit; + node = (struct esp_node *)sta->drv_priv; + ASSERT(node != NULL); + if(node->sta == NULL) + goto _exit; +#else + node = esp_get_node_by_addr(sip->epub, wh->addr1); + if(node == NULL) + goto _exit; +#endif + tid = &node->tid[tidno]; + spin_lock_bh(&sip->epub->tx_ampdu_lock); + //start session + ASSERT(tid != NULL); + if ((tid->state == ESP_TID_STATE_INIT) && + (TID_TO_AC(tidno) != WME_AC_VO) && tid->cnt >= 10) { + tid->state = ESP_TID_STATE_TRIGGER; + esp_sip_dbg(ESP_DBG_ERROR, "start tx ba session,addr:%pM,tid:%u\n", wh->addr1, tidno); + spin_unlock_bh(&sip->epub->tx_ampdu_lock); +#if (LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 28)) + ieee80211_start_tx_ba_session(sip->epub->hw, wh->addr1, tidno); +#elif (LINUX_VERSION_CODE <= KERNEL_VERSION(2, 6, 32)) + ieee80211_start_tx_ba_session(sip->epub->hw, sta->addr, tidno); +#elif (LINUX_VERSION_CODE <= KERNEL_VERSION(2, 6, 37)) + ieee80211_start_tx_ba_session(sta, tidno); +#else + ieee80211_start_tx_ba_session(sta, tidno, 0); +#endif + } else { + if(tid->state == ESP_TID_STATE_INIT) + tid->cnt++; + else + tid->cnt = 0; + spin_unlock_bh(&sip->epub->tx_ampdu_lock); + } + } + } + } +_exit: +#ifndef FAST_TX_NOWAIT + skb_queue_tail(&sip->epub->txdoneq, skb); +#else + ieee80211_tx_status(sip->epub->hw, skb); +#endif +} +#endif /* FAST_TX_STATUS */ + +/* + * NB: this routine should be locked when calling + */ +void +sip_txq_process(struct esp_pub *epub) +{ + struct sk_buff *skb; + struct esp_sip *sip = epub->sip; + u32 pkt_len = 0, tx_len = 0, blknum = 0; + bool queued_back = false; + bool out_of_credits = false; + struct ieee80211_tx_info *itx_info; + int pm_state = 0; + + while ((skb = skb_dequeue(&epub->txq))) { + + /* cmd skb->len does not include sip_hdr too */ + pkt_len = skb->len; + itx_info = IEEE80211_SKB_CB(skb); + if (itx_info->flags != 0xffffffff) { + pkt_len += roundup(sizeof(struct sip_hdr), 4); + if ((itx_info->flags & IEEE80211_TX_CTL_AMPDU) && (true || esp_is_ip_pkt(skb))) + pkt_len += roundup(sizeof(struct esp_tx_ampdu_entry), 4); + } + + /* current design simply requires every sip_hdr must be at the begin of mblk, that definitely + * need to be optimized, e.g. calulate remain length in the previous mblk, if it larger than + * certain threshold (e.g, whole pkt or > 50% of pkt or 2 x sizeof(struct sip_hdr), append pkt + * to the previous mblk. This might be done in sip_pack_pkt() + */ + pkt_len = roundup(pkt_len, sip->tx_blksz); + blknum = pkt_len / sip->tx_blksz; + esp_dbg(ESP_DBG_TRACE, "%s skb_len %d pkt_len %d blknum %d\n", __func__, skb->len, pkt_len, blknum); + + if (unlikely(blknum > atomic_read(&sip->tx_credits) - sip->credit_to_reserve)) { + esp_dbg(ESP_DBG_TRACE, "%s out of credits!\n", __func__); + STRACE_TX_OUT_OF_CREDIT_INC(); + queued_back = true; + out_of_credits = true; + /* + if (epub->hw) { + ieee80211_stop_queues(epub->hw); + atomic_set(&epub->txq_stopped, true); + } + */ + /* we will be back */ + break; + } + + tx_len += pkt_len; + if (tx_len >= SIP_TX_AGGR_BUF_SIZE) { + /* do we need to have limitation likemax 8 pkts in a row? */ + esp_dbg(ESP_DBG_TRACE, "%s too much pkts in one shot!\n", __func__); + STRACE_TX_ONE_SHOT_INC(); + tx_len -= pkt_len; + queued_back = true; + break; + } + + if (sip_pack_pkt(sip, skb, &pm_state) != 0) { + /* wrong pkt, won't send to target */ + tx_len -= pkt_len; + continue; + } + + esp_sip_dbg(ESP_DBG_TRACE, "%s:before sub, credits is %d\n", __func__, atomic_read(&sip->tx_credits)); + atomic_sub(blknum, &sip->tx_credits); + esp_sip_dbg(ESP_DBG_TRACE, "%s:after sub %d,credits remains %d\n", __func__, blknum, atomic_read(&sip->tx_credits)); + + } + + if (queued_back) { + skb_queue_head(&epub->txq, skb); + } + + if (atomic_read(&sip->state) == SIP_STOP +#ifdef HOST_RESET_BUG + || atomic_read(&epub->wl.off) == 1 +#endif + ) + { + queued_back = 1; + tx_len = 0; + sip_after_write_pkts(sip); + } + + if (tx_len) { + + sip_write_pkts(sip, pm_state); + + sip_after_write_pkts(sip); + } + + if (queued_back && !out_of_credits) { + + /* skb pending, do async process again */ + // if (!skb_queue_empty(&epub->txq)) + sip_trigger_txq_process(sip); + } +} + +static void sip_after_write_pkts(struct esp_sip *sip) +{ + //enable txq +#if 0 + if (atomic_read(&sip->epub->txq_stopped) == true && sip_queue_may_resume(sip)) { + atomic_set(&sip->epub->txq_stopped, false); + ieee80211_wake_queues(sip->epub->hw); + esp_sip_dbg(ESP_DBG_TRACE, "%s resume ieee80211 tx \n", __func__); + } +#endif + +#ifndef FAST_TX_NOWAIT + sip_txdoneq_process(sip); +#endif + //disable tx_data +#ifndef FAST_TX_STATUS + if (atomic_read(&sip->data_tx_stopped) == false && sip_tx_data_need_stop(sip)) { + esp_sip_dbg(ESP_DBG_TRACE, "%s data_tx_stopped \n", __func__); + atomic_set(&sip->data_tx_stopped, true); + } +#endif /* FAST_TX_STATUS */ +} + +#if (LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 35)) +/* + * old mac80211 (2.6.32.x) needs payload is 4 byte aligned, thus we need this hack. + * TBD: However, the latest mac80211 stack does not need this. we may + * need to check kernel version here... + */ +static void sip_check_skb_alignment(struct sk_buff *skb) +{ + struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)skb->data; + int hdrlen; + + hdrlen = ieee80211_hdrlen(hdr->frame_control); + +#if 0 + + /* TBD */ + if (rx->flags & IEEE80211_RX_AMSDU) + hdrlen += ETH_HLEN; + +#endif + + if (unlikely(((unsigned long)(skb->data + hdrlen)) & 3)) { + + esp_sip_dbg(ESP_DBG_TRACE, "%s adjust skb data postion \n", __func__); + skb_push(skb, 2); + memmove(skb->data, skb->data+2, skb->len-2); + skb_trim(skb, skb->len-2); + } +} +#endif /* !NEW_KERNEL */ + +int sip_channel_value_inconsistency(u8 *start, size_t len, unsigned channel) +{ + size_t left = len; + u8 *pos = start; + u8 *DS_Param = NULL; + u8 channel_parsed = 0xff; + bool found = false; + u8 ssid[33]; + + while(left >=2 && !found) { + u8 id, elen; + id = *pos++; + elen = *pos++; + left -= 2; + + if(elen > left) + break; + + switch (id) { + case WLAN_EID_SSID: + ASSERT(elen < 33); + memcpy(ssid, pos, elen); + ssid[elen] = 0; + esp_sip_dbg(ESP_DBG_TRACE, "ssid:%s\n", ssid); + break; + case WLAN_EID_SUPP_RATES: + break; + case WLAN_EID_FH_PARAMS: + break; + case WLAN_EID_DS_PARAMS: + DS_Param = pos; + found = true; + break; + default: + break; + } + + left -= elen; + pos += elen; + } + + if (DS_Param) { + channel_parsed = DS_Param[0]; + } else { + esp_dbg(ESP_DBG_ERROR, "DS_Param not found\n"); + return -1; + } + + return channel_parsed != channel; +} + +/* parse mac_rx_ctrl and return length */ +static int sip_parse_mac_rx_info(struct esp_sip *sip, struct esp_mac_rx_ctrl * mac_ctrl, struct sk_buff *skb) +{ + struct ieee80211_rx_status *rx_status = NULL; + struct ieee80211_hdr *hdr; + +#if (LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 32)) + rx_status = IEEE80211_SKB_RXCB(skb); +#else + rx_status = (struct ieee80211_rx_status *)skb->cb; +#endif + rx_status->freq = esp_ieee2mhz(mac_ctrl->channel); + +#if (LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 39)) + rx_status->signal = mac_ctrl->rssi + atomic_read(&sip->noise_floor); /* snr actually, need to offset noise floor e.g. -85 */ +#else + rx_status->signal = mac_ctrl->rssi; /* snr actually, need to offset noise floor e.g. -85 */ +#endif /* NEW_KERNEL */ + + hdr = (struct ieee80211_hdr *)skb->data; + if (mac_ctrl->damatch0 == 1 && mac_ctrl->bssidmatch0 == 1 /*match bssid and da, but beacon package contain other bssid*/ + && strncmp(hdr->addr2, sip->epub->wl.bssid, ETH_ALEN) == 0) { /* force match addr2 */ + if (++signal_loop >= SIGNAL_COUNT) { + avg_signal += rx_status->signal; + avg_signal /= SIGNAL_COUNT; + old_signal = rx_status->signal = (avg_signal + 5); + signal_loop = 0; + avg_signal = 0; + } else { + avg_signal += rx_status->signal; + rx_status->signal = old_signal; + } + } + +#if (LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 35)) +#define ESP_RSSI_MIN_RSSI (-90) +#define ESP_RSSI_MAX_RSSI (-45) + rx_status->noise = 0; /* TBD */ + rx_status->qual = (mac_ctrl->rssi - ESP_RSSI_MIN_RSSI)* 100/(ESP_RSSI_MAX_RSSI - ESP_RSSI_MIN_RSSI); + rx_status->qual = min(rx_status->qual, 100); + rx_status->qual = max(rx_status->qual, 0); +#undef ESP_RSSI_MAX_RSSI +#undef ESP_RSSI_MIN_RSSI +#endif /* !NEW_KERNEL && KERNEL_35*/ + rx_status->antenna = 0; /* one antenna for now */ + rx_status->band = IEEE80211_BAND_2GHZ; + rx_status->flag = RX_FLAG_DECRYPTED | RX_FLAG_MMIC_STRIPPED; + if (mac_ctrl->sig_mode) { + // 2.6.27 has RX_FLAG_RADIOTAP in enum mac80211_rx_flags in include/net/mac80211.h +#if (LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 29)) + rx_status->flag |= RX_FLAG_HT; + rx_status->rate_idx = mac_ctrl->MCS; + if(mac_ctrl->SGI) + rx_status->flag |= RX_FLAG_SHORT_GI; +#else + rx_status->rate_idx = esp_wmac_rate2idx(0xc);//ESP_RATE_54 +#endif + } else { + rx_status->rate_idx = esp_wmac_rate2idx(mac_ctrl->rate); + } + if (mac_ctrl->rxend_state == RX_FCS_ERR) + rx_status->flag |= RX_FLAG_FAILED_FCS_CRC; + + /* Mic error frame flag */ + if (mac_ctrl->rxend_state == RX_TKIPMIC_ERR || mac_ctrl->rxend_state == RX_CCMPMIC_ERR){ + if(atomic_read(&sip->epub->wl.tkip_key_set) == 1){ + rx_status->flag|= RX_FLAG_MMIC_ERROR; + atomic_set(&sip->epub->wl.tkip_key_set, 0); + printk("mic err\n"); + } else { + printk("mic err discard\n"); + } + } + + //esp_dbg(ESP_DBG_LOG, "%s freq: %u; signal: %d; rate_idx %d; flag: %d \n", __func__, rx_status->freq, rx_status->signal, rx_status->rate_idx, rx_status->flag); + + do { + struct ieee80211_hdr * wh = (struct ieee80211_hdr *)((u8 *)skb->data); + + if (ieee80211_is_beacon(wh->frame_control) || ieee80211_is_probe_resp(wh->frame_control)) { + struct ieee80211_mgmt * mgmt = (struct ieee80211_mgmt *)((u8 *)skb->data); + u8 *start = NULL; + size_t baselen, len = skb->len; + int inconsistency = 0; + + if (ieee80211_is_beacon(wh->frame_control)) { + start = mgmt->u.beacon.variable; + baselen = (u8 *) mgmt->u.beacon.variable - (u8 *) mgmt; + bcn_counter++; + } else if(ieee80211_is_probe_resp(wh->frame_control)) { + start = mgmt->u.probe_resp.variable; + baselen = (u8 *) mgmt->u.probe_resp.variable - (u8 *) mgmt; + probe_rsp_counter++; + } + + if (baselen > len) + return -1; + + inconsistency = sip_channel_value_inconsistency(start, len-baselen, mac_ctrl->channel); + + if (inconsistency) { + return -1; + } + } + +#ifdef KERNEL_IV_WAR + /* some kernel e.g. 3.0.8 wrongly handles non-encrypted pkt like eapol */ + if (ieee80211_is_data(wh->frame_control)) { + if( !ieee80211_has_protected(wh->frame_control)) { + esp_sip_dbg(ESP_DBG_TRACE, "%s kiv_war, add iv_stripped flag \n", __func__); + rx_status->flag |= RX_FLAG_IV_STRIPPED; + } else { + if ((atomic_read(&sip->epub->wl.ptk_cnt) == 0 && !(wh->addr1[0] & 0x1)) || + (atomic_read(&sip->epub->wl.gtk_cnt) == 0 && (wh->addr1[0] & 0x1))) + { + esp_dbg(ESP_DBG_TRACE, "%s ==kiv_war, got bogus enc pkt==\n", __func__); + rx_status->flag |= RX_FLAG_IV_STRIPPED; + //show_buf(skb->data, 32); + } + + esp_sip_dbg(ESP_DBG_TRACE, "%s kiv_war, got enc pkt \n", __func__); + } + } +#endif /* KERNEL_IV_WAR*/ + } while (0); + + return 0; +} + +static struct esp_mac_rx_ctrl *sip_parse_normal_mac_ctrl(struct sk_buff *skb, int * pkt_len_enc, int *buf_len, int *pulled_len) +{ + struct esp_mac_rx_ctrl *mac_ctrl = NULL; + struct sip_hdr *hdr =(struct sip_hdr *)skb->data; + int len_in_hdr = hdr->len; + + ASSERT(skb != NULL); + ASSERT(skb->len > SIP_MIN_DATA_PKT_LEN); + + skb_pull(skb, sizeof(struct sip_hdr)); + *pulled_len += sizeof(struct sip_hdr); + mac_ctrl = (struct esp_mac_rx_ctrl *)skb->data; + if(!mac_ctrl->Aggregation) { + ASSERT(pkt_len_enc != NULL); + ASSERT(buf_len != NULL); + *pkt_len_enc = (mac_ctrl->sig_mode?mac_ctrl->HT_length:mac_ctrl->legacy_length) - FCS_LEN; + *buf_len = len_in_hdr - sizeof(struct sip_hdr) - sizeof(struct esp_mac_rx_ctrl); + } + skb_pull(skb, sizeof(struct esp_mac_rx_ctrl)); + *pulled_len += sizeof(struct esp_mac_rx_ctrl); + + return mac_ctrl; +} + +/* + * for one MPDU (including subframe in AMPDU) + * + */ +static struct sk_buff * sip_parse_data_rx_info(struct esp_sip *sip, struct sk_buff *skb, int pkt_len_enc, int buf_len, struct esp_mac_rx_ctrl *mac_ctrl, int *pulled_len) { + /* + * | mac_rx_ctrl | real_data_payload | ampdu_entries | + */ + //without enc + int pkt_len = 0; + struct sk_buff *rskb = NULL; + int ret; + + if (mac_ctrl->Aggregation) { + struct ieee80211_hdr * wh = (struct ieee80211_hdr *)skb->data; + pkt_len = pkt_len_enc; + if (ieee80211_has_protected(wh->frame_control))//ampdu, it is CCMP enc + pkt_len -= 8; + buf_len = roundup(pkt_len, 4); + } else + pkt_len = buf_len - 3 + ((pkt_len_enc - 1) & 0x3); + esp_dbg(ESP_DBG_TRACE, "%s pkt_len %u, pkt_len_enc %u!, delta %d \n", __func__, pkt_len, pkt_len_enc, pkt_len_enc - pkt_len); + do { +#if (LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 39)) + rskb = __dev_alloc_skb(pkt_len_enc + 2, GFP_ATOMIC); +#else + rskb = __dev_alloc_skb(pkt_len_enc, GFP_ATOMIC); +#endif/* NEW_KERNEL */ + if (unlikely(rskb == NULL)) { + esp_sip_dbg(ESP_DBG_ERROR, "%s no mem for rskb\n", __func__); + return NULL; + } + skb_put(rskb, pkt_len_enc); + } while(0); + + do { +#if (LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 39)) + do { + struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)skb->data; + int hdrlen; + + hdrlen = ieee80211_hdrlen(hdr->frame_control); + if (unlikely(((unsigned long)(rskb->data + hdrlen)) & 3)) { + skb_put(rskb, 2); + skb_pull(rskb, 2); + } + } while(0); +#endif /* < KERNEL_VERSION(2, 6, 39) */ + memcpy(rskb->data, skb->data, pkt_len); + if (pkt_len_enc > pkt_len) { + memset(rskb->data + pkt_len, 0, pkt_len_enc - pkt_len); + } + /* strip out current pkt, move to the next one */ + skb_pull(skb, buf_len); + *pulled_len += buf_len; + } while (0); + + ret = sip_parse_mac_rx_info(sip, mac_ctrl, rskb); + if(ret == -1 && !mac_ctrl->Aggregation) { + kfree_skb(rskb); + return NULL; + } + + esp_dbg(ESP_DBG_LOG, "%s after pull headers, skb->len %d rskb->len %d \n", __func__, skb->len, rskb->len); + + return rskb; +} + +struct esp_sip * sip_attach(struct esp_pub *epub) +{ + struct esp_sip *sip = NULL; + struct sip_pkt *pkt = NULL; + int i; + + sip = kzalloc(sizeof(struct esp_sip), GFP_KERNEL); + ASSERT(sip != NULL); + +#ifdef ESP_RX_COPYBACK_TEST + /* alloc 64KB for rx test */ + copyback_buf = kzalloc(0x10000, GFP_KERNEL); +#endif /* ESP_RX_COPYBACK_TEST */ + + spin_lock_init(&sip->lock); + + INIT_LIST_HEAD(&sip->free_ctrl_txbuf); + INIT_LIST_HEAD(&sip->free_ctrl_rxbuf); + + for (i = 0; i < SIP_CTRL_BUF_N; i++) { + pkt = kzalloc(sizeof(struct sip_pkt), GFP_KERNEL); + + if (!pkt) break; + + pkt->buf_begin = kzalloc(SIP_CTRL_BUF_SZ, GFP_KERNEL); + + if (pkt->buf_begin == NULL) { + kfree(pkt); + break; + } + + pkt->buf_len = SIP_CTRL_BUF_SZ; + pkt->buf = pkt->buf_begin; + + if (i < SIP_CTRL_TXBUF_N) { + list_add_tail(&pkt->list, &sip->free_ctrl_txbuf); + } else { + list_add_tail(&pkt->list, &sip->free_ctrl_rxbuf); + } + } + + spin_lock_init(&sip->rx_lock); + skb_queue_head_init(&sip->rxq); +#ifndef RX_SYNC + INIT_WORK(&sip->rx_process_work, sip_rxq_process); +#endif/* RX_SYNC */ + + sip->epub = epub; + atomic_set(&sip->noise_floor, -96); + + atomic_set(&sip->state, SIP_INIT); + atomic_set(&sip->tx_credits, 0); + + return sip; +} + +static void sip_free_init_ctrl_buf(struct esp_sip *sip) +{ + struct sip_pkt *pkt, *tpkt; + + list_for_each_entry_safe(pkt, tpkt, + &sip->free_ctrl_txbuf, list) { + list_del(&pkt->list); + kfree(pkt->buf_begin); + kfree(pkt); + } + + list_for_each_entry_safe(pkt, tpkt, + &sip->free_ctrl_rxbuf, list) { + list_del(&pkt->list); + kfree(pkt->buf_begin); + kfree(pkt); + } +} + +void sip_detach(struct esp_sip *sip) +{ + int po; + + sip_free_init_ctrl_buf(sip); + + if (atomic_read(&sip->state) == SIP_RUN) { + + sif_disable_target_interrupt(sip->epub); + + atomic_set(&sip->state, SIP_STOP); + + /* disable irq here */ + sif_disable_irq(sip->epub); +#ifndef RX_SYNC + cancel_work_sync(&sip->rx_process_work); +#endif/* RX_SYNC */ + + skb_queue_purge(&sip->rxq); + cancel_work_sync(&sip->epub->sendup_work); + skb_queue_purge(&sip->epub->rxq); + +#ifdef ESP_NO_MAC80211 + unregister_netdev(sip->epub->net_dev); + wiphy_unregister(sip->epub->wdev->wiphy); +#else + if (test_and_clear_bit(ESP_WL_FLAG_HW_REGISTERED, &sip->epub->wl.flags)) { + ieee80211_unregister_hw(sip->epub->hw); + } +#endif + + /* cancel all worker/timer */ + cancel_work_sync(&sip->epub->tx_work); + skb_queue_purge(&sip->epub->txq); + skb_queue_purge(&sip->epub->txdoneq); + + po = get_order(SIP_TX_AGGR_BUF_SIZE); + free_pages((unsigned long)sip->tx_aggr_buf, po); + sip->tx_aggr_buf = NULL; +#if 0 + po = get_order(SIP_RX_AGGR_BUF_SIZE); + free_pages((unsigned long)sip->rx_aggr_buf, po); + sip->rx_aggr_buf = NULL; +#endif + atomic_set(&sip->state, SIP_INIT); + } else if (atomic_read(&sip->state) >= SIP_BOOT && atomic_read(&sip->state) <= SIP_WAIT_BOOTUP) { + + sif_disable_target_interrupt(sip->epub); + atomic_set(&sip->state, SIP_STOP); + sif_disable_irq(sip->epub); + + if (sip->rawbuf) + kfree(sip->rawbuf); + + if (atomic_read(&sip->state) == SIP_SEND_INIT) { +#ifndef RX_SYNC + cancel_work_sync(&sip->rx_process_work); +#endif/* RX_SYNC */ + skb_queue_purge(&sip->rxq); + cancel_work_sync(&sip->epub->sendup_work); + skb_queue_purge(&sip->epub->rxq); + } + +#ifdef ESP_NO_MAC80211 + unregister_netdev(sip->epub->net_dev); + wiphy_unregister(sip->epub->wdev->wiphy); +#else + if (test_and_clear_bit(ESP_WL_FLAG_HW_REGISTERED, &sip->epub->wl.flags)) { + ieee80211_unregister_hw(sip->epub->hw); + } +#endif + atomic_set(&sip->state, SIP_INIT); + } else + esp_dbg(ESP_DBG_ERROR, "%s wrong state %d\n", __func__, atomic_read(&sip->state)); + + kfree(sip); +} + +int sip_prepare_boot(struct esp_sip *sip) +{ + if (atomic_read(&sip->state) != SIP_INIT) { + esp_dbg(ESP_DBG_ERROR, "%s wrong state %d\n", __func__, atomic_read(&sip->state)); + return -ENOTRECOVERABLE; + } + + if (sip->rawbuf == NULL) { + sip->rawbuf = kzalloc(SIP_BOOT_BUF_SIZE, GFP_KERNEL); + + if (sip->rawbuf == NULL) + return -ENOMEM; + } + + atomic_set(&sip->state, SIP_PREPARE_BOOT); + + return 0; +} + +int sip_write_memory(struct esp_sip *sip, u32 addr, u8 *buf, u16 len) +{ + struct sip_cmd_write_memory *cmd; + struct sip_hdr *chdr; + u16 remains, hdrs, bufsize; + u32 loadaddr; + u8 *src; + int err = 0; + u32 *t = 0; + + ASSERT(sip->rawbuf != NULL); + + memset(sip->rawbuf, 0, SIP_BOOT_BUF_SIZE); + + chdr = (struct sip_hdr *)sip->rawbuf; + SIP_HDR_SET_TYPE(chdr->fc[0], SIP_CTRL); + chdr->c_cmdid = SIP_CMD_WRITE_MEMORY; + + remains = len; + hdrs = sizeof(struct sip_hdr) + sizeof(struct sip_cmd_write_memory); + + while (remains) { + src = &buf[len - remains]; + loadaddr = addr + (len - remains); + + if (remains < (SIP_BOOT_BUF_SIZE - hdrs)) { + /* aligned with 4 bytes */ + bufsize = roundup(remains, 4); + memset(sip->rawbuf + hdrs, 0, bufsize); + remains = 0; + } else { + bufsize = SIP_BOOT_BUF_SIZE - hdrs; + remains -= bufsize; + } + + chdr->len = bufsize + hdrs; + chdr->seq = sip->txseq++; + cmd = (struct sip_cmd_write_memory *)(sip->rawbuf + SIP_CTRL_HDR_LEN); + cmd->len = bufsize; + cmd->addr = loadaddr; + memcpy(sip->rawbuf+hdrs, src, bufsize); + + t = (u32 *)sip->rawbuf; + esp_dbg(ESP_DBG_TRACE, "%s t0: 0x%08x t1: 0x%08x t2:0x%08x loadaddr 0x%08x \n", __func__, t[0], t[1], t[2], loadaddr); + + err = esp_common_write(sip->epub, sip->rawbuf, chdr->len, ESP_SIF_SYNC); + + if (err) { + esp_dbg(ESP_DBG_ERROR, "%s send buffer failed\n", __func__); + return err; + } + + // 1ms is enough, in fact on dell-d430, need not delay at all. + mdelay(1); + + sip_dec_credit(sip); + } + + return err; +} + +int sip_send_cmd(struct esp_sip *sip, int cid, u32 cmdlen, void *cmd) +{ + struct sip_hdr *chdr; + struct sip_pkt *pkt = NULL; + int ret = 0; + + pkt = sip_get_ctrl_buf(sip, SIP_TX_CTRL_BUF); + + if (pkt == NULL) + return -ENOMEM; + + chdr = (struct sip_hdr *)pkt->buf_begin; + chdr->len = SIP_CTRL_HDR_LEN + cmdlen; + chdr->seq = sip->txseq++; + chdr->c_cmdid = cid; + + + if (cmd) { + memset(pkt->buf, 0, cmdlen); + memcpy(pkt->buf, (u8 *)cmd, cmdlen); + } + + esp_dbg(ESP_DBG_TRACE, "cid %d, len %u, seq %u \n", chdr->c_cmdid, chdr->len, chdr->seq); + + esp_dbg(ESP_DBG_TRACE, "c1 0x%08x c2 0x%08x\n", *(u32 *)&pkt->buf[0], *(u32 *)&pkt->buf[4]); + + ret = esp_common_write(sip->epub, pkt->buf_begin, chdr->len, ESP_SIF_SYNC); + + if (ret) + esp_dbg(ESP_DBG_ERROR, "%s send cmd %d failed \n", __func__, cid); + + sip_dec_credit(sip); + + sip_reclaim_ctrl_buf(sip, pkt, SIP_TX_CTRL_BUF); + + /* + * Hack here: reset tx/rx seq before target ram code is up... + */ + if (cid == SIP_CMD_BOOTUP) { + sip->rxseq = 0; + sip->txseq = 0; + sip->txdataseq = 0; + } + + return ret; +} + +struct sk_buff * +sip_alloc_ctrl_skbuf(struct esp_sip *sip, u16 len, u32 cid) { + struct sip_hdr *si = NULL; + struct ieee80211_tx_info *ti = NULL; + struct sk_buff *skb = NULL; + + ASSERT(len <= sip->tx_blksz); + + /* no need to reserve space for net stack */ + skb = __dev_alloc_skb(len, GFP_KERNEL); + + if (skb == NULL) { + esp_dbg(ESP_DBG_ERROR, "no skb for ctrl !\n"); + return NULL; + } + + skb->len = len; + + ti = IEEE80211_SKB_CB(skb); + /* set tx_info flags to 0xffffffff to indicate sip_ctrl pkt */ + ti->flags = 0xffffffff; + si = (struct sip_hdr *)skb->data; + memset(si, 0, sizeof(struct sip_hdr)); + SIP_HDR_SET_TYPE(si->fc[0], SIP_CTRL); + si->len = len; + si->c_cmdid = cid; + + return skb; +} + +void +sip_free_ctrl_skbuff(struct esp_sip *sip, struct sk_buff *skb) +{ + memset(IEEE80211_SKB_CB(skb), 0, sizeof(struct ieee80211_tx_info)); + kfree_skb(skb); +} + +static struct sip_pkt * +sip_get_ctrl_buf(struct esp_sip *sip, SIP_BUF_TYPE bftype) { + struct sip_pkt *pkt = NULL; + struct list_head *bflist; + struct sip_hdr *chdr; + + bflist = (bftype == SIP_TX_CTRL_BUF) ? &sip->free_ctrl_txbuf :&sip->free_ctrl_rxbuf; + + spin_lock_bh(&sip->lock); + + if (list_empty(bflist)) { + spin_unlock_bh(&sip->lock); + return NULL; + } + + pkt = list_first_entry(bflist, struct sip_pkt, list); + list_del(&pkt->list); + spin_unlock_bh(&sip->lock); + + if (bftype == SIP_TX_CTRL_BUF) { + chdr = (struct sip_hdr *)pkt->buf_begin; + SIP_HDR_SET_TYPE(chdr->fc[0], SIP_CTRL); + pkt->buf = pkt->buf_begin + SIP_CTRL_HDR_LEN; + } else { + pkt->buf = pkt->buf_begin; + } + + return pkt; +} + +static void +sip_reclaim_ctrl_buf(struct esp_sip *sip, struct sip_pkt *pkt, SIP_BUF_TYPE bftype) +{ + struct list_head *bflist = NULL; + + if (bftype == SIP_TX_CTRL_BUF) + bflist = &sip->free_ctrl_txbuf; + else if (bftype == SIP_RX_CTRL_BUF) + bflist = &sip->free_ctrl_rxbuf; + else return; + + pkt->buf = pkt->buf_begin; + pkt->payload_len = 0; + + spin_lock_bh(&sip->lock); + list_add_tail(&pkt->list, bflist); + spin_unlock_bh(&sip->lock); +} + +int +sip_poll_bootup_event(struct esp_sip *sip) +{ + int ret = 0; + + esp_dbg(ESP_DBG_TRACE, "polling bootup event... \n"); + + if (gl_bootup_cplx) + ret = wait_for_completion_timeout(gl_bootup_cplx, 2 * HZ); + + esp_dbg(ESP_DBG_TRACE, "******time remain****** = [%d]\n", ret); + if (ret <= 0) { + esp_dbg(ESP_DBG_ERROR, "bootup event timeout\n"); + return -ETIMEDOUT; + } + +#ifndef FPGA_LOOPBACK + ret = esp_register_mac80211(sip->epub); +#endif /* !FPGA_LOOPBACK */ + +#ifdef TEST_MODE + ret = test_init_netlink(sip); + if (ret < 0) { + esp_sip_dbg(ESP_DBG_TRACE, "esp_sdio: failed initializing netlink\n"); + return ret; + } +#endif + + atomic_set(&sip->state, SIP_RUN); + esp_dbg(ESP_DBG_TRACE, "target booted up\n"); + + return ret; +} + +int +sip_poll_resetting_event(struct esp_sip *sip) +{ + int ret = 0; + + esp_dbg(ESP_DBG_TRACE, "polling resetting event... \n"); + + if (gl_bootup_cplx) + ret = wait_for_completion_timeout(gl_bootup_cplx, 10 * HZ); + + esp_dbg(ESP_DBG_TRACE, "******time remain****** = [%d]\n", ret); + if (ret <= 0) { + esp_dbg(ESP_DBG_ERROR, "resetting event timeout\n"); + return -ETIMEDOUT; + } + + esp_dbg(ESP_DBG_TRACE, "target resetting %d %p\n", ret, gl_bootup_cplx); + + return 0; +} + + +#ifdef FPGA_DEBUG + +/* bogus bootup cmd for FPGA debugging */ +int +sip_send_bootup(struct esp_sip *sip) +{ + int ret; + struct sip_cmd_bootup bootcmd; + + esp_dbg(ESP_DBG_LOG, "sending bootup\n"); + + bootcmd.boot_addr = 0; + ret = sip_send_cmd(sip, SIP_CMD_BOOTUP, sizeof(struct sip_cmd_bootup), &bootcmd); + + return ret; +} + +#endif /* FPGA_DEBUG */ + +bool +sip_queue_need_stop(struct esp_sip *sip) +{ + return atomic_read(&sip->tx_data_pkt_queued) >= SIP_STOP_QUEUE_THRESHOLD + || (atomic_read(&sip->tx_credits) < 8 + && atomic_read(&sip->tx_data_pkt_queued) >= SIP_STOP_QUEUE_THRESHOLD / 4 * 3); +} + +bool +sip_queue_may_resume(struct esp_sip *sip) +{ + return atomic_read(&sip->epub->txq_stopped) + && !test_bit(ESP_WL_FLAG_STOP_TXQ, &sip->epub->wl.flags) + && ((atomic_read(&sip->tx_credits) >= 16 + && atomic_read(&sip->tx_data_pkt_queued) < SIP_RESUME_QUEUE_THRESHOLD * 2) + || atomic_read(&sip->tx_data_pkt_queued) < SIP_RESUME_QUEUE_THRESHOLD); +} + +#ifndef FAST_TX_STATUS +bool +sip_tx_data_need_stop(struct esp_sip *sip) +{ + return atomic_read(&sip->pending_tx_status) >= SIP_PENDING_STOP_TX_THRESHOLD; +} + +bool +sip_tx_data_may_resume(struct esp_sip *sip) +{ + return atomic_read(&sip->pending_tx_status) < SIP_PENDING_RESUME_TX_THRESHOLD; +} +#endif /* FAST_TX_STATUS */ + +int +sip_cmd_enqueue(struct esp_sip *sip, struct sk_buff *skb) +{ + if (!sip || !sip->epub) { + esp_dbg(ESP_DBG_ERROR, "func %s, sip->epub->txq is NULL\n", __func__); + return -EINVAL; + } + + if (!skb) { + esp_dbg(ESP_DBG_ERROR, "func %s, skb is NULL\n", __func__); + return -EINVAL; + } + + skb_queue_tail(&sip->epub->txq, skb); + +#if !defined(FPGA_LOOPBACK) && (LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 32)) + ieee80211_queue_work(sip->epub->hw, &sip->epub->tx_work); +#else + queue_work(sip->epub->esp_wkq, &sip->epub->tx_work); +#endif + return 0; +} + +void sip_tx_data_pkt_enqueue(struct esp_pub *epub, struct sk_buff *skb) +{ + if(!epub || !epub->sip) { + if (!epub) + esp_dbg(ESP_DBG_ERROR, "func %s, epub is NULL\n", __func__); + else + esp_dbg(ESP_DBG_ERROR, "func %s, epub->sip is NULL\n", __func__); + + return; + } + if (!skb) { + esp_dbg(ESP_DBG_ERROR, "func %s, skb is NULL\n", __func__); + return; + } + skb_queue_tail(&epub->txq, skb); + atomic_inc(&epub->sip->tx_data_pkt_queued); + if(sip_queue_need_stop(epub->sip)){ + if (epub->hw) { + ieee80211_stop_queues(epub->hw); + atomic_set(&epub->txq_stopped, true); + } + + } +} + +#ifdef FPGA_TXDATA +int sip_send_tx_data(struct esp_sip *sip) +{ + struct sk_buff *skb = NULL; + struct sip_cmd_bss_info_update*bsscmd; + + skb = sip_alloc_ctrl_skbuf(epub->sip, sizeof(struct sip_cmd_bss_info_update), SIP_CMD_BSS_INFO_UPDATE); + if (!skb) + return -EINVAL; + + bsscmd = (struct sip_cmd_bss_info_update *)(skb->data + sizeof(struct sip_tx_info)); + bsscmd->isassoc= (assoc==true)? 1: 0; + memcpy(bsscmd->bssid, bssid, ETH_ALEN); + STRACE_SHOW(epub->sip); + return sip_cmd_enqueue(epub->sip, skb); +} +#endif /* FPGA_TXDATA */ + +#ifdef SIP_DEBUG +void sip_dump_pending_data(struct esp_pub *epub) +{ +#if 0 + struct sk_buff *tskb, *tmp; + + skb_queue_walk_safe(&epub->txdoneq, tskb, tmp) { + show_buf(tskb->data, 32); + } +#endif //0000 +} +#else +void sip_dump_pending_data(struct esp_pub *epub) +{} +#endif /* SIP_DEBUG */ + |