mirror of
https://github.com/Qortal/Brooklyn.git
synced 2025-02-01 07:42:18 +00:00
379 lines
7.5 KiB
C
379 lines
7.5 KiB
C
// SPDX-License-Identifier: ISC
|
|
/* Copyright (C) 2020 MediaTek Inc.
|
|
*
|
|
* This file is written based on mt76/usb.c.
|
|
*
|
|
* Author: Felix Fietkau <nbd@nbd.name>
|
|
* Lorenzo Bianconi <lorenzo@kernel.org>
|
|
* Sean Wang <sean.wang@mediatek.com>
|
|
*/
|
|
|
|
#include <linux/iopoll.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/module.h>
|
|
#include <linux/mmc/sdio_func.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/kthread.h>
|
|
|
|
#include "mt76.h"
|
|
|
|
static int
|
|
mt76s_alloc_rx_queue(struct mt76_dev *dev, enum mt76_rxq_id qid)
|
|
{
|
|
struct mt76_queue *q = &dev->q_rx[qid];
|
|
|
|
spin_lock_init(&q->lock);
|
|
q->entry = devm_kcalloc(dev->dev,
|
|
MT_NUM_RX_ENTRIES, sizeof(*q->entry),
|
|
GFP_KERNEL);
|
|
if (!q->entry)
|
|
return -ENOMEM;
|
|
|
|
q->ndesc = MT_NUM_RX_ENTRIES;
|
|
q->head = q->tail = 0;
|
|
q->queued = 0;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static struct mt76_queue *mt76s_alloc_tx_queue(struct mt76_dev *dev)
|
|
{
|
|
struct mt76_queue *q;
|
|
|
|
q = devm_kzalloc(dev->dev, sizeof(*q), GFP_KERNEL);
|
|
if (!q)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
spin_lock_init(&q->lock);
|
|
q->entry = devm_kcalloc(dev->dev,
|
|
MT_NUM_TX_ENTRIES, sizeof(*q->entry),
|
|
GFP_KERNEL);
|
|
if (!q->entry)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
q->ndesc = MT_NUM_TX_ENTRIES;
|
|
|
|
return q;
|
|
}
|
|
|
|
static int mt76s_alloc_tx(struct mt76_dev *dev)
|
|
{
|
|
struct mt76_queue *q;
|
|
int i;
|
|
|
|
for (i = 0; i <= MT_TXQ_PSD; i++) {
|
|
q = mt76s_alloc_tx_queue(dev);
|
|
if (IS_ERR(q))
|
|
return PTR_ERR(q);
|
|
|
|
q->qid = i;
|
|
dev->phy.q_tx[i] = q;
|
|
}
|
|
|
|
q = mt76s_alloc_tx_queue(dev);
|
|
if (IS_ERR(q))
|
|
return PTR_ERR(q);
|
|
|
|
q->qid = MT_MCUQ_WM;
|
|
dev->q_mcu[MT_MCUQ_WM] = q;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int mt76s_alloc_queues(struct mt76_dev *dev)
|
|
{
|
|
int err;
|
|
|
|
err = mt76s_alloc_rx_queue(dev, MT_RXQ_MAIN);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
return mt76s_alloc_tx(dev);
|
|
}
|
|
EXPORT_SYMBOL_GPL(mt76s_alloc_queues);
|
|
|
|
static struct mt76_queue_entry *
|
|
mt76s_get_next_rx_entry(struct mt76_queue *q)
|
|
{
|
|
struct mt76_queue_entry *e = NULL;
|
|
|
|
spin_lock_bh(&q->lock);
|
|
if (q->queued > 0) {
|
|
e = &q->entry[q->tail];
|
|
q->tail = (q->tail + 1) % q->ndesc;
|
|
q->queued--;
|
|
}
|
|
spin_unlock_bh(&q->lock);
|
|
|
|
return e;
|
|
}
|
|
|
|
static int
|
|
mt76s_process_rx_queue(struct mt76_dev *dev, struct mt76_queue *q)
|
|
{
|
|
int qid = q - &dev->q_rx[MT_RXQ_MAIN];
|
|
int nframes = 0;
|
|
|
|
while (true) {
|
|
struct mt76_queue_entry *e;
|
|
|
|
if (!test_bit(MT76_STATE_INITIALIZED, &dev->phy.state))
|
|
break;
|
|
|
|
e = mt76s_get_next_rx_entry(q);
|
|
if (!e || !e->skb)
|
|
break;
|
|
|
|
dev->drv->rx_skb(dev, MT_RXQ_MAIN, e->skb);
|
|
e->skb = NULL;
|
|
nframes++;
|
|
}
|
|
if (qid == MT_RXQ_MAIN)
|
|
mt76_rx_poll_complete(dev, MT_RXQ_MAIN, NULL);
|
|
|
|
return nframes;
|
|
}
|
|
|
|
static void mt76s_net_worker(struct mt76_worker *w)
|
|
{
|
|
struct mt76_sdio *sdio = container_of(w, struct mt76_sdio,
|
|
net_worker);
|
|
struct mt76_dev *dev = container_of(sdio, struct mt76_dev, sdio);
|
|
int i, nframes;
|
|
|
|
do {
|
|
nframes = 0;
|
|
|
|
local_bh_disable();
|
|
rcu_read_lock();
|
|
|
|
mt76_for_each_q_rx(dev, i)
|
|
nframes += mt76s_process_rx_queue(dev, &dev->q_rx[i]);
|
|
|
|
rcu_read_unlock();
|
|
local_bh_enable();
|
|
} while (nframes > 0);
|
|
}
|
|
|
|
static int mt76s_process_tx_queue(struct mt76_dev *dev, struct mt76_queue *q)
|
|
{
|
|
struct mt76_queue_entry entry;
|
|
int nframes = 0;
|
|
bool mcu;
|
|
|
|
if (!q)
|
|
return 0;
|
|
|
|
mcu = q == dev->q_mcu[MT_MCUQ_WM];
|
|
while (q->queued > 0) {
|
|
if (!q->entry[q->tail].done)
|
|
break;
|
|
|
|
entry = q->entry[q->tail];
|
|
q->entry[q->tail].done = false;
|
|
|
|
if (mcu) {
|
|
dev_kfree_skb(entry.skb);
|
|
entry.skb = NULL;
|
|
}
|
|
|
|
mt76_queue_tx_complete(dev, q, &entry);
|
|
nframes++;
|
|
}
|
|
|
|
if (!q->queued)
|
|
wake_up(&dev->tx_wait);
|
|
|
|
if (!mcu)
|
|
mt76_txq_schedule(&dev->phy, q->qid);
|
|
|
|
return nframes;
|
|
}
|
|
|
|
static void mt76s_status_worker(struct mt76_worker *w)
|
|
{
|
|
struct mt76_sdio *sdio = container_of(w, struct mt76_sdio,
|
|
status_worker);
|
|
struct mt76_dev *dev = container_of(sdio, struct mt76_dev, sdio);
|
|
int i, nframes;
|
|
|
|
do {
|
|
nframes = mt76s_process_tx_queue(dev, dev->q_mcu[MT_MCUQ_WM]);
|
|
|
|
for (i = 0; i <= MT_TXQ_PSD; i++)
|
|
nframes += mt76s_process_tx_queue(dev,
|
|
dev->phy.q_tx[i]);
|
|
|
|
if (dev->drv->tx_status_data &&
|
|
!test_and_set_bit(MT76_READING_STATS, &dev->phy.state))
|
|
queue_work(dev->wq, &dev->sdio.stat_work);
|
|
} while (nframes > 0);
|
|
}
|
|
|
|
static void mt76s_tx_status_data(struct work_struct *work)
|
|
{
|
|
struct mt76_sdio *sdio;
|
|
struct mt76_dev *dev;
|
|
u8 update = 1;
|
|
u16 count = 0;
|
|
|
|
sdio = container_of(work, struct mt76_sdio, stat_work);
|
|
dev = container_of(sdio, struct mt76_dev, sdio);
|
|
|
|
while (true) {
|
|
if (test_bit(MT76_REMOVED, &dev->phy.state))
|
|
break;
|
|
|
|
if (!dev->drv->tx_status_data(dev, &update))
|
|
break;
|
|
count++;
|
|
}
|
|
|
|
if (count && test_bit(MT76_STATE_RUNNING, &dev->phy.state))
|
|
queue_work(dev->wq, &sdio->stat_work);
|
|
else
|
|
clear_bit(MT76_READING_STATS, &dev->phy.state);
|
|
}
|
|
|
|
static int
|
|
mt76s_tx_queue_skb(struct mt76_dev *dev, struct mt76_queue *q,
|
|
struct sk_buff *skb, struct mt76_wcid *wcid,
|
|
struct ieee80211_sta *sta)
|
|
{
|
|
struct mt76_tx_info tx_info = {
|
|
.skb = skb,
|
|
};
|
|
int err, len = skb->len;
|
|
u16 idx = q->head;
|
|
|
|
if (q->queued == q->ndesc)
|
|
return -ENOSPC;
|
|
|
|
skb->prev = skb->next = NULL;
|
|
err = dev->drv->tx_prepare_skb(dev, NULL, q->qid, wcid, sta, &tx_info);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
q->entry[q->head].skb = tx_info.skb;
|
|
q->entry[q->head].buf_sz = len;
|
|
|
|
smp_wmb();
|
|
|
|
q->head = (q->head + 1) % q->ndesc;
|
|
q->queued++;
|
|
|
|
return idx;
|
|
}
|
|
|
|
static int
|
|
mt76s_tx_queue_skb_raw(struct mt76_dev *dev, struct mt76_queue *q,
|
|
struct sk_buff *skb, u32 tx_info)
|
|
{
|
|
int ret = -ENOSPC, len = skb->len, pad;
|
|
|
|
if (q->queued == q->ndesc)
|
|
goto error;
|
|
|
|
pad = round_up(skb->len, 4) - skb->len;
|
|
ret = mt76_skb_adjust_pad(skb, pad);
|
|
if (ret)
|
|
goto error;
|
|
|
|
spin_lock_bh(&q->lock);
|
|
|
|
q->entry[q->head].buf_sz = len;
|
|
q->entry[q->head].skb = skb;
|
|
q->head = (q->head + 1) % q->ndesc;
|
|
q->queued++;
|
|
|
|
spin_unlock_bh(&q->lock);
|
|
|
|
return 0;
|
|
|
|
error:
|
|
dev_kfree_skb(skb);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void mt76s_tx_kick(struct mt76_dev *dev, struct mt76_queue *q)
|
|
{
|
|
struct mt76_sdio *sdio = &dev->sdio;
|
|
|
|
mt76_worker_schedule(&sdio->txrx_worker);
|
|
}
|
|
|
|
static const struct mt76_queue_ops sdio_queue_ops = {
|
|
.tx_queue_skb = mt76s_tx_queue_skb,
|
|
.kick = mt76s_tx_kick,
|
|
.tx_queue_skb_raw = mt76s_tx_queue_skb_raw,
|
|
};
|
|
|
|
void mt76s_deinit(struct mt76_dev *dev)
|
|
{
|
|
struct mt76_sdio *sdio = &dev->sdio;
|
|
int i;
|
|
|
|
mt76_worker_teardown(&sdio->txrx_worker);
|
|
mt76_worker_teardown(&sdio->status_worker);
|
|
mt76_worker_teardown(&sdio->net_worker);
|
|
|
|
cancel_work_sync(&sdio->stat_work);
|
|
clear_bit(MT76_READING_STATS, &dev->phy.state);
|
|
|
|
mt76_tx_status_check(dev, NULL, true);
|
|
|
|
sdio_claim_host(sdio->func);
|
|
sdio_release_irq(sdio->func);
|
|
sdio_release_host(sdio->func);
|
|
|
|
mt76_for_each_q_rx(dev, i) {
|
|
struct mt76_queue *q = &dev->q_rx[i];
|
|
int j;
|
|
|
|
for (j = 0; j < q->ndesc; j++) {
|
|
struct mt76_queue_entry *e = &q->entry[j];
|
|
|
|
if (!e->skb)
|
|
continue;
|
|
|
|
dev_kfree_skb(e->skb);
|
|
e->skb = NULL;
|
|
}
|
|
}
|
|
}
|
|
EXPORT_SYMBOL_GPL(mt76s_deinit);
|
|
|
|
int mt76s_init(struct mt76_dev *dev, struct sdio_func *func,
|
|
const struct mt76_bus_ops *bus_ops)
|
|
{
|
|
struct mt76_sdio *sdio = &dev->sdio;
|
|
int err;
|
|
|
|
err = mt76_worker_setup(dev->hw, &sdio->status_worker,
|
|
mt76s_status_worker, "sdio-status");
|
|
if (err)
|
|
return err;
|
|
|
|
err = mt76_worker_setup(dev->hw, &sdio->net_worker, mt76s_net_worker,
|
|
"sdio-net");
|
|
if (err)
|
|
return err;
|
|
|
|
sched_set_fifo_low(sdio->status_worker.task);
|
|
sched_set_fifo_low(sdio->net_worker.task);
|
|
|
|
INIT_WORK(&sdio->stat_work, mt76s_tx_status_data);
|
|
|
|
dev->queue_ops = &sdio_queue_ops;
|
|
dev->bus = bus_ops;
|
|
dev->sdio.func = func;
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(mt76s_init);
|
|
|
|
MODULE_AUTHOR("Sean Wang <sean.wang@mediatek.com>");
|
|
MODULE_AUTHOR("Lorenzo Bianconi <lorenzo@kernel.org>");
|
|
MODULE_LICENSE("Dual BSD/GPL");
|