Commit d7d4ea9a authored by Stanislaw Gruszka's avatar Stanislaw Gruszka Committed by Felix Fietkau
Browse files

mt76usb: remove mt76u_buf and use urb directly



Put urb pointer in mt76_queue_entry directly instead of mt76u_buf
structure.

Signed-off-by: default avatarStanislaw Gruszka <sgruszka@redhat.com>
Signed-off-by: default avatarFelix Fietkau <nbd@nbd.name>
parent 279ade99
Loading
Loading
Loading
Loading
+1 −5
Original line number Diff line number Diff line
@@ -89,10 +89,6 @@ struct mt76_tx_info {
	u32 info;
};

struct mt76u_buf {
	struct urb *urb;
};

struct mt76_queue_entry {
	union {
		void *buf;
@@ -100,7 +96,7 @@ struct mt76_queue_entry {
	};
	union {
		struct mt76_txwi_cache *txwi;
		struct mt76u_buf ubuf;
		struct urb *urb;
	};
	enum mt76_txq_id qid;
	bool schedule;
+63 −67
Original line number Diff line number Diff line
@@ -283,12 +283,11 @@ mt76u_set_endpoints(struct usb_interface *intf,
}

static int
mt76u_fill_rx_sg(struct mt76_dev *dev, struct mt76u_buf *buf, int nsgs,
mt76u_fill_rx_sg(struct mt76_dev *dev, struct urb *urb, int nsgs,
		 gfp_t gfp)
{
	struct mt76_queue *q = &dev->q_rx[MT_RXQ_MAIN];
	int sglen = SKB_WITH_OVERHEAD(q->buf_size);
	struct urb *urb = buf->urb;

	int i;

@@ -323,44 +322,43 @@ mt76u_fill_rx_sg(struct mt76_dev *dev, struct mt76u_buf *buf, int nsgs,

static int
mt76u_refill_rx(struct mt76_dev *dev, struct mt76_queue *q,
		struct mt76u_buf *buf, int nsgs, gfp_t gfp)
		struct urb *urb, int nsgs, gfp_t gfp)
{
	if (dev->usb.sg_en) {
		return mt76u_fill_rx_sg(dev, buf, nsgs, gfp);
		return mt76u_fill_rx_sg(dev, urb, nsgs, gfp);
	} else {
		buf->urb->transfer_buffer_length =
			SKB_WITH_OVERHEAD(q->buf_size);
		buf->urb->transfer_buffer =
			page_frag_alloc(&q->rx_page, q->buf_size, gfp);
		return buf->urb->transfer_buffer ? 0 : -ENOMEM;
		urb->transfer_buffer_length = SKB_WITH_OVERHEAD(q->buf_size);
		urb->transfer_buffer = page_frag_alloc(&q->rx_page,
						       q->buf_size, gfp);
		return urb->transfer_buffer ? 0 : -ENOMEM;
	}
}

static int
mt76u_buf_alloc(struct mt76_dev *dev, struct mt76u_buf *buf)
mt76u_urb_alloc(struct mt76_dev *dev, struct mt76_queue_entry *e)
{
	struct mt76_queue *q = &dev->q_rx[MT_RXQ_MAIN];
	struct urb *urb;

	buf->urb = usb_alloc_urb(0, GFP_KERNEL);
	if (!buf->urb)
	urb = usb_alloc_urb(0, GFP_KERNEL);
	if (!urb)
		return -ENOMEM;
	e->urb = urb;

	if (dev->usb.sg_en) {
		buf->urb->sg = devm_kcalloc(dev->dev, MT_SG_MAX_SIZE,
					    sizeof(*buf->urb->sg),
					    GFP_KERNEL);
		if (!buf->urb->sg)
		urb->sg = devm_kcalloc(dev->dev, MT_SG_MAX_SIZE,
				       sizeof(*urb->sg), GFP_KERNEL);
		if (!urb->sg)
			return -ENOMEM;

		sg_init_table(buf->urb->sg, MT_SG_MAX_SIZE);
		sg_init_table(urb->sg, MT_SG_MAX_SIZE);
	}

	return mt76u_refill_rx(dev, q, buf, MT_SG_MAX_SIZE, GFP_KERNEL);
	return mt76u_refill_rx(dev, q, urb, MT_SG_MAX_SIZE, GFP_KERNEL);
}

static void mt76u_buf_free(struct mt76u_buf *buf)
static void mt76u_urb_free(struct urb *urb)
{
	struct urb *urb = buf->urb;
	int i;

	for (i = 0; i < urb->num_sgs; i++)
@@ -369,12 +367,12 @@ static void mt76u_buf_free(struct mt76u_buf *buf)
	if (urb->transfer_buffer)
		skb_free_frag(urb->transfer_buffer);

	usb_free_urb(buf->urb);
	usb_free_urb(urb);
}

static void
mt76u_fill_bulk_urb(struct mt76_dev *dev, int dir, int index,
		    struct mt76u_buf *buf, usb_complete_t complete_fn,
		    struct urb *urb, usb_complete_t complete_fn,
		    void *context)
{
	struct usb_device *udev = to_usb_device(dev->dev);
@@ -385,27 +383,27 @@ mt76u_fill_bulk_urb(struct mt76_dev *dev, int dir, int index,
	else
		pipe = usb_sndbulkpipe(udev, dev->usb.out_ep[index]);

	buf->urb->dev = udev;
	buf->urb->pipe = pipe;
	buf->urb->complete = complete_fn;
	buf->urb->context = context;
	urb->dev = udev;
	urb->pipe = pipe;
	urb->complete = complete_fn;
	urb->context = context;
}

static inline struct mt76u_buf
*mt76u_get_next_rx_entry(struct mt76_queue *q)
static inline struct urb *
mt76u_get_next_rx_entry(struct mt76_queue *q)
{
	struct mt76u_buf *buf = NULL;
	struct urb *urb = NULL;
	unsigned long flags;

	spin_lock_irqsave(&q->lock, flags);
	if (q->queued > 0) {
		buf = &q->entry[q->head].ubuf;
		urb = q->entry[q->head].urb;
		q->head = (q->head + 1) % q->ndesc;
		q->queued--;
	}
	spin_unlock_irqrestore(&q->lock, flags);

	return buf;
	return urb;
}

static int mt76u_get_rx_entry_len(u8 *data, u32 data_len)
@@ -424,10 +422,9 @@ static int mt76u_get_rx_entry_len(u8 *data, u32 data_len)
}

static int
mt76u_process_rx_entry(struct mt76_dev *dev, struct mt76u_buf *buf)
mt76u_process_rx_entry(struct mt76_dev *dev, struct urb *urb)
{
	struct mt76_queue *q = &dev->q_rx[MT_RXQ_MAIN];
	struct urb *urb = buf->urb;
	u8 *data = urb->num_sgs ? sg_virt(&urb->sg[0]) : urb->transfer_buffer;
	int data_len = urb->num_sgs ? urb->sg[0].length : urb->actual_length;
	int len, nsgs = 1;
@@ -488,7 +485,7 @@ static void mt76u_complete_rx(struct urb *urb)
	}

	spin_lock_irqsave(&q->lock, flags);
	if (WARN_ONCE(q->entry[q->tail].ubuf.urb != urb, "rx urb mismatch"))
	if (WARN_ONCE(q->entry[q->tail].urb != urb, "rx urb mismatch"))
		goto out;

	q->tail = (q->tail + 1) % q->ndesc;
@@ -499,37 +496,37 @@ out:
}

static int
mt76u_submit_rx_buf(struct mt76_dev *dev, struct mt76u_buf *buf)
mt76u_submit_rx_buf(struct mt76_dev *dev, struct urb *urb)
{
	mt76u_fill_bulk_urb(dev, USB_DIR_IN, MT_EP_IN_PKT_RX, buf,
	mt76u_fill_bulk_urb(dev, USB_DIR_IN, MT_EP_IN_PKT_RX, urb,
			    mt76u_complete_rx, dev);
	trace_submit_urb(dev, buf->urb);
	trace_submit_urb(dev, urb);

	return usb_submit_urb(buf->urb, GFP_ATOMIC);
	return usb_submit_urb(urb, GFP_ATOMIC);
}

static void mt76u_rx_tasklet(unsigned long data)
{
	struct mt76_dev *dev = (struct mt76_dev *)data;
	struct mt76_queue *q = &dev->q_rx[MT_RXQ_MAIN];
	struct mt76u_buf *buf;
	struct urb *urb;
	int err, count;

	rcu_read_lock();

	while (true) {
		buf = mt76u_get_next_rx_entry(q);
		if (!buf)
		urb = mt76u_get_next_rx_entry(q);
		if (!urb)
			break;

		count = mt76u_process_rx_entry(dev, buf);
		count = mt76u_process_rx_entry(dev, urb);
		if (count > 0) {
			err = mt76u_refill_rx(dev, q, buf, count,
			err = mt76u_refill_rx(dev, q, urb, count,
					      GFP_ATOMIC);
			if (err < 0)
				break;
		}
		mt76u_submit_rx_buf(dev, buf);
		mt76u_submit_rx_buf(dev, urb);
	}
	mt76_rx_poll_complete(dev, MT_RXQ_MAIN, NULL);

@@ -544,7 +541,7 @@ int mt76u_submit_rx_buffers(struct mt76_dev *dev)

	spin_lock_irqsave(&q->lock, flags);
	for (i = 0; i < q->ndesc; i++) {
		err = mt76u_submit_rx_buf(dev, &q->entry[i].ubuf);
		err = mt76u_submit_rx_buf(dev, q->entry[i].urb);
		if (err < 0)
			break;
	}
@@ -576,7 +573,7 @@ static int mt76u_alloc_rx(struct mt76_dev *dev)
	q->buf_size = dev->usb.sg_en ? MT_RX_BUF_SIZE : PAGE_SIZE;
	q->ndesc = MT_NUM_RX_ENTRIES;
	for (i = 0; i < q->ndesc; i++) {
		err = mt76u_buf_alloc(dev, &q->entry[i].ubuf);
		err = mt76u_urb_alloc(dev, &q->entry[i]);
		if (err < 0)
			return err;
	}
@@ -591,7 +588,7 @@ static void mt76u_free_rx(struct mt76_dev *dev)
	int i;

	for (i = 0; i < q->ndesc; i++)
		mt76u_buf_free(&q->entry[i].ubuf);
		mt76u_urb_free(q->entry[i].urb);

	if (!q->rx_page.va)
		return;
@@ -607,7 +604,7 @@ static void mt76u_stop_rx(struct mt76_dev *dev)
	int i;

	for (i = 0; i < q->ndesc; i++)
		usb_kill_urb(q->entry[i].ubuf.urb);
		usb_kill_urb(q->entry[i].urb);
}

static void mt76u_tx_tasklet(unsigned long data)
@@ -724,7 +721,7 @@ mt76u_tx_queue_skb(struct mt76_dev *dev, enum mt76_txq_id qid,
		   struct ieee80211_sta *sta)
{
	struct mt76_queue *q = dev->q_tx[qid].q;
	struct mt76u_buf *buf;
	struct urb *urb;
	u16 idx = q->tail;
	int err;

@@ -737,13 +734,13 @@ mt76u_tx_queue_skb(struct mt76_dev *dev, enum mt76_txq_id qid,
		return err;

	q->entry[idx].done = false;
	buf = &q->entry[idx].ubuf;
	err = mt76u_tx_setup_buffers(dev, skb, buf->urb);
	urb = q->entry[idx].urb;
	err = mt76u_tx_setup_buffers(dev, skb, urb);
	if (err < 0)
		return err;

	mt76u_fill_bulk_urb(dev, USB_DIR_OUT, q2ep(q->hw_idx),
			    buf, mt76u_complete_tx, &q->entry[idx]);
			    urb, mt76u_complete_tx, &q->entry[idx]);

	q->tail = (q->tail + 1) % q->ndesc;
	q->entry[idx].skb = skb;
@@ -754,14 +751,14 @@ mt76u_tx_queue_skb(struct mt76_dev *dev, enum mt76_txq_id qid,

static void mt76u_tx_kick(struct mt76_dev *dev, struct mt76_queue *q)
{
	struct mt76u_buf *buf;
	struct urb *urb;
	int err;

	while (q->first != q->tail) {
		buf = &q->entry[q->first].ubuf;
		urb = q->entry[q->first].urb;

		trace_submit_urb(dev, buf->urb);
		err = usb_submit_urb(buf->urb, GFP_ATOMIC);
		trace_submit_urb(dev, urb);
		err = usb_submit_urb(urb, GFP_ATOMIC);
		if (err < 0) {
			if (err == -ENODEV)
				set_bit(MT76_REMOVED, &dev->state);
@@ -776,7 +773,7 @@ static void mt76u_tx_kick(struct mt76_dev *dev, struct mt76_queue *q)

static int mt76u_alloc_tx(struct mt76_dev *dev)
{
	struct mt76u_buf *buf;
	struct urb *urb;
	struct mt76_queue *q;
	int i, j;

@@ -804,19 +801,18 @@ static int mt76u_alloc_tx(struct mt76_dev *dev)

		q->ndesc = MT_NUM_TX_ENTRIES;
		for (j = 0; j < q->ndesc; j++) {
			buf = &q->entry[j].ubuf;

			buf->urb = usb_alloc_urb(0, GFP_KERNEL);
			if (!buf->urb)
			urb = usb_alloc_urb(0, GFP_KERNEL);
			if (!urb)
				return -ENOMEM;
			q->entry[j].urb = urb;

			if (!dev->usb.sg_en)
				continue;

			buf->urb->sg = devm_kcalloc(dev->dev, MT_SG_MAX_SIZE,
			urb->sg = devm_kcalloc(dev->dev, MT_SG_MAX_SIZE,
					       sizeof(struct scatterlist),
					       GFP_KERNEL);
			if (!buf->urb->sg)
			if (!urb->sg)
				return -ENOMEM;
		}
	}
@@ -831,7 +827,7 @@ static void mt76u_free_tx(struct mt76_dev *dev)
	for (i = 0; i < IEEE80211_NUM_ACS; i++) {
		q = dev->q_tx[i].q;
		for (j = 0; j < q->ndesc; j++)
			usb_free_urb(q->entry[j].ubuf.urb);
			usb_free_urb(q->entry[j].urb);
	}
}

@@ -843,7 +839,7 @@ static void mt76u_stop_tx(struct mt76_dev *dev)
	for (i = 0; i < IEEE80211_NUM_ACS; i++) {
		q = dev->q_tx[i].q;
		for (j = 0; j < q->ndesc; j++)
			usb_kill_urb(q->entry[j].ubuf.urb);
			usb_kill_urb(q->entry[j].urb);
	}
}