Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 7262d593 authored by Christian Lamparter's avatar Christian Lamparter Committed by John W. Linville
Browse files

p54pci: rx tasklet refactoring



This patch moves the all of p54pci's receiver code out of the
bloated interrupt handler routine and into a less critical tasklet.

Signed-off-by: default avatarChristian Lamparter <chunkeey@web.de>
Signed-off-by: default avatarJohn W. Linville <linville@tuxdriver.com>
parent 84df3ed3
Loading
Loading
Loading
Loading
+161 −81
Original line number Original line Diff line number Diff line
@@ -3,6 +3,7 @@
 * Linux device driver for PCI based Prism54
 * Linux device driver for PCI based Prism54
 *
 *
 * Copyright (c) 2006, Michael Wu <flamingice@sourmilk.net>
 * Copyright (c) 2006, Michael Wu <flamingice@sourmilk.net>
 * Copyright (c) 2008, Christian Lamparter <chunkeey@web.de>
 *
 *
 * Based on the islsm (softmac prism54) driver, which is:
 * Based on the islsm (softmac prism54) driver, which is:
 * Copyright 2004-2006 Jean-Baptiste Note <jean-baptiste.note@m4x.org>, et al.
 * Copyright 2004-2006 Jean-Baptiste Note <jean-baptiste.note@m4x.org>, et al.
@@ -237,20 +238,22 @@ static int p54p_read_eeprom(struct ieee80211_hw *dev)
	return err;
	return err;
}
}


static void p54p_refill_rx_ring(struct ieee80211_hw *dev)
static void p54p_refill_rx_ring(struct ieee80211_hw *dev,
	int ring_index, struct p54p_desc *ring, u32 ring_limit,
	struct sk_buff **rx_buf)
{
{
	struct p54p_priv *priv = dev->priv;
	struct p54p_priv *priv = dev->priv;
	struct p54p_ring_control *ring_control = priv->ring_control;
	struct p54p_ring_control *ring_control = priv->ring_control;
	u32 limit, host_idx, idx;
	u32 limit, idx, i;


	host_idx = le32_to_cpu(ring_control->host_idx[0]);
	idx = le32_to_cpu(ring_control->host_idx[ring_index]);
	limit = host_idx;
	limit = idx;
	limit -= le32_to_cpu(ring_control->device_idx[0]);
	limit -= le32_to_cpu(ring_control->device_idx[ring_index]);
	limit = ARRAY_SIZE(ring_control->rx_data) - limit;
	limit = ring_limit - limit;


	idx = host_idx % ARRAY_SIZE(ring_control->rx_data);
	i = idx % ring_limit;
	while (limit-- > 1) {
	while (limit-- > 1) {
		struct p54p_desc *desc = &ring_control->rx_data[idx];
		struct p54p_desc *desc = &ring[i];


		if (!desc->host_addr) {
		if (!desc->host_addr) {
			struct sk_buff *skb;
			struct sk_buff *skb;
@@ -267,50 +270,78 @@ static void p54p_refill_rx_ring(struct ieee80211_hw *dev)
			desc->device_addr = 0;	// FIXME: necessary?
			desc->device_addr = 0;	// FIXME: necessary?
			desc->len = cpu_to_le16(MAX_RX_SIZE);
			desc->len = cpu_to_le16(MAX_RX_SIZE);
			desc->flags = 0;
			desc->flags = 0;
			priv->rx_buf[idx] = skb;
			rx_buf[i] = skb;
		}
		}


		i++;
		idx++;
		idx++;
		host_idx++;
		i %= ring_limit;
		idx %= ARRAY_SIZE(ring_control->rx_data);
	}
	}


	wmb();
	wmb();
	ring_control->host_idx[0] = cpu_to_le32(host_idx);
	ring_control->host_idx[ring_index] = cpu_to_le32(idx);
}
}


static irqreturn_t p54p_interrupt(int irq, void *dev_id)
static void p54p_check_rx_ring(struct ieee80211_hw *dev, u32 *index,
	int ring_index, struct p54p_desc *ring, u32 ring_limit,
	struct sk_buff **rx_buf)
{
{
	struct ieee80211_hw *dev = dev_id;
	struct p54p_priv *priv = dev->priv;
	struct p54p_priv *priv = dev->priv;
	struct p54p_ring_control *ring_control = priv->ring_control;
	struct p54p_ring_control *ring_control = priv->ring_control;
	__le32 reg;
	struct p54p_desc *desc;
	u32 idx, i;


	spin_lock(&priv->lock);
	i = (*index) % ring_limit;
	reg = P54P_READ(int_ident);
	(*index) = idx = le32_to_cpu(ring_control->device_idx[ring_index]);
	if (unlikely(reg == cpu_to_le32(0xFFFFFFFF))) {
	idx %= ring_limit;
		spin_unlock(&priv->lock);
	while (i != idx) {
		return IRQ_HANDLED;
		u16 len;
		struct sk_buff *skb;
		desc = &ring[i];
		len = le16_to_cpu(desc->len);
		skb = rx_buf[i];

		if (!skb)
			continue;

		skb_put(skb, len);

		if (p54_rx(dev, skb)) {
			pci_unmap_single(priv->pdev,
					 le32_to_cpu(desc->host_addr),
					 MAX_RX_SIZE, PCI_DMA_FROMDEVICE);
			rx_buf[i] = NULL;
			desc->host_addr = 0;
		} else {
			skb_trim(skb, 0);
			desc->len = cpu_to_le16(MAX_RX_SIZE);
		}
		}


	P54P_WRITE(int_ack, reg);
		i++;
		i %= ring_limit;
	}


	reg &= P54P_READ(int_enable);
	p54p_refill_rx_ring(dev, ring_index, ring, ring_limit, rx_buf);
}


	if (reg & cpu_to_le32(ISL38XX_INT_IDENT_UPDATE)) {
/* caller must hold priv->lock */
static void p54p_check_tx_ring(struct ieee80211_hw *dev, u32 *index,
	int ring_index, struct p54p_desc *ring, u32 ring_limit,
	void **tx_buf)
{
	struct p54p_priv *priv = dev->priv;
	struct p54p_ring_control *ring_control = priv->ring_control;
	struct p54p_desc *desc;
	struct p54p_desc *desc;
	u32 idx, i;
	u32 idx, i;
		i = priv->tx_idx;

		i %= ARRAY_SIZE(ring_control->tx_data);
	i = (*index) % ring_limit;
		priv->tx_idx = idx = le32_to_cpu(ring_control->device_idx[1]);
	(*index) = idx = le32_to_cpu(ring_control->device_idx[1]);
		idx %= ARRAY_SIZE(ring_control->tx_data);
	idx %= ring_limit;


	while (i != idx) {
	while (i != idx) {
			desc = &ring_control->tx_data[i];
		desc = &ring[i];
			if (priv->tx_buf[i]) {
		kfree(tx_buf[i]);
				kfree(priv->tx_buf[i]);
		tx_buf[i] = NULL;
				priv->tx_buf[i] = NULL;
			}


		pci_unmap_single(priv->pdev, le32_to_cpu(desc->host_addr),
		pci_unmap_single(priv->pdev, le32_to_cpu(desc->host_addr),
				 le16_to_cpu(desc->len), PCI_DMA_TODEVICE);
				 le16_to_cpu(desc->len), PCI_DMA_TODEVICE);
@@ -321,42 +352,57 @@ static irqreturn_t p54p_interrupt(int irq, void *dev_id)
		desc->flags = 0;
		desc->flags = 0;


		i++;
		i++;
			i %= ARRAY_SIZE(ring_control->tx_data);
		i %= ring_limit;
	}
}
}


		i = priv->rx_idx;
static void p54p_rx_tasklet(unsigned long dev_id)
		i %= ARRAY_SIZE(ring_control->rx_data);
{
		priv->rx_idx = idx = le32_to_cpu(ring_control->device_idx[0]);
	struct ieee80211_hw *dev = (struct ieee80211_hw *)dev_id;
		idx %= ARRAY_SIZE(ring_control->rx_data);
	struct p54p_priv *priv = dev->priv;
		while (i != idx) {
	struct p54p_ring_control *ring_control = priv->ring_control;
			u16 len;
			struct sk_buff *skb;
			desc = &ring_control->rx_data[i];
			len = le16_to_cpu(desc->len);
			skb = priv->rx_buf[i];


			skb_put(skb, len);
	p54p_check_rx_ring(dev, &priv->rx_idx_mgmt, 2, ring_control->rx_mgmt,
		ARRAY_SIZE(ring_control->rx_mgmt), priv->rx_buf_mgmt);


			if (p54_rx(dev, skb)) {
	p54p_check_rx_ring(dev, &priv->rx_idx_data, 0, ring_control->rx_data,
				pci_unmap_single(priv->pdev,
		ARRAY_SIZE(ring_control->rx_data), priv->rx_buf_data);
						 le32_to_cpu(desc->host_addr),
						 MAX_RX_SIZE, PCI_DMA_FROMDEVICE);


				priv->rx_buf[i] = NULL;
	wmb();
				desc->host_addr = 0;
	P54P_WRITE(dev_int, cpu_to_le32(ISL38XX_DEV_INT_UPDATE));
			} else {
				skb_trim(skb, 0);
				desc->len = cpu_to_le16(MAX_RX_SIZE);
}
}


			i++;
static irqreturn_t p54p_interrupt(int irq, void *dev_id)
			i %= ARRAY_SIZE(ring_control->rx_data);
{
	struct ieee80211_hw *dev = dev_id;
	struct p54p_priv *priv = dev->priv;
	struct p54p_ring_control *ring_control = priv->ring_control;
	__le32 reg;

	spin_lock(&priv->lock);
	reg = P54P_READ(int_ident);
	if (unlikely(reg == cpu_to_le32(0xFFFFFFFF))) {
		spin_unlock(&priv->lock);
		return IRQ_HANDLED;
	}
	}


		p54p_refill_rx_ring(dev);
	P54P_WRITE(int_ack, reg);

	reg &= P54P_READ(int_enable);

	if (reg & cpu_to_le32(ISL38XX_INT_IDENT_UPDATE)) {
		p54p_check_tx_ring(dev, &priv->tx_idx_mgmt,
				   3, ring_control->tx_mgmt,
				   ARRAY_SIZE(ring_control->tx_mgmt),
				   priv->tx_buf_mgmt);

		p54p_check_tx_ring(dev, &priv->tx_idx_data,
				   1, ring_control->tx_data,
				   ARRAY_SIZE(ring_control->tx_data),
				   priv->tx_buf_data);

		tasklet_schedule(&priv->rx_tasklet);


		wmb();
		P54P_WRITE(dev_int, cpu_to_le32(ISL38XX_DEV_INT_UPDATE));
	} else if (reg & cpu_to_le32(ISL38XX_INT_IDENT_INIT))
	} else if (reg & cpu_to_le32(ISL38XX_INT_IDENT_INIT))
		complete(&priv->boot_comp);
		complete(&priv->boot_comp);


@@ -392,7 +438,7 @@ static void p54p_tx(struct ieee80211_hw *dev, struct p54_control_hdr *data,
	ring_control->host_idx[1] = cpu_to_le32(idx + 1);
	ring_control->host_idx[1] = cpu_to_le32(idx + 1);


	if (free_on_tx)
	if (free_on_tx)
		priv->tx_buf[i] = data;
		priv->tx_buf_data[i] = data;


	spin_unlock_irqrestore(&priv->lock, flags);
	spin_unlock_irqrestore(&priv->lock, flags);


@@ -420,8 +466,14 @@ static int p54p_open(struct ieee80211_hw *dev)
	}
	}


	memset(priv->ring_control, 0, sizeof(*priv->ring_control));
	memset(priv->ring_control, 0, sizeof(*priv->ring_control));
	priv->rx_idx = priv->tx_idx = 0;
	priv->rx_idx_data = priv->tx_idx_data = 0;
	p54p_refill_rx_ring(dev);
	priv->rx_idx_mgmt = priv->tx_idx_mgmt = 0;

	p54p_refill_rx_ring(dev, 0, priv->ring_control->rx_data,
		ARRAY_SIZE(priv->ring_control->rx_data), priv->rx_buf_data);

	p54p_refill_rx_ring(dev, 2, priv->ring_control->rx_mgmt,
		ARRAY_SIZE(priv->ring_control->rx_mgmt), priv->rx_buf_mgmt);


	p54p_upload_firmware(dev);
	p54p_upload_firmware(dev);


@@ -465,6 +517,8 @@ static void p54p_stop(struct ieee80211_hw *dev)
	unsigned int i;
	unsigned int i;
	struct p54p_desc *desc;
	struct p54p_desc *desc;


	tasklet_kill(&priv->rx_tasklet);

	P54P_WRITE(int_enable, cpu_to_le32(0));
	P54P_WRITE(int_enable, cpu_to_le32(0));
	P54P_READ(int_enable);
	P54P_READ(int_enable);
	udelay(10);
	udelay(10);
@@ -473,26 +527,51 @@ static void p54p_stop(struct ieee80211_hw *dev)


	P54P_WRITE(dev_int, cpu_to_le32(ISL38XX_DEV_INT_RESET));
	P54P_WRITE(dev_int, cpu_to_le32(ISL38XX_DEV_INT_RESET));


	for (i = 0; i < ARRAY_SIZE(priv->rx_buf); i++) {
	for (i = 0; i < ARRAY_SIZE(priv->rx_buf_data); i++) {
		desc = &ring_control->rx_data[i];
		desc = &ring_control->rx_data[i];
		if (desc->host_addr)
		if (desc->host_addr)
			pci_unmap_single(priv->pdev, le32_to_cpu(desc->host_addr),
			pci_unmap_single(priv->pdev,
					 le32_to_cpu(desc->host_addr),
					 MAX_RX_SIZE, PCI_DMA_FROMDEVICE);
		kfree_skb(priv->rx_buf_data[i]);
		priv->rx_buf_data[i] = NULL;
	}

	for (i = 0; i < ARRAY_SIZE(priv->rx_buf_mgmt); i++) {
		desc = &ring_control->rx_mgmt[i];
		if (desc->host_addr)
			pci_unmap_single(priv->pdev,
					 le32_to_cpu(desc->host_addr),
					 MAX_RX_SIZE, PCI_DMA_FROMDEVICE);
					 MAX_RX_SIZE, PCI_DMA_FROMDEVICE);
		kfree_skb(priv->rx_buf[i]);
		kfree_skb(priv->rx_buf_mgmt[i]);
		priv->rx_buf[i] = NULL;
		priv->rx_buf_mgmt[i] = NULL;
	}
	}


	for (i = 0; i < ARRAY_SIZE(priv->tx_buf); i++) {
	for (i = 0; i < ARRAY_SIZE(priv->tx_buf_data); i++) {
		desc = &ring_control->tx_data[i];
		desc = &ring_control->tx_data[i];
		if (desc->host_addr)
		if (desc->host_addr)
			pci_unmap_single(priv->pdev, le32_to_cpu(desc->host_addr),
			pci_unmap_single(priv->pdev,
					 le16_to_cpu(desc->len), PCI_DMA_TODEVICE);
					 le32_to_cpu(desc->host_addr),
					 le16_to_cpu(desc->len),
					 PCI_DMA_TODEVICE);


		kfree(priv->tx_buf[i]);
		kfree(priv->tx_buf_data[i]);
		priv->tx_buf[i] = NULL;
		priv->tx_buf_data[i] = NULL;
	}
	}


	memset(ring_control, 0, sizeof(ring_control));
	for (i = 0; i < ARRAY_SIZE(priv->tx_buf_mgmt); i++) {
		desc = &ring_control->tx_mgmt[i];
		if (desc->host_addr)
			pci_unmap_single(priv->pdev,
					 le32_to_cpu(desc->host_addr),
					 le16_to_cpu(desc->len),
					 PCI_DMA_TODEVICE);

		kfree(priv->tx_buf_mgmt[i]);
		priv->tx_buf_mgmt[i] = NULL;
	}

	memset(ring_control, 0, sizeof(*ring_control));
}
}


static int __devinit p54p_probe(struct pci_dev *pdev,
static int __devinit p54p_probe(struct pci_dev *pdev,
@@ -585,6 +664,7 @@ static int __devinit p54p_probe(struct pci_dev *pdev,
	priv->common.tx = p54p_tx;
	priv->common.tx = p54p_tx;


	spin_lock_init(&priv->lock);
	spin_lock_init(&priv->lock);
	tasklet_init(&priv->rx_tasklet, p54p_rx_tasklet, (unsigned long)dev);


	err = ieee80211_register_hw(dev);
	err = ieee80211_register_hw(dev);
	if (err) {
	if (err) {
+7 −3
Original line number Original line Diff line number Diff line
@@ -92,13 +92,17 @@ struct p54p_priv {
	struct p54_common common;
	struct p54_common common;
	struct pci_dev *pdev;
	struct pci_dev *pdev;
	struct p54p_csr __iomem *map;
	struct p54p_csr __iomem *map;
	struct tasklet_struct rx_tasklet;


	spinlock_t lock;
	spinlock_t lock;
	struct p54p_ring_control *ring_control;
	struct p54p_ring_control *ring_control;
	dma_addr_t ring_control_dma;
	dma_addr_t ring_control_dma;
	u32 rx_idx, tx_idx;
	u32 rx_idx_data, tx_idx_data;
	struct sk_buff *rx_buf[8];
	u32 rx_idx_mgmt, tx_idx_mgmt;
	void *tx_buf[32];
	struct sk_buff *rx_buf_data[8];
	struct sk_buff *rx_buf_mgmt[4];
	void *tx_buf_data[32];
	void *tx_buf_mgmt[4];
	struct completion boot_comp;
	struct completion boot_comp;
};
};