Commit 1cac6ec9 authored by Linus Torvalds's avatar Linus Torvalds

Merge git://git.kernel.org/pub/scm/linux/kernel/git/davem/net-2.6

* git://git.kernel.org/pub/scm/linux/kernel/git/davem/net-2.6:
  smc91x: let smc91x work well under netpoll
  pxaficp-ir: remove incorrect net_device_ops
  NET: llc, zero sockaddr_llc struct
  drivers/net: fixed drivers that support netpoll use ndo_start_xmit()
  netpoll: warning for ndo_start_xmit returns with interrupts enabled
  net: Fix Micrel KSZ8842 Kconfig description
  netfilter: xt_quota: fix wrong return value (error case)
  ipv6: Fix commit 63d9950b (ipv6: Make v4-mapped bindings consistent with IPv4)
  E100: fix interaction with swiotlb on X86.
  pkt_sched: Convert CBQ to tasklet_hrtimer.
  pkt_sched: Convert qdisc_watchdog to tasklet_hrtimer
  rtl8187: always set MSR_LINK_ENEDCA flag with RTL8187B
  ibm_newemac: emac_close() needs to call netif_carrier_off()
  net: fix ks8851 build errors
  net: Rename MAC platform driver for w90p910 platform
  yellowfin: Fix buffer underrun after dev_alloc_skb() failure
  orinoco: correct key bounds check in orinoco_hw_get_tkip_iv
  mac80211: fix todo lock
parents ce0cfd4c 8ff499e4
...@@ -1727,12 +1727,14 @@ config KS8842 ...@@ -1727,12 +1727,14 @@ config KS8842
tristate "Micrel KSZ8842" tristate "Micrel KSZ8842"
depends on HAS_IOMEM depends on HAS_IOMEM
help help
This platform driver is for Micrel KSZ8842 chip. This platform driver is for Micrel KSZ8842 / KS8842
2-port ethernet switch chip (managed, VLAN, QoS).
config KS8851 config KS8851
tristate "Micrel KS8851 SPI" tristate "Micrel KS8851 SPI"
depends on SPI depends on SPI
select MII select MII
select CRC32
help help
SPI driver for Micrel KS8851 SPI attached network chip. SPI driver for Micrel KS8851 SPI attached network chip.
......
...@@ -1080,7 +1080,7 @@ static struct platform_driver w90p910_ether_driver = { ...@@ -1080,7 +1080,7 @@ static struct platform_driver w90p910_ether_driver = {
.probe = w90p910_ether_probe, .probe = w90p910_ether_probe,
.remove = __devexit_p(w90p910_ether_remove), .remove = __devexit_p(w90p910_ether_remove),
.driver = { .driver = {
.name = "w90p910-emc", .name = "nuc900-emc",
.owner = THIS_MODULE, .owner = THIS_MODULE,
}, },
}; };
...@@ -1101,5 +1101,5 @@ module_exit(w90p910_ether_exit); ...@@ -1101,5 +1101,5 @@ module_exit(w90p910_ether_exit);
MODULE_AUTHOR("Wan ZongShun <mcuos.com@gmail.com>"); MODULE_AUTHOR("Wan ZongShun <mcuos.com@gmail.com>");
MODULE_DESCRIPTION("w90p910 MAC driver!"); MODULE_DESCRIPTION("w90p910 MAC driver!");
MODULE_LICENSE("GPL"); MODULE_LICENSE("GPL");
MODULE_ALIAS("platform:w90p910-emc"); MODULE_ALIAS("platform:nuc900-emc");
...@@ -1899,7 +1899,7 @@ static int e100_rx_indicate(struct nic *nic, struct rx *rx, ...@@ -1899,7 +1899,7 @@ static int e100_rx_indicate(struct nic *nic, struct rx *rx,
nic->ru_running = RU_SUSPENDED; nic->ru_running = RU_SUSPENDED;
pci_dma_sync_single_for_device(nic->pdev, rx->dma_addr, pci_dma_sync_single_for_device(nic->pdev, rx->dma_addr,
sizeof(struct rfd), sizeof(struct rfd),
PCI_DMA_BIDIRECTIONAL); PCI_DMA_FROMDEVICE);
return -ENODATA; return -ENODATA;
} }
......
...@@ -309,6 +309,7 @@ static int mpc52xx_fec_start_xmit(struct sk_buff *skb, struct net_device *dev) ...@@ -309,6 +309,7 @@ static int mpc52xx_fec_start_xmit(struct sk_buff *skb, struct net_device *dev)
{ {
struct mpc52xx_fec_priv *priv = netdev_priv(dev); struct mpc52xx_fec_priv *priv = netdev_priv(dev);
struct bcom_fec_bd *bd; struct bcom_fec_bd *bd;
unsigned long flags;
if (bcom_queue_full(priv->tx_dmatsk)) { if (bcom_queue_full(priv->tx_dmatsk)) {
if (net_ratelimit()) if (net_ratelimit())
...@@ -316,7 +317,7 @@ static int mpc52xx_fec_start_xmit(struct sk_buff *skb, struct net_device *dev) ...@@ -316,7 +317,7 @@ static int mpc52xx_fec_start_xmit(struct sk_buff *skb, struct net_device *dev)
return NETDEV_TX_BUSY; return NETDEV_TX_BUSY;
} }
spin_lock_irq(&priv->lock); spin_lock_irqsave(&priv->lock, flags);
dev->trans_start = jiffies; dev->trans_start = jiffies;
bd = (struct bcom_fec_bd *) bd = (struct bcom_fec_bd *)
...@@ -332,7 +333,7 @@ static int mpc52xx_fec_start_xmit(struct sk_buff *skb, struct net_device *dev) ...@@ -332,7 +333,7 @@ static int mpc52xx_fec_start_xmit(struct sk_buff *skb, struct net_device *dev)
netif_stop_queue(dev); netif_stop_queue(dev);
} }
spin_unlock_irq(&priv->lock); spin_unlock_irqrestore(&priv->lock, flags);
return NETDEV_TX_OK; return NETDEV_TX_OK;
} }
......
...@@ -1305,6 +1305,8 @@ static int emac_close(struct net_device *ndev) ...@@ -1305,6 +1305,8 @@ static int emac_close(struct net_device *ndev)
free_irq(dev->emac_irq, dev); free_irq(dev->emac_irq, dev);
netif_carrier_off(ndev);
return 0; return 0;
} }
......
...@@ -803,9 +803,6 @@ static const struct net_device_ops pxa_irda_netdev_ops = { ...@@ -803,9 +803,6 @@ static const struct net_device_ops pxa_irda_netdev_ops = {
.ndo_stop = pxa_irda_stop, .ndo_stop = pxa_irda_stop,
.ndo_start_xmit = pxa_irda_hard_xmit, .ndo_start_xmit = pxa_irda_hard_xmit,
.ndo_do_ioctl = pxa_irda_ioctl, .ndo_do_ioctl = pxa_irda_ioctl,
.ndo_change_mtu = eth_change_mtu,
.ndo_validate_addr = eth_validate_addr,
.ndo_set_mac_address = eth_mac_addr,
}; };
static int pxa_irda_probe(struct platform_device *pdev) static int pxa_irda_probe(struct platform_device *pdev)
...@@ -830,6 +827,7 @@ static int pxa_irda_probe(struct platform_device *pdev) ...@@ -830,6 +827,7 @@ static int pxa_irda_probe(struct platform_device *pdev)
if (!dev) if (!dev)
goto err_mem_3; goto err_mem_3;
SET_NETDEV_DEV(dev, &pdev->dev);
si = netdev_priv(dev); si = netdev_priv(dev);
si->dev = &pdev->dev; si->dev = &pdev->dev;
si->pdata = pdev->dev.platform_data; si->pdata = pdev->dev.platform_data;
......
...@@ -41,6 +41,7 @@ static int ixpdev_xmit(struct sk_buff *skb, struct net_device *dev) ...@@ -41,6 +41,7 @@ static int ixpdev_xmit(struct sk_buff *skb, struct net_device *dev)
struct ixpdev_priv *ip = netdev_priv(dev); struct ixpdev_priv *ip = netdev_priv(dev);
struct ixpdev_tx_desc *desc; struct ixpdev_tx_desc *desc;
int entry; int entry;
unsigned long flags;
if (unlikely(skb->len > PAGE_SIZE)) { if (unlikely(skb->len > PAGE_SIZE)) {
/* @@@ Count drops. */ /* @@@ Count drops. */
...@@ -63,11 +64,11 @@ static int ixpdev_xmit(struct sk_buff *skb, struct net_device *dev) ...@@ -63,11 +64,11 @@ static int ixpdev_xmit(struct sk_buff *skb, struct net_device *dev)
dev->trans_start = jiffies; dev->trans_start = jiffies;
local_irq_disable(); local_irq_save(flags);
ip->tx_queue_entries++; ip->tx_queue_entries++;
if (ip->tx_queue_entries == TX_BUF_COUNT_PER_CHAN) if (ip->tx_queue_entries == TX_BUF_COUNT_PER_CHAN)
netif_stop_queue(dev); netif_stop_queue(dev);
local_irq_enable(); local_irq_restore(flags);
return 0; return 0;
} }
......
...@@ -620,6 +620,7 @@ static int macb_start_xmit(struct sk_buff *skb, struct net_device *dev) ...@@ -620,6 +620,7 @@ static int macb_start_xmit(struct sk_buff *skb, struct net_device *dev)
dma_addr_t mapping; dma_addr_t mapping;
unsigned int len, entry; unsigned int len, entry;
u32 ctrl; u32 ctrl;
unsigned long flags;
#ifdef DEBUG #ifdef DEBUG
int i; int i;
...@@ -635,12 +636,12 @@ static int macb_start_xmit(struct sk_buff *skb, struct net_device *dev) ...@@ -635,12 +636,12 @@ static int macb_start_xmit(struct sk_buff *skb, struct net_device *dev)
#endif #endif
len = skb->len; len = skb->len;
spin_lock_irq(&bp->lock); spin_lock_irqsave(&bp->lock, flags);
/* This is a hard error, log it. */ /* This is a hard error, log it. */
if (TX_BUFFS_AVAIL(bp) < 1) { if (TX_BUFFS_AVAIL(bp) < 1) {
netif_stop_queue(dev); netif_stop_queue(dev);
spin_unlock_irq(&bp->lock); spin_unlock_irqrestore(&bp->lock, flags);
dev_err(&bp->pdev->dev, dev_err(&bp->pdev->dev,
"BUG! Tx Ring full when queue awake!\n"); "BUG! Tx Ring full when queue awake!\n");
dev_dbg(&bp->pdev->dev, "tx_head = %u, tx_tail = %u\n", dev_dbg(&bp->pdev->dev, "tx_head = %u, tx_tail = %u\n",
...@@ -674,7 +675,7 @@ static int macb_start_xmit(struct sk_buff *skb, struct net_device *dev) ...@@ -674,7 +675,7 @@ static int macb_start_xmit(struct sk_buff *skb, struct net_device *dev)
if (TX_BUFFS_AVAIL(bp) < 1) if (TX_BUFFS_AVAIL(bp) < 1)
netif_stop_queue(dev); netif_stop_queue(dev);
spin_unlock_irq(&bp->lock); spin_unlock_irqrestore(&bp->lock, flags);
dev->trans_start = jiffies; dev->trans_start = jiffies;
......
...@@ -437,6 +437,7 @@ static inline void mlx4_en_xmit_poll(struct mlx4_en_priv *priv, int tx_ind) ...@@ -437,6 +437,7 @@ static inline void mlx4_en_xmit_poll(struct mlx4_en_priv *priv, int tx_ind)
{ {
struct mlx4_en_cq *cq = &priv->tx_cq[tx_ind]; struct mlx4_en_cq *cq = &priv->tx_cq[tx_ind];
struct mlx4_en_tx_ring *ring = &priv->tx_ring[tx_ind]; struct mlx4_en_tx_ring *ring = &priv->tx_ring[tx_ind];
unsigned long flags;
/* If we don't have a pending timer, set one up to catch our recent /* If we don't have a pending timer, set one up to catch our recent
post in case the interface becomes idle */ post in case the interface becomes idle */
...@@ -445,9 +446,9 @@ static inline void mlx4_en_xmit_poll(struct mlx4_en_priv *priv, int tx_ind) ...@@ -445,9 +446,9 @@ static inline void mlx4_en_xmit_poll(struct mlx4_en_priv *priv, int tx_ind)
/* Poll the CQ every mlx4_en_TX_MODER_POLL packets */ /* Poll the CQ every mlx4_en_TX_MODER_POLL packets */
if ((++ring->poll_cnt & (MLX4_EN_TX_POLL_MODER - 1)) == 0) if ((++ring->poll_cnt & (MLX4_EN_TX_POLL_MODER - 1)) == 0)
if (spin_trylock_irq(&ring->comp_lock)) { if (spin_trylock_irqsave(&ring->comp_lock, flags)) {
mlx4_en_process_tx_cq(priv->dev, cq); mlx4_en_process_tx_cq(priv->dev, cq);
spin_unlock_irq(&ring->comp_lock); spin_unlock_irqrestore(&ring->comp_lock, flags);
} }
} }
......
...@@ -196,21 +196,23 @@ static void PRINT_PKT(u_char *buf, int length) ...@@ -196,21 +196,23 @@ static void PRINT_PKT(u_char *buf, int length)
/* this enables an interrupt in the interrupt mask register */ /* this enables an interrupt in the interrupt mask register */
#define SMC_ENABLE_INT(lp, x) do { \ #define SMC_ENABLE_INT(lp, x) do { \
unsigned char mask; \ unsigned char mask; \
spin_lock_irq(&lp->lock); \ unsigned long smc_enable_flags; \
spin_lock_irqsave(&lp->lock, smc_enable_flags); \
mask = SMC_GET_INT_MASK(lp); \ mask = SMC_GET_INT_MASK(lp); \
mask |= (x); \ mask |= (x); \
SMC_SET_INT_MASK(lp, mask); \ SMC_SET_INT_MASK(lp, mask); \
spin_unlock_irq(&lp->lock); \ spin_unlock_irqrestore(&lp->lock, smc_enable_flags); \
} while (0) } while (0)
/* this disables an interrupt from the interrupt mask register */ /* this disables an interrupt from the interrupt mask register */
#define SMC_DISABLE_INT(lp, x) do { \ #define SMC_DISABLE_INT(lp, x) do { \
unsigned char mask; \ unsigned char mask; \
spin_lock_irq(&lp->lock); \ unsigned long smc_disable_flags; \
spin_lock_irqsave(&lp->lock, smc_disable_flags); \
mask = SMC_GET_INT_MASK(lp); \ mask = SMC_GET_INT_MASK(lp); \
mask &= ~(x); \ mask &= ~(x); \
SMC_SET_INT_MASK(lp, mask); \ SMC_SET_INT_MASK(lp, mask); \
spin_unlock_irq(&lp->lock); \ spin_unlock_irqrestore(&lp->lock, smc_disable_flags); \
} while (0) } while (0)
/* /*
...@@ -520,21 +522,21 @@ static inline void smc_rcv(struct net_device *dev) ...@@ -520,21 +522,21 @@ static inline void smc_rcv(struct net_device *dev)
* any other concurrent access and C would always interrupt B. But life * any other concurrent access and C would always interrupt B. But life
* isn't that easy in a SMP world... * isn't that easy in a SMP world...
*/ */
#define smc_special_trylock(lock) \ #define smc_special_trylock(lock, flags) \
({ \ ({ \
int __ret; \ int __ret; \
local_irq_disable(); \ local_irq_save(flags); \
__ret = spin_trylock(lock); \ __ret = spin_trylock(lock); \
if (!__ret) \ if (!__ret) \
local_irq_enable(); \ local_irq_restore(flags); \
__ret; \ __ret; \
}) })
#define smc_special_lock(lock) spin_lock_irq(lock) #define smc_special_lock(lock, flags) spin_lock_irq(lock, flags)
#define smc_special_unlock(lock) spin_unlock_irq(lock) #define smc_special_unlock(lock, flags) spin_unlock_irqrestore(lock, flags)
#else #else
#define smc_special_trylock(lock) (1) #define smc_special_trylock(lock, flags) (1)
#define smc_special_lock(lock) do { } while (0) #define smc_special_lock(lock, flags) do { } while (0)
#define smc_special_unlock(lock) do { } while (0) #define smc_special_unlock(lock, flags) do { } while (0)
#endif #endif
/* /*
...@@ -548,10 +550,11 @@ static void smc_hardware_send_pkt(unsigned long data) ...@@ -548,10 +550,11 @@ static void smc_hardware_send_pkt(unsigned long data)
struct sk_buff *skb; struct sk_buff *skb;
unsigned int packet_no, len; unsigned int packet_no, len;
unsigned char *buf; unsigned char *buf;
unsigned long flags;
DBG(3, "%s: %s\n", dev->name, __func__); DBG(3, "%s: %s\n", dev->name, __func__);
if (!smc_special_trylock(&lp->lock)) { if (!smc_special_trylock(&lp->lock, flags)) {
netif_stop_queue(dev); netif_stop_queue(dev);
tasklet_schedule(&lp->tx_task); tasklet_schedule(&lp->tx_task);
return; return;
...@@ -559,7 +562,7 @@ static void smc_hardware_send_pkt(unsigned long data) ...@@ -559,7 +562,7 @@ static void smc_hardware_send_pkt(unsigned long data)
skb = lp->pending_tx_skb; skb = lp->pending_tx_skb;
if (unlikely(!skb)) { if (unlikely(!skb)) {
smc_special_unlock(&lp->lock); smc_special_unlock(&lp->lock, flags);
return; return;
} }
lp->pending_tx_skb = NULL; lp->pending_tx_skb = NULL;
...@@ -569,7 +572,7 @@ static void smc_hardware_send_pkt(unsigned long data) ...@@ -569,7 +572,7 @@ static void smc_hardware_send_pkt(unsigned long data)
printk("%s: Memory allocation failed.\n", dev->name); printk("%s: Memory allocation failed.\n", dev->name);
dev->stats.tx_errors++; dev->stats.tx_errors++;
dev->stats.tx_fifo_errors++; dev->stats.tx_fifo_errors++;
smc_special_unlock(&lp->lock); smc_special_unlock(&lp->lock, flags);
goto done; goto done;
} }
...@@ -608,7 +611,7 @@ static void smc_hardware_send_pkt(unsigned long data) ...@@ -608,7 +611,7 @@ static void smc_hardware_send_pkt(unsigned long data)
/* queue the packet for TX */ /* queue the packet for TX */
SMC_SET_MMU_CMD(lp, MC_ENQUEUE); SMC_SET_MMU_CMD(lp, MC_ENQUEUE);
smc_special_unlock(&lp->lock); smc_special_unlock(&lp->lock, flags);
dev->trans_start = jiffies; dev->trans_start = jiffies;
dev->stats.tx_packets++; dev->stats.tx_packets++;
...@@ -633,6 +636,7 @@ static int smc_hard_start_xmit(struct sk_buff *skb, struct net_device *dev) ...@@ -633,6 +636,7 @@ static int smc_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
struct smc_local *lp = netdev_priv(dev); struct smc_local *lp = netdev_priv(dev);
void __iomem *ioaddr = lp->base; void __iomem *ioaddr = lp->base;
unsigned int numPages, poll_count, status; unsigned int numPages, poll_count, status;
unsigned long flags;
DBG(3, "%s: %s\n", dev->name, __func__); DBG(3, "%s: %s\n", dev->name, __func__);
...@@ -658,7 +662,7 @@ static int smc_hard_start_xmit(struct sk_buff *skb, struct net_device *dev) ...@@ -658,7 +662,7 @@ static int smc_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
return 0; return 0;
} }
smc_special_lock(&lp->lock); smc_special_lock(&lp->lock, flags);
/* now, try to allocate the memory */ /* now, try to allocate the memory */
SMC_SET_MMU_CMD(lp, MC_ALLOC | numPages); SMC_SET_MMU_CMD(lp, MC_ALLOC | numPages);
...@@ -676,7 +680,7 @@ static int smc_hard_start_xmit(struct sk_buff *skb, struct net_device *dev) ...@@ -676,7 +680,7 @@ static int smc_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
} }
} while (--poll_count); } while (--poll_count);
smc_special_unlock(&lp->lock); smc_special_unlock(&lp->lock, flags);
lp->pending_tx_skb = skb; lp->pending_tx_skb = skb;
if (!poll_count) { if (!poll_count) {
......
...@@ -70,7 +70,7 @@ int orinoco_hw_get_tkip_iv(struct orinoco_private *priv, int key, u8 *tsc) ...@@ -70,7 +70,7 @@ int orinoco_hw_get_tkip_iv(struct orinoco_private *priv, int key, u8 *tsc)
int err = 0; int err = 0;
u8 tsc_arr[4][IW_ENCODE_SEQ_MAX_SIZE]; u8 tsc_arr[4][IW_ENCODE_SEQ_MAX_SIZE];
if ((key < 0) || (key > 4)) if ((key < 0) || (key >= 4))
return -EINVAL; return -EINVAL;
err = hermes_read_ltv(hw, USER_BAP, HERMES_RID_CURRENT_TKIP_IV, err = hermes_read_ltv(hw, USER_BAP, HERMES_RID_CURRENT_TKIP_IV,
......
...@@ -869,6 +869,9 @@ static int rtl8187b_init_hw(struct ieee80211_hw *dev) ...@@ -869,6 +869,9 @@ static int rtl8187b_init_hw(struct ieee80211_hw *dev)
priv->aifsn[3] = 3; /* AIFSN[AC_BE] */ priv->aifsn[3] = 3; /* AIFSN[AC_BE] */
rtl818x_iowrite8(priv, &priv->map->ACM_CONTROL, 0); rtl818x_iowrite8(priv, &priv->map->ACM_CONTROL, 0);
/* ENEDCA flag must always be set, transmit issues? */
rtl818x_iowrite8(priv, &priv->map->MSR, RTL818X_MSR_ENEDCA);
return 0; return 0;
} }
...@@ -1173,13 +1176,16 @@ static void rtl8187_bss_info_changed(struct ieee80211_hw *dev, ...@@ -1173,13 +1176,16 @@ static void rtl8187_bss_info_changed(struct ieee80211_hw *dev,
rtl818x_iowrite8(priv, &priv->map->BSSID[i], rtl818x_iowrite8(priv, &priv->map->BSSID[i],
info->bssid[i]); info->bssid[i]);
if (is_valid_ether_addr(info->bssid)) {
reg = RTL818X_MSR_INFRA;
if (priv->is_rtl8187b) if (priv->is_rtl8187b)
reg |= RTL818X_MSR_ENEDCA; reg = RTL818X_MSR_ENEDCA;
else
reg = 0;
if (is_valid_ether_addr(info->bssid)) {
reg |= RTL818X_MSR_INFRA;
rtl818x_iowrite8(priv, &priv->map->MSR, reg); rtl818x_iowrite8(priv, &priv->map->MSR, reg);
} else { } else {
reg = RTL818X_MSR_NO_LINK; reg |= RTL818X_MSR_NO_LINK;
rtl818x_iowrite8(priv, &priv->map->MSR, reg); rtl818x_iowrite8(priv, &priv->map->MSR, reg);
} }
......
...@@ -346,7 +346,7 @@ static int netdev_ioctl(struct net_device *dev, struct ifreq *rq, int cmd); ...@@ -346,7 +346,7 @@ static int netdev_ioctl(struct net_device *dev, struct ifreq *rq, int cmd);
static int yellowfin_open(struct net_device *dev); static int yellowfin_open(struct net_device *dev);
static void yellowfin_timer(unsigned long data); static void yellowfin_timer(unsigned long data);
static void yellowfin_tx_timeout(struct net_device *dev); static void yellowfin_tx_timeout(struct net_device *dev);
static void yellowfin_init_ring(struct net_device *dev); static int yellowfin_init_ring(struct net_device *dev);
static int yellowfin_start_xmit(struct sk_buff *skb, struct net_device *dev); static int yellowfin_start_xmit(struct sk_buff *skb, struct net_device *dev);
static irqreturn_t yellowfin_interrupt(int irq, void *dev_instance); static irqreturn_t yellowfin_interrupt(int irq, void *dev_instance);
static int yellowfin_rx(struct net_device *dev); static int yellowfin_rx(struct net_device *dev);
...@@ -573,19 +573,24 @@ static int yellowfin_open(struct net_device *dev) ...@@ -573,19 +573,24 @@ static int yellowfin_open(struct net_device *dev)
{ {
struct yellowfin_private *yp = netdev_priv(dev); struct yellowfin_private *yp = netdev_priv(dev);
void __iomem *ioaddr = yp->base; void __iomem *ioaddr = yp->base;
int i; int i, ret;
/* Reset the chip. */ /* Reset the chip. */
iowrite32(0x80000000, ioaddr + DMACtrl); iowrite32(0x80000000, ioaddr + DMACtrl);
i = request_irq(dev->irq, &yellowfin_interrupt, IRQF_SHARED, dev->name, dev); ret = request_irq(dev->irq, &yellowfin_interrupt, IRQF_SHARED, dev->name, dev);
if (i) return i; if (ret)
return ret;
if (yellowfin_debug > 1) if (yellowfin_debug > 1)
printk(KERN_DEBUG "%s: yellowfin_open() irq %d.\n", printk(KERN_DEBUG "%s: yellowfin_open() irq %d.\n",
dev->name, dev->irq); dev->name, dev->irq);
yellowfin_init_ring(dev); ret = yellowfin_init_ring(dev);
if (ret) {
free_irq(dev->irq, dev);
return ret;
}
iowrite32(yp->rx_ring_dma, ioaddr + RxPtr); iowrite32(yp->rx_ring_dma, ioaddr + RxPtr);
iowrite32(yp->tx_ring_dma, ioaddr + TxPtr); iowrite32(yp->tx_ring_dma, ioaddr + TxPtr);
...@@ -725,10 +730,10 @@ static void yellowfin_tx_timeout(struct net_device *dev) ...@@ -725,10 +730,10 @@ static void yellowfin_tx_timeout(struct net_device *dev)
} }
/* Initialize the Rx and Tx rings, along with various 'dev' bits. */ /* Initialize the Rx and Tx rings, along with various 'dev' bits. */
static void yellowfin_init_ring(struct net_device *dev) static int yellowfin_init_ring(struct net_device *dev)
{ {
struct yellowfin_private *yp = netdev_priv(dev); struct yellowfin_private *yp = netdev_priv(dev);
int i; int i, j;
yp->tx_full = 0; yp->tx_full = 0;
yp->cur_rx = yp->cur_tx = 0; yp->cur_rx = yp->cur_tx = 0;
...@@ -753,6 +758,11 @@ static void yellowfin_init_ring(struct net_device *dev) ...@@ -753,6 +758,11 @@ static void yellowfin_init_ring(struct net_device *dev)
yp->rx_ring[i].addr = cpu_to_le32(pci_map_single(yp->pci_dev, yp->rx_ring[i].addr = cpu_to_le32(pci_map_single(yp->pci_dev,
skb->data, yp->rx_buf_sz, PCI_DMA_FROMDEVICE)); skb->data, yp->rx_buf_sz, PCI_DMA_FROMDEVICE));
} }
if (i != RX_RING_SIZE) {
for (j = 0; j < i; j++)
dev_kfree_skb(yp->rx_skbuff[j]);
return -ENOMEM;
}
yp->rx_ring[i-1].dbdma_cmd = cpu_to_le32(CMD_STOP); yp->rx_ring[i-1].dbdma_cmd = cpu_to_le32(CMD_STOP);
yp->dirty_rx = (unsigned int)(i - RX_RING_SIZE); yp->dirty_rx = (unsigned int)(i - RX_RING_SIZE);
...@@ -769,8 +779,6 @@ static void yellowfin_init_ring(struct net_device *dev) ...@@ -769,8 +779,6 @@ static void yellowfin_init_ring(struct net_device *dev)
yp->tx_ring[--i].dbdma_cmd = cpu_to_le32(CMD_STOP | BRANCH_ALWAYS); yp->tx_ring[--i].dbdma_cmd = cpu_to_le32(CMD_STOP | BRANCH_ALWAYS);
#else #else
{ {
int j;
/* Tx ring needs a pair of descriptors, the second for the status. */ /* Tx ring needs a pair of descriptors, the second for the status. */
for (i = 0; i < TX_RING_SIZE; i++) { for (i = 0; i < TX_RING_SIZE; i++) {
j = 2*i; j = 2*i;
...@@ -805,7 +813,7 @@ static void yellowfin_init_ring(struct net_device *dev) ...@@ -805,7 +813,7 @@ static void yellowfin_init_ring(struct net_device *dev)
} }
#endif #endif
yp->tx_tail_desc = &yp->tx_status[0]; yp->tx_tail_desc = &yp->tx_status[0];
return; return 0;
} }
static int yellowfin_start_xmit(struct sk_buff *skb, struct net_device *dev) static int yellowfin_start_xmit(struct sk_buff *skb, struct net_device *dev)
......
...@@ -61,7 +61,7 @@ psched_tdiff_bounded(psched_time_t tv1, psched_time_t tv2, psched_time_t bound) ...@@ -61,7 +61,7 @@ psched_tdiff_bounded(psched_time_t tv1, psched_time_t tv2, psched_time_t bound)
} }
struct qdisc_watchdog { struct qdisc_watchdog {
struct hrtimer timer; struct tasklet_hrtimer timer;
struct Qdisc *qdisc; struct Qdisc *qdisc;
}; };
......
...@@ -319,6 +319,11 @@ static void netpoll_send_skb(struct netpoll *np, struct sk_buff *skb) ...@@ -319,6 +319,11 @@ static void netpoll_send_skb(struct netpoll *np, struct sk_buff *skb)
udelay(USEC_PER_POLL); udelay(USEC_PER_POLL);
} }
WARN_ONCE(!irqs_disabled(),
"netpoll_send_skb(): %s enabled interrupts in poll (%pF)\n",
dev->name, ops->ndo_start_xmit);
local_irq_restore(flags); local_irq_restore(flags);
} }
......
...@@ -306,8 +306,10 @@ int inet6_bind(struct socket *sock, struct sockaddr *uaddr, int addr_len) ...@@ -306,8 +306,10 @@ int inet6_bind(struct socket *sock, struct sockaddr *uaddr, int addr_len)
v4addr != htonl(INADDR_ANY) && v4addr != htonl(INADDR_ANY) &&
chk_addr_ret != RTN_LOCAL && chk_addr_ret != RTN_LOCAL &&
chk_addr_ret != RTN_MULTICAST && chk_addr_ret != RTN_MULTICAST &&
chk_addr_ret != RTN_BROADCAST) chk_addr_ret != RTN_BROADCAST) {
err = -EADDRNOTAVAIL;
goto out; goto out;
}
} else { } else {
if (addr_type != IPV6_ADDR_ANY) { if (addr_type != IPV6_ADDR_ANY) {
struct net_device *dev = NULL; struct net_device *dev = NULL;
......
...@@ -914,6 +914,7 @@ static int llc_ui_getname(struct socket *sock, struct sockaddr *uaddr, ...@@ -914,6 +914,7 @@ static int llc_ui_getname(struct socket *sock, struct sockaddr *uaddr,
struct llc_sock *llc = llc_sk(sk); struct llc_sock *llc = llc_sk(sk);
int rc = 0; int rc = 0;
memset(&sllc, 0, sizeof(sllc));
lock_sock(sk); lock_sock(sk);
if (sock_flag(sk, SOCK_ZAPPED)) if (sock_flag(sk, SOCK_ZAPPED))
goto out; goto out;
......
...@@ -67,6 +67,8 @@ static DECLARE_WORK(todo_work, key_todo); ...@@ -67,6 +67,8 @@ static DECLARE_WORK(todo_work, key_todo);
* *
* @key: key to add to do item for * @key: key to add to do item for
* @flag: todo flag(s) * @flag: todo flag(s)
*
* Must be called with IRQs or softirqs disabled.
*/ */
static void add_todo(struct ieee80211_key *key, u32 flag) static void add_todo(struct ieee80211_key *key, u32 flag)
{ {
...@@ -140,9 +142,9 @@ static void ieee80211_key_enable_hw_accel(struct ieee80211_key *key) ...@@ -140,9 +142,9 @@ static void ieee80211_key_enable_hw_accel(struct ieee80211_key *key)
ret = drv_set_key(key->local, SET_KEY, &sdata->vif, sta, &key->conf); ret = drv_set_key(key->local, SET_KEY, &sdata->vif, sta, &key->conf);
if (!ret) { if (!ret) {
spin_lock(&todo_lock); spin_lock_bh(&todo_lock);
key->flags |= KEY_FLAG_UPLOADED_TO_HARDWARE; key->flags |= KEY_FLAG_UPLOADED_TO_HARDWARE;
spin_unlock(&todo_lock); spin_unlock_bh(&todo_lock);
} }
if (ret && ret != -ENOSPC && ret != -EOPNOTSUPP) if (ret && ret != -ENOSPC && ret != -EOPNOTSUPP)
...@@ -164,12 +166,12 @@ static void ieee80211_key_disable_hw_accel(struct ieee80211_key *key) ...@@ -164,12 +166,12 @@ static void ieee80211_key_disable_hw_accel(struct ieee80211_key *key)
if (!key || !key->local->ops->set_key) if (!key || !key->local->ops->set_key)
return; return;
spin_lock(&todo_lock); spin_lock_bh(&todo_lock);
if (!(key->flags & KEY_FLAG_UPLOADED_TO_HARDWARE)) { if (!(key->flags & KEY_FLAG_UPLOADED_TO_HARDWARE)) {
spin_unlock(&todo_lock); spin_unlock_bh(&todo_lock);
return; return;
} }
spin_unlock(&todo_lock); spin_unlock_bh(&todo_lock);
sta = get_sta_for_key(key); sta = get_sta_for_key(key);
sdata = key->sdata; sdata = key->sdata;
...@@ -188,9 +190,9 @@ static void ieee80211_key_disable_hw_accel(struct ieee80211_key *key) ...@@ -188,9 +190,9 @@ static void ieee80211_key_disable_hw_accel(struct ieee80211_key *key)
wiphy_name(key->local->hw.wiphy), wiphy_name(key->local->hw.wiphy),
key->conf.keyidx, sta ? sta->addr : bcast_addr, ret); key->conf.keyidx, sta ? sta->addr : bcast_addr, ret);
spin_lock(&todo_lock); spin_lock_bh(&todo_lock);
key->flags &= ~KEY_FLAG_UPLOADED_TO_HARDWARE; key->flags &= ~KEY_FLAG_UPLOADED_TO_HARDWARE;
spin_unlock(&todo_lock); spin_unlock_bh(&todo_lock);
} }
static void __ieee80211_set_default_key(struct ieee80211_sub_if_data *sdata, static void __ieee80211_set_default_key(struct ieee80211_sub_if_data *sdata,
...@@ -437,14 +439,14 @@ void ieee80211_key_link(struct ieee80211_key *key, ...@@ -437,14 +439,14 @@ void ieee80211_key_link(struct ieee80211_key *key,
__ieee80211_key_replace(sdata, sta, old_key, key); __ieee80211_key_replace(sdata, sta, old_key, key);
spin_unlock_irqrestore(&sdata->local->key_lock, flags);
/* free old key later */ /* free old key later */
add_todo(old_key, KEY_FLAG_TODO_DELETE); add_todo(old_key, KEY_FLAG_TODO_DELETE);
add_todo(key, KEY_FLAG_TODO_ADD_DEBUGFS); add_todo(key, KEY_FLAG_TODO_ADD_DEBUGFS);
if (netif_running(sdata->dev)) if (netif_running(sdata->dev))
add_todo(key, KEY_FLAG_TODO_HWACCEL_ADD); add_todo(key, KEY_FLAG_TODO_HWACCEL_ADD);
spin_unlock_irqrestore(&sdata->local->key_lock, flags);
} }
static void __ieee80211_key_free(struct ieee80211_key *key) static void __ieee80211_key_free(struct ieee80211_key *key)
...@@ -547,7 +549,7 @@ static void __ieee80211_key_todo(void) ...@@ -547,7 +549,7 @@ static void __ieee80211_key_todo(void)
*/ */
synchronize_rcu(); synchronize_rcu();
spin_lock(&todo_lock); spin_lock_bh(&todo_lock);
while (!list_empty(&todo_list)) { while (!list_empty(&todo_list)) {
key = list_first_entry(&todo_list, struct ieee80211_key, todo); key = list_first_entry(&todo_list, struct ieee80211_key, todo);
list_del_init(&key->todo); list_del_init(&key->todo);
...@@ -558,7 +560,7 @@ static void __ieee80211_key_todo(void) ...@@ -558,7 +560,7 @@ static void __ieee80211_key_todo(void)
KEY_FLAG_TODO_HWACCEL_REMOVE | KEY_FLAG_TODO_HWACCEL_REMOVE |
KEY_FLAG_TODO_DELETE); KEY_FLAG_TODO_DELETE);
key->flags &= ~todoflags; key->flags &= ~todoflags;
spin_unlock(&todo_lock); spin_unlock_bh(&todo_lock);
work_done = false; work_done = false;
...@@ -591,9 +593,9 @@ static void __ieee80211_key_todo(void) ...@@ -591,9 +593,9 @@ static void __ieee80211_key_todo(void)
WARN_ON(!work_done); WARN_ON(!work_done);
spin_lock(&todo_lock); spin_lock_bh(&todo_lock);
} }
spin_unlock(&todo_lock); spin_unlock_bh(&todo_lock);
} }
void ieee80211_key_todo(void) void ieee80211_key_todo(void)
......
...@@ -52,7 +52,7 @@ static bool quota_mt_check(const struct xt_mtchk_param *par) ...@@ -52,7 +52,7 @@ static bool quota_mt_check(const struct xt_mtchk_param *par)
q->master = kmalloc(sizeof(*q->master), GFP_KERNEL); q->master = kmalloc(sizeof(*q->master), GFP_KERNEL);
if (q->master == NULL) if (q->master == NULL)
return -ENOMEM; return false;
q->master->quota = q->quota; q->master->quota = q->quota;
return true; return true;
......
...@@ -468,8 +468,8 @@ static enum hrtimer_restart qdisc_watchdog(struct hrtimer *timer) ...@@ -468,8 +468,8 @@ static enum hrtimer_restart qdisc_watchdog(struct hrtimer *timer)
void qdisc_watchdog_init(struct qdisc_watchdog *wd, struct Qdisc *qdisc) void qdisc_watchdog_init(struct qdisc_watchdog *wd, struct Qdisc *qdisc)
{ {
hrtimer_init(&wd->timer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS); tasklet_hrtimer_init(&wd->timer, qdisc_watchdog,
wd->timer.function = qdisc_watchdog; CLOCK_MONOTONIC, HRTIMER_MODE_ABS);
wd->qdisc = qdisc; wd->qdisc = qdisc;
} }
EXPORT_SYMBOL(qdisc_watchdog_init); EXPORT_SYMBOL(qdisc_watchdog_init);
...@@ -485,13 +485,13 @@ void qdisc_watchdog_schedule(struct qdisc_watchdog *wd, psched_time_t expires) ...@@ -485,13 +485,13 @@ void qdisc_watchdog_schedule(struct qdisc_watchdog *wd, psched_time_t expires)
wd->qdisc->flags |= TCQ_F_THROTTLED; wd->qdisc->flags |= TCQ_F_THROTTLED;
time = ktime_set(0, 0); time = ktime_set(0, 0);
time = ktime_add_ns(time, PSCHED_TICKS2NS(expires)); time = ktime_add_ns(time, PSCHED_TICKS2NS(expires));
hrtimer_start(&wd->timer, time, HRTIMER_MODE_ABS); tasklet_hrtimer_start(&wd->timer, time, HRTIMER_MODE_ABS);
} }
EXPORT_SYMBOL(qdisc_watchdog_schedule); EXPORT_SYMBOL(qdisc_watchdog_schedule);
void qdisc_watchdog_cancel(struct qdisc_watchdog *wd) void qdisc_watchdog_cancel(struct qdisc_watchdog *wd)
{ {
hrtimer_cancel(&wd->timer); tasklet_hrtimer_cancel(&wd->timer);
wd->qdisc->flags &= ~TCQ_F_THROTTLED; wd->qdisc->flags &= ~TCQ_F_THROTTLED;
} }
EXPORT_SYMBOL(qdisc_watchdog_cancel); EXPORT_SYMBOL(qdisc_watchdog_cancel);
......
...@@ -163,7 +163,7 @@ struct cbq_sched_data ...@@ -163,7 +163,7 @@ struct cbq_sched_data
psched_time_t now_rt; /* Cached real time */ psched_time_t now_rt; /* Cached real time */
unsigned pmask; unsigned pmask;
struct hrtimer delay_timer; struct tasklet_hrtimer delay_timer;
struct qdisc_watchdog watchdog; /* Watchdog timer, struct qdisc_watchdog watchdog; /* Watchdog timer,
started when CBQ has started when CBQ has
backlog, but cannot backlog, but cannot
...@@ -503,6 +503,8 @@ static void cbq_ovl_delay(struct cbq_class *cl) ...@@ -503,6 +503,8 @@ static void cbq_ovl_delay(struct cbq_class *cl)
cl->undertime = q->now + delay; cl->undertime = q->now + delay;
if (delay > 0) { if (delay > 0) {
struct hrtimer *ht;
sched += delay + cl->penalty; sched += delay + cl->penalty;
cl->penalized = sched; cl->penalized = sched;
cl->cpriority = TC_CBQ_MAXPRIO; cl->cpriority = TC_CBQ_MAXPRIO;
...@@ -510,12 +512,12 @@ static void cbq_ovl_delay(struct cbq_class *cl) ...@@ -510,12 +512,12 @@ static void cbq_ovl_delay(struct cbq_class *cl)
expires = ktime_set(0, 0); expires = ktime_set(0, 0);
expires = ktime_add_ns(expires, PSCHED_TICKS2NS(sched)); expires = ktime_add_ns(expires, PSCHED_TICKS2NS(sched));
if (hrtimer_try_to_cancel(&q->delay_timer) && ht = &q->delay_timer.timer;
ktime_to_ns(ktime_sub( if (hrtimer_try_to_cancel(ht) &&
hrtimer_get_expires(&q->delay_timer), ktime_to_ns(ktime_sub(hrtimer_get_expires(ht),
expires)) > 0) expires)) > 0)
hrtimer_set_expires(&q->delay_timer, expires); hrtimer_set_expires(ht, expires);
hrtimer_restart(&q->delay_timer); hrtimer_restart(ht);
cl->delayed = 1; cl->delayed = 1;
cl->xstats.overactions++; cl->xstats.overactions++;
return; return;
...@@ -621,7 +623,7 @@ static enum hrtimer_restart cbq_undelay(struct hrtimer *timer) ...@@ -621,7 +623,7 @@ static enum hrtimer_restart cbq_undelay(struct hrtimer *timer)
time = ktime_set(0, 0); time = ktime_set(0, 0);
time = ktime_add_ns(time, PSCHED_TICKS2NS(now + delay)); time = ktime_add_ns(time, PSCHED_TICKS2NS(now + delay));
hrtimer_start(&q->delay_timer, time, HRTIMER_MODE_ABS); tasklet_hrtimer_start(&q->delay_timer, time, HRTIMER_MODE_ABS);
} }
sch->flags &= ~TCQ_F_THROTTLED; sch->flags &= ~TCQ_F_THROTTLED;
...@@ -1214,7 +1216,7 @@ cbq_reset(struct Qdisc* sch) ...@@ -1214,7 +1216,7 @@ cbq_reset(struct Qdisc* sch)
q->tx_class = NULL; q->tx_class = NULL;
q->tx_borrowed = NULL; q->tx_borrowed = NULL;
qdisc_watchdog_cancel(&q->watchdog); qdisc_watchdog_cancel(&q->watchdog);
hrtimer_cancel(&q->delay_timer); tasklet_hrtimer_cancel(&q->delay_timer);
q->toplevel = TC_CBQ_MAXLEVEL; q->toplevel = TC_CBQ_MAXLEVEL;
q->now = psched_get_time(); q->now = psched_get_time();
q->now_rt = q->now; q->now_rt = q->now;
...@@ -1397,7 +1399,8 @@ static int cbq_init(struct Qdisc *sch, struct nlattr *opt) ...@@ -1397,7 +1399,8 @@ static int cbq_init(struct Qdisc *sch, struct nlattr *opt)
q->link.minidle = -0x7FFFFFFF; q->link.minidle = -0x7FFFFFFF;
qdisc_watchdog_init(&q->watchdog, sch); qdisc_watchdog_init(&q->watchdog, sch);
hrtimer_init(&q->delay_timer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS); tasklet_hrtimer_init(&q->delay_timer, cbq_undelay,
CLOCK_MONOTONIC, HRTIMER_MODE_ABS);
q->delay_timer.function = cbq_undelay; q->delay_timer.function = cbq_undelay;
q->toplevel = TC_CBQ_MAXLEVEL; q->toplevel = TC_CBQ_MAXLEVEL;
q->now = psched_get_time(); q->now = psched_get_time();
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment