Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
Kirill Smelkov
linux
Commits
6cb553d2
Commit
6cb553d2
authored
Sep 06, 2004
by
Eric Lemoine
Committed by
Patrick McHardy
Sep 06, 2004
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
[SUNGEM]: Add NAPI support.
Signed-off-by:
David S. Miller
<
davem@davemloft.net
>
parent
f77bdc6f
Changes
2
Show whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
111 additions
and
30 deletions
+111
-30
drivers/net/sungem.c
drivers/net/sungem.c
+107
-30
drivers/net/sungem.h
drivers/net/sungem.h
+4
-0
No files found.
drivers/net/sungem.c
View file @
6cb553d2
...
@@ -181,6 +181,18 @@ static inline void phy_write(struct gem *gp, int reg, u16 val)
...
@@ -181,6 +181,18 @@ static inline void phy_write(struct gem *gp, int reg, u16 val)
__phy_write
(
gp
,
gp
->
mii_phy_addr
,
reg
,
val
);
__phy_write
(
gp
,
gp
->
mii_phy_addr
,
reg
,
val
);
}
}
static
inline
void
gem_enable_ints
(
struct
gem
*
gp
)
{
/* Enable all interrupts but TXDONE */
writel
(
GREG_STAT_TXDONE
,
gp
->
regs
+
GREG_IMASK
);
}
static
inline
void
gem_disable_ints
(
struct
gem
*
gp
)
{
/* Disable all interrupts, including TXDONE */
writel
(
GREG_STAT_NAPI
|
GREG_STAT_TXDONE
,
gp
->
regs
+
GREG_IMASK
);
}
static
void
gem_handle_mif_event
(
struct
gem
*
gp
,
u32
reg_val
,
u32
changed_bits
)
static
void
gem_handle_mif_event
(
struct
gem
*
gp
,
u32
reg_val
,
u32
changed_bits
)
{
{
if
(
netif_msg_intr
(
gp
))
if
(
netif_msg_intr
(
gp
))
...
@@ -639,7 +651,7 @@ static __inline__ void gem_tx(struct net_device *dev, struct gem *gp, u32 gem_st
...
@@ -639,7 +651,7 @@ static __inline__ void gem_tx(struct net_device *dev, struct gem *gp, u32 gem_st
}
}
gp
->
net_stats
.
tx_packets
++
;
gp
->
net_stats
.
tx_packets
++
;
dev_kfree_skb
_irq
(
skb
);
dev_kfree_skb
(
skb
);
}
}
gp
->
tx_old
=
entry
;
gp
->
tx_old
=
entry
;
...
@@ -678,12 +690,12 @@ static __inline__ void gem_post_rxds(struct gem *gp, int limit)
...
@@ -678,12 +690,12 @@ static __inline__ void gem_post_rxds(struct gem *gp, int limit)
}
}
}
}
static
void
gem_rx
(
struct
gem
*
gp
)
static
int
gem_rx
(
struct
gem
*
gp
,
int
work_to_do
)
{
{
int
entry
,
drops
;
int
entry
,
drops
,
work_done
=
0
;
u32
done
;
u32
done
;
if
(
netif_msg_
intr
(
gp
))
if
(
netif_msg_
rx_status
(
gp
))
printk
(
KERN_DEBUG
"%s: rx interrupt, done: %d, rx_new: %d
\n
"
,
printk
(
KERN_DEBUG
"%s: rx interrupt, done: %d, rx_new: %d
\n
"
,
gp
->
dev
->
name
,
readl
(
gp
->
regs
+
RXDMA_DONE
),
gp
->
rx_new
);
gp
->
dev
->
name
,
readl
(
gp
->
regs
+
RXDMA_DONE
),
gp
->
rx_new
);
...
@@ -700,6 +712,9 @@ static void gem_rx(struct gem *gp)
...
@@ -700,6 +712,9 @@ static void gem_rx(struct gem *gp)
if
((
status
&
RXDCTRL_OWN
)
!=
0
)
if
((
status
&
RXDCTRL_OWN
)
!=
0
)
break
;
break
;
if
(
work_done
>=
RX_RING_SIZE
||
work_done
>=
work_to_do
)
break
;
/* When writing back RX descriptor, GEM writes status
/* When writing back RX descriptor, GEM writes status
* then buffer address, possibly in seperate transactions.
* then buffer address, possibly in seperate transactions.
* If we don't wait for the chip to write both, we could
* If we don't wait for the chip to write both, we could
...
@@ -713,6 +728,9 @@ static void gem_rx(struct gem *gp)
...
@@ -713,6 +728,9 @@ static void gem_rx(struct gem *gp)
break
;
break
;
}
}
/* We can now account for the work we're about to do */
work_done
++
;
skb
=
gp
->
rx_skbs
[
entry
];
skb
=
gp
->
rx_skbs
[
entry
];
len
=
(
status
&
RXDCTRL_BUFSZ
)
>>
16
;
len
=
(
status
&
RXDCTRL_BUFSZ
)
>>
16
;
...
@@ -775,7 +793,8 @@ static void gem_rx(struct gem *gp)
...
@@ -775,7 +793,8 @@ static void gem_rx(struct gem *gp)
skb
->
csum
=
ntohs
((
status
&
RXDCTRL_TCPCSUM
)
^
0xffff
);
skb
->
csum
=
ntohs
((
status
&
RXDCTRL_TCPCSUM
)
^
0xffff
);
skb
->
ip_summed
=
CHECKSUM_HW
;
skb
->
ip_summed
=
CHECKSUM_HW
;
skb
->
protocol
=
eth_type_trans
(
skb
,
gp
->
dev
);
skb
->
protocol
=
eth_type_trans
(
skb
,
gp
->
dev
);
netif_rx
(
skb
);
netif_receive_skb
(
skb
);
gp
->
net_stats
.
rx_packets
++
;
gp
->
net_stats
.
rx_packets
++
;
gp
->
net_stats
.
rx_bytes
+=
len
;
gp
->
net_stats
.
rx_bytes
+=
len
;
...
@@ -792,32 +811,88 @@ static void gem_rx(struct gem *gp)
...
@@ -792,32 +811,88 @@ static void gem_rx(struct gem *gp)
if
(
drops
)
if
(
drops
)
printk
(
KERN_INFO
"%s: Memory squeeze, deferring packet.
\n
"
,
printk
(
KERN_INFO
"%s: Memory squeeze, deferring packet.
\n
"
,
gp
->
dev
->
name
);
gp
->
dev
->
name
);
return
work_done
;
}
static
int
gem_poll
(
struct
net_device
*
dev
,
int
*
budget
)
{
struct
gem
*
gp
=
dev
->
priv
;
unsigned
long
flags
;
spin_lock_irqsave
(
&
gp
->
lock
,
flags
);
do
{
int
work_to_do
,
work_done
;
/* Handle anomalies */
if
(
gp
->
status
&
GREG_STAT_ABNORMAL
)
{
if
(
gem_abnormal_irq
(
dev
,
gp
,
gp
->
status
))
break
;
}
/* Run TX completion thread */
gem_tx
(
dev
,
gp
,
gp
->
status
);
spin_unlock_irqrestore
(
&
gp
->
lock
,
flags
);
/* Run RX thread. We don't use any locking here,
* code willing to do bad things - like cleaning the
* rx ring - must call netif_poll_disable(), which
* schedule_timeout()'s if polling is already disabled.
*/
work_to_do
=
min
(
*
budget
,
dev
->
quota
);
work_done
=
gem_rx
(
gp
,
work_to_do
);
*
budget
-=
work_done
;
dev
->
quota
-=
work_done
;
if
(
work_done
>=
work_to_do
)
return
1
;
spin_lock_irqsave
(
&
gp
->
lock
,
flags
);
gp
->
status
=
readl
(
gp
->
regs
+
GREG_STAT
);
}
while
(
gp
->
status
&
GREG_STAT_NAPI
);
__netif_rx_complete
(
dev
);
gem_enable_ints
(
gp
);
spin_unlock_irqrestore
(
&
gp
->
lock
,
flags
);
return
0
;
}
}
static
irqreturn_t
gem_interrupt
(
int
irq
,
void
*
dev_id
,
struct
pt_regs
*
regs
)
static
irqreturn_t
gem_interrupt
(
int
irq
,
void
*
dev_id
,
struct
pt_regs
*
regs
)
{
{
struct
net_device
*
dev
=
dev_id
;
struct
net_device
*
dev
=
dev_id
;
struct
gem
*
gp
=
dev
->
priv
;
struct
gem
*
gp
=
dev
->
priv
;
u
32
gem_status
=
readl
(
gp
->
regs
+
GREG_STAT
)
;
u
nsigned
long
flags
;
/* Swallow interrupts when shutting the chip down */
/* Swallow interrupts when shutting the chip down */
if
(
gp
->
hw_running
==
0
)
if
(
!
gp
->
hw_running
)
goto
out
;
return
IRQ_HANDLED
;
spin_lock
(
&
gp
->
lock
);
spin_lock
_irqsave
(
&
gp
->
lock
,
flags
);
if
(
gem_status
&
GREG_STAT_ABNORMAL
)
{
if
(
netif_rx_schedule_prep
(
dev
))
{
if
(
gem_abnormal_irq
(
dev
,
gp
,
gem_status
))
u32
gem_status
=
readl
(
gp
->
regs
+
GREG_STAT
);
goto
out_unlock
;
if
(
gem_status
==
0
)
{
spin_unlock_irqrestore
(
&
gp
->
lock
,
flags
);
return
IRQ_NONE
;
}
gp
->
status
=
gem_status
;
gem_disable_ints
(
gp
);
__netif_rx_schedule
(
dev
);
}
}
if
(
gem_status
&
(
GREG_STAT_TXALL
|
GREG_STAT_TXINTME
))
gem_tx
(
dev
,
gp
,
gem_status
);
if
(
gem_status
&
GREG_STAT_RXDONE
)
gem_rx
(
gp
);
out_unlock:
spin_unlock_irqrestore
(
&
gp
->
lock
,
flags
);
spin_unlock
(
&
gp
->
lock
);
out:
/* If polling was disabled at the time we received that
* interrupt, we may return IRQ_HANDLED here while we
* should return IRQ_NONE. No big deal...
*/
return
IRQ_HANDLED
;
return
IRQ_HANDLED
;
}
}
...
@@ -1312,19 +1387,12 @@ static void gem_reset_task(void *data)
...
@@ -1312,19 +1387,12 @@ static void gem_reset_task(void *data)
{
{
struct
gem
*
gp
=
(
struct
gem
*
)
data
;
struct
gem
*
gp
=
(
struct
gem
*
)
data
;
/* The link went down, we reset the ring, but keep
netif_poll_disable
(
gp
->
dev
);
* DMA stopped. Todo: Use this function for reset
* on error as well.
*/
spin_lock_irq
(
&
gp
->
lock
);
spin_lock_irq
(
&
gp
->
lock
);
if
(
gp
->
hw_running
&&
gp
->
opened
)
{
if
(
gp
->
hw_running
&&
gp
->
opened
)
{
/* Make sure we don't get interrupts or tx packets */
netif_stop_queue
(
gp
->
dev
);
netif_stop_queue
(
gp
->
dev
);
writel
(
0xffffffff
,
gp
->
regs
+
GREG_IMASK
);
/* Reset the chip & rings */
/* Reset the chip & rings */
gem_stop
(
gp
);
gem_stop
(
gp
);
gem_init_rings
(
gp
);
gem_init_rings
(
gp
);
...
@@ -1337,6 +1405,7 @@ static void gem_reset_task(void *data)
...
@@ -1337,6 +1405,7 @@ static void gem_reset_task(void *data)
gp
->
reset_task_pending
=
0
;
gp
->
reset_task_pending
=
0
;
spin_unlock_irq
(
&
gp
->
lock
);
spin_unlock_irq
(
&
gp
->
lock
);
netif_poll_enable
(
gp
->
dev
);
}
}
static
void
gem_link_timer
(
unsigned
long
data
)
static
void
gem_link_timer
(
unsigned
long
data
)
...
@@ -2214,11 +2283,15 @@ static int gem_close(struct net_device *dev)
...
@@ -2214,11 +2283,15 @@ static int gem_close(struct net_device *dev)
/* Make sure we don't get distracted by suspend/resume */
/* Make sure we don't get distracted by suspend/resume */
down
(
&
gp
->
pm_sem
);
down
(
&
gp
->
pm_sem
);
/* Note: we don't need to call netif_poll_disable() here because
* our caller (dev_close) already did it for us
*/
/* Stop traffic, mark us closed */
/* Stop traffic, mark us closed */
spin_lock_irq
(
&
gp
->
lock
);
spin_lock_irq
(
&
gp
->
lock
);
gp
->
opened
=
0
;
gp
->
opened
=
0
;
writel
(
0xffffffff
,
gp
->
regs
+
GREG_IMASK
);
netif_stop_queue
(
dev
);
netif_stop_queue
(
dev
);
/* Stop chip */
/* Stop chip */
...
@@ -2247,6 +2320,8 @@ static int gem_suspend(struct pci_dev *pdev, u32 state)
...
@@ -2247,6 +2320,8 @@ static int gem_suspend(struct pci_dev *pdev, u32 state)
struct
net_device
*
dev
=
pci_get_drvdata
(
pdev
);
struct
net_device
*
dev
=
pci_get_drvdata
(
pdev
);
struct
gem
*
gp
=
dev
->
priv
;
struct
gem
*
gp
=
dev
->
priv
;
netif_poll_disable
(
dev
);
/* We hold the PM semaphore during entire driver
/* We hold the PM semaphore during entire driver
* sleep time
* sleep time
*/
*/
...
@@ -2262,8 +2337,6 @@ static int gem_suspend(struct pci_dev *pdev, u32 state)
...
@@ -2262,8 +2337,6 @@ static int gem_suspend(struct pci_dev *pdev, u32 state)
/* Stop traffic, mark us closed */
/* Stop traffic, mark us closed */
netif_device_detach
(
dev
);
netif_device_detach
(
dev
);
writel
(
0xffffffff
,
gp
->
regs
+
GREG_IMASK
);
/* Stop chip */
/* Stop chip */
gem_stop
(
gp
);
gem_stop
(
gp
);
...
@@ -2317,6 +2390,8 @@ static int gem_resume(struct pci_dev *pdev)
...
@@ -2317,6 +2390,8 @@ static int gem_resume(struct pci_dev *pdev)
}
}
up
(
&
gp
->
pm_sem
);
up
(
&
gp
->
pm_sem
);
netif_poll_enable
(
dev
);
return
0
;
return
0
;
}
}
#endif
/* CONFIG_PM */
#endif
/* CONFIG_PM */
...
@@ -2806,6 +2881,8 @@ static int __devinit gem_init_one(struct pci_dev *pdev,
...
@@ -2806,6 +2881,8 @@ static int __devinit gem_init_one(struct pci_dev *pdev,
dev
->
get_stats
=
gem_get_stats
;
dev
->
get_stats
=
gem_get_stats
;
dev
->
set_multicast_list
=
gem_set_multicast
;
dev
->
set_multicast_list
=
gem_set_multicast
;
dev
->
do_ioctl
=
gem_ioctl
;
dev
->
do_ioctl
=
gem_ioctl
;
dev
->
poll
=
gem_poll
;
dev
->
weight
=
64
;
dev
->
ethtool_ops
=
&
gem_ethtool_ops
;
dev
->
ethtool_ops
=
&
gem_ethtool_ops
;
dev
->
tx_timeout
=
gem_tx_timeout
;
dev
->
tx_timeout
=
gem_tx_timeout
;
dev
->
watchdog_timeo
=
5
*
HZ
;
dev
->
watchdog_timeo
=
5
*
HZ
;
...
...
drivers/net/sungem.h
View file @
6cb553d2
...
@@ -60,6 +60,9 @@
...
@@ -60,6 +60,9 @@
GREG_STAT_PCS | GREG_STAT_TXMAC | GREG_STAT_RXMAC | \
GREG_STAT_PCS | GREG_STAT_TXMAC | GREG_STAT_RXMAC | \
GREG_STAT_MAC | GREG_STAT_MIF | GREG_STAT_PCIERR)
GREG_STAT_MAC | GREG_STAT_MIF | GREG_STAT_PCIERR)
#define GREG_STAT_NAPI (GREG_STAT_TXALL | GREG_STAT_TXINTME | \
GREG_STAT_RXDONE | GREG_STAT_ABNORMAL)
/* The layout of GREG_IMASK and GREG_IACK is identical to GREG_STAT.
/* The layout of GREG_IMASK and GREG_IACK is identical to GREG_STAT.
* Bits set in GREG_IMASK will prevent that interrupt type from being
* Bits set in GREG_IMASK will prevent that interrupt type from being
* signalled to the cpu. GREG_IACK can be used to clear specific top-level
* signalled to the cpu. GREG_IACK can be used to clear specific top-level
...
@@ -969,6 +972,7 @@ struct gem {
...
@@ -969,6 +972,7 @@ struct gem {
struct
sk_buff
*
tx_skbs
[
RX_RING_SIZE
];
struct
sk_buff
*
tx_skbs
[
RX_RING_SIZE
];
u32
msg_enable
;
u32
msg_enable
;
u32
status
;
struct
net_device_stats
net_stats
;
struct
net_device_stats
net_stats
;
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment