Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
Kirill Smelkov
linux
Commits
44757223
Commit
44757223
authored
Sep 02, 2005
by
Linus Torvalds
Browse files
Options
Browse Files
Download
Plain Diff
Merge refs/heads/upstream from master.kernel.org:/pub/scm/linux/kernel/git/jgarzik/netdev-2.6
parents
712fbdd3
ee05f031
Changes
4
Show whitespace changes
Inline
Side-by-side
Showing
4 changed files
with
583 additions
and
335 deletions
+583
-335
drivers/net/iseries_veth.c
drivers/net/iseries_veth.c
+581
-288
drivers/net/iseries_veth.h
drivers/net/iseries_veth.h
+0
-46
drivers/net/tulip/de2104x.c
drivers/net/tulip/de2104x.c
+1
-1
drivers/net/tulip/tulip_core.c
drivers/net/tulip/tulip_core.c
+1
-0
No files found.
drivers/net/iseries_veth.c
View file @
44757223
...
@@ -79,12 +79,55 @@
...
@@ -79,12 +79,55 @@
#include <asm/iommu.h>
#include <asm/iommu.h>
#include <asm/vio.h>
#include <asm/vio.h>
#
include "iseries_veth.h"
#
undef DEBUG
MODULE_AUTHOR
(
"Kyle Lucke <klucke@us.ibm.com>"
);
MODULE_AUTHOR
(
"Kyle Lucke <klucke@us.ibm.com>"
);
MODULE_DESCRIPTION
(
"iSeries Virtual ethernet driver"
);
MODULE_DESCRIPTION
(
"iSeries Virtual ethernet driver"
);
MODULE_LICENSE
(
"GPL"
);
MODULE_LICENSE
(
"GPL"
);
#define VETH_EVENT_CAP (0)
#define VETH_EVENT_FRAMES (1)
#define VETH_EVENT_MONITOR (2)
#define VETH_EVENT_FRAMES_ACK (3)
#define VETH_MAX_ACKS_PER_MSG (20)
#define VETH_MAX_FRAMES_PER_MSG (6)
struct
veth_frames_data
{
u32
addr
[
VETH_MAX_FRAMES_PER_MSG
];
u16
len
[
VETH_MAX_FRAMES_PER_MSG
];
u32
eofmask
;
};
#define VETH_EOF_SHIFT (32-VETH_MAX_FRAMES_PER_MSG)
struct
veth_frames_ack_data
{
u16
token
[
VETH_MAX_ACKS_PER_MSG
];
};
struct
veth_cap_data
{
u8
caps_version
;
u8
rsvd1
;
u16
num_buffers
;
u16
ack_threshold
;
u16
rsvd2
;
u32
ack_timeout
;
u32
rsvd3
;
u64
rsvd4
[
3
];
};
struct
veth_lpevent
{
struct
HvLpEvent
base_event
;
union
{
struct
veth_cap_data
caps_data
;
struct
veth_frames_data
frames_data
;
struct
veth_frames_ack_data
frames_ack_data
;
}
u
;
};
#define DRV_NAME "iseries_veth"
#define DRV_VERSION "2.0"
#define VETH_NUMBUFFERS (120)
#define VETH_NUMBUFFERS (120)
#define VETH_ACKTIMEOUT (1000000)
/* microseconds */
#define VETH_ACKTIMEOUT (1000000)
/* microseconds */
#define VETH_MAX_MCAST (12)
#define VETH_MAX_MCAST (12)
...
@@ -113,9 +156,9 @@ MODULE_LICENSE("GPL");
...
@@ -113,9 +156,9 @@ MODULE_LICENSE("GPL");
struct
veth_msg
{
struct
veth_msg
{
struct
veth_msg
*
next
;
struct
veth_msg
*
next
;
struct
VethFramesD
ata
data
;
struct
veth_frames_d
ata
data
;
int
token
;
int
token
;
unsigned
long
in_use
;
int
in_use
;
struct
sk_buff
*
skb
;
struct
sk_buff
*
skb
;
struct
device
*
dev
;
struct
device
*
dev
;
};
};
...
@@ -125,23 +168,28 @@ struct veth_lpar_connection {
...
@@ -125,23 +168,28 @@ struct veth_lpar_connection {
struct
work_struct
statemachine_wq
;
struct
work_struct
statemachine_wq
;
struct
veth_msg
*
msgs
;
struct
veth_msg
*
msgs
;
int
num_events
;
int
num_events
;
struct
VethCapD
ata
local_caps
;
struct
veth_cap_d
ata
local_caps
;
struct
kobject
kobject
;
struct
timer_list
ack_timer
;
struct
timer_list
ack_timer
;
struct
timer_list
reset_timer
;
unsigned
int
reset_timeout
;
unsigned
long
last_contact
;
int
outstanding_tx
;
spinlock_t
lock
;
spinlock_t
lock
;
unsigned
long
state
;
unsigned
long
state
;
HvLpInstanceId
src_inst
;
HvLpInstanceId
src_inst
;
HvLpInstanceId
dst_inst
;
HvLpInstanceId
dst_inst
;
struct
VethLpE
vent
cap_event
,
cap_ack_event
;
struct
veth_lpe
vent
cap_event
,
cap_ack_event
;
u16
pending_acks
[
VETH_MAX_ACKS_PER_MSG
];
u16
pending_acks
[
VETH_MAX_ACKS_PER_MSG
];
u32
num_pending_acks
;
u32
num_pending_acks
;
int
num_ack_events
;
int
num_ack_events
;
struct
VethCapD
ata
remote_caps
;
struct
veth_cap_d
ata
remote_caps
;
u32
ack_timeout
;
u32
ack_timeout
;
spinlock_t
msg_stack_lock
;
struct
veth_msg
*
msg_stack_head
;
struct
veth_msg
*
msg_stack_head
;
};
};
...
@@ -151,15 +199,17 @@ struct veth_port {
...
@@ -151,15 +199,17 @@ struct veth_port {
u64
mac_addr
;
u64
mac_addr
;
HvLpIndexMap
lpar_map
;
HvLpIndexMap
lpar_map
;
spinlock_t
pending_gate
;
/* queue_lock protects the stopped_map and dev's queue. */
s
truct
sk_buff
*
pending_skb
;
s
pinlock_t
queue_lock
;
HvLpIndexMap
pending_lpmask
;
HvLpIndexMap
stopped_map
;
/* mcast_gate protects promiscuous, num_mcast & mcast_addr. */
rwlock_t
mcast_gate
;
rwlock_t
mcast_gate
;
int
promiscuous
;
int
promiscuous
;
int
all_mcast
;
int
num_mcast
;
int
num_mcast
;
u64
mcast_addr
[
VETH_MAX_MCAST
];
u64
mcast_addr
[
VETH_MAX_MCAST
];
struct
kobject
kobject
;
};
};
static
HvLpIndex
this_lp
;
static
HvLpIndex
this_lp
;
...
@@ -168,44 +218,56 @@ static struct net_device *veth_dev[HVMAXARCHITECTEDVIRTUALLANS]; /* = 0 */
...
@@ -168,44 +218,56 @@ static struct net_device *veth_dev[HVMAXARCHITECTEDVIRTUALLANS]; /* = 0 */
static
int
veth_start_xmit
(
struct
sk_buff
*
skb
,
struct
net_device
*
dev
);
static
int
veth_start_xmit
(
struct
sk_buff
*
skb
,
struct
net_device
*
dev
);
static
void
veth_recycle_msg
(
struct
veth_lpar_connection
*
,
struct
veth_msg
*
);
static
void
veth_recycle_msg
(
struct
veth_lpar_connection
*
,
struct
veth_msg
*
);
static
void
veth_flush_pending
(
struct
veth_lpar_connection
*
cnx
);
static
void
veth_wake_queues
(
struct
veth_lpar_connection
*
cnx
);
static
void
veth_receive
(
struct
veth_lpar_connection
*
,
struct
VethLpEvent
*
);
static
void
veth_stop_queues
(
struct
veth_lpar_connection
*
cnx
);
static
void
veth_timed_ack
(
unsigned
long
connectionPtr
);
static
void
veth_receive
(
struct
veth_lpar_connection
*
,
struct
veth_lpevent
*
);
static
void
veth_release_connection
(
struct
kobject
*
kobject
);
static
void
veth_timed_ack
(
unsigned
long
ptr
);
static
void
veth_timed_reset
(
unsigned
long
ptr
);
/*
/*
* Utility functions
* Utility functions
*/
*/
#define veth_
printk(prio,
fmt, args...) \
#define veth_
info(
fmt, args...) \
printk(
prio "%s: " fmt, __FILE__
, ## args)
printk(
KERN_INFO DRV_NAME ": " fmt
, ## args)
#define veth_error(fmt, args...) \
#define veth_error(fmt, args...) \
printk(KERN_ERR "(%s:%3.3d) ERROR: " fmt, __FILE__, __LINE__ , ## args)
printk(KERN_ERR DRV_NAME ": Error: " fmt, ## args)
#ifdef DEBUG
#define veth_debug(fmt, args...) \
printk(KERN_DEBUG DRV_NAME ": " fmt, ## args)
#else
#define veth_debug(fmt, args...) do {} while (0)
#endif
/* You must hold the connection's lock when you call this function. */
static
inline
void
veth_stack_push
(
struct
veth_lpar_connection
*
cnx
,
static
inline
void
veth_stack_push
(
struct
veth_lpar_connection
*
cnx
,
struct
veth_msg
*
msg
)
struct
veth_msg
*
msg
)
{
{
unsigned
long
flags
;
spin_lock_irqsave
(
&
cnx
->
msg_stack_lock
,
flags
);
msg
->
next
=
cnx
->
msg_stack_head
;
msg
->
next
=
cnx
->
msg_stack_head
;
cnx
->
msg_stack_head
=
msg
;
cnx
->
msg_stack_head
=
msg
;
spin_unlock_irqrestore
(
&
cnx
->
msg_stack_lock
,
flags
);
}
}
/* You must hold the connection's lock when you call this function. */
static
inline
struct
veth_msg
*
veth_stack_pop
(
struct
veth_lpar_connection
*
cnx
)
static
inline
struct
veth_msg
*
veth_stack_pop
(
struct
veth_lpar_connection
*
cnx
)
{
{
unsigned
long
flags
;
struct
veth_msg
*
msg
;
struct
veth_msg
*
msg
;
spin_lock_irqsave
(
&
cnx
->
msg_stack_lock
,
flags
);
msg
=
cnx
->
msg_stack_head
;
msg
=
cnx
->
msg_stack_head
;
if
(
msg
)
if
(
msg
)
cnx
->
msg_stack_head
=
cnx
->
msg_stack_head
->
next
;
cnx
->
msg_stack_head
=
cnx
->
msg_stack_head
->
next
;
spin_unlock_irqrestore
(
&
cnx
->
msg_stack_lock
,
flags
);
return
msg
;
return
msg
;
}
}
/* You must hold the connection's lock when you call this function. */
static
inline
int
veth_stack_is_empty
(
struct
veth_lpar_connection
*
cnx
)
{
return
cnx
->
msg_stack_head
==
NULL
;
}
static
inline
HvLpEvent_Rc
static
inline
HvLpEvent_Rc
veth_signalevent
(
struct
veth_lpar_connection
*
cnx
,
u16
subtype
,
veth_signalevent
(
struct
veth_lpar_connection
*
cnx
,
u16
subtype
,
HvLpEvent_AckInd
ackind
,
HvLpEvent_AckType
acktype
,
HvLpEvent_AckInd
ackind
,
HvLpEvent_AckType
acktype
,
...
@@ -249,13 +311,144 @@ static int veth_allocate_events(HvLpIndex rlp, int number)
...
@@ -249,13 +311,144 @@ static int veth_allocate_events(HvLpIndex rlp, int number)
struct
veth_allocation
vc
=
{
COMPLETION_INITIALIZER
(
vc
.
c
),
0
};
struct
veth_allocation
vc
=
{
COMPLETION_INITIALIZER
(
vc
.
c
),
0
};
mf_allocate_lp_events
(
rlp
,
HvLpEvent_Type_VirtualLan
,
mf_allocate_lp_events
(
rlp
,
HvLpEvent_Type_VirtualLan
,
sizeof
(
struct
VethLpE
vent
),
number
,
sizeof
(
struct
veth_lpe
vent
),
number
,
&
veth_complete_allocation
,
&
vc
);
&
veth_complete_allocation
,
&
vc
);
wait_for_completion
(
&
vc
.
c
);
wait_for_completion
(
&
vc
.
c
);
return
vc
.
num
;
return
vc
.
num
;
}
}
/*
* sysfs support
*/
struct
veth_cnx_attribute
{
struct
attribute
attr
;
ssize_t
(
*
show
)(
struct
veth_lpar_connection
*
,
char
*
buf
);
ssize_t
(
*
store
)(
struct
veth_lpar_connection
*
,
const
char
*
buf
);
};
static
ssize_t
veth_cnx_attribute_show
(
struct
kobject
*
kobj
,
struct
attribute
*
attr
,
char
*
buf
)
{
struct
veth_cnx_attribute
*
cnx_attr
;
struct
veth_lpar_connection
*
cnx
;
cnx_attr
=
container_of
(
attr
,
struct
veth_cnx_attribute
,
attr
);
cnx
=
container_of
(
kobj
,
struct
veth_lpar_connection
,
kobject
);
if
(
!
cnx_attr
->
show
)
return
-
EIO
;
return
cnx_attr
->
show
(
cnx
,
buf
);
}
#define CUSTOM_CNX_ATTR(_name, _format, _expression) \
static ssize_t _name##_show(struct veth_lpar_connection *cnx, char *buf)\
{ \
return sprintf(buf, _format, _expression); \
} \
struct veth_cnx_attribute veth_cnx_attr_##_name = __ATTR_RO(_name)
#define SIMPLE_CNX_ATTR(_name) \
CUSTOM_CNX_ATTR(_name, "%lu\n", (unsigned long)cnx->_name)
SIMPLE_CNX_ATTR
(
outstanding_tx
);
SIMPLE_CNX_ATTR
(
remote_lp
);
SIMPLE_CNX_ATTR
(
num_events
);
SIMPLE_CNX_ATTR
(
src_inst
);
SIMPLE_CNX_ATTR
(
dst_inst
);
SIMPLE_CNX_ATTR
(
num_pending_acks
);
SIMPLE_CNX_ATTR
(
num_ack_events
);
CUSTOM_CNX_ATTR
(
ack_timeout
,
"%d
\n
"
,
jiffies_to_msecs
(
cnx
->
ack_timeout
));
CUSTOM_CNX_ATTR
(
reset_timeout
,
"%d
\n
"
,
jiffies_to_msecs
(
cnx
->
reset_timeout
));
CUSTOM_CNX_ATTR
(
state
,
"0x%.4lX
\n
"
,
cnx
->
state
);
CUSTOM_CNX_ATTR
(
last_contact
,
"%d
\n
"
,
cnx
->
last_contact
?
jiffies_to_msecs
(
jiffies
-
cnx
->
last_contact
)
:
0
);
#define GET_CNX_ATTR(_name) (&veth_cnx_attr_##_name.attr)
static
struct
attribute
*
veth_cnx_default_attrs
[]
=
{
GET_CNX_ATTR
(
outstanding_tx
),
GET_CNX_ATTR
(
remote_lp
),
GET_CNX_ATTR
(
num_events
),
GET_CNX_ATTR
(
reset_timeout
),
GET_CNX_ATTR
(
last_contact
),
GET_CNX_ATTR
(
state
),
GET_CNX_ATTR
(
src_inst
),
GET_CNX_ATTR
(
dst_inst
),
GET_CNX_ATTR
(
num_pending_acks
),
GET_CNX_ATTR
(
num_ack_events
),
GET_CNX_ATTR
(
ack_timeout
),
NULL
};
static
struct
sysfs_ops
veth_cnx_sysfs_ops
=
{
.
show
=
veth_cnx_attribute_show
};
static
struct
kobj_type
veth_lpar_connection_ktype
=
{
.
release
=
veth_release_connection
,
.
sysfs_ops
=
&
veth_cnx_sysfs_ops
,
.
default_attrs
=
veth_cnx_default_attrs
};
struct
veth_port_attribute
{
struct
attribute
attr
;
ssize_t
(
*
show
)(
struct
veth_port
*
,
char
*
buf
);
ssize_t
(
*
store
)(
struct
veth_port
*
,
const
char
*
buf
);
};
static
ssize_t
veth_port_attribute_show
(
struct
kobject
*
kobj
,
struct
attribute
*
attr
,
char
*
buf
)
{
struct
veth_port_attribute
*
port_attr
;
struct
veth_port
*
port
;
port_attr
=
container_of
(
attr
,
struct
veth_port_attribute
,
attr
);
port
=
container_of
(
kobj
,
struct
veth_port
,
kobject
);
if
(
!
port_attr
->
show
)
return
-
EIO
;
return
port_attr
->
show
(
port
,
buf
);
}
#define CUSTOM_PORT_ATTR(_name, _format, _expression) \
static ssize_t _name##_show(struct veth_port *port, char *buf) \
{ \
return sprintf(buf, _format, _expression); \
} \
struct veth_port_attribute veth_port_attr_##_name = __ATTR_RO(_name)
#define SIMPLE_PORT_ATTR(_name) \
CUSTOM_PORT_ATTR(_name, "%lu\n", (unsigned long)port->_name)
SIMPLE_PORT_ATTR
(
promiscuous
);
SIMPLE_PORT_ATTR
(
num_mcast
);
CUSTOM_PORT_ATTR
(
lpar_map
,
"0x%X
\n
"
,
port
->
lpar_map
);
CUSTOM_PORT_ATTR
(
stopped_map
,
"0x%X
\n
"
,
port
->
stopped_map
);
CUSTOM_PORT_ATTR
(
mac_addr
,
"0x%lX
\n
"
,
port
->
mac_addr
);
#define GET_PORT_ATTR(_name) (&veth_port_attr_##_name.attr)
static
struct
attribute
*
veth_port_default_attrs
[]
=
{
GET_PORT_ATTR
(
mac_addr
),
GET_PORT_ATTR
(
lpar_map
),
GET_PORT_ATTR
(
stopped_map
),
GET_PORT_ATTR
(
promiscuous
),
GET_PORT_ATTR
(
num_mcast
),
NULL
};
static
struct
sysfs_ops
veth_port_sysfs_ops
=
{
.
show
=
veth_port_attribute_show
};
static
struct
kobj_type
veth_port_ktype
=
{
.
sysfs_ops
=
&
veth_port_sysfs_ops
,
.
default_attrs
=
veth_port_default_attrs
};
/*
/*
* LPAR connection code
* LPAR connection code
*/
*/
...
@@ -266,7 +459,7 @@ static inline void veth_kick_statemachine(struct veth_lpar_connection *cnx)
...
@@ -266,7 +459,7 @@ static inline void veth_kick_statemachine(struct veth_lpar_connection *cnx)
}
}
static
void
veth_take_cap
(
struct
veth_lpar_connection
*
cnx
,
static
void
veth_take_cap
(
struct
veth_lpar_connection
*
cnx
,
struct
VethLpE
vent
*
event
)
struct
veth_lpe
vent
*
event
)
{
{
unsigned
long
flags
;
unsigned
long
flags
;
...
@@ -278,7 +471,7 @@ static void veth_take_cap(struct veth_lpar_connection *cnx,
...
@@ -278,7 +471,7 @@ static void veth_take_cap(struct veth_lpar_connection *cnx,
HvLpEvent_Type_VirtualLan
);
HvLpEvent_Type_VirtualLan
);
if
(
cnx
->
state
&
VETH_STATE_GOTCAPS
)
{
if
(
cnx
->
state
&
VETH_STATE_GOTCAPS
)
{
veth_error
(
"Received a second capabilities from
lpar %d
\n
"
,
veth_error
(
"Received a second capabilities from
LPAR %d.
\n
"
,
cnx
->
remote_lp
);
cnx
->
remote_lp
);
event
->
base_event
.
xRc
=
HvLpEvent_Rc_BufferNotAvailable
;
event
->
base_event
.
xRc
=
HvLpEvent_Rc_BufferNotAvailable
;
HvCallEvent_ackLpEvent
((
struct
HvLpEvent
*
)
event
);
HvCallEvent_ackLpEvent
((
struct
HvLpEvent
*
)
event
);
...
@@ -291,13 +484,13 @@ static void veth_take_cap(struct veth_lpar_connection *cnx,
...
@@ -291,13 +484,13 @@ static void veth_take_cap(struct veth_lpar_connection *cnx,
}
}
static
void
veth_take_cap_ack
(
struct
veth_lpar_connection
*
cnx
,
static
void
veth_take_cap_ack
(
struct
veth_lpar_connection
*
cnx
,
struct
VethLpE
vent
*
event
)
struct
veth_lpe
vent
*
event
)
{
{
unsigned
long
flags
;
unsigned
long
flags
;
spin_lock_irqsave
(
&
cnx
->
lock
,
flags
);
spin_lock_irqsave
(
&
cnx
->
lock
,
flags
);
if
(
cnx
->
state
&
VETH_STATE_GOTCAPACK
)
{
if
(
cnx
->
state
&
VETH_STATE_GOTCAPACK
)
{
veth_error
(
"Received a second capabilities ack from
lpar %d
\n
"
,
veth_error
(
"Received a second capabilities ack from
LPAR %d.
\n
"
,
cnx
->
remote_lp
);
cnx
->
remote_lp
);
}
else
{
}
else
{
memcpy
(
&
cnx
->
cap_ack_event
,
event
,
memcpy
(
&
cnx
->
cap_ack_event
,
event
,
...
@@ -309,19 +502,24 @@ static void veth_take_cap_ack(struct veth_lpar_connection *cnx,
...
@@ -309,19 +502,24 @@ static void veth_take_cap_ack(struct veth_lpar_connection *cnx,
}
}
static
void
veth_take_monitor_ack
(
struct
veth_lpar_connection
*
cnx
,
static
void
veth_take_monitor_ack
(
struct
veth_lpar_connection
*
cnx
,
struct
VethLpE
vent
*
event
)
struct
veth_lpe
vent
*
event
)
{
{
unsigned
long
flags
;
unsigned
long
flags
;
spin_lock_irqsave
(
&
cnx
->
lock
,
flags
);
spin_lock_irqsave
(
&
cnx
->
lock
,
flags
);
veth_printk
(
KERN_DEBUG
,
"Monitor ack returned for lpar %d
\n
"
,
veth_debug
(
"cnx %d: lost connection.
\n
"
,
cnx
->
remote_lp
);
cnx
->
remote_lp
);
/* Avoid kicking the statemachine once we're shutdown.
* It's unnecessary and it could break veth_stop_connection(). */
if
(
!
(
cnx
->
state
&
VETH_STATE_SHUTDOWN
))
{
cnx
->
state
|=
VETH_STATE_RESET
;
cnx
->
state
|=
VETH_STATE_RESET
;
veth_kick_statemachine
(
cnx
);
veth_kick_statemachine
(
cnx
);
}
spin_unlock_irqrestore
(
&
cnx
->
lock
,
flags
);
spin_unlock_irqrestore
(
&
cnx
->
lock
,
flags
);
}
}
static
void
veth_handle_ack
(
struct
VethLpE
vent
*
event
)
static
void
veth_handle_ack
(
struct
veth_lpe
vent
*
event
)
{
{
HvLpIndex
rlp
=
event
->
base_event
.
xTargetLp
;
HvLpIndex
rlp
=
event
->
base_event
.
xTargetLp
;
struct
veth_lpar_connection
*
cnx
=
veth_cnx
[
rlp
];
struct
veth_lpar_connection
*
cnx
=
veth_cnx
[
rlp
];
...
@@ -329,58 +527,67 @@ static void veth_handle_ack(struct VethLpEvent *event)
...
@@ -329,58 +527,67 @@ static void veth_handle_ack(struct VethLpEvent *event)
BUG_ON
(
!
cnx
);
BUG_ON
(
!
cnx
);
switch
(
event
->
base_event
.
xSubtype
)
{
switch
(
event
->
base_event
.
xSubtype
)
{
case
V
ethEventTypeCap
:
case
V
ETH_EVENT_CAP
:
veth_take_cap_ack
(
cnx
,
event
);
veth_take_cap_ack
(
cnx
,
event
);
break
;
break
;
case
V
ethEventTypeMonitor
:
case
V
ETH_EVENT_MONITOR
:
veth_take_monitor_ack
(
cnx
,
event
);
veth_take_monitor_ack
(
cnx
,
event
);
break
;
break
;
default:
default:
veth_error
(
"Unknown ack type %d from
lpar %d
\n
"
,
veth_error
(
"Unknown ack type %d from
LPAR %d.
\n
"
,
event
->
base_event
.
xSubtype
,
rlp
);
event
->
base_event
.
xSubtype
,
rlp
);
};
};
}
}
static
void
veth_handle_int
(
struct
VethLpE
vent
*
event
)
static
void
veth_handle_int
(
struct
veth_lpe
vent
*
event
)
{
{
HvLpIndex
rlp
=
event
->
base_event
.
xSourceLp
;
HvLpIndex
rlp
=
event
->
base_event
.
xSourceLp
;
struct
veth_lpar_connection
*
cnx
=
veth_cnx
[
rlp
];
struct
veth_lpar_connection
*
cnx
=
veth_cnx
[
rlp
];
unsigned
long
flags
;
unsigned
long
flags
;
int
i
;
int
i
,
acked
=
0
;
BUG_ON
(
!
cnx
);
BUG_ON
(
!
cnx
);
switch
(
event
->
base_event
.
xSubtype
)
{
switch
(
event
->
base_event
.
xSubtype
)
{
case
V
ethEventTypeCap
:
case
V
ETH_EVENT_CAP
:
veth_take_cap
(
cnx
,
event
);
veth_take_cap
(
cnx
,
event
);
break
;
break
;
case
V
ethEventTypeMonitor
:
case
V
ETH_EVENT_MONITOR
:
/* do nothing... this'll hang out here til we're dead,
/* do nothing... this'll hang out here til we're dead,
* and the hypervisor will return it for us. */
* and the hypervisor will return it for us. */
break
;
break
;
case
V
ethEventTypeFramesAck
:
case
V
ETH_EVENT_FRAMES_ACK
:
spin_lock_irqsave
(
&
cnx
->
lock
,
flags
);
spin_lock_irqsave
(
&
cnx
->
lock
,
flags
);
for
(
i
=
0
;
i
<
VETH_MAX_ACKS_PER_MSG
;
++
i
)
{
for
(
i
=
0
;
i
<
VETH_MAX_ACKS_PER_MSG
;
++
i
)
{
u16
msgnum
=
event
->
u
.
frames_ack_data
.
token
[
i
];
u16
msgnum
=
event
->
u
.
frames_ack_data
.
token
[
i
];
if
(
msgnum
<
VETH_NUMBUFFERS
)
if
(
msgnum
<
VETH_NUMBUFFERS
)
{
veth_recycle_msg
(
cnx
,
cnx
->
msgs
+
msgnum
);
veth_recycle_msg
(
cnx
,
cnx
->
msgs
+
msgnum
);
cnx
->
outstanding_tx
--
;
acked
++
;
}
}
}
if
(
acked
>
0
)
{
cnx
->
last_contact
=
jiffies
;
veth_wake_queues
(
cnx
);
}
spin_unlock_irqrestore
(
&
cnx
->
lock
,
flags
);
spin_unlock_irqrestore
(
&
cnx
->
lock
,
flags
);
veth_flush_pending
(
cnx
);
break
;
break
;
case
V
ethEventTypeFrames
:
case
V
ETH_EVENT_FRAMES
:
veth_receive
(
cnx
,
event
);
veth_receive
(
cnx
,
event
);
break
;
break
;
default:
default:
veth_error
(
"Unknown interrupt type %d from
lpar %d
\n
"
,
veth_error
(
"Unknown interrupt type %d from
LPAR %d.
\n
"
,
event
->
base_event
.
xSubtype
,
rlp
);
event
->
base_event
.
xSubtype
,
rlp
);
};
};
}
}
static
void
veth_handle_event
(
struct
HvLpEvent
*
event
,
struct
pt_regs
*
regs
)
static
void
veth_handle_event
(
struct
HvLpEvent
*
event
,
struct
pt_regs
*
regs
)
{
{
struct
VethLpEvent
*
veth_event
=
(
struct
VethLpE
vent
*
)
event
;
struct
veth_lpevent
*
veth_event
=
(
struct
veth_lpe
vent
*
)
event
;
if
(
event
->
xFlags
.
xFunction
==
HvLpEvent_Function_Ack
)
if
(
event
->
xFlags
.
xFunction
==
HvLpEvent_Function_Ack
)
veth_handle_ack
(
veth_event
);
veth_handle_ack
(
veth_event
);
...
@@ -390,7 +597,7 @@ static void veth_handle_event(struct HvLpEvent *event, struct pt_regs *regs)
...
@@ -390,7 +597,7 @@ static void veth_handle_event(struct HvLpEvent *event, struct pt_regs *regs)
static
int
veth_process_caps
(
struct
veth_lpar_connection
*
cnx
)
static
int
veth_process_caps
(
struct
veth_lpar_connection
*
cnx
)
{
{
struct
VethCapD
ata
*
remote_caps
=
&
cnx
->
remote_caps
;
struct
veth_cap_d
ata
*
remote_caps
=
&
cnx
->
remote_caps
;
int
num_acks_needed
;
int
num_acks_needed
;
/* Convert timer to jiffies */
/* Convert timer to jiffies */
...
@@ -400,7 +607,7 @@ static int veth_process_caps(struct veth_lpar_connection *cnx)
...
@@ -400,7 +607,7 @@ static int veth_process_caps(struct veth_lpar_connection *cnx)
||
(
remote_caps
->
ack_threshold
>
VETH_MAX_ACKS_PER_MSG
)
||
(
remote_caps
->
ack_threshold
>
VETH_MAX_ACKS_PER_MSG
)
||
(
remote_caps
->
ack_threshold
==
0
)
||
(
remote_caps
->
ack_threshold
==
0
)
||
(
cnx
->
ack_timeout
==
0
)
)
{
||
(
cnx
->
ack_timeout
==
0
)
)
{
veth_error
(
"Received incompatible capabilities from
lpar %d
\n
"
,
veth_error
(
"Received incompatible capabilities from
LPAR %d.
\n
"
,
cnx
->
remote_lp
);
cnx
->
remote_lp
);
return
HvLpEvent_Rc_InvalidSubtypeData
;
return
HvLpEvent_Rc_InvalidSubtypeData
;
}
}
...
@@ -418,8 +625,8 @@ static int veth_process_caps(struct veth_lpar_connection *cnx)
...
@@ -418,8 +625,8 @@ static int veth_process_caps(struct veth_lpar_connection *cnx)
cnx
->
num_ack_events
+=
num
;
cnx
->
num_ack_events
+=
num
;
if
(
cnx
->
num_ack_events
<
num_acks_needed
)
{
if
(
cnx
->
num_ack_events
<
num_acks_needed
)
{
veth_error
(
"Couldn't allocate enough ack events
for lpar %d
\n
"
,
veth_error
(
"Couldn't allocate enough ack events
"
cnx
->
remote_lp
);
"for LPAR %d.
\n
"
,
cnx
->
remote_lp
);
return
HvLpEvent_Rc_BufferNotAvailable
;
return
HvLpEvent_Rc_BufferNotAvailable
;
}
}
...
@@ -440,15 +647,15 @@ static void veth_statemachine(void *p)
...
@@ -440,15 +647,15 @@ static void veth_statemachine(void *p)
restart:
restart:
if
(
cnx
->
state
&
VETH_STATE_RESET
)
{
if
(
cnx
->
state
&
VETH_STATE_RESET
)
{
int
i
;
del_timer
(
&
cnx
->
ack_timer
);
if
(
cnx
->
state
&
VETH_STATE_OPEN
)
if
(
cnx
->
state
&
VETH_STATE_OPEN
)
HvCallEvent_closeLpEventPath
(
cnx
->
remote_lp
,
HvCallEvent_closeLpEventPath
(
cnx
->
remote_lp
,
HvLpEvent_Type_VirtualLan
);
HvLpEvent_Type_VirtualLan
);
/* reset ack data */
/*
* Reset ack data. This prevents the ack_timer actually
* doing anything, even if it runs one more time when
* we drop the lock below.
*/
memset
(
&
cnx
->
pending_acks
,
0xff
,
sizeof
(
cnx
->
pending_acks
));
memset
(
&
cnx
->
pending_acks
,
0xff
,
sizeof
(
cnx
->
pending_acks
));
cnx
->
num_pending_acks
=
0
;
cnx
->
num_pending_acks
=
0
;
...
@@ -458,14 +665,32 @@ static void veth_statemachine(void *p)
...
@@ -458,14 +665,32 @@ static void veth_statemachine(void *p)
|
VETH_STATE_SENTCAPACK
|
VETH_STATE_READY
);
|
VETH_STATE_SENTCAPACK
|
VETH_STATE_READY
);
/* Clean up any leftover messages */
/* Clean up any leftover messages */
if
(
cnx
->
msgs
)
if
(
cnx
->
msgs
)
{
int
i
;
for
(
i
=
0
;
i
<
VETH_NUMBUFFERS
;
++
i
)
for
(
i
=
0
;
i
<
VETH_NUMBUFFERS
;
++
i
)
veth_recycle_msg
(
cnx
,
cnx
->
msgs
+
i
);
veth_recycle_msg
(
cnx
,
cnx
->
msgs
+
i
);
}
cnx
->
outstanding_tx
=
0
;
veth_wake_queues
(
cnx
);
/* Drop the lock so we can do stuff that might sleep or
* take other locks. */
spin_unlock_irq
(
&
cnx
->
lock
);
spin_unlock_irq
(
&
cnx
->
lock
);
veth_flush_pending
(
cnx
);
del_timer_sync
(
&
cnx
->
ack_timer
);
del_timer_sync
(
&
cnx
->
reset_timer
);
spin_lock_irq
(
&
cnx
->
lock
);
spin_lock_irq
(
&
cnx
->
lock
);
if
(
cnx
->
state
&
VETH_STATE_RESET
)
if
(
cnx
->
state
&
VETH_STATE_RESET
)
goto
restart
;
goto
restart
;
/* Hack, wait for the other end to reset itself. */
if
(
!
(
cnx
->
state
&
VETH_STATE_SHUTDOWN
))
{
schedule_delayed_work
(
&
cnx
->
statemachine_wq
,
5
*
HZ
);
goto
out
;
}
}
}
if
(
cnx
->
state
&
VETH_STATE_SHUTDOWN
)
if
(
cnx
->
state
&
VETH_STATE_SHUTDOWN
)
...
@@ -488,7 +713,7 @@ static void veth_statemachine(void *p)
...
@@ -488,7 +713,7 @@ static void veth_statemachine(void *p)
if
(
(
cnx
->
state
&
VETH_STATE_OPEN
)
if
(
(
cnx
->
state
&
VETH_STATE_OPEN
)
&&
!
(
cnx
->
state
&
VETH_STATE_SENTMON
)
)
{
&&
!
(
cnx
->
state
&
VETH_STATE_SENTMON
)
)
{
rc
=
veth_signalevent
(
cnx
,
V
ethEventTypeMonitor
,
rc
=
veth_signalevent
(
cnx
,
V
ETH_EVENT_MONITOR
,
HvLpEvent_AckInd_DoAck
,
HvLpEvent_AckInd_DoAck
,
HvLpEvent_AckType_DeferredAck
,
HvLpEvent_AckType_DeferredAck
,
0
,
0
,
0
,
0
,
0
,
0
);
0
,
0
,
0
,
0
,
0
,
0
);
...
@@ -498,9 +723,8 @@ static void veth_statemachine(void *p)
...
@@ -498,9 +723,8 @@ static void veth_statemachine(void *p)
}
else
{
}
else
{
if
(
(
rc
!=
HvLpEvent_Rc_PartitionDead
)
if
(
(
rc
!=
HvLpEvent_Rc_PartitionDead
)
&&
(
rc
!=
HvLpEvent_Rc_PathClosed
)
)
&&
(
rc
!=
HvLpEvent_Rc_PathClosed
)
)
veth_error
(
"Error sending monitor to "
veth_error
(
"Error sending monitor to LPAR %d, "
"lpar %d, rc=%x
\n
"
,
"rc = %d
\n
"
,
rlp
,
rc
);
rlp
,
(
int
)
rc
);
/* Oh well, hope we get a cap from the other
/* Oh well, hope we get a cap from the other
* end and do better when that kicks us */
* end and do better when that kicks us */
...
@@ -512,7 +736,7 @@ static void veth_statemachine(void *p)
...
@@ -512,7 +736,7 @@ static void veth_statemachine(void *p)
&&
!
(
cnx
->
state
&
VETH_STATE_SENTCAPS
))
{
&&
!
(
cnx
->
state
&
VETH_STATE_SENTCAPS
))
{
u64
*
rawcap
=
(
u64
*
)
&
cnx
->
local_caps
;
u64
*
rawcap
=
(
u64
*
)
&
cnx
->
local_caps
;
rc
=
veth_signalevent
(
cnx
,
V
ethEventTypeCap
,
rc
=
veth_signalevent
(
cnx
,
V
ETH_EVENT_CAP
,
HvLpEvent_AckInd_DoAck
,
HvLpEvent_AckInd_DoAck
,
HvLpEvent_AckType_ImmediateAck
,
HvLpEvent_AckType_ImmediateAck
,
0
,
rawcap
[
0
],
rawcap
[
1
],
rawcap
[
2
],
0
,
rawcap
[
0
],
rawcap
[
1
],
rawcap
[
2
],
...
@@ -523,9 +747,9 @@ static void veth_statemachine(void *p)
...
@@ -523,9 +747,9 @@ static void veth_statemachine(void *p)
}
else
{
}
else
{
if
(
(
rc
!=
HvLpEvent_Rc_PartitionDead
)
if
(
(
rc
!=
HvLpEvent_Rc_PartitionDead
)
&&
(
rc
!=
HvLpEvent_Rc_PathClosed
)
)
&&
(
rc
!=
HvLpEvent_Rc_PathClosed
)
)
veth_error
(
"Error sending caps to "
veth_error
(
"Error sending caps to
LPAR %d,
"
"lpar %d, rc=%x
\n
"
,
"rc = %d
\n
"
,
rlp
,
rc
);
rlp
,
(
int
)
rc
);
/* Oh well, hope we get a cap from the other
/* Oh well, hope we get a cap from the other
* end and do better when that kicks us */
* end and do better when that kicks us */
goto
out
;
goto
out
;
...
@@ -534,7 +758,7 @@ static void veth_statemachine(void *p)
...
@@ -534,7 +758,7 @@ static void veth_statemachine(void *p)
if
((
cnx
->
state
&
VETH_STATE_GOTCAPS
)
if
((
cnx
->
state
&
VETH_STATE_GOTCAPS
)
&&
!
(
cnx
->
state
&
VETH_STATE_SENTCAPACK
))
{
&&
!
(
cnx
->
state
&
VETH_STATE_SENTCAPACK
))
{
struct
VethCapD
ata
*
remote_caps
=
&
cnx
->
remote_caps
;
struct
veth_cap_d
ata
*
remote_caps
=
&
cnx
->
remote_caps
;
memcpy
(
remote_caps
,
&
cnx
->
cap_event
.
u
.
caps_data
,
memcpy
(
remote_caps
,
&
cnx
->
cap_event
.
u
.
caps_data
,
sizeof
(
*
remote_caps
));
sizeof
(
*
remote_caps
));
...
@@ -565,10 +789,8 @@ static void veth_statemachine(void *p)
...
@@ -565,10 +789,8 @@ static void veth_statemachine(void *p)
add_timer
(
&
cnx
->
ack_timer
);
add_timer
(
&
cnx
->
ack_timer
);
cnx
->
state
|=
VETH_STATE_READY
;
cnx
->
state
|=
VETH_STATE_READY
;
}
else
{
}
else
{
veth_printk
(
KERN_ERR
,
"Caps rejected (rc=%d) by "
veth_error
(
"Caps rejected by LPAR %d, rc = %d
\n
"
,
"lpar %d
\n
"
,
rlp
,
cnx
->
cap_ack_event
.
base_event
.
xRc
);
cnx
->
cap_ack_event
.
base_event
.
xRc
,
rlp
);
goto
cant_cope
;
goto
cant_cope
;
}
}
}
}
...
@@ -581,8 +803,8 @@ static void veth_statemachine(void *p)
...
@@ -581,8 +803,8 @@ static void veth_statemachine(void *p)
/* FIXME: we get here if something happens we really can't
/* FIXME: we get here if something happens we really can't
* cope with. The link will never work once we get here, and
* cope with. The link will never work once we get here, and
* all we can do is not lock the rest of the system up */
* all we can do is not lock the rest of the system up */
veth_error
(
"
Badness on connection to lpar %d (state=%04lx)
"
veth_error
(
"
Unrecoverable error on connection to LPAR %d, shutting down
"
" - shutting down
\n
"
,
rlp
,
cnx
->
state
);
" (state = 0x%04lx)
\n
"
,
rlp
,
cnx
->
state
);
cnx
->
state
|=
VETH_STATE_SHUTDOWN
;
cnx
->
state
|=
VETH_STATE_SHUTDOWN
;
spin_unlock_irq
(
&
cnx
->
lock
);
spin_unlock_irq
(
&
cnx
->
lock
);
}
}
...
@@ -591,7 +813,7 @@ static int veth_init_connection(u8 rlp)
...
@@ -591,7 +813,7 @@ static int veth_init_connection(u8 rlp)
{
{
struct
veth_lpar_connection
*
cnx
;
struct
veth_lpar_connection
*
cnx
;
struct
veth_msg
*
msgs
;
struct
veth_msg
*
msgs
;
int
i
;
int
i
,
rc
;
if
(
(
rlp
==
this_lp
)
if
(
(
rlp
==
this_lp
)
||
!
HvLpConfig_doLpsCommunicateOnVirtualLan
(
this_lp
,
rlp
)
)
||
!
HvLpConfig_doLpsCommunicateOnVirtualLan
(
this_lp
,
rlp
)
)
...
@@ -605,22 +827,36 @@ static int veth_init_connection(u8 rlp)
...
@@ -605,22 +827,36 @@ static int veth_init_connection(u8 rlp)
cnx
->
remote_lp
=
rlp
;
cnx
->
remote_lp
=
rlp
;
spin_lock_init
(
&
cnx
->
lock
);
spin_lock_init
(
&
cnx
->
lock
);
INIT_WORK
(
&
cnx
->
statemachine_wq
,
veth_statemachine
,
cnx
);
INIT_WORK
(
&
cnx
->
statemachine_wq
,
veth_statemachine
,
cnx
);
init_timer
(
&
cnx
->
ack_timer
);
init_timer
(
&
cnx
->
ack_timer
);
cnx
->
ack_timer
.
function
=
veth_timed_ack
;
cnx
->
ack_timer
.
function
=
veth_timed_ack
;
cnx
->
ack_timer
.
data
=
(
unsigned
long
)
cnx
;
cnx
->
ack_timer
.
data
=
(
unsigned
long
)
cnx
;
init_timer
(
&
cnx
->
reset_timer
);
cnx
->
reset_timer
.
function
=
veth_timed_reset
;
cnx
->
reset_timer
.
data
=
(
unsigned
long
)
cnx
;
cnx
->
reset_timeout
=
5
*
HZ
*
(
VETH_ACKTIMEOUT
/
1000000
);
memset
(
&
cnx
->
pending_acks
,
0xff
,
sizeof
(
cnx
->
pending_acks
));
memset
(
&
cnx
->
pending_acks
,
0xff
,
sizeof
(
cnx
->
pending_acks
));
veth_cnx
[
rlp
]
=
cnx
;
veth_cnx
[
rlp
]
=
cnx
;
/* This gets us 1 reference, which is held on behalf of the driver
* infrastructure. It's released at module unload. */
kobject_init
(
&
cnx
->
kobject
);
cnx
->
kobject
.
ktype
=
&
veth_lpar_connection_ktype
;
rc
=
kobject_set_name
(
&
cnx
->
kobject
,
"cnx%.2d"
,
rlp
);
if
(
rc
!=
0
)
return
rc
;
msgs
=
kmalloc
(
VETH_NUMBUFFERS
*
sizeof
(
struct
veth_msg
),
GFP_KERNEL
);
msgs
=
kmalloc
(
VETH_NUMBUFFERS
*
sizeof
(
struct
veth_msg
),
GFP_KERNEL
);
if
(
!
msgs
)
{
if
(
!
msgs
)
{
veth_error
(
"Can't allocate buffers for
lpar %d
\n
"
,
rlp
);
veth_error
(
"Can't allocate buffers for
LPAR %d.
\n
"
,
rlp
);
return
-
ENOMEM
;
return
-
ENOMEM
;
}
}
cnx
->
msgs
=
msgs
;
cnx
->
msgs
=
msgs
;
memset
(
msgs
,
0
,
VETH_NUMBUFFERS
*
sizeof
(
struct
veth_msg
));
memset
(
msgs
,
0
,
VETH_NUMBUFFERS
*
sizeof
(
struct
veth_msg
));
spin_lock_init
(
&
cnx
->
msg_stack_lock
);
for
(
i
=
0
;
i
<
VETH_NUMBUFFERS
;
i
++
)
{
for
(
i
=
0
;
i
<
VETH_NUMBUFFERS
;
i
++
)
{
msgs
[
i
].
token
=
i
;
msgs
[
i
].
token
=
i
;
...
@@ -630,8 +866,7 @@ static int veth_init_connection(u8 rlp)
...
@@ -630,8 +866,7 @@ static int veth_init_connection(u8 rlp)
cnx
->
num_events
=
veth_allocate_events
(
rlp
,
2
+
VETH_NUMBUFFERS
);
cnx
->
num_events
=
veth_allocate_events
(
rlp
,
2
+
VETH_NUMBUFFERS
);
if
(
cnx
->
num_events
<
(
2
+
VETH_NUMBUFFERS
))
{
if
(
cnx
->
num_events
<
(
2
+
VETH_NUMBUFFERS
))
{
veth_error
(
"Can't allocate events for lpar %d, only got %d
\n
"
,
veth_error
(
"Can't allocate enough events for LPAR %d.
\n
"
,
rlp
);
rlp
,
cnx
->
num_events
);
return
-
ENOMEM
;
return
-
ENOMEM
;
}
}
...
@@ -642,11 +877,9 @@ static int veth_init_connection(u8 rlp)
...
@@ -642,11 +877,9 @@ static int veth_init_connection(u8 rlp)
return
0
;
return
0
;
}
}
static
void
veth_stop_connection
(
u8
rlp
)
static
void
veth_stop_connection
(
struct
veth_lpar_connection
*
cnx
)
{
{
struct
veth_lpar_connection
*
cnx
=
veth_cnx
[
rlp
];
if
(
!
cnx
)
if
(
!
cnx
)
return
;
return
;
spin_lock_irq
(
&
cnx
->
lock
);
spin_lock_irq
(
&
cnx
->
lock
);
...
@@ -654,12 +887,23 @@ static void veth_stop_connection(u8 rlp)
...
@@ -654,12 +887,23 @@ static void veth_stop_connection(u8 rlp)
veth_kick_statemachine
(
cnx
);
veth_kick_statemachine
(
cnx
);
spin_unlock_irq
(
&
cnx
->
lock
);
spin_unlock_irq
(
&
cnx
->
lock
);
/* There's a slim chance the reset code has just queued the
* statemachine to run in five seconds. If so we need to cancel
* that and requeue the work to run now. */
if
(
cancel_delayed_work
(
&
cnx
->
statemachine_wq
))
{
spin_lock_irq
(
&
cnx
->
lock
);
veth_kick_statemachine
(
cnx
);
spin_unlock_irq
(
&
cnx
->
lock
);
}
/* Wait for the state machine to run. */
flush_scheduled_work
();
flush_scheduled_work
();
}
/* FIXME: not sure if this is necessary - will already have
static
void
veth_destroy_connection
(
struct
veth_lpar_connection
*
cnx
)
* been deleted by the state machine, just want to make sure
{
* its not running any more */
if
(
!
cnx
)
del_timer_sync
(
&
cnx
->
ack_timer
)
;
return
;
if
(
cnx
->
num_events
>
0
)
if
(
cnx
->
num_events
>
0
)
mf_deallocate_lp_events
(
cnx
->
remote_lp
,
mf_deallocate_lp_events
(
cnx
->
remote_lp
,
...
@@ -671,18 +915,18 @@ static void veth_stop_connection(u8 rlp)
...
@@ -671,18 +915,18 @@ static void veth_stop_connection(u8 rlp)
HvLpEvent_Type_VirtualLan
,
HvLpEvent_Type_VirtualLan
,
cnx
->
num_ack_events
,
cnx
->
num_ack_events
,
NULL
,
NULL
);
NULL
,
NULL
);
}
static
void
veth_destroy_connection
(
u8
rlp
)
{
struct
veth_lpar_connection
*
cnx
=
veth_cnx
[
rlp
];
if
(
!
cnx
)
return
;
kfree
(
cnx
->
msgs
);
kfree
(
cnx
->
msgs
);
veth_cnx
[
cnx
->
remote_lp
]
=
NULL
;
kfree
(
cnx
);
kfree
(
cnx
);
veth_cnx
[
rlp
]
=
NULL
;
}
static
void
veth_release_connection
(
struct
kobject
*
kobj
)
{
struct
veth_lpar_connection
*
cnx
;
cnx
=
container_of
(
kobj
,
struct
veth_lpar_connection
,
kobject
);
veth_stop_connection
(
cnx
);
veth_destroy_connection
(
cnx
);
}
}
/*
/*
...
@@ -726,17 +970,15 @@ static void veth_set_multicast_list(struct net_device *dev)
...
@@ -726,17 +970,15 @@ static void veth_set_multicast_list(struct net_device *dev)
write_lock_irqsave
(
&
port
->
mcast_gate
,
flags
);
write_lock_irqsave
(
&
port
->
mcast_gate
,
flags
);
if
(
dev
->
flags
&
IFF_PROMISC
)
{
/* set promiscuous mode */
if
((
dev
->
flags
&
IFF_PROMISC
)
||
(
dev
->
flags
&
IFF_ALLMULTI
)
||
printk
(
KERN_INFO
"%s: Promiscuous mode enabled.
\n
"
,
(
dev
->
mc_count
>
VETH_MAX_MCAST
))
{
dev
->
name
);
port
->
promiscuous
=
1
;
port
->
promiscuous
=
1
;
}
else
if
(
(
dev
->
flags
&
IFF_ALLMULTI
)
||
(
dev
->
mc_count
>
VETH_MAX_MCAST
)
)
{
port
->
all_mcast
=
1
;
}
else
{
}
else
{
struct
dev_mc_list
*
dmi
=
dev
->
mc_list
;
struct
dev_mc_list
*
dmi
=
dev
->
mc_list
;
int
i
;
int
i
;
port
->
promiscuous
=
0
;
/* Update table */
/* Update table */
port
->
num_mcast
=
0
;
port
->
num_mcast
=
0
;
...
@@ -758,9 +1000,10 @@ static void veth_set_multicast_list(struct net_device *dev)
...
@@ -758,9 +1000,10 @@ static void veth_set_multicast_list(struct net_device *dev)
static
void
veth_get_drvinfo
(
struct
net_device
*
dev
,
struct
ethtool_drvinfo
*
info
)
static
void
veth_get_drvinfo
(
struct
net_device
*
dev
,
struct
ethtool_drvinfo
*
info
)
{
{
strncpy
(
info
->
driver
,
"veth"
,
sizeof
(
info
->
driver
)
-
1
);
strncpy
(
info
->
driver
,
DRV_NAME
,
sizeof
(
info
->
driver
)
-
1
);
info
->
driver
[
sizeof
(
info
->
driver
)
-
1
]
=
'\0'
;
info
->
driver
[
sizeof
(
info
->
driver
)
-
1
]
=
'\0'
;
strncpy
(
info
->
version
,
"1.0"
,
sizeof
(
info
->
version
)
-
1
);
strncpy
(
info
->
version
,
DRV_VERSION
,
sizeof
(
info
->
version
)
-
1
);
info
->
version
[
sizeof
(
info
->
version
)
-
1
]
=
'\0'
;
}
}
static
int
veth_get_settings
(
struct
net_device
*
dev
,
struct
ethtool_cmd
*
ecmd
)
static
int
veth_get_settings
(
struct
net_device
*
dev
,
struct
ethtool_cmd
*
ecmd
)
...
@@ -791,49 +1034,6 @@ static struct ethtool_ops ops = {
...
@@ -791,49 +1034,6 @@ static struct ethtool_ops ops = {
.
get_link
=
veth_get_link
,
.
get_link
=
veth_get_link
,
};
};
static
void
veth_tx_timeout
(
struct
net_device
*
dev
)
{
struct
veth_port
*
port
=
(
struct
veth_port
*
)
dev
->
priv
;
struct
net_device_stats
*
stats
=
&
port
->
stats
;
unsigned
long
flags
;
int
i
;
stats
->
tx_errors
++
;
spin_lock_irqsave
(
&
port
->
pending_gate
,
flags
);
if
(
!
port
->
pending_lpmask
)
{
spin_unlock_irqrestore
(
&
port
->
pending_gate
,
flags
);
return
;
}
printk
(
KERN_WARNING
"%s: Tx timeout! Resetting lp connections: %08x
\n
"
,
dev
->
name
,
port
->
pending_lpmask
);
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
i
++
)
{
struct
veth_lpar_connection
*
cnx
=
veth_cnx
[
i
];
if
(
!
(
port
->
pending_lpmask
&
(
1
<<
i
)))
continue
;
/* If we're pending on it, we must be connected to it,
* so we should certainly have a structure for it. */
BUG_ON
(
!
cnx
);
/* Theoretically we could be kicking a connection
* which doesn't deserve it, but in practice if we've
* had a Tx timeout, the pending_lpmask will have
* exactly one bit set - the connection causing the
* problem. */
spin_lock
(
&
cnx
->
lock
);
cnx
->
state
|=
VETH_STATE_RESET
;
veth_kick_statemachine
(
cnx
);
spin_unlock
(
&
cnx
->
lock
);
}
spin_unlock_irqrestore
(
&
port
->
pending_gate
,
flags
);
}
static
struct
net_device
*
__init
veth_probe_one
(
int
vlan
,
struct
device
*
vdev
)
static
struct
net_device
*
__init
veth_probe_one
(
int
vlan
,
struct
device
*
vdev
)
{
{
struct
net_device
*
dev
;
struct
net_device
*
dev
;
...
@@ -848,8 +1048,9 @@ static struct net_device * __init veth_probe_one(int vlan, struct device *vdev)
...
@@ -848,8 +1048,9 @@ static struct net_device * __init veth_probe_one(int vlan, struct device *vdev)
port
=
(
struct
veth_port
*
)
dev
->
priv
;
port
=
(
struct
veth_port
*
)
dev
->
priv
;
spin_lock_init
(
&
port
->
pending_gate
);
spin_lock_init
(
&
port
->
queue_lock
);
rwlock_init
(
&
port
->
mcast_gate
);
rwlock_init
(
&
port
->
mcast_gate
);
port
->
stopped_map
=
0
;
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
i
++
)
{
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
i
++
)
{
HvLpVirtualLanIndexMap
map
;
HvLpVirtualLanIndexMap
map
;
...
@@ -882,21 +1083,23 @@ static struct net_device * __init veth_probe_one(int vlan, struct device *vdev)
...
@@ -882,21 +1083,23 @@ static struct net_device * __init veth_probe_one(int vlan, struct device *vdev)
dev
->
set_multicast_list
=
veth_set_multicast_list
;
dev
->
set_multicast_list
=
veth_set_multicast_list
;
SET_ETHTOOL_OPS
(
dev
,
&
ops
);
SET_ETHTOOL_OPS
(
dev
,
&
ops
);
dev
->
watchdog_timeo
=
2
*
(
VETH_ACKTIMEOUT
*
HZ
/
1000000
);
dev
->
tx_timeout
=
veth_tx_timeout
;
SET_NETDEV_DEV
(
dev
,
vdev
);
SET_NETDEV_DEV
(
dev
,
vdev
);
rc
=
register_netdev
(
dev
);
rc
=
register_netdev
(
dev
);
if
(
rc
!=
0
)
{
if
(
rc
!=
0
)
{
veth_printk
(
KERN_ERR
,
veth_error
(
"Failed registering net device for vlan%d.
\n
"
,
vlan
);
"Failed to register ethernet device for vlan %d
\n
"
,
vlan
);
free_netdev
(
dev
);
free_netdev
(
dev
);
return
NULL
;
return
NULL
;
}
}
veth_printk
(
KERN_DEBUG
,
"%s attached to iSeries vlan %d (lpar_map=0x%04x)
\n
"
,
kobject_init
(
&
port
->
kobject
);
port
->
kobject
.
parent
=
&
dev
->
class_dev
.
kobj
;
port
->
kobject
.
ktype
=
&
veth_port_ktype
;
kobject_set_name
(
&
port
->
kobject
,
"veth_port"
);
if
(
0
!=
kobject_add
(
&
port
->
kobject
))
veth_error
(
"Failed adding port for %s to sysfs.
\n
"
,
dev
->
name
);
veth_info
(
"%s attached to iSeries vlan %d (LPAR map = 0x%.4X)
\n
"
,
dev
->
name
,
vlan
,
port
->
lpar_map
);
dev
->
name
,
vlan
,
port
->
lpar_map
);
return
dev
;
return
dev
;
...
@@ -912,98 +1115,95 @@ static int veth_transmit_to_one(struct sk_buff *skb, HvLpIndex rlp,
...
@@ -912,98 +1115,95 @@ static int veth_transmit_to_one(struct sk_buff *skb, HvLpIndex rlp,
struct
veth_lpar_connection
*
cnx
=
veth_cnx
[
rlp
];
struct
veth_lpar_connection
*
cnx
=
veth_cnx
[
rlp
];
struct
veth_port
*
port
=
(
struct
veth_port
*
)
dev
->
priv
;
struct
veth_port
*
port
=
(
struct
veth_port
*
)
dev
->
priv
;
HvLpEvent_Rc
rc
;
HvLpEvent_Rc
rc
;
u32
dma_address
,
dma_length
;
struct
veth_msg
*
msg
=
NULL
;
struct
veth_msg
*
msg
=
NULL
;
int
err
=
0
;
unsigned
long
flags
;
unsigned
long
flags
;
if
(
!
cnx
)
{
if
(
!
cnx
)
port
->
stats
.
tx_errors
++
;
dev_kfree_skb
(
skb
);
return
0
;
return
0
;
}
spin_lock_irqsave
(
&
cnx
->
lock
,
flags
);
spin_lock_irqsave
(
&
cnx
->
lock
,
flags
);
if
(
!
(
cnx
->
state
&
VETH_STATE_READY
))
if
(
!
(
cnx
->
state
&
VETH_STATE_READY
))
goto
drop
;
goto
no_error
;
if
((
skb
->
len
-
14
)
>
VETH_MAX_MTU
)
if
((
skb
->
len
-
ETH_HLEN
)
>
VETH_MAX_MTU
)
goto
drop
;
goto
drop
;
msg
=
veth_stack_pop
(
cnx
);
msg
=
veth_stack_pop
(
cnx
);
if
(
!
msg
)
if
(
!
msg
)
{
err
=
1
;
goto
drop
;
goto
drop
;
}
dma_length
=
skb
->
len
;
msg
->
in_use
=
1
;
dma_address
=
dma_map_single
(
port
->
dev
,
skb
->
data
,
msg
->
skb
=
skb_get
(
skb
);
dma_length
,
DMA_TO_DEVICE
);
if
(
dma_mapping_error
(
dma_address
))
msg
->
data
.
addr
[
0
]
=
dma_map_single
(
port
->
dev
,
skb
->
data
,
skb
->
len
,
DMA_TO_DEVICE
);
if
(
dma_mapping_error
(
msg
->
data
.
addr
[
0
]))
goto
recycle_and_drop
;
goto
recycle_and_drop
;
/* Is it really necessary to check the length and address
* fields of the first entry here? */
msg
->
skb
=
skb
;
msg
->
dev
=
port
->
dev
;
msg
->
dev
=
port
->
dev
;
msg
->
data
.
addr
[
0
]
=
dma_address
;
msg
->
data
.
len
[
0
]
=
skb
->
len
;
msg
->
data
.
len
[
0
]
=
dma_length
;
msg
->
data
.
eofmask
=
1
<<
VETH_EOF_SHIFT
;
msg
->
data
.
eofmask
=
1
<<
VETH_EOF_SHIFT
;
set_bit
(
0
,
&
(
msg
->
in_use
));
rc
=
veth_signaldata
(
cnx
,
V
ethEventTypeFrames
,
msg
->
token
,
&
msg
->
data
);
rc
=
veth_signaldata
(
cnx
,
V
ETH_EVENT_FRAMES
,
msg
->
token
,
&
msg
->
data
);
if
(
rc
!=
HvLpEvent_Rc_Good
)
if
(
rc
!=
HvLpEvent_Rc_Good
)
goto
recycle_and_drop
;
goto
recycle_and_drop
;
/* If the timer's not already running, start it now. */
if
(
0
==
cnx
->
outstanding_tx
)
mod_timer
(
&
cnx
->
reset_timer
,
jiffies
+
cnx
->
reset_timeout
);
cnx
->
last_contact
=
jiffies
;
cnx
->
outstanding_tx
++
;
if
(
veth_stack_is_empty
(
cnx
))
veth_stop_queues
(
cnx
);
no_error:
spin_unlock_irqrestore
(
&
cnx
->
lock
,
flags
);
spin_unlock_irqrestore
(
&
cnx
->
lock
,
flags
);
return
0
;
return
0
;
recycle_and_drop:
recycle_and_drop:
msg
->
skb
=
NULL
;
/* need to set in use to make veth_recycle_msg in case this
* was a mapping failure */
set_bit
(
0
,
&
msg
->
in_use
);
veth_recycle_msg
(
cnx
,
msg
);
veth_recycle_msg
(
cnx
,
msg
);
drop:
drop:
port
->
stats
.
tx_errors
++
;
dev_kfree_skb
(
skb
);
spin_unlock_irqrestore
(
&
cnx
->
lock
,
flags
);
spin_unlock_irqrestore
(
&
cnx
->
lock
,
flags
);
return
err
;
return
1
;
}
}
static
HvLpIndexMap
veth_transmit_to_many
(
struct
sk_buff
*
skb
,
static
void
veth_transmit_to_many
(
struct
sk_buff
*
skb
,
HvLpIndexMap
lpmask
,
HvLpIndexMap
lpmask
,
struct
net_device
*
dev
)
struct
net_device
*
dev
)
{
{
struct
veth_port
*
port
=
(
struct
veth_port
*
)
dev
->
priv
;
struct
veth_port
*
port
=
(
struct
veth_port
*
)
dev
->
priv
;
int
i
;
int
i
,
success
,
error
;
int
rc
;
success
=
error
=
0
;
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
i
++
)
{
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
i
++
)
{
if
((
lpmask
&
(
1
<<
i
))
==
0
)
if
((
lpmask
&
(
1
<<
i
))
==
0
)
continue
;
continue
;
rc
=
veth_transmit_to_one
(
skb_get
(
skb
),
i
,
dev
);
if
(
veth_transmit_to_one
(
skb
,
i
,
dev
))
if
(
!
rc
)
error
=
1
;
lpmask
&=
~
(
1
<<
i
);
else
success
=
1
;
}
}
if
(
!
lpmask
)
{
if
(
error
)
port
->
stats
.
tx_errors
++
;
if
(
success
)
{
port
->
stats
.
tx_packets
++
;
port
->
stats
.
tx_packets
++
;
port
->
stats
.
tx_bytes
+=
skb
->
len
;
port
->
stats
.
tx_bytes
+=
skb
->
len
;
}
}
return
lpmask
;
}
}
static
int
veth_start_xmit
(
struct
sk_buff
*
skb
,
struct
net_device
*
dev
)
static
int
veth_start_xmit
(
struct
sk_buff
*
skb
,
struct
net_device
*
dev
)
{
{
unsigned
char
*
frame
=
skb
->
data
;
unsigned
char
*
frame
=
skb
->
data
;
struct
veth_port
*
port
=
(
struct
veth_port
*
)
dev
->
priv
;
struct
veth_port
*
port
=
(
struct
veth_port
*
)
dev
->
priv
;
unsigned
long
flags
;
HvLpIndexMap
lpmask
;
HvLpIndexMap
lpmask
;
if
(
!
(
frame
[
0
]
&
0x01
))
{
if
(
!
(
frame
[
0
]
&
0x01
))
{
...
@@ -1020,42 +1220,25 @@ static int veth_start_xmit(struct sk_buff *skb, struct net_device *dev)
...
@@ -1020,42 +1220,25 @@ static int veth_start_xmit(struct sk_buff *skb, struct net_device *dev)
lpmask
=
port
->
lpar_map
;
lpmask
=
port
->
lpar_map
;
}
}
spin_lock_irqsave
(
&
port
->
pending_gate
,
flags
);
veth_transmit_to_many
(
skb
,
lpmask
,
dev
);
lpmask
=
veth_transmit_to_many
(
skb
,
lpmask
,
dev
);
dev
->
trans_start
=
jiffies
;
if
(
!
lpmask
)
{
dev_kfree_skb
(
skb
);
}
else
{
if
(
port
->
pending_skb
)
{
veth_error
(
"%s: Tx while skb was pending!
\n
"
,
dev
->
name
);
dev_kfree_skb
(
skb
);
dev_kfree_skb
(
skb
);
spin_unlock_irqrestore
(
&
port
->
pending_gate
,
flags
);
return
1
;
}
port
->
pending_skb
=
skb
;
port
->
pending_lpmask
=
lpmask
;
netif_stop_queue
(
dev
);
}
spin_unlock_irqrestore
(
&
port
->
pending_gate
,
flags
);
return
0
;
return
0
;
}
}
/* You must hold the connection's lock when you call this function. */
static
void
veth_recycle_msg
(
struct
veth_lpar_connection
*
cnx
,
static
void
veth_recycle_msg
(
struct
veth_lpar_connection
*
cnx
,
struct
veth_msg
*
msg
)
struct
veth_msg
*
msg
)
{
{
u32
dma_address
,
dma_length
;
u32
dma_address
,
dma_length
;
if
(
test_and_clear_bit
(
0
,
&
msg
->
in_use
))
{
if
(
msg
->
in_use
)
{
msg
->
in_use
=
0
;
dma_address
=
msg
->
data
.
addr
[
0
];
dma_address
=
msg
->
data
.
addr
[
0
];
dma_length
=
msg
->
data
.
len
[
0
];
dma_length
=
msg
->
data
.
len
[
0
];
if
(
!
dma_mapping_error
(
dma_address
))
dma_unmap_single
(
msg
->
dev
,
dma_address
,
dma_length
,
dma_unmap_single
(
msg
->
dev
,
dma_address
,
dma_length
,
DMA_TO_DEVICE
);
DMA_TO_DEVICE
);
...
@@ -1066,15 +1249,16 @@ static void veth_recycle_msg(struct veth_lpar_connection *cnx,
...
@@ -1066,15 +1249,16 @@ static void veth_recycle_msg(struct veth_lpar_connection *cnx,
memset
(
&
msg
->
data
,
0
,
sizeof
(
msg
->
data
));
memset
(
&
msg
->
data
,
0
,
sizeof
(
msg
->
data
));
veth_stack_push
(
cnx
,
msg
);
veth_stack_push
(
cnx
,
msg
);
}
else
}
else
if
(
cnx
->
state
&
VETH_STATE_OPEN
)
{
if
(
cnx
->
state
&
VETH_STATE_OPEN
)
veth_error
(
"Non-pending frame (# %d) acked by LPAR %d.
\n
"
,
veth_error
(
"Bogus frames ack from lpar %d (#%d)
\n
"
,
cnx
->
remote_lp
,
msg
->
token
);
cnx
->
remote_lp
,
msg
->
token
);
}
}
}
static
void
veth_
flush_pending
(
struct
veth_lpar_connection
*
cnx
)
static
void
veth_
wake_queues
(
struct
veth_lpar_connection
*
cnx
)
{
{
int
i
;
int
i
;
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDVIRTUALLANS
;
i
++
)
{
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDVIRTUALLANS
;
i
++
)
{
struct
net_device
*
dev
=
veth_dev
[
i
];
struct
net_device
*
dev
=
veth_dev
[
i
];
struct
veth_port
*
port
;
struct
veth_port
*
port
;
...
@@ -1088,22 +1272,79 @@ static void veth_flush_pending(struct veth_lpar_connection *cnx)
...
@@ -1088,22 +1272,79 @@ static void veth_flush_pending(struct veth_lpar_connection *cnx)
if
(
!
(
port
->
lpar_map
&
(
1
<<
cnx
->
remote_lp
)))
if
(
!
(
port
->
lpar_map
&
(
1
<<
cnx
->
remote_lp
)))
continue
;
continue
;
spin_lock_irqsave
(
&
port
->
pending_gate
,
flags
);
spin_lock_irqsave
(
&
port
->
queue_lock
,
flags
);
if
(
port
->
pending_skb
)
{
port
->
pending_lpmask
=
port
->
stopped_map
&=
~
(
1
<<
cnx
->
remote_lp
);
veth_transmit_to_many
(
port
->
pending_skb
,
port
->
pending_lpmask
,
if
(
0
==
port
->
stopped_map
&&
netif_queue_stopped
(
dev
))
{
dev
);
veth_debug
(
"cnx %d: woke queue for %s.
\n
"
,
if
(
!
port
->
pending_lpmask
)
{
cnx
->
remote_lp
,
dev
->
name
);
dev_kfree_skb_any
(
port
->
pending_skb
);
port
->
pending_skb
=
NULL
;
netif_wake_queue
(
dev
);
netif_wake_queue
(
dev
);
}
}
spin_unlock_irqrestore
(
&
port
->
queue_lock
,
flags
);
}
}
spin_unlock_irqrestore
(
&
port
->
pending_gate
,
flags
);
}
static
void
veth_stop_queues
(
struct
veth_lpar_connection
*
cnx
)
{
int
i
;
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDVIRTUALLANS
;
i
++
)
{
struct
net_device
*
dev
=
veth_dev
[
i
];
struct
veth_port
*
port
;
if
(
!
dev
)
continue
;
port
=
(
struct
veth_port
*
)
dev
->
priv
;
/* If this cnx is not on the vlan for this port, continue */
if
(
!
(
port
->
lpar_map
&
(
1
<<
cnx
->
remote_lp
)))
continue
;
spin_lock
(
&
port
->
queue_lock
);
netif_stop_queue
(
dev
);
port
->
stopped_map
|=
(
1
<<
cnx
->
remote_lp
);
veth_debug
(
"cnx %d: stopped queue for %s, map = 0x%x.
\n
"
,
cnx
->
remote_lp
,
dev
->
name
,
port
->
stopped_map
);
spin_unlock
(
&
port
->
queue_lock
);
}
}
}
}
static
void
veth_timed_reset
(
unsigned
long
ptr
)
{
struct
veth_lpar_connection
*
cnx
=
(
struct
veth_lpar_connection
*
)
ptr
;
unsigned
long
trigger_time
,
flags
;
/* FIXME is it possible this fires after veth_stop_connection()?
* That would reschedule the statemachine for 5 seconds and probably
* execute it after the module's been unloaded. Hmm. */
spin_lock_irqsave
(
&
cnx
->
lock
,
flags
);
if
(
cnx
->
outstanding_tx
>
0
)
{
trigger_time
=
cnx
->
last_contact
+
cnx
->
reset_timeout
;
if
(
trigger_time
<
jiffies
)
{
cnx
->
state
|=
VETH_STATE_RESET
;
veth_kick_statemachine
(
cnx
);
veth_error
(
"%d packets not acked by LPAR %d within %d "
"seconds, resetting.
\n
"
,
cnx
->
outstanding_tx
,
cnx
->
remote_lp
,
cnx
->
reset_timeout
/
HZ
);
}
else
{
/* Reschedule the timer */
trigger_time
=
jiffies
+
cnx
->
reset_timeout
;
mod_timer
(
&
cnx
->
reset_timer
,
trigger_time
);
}
}
spin_unlock_irqrestore
(
&
cnx
->
lock
,
flags
);
}
/*
/*
* Rx path
* Rx path
*/
*/
...
@@ -1117,12 +1358,9 @@ static inline int veth_frame_wanted(struct veth_port *port, u64 mac_addr)
...
@@ -1117,12 +1358,9 @@ static inline int veth_frame_wanted(struct veth_port *port, u64 mac_addr)
if
(
(
mac_addr
==
port
->
mac_addr
)
||
(
mac_addr
==
0xffffffffffff0000
)
)
if
(
(
mac_addr
==
port
->
mac_addr
)
||
(
mac_addr
==
0xffffffffffff0000
)
)
return
1
;
return
1
;
if
(
!
(((
char
*
)
&
mac_addr
)[
0
]
&
0x01
))
return
0
;
read_lock_irqsave
(
&
port
->
mcast_gate
,
flags
);
read_lock_irqsave
(
&
port
->
mcast_gate
,
flags
);
if
(
port
->
promiscuous
||
port
->
all_mcast
)
{
if
(
port
->
promiscuous
)
{
wanted
=
1
;
wanted
=
1
;
goto
out
;
goto
out
;
}
}
...
@@ -1175,21 +1413,21 @@ static void veth_flush_acks(struct veth_lpar_connection *cnx)
...
@@ -1175,21 +1413,21 @@ static void veth_flush_acks(struct veth_lpar_connection *cnx)
{
{
HvLpEvent_Rc
rc
;
HvLpEvent_Rc
rc
;
rc
=
veth_signaldata
(
cnx
,
V
ethEventTypeFramesAck
,
rc
=
veth_signaldata
(
cnx
,
V
ETH_EVENT_FRAMES_ACK
,
0
,
&
cnx
->
pending_acks
);
0
,
&
cnx
->
pending_acks
);
if
(
rc
!=
HvLpEvent_Rc_Good
)
if
(
rc
!=
HvLpEvent_Rc_Good
)
veth_error
(
"
Error 0x%x acking frames from lpar %d!
\n
"
,
veth_error
(
"
Failed acking frames from LPAR %d, rc = %d
\n
"
,
(
unsigned
)
rc
,
cnx
->
remote_lp
);
cnx
->
remote_lp
,
(
int
)
rc
);
cnx
->
num_pending_acks
=
0
;
cnx
->
num_pending_acks
=
0
;
memset
(
&
cnx
->
pending_acks
,
0xff
,
sizeof
(
cnx
->
pending_acks
));
memset
(
&
cnx
->
pending_acks
,
0xff
,
sizeof
(
cnx
->
pending_acks
));
}
}
static
void
veth_receive
(
struct
veth_lpar_connection
*
cnx
,
static
void
veth_receive
(
struct
veth_lpar_connection
*
cnx
,
struct
VethLpE
vent
*
event
)
struct
veth_lpe
vent
*
event
)
{
{
struct
VethFramesD
ata
*
senddata
=
&
event
->
u
.
frames_data
;
struct
veth_frames_d
ata
*
senddata
=
&
event
->
u
.
frames_data
;
int
startchunk
=
0
;
int
startchunk
=
0
;
int
nchunks
;
int
nchunks
;
unsigned
long
flags
;
unsigned
long
flags
;
...
@@ -1216,9 +1454,10 @@ static void veth_receive(struct veth_lpar_connection *cnx,
...
@@ -1216,9 +1454,10 @@ static void veth_receive(struct veth_lpar_connection *cnx,
/* make sure that we have at least 1 EOF entry in the
/* make sure that we have at least 1 EOF entry in the
* remaining entries */
* remaining entries */
if
(
!
(
senddata
->
eofmask
>>
(
startchunk
+
VETH_EOF_SHIFT
)))
{
if
(
!
(
senddata
->
eofmask
>>
(
startchunk
+
VETH_EOF_SHIFT
)))
{
veth_error
(
"missing EOF frag in event "
veth_error
(
"Missing EOF fragment in event "
"eofmask=0x%x startchunk=%d
\n
"
,
"eofmask = 0x%x startchunk = %d
\n
"
,
(
unsigned
)
senddata
->
eofmask
,
startchunk
);
(
unsigned
)
senddata
->
eofmask
,
startchunk
);
break
;
break
;
}
}
...
@@ -1237,8 +1476,9 @@ static void veth_receive(struct veth_lpar_connection *cnx,
...
@@ -1237,8 +1476,9 @@ static void veth_receive(struct veth_lpar_connection *cnx,
/* nchunks == # of chunks in this frame */
/* nchunks == # of chunks in this frame */
if
((
length
-
ETH_HLEN
)
>
VETH_MAX_MTU
)
{
if
((
length
-
ETH_HLEN
)
>
VETH_MAX_MTU
)
{
veth_error
(
"Received oversize frame from lpar %d "
veth_error
(
"Received oversize frame from LPAR %d "
"(length=%d)
\n
"
,
cnx
->
remote_lp
,
length
);
"(length = %d)
\n
"
,
cnx
->
remote_lp
,
length
);
continue
;
continue
;
}
}
...
@@ -1331,15 +1571,33 @@ static void veth_timed_ack(unsigned long ptr)
...
@@ -1331,15 +1571,33 @@ static void veth_timed_ack(unsigned long ptr)
static
int
veth_remove
(
struct
vio_dev
*
vdev
)
static
int
veth_remove
(
struct
vio_dev
*
vdev
)
{
{
int
i
=
vdev
->
unit_address
;
struct
veth_lpar_connection
*
cnx
;
struct
net_device
*
dev
;
struct
net_device
*
dev
;
struct
veth_port
*
port
;
int
i
;
dev
=
veth_dev
[
vdev
->
unit_address
];
if
(
!
dev
)
return
0
;
port
=
netdev_priv
(
dev
);
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
i
++
)
{
cnx
=
veth_cnx
[
i
];
dev
=
veth_dev
[
i
];
if
(
cnx
&&
(
port
->
lpar_map
&
(
1
<<
i
)))
{
if
(
dev
!=
NULL
)
{
/* Drop our reference to connections on our VLAN */
veth_dev
[
i
]
=
NULL
;
kobject_put
(
&
cnx
->
kobject
);
}
}
veth_dev
[
vdev
->
unit_address
]
=
NULL
;
kobject_del
(
&
port
->
kobject
);
kobject_put
(
&
port
->
kobject
);
unregister_netdev
(
dev
);
unregister_netdev
(
dev
);
free_netdev
(
dev
);
free_netdev
(
dev
);
}
return
0
;
return
0
;
}
}
...
@@ -1347,6 +1605,7 @@ static int veth_probe(struct vio_dev *vdev, const struct vio_device_id *id)
...
@@ -1347,6 +1605,7 @@ static int veth_probe(struct vio_dev *vdev, const struct vio_device_id *id)
{
{
int
i
=
vdev
->
unit_address
;
int
i
=
vdev
->
unit_address
;
struct
net_device
*
dev
;
struct
net_device
*
dev
;
struct
veth_port
*
port
;
dev
=
veth_probe_one
(
i
,
&
vdev
->
dev
);
dev
=
veth_probe_one
(
i
,
&
vdev
->
dev
);
if
(
dev
==
NULL
)
{
if
(
dev
==
NULL
)
{
...
@@ -1355,11 +1614,23 @@ static int veth_probe(struct vio_dev *vdev, const struct vio_device_id *id)
...
@@ -1355,11 +1614,23 @@ static int veth_probe(struct vio_dev *vdev, const struct vio_device_id *id)
}
}
veth_dev
[
i
]
=
dev
;
veth_dev
[
i
]
=
dev
;
/* Start the state machine on each connection, to commence
port
=
(
struct
veth_port
*
)
netdev_priv
(
dev
);
* link negotiation */
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
i
++
)
/* Start the state machine on each connection on this vlan. If we're
if
(
veth_cnx
[
i
])
* the first dev to do so this will commence link negotiation */
veth_kick_statemachine
(
veth_cnx
[
i
]);
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
i
++
)
{
struct
veth_lpar_connection
*
cnx
;
if
(
!
(
port
->
lpar_map
&
(
1
<<
i
)))
continue
;
cnx
=
veth_cnx
[
i
];
if
(
!
cnx
)
continue
;
kobject_get
(
&
cnx
->
kobject
);
veth_kick_statemachine
(
cnx
);
}
return
0
;
return
0
;
}
}
...
@@ -1375,7 +1646,7 @@ static struct vio_device_id veth_device_table[] __devinitdata = {
...
@@ -1375,7 +1646,7 @@ static struct vio_device_id veth_device_table[] __devinitdata = {
MODULE_DEVICE_TABLE
(
vio
,
veth_device_table
);
MODULE_DEVICE_TABLE
(
vio
,
veth_device_table
);
static
struct
vio_driver
veth_driver
=
{
static
struct
vio_driver
veth_driver
=
{
.
name
=
"iseries_veth"
,
.
name
=
DRV_NAME
,
.
id_table
=
veth_device_table
,
.
id_table
=
veth_device_table
,
.
probe
=
veth_probe
,
.
probe
=
veth_probe
,
.
remove
=
veth_remove
.
remove
=
veth_remove
...
@@ -1388,29 +1659,29 @@ static struct vio_driver veth_driver = {
...
@@ -1388,29 +1659,29 @@ static struct vio_driver veth_driver = {
void
__exit
veth_module_cleanup
(
void
)
void
__exit
veth_module_cleanup
(
void
)
{
{
int
i
;
int
i
;
struct
veth_lpar_connection
*
cnx
;
/* Stop the queues first to stop any new packets being sent. */
/* Disconnect our "irq" to stop events coming from the Hypervisor. */
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDVIRTUALLANS
;
i
++
)
if
(
veth_dev
[
i
])
netif_stop_queue
(
veth_dev
[
i
]);
/* Stop the connections before we unregister the driver. This
* ensures there's no skbs lying around holding the device open. */
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
++
i
)
veth_stop_connection
(
i
);
HvLpEvent_unregisterHandler
(
HvLpEvent_Type_VirtualLan
);
HvLpEvent_unregisterHandler
(
HvLpEvent_Type_VirtualLan
);
/* Hypervisor callbacks may have scheduled more work while we
/* Make sure any work queued from Hypervisor callbacks is finished. */
* were stoping connections. Now that we've disconnected from
* the hypervisor make sure everything's finished. */
flush_scheduled_work
();
flush_scheduled_work
();
vio_unregister_driver
(
&
veth_driver
);
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
++
i
)
{
cnx
=
veth_cnx
[
i
];
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
++
i
)
if
(
!
cnx
)
veth_destroy_connection
(
i
);
continue
;
/* Remove the connection from sysfs */
kobject_del
(
&
cnx
->
kobject
);
/* Drop the driver's reference to the connection */
kobject_put
(
&
cnx
->
kobject
);
}
/* Unregister the driver, which will close all the netdevs and stop
* the connections when they're no longer referenced. */
vio_unregister_driver
(
&
veth_driver
);
}
}
module_exit
(
veth_module_cleanup
);
module_exit
(
veth_module_cleanup
);
...
@@ -1423,15 +1694,37 @@ int __init veth_module_init(void)
...
@@ -1423,15 +1694,37 @@ int __init veth_module_init(void)
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
++
i
)
{
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
++
i
)
{
rc
=
veth_init_connection
(
i
);
rc
=
veth_init_connection
(
i
);
if
(
rc
!=
0
)
{
if
(
rc
!=
0
)
veth_module_cleanup
();
goto
error
;
return
rc
;
}
}
}
HvLpEvent_registerHandler
(
HvLpEvent_Type_VirtualLan
,
HvLpEvent_registerHandler
(
HvLpEvent_Type_VirtualLan
,
&
veth_handle_event
);
&
veth_handle_event
);
return
vio_register_driver
(
&
veth_driver
);
rc
=
vio_register_driver
(
&
veth_driver
);
if
(
rc
!=
0
)
goto
error
;
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
++
i
)
{
struct
kobject
*
kobj
;
if
(
!
veth_cnx
[
i
])
continue
;
kobj
=
&
veth_cnx
[
i
]
->
kobject
;
kobj
->
parent
=
&
veth_driver
.
driver
.
kobj
;
/* If the add failes, complain but otherwise continue */
if
(
0
!=
kobject_add
(
kobj
))
veth_error
(
"cnx %d: Failed adding to sysfs.
\n
"
,
i
);
}
return
0
;
error:
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
++
i
)
{
veth_destroy_connection
(
veth_cnx
[
i
]);
}
return
rc
;
}
}
module_init
(
veth_module_init
);
module_init
(
veth_module_init
);
drivers/net/iseries_veth.h
deleted
100644 → 0
View file @
712fbdd3
/* File veth.h created by Kyle A. Lucke on Mon Aug 7 2000. */
#ifndef _ISERIES_VETH_H
#define _ISERIES_VETH_H
#define VethEventTypeCap (0)
#define VethEventTypeFrames (1)
#define VethEventTypeMonitor (2)
#define VethEventTypeFramesAck (3)
#define VETH_MAX_ACKS_PER_MSG (20)
#define VETH_MAX_FRAMES_PER_MSG (6)
struct
VethFramesData
{
u32
addr
[
VETH_MAX_FRAMES_PER_MSG
];
u16
len
[
VETH_MAX_FRAMES_PER_MSG
];
u32
eofmask
;
};
#define VETH_EOF_SHIFT (32-VETH_MAX_FRAMES_PER_MSG)
struct
VethFramesAckData
{
u16
token
[
VETH_MAX_ACKS_PER_MSG
];
};
struct
VethCapData
{
u8
caps_version
;
u8
rsvd1
;
u16
num_buffers
;
u16
ack_threshold
;
u16
rsvd2
;
u32
ack_timeout
;
u32
rsvd3
;
u64
rsvd4
[
3
];
};
struct
VethLpEvent
{
struct
HvLpEvent
base_event
;
union
{
struct
VethCapData
caps_data
;
struct
VethFramesData
frames_data
;
struct
VethFramesAckData
frames_ack_data
;
}
u
;
};
#endif
/* _ISERIES_VETH_H */
drivers/net/tulip/de2104x.c
View file @
44757223
...
@@ -1934,7 +1934,7 @@ static int __init de_init_one (struct pci_dev *pdev,
...
@@ -1934,7 +1934,7 @@ static int __init de_init_one (struct pci_dev *pdev,
struct
de_private
*
de
;
struct
de_private
*
de
;
int
rc
;
int
rc
;
void
__iomem
*
regs
;
void
__iomem
*
regs
;
long
pciaddr
;
unsigned
long
pciaddr
;
static
int
board_idx
=
-
1
;
static
int
board_idx
=
-
1
;
board_idx
++
;
board_idx
++
;
...
...
drivers/net/tulip/tulip_core.c
View file @
44757223
...
@@ -238,6 +238,7 @@ static struct pci_device_id tulip_pci_tbl[] = {
...
@@ -238,6 +238,7 @@ static struct pci_device_id tulip_pci_tbl[] = {
{
0x17B3
,
0xAB08
,
PCI_ANY_ID
,
PCI_ANY_ID
,
0
,
0
,
COMET
},
{
0x17B3
,
0xAB08
,
PCI_ANY_ID
,
PCI_ANY_ID
,
0
,
0
,
COMET
},
{
0x10b7
,
0x9300
,
PCI_ANY_ID
,
PCI_ANY_ID
,
0
,
0
,
COMET
},
/* 3Com 3CSOHO100B-TX */
{
0x10b7
,
0x9300
,
PCI_ANY_ID
,
PCI_ANY_ID
,
0
,
0
,
COMET
},
/* 3Com 3CSOHO100B-TX */
{
0x14ea
,
0xab08
,
PCI_ANY_ID
,
PCI_ANY_ID
,
0
,
0
,
COMET
},
/* Planex FNW-3602-TX */
{
0x14ea
,
0xab08
,
PCI_ANY_ID
,
PCI_ANY_ID
,
0
,
0
,
COMET
},
/* Planex FNW-3602-TX */
{
0x1414
,
0x0002
,
PCI_ANY_ID
,
PCI_ANY_ID
,
0
,
0
,
COMET
},
{
}
/* terminate list */
{
}
/* terminate list */
};
};
MODULE_DEVICE_TABLE
(
pci
,
tulip_pci_tbl
);
MODULE_DEVICE_TABLE
(
pci
,
tulip_pci_tbl
);
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment