Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in
Toggle navigation
L
linux-davinci-2.6.23
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Redmine
Redmine
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Operations
Operations
Metrics
Environments
Analytics
Analytics
CI / CD
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
linux
linux-davinci-2.6.23
Commits
44757223
Commit
44757223
authored
Sep 02, 2005
by
Linus Torvalds
Browse files
Options
Browse Files
Download
Plain Diff
Merge refs/heads/upstream from master.kernel.org:/pub/scm/linux/kernel/git/jgarzik/netdev-2.6
parents
712fbdd3
ee05f031
Changes
4
Show whitespace changes
Inline
Side-by-side
Showing
4 changed files
with
583 additions
and
335 deletions
+583
-335
drivers/net/iseries_veth.c
drivers/net/iseries_veth.c
+581
-288
drivers/net/iseries_veth.h
drivers/net/iseries_veth.h
+0
-46
drivers/net/tulip/de2104x.c
drivers/net/tulip/de2104x.c
+1
-1
drivers/net/tulip/tulip_core.c
drivers/net/tulip/tulip_core.c
+1
-0
No files found.
drivers/net/iseries_veth.c
View file @
44757223
...
@@ -79,12 +79,55 @@
...
@@ -79,12 +79,55 @@
#include <asm/iommu.h>
#include <asm/iommu.h>
#include <asm/vio.h>
#include <asm/vio.h>
#
include "iseries_veth.h"
#
undef DEBUG
MODULE_AUTHOR
(
"Kyle Lucke <klucke@us.ibm.com>"
);
MODULE_AUTHOR
(
"Kyle Lucke <klucke@us.ibm.com>"
);
MODULE_DESCRIPTION
(
"iSeries Virtual ethernet driver"
);
MODULE_DESCRIPTION
(
"iSeries Virtual ethernet driver"
);
MODULE_LICENSE
(
"GPL"
);
MODULE_LICENSE
(
"GPL"
);
#define VETH_EVENT_CAP (0)
#define VETH_EVENT_FRAMES (1)
#define VETH_EVENT_MONITOR (2)
#define VETH_EVENT_FRAMES_ACK (3)
#define VETH_MAX_ACKS_PER_MSG (20)
#define VETH_MAX_FRAMES_PER_MSG (6)
struct
veth_frames_data
{
u32
addr
[
VETH_MAX_FRAMES_PER_MSG
];
u16
len
[
VETH_MAX_FRAMES_PER_MSG
];
u32
eofmask
;
};
#define VETH_EOF_SHIFT (32-VETH_MAX_FRAMES_PER_MSG)
struct
veth_frames_ack_data
{
u16
token
[
VETH_MAX_ACKS_PER_MSG
];
};
struct
veth_cap_data
{
u8
caps_version
;
u8
rsvd1
;
u16
num_buffers
;
u16
ack_threshold
;
u16
rsvd2
;
u32
ack_timeout
;
u32
rsvd3
;
u64
rsvd4
[
3
];
};
struct
veth_lpevent
{
struct
HvLpEvent
base_event
;
union
{
struct
veth_cap_data
caps_data
;
struct
veth_frames_data
frames_data
;
struct
veth_frames_ack_data
frames_ack_data
;
}
u
;
};
#define DRV_NAME "iseries_veth"
#define DRV_VERSION "2.0"
#define VETH_NUMBUFFERS (120)
#define VETH_NUMBUFFERS (120)
#define VETH_ACKTIMEOUT (1000000)
/* microseconds */
#define VETH_ACKTIMEOUT (1000000)
/* microseconds */
#define VETH_MAX_MCAST (12)
#define VETH_MAX_MCAST (12)
...
@@ -113,9 +156,9 @@ MODULE_LICENSE("GPL");
...
@@ -113,9 +156,9 @@ MODULE_LICENSE("GPL");
struct
veth_msg
{
struct
veth_msg
{
struct
veth_msg
*
next
;
struct
veth_msg
*
next
;
struct
VethFramesD
ata
data
;
struct
veth_frames_d
ata
data
;
int
token
;
int
token
;
unsigned
long
in_use
;
int
in_use
;
struct
sk_buff
*
skb
;
struct
sk_buff
*
skb
;
struct
device
*
dev
;
struct
device
*
dev
;
};
};
...
@@ -125,23 +168,28 @@ struct veth_lpar_connection {
...
@@ -125,23 +168,28 @@ struct veth_lpar_connection {
struct
work_struct
statemachine_wq
;
struct
work_struct
statemachine_wq
;
struct
veth_msg
*
msgs
;
struct
veth_msg
*
msgs
;
int
num_events
;
int
num_events
;
struct
VethCapD
ata
local_caps
;
struct
veth_cap_d
ata
local_caps
;
struct
kobject
kobject
;
struct
timer_list
ack_timer
;
struct
timer_list
ack_timer
;
struct
timer_list
reset_timer
;
unsigned
int
reset_timeout
;
unsigned
long
last_contact
;
int
outstanding_tx
;
spinlock_t
lock
;
spinlock_t
lock
;
unsigned
long
state
;
unsigned
long
state
;
HvLpInstanceId
src_inst
;
HvLpInstanceId
src_inst
;
HvLpInstanceId
dst_inst
;
HvLpInstanceId
dst_inst
;
struct
VethLpE
vent
cap_event
,
cap_ack_event
;
struct
veth_lpe
vent
cap_event
,
cap_ack_event
;
u16
pending_acks
[
VETH_MAX_ACKS_PER_MSG
];
u16
pending_acks
[
VETH_MAX_ACKS_PER_MSG
];
u32
num_pending_acks
;
u32
num_pending_acks
;
int
num_ack_events
;
int
num_ack_events
;
struct
VethCapD
ata
remote_caps
;
struct
veth_cap_d
ata
remote_caps
;
u32
ack_timeout
;
u32
ack_timeout
;
spinlock_t
msg_stack_lock
;
struct
veth_msg
*
msg_stack_head
;
struct
veth_msg
*
msg_stack_head
;
};
};
...
@@ -151,15 +199,17 @@ struct veth_port {
...
@@ -151,15 +199,17 @@ struct veth_port {
u64
mac_addr
;
u64
mac_addr
;
HvLpIndexMap
lpar_map
;
HvLpIndexMap
lpar_map
;
spinlock_t
pending_gate
;
/* queue_lock protects the stopped_map and dev's queue. */
s
truct
sk_buff
*
pending_skb
;
s
pinlock_t
queue_lock
;
HvLpIndexMap
pending_lpmask
;
HvLpIndexMap
stopped_map
;
/* mcast_gate protects promiscuous, num_mcast & mcast_addr. */
rwlock_t
mcast_gate
;
rwlock_t
mcast_gate
;
int
promiscuous
;
int
promiscuous
;
int
all_mcast
;
int
num_mcast
;
int
num_mcast
;
u64
mcast_addr
[
VETH_MAX_MCAST
];
u64
mcast_addr
[
VETH_MAX_MCAST
];
struct
kobject
kobject
;
};
};
static
HvLpIndex
this_lp
;
static
HvLpIndex
this_lp
;
...
@@ -168,44 +218,56 @@ static struct net_device *veth_dev[HVMAXARCHITECTEDVIRTUALLANS]; /* = 0 */
...
@@ -168,44 +218,56 @@ static struct net_device *veth_dev[HVMAXARCHITECTEDVIRTUALLANS]; /* = 0 */
static
int
veth_start_xmit
(
struct
sk_buff
*
skb
,
struct
net_device
*
dev
);
static
int
veth_start_xmit
(
struct
sk_buff
*
skb
,
struct
net_device
*
dev
);
static
void
veth_recycle_msg
(
struct
veth_lpar_connection
*
,
struct
veth_msg
*
);
static
void
veth_recycle_msg
(
struct
veth_lpar_connection
*
,
struct
veth_msg
*
);
static
void
veth_flush_pending
(
struct
veth_lpar_connection
*
cnx
);
static
void
veth_wake_queues
(
struct
veth_lpar_connection
*
cnx
);
static
void
veth_receive
(
struct
veth_lpar_connection
*
,
struct
VethLpEvent
*
);
static
void
veth_stop_queues
(
struct
veth_lpar_connection
*
cnx
);
static
void
veth_timed_ack
(
unsigned
long
connectionPtr
);
static
void
veth_receive
(
struct
veth_lpar_connection
*
,
struct
veth_lpevent
*
);
static
void
veth_release_connection
(
struct
kobject
*
kobject
);
static
void
veth_timed_ack
(
unsigned
long
ptr
);
static
void
veth_timed_reset
(
unsigned
long
ptr
);
/*
/*
* Utility functions
* Utility functions
*/
*/
#define veth_
printk(prio,
fmt, args...) \
#define veth_
info(
fmt, args...) \
printk(
prio "%s: " fmt, __FILE__
, ## args)
printk(
KERN_INFO DRV_NAME ": " fmt
, ## args)
#define veth_error(fmt, args...) \
#define veth_error(fmt, args...) \
printk(KERN_ERR "(%s:%3.3d) ERROR: " fmt, __FILE__, __LINE__ , ## args)
printk(KERN_ERR DRV_NAME ": Error: " fmt, ## args)
#ifdef DEBUG
#define veth_debug(fmt, args...) \
printk(KERN_DEBUG DRV_NAME ": " fmt, ## args)
#else
#define veth_debug(fmt, args...) do {} while (0)
#endif
/* You must hold the connection's lock when you call this function. */
static
inline
void
veth_stack_push
(
struct
veth_lpar_connection
*
cnx
,
static
inline
void
veth_stack_push
(
struct
veth_lpar_connection
*
cnx
,
struct
veth_msg
*
msg
)
struct
veth_msg
*
msg
)
{
{
unsigned
long
flags
;
spin_lock_irqsave
(
&
cnx
->
msg_stack_lock
,
flags
);
msg
->
next
=
cnx
->
msg_stack_head
;
msg
->
next
=
cnx
->
msg_stack_head
;
cnx
->
msg_stack_head
=
msg
;
cnx
->
msg_stack_head
=
msg
;
spin_unlock_irqrestore
(
&
cnx
->
msg_stack_lock
,
flags
);
}
}
/* You must hold the connection's lock when you call this function. */
static
inline
struct
veth_msg
*
veth_stack_pop
(
struct
veth_lpar_connection
*
cnx
)
static
inline
struct
veth_msg
*
veth_stack_pop
(
struct
veth_lpar_connection
*
cnx
)
{
{
unsigned
long
flags
;
struct
veth_msg
*
msg
;
struct
veth_msg
*
msg
;
spin_lock_irqsave
(
&
cnx
->
msg_stack_lock
,
flags
);
msg
=
cnx
->
msg_stack_head
;
msg
=
cnx
->
msg_stack_head
;
if
(
msg
)
if
(
msg
)
cnx
->
msg_stack_head
=
cnx
->
msg_stack_head
->
next
;
cnx
->
msg_stack_head
=
cnx
->
msg_stack_head
->
next
;
spin_unlock_irqrestore
(
&
cnx
->
msg_stack_lock
,
flags
);
return
msg
;
return
msg
;
}
}
/* You must hold the connection's lock when you call this function. */
static
inline
int
veth_stack_is_empty
(
struct
veth_lpar_connection
*
cnx
)
{
return
cnx
->
msg_stack_head
==
NULL
;
}
static
inline
HvLpEvent_Rc
static
inline
HvLpEvent_Rc
veth_signalevent
(
struct
veth_lpar_connection
*
cnx
,
u16
subtype
,
veth_signalevent
(
struct
veth_lpar_connection
*
cnx
,
u16
subtype
,
HvLpEvent_AckInd
ackind
,
HvLpEvent_AckType
acktype
,
HvLpEvent_AckInd
ackind
,
HvLpEvent_AckType
acktype
,
...
@@ -249,13 +311,144 @@ static int veth_allocate_events(HvLpIndex rlp, int number)
...
@@ -249,13 +311,144 @@ static int veth_allocate_events(HvLpIndex rlp, int number)
struct
veth_allocation
vc
=
{
COMPLETION_INITIALIZER
(
vc
.
c
),
0
};
struct
veth_allocation
vc
=
{
COMPLETION_INITIALIZER
(
vc
.
c
),
0
};
mf_allocate_lp_events
(
rlp
,
HvLpEvent_Type_VirtualLan
,
mf_allocate_lp_events
(
rlp
,
HvLpEvent_Type_VirtualLan
,
sizeof
(
struct
VethLpE
vent
),
number
,
sizeof
(
struct
veth_lpe
vent
),
number
,
&
veth_complete_allocation
,
&
vc
);
&
veth_complete_allocation
,
&
vc
);
wait_for_completion
(
&
vc
.
c
);
wait_for_completion
(
&
vc
.
c
);
return
vc
.
num
;
return
vc
.
num
;
}
}
/*
* sysfs support
*/
struct
veth_cnx_attribute
{
struct
attribute
attr
;
ssize_t
(
*
show
)(
struct
veth_lpar_connection
*
,
char
*
buf
);
ssize_t
(
*
store
)(
struct
veth_lpar_connection
*
,
const
char
*
buf
);
};
static
ssize_t
veth_cnx_attribute_show
(
struct
kobject
*
kobj
,
struct
attribute
*
attr
,
char
*
buf
)
{
struct
veth_cnx_attribute
*
cnx_attr
;
struct
veth_lpar_connection
*
cnx
;
cnx_attr
=
container_of
(
attr
,
struct
veth_cnx_attribute
,
attr
);
cnx
=
container_of
(
kobj
,
struct
veth_lpar_connection
,
kobject
);
if
(
!
cnx_attr
->
show
)
return
-
EIO
;
return
cnx_attr
->
show
(
cnx
,
buf
);
}
#define CUSTOM_CNX_ATTR(_name, _format, _expression) \
static ssize_t _name##_show(struct veth_lpar_connection *cnx, char *buf)\
{ \
return sprintf(buf, _format, _expression); \
} \
struct veth_cnx_attribute veth_cnx_attr_##_name = __ATTR_RO(_name)
#define SIMPLE_CNX_ATTR(_name) \
CUSTOM_CNX_ATTR(_name, "%lu\n", (unsigned long)cnx->_name)
SIMPLE_CNX_ATTR
(
outstanding_tx
);
SIMPLE_CNX_ATTR
(
remote_lp
);
SIMPLE_CNX_ATTR
(
num_events
);
SIMPLE_CNX_ATTR
(
src_inst
);
SIMPLE_CNX_ATTR
(
dst_inst
);
SIMPLE_CNX_ATTR
(
num_pending_acks
);
SIMPLE_CNX_ATTR
(
num_ack_events
);
CUSTOM_CNX_ATTR
(
ack_timeout
,
"%d
\n
"
,
jiffies_to_msecs
(
cnx
->
ack_timeout
));
CUSTOM_CNX_ATTR
(
reset_timeout
,
"%d
\n
"
,
jiffies_to_msecs
(
cnx
->
reset_timeout
));
CUSTOM_CNX_ATTR
(
state
,
"0x%.4lX
\n
"
,
cnx
->
state
);
CUSTOM_CNX_ATTR
(
last_contact
,
"%d
\n
"
,
cnx
->
last_contact
?
jiffies_to_msecs
(
jiffies
-
cnx
->
last_contact
)
:
0
);
#define GET_CNX_ATTR(_name) (&veth_cnx_attr_##_name.attr)
static
struct
attribute
*
veth_cnx_default_attrs
[]
=
{
GET_CNX_ATTR
(
outstanding_tx
),
GET_CNX_ATTR
(
remote_lp
),
GET_CNX_ATTR
(
num_events
),
GET_CNX_ATTR
(
reset_timeout
),
GET_CNX_ATTR
(
last_contact
),
GET_CNX_ATTR
(
state
),
GET_CNX_ATTR
(
src_inst
),
GET_CNX_ATTR
(
dst_inst
),
GET_CNX_ATTR
(
num_pending_acks
),
GET_CNX_ATTR
(
num_ack_events
),
GET_CNX_ATTR
(
ack_timeout
),
NULL
};
static
struct
sysfs_ops
veth_cnx_sysfs_ops
=
{
.
show
=
veth_cnx_attribute_show
};
static
struct
kobj_type
veth_lpar_connection_ktype
=
{
.
release
=
veth_release_connection
,
.
sysfs_ops
=
&
veth_cnx_sysfs_ops
,
.
default_attrs
=
veth_cnx_default_attrs
};
struct
veth_port_attribute
{
struct
attribute
attr
;
ssize_t
(
*
show
)(
struct
veth_port
*
,
char
*
buf
);
ssize_t
(
*
store
)(
struct
veth_port
*
,
const
char
*
buf
);
};
static
ssize_t
veth_port_attribute_show
(
struct
kobject
*
kobj
,
struct
attribute
*
attr
,
char
*
buf
)
{
struct
veth_port_attribute
*
port_attr
;
struct
veth_port
*
port
;
port_attr
=
container_of
(
attr
,
struct
veth_port_attribute
,
attr
);
port
=
container_of
(
kobj
,
struct
veth_port
,
kobject
);
if
(
!
port_attr
->
show
)
return
-
EIO
;
return
port_attr
->
show
(
port
,
buf
);
}
#define CUSTOM_PORT_ATTR(_name, _format, _expression) \
static ssize_t _name##_show(struct veth_port *port, char *buf) \
{ \
return sprintf(buf, _format, _expression); \
} \
struct veth_port_attribute veth_port_attr_##_name = __ATTR_RO(_name)
#define SIMPLE_PORT_ATTR(_name) \
CUSTOM_PORT_ATTR(_name, "%lu\n", (unsigned long)port->_name)
SIMPLE_PORT_ATTR
(
promiscuous
);
SIMPLE_PORT_ATTR
(
num_mcast
);
CUSTOM_PORT_ATTR
(
lpar_map
,
"0x%X
\n
"
,
port
->
lpar_map
);
CUSTOM_PORT_ATTR
(
stopped_map
,
"0x%X
\n
"
,
port
->
stopped_map
);
CUSTOM_PORT_ATTR
(
mac_addr
,
"0x%lX
\n
"
,
port
->
mac_addr
);
#define GET_PORT_ATTR(_name) (&veth_port_attr_##_name.attr)
static
struct
attribute
*
veth_port_default_attrs
[]
=
{
GET_PORT_ATTR
(
mac_addr
),
GET_PORT_ATTR
(
lpar_map
),
GET_PORT_ATTR
(
stopped_map
),
GET_PORT_ATTR
(
promiscuous
),
GET_PORT_ATTR
(
num_mcast
),
NULL
};
static
struct
sysfs_ops
veth_port_sysfs_ops
=
{
.
show
=
veth_port_attribute_show
};
static
struct
kobj_type
veth_port_ktype
=
{
.
sysfs_ops
=
&
veth_port_sysfs_ops
,
.
default_attrs
=
veth_port_default_attrs
};
/*
/*
* LPAR connection code
* LPAR connection code
*/
*/
...
@@ -266,7 +459,7 @@ static inline void veth_kick_statemachine(struct veth_lpar_connection *cnx)
...
@@ -266,7 +459,7 @@ static inline void veth_kick_statemachine(struct veth_lpar_connection *cnx)
}
}
static
void
veth_take_cap
(
struct
veth_lpar_connection
*
cnx
,
static
void
veth_take_cap
(
struct
veth_lpar_connection
*
cnx
,
struct
VethLpE
vent
*
event
)
struct
veth_lpe
vent
*
event
)
{
{
unsigned
long
flags
;
unsigned
long
flags
;
...
@@ -278,7 +471,7 @@ static void veth_take_cap(struct veth_lpar_connection *cnx,
...
@@ -278,7 +471,7 @@ static void veth_take_cap(struct veth_lpar_connection *cnx,
HvLpEvent_Type_VirtualLan
);
HvLpEvent_Type_VirtualLan
);
if
(
cnx
->
state
&
VETH_STATE_GOTCAPS
)
{
if
(
cnx
->
state
&
VETH_STATE_GOTCAPS
)
{
veth_error
(
"Received a second capabilities from
lpar %d
\n
"
,
veth_error
(
"Received a second capabilities from
LPAR %d.
\n
"
,
cnx
->
remote_lp
);
cnx
->
remote_lp
);
event
->
base_event
.
xRc
=
HvLpEvent_Rc_BufferNotAvailable
;
event
->
base_event
.
xRc
=
HvLpEvent_Rc_BufferNotAvailable
;
HvCallEvent_ackLpEvent
((
struct
HvLpEvent
*
)
event
);
HvCallEvent_ackLpEvent
((
struct
HvLpEvent
*
)
event
);
...
@@ -291,13 +484,13 @@ static void veth_take_cap(struct veth_lpar_connection *cnx,
...
@@ -291,13 +484,13 @@ static void veth_take_cap(struct veth_lpar_connection *cnx,
}
}
static
void
veth_take_cap_ack
(
struct
veth_lpar_connection
*
cnx
,
static
void
veth_take_cap_ack
(
struct
veth_lpar_connection
*
cnx
,
struct
VethLpE
vent
*
event
)
struct
veth_lpe
vent
*
event
)
{
{
unsigned
long
flags
;
unsigned
long
flags
;
spin_lock_irqsave
(
&
cnx
->
lock
,
flags
);
spin_lock_irqsave
(
&
cnx
->
lock
,
flags
);
if
(
cnx
->
state
&
VETH_STATE_GOTCAPACK
)
{
if
(
cnx
->
state
&
VETH_STATE_GOTCAPACK
)
{
veth_error
(
"Received a second capabilities ack from
lpar %d
\n
"
,
veth_error
(
"Received a second capabilities ack from
LPAR %d.
\n
"
,
cnx
->
remote_lp
);
cnx
->
remote_lp
);
}
else
{
}
else
{
memcpy
(
&
cnx
->
cap_ack_event
,
event
,
memcpy
(
&
cnx
->
cap_ack_event
,
event
,
...
@@ -309,19 +502,24 @@ static void veth_take_cap_ack(struct veth_lpar_connection *cnx,
...
@@ -309,19 +502,24 @@ static void veth_take_cap_ack(struct veth_lpar_connection *cnx,
}
}
static
void
veth_take_monitor_ack
(
struct
veth_lpar_connection
*
cnx
,
static
void
veth_take_monitor_ack
(
struct
veth_lpar_connection
*
cnx
,
struct
VethLpE
vent
*
event
)
struct
veth_lpe
vent
*
event
)
{
{
unsigned
long
flags
;
unsigned
long
flags
;
spin_lock_irqsave
(
&
cnx
->
lock
,
flags
);
spin_lock_irqsave
(
&
cnx
->
lock
,
flags
);
veth_printk
(
KERN_DEBUG
,
"Monitor ack returned for lpar %d
\n
"
,
veth_debug
(
"cnx %d: lost connection.
\n
"
,
cnx
->
remote_lp
);
cnx
->
remote_lp
);
/* Avoid kicking the statemachine once we're shutdown.
* It's unnecessary and it could break veth_stop_connection(). */
if
(
!
(
cnx
->
state
&
VETH_STATE_SHUTDOWN
))
{
cnx
->
state
|=
VETH_STATE_RESET
;
cnx
->
state
|=
VETH_STATE_RESET
;
veth_kick_statemachine
(
cnx
);
veth_kick_statemachine
(
cnx
);
}
spin_unlock_irqrestore
(
&
cnx
->
lock
,
flags
);
spin_unlock_irqrestore
(
&
cnx
->
lock
,
flags
);
}
}
static
void
veth_handle_ack
(
struct
VethLpE
vent
*
event
)
static
void
veth_handle_ack
(
struct
veth_lpe
vent
*
event
)
{
{
HvLpIndex
rlp
=
event
->
base_event
.
xTargetLp
;
HvLpIndex
rlp
=
event
->
base_event
.
xTargetLp
;
struct
veth_lpar_connection
*
cnx
=
veth_cnx
[
rlp
];
struct
veth_lpar_connection
*
cnx
=
veth_cnx
[
rlp
];
...
@@ -329,58 +527,67 @@ static void veth_handle_ack(struct VethLpEvent *event)
...
@@ -329,58 +527,67 @@ static void veth_handle_ack(struct VethLpEvent *event)
BUG_ON
(
!
cnx
);
BUG_ON
(
!
cnx
);
switch
(
event
->
base_event
.
xSubtype
)
{
switch
(
event
->
base_event
.
xSubtype
)
{
case
V
ethEventTypeCap
:
case
V
ETH_EVENT_CAP
:
veth_take_cap_ack
(
cnx
,
event
);
veth_take_cap_ack
(
cnx
,
event
);
break
;
break
;
case
V
ethEventTypeMonitor
:
case
V
ETH_EVENT_MONITOR
:
veth_take_monitor_ack
(
cnx
,
event
);
veth_take_monitor_ack
(
cnx
,
event
);
break
;
break
;
default:
default:
veth_error
(
"Unknown ack type %d from
lpar %d
\n
"
,
veth_error
(
"Unknown ack type %d from
LPAR %d.
\n
"
,
event
->
base_event
.
xSubtype
,
rlp
);
event
->
base_event
.
xSubtype
,
rlp
);
};
};
}
}
static
void
veth_handle_int
(
struct
VethLpE
vent
*
event
)
static
void
veth_handle_int
(
struct
veth_lpe
vent
*
event
)
{
{
HvLpIndex
rlp
=
event
->
base_event
.
xSourceLp
;
HvLpIndex
rlp
=
event
->
base_event
.
xSourceLp
;
struct
veth_lpar_connection
*
cnx
=
veth_cnx
[
rlp
];
struct
veth_lpar_connection
*
cnx
=
veth_cnx
[
rlp
];
unsigned
long
flags
;
unsigned
long
flags
;
int
i
;
int
i
,
acked
=
0
;
BUG_ON
(
!
cnx
);
BUG_ON
(
!
cnx
);
switch
(
event
->
base_event
.
xSubtype
)
{
switch
(
event
->
base_event
.
xSubtype
)
{
case
V
ethEventTypeCap
:
case
V
ETH_EVENT_CAP
:
veth_take_cap
(
cnx
,
event
);
veth_take_cap
(
cnx
,
event
);
break
;
break
;
case
V
ethEventTypeMonitor
:
case
V
ETH_EVENT_MONITOR
:
/* do nothing... this'll hang out here til we're dead,
/* do nothing... this'll hang out here til we're dead,
* and the hypervisor will return it for us. */
* and the hypervisor will return it for us. */
break
;
break
;
case
V
ethEventTypeFramesAck
:
case
V
ETH_EVENT_FRAMES_ACK
:
spin_lock_irqsave
(
&
cnx
->
lock
,
flags
);
spin_lock_irqsave
(
&
cnx
->
lock
,
flags
);
for
(
i
=
0
;
i
<
VETH_MAX_ACKS_PER_MSG
;
++
i
)
{
for
(
i
=
0
;
i
<
VETH_MAX_ACKS_PER_MSG
;
++
i
)
{
u16
msgnum
=
event
->
u
.
frames_ack_data
.
token
[
i
];
u16
msgnum
=
event
->
u
.
frames_ack_data
.
token
[
i
];
if
(
msgnum
<
VETH_NUMBUFFERS
)
if
(
msgnum
<
VETH_NUMBUFFERS
)
{
veth_recycle_msg
(
cnx
,
cnx
->
msgs
+
msgnum
);
veth_recycle_msg
(
cnx
,
cnx
->
msgs
+
msgnum
);
cnx
->
outstanding_tx
--
;
acked
++
;
}
}
}
if
(
acked
>
0
)
{
cnx
->
last_contact
=
jiffies
;
veth_wake_queues
(
cnx
);
}
spin_unlock_irqrestore
(
&
cnx
->
lock
,
flags
);
spin_unlock_irqrestore
(
&
cnx
->
lock
,
flags
);
veth_flush_pending
(
cnx
);
break
;
break
;
case
V
ethEventTypeFrames
:
case
V
ETH_EVENT_FRAMES
:
veth_receive
(
cnx
,
event
);
veth_receive
(
cnx
,
event
);
break
;
break
;
default:
default:
veth_error
(
"Unknown interrupt type %d from
lpar %d
\n
"
,
veth_error
(
"Unknown interrupt type %d from
LPAR %d.
\n
"
,
event
->
base_event
.
xSubtype
,
rlp
);
event
->
base_event
.
xSubtype
,
rlp
);
};
};
}
}
static
void
veth_handle_event
(
struct
HvLpEvent
*
event
,
struct
pt_regs
*
regs
)
static
void
veth_handle_event
(
struct
HvLpEvent
*
event
,
struct
pt_regs
*
regs
)
{
{
struct
VethLpEvent
*
veth_event
=
(
struct
VethLpE
vent
*
)
event
;
struct
veth_lpevent
*
veth_event
=
(
struct
veth_lpe
vent
*
)
event
;
if
(
event
->
xFlags
.
xFunction
==
HvLpEvent_Function_Ack
)
if
(
event
->
xFlags
.
xFunction
==
HvLpEvent_Function_Ack
)
veth_handle_ack
(
veth_event
);
veth_handle_ack
(
veth_event
);
...
@@ -390,7 +597,7 @@ static void veth_handle_event(struct HvLpEvent *event, struct pt_regs *regs)
...
@@ -390,7 +597,7 @@ static void veth_handle_event(struct HvLpEvent *event, struct pt_regs *regs)
static
int
veth_process_caps
(
struct
veth_lpar_connection
*
cnx
)
static
int
veth_process_caps
(
struct
veth_lpar_connection
*
cnx
)
{
{
struct
VethCapD
ata
*
remote_caps
=
&
cnx
->
remote_caps
;
struct
veth_cap_d
ata
*
remote_caps
=
&
cnx
->
remote_caps
;
int
num_acks_needed
;
int
num_acks_needed
;
/* Convert timer to jiffies */
/* Convert timer to jiffies */
...
@@ -400,7 +607,7 @@ static int veth_process_caps(struct veth_lpar_connection *cnx)
...
@@ -400,7 +607,7 @@ static int veth_process_caps(struct veth_lpar_connection *cnx)
||
(
remote_caps
->
ack_threshold
>
VETH_MAX_ACKS_PER_MSG
)
||
(
remote_caps
->
ack_threshold
>
VETH_MAX_ACKS_PER_MSG
)
||
(
remote_caps
->
ack_threshold
==
0
)
||
(
remote_caps
->
ack_threshold
==
0
)
||
(
cnx
->
ack_timeout
==
0
)
)
{
||
(
cnx
->
ack_timeout
==
0
)
)
{
veth_error
(
"Received incompatible capabilities from
lpar %d
\n
"
,
veth_error
(
"Received incompatible capabilities from
LPAR %d.
\n
"
,
cnx
->
remote_lp
);
cnx
->
remote_lp
);
return
HvLpEvent_Rc_InvalidSubtypeData
;
return
HvLpEvent_Rc_InvalidSubtypeData
;
}
}
...
@@ -418,8 +625,8 @@ static int veth_process_caps(struct veth_lpar_connection *cnx)
...
@@ -418,8 +625,8 @@ static int veth_process_caps(struct veth_lpar_connection *cnx)
cnx
->
num_ack_events
+=
num
;
cnx
->
num_ack_events
+=
num
;
if
(
cnx
->
num_ack_events
<
num_acks_needed
)
{
if
(
cnx
->
num_ack_events
<
num_acks_needed
)
{
veth_error
(
"Couldn't allocate enough ack events
for lpar %d
\n
"
,
veth_error
(
"Couldn't allocate enough ack events
"
cnx
->
remote_lp
);
"for LPAR %d.
\n
"
,
cnx
->
remote_lp
);
return
HvLpEvent_Rc_BufferNotAvailable
;
return
HvLpEvent_Rc_BufferNotAvailable
;
}
}
...
@@ -440,15 +647,15 @@ static void veth_statemachine(void *p)
...
@@ -440,15 +647,15 @@ static void veth_statemachine(void *p)
restart:
restart:
if
(
cnx
->
state
&
VETH_STATE_RESET
)
{
if
(
cnx
->
state
&
VETH_STATE_RESET
)
{
int
i
;
del_timer
(
&
cnx
->
ack_timer
);
if
(
cnx
->
state
&
VETH_STATE_OPEN
)
if
(
cnx
->
state
&
VETH_STATE_OPEN
)
HvCallEvent_closeLpEventPath
(
cnx
->
remote_lp
,
HvCallEvent_closeLpEventPath
(
cnx
->
remote_lp
,
HvLpEvent_Type_VirtualLan
);
HvLpEvent_Type_VirtualLan
);
/* reset ack data */
/*
* Reset ack data. This prevents the ack_timer actually
* doing anything, even if it runs one more time when
* we drop the lock below.
*/
memset
(
&
cnx
->
pending_acks
,
0xff
,
sizeof
(
cnx
->
pending_acks
));
memset
(
&
cnx
->
pending_acks
,
0xff
,
sizeof
(
cnx
->
pending_acks
));
cnx
->
num_pending_acks
=
0
;
cnx
->
num_pending_acks
=
0
;
...
@@ -458,14 +665,32 @@ static void veth_statemachine(void *p)
...
@@ -458,14 +665,32 @@ static void veth_statemachine(void *p)
|
VETH_STATE_SENTCAPACK
|
VETH_STATE_READY
);
|
VETH_STATE_SENTCAPACK
|
VETH_STATE_READY
);
/* Clean up any leftover messages */
/* Clean up any leftover messages */
if
(
cnx
->
msgs
)
if
(
cnx
->
msgs
)
{
int
i
;
for
(
i
=
0
;
i
<
VETH_NUMBUFFERS
;
++
i
)
for
(
i
=
0
;
i
<
VETH_NUMBUFFERS
;
++
i
)
veth_recycle_msg
(
cnx
,
cnx
->
msgs
+
i
);
veth_recycle_msg
(
cnx
,
cnx
->
msgs
+
i
);
}
cnx
->
outstanding_tx
=
0
;
veth_wake_queues
(
cnx
);
/* Drop the lock so we can do stuff that might sleep or
* take other locks. */
spin_unlock_irq
(
&
cnx
->
lock
);
spin_unlock_irq
(
&
cnx
->
lock
);
veth_flush_pending
(
cnx
);
del_timer_sync
(
&
cnx
->
ack_timer
);
del_timer_sync
(
&
cnx
->
reset_timer
);
spin_lock_irq
(
&
cnx
->
lock
);
spin_lock_irq
(
&
cnx
->
lock
);
if
(
cnx
->
state
&
VETH_STATE_RESET
)
if
(
cnx
->
state
&
VETH_STATE_RESET
)
goto
restart
;
goto
restart
;
/* Hack, wait for the other end to reset itself. */
if
(
!
(
cnx
->
state
&
VETH_STATE_SHUTDOWN
))
{
schedule_delayed_work
(
&
cnx
->
statemachine_wq
,
5
*
HZ
);
goto
out
;
}
}
}
if
(
cnx
->
state
&
VETH_STATE_SHUTDOWN
)
if
(
cnx
->
state
&
VETH_STATE_SHUTDOWN
)
...
@@ -488,7 +713,7 @@ static void veth_statemachine(void *p)
...
@@ -488,7 +713,7 @@ static void veth_statemachine(void *p)
if
(
(
cnx
->
state
&
VETH_STATE_OPEN
)
if
(
(
cnx
->
state
&
VETH_STATE_OPEN
)
&&
!
(
cnx
->
state
&
VETH_STATE_SENTMON
)
)
{
&&
!
(
cnx
->
state
&
VETH_STATE_SENTMON
)
)
{
rc
=
veth_signalevent
(
cnx
,
V
ethEventTypeMonitor
,
rc
=
veth_signalevent
(
cnx
,
V
ETH_EVENT_MONITOR
,
HvLpEvent_AckInd_DoAck
,
HvLpEvent_AckInd_DoAck
,
HvLpEvent_AckType_DeferredAck
,
HvLpEvent_AckType_DeferredAck
,
0
,
0
,
0
,
0
,
0
,
0
);
0
,
0
,
0
,
0
,
0
,
0
);
...
@@ -498,9 +723,8 @@ static void veth_statemachine(void *p)
...
@@ -498,9 +723,8 @@ static void veth_statemachine(void *p)
}
else
{
}
else
{
if
(
(
rc
!=
HvLpEvent_Rc_PartitionDead
)
if
(
(
rc
!=
HvLpEvent_Rc_PartitionDead
)
&&
(
rc
!=
HvLpEvent_Rc_PathClosed
)
)
&&
(
rc
!=
HvLpEvent_Rc_PathClosed
)
)
veth_error
(
"Error sending monitor to "
veth_error
(
"Error sending monitor to LPAR %d, "
"lpar %d, rc=%x
\n
"
,
"rc = %d
\n
"
,
rlp
,
rc
);
rlp
,
(
int
)
rc
);
/* Oh well, hope we get a cap from the other
/* Oh well, hope we get a cap from the other
* end and do better when that kicks us */
* end and do better when that kicks us */
...
@@ -512,7 +736,7 @@ static void veth_statemachine(void *p)
...
@@ -512,7 +736,7 @@ static void veth_statemachine(void *p)
&&
!
(
cnx
->
state
&
VETH_STATE_SENTCAPS
))
{
&&
!
(
cnx
->
state
&
VETH_STATE_SENTCAPS
))
{
u64
*
rawcap
=
(
u64
*
)
&
cnx
->
local_caps
;
u64
*
rawcap
=
(
u64
*
)
&
cnx
->
local_caps
;
rc
=
veth_signalevent
(
cnx
,
V
ethEventTypeCap
,
rc
=
veth_signalevent
(
cnx
,
V
ETH_EVENT_CAP
,
HvLpEvent_AckInd_DoAck
,
HvLpEvent_AckInd_DoAck
,
HvLpEvent_AckType_ImmediateAck
,
HvLpEvent_AckType_ImmediateAck
,
0
,
rawcap
[
0
],
rawcap
[
1
],
rawcap
[
2
],
0
,
rawcap
[
0
],
rawcap
[
1
],
rawcap
[
2
],
...
@@ -523,9 +747,9 @@ static void veth_statemachine(void *p)
...
@@ -523,9 +747,9 @@ static void veth_statemachine(void *p)
}
else
{
}
else
{
if
(
(
rc
!=
HvLpEvent_Rc_PartitionDead
)
if
(
(
rc
!=
HvLpEvent_Rc_PartitionDead
)
&&
(
rc
!=
HvLpEvent_Rc_PathClosed
)
)
&&
(
rc
!=
HvLpEvent_Rc_PathClosed
)
)
veth_error
(
"Error sending caps to "
veth_error
(
"Error sending caps to
LPAR %d,
"
"lpar %d, rc=%x
\n
"
,
"rc = %d
\n
"
,
rlp
,
rc
);
rlp
,
(
int
)
rc
);
/* Oh well, hope we get a cap from the other
/* Oh well, hope we get a cap from the other
* end and do better when that kicks us */
* end and do better when that kicks us */
goto
out
;
goto
out
;
...
@@ -534,7 +758,7 @@ static void veth_statemachine(void *p)
...
@@ -534,7 +758,7 @@ static void veth_statemachine(void *p)
if
((
cnx
->
state
&
VETH_STATE_GOTCAPS
)
if
((
cnx
->
state
&
VETH_STATE_GOTCAPS
)
&&
!
(
cnx
->
state
&
VETH_STATE_SENTCAPACK
))
{
&&
!
(
cnx
->
state
&
VETH_STATE_SENTCAPACK
))
{
struct
VethCapD
ata
*
remote_caps
=
&
cnx
->
remote_caps
;
struct
veth_cap_d
ata
*
remote_caps
=
&
cnx
->
remote_caps
;
memcpy
(
remote_caps
,
&
cnx
->
cap_event
.
u
.
caps_data
,
memcpy
(
remote_caps
,
&
cnx
->
cap_event
.
u
.
caps_data
,
sizeof
(
*
remote_caps
));
sizeof
(
*
remote_caps
));
...
@@ -565,10 +789,8 @@ static void veth_statemachine(void *p)
...
@@ -565,10 +789,8 @@ static void veth_statemachine(void *p)
add_timer
(
&
cnx
->
ack_timer
);
add_timer
(
&
cnx
->
ack_timer
);
cnx
->
state
|=
VETH_STATE_READY
;
cnx
->
state
|=
VETH_STATE_READY
;
}
else
{
}
else
{
veth_printk
(
KERN_ERR
,
"Caps rejected (rc=%d) by "
veth_error
(
"Caps rejected by LPAR %d, rc = %d
\n
"
,
"lpar %d
\n
"
,
rlp
,
cnx
->
cap_ack_event
.
base_event
.
xRc
);
cnx
->
cap_ack_event
.
base_event
.
xRc
,
rlp
);
goto
cant_cope
;
goto
cant_cope
;
}
}
}
}
...
@@ -581,8 +803,8 @@ static void veth_statemachine(void *p)
...
@@ -581,8 +803,8 @@ static void veth_statemachine(void *p)
/* FIXME: we get here if something happens we really can't
/* FIXME: we get here if something happens we really can't
* cope with. The link will never work once we get here, and
* cope with. The link will never work once we get here, and
* all we can do is not lock the rest of the system up */
* all we can do is not lock the rest of the system up */
veth_error
(
"
Badness on connection to lpar %d (state=%04lx)
"
veth_error
(
"
Unrecoverable error on connection to LPAR %d, shutting down
"
" - shutting down
\n
"
,
rlp
,
cnx
->
state
);
" (state = 0x%04lx)
\n
"
,
rlp
,
cnx
->
state
);
cnx
->
state
|=
VETH_STATE_SHUTDOWN
;
cnx
->
state
|=
VETH_STATE_SHUTDOWN
;
spin_unlock_irq
(
&
cnx
->
lock
);
spin_unlock_irq
(
&
cnx
->
lock
);
}
}
...
@@ -591,7 +813,7 @@ static int veth_init_connection(u8 rlp)
...
@@ -591,7 +813,7 @@ static int veth_init_connection(u8 rlp)
{
{
struct
veth_lpar_connection
*
cnx
;
struct
veth_lpar_connection
*
cnx
;
struct
veth_msg
*
msgs
;
struct
veth_msg
*
msgs
;
int
i
;
int
i
,
rc
;
if
(
(
rlp
==
this_lp
)
if
(
(
rlp
==
this_lp
)
||
!
HvLpConfig_doLpsCommunicateOnVirtualLan
(
this_lp
,
rlp
)
)
||
!
HvLpConfig_doLpsCommunicateOnVirtualLan
(
this_lp
,
rlp
)
)
...
@@ -605,22 +827,36 @@ static int veth_init_connection(u8 rlp)
...
@@ -605,22 +827,36 @@ static int veth_init_connection(u8 rlp)
cnx
->
remote_lp
=
rlp
;
cnx
->
remote_lp
=
rlp
;
spin_lock_init
(
&
cnx
->
lock
);
spin_lock_init
(
&
cnx
->
lock
);
INIT_WORK
(
&
cnx
->
statemachine_wq
,
veth_statemachine
,
cnx
);
INIT_WORK
(
&
cnx
->
statemachine_wq
,
veth_statemachine
,
cnx
);
init_timer
(
&
cnx
->
ack_timer
);
init_timer
(
&
cnx
->
ack_timer
);
cnx
->
ack_timer
.
function
=
veth_timed_ack
;
cnx
->
ack_timer
.
function
=
veth_timed_ack
;
cnx
->
ack_timer
.
data
=
(
unsigned
long
)
cnx
;
cnx
->
ack_timer
.
data
=
(
unsigned
long
)
cnx
;
init_timer
(
&
cnx
->
reset_timer
);
cnx
->
reset_timer
.
function
=
veth_timed_reset
;
cnx
->
reset_timer
.
data
=
(
unsigned
long
)
cnx
;
cnx
->
reset_timeout
=
5
*
HZ
*
(
VETH_ACKTIMEOUT
/
1000000
);
memset
(
&
cnx
->
pending_acks
,
0xff
,
sizeof
(
cnx
->
pending_acks
));
memset
(
&
cnx
->
pending_acks
,
0xff
,
sizeof
(
cnx
->
pending_acks
));
veth_cnx
[
rlp
]
=
cnx
;
veth_cnx
[
rlp
]
=
cnx
;
/* This gets us 1 reference, which is held on behalf of the driver
* infrastructure. It's released at module unload. */
kobject_init
(
&
cnx
->
kobject
);
cnx
->
kobject
.
ktype
=
&
veth_lpar_connection_ktype
;
rc
=
kobject_set_name
(
&
cnx
->
kobject
,
"cnx%.2d"
,
rlp
);
if
(
rc
!=
0
)
return
rc
;
msgs
=
kmalloc
(
VETH_NUMBUFFERS
*
sizeof
(
struct
veth_msg
),
GFP_KERNEL
);
msgs
=
kmalloc
(
VETH_NUMBUFFERS
*
sizeof
(
struct
veth_msg
),
GFP_KERNEL
);
if
(
!
msgs
)
{
if
(
!
msgs
)
{
veth_error
(
"Can't allocate buffers for
lpar %d
\n
"
,
rlp
);
veth_error
(
"Can't allocate buffers for
LPAR %d.
\n
"
,
rlp
);
return
-
ENOMEM
;
return
-
ENOMEM
;
}
}
cnx
->
msgs
=
msgs
;
cnx
->
msgs
=
msgs
;
memset
(
msgs
,
0
,
VETH_NUMBUFFERS
*
sizeof
(
struct
veth_msg
));
memset
(
msgs
,
0
,
VETH_NUMBUFFERS
*
sizeof
(
struct
veth_msg
));
spin_lock_init
(
&
cnx
->
msg_stack_lock
);
for
(
i
=
0
;
i
<
VETH_NUMBUFFERS
;
i
++
)
{
for
(
i
=
0
;
i
<
VETH_NUMBUFFERS
;
i
++
)
{
msgs
[
i
].
token
=
i
;
msgs
[
i
].
token
=
i
;
...
@@ -630,8 +866,7 @@ static int veth_init_connection(u8 rlp)
...
@@ -630,8 +866,7 @@ static int veth_init_connection(u8 rlp)
cnx
->
num_events
=
veth_allocate_events
(
rlp
,
2
+
VETH_NUMBUFFERS
);
cnx
->
num_events
=
veth_allocate_events
(
rlp
,
2
+
VETH_NUMBUFFERS
);
if
(
cnx
->
num_events
<
(
2
+
VETH_NUMBUFFERS
))
{
if
(
cnx
->
num_events
<
(
2
+
VETH_NUMBUFFERS
))
{
veth_error
(
"Can't allocate events for lpar %d, only got %d
\n
"
,
veth_error
(
"Can't allocate enough events for LPAR %d.
\n
"
,
rlp
);
rlp
,
cnx
->
num_events
);
return
-
ENOMEM
;
return
-
ENOMEM
;
}
}
...
@@ -642,11 +877,9 @@ static int veth_init_connection(u8 rlp)
...
@@ -642,11 +877,9 @@ static int veth_init_connection(u8 rlp)
return
0
;
return
0
;
}
}
static
void
veth_stop_connection
(
u8
rlp
)
static
void
veth_stop_connection
(
struct
veth_lpar_connection
*
cnx
)
{
{
struct
veth_lpar_connection
*
cnx
=
veth_cnx
[
rlp
];
if
(
!
cnx
)
if
(
!
cnx
)
return
;
return
;
spin_lock_irq
(
&
cnx
->
lock
);
spin_lock_irq
(
&
cnx
->
lock
);
...
@@ -654,12 +887,23 @@ static void veth_stop_connection(u8 rlp)
...
@@ -654,12 +887,23 @@ static void veth_stop_connection(u8 rlp)
veth_kick_statemachine
(
cnx
);
veth_kick_statemachine
(
cnx
);
spin_unlock_irq
(
&
cnx
->
lock
);
spin_unlock_irq
(
&
cnx
->
lock
);
/* There's a slim chance the reset code has just queued the
* statemachine to run in five seconds. If so we need to cancel
* that and requeue the work to run now. */
if
(
cancel_delayed_work
(
&
cnx
->
statemachine_wq
))
{
spin_lock_irq
(
&
cnx
->
lock
);
veth_kick_statemachine
(
cnx
);
spin_unlock_irq
(
&
cnx
->
lock
);
}
/* Wait for the state machine to run. */
flush_scheduled_work
();
flush_scheduled_work
();
}
/* FIXME: not sure if this is necessary - will already have
static
void
veth_destroy_connection
(
struct
veth_lpar_connection
*
cnx
)
* been deleted by the state machine, just want to make sure
{
* its not running any more */
if
(
!
cnx
)
del_timer_sync
(
&
cnx
->
ack_timer
)
;
return
;
if
(
cnx
->
num_events
>
0
)
if
(
cnx
->
num_events
>
0
)
mf_deallocate_lp_events
(
cnx
->
remote_lp
,
mf_deallocate_lp_events
(
cnx
->
remote_lp
,
...
@@ -671,18 +915,18 @@ static void veth_stop_connection(u8 rlp)
...
@@ -671,18 +915,18 @@ static void veth_stop_connection(u8 rlp)
HvLpEvent_Type_VirtualLan
,
HvLpEvent_Type_VirtualLan
,
cnx
->
num_ack_events
,
cnx
->
num_ack_events
,
NULL
,
NULL
);
NULL
,
NULL
);
}
static
void
veth_destroy_connection
(
u8
rlp
)
{
struct
veth_lpar_connection
*
cnx
=
veth_cnx
[
rlp
];
if
(
!
cnx
)
return
;
kfree
(
cnx
->
msgs
);
kfree
(
cnx
->
msgs
);
veth_cnx
[
cnx
->
remote_lp
]
=
NULL
;
kfree
(
cnx
);
kfree
(
cnx
);
veth_cnx
[
rlp
]
=
NULL
;
}
static
void
veth_release_connection
(
struct
kobject
*
kobj
)
{
struct
veth_lpar_connection
*
cnx
;
cnx
=
container_of
(
kobj
,
struct
veth_lpar_connection
,
kobject
);
veth_stop_connection
(
cnx
);
veth_destroy_connection
(
cnx
);
}
}
/*
/*
...
@@ -726,17 +970,15 @@ static void veth_set_multicast_list(struct net_device *dev)
...
@@ -726,17 +970,15 @@ static void veth_set_multicast_list(struct net_device *dev)
write_lock_irqsave
(
&
port
->
mcast_gate
,
flags
);
write_lock_irqsave
(
&
port
->
mcast_gate
,
flags
);
if
(
dev
->
flags
&
IFF_PROMISC
)
{
/* set promiscuous mode */
if
((
dev
->
flags
&
IFF_PROMISC
)
||
(
dev
->
flags
&
IFF_ALLMULTI
)
||
printk
(
KERN_INFO
"%s: Promiscuous mode enabled.
\n
"
,
(
dev
->
mc_count
>
VETH_MAX_MCAST
))
{
dev
->
name
);
port
->
promiscuous
=
1
;
port
->
promiscuous
=
1
;
}
else
if
(
(
dev
->
flags
&
IFF_ALLMULTI
)
||
(
dev
->
mc_count
>
VETH_MAX_MCAST
)
)
{
port
->
all_mcast
=
1
;
}
else
{
}
else
{
struct
dev_mc_list
*
dmi
=
dev
->
mc_list
;
struct
dev_mc_list
*
dmi
=
dev
->
mc_list
;
int
i
;
int
i
;
port
->
promiscuous
=
0
;
/* Update table */
/* Update table */
port
->
num_mcast
=
0
;
port
->
num_mcast
=
0
;
...
@@ -758,9 +1000,10 @@ static void veth_set_multicast_list(struct net_device *dev)
...
@@ -758,9 +1000,10 @@ static void veth_set_multicast_list(struct net_device *dev)
static
void
veth_get_drvinfo
(
struct
net_device
*
dev
,
struct
ethtool_drvinfo
*
info
)
static
void
veth_get_drvinfo
(
struct
net_device
*
dev
,
struct
ethtool_drvinfo
*
info
)
{
{
strncpy
(
info
->
driver
,
"veth"
,
sizeof
(
info
->
driver
)
-
1
);
strncpy
(
info
->
driver
,
DRV_NAME
,
sizeof
(
info
->
driver
)
-
1
);
info
->
driver
[
sizeof
(
info
->
driver
)
-
1
]
=
'\0'
;
info
->
driver
[
sizeof
(
info
->
driver
)
-
1
]
=
'\0'
;
strncpy
(
info
->
version
,
"1.0"
,
sizeof
(
info
->
version
)
-
1
);
strncpy
(
info
->
version
,
DRV_VERSION
,
sizeof
(
info
->
version
)
-
1
);
info
->
version
[
sizeof
(
info
->
version
)
-
1
]
=
'\0'
;
}
}
static
int
veth_get_settings
(
struct
net_device
*
dev
,
struct
ethtool_cmd
*
ecmd
)
static
int
veth_get_settings
(
struct
net_device
*
dev
,
struct
ethtool_cmd
*
ecmd
)
...
@@ -791,49 +1034,6 @@ static struct ethtool_ops ops = {
...
@@ -791,49 +1034,6 @@ static struct ethtool_ops ops = {
.
get_link
=
veth_get_link
,
.
get_link
=
veth_get_link
,
};
};
static
void
veth_tx_timeout
(
struct
net_device
*
dev
)
{
struct
veth_port
*
port
=
(
struct
veth_port
*
)
dev
->
priv
;
struct
net_device_stats
*
stats
=
&
port
->
stats
;
unsigned
long
flags
;
int
i
;
stats
->
tx_errors
++
;
spin_lock_irqsave
(
&
port
->
pending_gate
,
flags
);
if
(
!
port
->
pending_lpmask
)
{
spin_unlock_irqrestore
(
&
port
->
pending_gate
,
flags
);
return
;
}
printk
(
KERN_WARNING
"%s: Tx timeout! Resetting lp connections: %08x
\n
"
,
dev
->
name
,
port
->
pending_lpmask
);
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
i
++
)
{
struct
veth_lpar_connection
*
cnx
=
veth_cnx
[
i
];
if
(
!
(
port
->
pending_lpmask
&
(
1
<<
i
)))
continue
;
/* If we're pending on it, we must be connected to it,
* so we should certainly have a structure for it. */
BUG_ON
(
!
cnx
);
/* Theoretically we could be kicking a connection
* which doesn't deserve it, but in practice if we've
* had a Tx timeout, the pending_lpmask will have
* exactly one bit set - the connection causing the
* problem. */
spin_lock
(
&
cnx
->
lock
);
cnx
->
state
|=
VETH_STATE_RESET
;
veth_kick_statemachine
(
cnx
);
spin_unlock
(
&
cnx
->
lock
);
}
spin_unlock_irqrestore
(
&
port
->
pending_gate
,
flags
);
}
static
struct
net_device
*
__init
veth_probe_one
(
int
vlan
,
struct
device
*
vdev
)
static
struct
net_device
*
__init
veth_probe_one
(
int
vlan
,
struct
device
*
vdev
)
{
{
struct
net_device
*
dev
;
struct
net_device
*
dev
;
...
@@ -848,8 +1048,9 @@ static struct net_device * __init veth_probe_one(int vlan, struct device *vdev)
...
@@ -848,8 +1048,9 @@ static struct net_device * __init veth_probe_one(int vlan, struct device *vdev)
port
=
(
struct
veth_port
*
)
dev
->
priv
;
port
=
(
struct
veth_port
*
)
dev
->
priv
;
spin_lock_init
(
&
port
->
pending_gate
);
spin_lock_init
(
&
port
->
queue_lock
);
rwlock_init
(
&
port
->
mcast_gate
);
rwlock_init
(
&
port
->
mcast_gate
);
port
->
stopped_map
=
0
;
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
i
++
)
{
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
i
++
)
{
HvLpVirtualLanIndexMap
map
;
HvLpVirtualLanIndexMap
map
;
...
@@ -882,21 +1083,23 @@ static struct net_device * __init veth_probe_one(int vlan, struct device *vdev)
...
@@ -882,21 +1083,23 @@ static struct net_device * __init veth_probe_one(int vlan, struct device *vdev)
dev
->
set_multicast_list
=
veth_set_multicast_list
;
dev
->
set_multicast_list
=
veth_set_multicast_list
;
SET_ETHTOOL_OPS
(
dev
,
&
ops
);
SET_ETHTOOL_OPS
(
dev
,
&
ops
);
dev
->
watchdog_timeo
=
2
*
(
VETH_ACKTIMEOUT
*
HZ
/
1000000
);
dev
->
tx_timeout
=
veth_tx_timeout
;
SET_NETDEV_DEV
(
dev
,
vdev
);
SET_NETDEV_DEV
(
dev
,
vdev
);
rc
=
register_netdev
(
dev
);
rc
=
register_netdev
(
dev
);
if
(
rc
!=
0
)
{
if
(
rc
!=
0
)
{
veth_printk
(
KERN_ERR
,
veth_error
(
"Failed registering net device for vlan%d.
\n
"
,
vlan
);
"Failed to register ethernet device for vlan %d
\n
"
,
vlan
);
free_netdev
(
dev
);
free_netdev
(
dev
);
return
NULL
;
return
NULL
;
}
}
veth_printk
(
KERN_DEBUG
,
"%s attached to iSeries vlan %d (lpar_map=0x%04x)
\n
"
,
kobject_init
(
&
port
->
kobject
);
port
->
kobject
.
parent
=
&
dev
->
class_dev
.
kobj
;
port
->
kobject
.
ktype
=
&
veth_port_ktype
;
kobject_set_name
(
&
port
->
kobject
,
"veth_port"
);
if
(
0
!=
kobject_add
(
&
port
->
kobject
))
veth_error
(
"Failed adding port for %s to sysfs.
\n
"
,
dev
->
name
);
veth_info
(
"%s attached to iSeries vlan %d (LPAR map = 0x%.4X)
\n
"
,
dev
->
name
,
vlan
,
port
->
lpar_map
);
dev
->
name
,
vlan
,
port
->
lpar_map
);
return
dev
;
return
dev
;
...
@@ -912,98 +1115,95 @@ static int veth_transmit_to_one(struct sk_buff *skb, HvLpIndex rlp,
...
@@ -912,98 +1115,95 @@ static int veth_transmit_to_one(struct sk_buff *skb, HvLpIndex rlp,
struct
veth_lpar_connection
*
cnx
=
veth_cnx
[
rlp
];
struct
veth_lpar_connection
*
cnx
=
veth_cnx
[
rlp
];
struct
veth_port
*
port
=
(
struct
veth_port
*
)
dev
->
priv
;
struct
veth_port
*
port
=
(
struct
veth_port
*
)
dev
->
priv
;
HvLpEvent_Rc
rc
;
HvLpEvent_Rc
rc
;
u32
dma_address
,
dma_length
;
struct
veth_msg
*
msg
=
NULL
;
struct
veth_msg
*
msg
=
NULL
;
int
err
=
0
;
unsigned
long
flags
;
unsigned
long
flags
;
if
(
!
cnx
)
{
if
(
!
cnx
)
port
->
stats
.
tx_errors
++
;
dev_kfree_skb
(
skb
);
return
0
;
return
0
;
}
spin_lock_irqsave
(
&
cnx
->
lock
,
flags
);
spin_lock_irqsave
(
&
cnx
->
lock
,
flags
);
if
(
!
(
cnx
->
state
&
VETH_STATE_READY
))
if
(
!
(
cnx
->
state
&
VETH_STATE_READY
))
goto
drop
;
goto
no_error
;
if
((
skb
->
len
-
14
)
>
VETH_MAX_MTU
)
if
((
skb
->
len
-
ETH_HLEN
)
>
VETH_MAX_MTU
)
goto
drop
;
goto
drop
;
msg
=
veth_stack_pop
(
cnx
);
msg
=
veth_stack_pop
(
cnx
);
if
(
!
msg
)
if
(
!
msg
)
{
err
=
1
;
goto
drop
;
goto
drop
;
}
dma_length
=
skb
->
len
;
msg
->
in_use
=
1
;
dma_address
=
dma_map_single
(
port
->
dev
,
skb
->
data
,
msg
->
skb
=
skb_get
(
skb
);
dma_length
,
DMA_TO_DEVICE
);
if
(
dma_mapping_error
(
dma_address
))
msg
->
data
.
addr
[
0
]
=
dma_map_single
(
port
->
dev
,
skb
->
data
,
skb
->
len
,
DMA_TO_DEVICE
);
if
(
dma_mapping_error
(
msg
->
data
.
addr
[
0
]))
goto
recycle_and_drop
;
goto
recycle_and_drop
;
/* Is it really necessary to check the length and address
* fields of the first entry here? */
msg
->
skb
=
skb
;
msg
->
dev
=
port
->
dev
;
msg
->
dev
=
port
->
dev
;
msg
->
data
.
addr
[
0
]
=
dma_address
;
msg
->
data
.
len
[
0
]
=
skb
->
len
;
msg
->
data
.
len
[
0
]
=
dma_length
;
msg
->
data
.
eofmask
=
1
<<
VETH_EOF_SHIFT
;
msg
->
data
.
eofmask
=
1
<<
VETH_EOF_SHIFT
;
set_bit
(
0
,
&
(
msg
->
in_use
));
rc
=
veth_signaldata
(
cnx
,
V
ethEventTypeFrames
,
msg
->
token
,
&
msg
->
data
);
rc
=
veth_signaldata
(
cnx
,
V
ETH_EVENT_FRAMES
,
msg
->
token
,
&
msg
->
data
);
if
(
rc
!=
HvLpEvent_Rc_Good
)
if
(
rc
!=
HvLpEvent_Rc_Good
)
goto
recycle_and_drop
;
goto
recycle_and_drop
;
/* If the timer's not already running, start it now. */
if
(
0
==
cnx
->
outstanding_tx
)
mod_timer
(
&
cnx
->
reset_timer
,
jiffies
+
cnx
->
reset_timeout
);
cnx
->
last_contact
=
jiffies
;
cnx
->
outstanding_tx
++
;
if
(
veth_stack_is_empty
(
cnx
))
veth_stop_queues
(
cnx
);
no_error:
spin_unlock_irqrestore
(
&
cnx
->
lock
,
flags
);
spin_unlock_irqrestore
(
&
cnx
->
lock
,
flags
);
return
0
;
return
0
;
recycle_and_drop:
recycle_and_drop:
msg
->
skb
=
NULL
;
/* need to set in use to make veth_recycle_msg in case this
* was a mapping failure */
set_bit
(
0
,
&
msg
->
in_use
);
veth_recycle_msg
(
cnx
,
msg
);
veth_recycle_msg
(
cnx
,
msg
);
drop:
drop:
port
->
stats
.
tx_errors
++
;
dev_kfree_skb
(
skb
);
spin_unlock_irqrestore
(
&
cnx
->
lock
,
flags
);
spin_unlock_irqrestore
(
&
cnx
->
lock
,
flags
);
return
err
;
return
1
;
}
}
static
HvLpIndexMap
veth_transmit_to_many
(
struct
sk_buff
*
skb
,
static
void
veth_transmit_to_many
(
struct
sk_buff
*
skb
,
HvLpIndexMap
lpmask
,
HvLpIndexMap
lpmask
,
struct
net_device
*
dev
)
struct
net_device
*
dev
)
{
{
struct
veth_port
*
port
=
(
struct
veth_port
*
)
dev
->
priv
;
struct
veth_port
*
port
=
(
struct
veth_port
*
)
dev
->
priv
;
int
i
;
int
i
,
success
,
error
;
int
rc
;
success
=
error
=
0
;
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
i
++
)
{
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
i
++
)
{
if
((
lpmask
&
(
1
<<
i
))
==
0
)
if
((
lpmask
&
(
1
<<
i
))
==
0
)
continue
;
continue
;
rc
=
veth_transmit_to_one
(
skb_get
(
skb
),
i
,
dev
);
if
(
veth_transmit_to_one
(
skb
,
i
,
dev
))
if
(
!
rc
)
error
=
1
;
lpmask
&=
~
(
1
<<
i
);
else
success
=
1
;
}
}
if
(
!
lpmask
)
{
if
(
error
)
port
->
stats
.
tx_errors
++
;
if
(
success
)
{
port
->
stats
.
tx_packets
++
;
port
->
stats
.
tx_packets
++
;
port
->
stats
.
tx_bytes
+=
skb
->
len
;
port
->
stats
.
tx_bytes
+=
skb
->
len
;
}
}
return
lpmask
;
}
}
static
int
veth_start_xmit
(
struct
sk_buff
*
skb
,
struct
net_device
*
dev
)
static
int
veth_start_xmit
(
struct
sk_buff
*
skb
,
struct
net_device
*
dev
)
{
{
unsigned
char
*
frame
=
skb
->
data
;
unsigned
char
*
frame
=
skb
->
data
;
struct
veth_port
*
port
=
(
struct
veth_port
*
)
dev
->
priv
;
struct
veth_port
*
port
=
(
struct
veth_port
*
)
dev
->
priv
;
unsigned
long
flags
;
HvLpIndexMap
lpmask
;
HvLpIndexMap
lpmask
;
if
(
!
(
frame
[
0
]
&
0x01
))
{
if
(
!
(
frame
[
0
]
&
0x01
))
{
...
@@ -1020,42 +1220,25 @@ static int veth_start_xmit(struct sk_buff *skb, struct net_device *dev)
...
@@ -1020,42 +1220,25 @@ static int veth_start_xmit(struct sk_buff *skb, struct net_device *dev)
lpmask
=
port
->
lpar_map
;
lpmask
=
port
->
lpar_map
;
}
}
spin_lock_irqsave
(
&
port
->
pending_gate
,
flags
);
veth_transmit_to_many
(
skb
,
lpmask
,
dev
);
lpmask
=
veth_transmit_to_many
(
skb
,
lpmask
,
dev
);
dev
->
trans_start
=
jiffies
;
if
(
!
lpmask
)
{
dev_kfree_skb
(
skb
);
}
else
{
if
(
port
->
pending_skb
)
{
veth_error
(
"%s: Tx while skb was pending!
\n
"
,
dev
->
name
);
dev_kfree_skb
(
skb
);
dev_kfree_skb
(
skb
);
spin_unlock_irqrestore
(
&
port
->
pending_gate
,
flags
);
return
1
;
}
port
->
pending_skb
=
skb
;
port
->
pending_lpmask
=
lpmask
;
netif_stop_queue
(
dev
);
}
spin_unlock_irqrestore
(
&
port
->
pending_gate
,
flags
);
return
0
;
return
0
;
}
}
/* You must hold the connection's lock when you call this function. */
static
void
veth_recycle_msg
(
struct
veth_lpar_connection
*
cnx
,
static
void
veth_recycle_msg
(
struct
veth_lpar_connection
*
cnx
,
struct
veth_msg
*
msg
)
struct
veth_msg
*
msg
)
{
{
u32
dma_address
,
dma_length
;
u32
dma_address
,
dma_length
;
if
(
test_and_clear_bit
(
0
,
&
msg
->
in_use
))
{
if
(
msg
->
in_use
)
{
msg
->
in_use
=
0
;
dma_address
=
msg
->
data
.
addr
[
0
];
dma_address
=
msg
->
data
.
addr
[
0
];
dma_length
=
msg
->
data
.
len
[
0
];
dma_length
=
msg
->
data
.
len
[
0
];
if
(
!
dma_mapping_error
(
dma_address
))
dma_unmap_single
(
msg
->
dev
,
dma_address
,
dma_length
,
dma_unmap_single
(
msg
->
dev
,
dma_address
,
dma_length
,
DMA_TO_DEVICE
);
DMA_TO_DEVICE
);
...
@@ -1066,15 +1249,16 @@ static void veth_recycle_msg(struct veth_lpar_connection *cnx,
...
@@ -1066,15 +1249,16 @@ static void veth_recycle_msg(struct veth_lpar_connection *cnx,
memset
(
&
msg
->
data
,
0
,
sizeof
(
msg
->
data
));
memset
(
&
msg
->
data
,
0
,
sizeof
(
msg
->
data
));
veth_stack_push
(
cnx
,
msg
);
veth_stack_push
(
cnx
,
msg
);
}
else
}
else
if
(
cnx
->
state
&
VETH_STATE_OPEN
)
{
if
(
cnx
->
state
&
VETH_STATE_OPEN
)
veth_error
(
"Non-pending frame (# %d) acked by LPAR %d.
\n
"
,
veth_error
(
"Bogus frames ack from lpar %d (#%d)
\n
"
,
cnx
->
remote_lp
,
msg
->
token
);
cnx
->
remote_lp
,
msg
->
token
);
}
}
}
static
void
veth_
flush_pending
(
struct
veth_lpar_connection
*
cnx
)
static
void
veth_
wake_queues
(
struct
veth_lpar_connection
*
cnx
)
{
{
int
i
;
int
i
;
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDVIRTUALLANS
;
i
++
)
{
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDVIRTUALLANS
;
i
++
)
{
struct
net_device
*
dev
=
veth_dev
[
i
];
struct
net_device
*
dev
=
veth_dev
[
i
];
struct
veth_port
*
port
;
struct
veth_port
*
port
;
...
@@ -1088,22 +1272,79 @@ static void veth_flush_pending(struct veth_lpar_connection *cnx)
...
@@ -1088,22 +1272,79 @@ static void veth_flush_pending(struct veth_lpar_connection *cnx)
if
(
!
(
port
->
lpar_map
&
(
1
<<
cnx
->
remote_lp
)))
if
(
!
(
port
->
lpar_map
&
(
1
<<
cnx
->
remote_lp
)))
continue
;
continue
;
spin_lock_irqsave
(
&
port
->
pending_gate
,
flags
);
spin_lock_irqsave
(
&
port
->
queue_lock
,
flags
);
if
(
port
->
pending_skb
)
{
port
->
pending_lpmask
=
port
->
stopped_map
&=
~
(
1
<<
cnx
->
remote_lp
);
veth_transmit_to_many
(
port
->
pending_skb
,
port
->
pending_lpmask
,
if
(
0
==
port
->
stopped_map
&&
netif_queue_stopped
(
dev
))
{
dev
);
veth_debug
(
"cnx %d: woke queue for %s.
\n
"
,
if
(
!
port
->
pending_lpmask
)
{
cnx
->
remote_lp
,
dev
->
name
);
dev_kfree_skb_any
(
port
->
pending_skb
);
port
->
pending_skb
=
NULL
;
netif_wake_queue
(
dev
);
netif_wake_queue
(
dev
);
}
}
spin_unlock_irqrestore
(
&
port
->
queue_lock
,
flags
);
}
}
spin_unlock_irqrestore
(
&
port
->
pending_gate
,
flags
);
}
static
void
veth_stop_queues
(
struct
veth_lpar_connection
*
cnx
)
{
int
i
;
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDVIRTUALLANS
;
i
++
)
{
struct
net_device
*
dev
=
veth_dev
[
i
];
struct
veth_port
*
port
;
if
(
!
dev
)
continue
;
port
=
(
struct
veth_port
*
)
dev
->
priv
;
/* If this cnx is not on the vlan for this port, continue */
if
(
!
(
port
->
lpar_map
&
(
1
<<
cnx
->
remote_lp
)))
continue
;
spin_lock
(
&
port
->
queue_lock
);
netif_stop_queue
(
dev
);
port
->
stopped_map
|=
(
1
<<
cnx
->
remote_lp
);
veth_debug
(
"cnx %d: stopped queue for %s, map = 0x%x.
\n
"
,
cnx
->
remote_lp
,
dev
->
name
,
port
->
stopped_map
);
spin_unlock
(
&
port
->
queue_lock
);
}
}
}
}
static
void
veth_timed_reset
(
unsigned
long
ptr
)
{
struct
veth_lpar_connection
*
cnx
=
(
struct
veth_lpar_connection
*
)
ptr
;
unsigned
long
trigger_time
,
flags
;
/* FIXME is it possible this fires after veth_stop_connection()?
* That would reschedule the statemachine for 5 seconds and probably
* execute it after the module's been unloaded. Hmm. */
spin_lock_irqsave
(
&
cnx
->
lock
,
flags
);
if
(
cnx
->
outstanding_tx
>
0
)
{
trigger_time
=
cnx
->
last_contact
+
cnx
->
reset_timeout
;
if
(
trigger_time
<
jiffies
)
{
cnx
->
state
|=
VETH_STATE_RESET
;
veth_kick_statemachine
(
cnx
);
veth_error
(
"%d packets not acked by LPAR %d within %d "
"seconds, resetting.
\n
"
,
cnx
->
outstanding_tx
,
cnx
->
remote_lp
,
cnx
->
reset_timeout
/
HZ
);
}
else
{
/* Reschedule the timer */
trigger_time
=
jiffies
+
cnx
->
reset_timeout
;
mod_timer
(
&
cnx
->
reset_timer
,
trigger_time
);
}
}
spin_unlock_irqrestore
(
&
cnx
->
lock
,
flags
);
}
/*
/*
* Rx path
* Rx path
*/
*/
...
@@ -1117,12 +1358,9 @@ static inline int veth_frame_wanted(struct veth_port *port, u64 mac_addr)
...
@@ -1117,12 +1358,9 @@ static inline int veth_frame_wanted(struct veth_port *port, u64 mac_addr)
if
(
(
mac_addr
==
port
->
mac_addr
)
||
(
mac_addr
==
0xffffffffffff0000
)
)
if
(
(
mac_addr
==
port
->
mac_addr
)
||
(
mac_addr
==
0xffffffffffff0000
)
)
return
1
;
return
1
;
if
(
!
(((
char
*
)
&
mac_addr
)[
0
]
&
0x01
))
return
0
;
read_lock_irqsave
(
&
port
->
mcast_gate
,
flags
);
read_lock_irqsave
(
&
port
->
mcast_gate
,
flags
);
if
(
port
->
promiscuous
||
port
->
all_mcast
)
{
if
(
port
->
promiscuous
)
{
wanted
=
1
;
wanted
=
1
;
goto
out
;
goto
out
;
}
}
...
@@ -1175,21 +1413,21 @@ static void veth_flush_acks(struct veth_lpar_connection *cnx)
...
@@ -1175,21 +1413,21 @@ static void veth_flush_acks(struct veth_lpar_connection *cnx)
{
{
HvLpEvent_Rc
rc
;
HvLpEvent_Rc
rc
;
rc
=
veth_signaldata
(
cnx
,
V
ethEventTypeFramesAck
,
rc
=
veth_signaldata
(
cnx
,
V
ETH_EVENT_FRAMES_ACK
,
0
,
&
cnx
->
pending_acks
);
0
,
&
cnx
->
pending_acks
);
if
(
rc
!=
HvLpEvent_Rc_Good
)
if
(
rc
!=
HvLpEvent_Rc_Good
)
veth_error
(
"
Error 0x%x acking frames from lpar %d!
\n
"
,
veth_error
(
"
Failed acking frames from LPAR %d, rc = %d
\n
"
,
(
unsigned
)
rc
,
cnx
->
remote_lp
);
cnx
->
remote_lp
,
(
int
)
rc
);
cnx
->
num_pending_acks
=
0
;
cnx
->
num_pending_acks
=
0
;
memset
(
&
cnx
->
pending_acks
,
0xff
,
sizeof
(
cnx
->
pending_acks
));
memset
(
&
cnx
->
pending_acks
,
0xff
,
sizeof
(
cnx
->
pending_acks
));
}
}
static
void
veth_receive
(
struct
veth_lpar_connection
*
cnx
,
static
void
veth_receive
(
struct
veth_lpar_connection
*
cnx
,
struct
VethLpE
vent
*
event
)
struct
veth_lpe
vent
*
event
)
{
{
struct
VethFramesD
ata
*
senddata
=
&
event
->
u
.
frames_data
;
struct
veth_frames_d
ata
*
senddata
=
&
event
->
u
.
frames_data
;
int
startchunk
=
0
;
int
startchunk
=
0
;
int
nchunks
;
int
nchunks
;
unsigned
long
flags
;
unsigned
long
flags
;
...
@@ -1216,9 +1454,10 @@ static void veth_receive(struct veth_lpar_connection *cnx,
...
@@ -1216,9 +1454,10 @@ static void veth_receive(struct veth_lpar_connection *cnx,
/* make sure that we have at least 1 EOF entry in the
/* make sure that we have at least 1 EOF entry in the
* remaining entries */
* remaining entries */
if
(
!
(
senddata
->
eofmask
>>
(
startchunk
+
VETH_EOF_SHIFT
)))
{
if
(
!
(
senddata
->
eofmask
>>
(
startchunk
+
VETH_EOF_SHIFT
)))
{
veth_error
(
"missing EOF frag in event "
veth_error
(
"Missing EOF fragment in event "
"eofmask=0x%x startchunk=%d
\n
"
,
"eofmask = 0x%x startchunk = %d
\n
"
,
(
unsigned
)
senddata
->
eofmask
,
startchunk
);
(
unsigned
)
senddata
->
eofmask
,
startchunk
);
break
;
break
;
}
}
...
@@ -1237,8 +1476,9 @@ static void veth_receive(struct veth_lpar_connection *cnx,
...
@@ -1237,8 +1476,9 @@ static void veth_receive(struct veth_lpar_connection *cnx,
/* nchunks == # of chunks in this frame */
/* nchunks == # of chunks in this frame */
if
((
length
-
ETH_HLEN
)
>
VETH_MAX_MTU
)
{
if
((
length
-
ETH_HLEN
)
>
VETH_MAX_MTU
)
{
veth_error
(
"Received oversize frame from lpar %d "
veth_error
(
"Received oversize frame from LPAR %d "
"(length=%d)
\n
"
,
cnx
->
remote_lp
,
length
);
"(length = %d)
\n
"
,
cnx
->
remote_lp
,
length
);
continue
;
continue
;
}
}
...
@@ -1331,15 +1571,33 @@ static void veth_timed_ack(unsigned long ptr)
...
@@ -1331,15 +1571,33 @@ static void veth_timed_ack(unsigned long ptr)
static
int
veth_remove
(
struct
vio_dev
*
vdev
)
static
int
veth_remove
(
struct
vio_dev
*
vdev
)
{
{
int
i
=
vdev
->
unit_address
;
struct
veth_lpar_connection
*
cnx
;
struct
net_device
*
dev
;
struct
net_device
*
dev
;
struct
veth_port
*
port
;
int
i
;
dev
=
veth_dev
[
vdev
->
unit_address
];
if
(
!
dev
)
return
0
;
port
=
netdev_priv
(
dev
);
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
i
++
)
{
cnx
=
veth_cnx
[
i
];
dev
=
veth_dev
[
i
];
if
(
cnx
&&
(
port
->
lpar_map
&
(
1
<<
i
)))
{
if
(
dev
!=
NULL
)
{
/* Drop our reference to connections on our VLAN */
veth_dev
[
i
]
=
NULL
;
kobject_put
(
&
cnx
->
kobject
);
}
}
veth_dev
[
vdev
->
unit_address
]
=
NULL
;
kobject_del
(
&
port
->
kobject
);
kobject_put
(
&
port
->
kobject
);
unregister_netdev
(
dev
);
unregister_netdev
(
dev
);
free_netdev
(
dev
);
free_netdev
(
dev
);
}
return
0
;
return
0
;
}
}
...
@@ -1347,6 +1605,7 @@ static int veth_probe(struct vio_dev *vdev, const struct vio_device_id *id)
...
@@ -1347,6 +1605,7 @@ static int veth_probe(struct vio_dev *vdev, const struct vio_device_id *id)
{
{
int
i
=
vdev
->
unit_address
;
int
i
=
vdev
->
unit_address
;
struct
net_device
*
dev
;
struct
net_device
*
dev
;
struct
veth_port
*
port
;
dev
=
veth_probe_one
(
i
,
&
vdev
->
dev
);
dev
=
veth_probe_one
(
i
,
&
vdev
->
dev
);
if
(
dev
==
NULL
)
{
if
(
dev
==
NULL
)
{
...
@@ -1355,11 +1614,23 @@ static int veth_probe(struct vio_dev *vdev, const struct vio_device_id *id)
...
@@ -1355,11 +1614,23 @@ static int veth_probe(struct vio_dev *vdev, const struct vio_device_id *id)
}
}
veth_dev
[
i
]
=
dev
;
veth_dev
[
i
]
=
dev
;
/* Start the state machine on each connection, to commence
port
=
(
struct
veth_port
*
)
netdev_priv
(
dev
);
* link negotiation */
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
i
++
)
/* Start the state machine on each connection on this vlan. If we're
if
(
veth_cnx
[
i
])
* the first dev to do so this will commence link negotiation */
veth_kick_statemachine
(
veth_cnx
[
i
]);
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
i
++
)
{
struct
veth_lpar_connection
*
cnx
;
if
(
!
(
port
->
lpar_map
&
(
1
<<
i
)))
continue
;
cnx
=
veth_cnx
[
i
];
if
(
!
cnx
)
continue
;
kobject_get
(
&
cnx
->
kobject
);
veth_kick_statemachine
(
cnx
);
}
return
0
;
return
0
;
}
}
...
@@ -1375,7 +1646,7 @@ static struct vio_device_id veth_device_table[] __devinitdata = {
...
@@ -1375,7 +1646,7 @@ static struct vio_device_id veth_device_table[] __devinitdata = {
MODULE_DEVICE_TABLE
(
vio
,
veth_device_table
);
MODULE_DEVICE_TABLE
(
vio
,
veth_device_table
);
static
struct
vio_driver
veth_driver
=
{
static
struct
vio_driver
veth_driver
=
{
.
name
=
"iseries_veth"
,
.
name
=
DRV_NAME
,
.
id_table
=
veth_device_table
,
.
id_table
=
veth_device_table
,
.
probe
=
veth_probe
,
.
probe
=
veth_probe
,
.
remove
=
veth_remove
.
remove
=
veth_remove
...
@@ -1388,29 +1659,29 @@ static struct vio_driver veth_driver = {
...
@@ -1388,29 +1659,29 @@ static struct vio_driver veth_driver = {
void
__exit
veth_module_cleanup
(
void
)
void
__exit
veth_module_cleanup
(
void
)
{
{
int
i
;
int
i
;
struct
veth_lpar_connection
*
cnx
;
/* Stop the queues first to stop any new packets being sent. */
/* Disconnect our "irq" to stop events coming from the Hypervisor. */
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDVIRTUALLANS
;
i
++
)
if
(
veth_dev
[
i
])
netif_stop_queue
(
veth_dev
[
i
]);
/* Stop the connections before we unregister the driver. This
* ensures there's no skbs lying around holding the device open. */
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
++
i
)
veth_stop_connection
(
i
);
HvLpEvent_unregisterHandler
(
HvLpEvent_Type_VirtualLan
);
HvLpEvent_unregisterHandler
(
HvLpEvent_Type_VirtualLan
);
/* Hypervisor callbacks may have scheduled more work while we
/* Make sure any work queued from Hypervisor callbacks is finished. */
* were stoping connections. Now that we've disconnected from
* the hypervisor make sure everything's finished. */
flush_scheduled_work
();
flush_scheduled_work
();
vio_unregister_driver
(
&
veth_driver
);
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
++
i
)
{
cnx
=
veth_cnx
[
i
];
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
++
i
)
if
(
!
cnx
)
veth_destroy_connection
(
i
);
continue
;
/* Remove the connection from sysfs */
kobject_del
(
&
cnx
->
kobject
);
/* Drop the driver's reference to the connection */
kobject_put
(
&
cnx
->
kobject
);
}
/* Unregister the driver, which will close all the netdevs and stop
* the connections when they're no longer referenced. */
vio_unregister_driver
(
&
veth_driver
);
}
}
module_exit
(
veth_module_cleanup
);
module_exit
(
veth_module_cleanup
);
...
@@ -1423,15 +1694,37 @@ int __init veth_module_init(void)
...
@@ -1423,15 +1694,37 @@ int __init veth_module_init(void)
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
++
i
)
{
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
++
i
)
{
rc
=
veth_init_connection
(
i
);
rc
=
veth_init_connection
(
i
);
if
(
rc
!=
0
)
{
if
(
rc
!=
0
)
veth_module_cleanup
();
goto
error
;
return
rc
;
}
}
}
HvLpEvent_registerHandler
(
HvLpEvent_Type_VirtualLan
,
HvLpEvent_registerHandler
(
HvLpEvent_Type_VirtualLan
,
&
veth_handle_event
);
&
veth_handle_event
);
return
vio_register_driver
(
&
veth_driver
);
rc
=
vio_register_driver
(
&
veth_driver
);
if
(
rc
!=
0
)
goto
error
;
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
++
i
)
{
struct
kobject
*
kobj
;
if
(
!
veth_cnx
[
i
])
continue
;
kobj
=
&
veth_cnx
[
i
]
->
kobject
;
kobj
->
parent
=
&
veth_driver
.
driver
.
kobj
;
/* If the add failes, complain but otherwise continue */
if
(
0
!=
kobject_add
(
kobj
))
veth_error
(
"cnx %d: Failed adding to sysfs.
\n
"
,
i
);
}
return
0
;
error:
for
(
i
=
0
;
i
<
HVMAXARCHITECTEDLPS
;
++
i
)
{
veth_destroy_connection
(
veth_cnx
[
i
]);
}
return
rc
;
}
}
module_init
(
veth_module_init
);
module_init
(
veth_module_init
);
drivers/net/iseries_veth.h
deleted
100644 → 0
View file @
712fbdd3
/* File veth.h created by Kyle A. Lucke on Mon Aug 7 2000. */
#ifndef _ISERIES_VETH_H
#define _ISERIES_VETH_H
#define VethEventTypeCap (0)
#define VethEventTypeFrames (1)
#define VethEventTypeMonitor (2)
#define VethEventTypeFramesAck (3)
#define VETH_MAX_ACKS_PER_MSG (20)
#define VETH_MAX_FRAMES_PER_MSG (6)
struct
VethFramesData
{
u32
addr
[
VETH_MAX_FRAMES_PER_MSG
];
u16
len
[
VETH_MAX_FRAMES_PER_MSG
];
u32
eofmask
;
};
#define VETH_EOF_SHIFT (32-VETH_MAX_FRAMES_PER_MSG)
struct
VethFramesAckData
{
u16
token
[
VETH_MAX_ACKS_PER_MSG
];
};
struct
VethCapData
{
u8
caps_version
;
u8
rsvd1
;
u16
num_buffers
;
u16
ack_threshold
;
u16
rsvd2
;
u32
ack_timeout
;
u32
rsvd3
;
u64
rsvd4
[
3
];
};
struct
VethLpEvent
{
struct
HvLpEvent
base_event
;
union
{
struct
VethCapData
caps_data
;
struct
VethFramesData
frames_data
;
struct
VethFramesAckData
frames_ack_data
;
}
u
;
};
#endif
/* _ISERIES_VETH_H */
drivers/net/tulip/de2104x.c
View file @
44757223
...
@@ -1934,7 +1934,7 @@ static int __init de_init_one (struct pci_dev *pdev,
...
@@ -1934,7 +1934,7 @@ static int __init de_init_one (struct pci_dev *pdev,
struct
de_private
*
de
;
struct
de_private
*
de
;
int
rc
;
int
rc
;
void
__iomem
*
regs
;
void
__iomem
*
regs
;
long
pciaddr
;
unsigned
long
pciaddr
;
static
int
board_idx
=
-
1
;
static
int
board_idx
=
-
1
;
board_idx
++
;
board_idx
++
;
...
...
drivers/net/tulip/tulip_core.c
View file @
44757223
...
@@ -238,6 +238,7 @@ static struct pci_device_id tulip_pci_tbl[] = {
...
@@ -238,6 +238,7 @@ static struct pci_device_id tulip_pci_tbl[] = {
{
0x17B3
,
0xAB08
,
PCI_ANY_ID
,
PCI_ANY_ID
,
0
,
0
,
COMET
},
{
0x17B3
,
0xAB08
,
PCI_ANY_ID
,
PCI_ANY_ID
,
0
,
0
,
COMET
},
{
0x10b7
,
0x9300
,
PCI_ANY_ID
,
PCI_ANY_ID
,
0
,
0
,
COMET
},
/* 3Com 3CSOHO100B-TX */
{
0x10b7
,
0x9300
,
PCI_ANY_ID
,
PCI_ANY_ID
,
0
,
0
,
COMET
},
/* 3Com 3CSOHO100B-TX */
{
0x14ea
,
0xab08
,
PCI_ANY_ID
,
PCI_ANY_ID
,
0
,
0
,
COMET
},
/* Planex FNW-3602-TX */
{
0x14ea
,
0xab08
,
PCI_ANY_ID
,
PCI_ANY_ID
,
0
,
0
,
COMET
},
/* Planex FNW-3602-TX */
{
0x1414
,
0x0002
,
PCI_ANY_ID
,
PCI_ANY_ID
,
0
,
0
,
COMET
},
{
}
/* terminate list */
{
}
/* terminate list */
};
};
MODULE_DEVICE_TABLE
(
pci
,
tulip_pci_tbl
);
MODULE_DEVICE_TABLE
(
pci
,
tulip_pci_tbl
);
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment