e1000e: Perform software segmentation for loopback
e1000e didn't perform software segmentation for loopback if virtio-net header is enabled, which is wrong. To fix the problem, introduce net_tx_pkt_send_custom(), which allows the caller to specify whether offloading should be assumed or not. net_tx_pkt_send_custom() also allows the caller to provide a custom sending function. Packets with virtio-net headers and ones without virtio-net headers will be provided at the same time so the function can choose the preferred version. In case of e1000e loopback, it prefers to have virtio-net headers as they allows to skip the checksum verification if VIRTIO_NET_HDR_F_DATA_VALID is set. Signed-off-by: Akihiko Odaki <akihiko.odaki@daynix.com> Signed-off-by: Jason Wang <jasowang@redhat.com>
This commit is contained in:
parent
aac8f89dba
commit
ffbd2dbd8e
@ -61,6 +61,10 @@ union e1000_rx_desc_union {
|
|||||||
union e1000_rx_desc_packet_split packet_split;
|
union e1000_rx_desc_packet_split packet_split;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
static ssize_t
|
||||||
|
e1000e_receive_internal(E1000ECore *core, const struct iovec *iov, int iovcnt,
|
||||||
|
bool has_vnet);
|
||||||
|
|
||||||
static inline void
|
static inline void
|
||||||
e1000e_set_interrupt_cause(E1000ECore *core, uint32_t val);
|
e1000e_set_interrupt_cause(E1000ECore *core, uint32_t val);
|
||||||
|
|
||||||
@ -655,6 +659,15 @@ e1000e_setup_tx_offloads(E1000ECore *core, struct e1000e_tx *tx)
|
|||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void e1000e_tx_pkt_callback(void *core,
|
||||||
|
const struct iovec *iov,
|
||||||
|
int iovcnt,
|
||||||
|
const struct iovec *virt_iov,
|
||||||
|
int virt_iovcnt)
|
||||||
|
{
|
||||||
|
e1000e_receive_internal(core, virt_iov, virt_iovcnt, true);
|
||||||
|
}
|
||||||
|
|
||||||
static bool
|
static bool
|
||||||
e1000e_tx_pkt_send(E1000ECore *core, struct e1000e_tx *tx, int queue_index)
|
e1000e_tx_pkt_send(E1000ECore *core, struct e1000e_tx *tx, int queue_index)
|
||||||
{
|
{
|
||||||
@ -669,7 +682,8 @@ e1000e_tx_pkt_send(E1000ECore *core, struct e1000e_tx *tx, int queue_index)
|
|||||||
|
|
||||||
if ((core->phy[0][MII_BMCR] & MII_BMCR_LOOPBACK) ||
|
if ((core->phy[0][MII_BMCR] & MII_BMCR_LOOPBACK) ||
|
||||||
((core->mac[RCTL] & E1000_RCTL_LBM_MAC) == E1000_RCTL_LBM_MAC)) {
|
((core->mac[RCTL] & E1000_RCTL_LBM_MAC) == E1000_RCTL_LBM_MAC)) {
|
||||||
return net_tx_pkt_send_loopback(tx->tx_pkt, queue);
|
return net_tx_pkt_send_custom(tx->tx_pkt, false,
|
||||||
|
e1000e_tx_pkt_callback, core);
|
||||||
} else {
|
} else {
|
||||||
return net_tx_pkt_send(tx->tx_pkt, queue);
|
return net_tx_pkt_send(tx->tx_pkt, queue);
|
||||||
}
|
}
|
||||||
@ -1674,6 +1688,13 @@ e1000e_rx_fix_l4_csum(E1000ECore *core, struct NetRxPkt *pkt)
|
|||||||
|
|
||||||
ssize_t
|
ssize_t
|
||||||
e1000e_receive_iov(E1000ECore *core, const struct iovec *iov, int iovcnt)
|
e1000e_receive_iov(E1000ECore *core, const struct iovec *iov, int iovcnt)
|
||||||
|
{
|
||||||
|
return e1000e_receive_internal(core, iov, iovcnt, core->has_vnet);
|
||||||
|
}
|
||||||
|
|
||||||
|
static ssize_t
|
||||||
|
e1000e_receive_internal(E1000ECore *core, const struct iovec *iov, int iovcnt,
|
||||||
|
bool has_vnet)
|
||||||
{
|
{
|
||||||
static const int maximum_ethernet_hdr_len = (ETH_HLEN + 4);
|
static const int maximum_ethernet_hdr_len = (ETH_HLEN + 4);
|
||||||
|
|
||||||
@ -1696,9 +1717,11 @@ e1000e_receive_iov(E1000ECore *core, const struct iovec *iov, int iovcnt)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/* Pull virtio header in */
|
/* Pull virtio header in */
|
||||||
if (core->has_vnet) {
|
if (has_vnet) {
|
||||||
net_rx_pkt_set_vhdr_iovec(core->rx_pkt, iov, iovcnt);
|
net_rx_pkt_set_vhdr_iovec(core->rx_pkt, iov, iovcnt);
|
||||||
iov_ofs = sizeof(struct virtio_net_hdr);
|
iov_ofs = sizeof(struct virtio_net_hdr);
|
||||||
|
} else {
|
||||||
|
net_rx_pkt_unset_vhdr(core->rx_pkt);
|
||||||
}
|
}
|
||||||
|
|
||||||
filter_buf = iov->iov_base + iov_ofs;
|
filter_buf = iov->iov_base + iov_ofs;
|
||||||
|
@ -463,6 +463,13 @@ void net_rx_pkt_set_vhdr_iovec(struct NetRxPkt *pkt,
|
|||||||
iov_to_buf(iov, iovcnt, 0, &pkt->virt_hdr, sizeof pkt->virt_hdr);
|
iov_to_buf(iov, iovcnt, 0, &pkt->virt_hdr, sizeof pkt->virt_hdr);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void net_rx_pkt_unset_vhdr(struct NetRxPkt *pkt)
|
||||||
|
{
|
||||||
|
assert(pkt);
|
||||||
|
|
||||||
|
memset(&pkt->virt_hdr, 0, sizeof(pkt->virt_hdr));
|
||||||
|
}
|
||||||
|
|
||||||
bool net_rx_pkt_is_vlan_stripped(struct NetRxPkt *pkt)
|
bool net_rx_pkt_is_vlan_stripped(struct NetRxPkt *pkt)
|
||||||
{
|
{
|
||||||
assert(pkt);
|
assert(pkt);
|
||||||
|
@ -312,6 +312,14 @@ void net_rx_pkt_set_vhdr(struct NetRxPkt *pkt,
|
|||||||
void net_rx_pkt_set_vhdr_iovec(struct NetRxPkt *pkt,
|
void net_rx_pkt_set_vhdr_iovec(struct NetRxPkt *pkt,
|
||||||
const struct iovec *iov, int iovcnt);
|
const struct iovec *iov, int iovcnt);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* unset vhdr data from packet context
|
||||||
|
*
|
||||||
|
* @pkt: packet
|
||||||
|
*
|
||||||
|
*/
|
||||||
|
void net_rx_pkt_unset_vhdr(struct NetRxPkt *pkt);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* save packet type in packet context
|
* save packet type in packet context
|
||||||
*
|
*
|
||||||
|
@ -53,8 +53,6 @@ struct NetTxPkt {
|
|||||||
uint16_t hdr_len;
|
uint16_t hdr_len;
|
||||||
eth_pkt_types_e packet_type;
|
eth_pkt_types_e packet_type;
|
||||||
uint8_t l4proto;
|
uint8_t l4proto;
|
||||||
|
|
||||||
bool is_loopback;
|
|
||||||
};
|
};
|
||||||
|
|
||||||
void net_tx_pkt_init(struct NetTxPkt **pkt, PCIDevice *pci_dev,
|
void net_tx_pkt_init(struct NetTxPkt **pkt, PCIDevice *pci_dev,
|
||||||
@ -508,12 +506,6 @@ static void net_tx_pkt_do_sw_csum(struct NetTxPkt *pkt)
|
|||||||
iov_from_buf(iov, iov_len, csum_offset, &csum, sizeof csum);
|
iov_from_buf(iov, iov_len, csum_offset, &csum, sizeof csum);
|
||||||
}
|
}
|
||||||
|
|
||||||
enum {
|
|
||||||
NET_TX_PKT_FRAGMENT_L2_HDR_POS = 0,
|
|
||||||
NET_TX_PKT_FRAGMENT_L3_HDR_POS,
|
|
||||||
NET_TX_PKT_FRAGMENT_HEADER_NUM
|
|
||||||
};
|
|
||||||
|
|
||||||
#define NET_MAX_FRAG_SG_LIST (64)
|
#define NET_MAX_FRAG_SG_LIST (64)
|
||||||
|
|
||||||
static size_t net_tx_pkt_fetch_fragment(struct NetTxPkt *pkt,
|
static size_t net_tx_pkt_fetch_fragment(struct NetTxPkt *pkt,
|
||||||
@ -522,7 +514,7 @@ static size_t net_tx_pkt_fetch_fragment(struct NetTxPkt *pkt,
|
|||||||
size_t fetched = 0;
|
size_t fetched = 0;
|
||||||
struct iovec *src = pkt->vec;
|
struct iovec *src = pkt->vec;
|
||||||
|
|
||||||
*dst_idx = NET_TX_PKT_FRAGMENT_HEADER_NUM;
|
*dst_idx = NET_TX_PKT_PL_START_FRAG;
|
||||||
|
|
||||||
while (fetched < IP_FRAG_ALIGN_SIZE(pkt->virt_hdr.gso_size)) {
|
while (fetched < IP_FRAG_ALIGN_SIZE(pkt->virt_hdr.gso_size)) {
|
||||||
|
|
||||||
@ -555,18 +547,22 @@ static size_t net_tx_pkt_fetch_fragment(struct NetTxPkt *pkt,
|
|||||||
return fetched;
|
return fetched;
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline void net_tx_pkt_sendv(struct NetTxPkt *pkt,
|
static void net_tx_pkt_sendv(
|
||||||
NetClientState *nc, const struct iovec *iov, int iov_cnt)
|
void *opaque, const struct iovec *iov, int iov_cnt,
|
||||||
|
const struct iovec *virt_iov, int virt_iov_cnt)
|
||||||
{
|
{
|
||||||
if (pkt->is_loopback) {
|
NetClientState *nc = opaque;
|
||||||
qemu_receive_packet_iov(nc, iov, iov_cnt);
|
|
||||||
|
if (qemu_get_using_vnet_hdr(nc->peer)) {
|
||||||
|
qemu_sendv_packet(nc, virt_iov, virt_iov_cnt);
|
||||||
} else {
|
} else {
|
||||||
qemu_sendv_packet(nc, iov, iov_cnt);
|
qemu_sendv_packet(nc, iov, iov_cnt);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static bool net_tx_pkt_do_sw_fragmentation(struct NetTxPkt *pkt,
|
static bool net_tx_pkt_do_sw_fragmentation(struct NetTxPkt *pkt,
|
||||||
NetClientState *nc)
|
NetTxPktCallback callback,
|
||||||
|
void *context)
|
||||||
{
|
{
|
||||||
struct iovec fragment[NET_MAX_FRAG_SG_LIST];
|
struct iovec fragment[NET_MAX_FRAG_SG_LIST];
|
||||||
size_t fragment_len = 0;
|
size_t fragment_len = 0;
|
||||||
@ -578,6 +574,10 @@ static bool net_tx_pkt_do_sw_fragmentation(struct NetTxPkt *pkt,
|
|||||||
int src_idx = NET_TX_PKT_PL_START_FRAG, dst_idx;
|
int src_idx = NET_TX_PKT_PL_START_FRAG, dst_idx;
|
||||||
size_t src_offset = 0;
|
size_t src_offset = 0;
|
||||||
size_t fragment_offset = 0;
|
size_t fragment_offset = 0;
|
||||||
|
struct virtio_net_hdr virt_hdr = {
|
||||||
|
.flags = pkt->virt_hdr.flags & VIRTIO_NET_HDR_F_NEEDS_CSUM ?
|
||||||
|
VIRTIO_NET_HDR_F_DATA_VALID : 0
|
||||||
|
};
|
||||||
|
|
||||||
l2_iov_base = pkt->vec[NET_TX_PKT_L2HDR_FRAG].iov_base;
|
l2_iov_base = pkt->vec[NET_TX_PKT_L2HDR_FRAG].iov_base;
|
||||||
l2_iov_len = pkt->vec[NET_TX_PKT_L2HDR_FRAG].iov_len;
|
l2_iov_len = pkt->vec[NET_TX_PKT_L2HDR_FRAG].iov_len;
|
||||||
@ -585,10 +585,12 @@ static bool net_tx_pkt_do_sw_fragmentation(struct NetTxPkt *pkt,
|
|||||||
l3_iov_len = pkt->vec[NET_TX_PKT_L3HDR_FRAG].iov_len;
|
l3_iov_len = pkt->vec[NET_TX_PKT_L3HDR_FRAG].iov_len;
|
||||||
|
|
||||||
/* Copy headers */
|
/* Copy headers */
|
||||||
fragment[NET_TX_PKT_FRAGMENT_L2_HDR_POS].iov_base = l2_iov_base;
|
fragment[NET_TX_PKT_VHDR_FRAG].iov_base = &virt_hdr;
|
||||||
fragment[NET_TX_PKT_FRAGMENT_L2_HDR_POS].iov_len = l2_iov_len;
|
fragment[NET_TX_PKT_VHDR_FRAG].iov_len = sizeof(virt_hdr);
|
||||||
fragment[NET_TX_PKT_FRAGMENT_L3_HDR_POS].iov_base = l3_iov_base;
|
fragment[NET_TX_PKT_L2HDR_FRAG].iov_base = l2_iov_base;
|
||||||
fragment[NET_TX_PKT_FRAGMENT_L3_HDR_POS].iov_len = l3_iov_len;
|
fragment[NET_TX_PKT_L2HDR_FRAG].iov_len = l2_iov_len;
|
||||||
|
fragment[NET_TX_PKT_L3HDR_FRAG].iov_base = l3_iov_base;
|
||||||
|
fragment[NET_TX_PKT_L3HDR_FRAG].iov_len = l3_iov_len;
|
||||||
|
|
||||||
|
|
||||||
/* Put as much data as possible and send */
|
/* Put as much data as possible and send */
|
||||||
@ -603,7 +605,9 @@ static bool net_tx_pkt_do_sw_fragmentation(struct NetTxPkt *pkt,
|
|||||||
|
|
||||||
eth_fix_ip4_checksum(l3_iov_base, l3_iov_len);
|
eth_fix_ip4_checksum(l3_iov_base, l3_iov_len);
|
||||||
|
|
||||||
net_tx_pkt_sendv(pkt, nc, fragment, dst_idx);
|
callback(context,
|
||||||
|
fragment + NET_TX_PKT_L2HDR_FRAG, dst_idx - NET_TX_PKT_L2HDR_FRAG,
|
||||||
|
fragment + NET_TX_PKT_VHDR_FRAG, dst_idx - NET_TX_PKT_VHDR_FRAG);
|
||||||
|
|
||||||
fragment_offset += fragment_len;
|
fragment_offset += fragment_len;
|
||||||
|
|
||||||
@ -614,12 +618,16 @@ static bool net_tx_pkt_do_sw_fragmentation(struct NetTxPkt *pkt,
|
|||||||
|
|
||||||
bool net_tx_pkt_send(struct NetTxPkt *pkt, NetClientState *nc)
|
bool net_tx_pkt_send(struct NetTxPkt *pkt, NetClientState *nc)
|
||||||
{
|
{
|
||||||
bool using_vnet_hdr = qemu_get_using_vnet_hdr(nc->peer);
|
bool offload = qemu_get_using_vnet_hdr(nc->peer);
|
||||||
|
return net_tx_pkt_send_custom(pkt, offload, net_tx_pkt_sendv, nc);
|
||||||
|
}
|
||||||
|
|
||||||
|
bool net_tx_pkt_send_custom(struct NetTxPkt *pkt, bool offload,
|
||||||
|
NetTxPktCallback callback, void *context)
|
||||||
|
{
|
||||||
assert(pkt);
|
assert(pkt);
|
||||||
|
|
||||||
if (!using_vnet_hdr &&
|
if (!offload && pkt->virt_hdr.flags & VIRTIO_NET_HDR_F_NEEDS_CSUM) {
|
||||||
pkt->virt_hdr.flags & VIRTIO_NET_HDR_F_NEEDS_CSUM) {
|
|
||||||
net_tx_pkt_do_sw_csum(pkt);
|
net_tx_pkt_do_sw_csum(pkt);
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -635,28 +643,16 @@ bool net_tx_pkt_send(struct NetTxPkt *pkt, NetClientState *nc)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if (using_vnet_hdr ||
|
if (offload || pkt->virt_hdr.gso_type == VIRTIO_NET_HDR_GSO_NONE) {
|
||||||
pkt->virt_hdr.gso_type == VIRTIO_NET_HDR_GSO_NONE) {
|
|
||||||
int index = using_vnet_hdr ?
|
|
||||||
NET_TX_PKT_VHDR_FRAG : NET_TX_PKT_L2HDR_FRAG;
|
|
||||||
net_tx_pkt_fix_ip6_payload_len(pkt);
|
net_tx_pkt_fix_ip6_payload_len(pkt);
|
||||||
net_tx_pkt_sendv(pkt, nc, pkt->vec + index,
|
callback(context, pkt->vec + NET_TX_PKT_L2HDR_FRAG,
|
||||||
pkt->payload_frags + NET_TX_PKT_PL_START_FRAG - index);
|
pkt->payload_frags + NET_TX_PKT_PL_START_FRAG - NET_TX_PKT_L2HDR_FRAG,
|
||||||
|
pkt->vec + NET_TX_PKT_VHDR_FRAG,
|
||||||
|
pkt->payload_frags + NET_TX_PKT_PL_START_FRAG - NET_TX_PKT_VHDR_FRAG);
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
return net_tx_pkt_do_sw_fragmentation(pkt, nc);
|
return net_tx_pkt_do_sw_fragmentation(pkt, callback, context);
|
||||||
}
|
|
||||||
|
|
||||||
bool net_tx_pkt_send_loopback(struct NetTxPkt *pkt, NetClientState *nc)
|
|
||||||
{
|
|
||||||
bool res;
|
|
||||||
|
|
||||||
pkt->is_loopback = true;
|
|
||||||
res = net_tx_pkt_send(pkt, nc);
|
|
||||||
pkt->is_loopback = false;
|
|
||||||
|
|
||||||
return res;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
void net_tx_pkt_fix_ip6_payload_len(struct NetTxPkt *pkt)
|
void net_tx_pkt_fix_ip6_payload_len(struct NetTxPkt *pkt)
|
||||||
|
@ -26,6 +26,8 @@
|
|||||||
|
|
||||||
struct NetTxPkt;
|
struct NetTxPkt;
|
||||||
|
|
||||||
|
typedef void (* NetTxPktCallback)(void *, const struct iovec *, int, const struct iovec *, int);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Init function for tx packet functionality
|
* Init function for tx packet functionality
|
||||||
*
|
*
|
||||||
@ -161,15 +163,16 @@ void net_tx_pkt_reset(struct NetTxPkt *pkt);
|
|||||||
bool net_tx_pkt_send(struct NetTxPkt *pkt, NetClientState *nc);
|
bool net_tx_pkt_send(struct NetTxPkt *pkt, NetClientState *nc);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Redirect packet directly to receive path (emulate loopback phy).
|
* Send packet with a custom function.
|
||||||
* Handles sw offloads if vhdr is not supported.
|
|
||||||
*
|
*
|
||||||
* @pkt: packet
|
* @pkt: packet
|
||||||
* @nc: NetClientState
|
* @offload: whether the callback implements offloading
|
||||||
|
* @callback: a function to be called back for each transformed packet
|
||||||
|
* @context: a pointer to be passed to the callback.
|
||||||
* @ret: operation result
|
* @ret: operation result
|
||||||
*
|
|
||||||
*/
|
*/
|
||||||
bool net_tx_pkt_send_loopback(struct NetTxPkt *pkt, NetClientState *nc);
|
bool net_tx_pkt_send_custom(struct NetTxPkt *pkt, bool offload,
|
||||||
|
NetTxPktCallback callback, void *context);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* parse raw packet data and analyze offload requirements.
|
* parse raw packet data and analyze offload requirements.
|
||||||
|
Loading…
Reference in New Issue
Block a user