mirror of
https://github.com/torvalds/linux.git
synced 2024-12-29 06:12:08 +00:00
tcp: change tcp_skb_pcount() location
Our goal is to access no more than one cache line access per skb in a write or receive queue when doing the various walks. After recent TCP_SKB_CB() reorganizations, it is almost done. Last part is tcp_skb_pcount() which currently uses skb_shinfo(skb)->gso_segs, which is a terrible choice, because it needs 3 cache lines in current kernel (skb->head, skb->end, and shinfo->gso_segs are all in 3 different cache lines, far from skb->cb) This very simple patch reuses space currently taken by tcp_tw_isn only in input path, as tcp_skb_pcount is only needed for skb stored in write queue. This considerably speeds up tcp_ack(), granted we avoid shinfo->tx_flags to get SKBTX_ACK_TSTAMP, which seems possible. This also speeds up all sack processing in general. This speeds up tcp_sendmsg() because it no longer has to access/dirty shinfo. Signed-off-by: Eric Dumazet <edumazet@google.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
dc83d4d8f6
commit
cd7d8498c9
@ -698,7 +698,16 @@ static inline u32 tcp_skb_timestamp(const struct sk_buff *skb)
|
||||
struct tcp_skb_cb {
|
||||
__u32 seq; /* Starting sequence number */
|
||||
__u32 end_seq; /* SEQ + FIN + SYN + datalen */
|
||||
__u32 tcp_tw_isn; /* isn chosen by tcp_timewait_state_process() */
|
||||
union {
|
||||
/* Note : tcp_tw_isn is used in input path only
|
||||
* (isn chosen by tcp_timewait_state_process())
|
||||
*
|
||||
* tcp_gso_segs is used in write queue only,
|
||||
* cf tcp_skb_pcount()
|
||||
*/
|
||||
__u32 tcp_tw_isn;
|
||||
__u32 tcp_gso_segs;
|
||||
};
|
||||
__u8 tcp_flags; /* TCP header flags. (tcp[13]) */
|
||||
|
||||
__u8 sacked; /* State flags for SACK/FACK. */
|
||||
@ -746,7 +755,17 @@ TCP_ECN_create_request(struct request_sock *req, const struct sk_buff *skb,
|
||||
*/
|
||||
static inline int tcp_skb_pcount(const struct sk_buff *skb)
|
||||
{
|
||||
return skb_shinfo(skb)->gso_segs;
|
||||
return TCP_SKB_CB(skb)->tcp_gso_segs;
|
||||
}
|
||||
|
||||
static inline void tcp_skb_pcount_set(struct sk_buff *skb, int segs)
|
||||
{
|
||||
TCP_SKB_CB(skb)->tcp_gso_segs = segs;
|
||||
}
|
||||
|
||||
static inline void tcp_skb_pcount_add(struct sk_buff *skb, int segs)
|
||||
{
|
||||
TCP_SKB_CB(skb)->tcp_gso_segs += segs;
|
||||
}
|
||||
|
||||
/* This is valid iff tcp_skb_pcount() > 1. */
|
||||
|
@ -963,7 +963,7 @@ new_segment:
|
||||
skb->ip_summed = CHECKSUM_PARTIAL;
|
||||
tp->write_seq += copy;
|
||||
TCP_SKB_CB(skb)->end_seq += copy;
|
||||
skb_shinfo(skb)->gso_segs = 0;
|
||||
tcp_skb_pcount_set(skb, 0);
|
||||
|
||||
if (!copied)
|
||||
TCP_SKB_CB(skb)->tcp_flags &= ~TCPHDR_PSH;
|
||||
@ -1261,7 +1261,7 @@ new_segment:
|
||||
|
||||
tp->write_seq += copy;
|
||||
TCP_SKB_CB(skb)->end_seq += copy;
|
||||
skb_shinfo(skb)->gso_segs = 0;
|
||||
tcp_skb_pcount_set(skb, 0);
|
||||
|
||||
from += copy;
|
||||
copied += copy;
|
||||
|
@ -1295,9 +1295,9 @@ static bool tcp_shifted_skb(struct sock *sk, struct sk_buff *skb,
|
||||
TCP_SKB_CB(prev)->end_seq += shifted;
|
||||
TCP_SKB_CB(skb)->seq += shifted;
|
||||
|
||||
skb_shinfo(prev)->gso_segs += pcount;
|
||||
BUG_ON(skb_shinfo(skb)->gso_segs < pcount);
|
||||
skb_shinfo(skb)->gso_segs -= pcount;
|
||||
tcp_skb_pcount_add(prev, pcount);
|
||||
BUG_ON(tcp_skb_pcount(skb) < pcount);
|
||||
tcp_skb_pcount_add(skb, -pcount);
|
||||
|
||||
/* When we're adding to gso_segs == 1, gso_size will be zero,
|
||||
* in theory this shouldn't be necessary but as long as DSACK
|
||||
@ -1310,7 +1310,7 @@ static bool tcp_shifted_skb(struct sock *sk, struct sk_buff *skb,
|
||||
}
|
||||
|
||||
/* CHECKME: To clear or not to clear? Mimics normal skb currently */
|
||||
if (skb_shinfo(skb)->gso_segs <= 1) {
|
||||
if (tcp_skb_pcount(skb) <= 1) {
|
||||
skb_shinfo(skb)->gso_size = 0;
|
||||
skb_shinfo(skb)->gso_type = 0;
|
||||
}
|
||||
|
@ -384,7 +384,7 @@ static void tcp_init_nondata_skb(struct sk_buff *skb, u32 seq, u8 flags)
|
||||
TCP_SKB_CB(skb)->tcp_flags = flags;
|
||||
TCP_SKB_CB(skb)->sacked = 0;
|
||||
|
||||
shinfo->gso_segs = 1;
|
||||
tcp_skb_pcount_set(skb, 1);
|
||||
shinfo->gso_size = 0;
|
||||
shinfo->gso_type = 0;
|
||||
|
||||
@ -972,6 +972,9 @@ static int tcp_transmit_skb(struct sock *sk, struct sk_buff *skb, int clone_it,
|
||||
TCP_ADD_STATS(sock_net(sk), TCP_MIB_OUTSEGS,
|
||||
tcp_skb_pcount(skb));
|
||||
|
||||
/* OK, its time to fill skb_shinfo(skb)->gso_segs */
|
||||
skb_shinfo(skb)->gso_segs = tcp_skb_pcount(skb);
|
||||
|
||||
/* Our usage of tstamp should remain private */
|
||||
skb->tstamp.tv64 = 0;
|
||||
|
||||
@ -1019,11 +1022,11 @@ static void tcp_set_skb_tso_segs(const struct sock *sk, struct sk_buff *skb,
|
||||
/* Avoid the costly divide in the normal
|
||||
* non-TSO case.
|
||||
*/
|
||||
shinfo->gso_segs = 1;
|
||||
tcp_skb_pcount_set(skb, 1);
|
||||
shinfo->gso_size = 0;
|
||||
shinfo->gso_type = 0;
|
||||
} else {
|
||||
shinfo->gso_segs = DIV_ROUND_UP(skb->len, mss_now);
|
||||
tcp_skb_pcount_set(skb, DIV_ROUND_UP(skb->len, mss_now));
|
||||
shinfo->gso_size = mss_now;
|
||||
shinfo->gso_type = sk->sk_gso_type;
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user