forked from Minki/linux
[NET]: prot_inuse cleanups and optimizations
1) Cleanups (all functions are prefixed by sock_prot_inuse) sock_prot_inc_use(prot) -> sock_prot_inuse_add(prot,-1) sock_prot_dec_use(prot) -> sock_prot_inuse_add(prot,-1) sock_prot_inuse() -> sock_prot_inuse_get() New functions : sock_prot_inuse_init() and sock_prot_inuse_free() to abstract pcounter use. 2) if CONFIG_PROC_FS=n, we can zap 'inuse' member from "struct proto", since nobody wants to read the inuse value. This saves 1372 bytes on i386/SMP and some cpu cycles. Signed-off-by: Eric Dumazet <dada1@cosmosbay.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
571e768202
commit
65f7651788
@ -293,7 +293,7 @@ static inline void inet_unhash(struct inet_hashinfo *hashinfo, struct sock *sk)
|
||||
}
|
||||
|
||||
if (__sk_del_node_init(sk))
|
||||
sock_prot_dec_use(sk->sk_prot);
|
||||
sock_prot_inuse_add(sk->sk_prot, -1);
|
||||
write_unlock_bh(lock);
|
||||
out:
|
||||
if (sk->sk_state == TCP_LISTEN)
|
||||
|
@ -548,7 +548,9 @@ struct proto {
|
||||
int (*get_port)(struct sock *sk, unsigned short snum);
|
||||
|
||||
/* Keeping track of sockets in use */
|
||||
#ifdef CONFIG_PROC_FS
|
||||
struct pcounter inuse;
|
||||
#endif
|
||||
|
||||
/* Memory pressure */
|
||||
void (*enter_memory_pressure)(void);
|
||||
@ -584,9 +586,6 @@ struct proto {
|
||||
#endif
|
||||
};
|
||||
|
||||
#define DEFINE_PROTO_INUSE(NAME) DEFINE_PCOUNTER(NAME)
|
||||
#define REF_PROTO_INUSE(NAME) PCOUNTER_MEMBER_INITIALIZER(NAME, .inuse)
|
||||
|
||||
extern int proto_register(struct proto *prot, int alloc_slab);
|
||||
extern void proto_unregister(struct proto *prot);
|
||||
|
||||
@ -615,21 +614,42 @@ static inline void sk_refcnt_debug_release(const struct sock *sk)
|
||||
#define sk_refcnt_debug_release(sk) do { } while (0)
|
||||
#endif /* SOCK_REFCNT_DEBUG */
|
||||
|
||||
|
||||
#ifdef CONFIG_PROC_FS
|
||||
# define DEFINE_PROTO_INUSE(NAME) DEFINE_PCOUNTER(NAME)
|
||||
# define REF_PROTO_INUSE(NAME) PCOUNTER_MEMBER_INITIALIZER(NAME, .inuse)
|
||||
/* Called with local bh disabled */
|
||||
static __inline__ void sock_prot_inc_use(struct proto *prot)
|
||||
static inline void sock_prot_inuse_add(struct proto *prot, int inc)
|
||||
{
|
||||
pcounter_add(&prot->inuse, 1);
|
||||
pcounter_add(&prot->inuse, inc);
|
||||
}
|
||||
|
||||
static __inline__ void sock_prot_dec_use(struct proto *prot)
|
||||
static inline int sock_prot_inuse_init(struct proto *proto)
|
||||
{
|
||||
pcounter_add(&prot->inuse, -1);
|
||||
return pcounter_alloc(&proto->inuse);
|
||||
}
|
||||
|
||||
static __inline__ int sock_prot_inuse(struct proto *proto)
|
||||
static inline int sock_prot_inuse_get(struct proto *proto)
|
||||
{
|
||||
return pcounter_getval(&proto->inuse);
|
||||
}
|
||||
static inline void sock_prot_inuse_free(struct proto *proto)
|
||||
{
|
||||
pcounter_free(&proto->inuse);
|
||||
}
|
||||
#else
|
||||
# define DEFINE_PROTO_INUSE(NAME)
|
||||
# define REF_PROTO_INUSE(NAME)
|
||||
static void inline sock_prot_inuse_add(struct proto *prot, int inc)
|
||||
{
|
||||
}
|
||||
static int inline sock_prot_inuse_init(struct proto *proto)
|
||||
{
|
||||
return 0;
|
||||
}
|
||||
static void inline sock_prot_inuse_free(struct proto *proto)
|
||||
{
|
||||
}
|
||||
#endif
|
||||
|
||||
|
||||
/* With per-bucket locks this operation is not-atomic, so that
|
||||
* this version is not worse.
|
||||
|
@ -115,7 +115,7 @@ static inline void udp_lib_unhash(struct sock *sk)
|
||||
write_lock_bh(&udp_hash_lock);
|
||||
if (sk_del_node_init(sk)) {
|
||||
inet_sk(sk)->num = 0;
|
||||
sock_prot_dec_use(sk->sk_prot);
|
||||
sock_prot_inuse_add(sk->sk_prot, -1);
|
||||
}
|
||||
write_unlock_bh(&udp_hash_lock);
|
||||
}
|
||||
|
@ -1913,7 +1913,7 @@ int proto_register(struct proto *prot, int alloc_slab)
|
||||
char *request_sock_slab_name = NULL;
|
||||
char *timewait_sock_slab_name;
|
||||
|
||||
if (pcounter_alloc(&prot->inuse) != 0) {
|
||||
if (sock_prot_inuse_init(prot) != 0) {
|
||||
printk(KERN_CRIT "%s: Can't alloc inuse counters!\n", prot->name);
|
||||
goto out;
|
||||
}
|
||||
@ -1984,7 +1984,7 @@ out_free_sock_slab:
|
||||
kmem_cache_destroy(prot->slab);
|
||||
prot->slab = NULL;
|
||||
out_free_inuse:
|
||||
pcounter_free(&prot->inuse);
|
||||
sock_prot_inuse_free(prot);
|
||||
out:
|
||||
return -ENOBUFS;
|
||||
}
|
||||
@ -1997,7 +1997,7 @@ void proto_unregister(struct proto *prot)
|
||||
list_del(&prot->node);
|
||||
write_unlock(&proto_list_lock);
|
||||
|
||||
pcounter_free(&prot->inuse);
|
||||
sock_prot_inuse_free(prot);
|
||||
|
||||
if (prot->slab != NULL) {
|
||||
kmem_cache_destroy(prot->slab);
|
||||
|
@ -278,7 +278,7 @@ unique:
|
||||
sk->sk_hash = hash;
|
||||
BUG_TRAP(sk_unhashed(sk));
|
||||
__sk_add_node(sk, &head->chain);
|
||||
sock_prot_inc_use(sk->sk_prot);
|
||||
sock_prot_inuse_add(sk->sk_prot, 1);
|
||||
write_unlock(lock);
|
||||
|
||||
if (twp) {
|
||||
@ -321,7 +321,7 @@ void __inet_hash_nolisten(struct inet_hashinfo *hashinfo, struct sock *sk)
|
||||
|
||||
write_lock(lock);
|
||||
__sk_add_node(sk, list);
|
||||
sock_prot_inc_use(sk->sk_prot);
|
||||
sock_prot_inuse_add(sk->sk_prot, 1);
|
||||
write_unlock(lock);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(__inet_hash_nolisten);
|
||||
@ -342,7 +342,7 @@ void __inet_hash(struct inet_hashinfo *hashinfo, struct sock *sk)
|
||||
|
||||
inet_listen_wlock(hashinfo);
|
||||
__sk_add_node(sk, list);
|
||||
sock_prot_inc_use(sk->sk_prot);
|
||||
sock_prot_inuse_add(sk->sk_prot, 1);
|
||||
write_unlock(lock);
|
||||
wake_up(&hashinfo->lhash_wait);
|
||||
}
|
||||
|
@ -91,7 +91,7 @@ void __inet_twsk_hashdance(struct inet_timewait_sock *tw, struct sock *sk,
|
||||
|
||||
/* Step 2: Remove SK from established hash. */
|
||||
if (__sk_del_node_init(sk))
|
||||
sock_prot_dec_use(sk->sk_prot);
|
||||
sock_prot_inuse_add(sk->sk_prot, -1);
|
||||
|
||||
/* Step 3: Hash TW into TIMEWAIT chain. */
|
||||
inet_twsk_add_node(tw, &ehead->twchain);
|
||||
|
@ -53,13 +53,14 @@ static int sockstat_seq_show(struct seq_file *seq, void *v)
|
||||
{
|
||||
socket_seq_show(seq);
|
||||
seq_printf(seq, "TCP: inuse %d orphan %d tw %d alloc %d mem %d\n",
|
||||
sock_prot_inuse(&tcp_prot), atomic_read(&tcp_orphan_count),
|
||||
sock_prot_inuse_get(&tcp_prot),
|
||||
atomic_read(&tcp_orphan_count),
|
||||
tcp_death_row.tw_count, atomic_read(&tcp_sockets_allocated),
|
||||
atomic_read(&tcp_memory_allocated));
|
||||
seq_printf(seq, "UDP: inuse %d mem %d\n", sock_prot_inuse(&udp_prot),
|
||||
seq_printf(seq, "UDP: inuse %d mem %d\n", sock_prot_inuse_get(&udp_prot),
|
||||
atomic_read(&udp_memory_allocated));
|
||||
seq_printf(seq, "UDPLITE: inuse %d\n", sock_prot_inuse(&udplite_prot));
|
||||
seq_printf(seq, "RAW: inuse %d\n", sock_prot_inuse(&raw_prot));
|
||||
seq_printf(seq, "UDPLITE: inuse %d\n", sock_prot_inuse_get(&udplite_prot));
|
||||
seq_printf(seq, "RAW: inuse %d\n", sock_prot_inuse_get(&raw_prot));
|
||||
seq_printf(seq, "FRAG: inuse %d memory %d\n",
|
||||
ip_frag_nqueues(), ip_frag_mem());
|
||||
return 0;
|
||||
|
@ -92,7 +92,7 @@ void raw_hash_sk(struct sock *sk, struct raw_hashinfo *h)
|
||||
|
||||
write_lock_bh(&h->lock);
|
||||
sk_add_node(sk, head);
|
||||
sock_prot_inc_use(sk->sk_prot);
|
||||
sock_prot_inuse_add(sk->sk_prot, 1);
|
||||
write_unlock_bh(&h->lock);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(raw_hash_sk);
|
||||
@ -101,7 +101,7 @@ void raw_unhash_sk(struct sock *sk, struct raw_hashinfo *h)
|
||||
{
|
||||
write_lock_bh(&h->lock);
|
||||
if (sk_del_node_init(sk))
|
||||
sock_prot_dec_use(sk->sk_prot);
|
||||
sock_prot_inuse_add(sk->sk_prot, -1);
|
||||
write_unlock_bh(&h->lock);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(raw_unhash_sk);
|
||||
|
@ -230,7 +230,7 @@ gotit:
|
||||
if (sk_unhashed(sk)) {
|
||||
head = &udptable[snum & (UDP_HTABLE_SIZE - 1)];
|
||||
sk_add_node(sk, head);
|
||||
sock_prot_inc_use(sk->sk_prot);
|
||||
sock_prot_inuse_add(sk->sk_prot, 1);
|
||||
}
|
||||
error = 0;
|
||||
fail:
|
||||
|
@ -43,7 +43,7 @@ void __inet6_hash(struct inet_hashinfo *hashinfo,
|
||||
}
|
||||
|
||||
__sk_add_node(sk, list);
|
||||
sock_prot_inc_use(sk->sk_prot);
|
||||
sock_prot_inuse_add(sk->sk_prot, 1);
|
||||
write_unlock(lock);
|
||||
}
|
||||
EXPORT_SYMBOL(__inet6_hash);
|
||||
@ -216,7 +216,7 @@ unique:
|
||||
BUG_TRAP(sk_unhashed(sk));
|
||||
__sk_add_node(sk, &head->chain);
|
||||
sk->sk_hash = hash;
|
||||
sock_prot_inc_use(sk->sk_prot);
|
||||
sock_prot_inuse_add(sk->sk_prot, 1);
|
||||
write_unlock(lock);
|
||||
|
||||
if (twp != NULL) {
|
||||
|
@ -268,8 +268,8 @@ static int do_ipv6_setsockopt(struct sock *sk, int level, int optname,
|
||||
struct inet_connection_sock *icsk = inet_csk(sk);
|
||||
|
||||
local_bh_disable();
|
||||
sock_prot_dec_use(sk->sk_prot);
|
||||
sock_prot_inc_use(&tcp_prot);
|
||||
sock_prot_inuse_add(sk->sk_prot, -1);
|
||||
sock_prot_inuse_add(&tcp_prot, 1);
|
||||
local_bh_enable();
|
||||
sk->sk_prot = &tcp_prot;
|
||||
icsk->icsk_af_ops = &ipv4_specific;
|
||||
@ -282,8 +282,8 @@ static int do_ipv6_setsockopt(struct sock *sk, int level, int optname,
|
||||
if (sk->sk_protocol == IPPROTO_UDPLITE)
|
||||
prot = &udplite_prot;
|
||||
local_bh_disable();
|
||||
sock_prot_dec_use(sk->sk_prot);
|
||||
sock_prot_inc_use(prot);
|
||||
sock_prot_inuse_add(sk->sk_prot, -1);
|
||||
sock_prot_inuse_add(prot, 1);
|
||||
local_bh_enable();
|
||||
sk->sk_prot = prot;
|
||||
sk->sk_socket->ops = &inet_dgram_ops;
|
||||
|
@ -36,13 +36,13 @@ static struct proc_dir_entry *proc_net_devsnmp6;
|
||||
static int sockstat6_seq_show(struct seq_file *seq, void *v)
|
||||
{
|
||||
seq_printf(seq, "TCP6: inuse %d\n",
|
||||
sock_prot_inuse(&tcpv6_prot));
|
||||
sock_prot_inuse_get(&tcpv6_prot));
|
||||
seq_printf(seq, "UDP6: inuse %d\n",
|
||||
sock_prot_inuse(&udpv6_prot));
|
||||
sock_prot_inuse_get(&udpv6_prot));
|
||||
seq_printf(seq, "UDPLITE6: inuse %d\n",
|
||||
sock_prot_inuse(&udplitev6_prot));
|
||||
sock_prot_inuse_get(&udplitev6_prot));
|
||||
seq_printf(seq, "RAW6: inuse %d\n",
|
||||
sock_prot_inuse(&rawv6_prot));
|
||||
sock_prot_inuse_get(&rawv6_prot));
|
||||
seq_printf(seq, "FRAG6: inuse %d memory %d\n",
|
||||
ip6_frag_nqueues(), ip6_frag_mem());
|
||||
return 0;
|
||||
|
Loading…
Reference in New Issue
Block a user