157 lines
5.0 KiB
Diff
157 lines
5.0 KiB
Diff
From: Peter Zijlstra <a.p.zijlstra@chello.nl>
|
|
Subject: [PATCH 14/31] net: sk_allocation() - concentrate socket related allocations
|
|
Patch-mainline: Not yet
|
|
|
|
Introduce sk_allocation(), this function allows to inject sock specific
|
|
flags to each sock related allocation.
|
|
|
|
Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
|
|
Signed-off-by: Suresh Jayaraman <sjayaraman@suse.de>
|
|
---
|
|
include/net/sock.h | 5 +++++
|
|
net/ipv4/tcp.c | 3 ++-
|
|
net/ipv4/tcp_output.c | 11 ++++++-----
|
|
net/ipv6/tcp_ipv6.c | 15 +++++++++++----
|
|
4 files changed, 24 insertions(+), 10 deletions(-)
|
|
|
|
--- a/include/net/sock.h
|
|
+++ b/include/net/sock.h
|
|
@@ -565,6 +565,11 @@ static inline int sock_flag(struct sock
|
|
return test_bit(flag, &sk->sk_flags);
|
|
}
|
|
|
|
+static inline gfp_t sk_allocation(struct sock *sk, gfp_t gfp_mask)
|
|
+{
|
|
+ return gfp_mask;
|
|
+}
|
|
+
|
|
static inline void sk_acceptq_removed(struct sock *sk)
|
|
{
|
|
sk->sk_ack_backlog--;
|
|
--- a/net/ipv4/tcp.c
|
|
+++ b/net/ipv4/tcp.c
|
|
@@ -685,7 +685,8 @@ struct sk_buff *sk_stream_alloc_skb(stru
|
|
/* The TCP header must be at least 32-bit aligned. */
|
|
size = ALIGN(size, 4);
|
|
|
|
- skb = alloc_skb_fclone(size + sk->sk_prot->max_header, gfp);
|
|
+ skb = alloc_skb_fclone(size + sk->sk_prot->max_header,
|
|
+ sk_allocation(sk, gfp));
|
|
if (skb) {
|
|
if (sk_wmem_schedule(sk, skb->truesize)) {
|
|
/*
|
|
--- a/net/ipv4/tcp_output.c
|
|
+++ b/net/ipv4/tcp_output.c
|
|
@@ -2313,7 +2313,7 @@ void tcp_send_fin(struct sock *sk)
|
|
/* Socket is locked, keep trying until memory is available. */
|
|
for (;;) {
|
|
skb = alloc_skb_fclone(MAX_TCP_HEADER,
|
|
- sk->sk_allocation);
|
|
+ sk_allocation(sk, GFP_KERNEL));
|
|
if (skb)
|
|
break;
|
|
yield();
|
|
@@ -2339,7 +2339,7 @@ void tcp_send_active_reset(struct sock *
|
|
struct sk_buff *skb;
|
|
|
|
/* NOTE: No TCP options attached and we never retransmit this. */
|
|
- skb = alloc_skb(MAX_TCP_HEADER, priority);
|
|
+ skb = alloc_skb(MAX_TCP_HEADER, sk_allocation(sk, priority));
|
|
if (!skb) {
|
|
NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTFAILED);
|
|
return;
|
|
@@ -2412,7 +2412,8 @@ struct sk_buff *tcp_make_synack(struct s
|
|
|
|
if (cvp != NULL && cvp->s_data_constant && cvp->s_data_desired)
|
|
s_data_desired = cvp->s_data_desired;
|
|
- skb = sock_wmalloc(sk, MAX_TCP_HEADER + 15 + s_data_desired, 1, GFP_ATOMIC);
|
|
+ skb = sock_wmalloc(sk, MAX_TCP_HEADER + 15 + s_data_desired, 1,
|
|
+ sk_allocation(sk, GFP_ATOMIC));
|
|
if (skb == NULL)
|
|
return NULL;
|
|
|
|
@@ -2694,7 +2695,7 @@ void tcp_send_ack(struct sock *sk)
|
|
* tcp_transmit_skb() will set the ownership to this
|
|
* sock.
|
|
*/
|
|
- buff = alloc_skb(MAX_TCP_HEADER, GFP_ATOMIC);
|
|
+ buff = alloc_skb(MAX_TCP_HEADER, sk_allocation(sk, GFP_ATOMIC));
|
|
if (buff == NULL) {
|
|
inet_csk_schedule_ack(sk);
|
|
inet_csk(sk)->icsk_ack.ato = TCP_ATO_MIN;
|
|
@@ -2729,7 +2730,7 @@ static int tcp_xmit_probe_skb(struct soc
|
|
struct sk_buff *skb;
|
|
|
|
/* We don't queue it, tcp_transmit_skb() sets ownership. */
|
|
- skb = alloc_skb(MAX_TCP_HEADER, GFP_ATOMIC);
|
|
+ skb = alloc_skb(MAX_TCP_HEADER, sk_allocation(sk, GFP_ATOMIC));
|
|
if (skb == NULL)
|
|
return -1;
|
|
|
|
--- a/net/ipv6/tcp_ipv6.c
|
|
+++ b/net/ipv6/tcp_ipv6.c
|
|
@@ -589,7 +589,8 @@ static int tcp_v6_md5_do_add(struct sock
|
|
} else {
|
|
/* reallocate new list if current one is full. */
|
|
if (!tp->md5sig_info) {
|
|
- tp->md5sig_info = kzalloc(sizeof(*tp->md5sig_info), GFP_ATOMIC);
|
|
+ tp->md5sig_info = kzalloc(sizeof(*tp->md5sig_info),
|
|
+ sk_allocation(sk, GFP_ATOMIC));
|
|
if (!tp->md5sig_info) {
|
|
kfree(newkey);
|
|
return -ENOMEM;
|
|
@@ -602,7 +603,8 @@ static int tcp_v6_md5_do_add(struct sock
|
|
}
|
|
if (tp->md5sig_info->alloced6 == tp->md5sig_info->entries6) {
|
|
keys = kmalloc((sizeof (tp->md5sig_info->keys6[0]) *
|
|
- (tp->md5sig_info->entries6 + 1)), GFP_ATOMIC);
|
|
+ (tp->md5sig_info->entries6 + 1)),
|
|
+ sk_allocation(sk, GFP_ATOMIC));
|
|
|
|
if (!keys) {
|
|
tcp_free_md5sig_pool();
|
|
@@ -726,7 +728,8 @@ static int tcp_v6_parse_md5_keys (struct
|
|
struct tcp_sock *tp = tcp_sk(sk);
|
|
struct tcp_md5sig_info *p;
|
|
|
|
- p = kzalloc(sizeof(struct tcp_md5sig_info), GFP_KERNEL);
|
|
+ p = kzalloc(sizeof(struct tcp_md5sig_info),
|
|
+ sk_allocation(sk, GFP_KERNEL));
|
|
if (!p)
|
|
return -ENOMEM;
|
|
|
|
@@ -997,6 +1000,7 @@ static void tcp_v6_send_response(struct
|
|
unsigned int tot_len = sizeof(struct tcphdr);
|
|
struct dst_entry *dst;
|
|
__be32 *topt;
|
|
+ gfp_t gfp_mask = GFP_ATOMIC;
|
|
|
|
if (ts)
|
|
tot_len += TCPOLEN_TSTAMP_ALIGNED;
|
|
@@ -1006,7 +1010,7 @@ static void tcp_v6_send_response(struct
|
|
#endif
|
|
|
|
buff = alloc_skb(MAX_HEADER + sizeof(struct ipv6hdr) + tot_len,
|
|
- GFP_ATOMIC);
|
|
+ gfp_mask);
|
|
if (buff == NULL)
|
|
return;
|
|
|
|
@@ -1083,6 +1087,7 @@ static void tcp_v6_send_reset(struct soc
|
|
struct tcphdr *th = tcp_hdr(skb);
|
|
u32 seq = 0, ack_seq = 0;
|
|
struct tcp_md5sig_key *key = NULL;
|
|
+ gfp_t gfp_mask = GFP_ATOMIC;
|
|
|
|
if (th->rst)
|
|
return;
|
|
@@ -1094,6 +1099,8 @@ static void tcp_v6_send_reset(struct soc
|
|
if (sk)
|
|
key = tcp_v6_md5_do_lookup(sk, &ipv6_hdr(skb)->daddr);
|
|
#endif
|
|
+ if (sk)
|
|
+ gfp_mask = sk_allocation(sk, gfp_mask);
|
|
|
|
if (th->ack)
|
|
seq = ntohl(th->ack_seq);
|