mirror of
https://github.com/Fishwaldo/Star64_linux.git
synced 2025-06-29 10:01:25 +00:00
[INET]: Consolidate the xxx_frag_destroy
To make in possible we need to know the exact frag queue size for inet_frags->mem management and two callbacks: * to destoy the skb (optional, used in conntracks only) * to free the queue itself (mandatory, but later I plan to move the allocation and the destruction of frag_queues into the common place, so this callback will most likely be optional too). Signed-off-by: Pavel Emelyanov <xemul@openvz.org> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
321a3a99e4
commit
1e4b82873a
5 changed files with 74 additions and 86 deletions
|
@ -33,16 +33,21 @@ struct inet_frags {
|
||||||
rwlock_t lock;
|
rwlock_t lock;
|
||||||
u32 rnd;
|
u32 rnd;
|
||||||
int nqueues;
|
int nqueues;
|
||||||
|
int qsize;
|
||||||
atomic_t mem;
|
atomic_t mem;
|
||||||
struct timer_list secret_timer;
|
struct timer_list secret_timer;
|
||||||
struct inet_frags_ctl *ctl;
|
struct inet_frags_ctl *ctl;
|
||||||
|
|
||||||
unsigned int (*hashfn)(struct inet_frag_queue *);
|
unsigned int (*hashfn)(struct inet_frag_queue *);
|
||||||
|
void (*destructor)(struct inet_frag_queue *);
|
||||||
|
void (*skb_free)(struct sk_buff *);
|
||||||
};
|
};
|
||||||
|
|
||||||
void inet_frags_init(struct inet_frags *);
|
void inet_frags_init(struct inet_frags *);
|
||||||
void inet_frags_fini(struct inet_frags *);
|
void inet_frags_fini(struct inet_frags *);
|
||||||
|
|
||||||
void inet_frag_kill(struct inet_frag_queue *q, struct inet_frags *f);
|
void inet_frag_kill(struct inet_frag_queue *q, struct inet_frags *f);
|
||||||
|
void inet_frag_destroy(struct inet_frag_queue *q,
|
||||||
|
struct inet_frags *f, int *work);
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
|
|
@ -17,6 +17,8 @@
|
||||||
#include <linux/timer.h>
|
#include <linux/timer.h>
|
||||||
#include <linux/mm.h>
|
#include <linux/mm.h>
|
||||||
#include <linux/random.h>
|
#include <linux/random.h>
|
||||||
|
#include <linux/skbuff.h>
|
||||||
|
#include <linux/rtnetlink.h>
|
||||||
|
|
||||||
#include <net/inet_frag.h>
|
#include <net/inet_frag.h>
|
||||||
|
|
||||||
|
@ -100,3 +102,41 @@ void inet_frag_kill(struct inet_frag_queue *fq, struct inet_frags *f)
|
||||||
}
|
}
|
||||||
|
|
||||||
EXPORT_SYMBOL(inet_frag_kill);
|
EXPORT_SYMBOL(inet_frag_kill);
|
||||||
|
|
||||||
|
static inline void frag_kfree_skb(struct inet_frags *f, struct sk_buff *skb,
|
||||||
|
int *work)
|
||||||
|
{
|
||||||
|
if (work)
|
||||||
|
*work -= skb->truesize;
|
||||||
|
|
||||||
|
atomic_sub(skb->truesize, &f->mem);
|
||||||
|
if (f->skb_free)
|
||||||
|
f->skb_free(skb);
|
||||||
|
kfree_skb(skb);
|
||||||
|
}
|
||||||
|
|
||||||
|
void inet_frag_destroy(struct inet_frag_queue *q, struct inet_frags *f,
|
||||||
|
int *work)
|
||||||
|
{
|
||||||
|
struct sk_buff *fp;
|
||||||
|
|
||||||
|
BUG_TRAP(q->last_in & COMPLETE);
|
||||||
|
BUG_TRAP(del_timer(&q->timer) == 0);
|
||||||
|
|
||||||
|
/* Release all fragment data. */
|
||||||
|
fp = q->fragments;
|
||||||
|
while (fp) {
|
||||||
|
struct sk_buff *xp = fp->next;
|
||||||
|
|
||||||
|
frag_kfree_skb(f, fp, work);
|
||||||
|
fp = xp;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (work)
|
||||||
|
*work -= f->qsize;
|
||||||
|
atomic_sub(f->qsize, &f->mem);
|
||||||
|
|
||||||
|
f->destructor(q);
|
||||||
|
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(inet_frag_destroy);
|
||||||
|
|
|
@ -132,11 +132,13 @@ static __inline__ void frag_kfree_skb(struct sk_buff *skb, int *work)
|
||||||
kfree_skb(skb);
|
kfree_skb(skb);
|
||||||
}
|
}
|
||||||
|
|
||||||
static __inline__ void frag_free_queue(struct ipq *qp, int *work)
|
static __inline__ void ip4_frag_free(struct inet_frag_queue *q)
|
||||||
{
|
{
|
||||||
if (work)
|
struct ipq *qp;
|
||||||
*work -= sizeof(struct ipq);
|
|
||||||
atomic_sub(sizeof(struct ipq), &ip4_frags.mem);
|
qp = container_of(q, struct ipq, q);
|
||||||
|
if (qp->peer)
|
||||||
|
inet_putpeer(qp->peer);
|
||||||
kfree(qp);
|
kfree(qp);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -153,34 +155,10 @@ static __inline__ struct ipq *frag_alloc_queue(void)
|
||||||
|
|
||||||
/* Destruction primitives. */
|
/* Destruction primitives. */
|
||||||
|
|
||||||
/* Complete destruction of ipq. */
|
|
||||||
static void ip_frag_destroy(struct ipq *qp, int *work)
|
|
||||||
{
|
|
||||||
struct sk_buff *fp;
|
|
||||||
|
|
||||||
BUG_TRAP(qp->q.last_in&COMPLETE);
|
|
||||||
BUG_TRAP(del_timer(&qp->q.timer) == 0);
|
|
||||||
|
|
||||||
if (qp->peer)
|
|
||||||
inet_putpeer(qp->peer);
|
|
||||||
|
|
||||||
/* Release all fragment data. */
|
|
||||||
fp = qp->q.fragments;
|
|
||||||
while (fp) {
|
|
||||||
struct sk_buff *xp = fp->next;
|
|
||||||
|
|
||||||
frag_kfree_skb(fp, work);
|
|
||||||
fp = xp;
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Finally, release the queue descriptor itself. */
|
|
||||||
frag_free_queue(qp, work);
|
|
||||||
}
|
|
||||||
|
|
||||||
static __inline__ void ipq_put(struct ipq *ipq, int *work)
|
static __inline__ void ipq_put(struct ipq *ipq, int *work)
|
||||||
{
|
{
|
||||||
if (atomic_dec_and_test(&ipq->q.refcnt))
|
if (atomic_dec_and_test(&ipq->q.refcnt))
|
||||||
ip_frag_destroy(ipq, work);
|
inet_frag_destroy(&ipq->q, &ip4_frags, work);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Kill ipq entry. It is not destroyed immediately,
|
/* Kill ipq entry. It is not destroyed immediately,
|
||||||
|
@ -721,6 +699,9 @@ void __init ipfrag_init(void)
|
||||||
{
|
{
|
||||||
ip4_frags.ctl = &ip4_frags_ctl;
|
ip4_frags.ctl = &ip4_frags_ctl;
|
||||||
ip4_frags.hashfn = ip4_hashfn;
|
ip4_frags.hashfn = ip4_hashfn;
|
||||||
|
ip4_frags.destructor = ip4_frag_free;
|
||||||
|
ip4_frags.skb_free = NULL;
|
||||||
|
ip4_frags.qsize = sizeof(struct ipq);
|
||||||
inet_frags_init(&ip4_frags);
|
inet_frags_init(&ip4_frags);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -114,25 +114,25 @@ static unsigned int nf_hashfn(struct inet_frag_queue *q)
|
||||||
return ip6qhashfn(nq->id, &nq->saddr, &nq->daddr);
|
return ip6qhashfn(nq->id, &nq->saddr, &nq->daddr);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void nf_skb_free(struct sk_buff *skb)
|
||||||
|
{
|
||||||
|
if (NFCT_FRAG6_CB(skb)->orig)
|
||||||
|
kfree_skb(NFCT_FRAG6_CB(skb)->orig);
|
||||||
|
}
|
||||||
|
|
||||||
/* Memory Tracking Functions. */
|
/* Memory Tracking Functions. */
|
||||||
static inline void frag_kfree_skb(struct sk_buff *skb, unsigned int *work)
|
static inline void frag_kfree_skb(struct sk_buff *skb, unsigned int *work)
|
||||||
{
|
{
|
||||||
if (work)
|
if (work)
|
||||||
*work -= skb->truesize;
|
*work -= skb->truesize;
|
||||||
atomic_sub(skb->truesize, &nf_frags.mem);
|
atomic_sub(skb->truesize, &nf_frags.mem);
|
||||||
if (NFCT_FRAG6_CB(skb)->orig)
|
nf_skb_free(skb);
|
||||||
kfree_skb(NFCT_FRAG6_CB(skb)->orig);
|
|
||||||
|
|
||||||
kfree_skb(skb);
|
kfree_skb(skb);
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline void frag_free_queue(struct nf_ct_frag6_queue *fq,
|
static void nf_frag_free(struct inet_frag_queue *q)
|
||||||
unsigned int *work)
|
|
||||||
{
|
{
|
||||||
if (work)
|
kfree(container_of(q, struct nf_ct_frag6_queue, q));
|
||||||
*work -= sizeof(struct nf_ct_frag6_queue);
|
|
||||||
atomic_sub(sizeof(struct nf_ct_frag6_queue), &nf_frags.mem);
|
|
||||||
kfree(fq);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline struct nf_ct_frag6_queue *frag_alloc_queue(void)
|
static inline struct nf_ct_frag6_queue *frag_alloc_queue(void)
|
||||||
|
@ -147,31 +147,10 @@ static inline struct nf_ct_frag6_queue *frag_alloc_queue(void)
|
||||||
|
|
||||||
/* Destruction primitives. */
|
/* Destruction primitives. */
|
||||||
|
|
||||||
/* Complete destruction of fq. */
|
|
||||||
static void nf_ct_frag6_destroy(struct nf_ct_frag6_queue *fq,
|
|
||||||
unsigned int *work)
|
|
||||||
{
|
|
||||||
struct sk_buff *fp;
|
|
||||||
|
|
||||||
BUG_TRAP(fq->q.last_in&COMPLETE);
|
|
||||||
BUG_TRAP(del_timer(&fq->q.timer) == 0);
|
|
||||||
|
|
||||||
/* Release all fragment data. */
|
|
||||||
fp = fq->q.fragments;
|
|
||||||
while (fp) {
|
|
||||||
struct sk_buff *xp = fp->next;
|
|
||||||
|
|
||||||
frag_kfree_skb(fp, work);
|
|
||||||
fp = xp;
|
|
||||||
}
|
|
||||||
|
|
||||||
frag_free_queue(fq, work);
|
|
||||||
}
|
|
||||||
|
|
||||||
static __inline__ void fq_put(struct nf_ct_frag6_queue *fq, unsigned int *work)
|
static __inline__ void fq_put(struct nf_ct_frag6_queue *fq, unsigned int *work)
|
||||||
{
|
{
|
||||||
if (atomic_dec_and_test(&fq->q.refcnt))
|
if (atomic_dec_and_test(&fq->q.refcnt))
|
||||||
nf_ct_frag6_destroy(fq, work);
|
inet_frag_destroy(&fq->q, &nf_frags, work);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Kill fq entry. It is not destroyed immediately,
|
/* Kill fq entry. It is not destroyed immediately,
|
||||||
|
@ -799,6 +778,9 @@ int nf_ct_frag6_init(void)
|
||||||
{
|
{
|
||||||
nf_frags.ctl = &nf_frags_ctl;
|
nf_frags.ctl = &nf_frags_ctl;
|
||||||
nf_frags.hashfn = nf_hashfn;
|
nf_frags.hashfn = nf_hashfn;
|
||||||
|
nf_frags.destructor = nf_frag_free;
|
||||||
|
nf_frags.skb_free = nf_skb_free;
|
||||||
|
nf_frags.qsize = sizeof(struct nf_ct_frag6_queue);
|
||||||
inet_frags_init(&nf_frags);
|
inet_frags_init(&nf_frags);
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
|
|
@ -152,12 +152,9 @@ static inline void frag_kfree_skb(struct sk_buff *skb, int *work)
|
||||||
kfree_skb(skb);
|
kfree_skb(skb);
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline void frag_free_queue(struct frag_queue *fq, int *work)
|
static void ip6_frag_free(struct inet_frag_queue *fq)
|
||||||
{
|
{
|
||||||
if (work)
|
kfree(container_of(fq, struct frag_queue, q));
|
||||||
*work -= sizeof(struct frag_queue);
|
|
||||||
atomic_sub(sizeof(struct frag_queue), &ip6_frags.mem);
|
|
||||||
kfree(fq);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline struct frag_queue *frag_alloc_queue(void)
|
static inline struct frag_queue *frag_alloc_queue(void)
|
||||||
|
@ -172,30 +169,10 @@ static inline struct frag_queue *frag_alloc_queue(void)
|
||||||
|
|
||||||
/* Destruction primitives. */
|
/* Destruction primitives. */
|
||||||
|
|
||||||
/* Complete destruction of fq. */
|
|
||||||
static void ip6_frag_destroy(struct frag_queue *fq, int *work)
|
|
||||||
{
|
|
||||||
struct sk_buff *fp;
|
|
||||||
|
|
||||||
BUG_TRAP(fq->q.last_in&COMPLETE);
|
|
||||||
BUG_TRAP(del_timer(&fq->q.timer) == 0);
|
|
||||||
|
|
||||||
/* Release all fragment data. */
|
|
||||||
fp = fq->q.fragments;
|
|
||||||
while (fp) {
|
|
||||||
struct sk_buff *xp = fp->next;
|
|
||||||
|
|
||||||
frag_kfree_skb(fp, work);
|
|
||||||
fp = xp;
|
|
||||||
}
|
|
||||||
|
|
||||||
frag_free_queue(fq, work);
|
|
||||||
}
|
|
||||||
|
|
||||||
static __inline__ void fq_put(struct frag_queue *fq, int *work)
|
static __inline__ void fq_put(struct frag_queue *fq, int *work)
|
||||||
{
|
{
|
||||||
if (atomic_dec_and_test(&fq->q.refcnt))
|
if (atomic_dec_and_test(&fq->q.refcnt))
|
||||||
ip6_frag_destroy(fq, work);
|
inet_frag_destroy(&fq->q, &ip6_frags, work);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Kill fq entry. It is not destroyed immediately,
|
/* Kill fq entry. It is not destroyed immediately,
|
||||||
|
@ -744,5 +721,8 @@ void __init ipv6_frag_init(void)
|
||||||
|
|
||||||
ip6_frags.ctl = &ip6_frags_ctl;
|
ip6_frags.ctl = &ip6_frags_ctl;
|
||||||
ip6_frags.hashfn = ip6_hashfn;
|
ip6_frags.hashfn = ip6_hashfn;
|
||||||
|
ip6_frags.destructor = ip6_frag_free;
|
||||||
|
ip6_frags.skb_free = NULL;
|
||||||
|
ip6_frags.qsize = sizeof(struct frag_queue);
|
||||||
inet_frags_init(&ip6_frags);
|
inet_frags_init(&ip6_frags);
|
||||||
}
|
}
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue