net: filter: split 'struct sk_filter' into socket and bpf parts
authorAlexei Starovoitov <ast@plumgrid.com>
Thu, 31 Jul 2014 03:34:16 +0000 (20:34 -0700)
committerDavid S. Miller <davem@davemloft.net>
Sat, 2 Aug 2014 22:03:58 +0000 (15:03 -0700)
clean up names related to socket filtering and bpf in the following way:
- everything that deals with sockets keeps 'sk_*' prefix
- everything that is pure BPF is changed to 'bpf_*' prefix

split 'struct sk_filter' into
struct sk_filter {
atomic_t        refcnt;
struct rcu_head rcu;
struct bpf_prog *prog;
};
and
struct bpf_prog {
        u32                     jited:1,
                                len:31;
        struct sock_fprog_kern  *orig_prog;
        unsigned int            (*bpf_func)(const struct sk_buff *skb,
                                            const struct bpf_insn *filter);
        union {
                struct sock_filter      insns[0];
                struct bpf_insn         insnsi[0];
                struct work_struct      work;
        };
};
so that 'struct bpf_prog' can be used independent of sockets and cleans up
'unattached' bpf use cases

split SK_RUN_FILTER macro into:
    SK_RUN_FILTER to be used with 'struct sk_filter *' and
    BPF_PROG_RUN to be used with 'struct bpf_prog *'

__sk_filter_release(struct sk_filter *) gains
__bpf_prog_release(struct bpf_prog *) helper function

also perform related renames for the functions that work
with 'struct bpf_prog *', since they're on the same lines:

sk_filter_size -> bpf_prog_size
sk_filter_select_runtime -> bpf_prog_select_runtime
sk_filter_free -> bpf_prog_free
sk_unattached_filter_create -> bpf_prog_create
sk_unattached_filter_destroy -> bpf_prog_destroy
sk_store_orig_filter -> bpf_prog_store_orig_filter
sk_release_orig_filter -> bpf_release_orig_filter
__sk_migrate_filter -> bpf_migrate_filter
__sk_prepare_filter -> bpf_prepare_filter

API for attaching classic BPF to a socket stays the same:
sk_attach_filter(prog, struct sock *)/sk_detach_filter(struct sock *)
and SK_RUN_FILTER(struct sk_filter *, ctx) to execute a program
which is used by sockets, tun, af_packet

API for 'unattached' BPF programs becomes:
bpf_prog_create(struct bpf_prog **)/bpf_prog_destroy(struct bpf_prog *)
and BPF_PROG_RUN(struct bpf_prog *, ctx) to execute a program
which is used by isdn, ppp, team, seccomp, ptp, xt_bpf, cls_bpf, test_bpf

Signed-off-by: Alexei Starovoitov <ast@plumgrid.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
21 files changed:
Documentation/networking/filter.txt
arch/arm/net/bpf_jit_32.c
arch/mips/net/bpf_jit.c
arch/powerpc/net/bpf_jit_comp.c
arch/s390/net/bpf_jit_comp.c
arch/sparc/net/bpf_jit_comp.c
arch/x86/net/bpf_jit_comp.c
drivers/isdn/i4l/isdn_ppp.c
drivers/net/ppp/ppp_generic.c
drivers/net/team/team_mode_loadbalance.c
include/linux/filter.h
include/linux/isdn_ppp.h
include/uapi/linux/netfilter/xt_bpf.h
kernel/bpf/core.c
kernel/seccomp.c
lib/test_bpf.c
net/core/filter.c
net/core/ptp_classifier.c
net/core/sock_diag.c
net/netfilter/xt_bpf.c
net/sched/cls_bpf.c

index 712068be8171de966f9d01b4dfa66f710d981abf..c48a9704bda8eaae5851bf54a0b0277cd01c88b4 100644 (file)
@@ -586,11 +586,11 @@ team driver's classifier for its load-balancing mode, netfilter's xt_bpf
 extension, PTP dissector/classifier, and much more. They are all internally
 converted by the kernel into the new instruction set representation and run
 in the eBPF interpreter. For in-kernel handlers, this all works transparently
-by using sk_unattached_filter_create() for setting up the filter, resp.
-sk_unattached_filter_destroy() for destroying it. The macro
-SK_RUN_FILTER(filter, ctx) transparently invokes eBPF interpreter or JITed
-code to run the filter. 'filter' is a pointer to struct sk_filter that we
-got from sk_unattached_filter_create(), and 'ctx' the given context (e.g.
+by using bpf_prog_create() for setting up the filter, resp.
+bpf_prog_destroy() for destroying it. The macro
+BPF_PROG_RUN(filter, ctx) transparently invokes eBPF interpreter or JITed
+code to run the filter. 'filter' is a pointer to struct bpf_prog that we
+got from bpf_prog_create(), and 'ctx' the given context (e.g.
 skb pointer). All constraints and restrictions from bpf_check_classic() apply
 before a conversion to the new layout is being done behind the scenes!
 
index fb5503ce016f0fa693f765f0157a2379422202dd..a37b989a2f91e302be654a7633082321e8f0c5bb 100644 (file)
@@ -56,7 +56,7 @@
 #define FLAG_NEED_X_RESET      (1 << 0)
 
 struct jit_ctx {
-       const struct sk_filter *skf;
+       const struct bpf_prog *skf;
        unsigned idx;
        unsigned prologue_bytes;
        int ret0_fp_idx;
@@ -465,7 +465,7 @@ static inline void update_on_xread(struct jit_ctx *ctx)
 static int build_body(struct jit_ctx *ctx)
 {
        void *load_func[] = {jit_get_skb_b, jit_get_skb_h, jit_get_skb_w};
-       const struct sk_filter *prog = ctx->skf;
+       const struct bpf_prog *prog = ctx->skf;
        const struct sock_filter *inst;
        unsigned i, load_order, off, condt;
        int imm12;
@@ -857,7 +857,7 @@ b_epilogue:
 }
 
 
-void bpf_jit_compile(struct sk_filter *fp)
+void bpf_jit_compile(struct bpf_prog *fp)
 {
        struct jit_ctx ctx;
        unsigned tmp_idx;
@@ -926,7 +926,7 @@ out:
        return;
 }
 
-void bpf_jit_free(struct sk_filter *fp)
+void bpf_jit_free(struct bpf_prog *fp)
 {
        if (fp->jited)
                module_free(NULL, fp->bpf_func);
index b87390a56a2fbdb42d13b3e194af6ca04127eadf..05a56619ece2f72044641c6aa1fd2d99f558bb69 100644 (file)
  * @target:            Memory location for the compiled filter
  */
 struct jit_ctx {
-       const struct sk_filter *skf;
+       const struct bpf_prog *skf;
        unsigned int prologue_bytes;
        u32 idx;
        u32 flags;
@@ -789,7 +789,7 @@ static int pkt_type_offset(void)
 static int build_body(struct jit_ctx *ctx)
 {
        void *load_func[] = {jit_get_skb_b, jit_get_skb_h, jit_get_skb_w};
-       const struct sk_filter *prog = ctx->skf;
+       const struct bpf_prog *prog = ctx->skf;
        const struct sock_filter *inst;
        unsigned int i, off, load_order, condt;
        u32 k, b_off __maybe_unused;
@@ -1369,7 +1369,7 @@ jmp_cmp:
 
 int bpf_jit_enable __read_mostly;
 
-void bpf_jit_compile(struct sk_filter *fp)
+void bpf_jit_compile(struct bpf_prog *fp)
 {
        struct jit_ctx ctx;
        unsigned int alloc_size, tmp_idx;
@@ -1423,7 +1423,7 @@ out:
        kfree(ctx.offsets);
 }
 
-void bpf_jit_free(struct sk_filter *fp)
+void bpf_jit_free(struct bpf_prog *fp)
 {
        if (fp->jited)
                module_free(NULL, fp->bpf_func);
index 82e82cadcde56589361ce04c98d554bcf4faec7d..3afa6f4c195705569726ef927ddc8f392af097ac 100644 (file)
@@ -25,7 +25,7 @@ static inline void bpf_flush_icache(void *start, void *end)
        flush_icache_range((unsigned long)start, (unsigned long)end);
 }
 
-static void bpf_jit_build_prologue(struct sk_filter *fp, u32 *image,
+static void bpf_jit_build_prologue(struct bpf_prog *fp, u32 *image,
                                   struct codegen_context *ctx)
 {
        int i;
@@ -121,7 +121,7 @@ static void bpf_jit_build_epilogue(u32 *image, struct codegen_context *ctx)
        ((int)K < 0 ? ((int)K >= SKF_LL_OFF ? func##_negative_offset : func) : func##_positive_offset)
 
 /* Assemble the body code between the prologue & epilogue. */
-static int bpf_jit_build_body(struct sk_filter *fp, u32 *image,
+static int bpf_jit_build_body(struct bpf_prog *fp, u32 *image,
                              struct codegen_context *ctx,
                              unsigned int *addrs)
 {
@@ -569,7 +569,7 @@ static int bpf_jit_build_body(struct sk_filter *fp, u32 *image,
        return 0;
 }
 
-void bpf_jit_compile(struct sk_filter *fp)
+void bpf_jit_compile(struct bpf_prog *fp)
 {
        unsigned int proglen;
        unsigned int alloclen;
@@ -693,7 +693,7 @@ out:
        return;
 }
 
-void bpf_jit_free(struct sk_filter *fp)
+void bpf_jit_free(struct bpf_prog *fp)
 {
        if (fp->jited)
                module_free(NULL, fp->bpf_func);
index a2cbd875543a3f206328221c521be3e3aeeab596..61e45b7c04d7bdf122a195903face00cfb6c2ce3 100644 (file)
@@ -812,7 +812,7 @@ static struct bpf_binary_header *bpf_alloc_binary(unsigned int bpfsize,
        return header;
 }
 
-void bpf_jit_compile(struct sk_filter *fp)
+void bpf_jit_compile(struct bpf_prog *fp)
 {
        struct bpf_binary_header *header = NULL;
        unsigned long size, prg_len, lit_len;
@@ -875,7 +875,7 @@ out:
        kfree(addrs);
 }
 
-void bpf_jit_free(struct sk_filter *fp)
+void bpf_jit_free(struct bpf_prog *fp)
 {
        unsigned long addr = (unsigned long)fp->bpf_func & PAGE_MASK;
        struct bpf_binary_header *header = (void *)addr;
index 892a102671adafc03950f907612f23b6902f934a..1f76c22a6a75d64a5dafd6f24da19eec6c816e96 100644 (file)
@@ -354,7 +354,7 @@ do {        *prog++ = BR_OPC | WDISP22(OFF);                \
  * emit_jump() calls with adjusted offsets.
  */
 
-void bpf_jit_compile(struct sk_filter *fp)
+void bpf_jit_compile(struct bpf_prog *fp)
 {
        unsigned int cleanup_addr, proglen, oldproglen = 0;
        u32 temp[8], *prog, *func, seen = 0, pass;
@@ -808,7 +808,7 @@ out:
        return;
 }
 
-void bpf_jit_free(struct sk_filter *fp)
+void bpf_jit_free(struct bpf_prog *fp)
 {
        if (fp->jited)
                module_free(NULL, fp->bpf_func);
index e2ecc1380b3de28b6e0726a285c76d3a6026687b..5c8cb8043c5a2b19df3e524d9d35ee3ec1949cd5 100644 (file)
@@ -211,7 +211,7 @@ struct jit_context {
        bool seen_ld_abs;
 };
 
-static int do_jit(struct sk_filter *bpf_prog, int *addrs, u8 *image,
+static int do_jit(struct bpf_prog *bpf_prog, int *addrs, u8 *image,
                  int oldproglen, struct jit_context *ctx)
 {
        struct bpf_insn *insn = bpf_prog->insnsi;
@@ -841,7 +841,7 @@ common_load:                ctx->seen_ld_abs = true;
                        /* By design x64 JIT should support all BPF instructions
                         * This error will be seen if new instruction was added
                         * to interpreter, but not to JIT
-                        * or if there is junk in sk_filter
+                        * or if there is junk in bpf_prog
                         */
                        pr_err("bpf_jit: unknown opcode %02x\n", insn->code);
                        return -EINVAL;
@@ -862,11 +862,11 @@ common_load:              ctx->seen_ld_abs = true;
        return proglen;
 }
 
-void bpf_jit_compile(struct sk_filter *prog)
+void bpf_jit_compile(struct bpf_prog *prog)
 {
 }
 
-void bpf_int_jit_compile(struct sk_filter *prog)
+void bpf_int_jit_compile(struct bpf_prog *prog)
 {
        struct bpf_binary_header *header = NULL;
        int proglen, oldproglen = 0;
@@ -932,7 +932,7 @@ out:
 
 static void bpf_jit_free_deferred(struct work_struct *work)
 {
-       struct sk_filter *fp = container_of(work, struct sk_filter, work);
+       struct bpf_prog *fp = container_of(work, struct bpf_prog, work);
        unsigned long addr = (unsigned long)fp->bpf_func & PAGE_MASK;
        struct bpf_binary_header *header = (void *)addr;
 
@@ -941,7 +941,7 @@ static void bpf_jit_free_deferred(struct work_struct *work)
        kfree(fp);
 }
 
-void bpf_jit_free(struct sk_filter *fp)
+void bpf_jit_free(struct bpf_prog *fp)
 {
        if (fp->jited) {
                INIT_WORK(&fp->work, bpf_jit_free_deferred);
index 62f0688d45a500530fd65014432be9fbfd5a7d69..c4198fa490bfac50dd3c05fa324502d4e6f228c8 100644 (file)
@@ -379,12 +379,12 @@ isdn_ppp_release(int min, struct file *file)
 #endif
 #ifdef CONFIG_IPPP_FILTER
        if (is->pass_filter) {
-               sk_unattached_filter_destroy(is->pass_filter);
+               bpf_prog_destroy(is->pass_filter);
                is->pass_filter = NULL;
        }
 
        if (is->active_filter) {
-               sk_unattached_filter_destroy(is->active_filter);
+               bpf_prog_destroy(is->active_filter);
                is->active_filter = NULL;
        }
 #endif
@@ -639,12 +639,11 @@ isdn_ppp_ioctl(int min, struct file *file, unsigned int cmd, unsigned long arg)
                fprog.filter = code;
 
                if (is->pass_filter) {
-                       sk_unattached_filter_destroy(is->pass_filter);
+                       bpf_prog_destroy(is->pass_filter);
                        is->pass_filter = NULL;
                }
                if (fprog.filter != NULL)
-                       err = sk_unattached_filter_create(&is->pass_filter,
-                                                         &fprog);
+                       err = bpf_prog_create(&is->pass_filter, &fprog);
                else
                        err = 0;
                kfree(code);
@@ -664,12 +663,11 @@ isdn_ppp_ioctl(int min, struct file *file, unsigned int cmd, unsigned long arg)
                fprog.filter = code;
 
                if (is->active_filter) {
-                       sk_unattached_filter_destroy(is->active_filter);
+                       bpf_prog_destroy(is->active_filter);
                        is->active_filter = NULL;
                }
                if (fprog.filter != NULL)
-                       err = sk_unattached_filter_create(&is->active_filter,
-                                                         &fprog);
+                       err = bpf_prog_create(&is->active_filter, &fprog);
                else
                        err = 0;
                kfree(code);
@@ -1174,14 +1172,14 @@ isdn_ppp_push_higher(isdn_net_dev *net_dev, isdn_net_local *lp, struct sk_buff *
        }
 
        if (is->pass_filter
-           && SK_RUN_FILTER(is->pass_filter, skb) == 0) {
+           && BPF_PROG_RUN(is->pass_filter, skb) == 0) {
                if (is->debug & 0x2)
                        printk(KERN_DEBUG "IPPP: inbound frame filtered.\n");
                kfree_skb(skb);
                return;
        }
        if (!(is->active_filter
-             && SK_RUN_FILTER(is->active_filter, skb) == 0)) {
+             && BPF_PROG_RUN(is->active_filter, skb) == 0)) {
                if (is->debug & 0x2)
                        printk(KERN_DEBUG "IPPP: link-active filter: resetting huptimer.\n");
                lp->huptimer = 0;
@@ -1320,14 +1318,14 @@ isdn_ppp_xmit(struct sk_buff *skb, struct net_device *netdev)
        }
 
        if (ipt->pass_filter
-           && SK_RUN_FILTER(ipt->pass_filter, skb) == 0) {
+           && BPF_PROG_RUN(ipt->pass_filter, skb) == 0) {
                if (ipt->debug & 0x4)
                        printk(KERN_DEBUG "IPPP: outbound frame filtered.\n");
                kfree_skb(skb);
                goto unlock;
        }
        if (!(ipt->active_filter
-             && SK_RUN_FILTER(ipt->active_filter, skb) == 0)) {
+             && BPF_PROG_RUN(ipt->active_filter, skb) == 0)) {
                if (ipt->debug & 0x4)
                        printk(KERN_DEBUG "IPPP: link-active filter: resetting huptimer.\n");
                lp->huptimer = 0;
@@ -1517,9 +1515,9 @@ int isdn_ppp_autodial_filter(struct sk_buff *skb, isdn_net_local *lp)
        }
 
        drop |= is->pass_filter
-               && SK_RUN_FILTER(is->pass_filter, skb) == 0;
+               && BPF_PROG_RUN(is->pass_filter, skb) == 0;
        drop |= is->active_filter
-               && SK_RUN_FILTER(is->active_filter, skb) == 0;
+               && BPF_PROG_RUN(is->active_filter, skb) == 0;
 
        skb_push(skb, IPPP_MAX_HEADER - 4);
        return drop;
index 765248b42a0aac6a42ad9ad0faf1c890f0f2ea1b..fa0d71727894a2ef8dc74f8e27c68bbce87f44fa 100644 (file)
@@ -143,8 +143,8 @@ struct ppp {
        struct sk_buff_head mrq;        /* MP: receive reconstruction queue */
 #endif /* CONFIG_PPP_MULTILINK */
 #ifdef CONFIG_PPP_FILTER
-       struct sk_filter *pass_filter;  /* filter for packets to pass */
-       struct sk_filter *active_filter;/* filter for pkts to reset idle */
+       struct bpf_prog *pass_filter;   /* filter for packets to pass */
+       struct bpf_prog *active_filter; /* filter for pkts to reset idle */
 #endif /* CONFIG_PPP_FILTER */
        struct net      *ppp_net;       /* the net we belong to */
        struct ppp_link_stats stats64;  /* 64 bit network stats */
@@ -762,12 +762,12 @@ static long ppp_ioctl(struct file *file, unsigned int cmd, unsigned long arg)
 
                        ppp_lock(ppp);
                        if (ppp->pass_filter) {
-                               sk_unattached_filter_destroy(ppp->pass_filter);
+                               bpf_prog_destroy(ppp->pass_filter);
                                ppp->pass_filter = NULL;
                        }
                        if (fprog.filter != NULL)
-                               err = sk_unattached_filter_create(&ppp->pass_filter,
-                                                                 &fprog);
+                               err = bpf_prog_create(&ppp->pass_filter,
+                                                     &fprog);
                        else
                                err = 0;
                        kfree(code);
@@ -788,12 +788,12 @@ static long ppp_ioctl(struct file *file, unsigned int cmd, unsigned long arg)
 
                        ppp_lock(ppp);
                        if (ppp->active_filter) {
-                               sk_unattached_filter_destroy(ppp->active_filter);
+                               bpf_prog_destroy(ppp->active_filter);
                                ppp->active_filter = NULL;
                        }
                        if (fprog.filter != NULL)
-                               err = sk_unattached_filter_create(&ppp->active_filter,
-                                                                 &fprog);
+                               err = bpf_prog_create(&ppp->active_filter,
+                                                     &fprog);
                        else
                                err = 0;
                        kfree(code);
@@ -1205,7 +1205,7 @@ ppp_send_frame(struct ppp *ppp, struct sk_buff *skb)
                   a four-byte PPP header on each packet */
                *skb_push(skb, 2) = 1;
                if (ppp->pass_filter &&
-                   SK_RUN_FILTER(ppp->pass_filter, skb) == 0) {
+                   BPF_PROG_RUN(ppp->pass_filter, skb) == 0) {
                        if (ppp->debug & 1)
                                netdev_printk(KERN_DEBUG, ppp->dev,
                                              "PPP: outbound frame "
@@ -1215,7 +1215,7 @@ ppp_send_frame(struct ppp *ppp, struct sk_buff *skb)
                }
                /* if this packet passes the active filter, record the time */
                if (!(ppp->active_filter &&
-                     SK_RUN_FILTER(ppp->active_filter, skb) == 0))
+                     BPF_PROG_RUN(ppp->active_filter, skb) == 0))
                        ppp->last_xmit = jiffies;
                skb_pull(skb, 2);
 #else
@@ -1839,7 +1839,7 @@ ppp_receive_nonmp_frame(struct ppp *ppp, struct sk_buff *skb)
 
                        *skb_push(skb, 2) = 0;
                        if (ppp->pass_filter &&
-                           SK_RUN_FILTER(ppp->pass_filter, skb) == 0) {
+                           BPF_PROG_RUN(ppp->pass_filter, skb) == 0) {
                                if (ppp->debug & 1)
                                        netdev_printk(KERN_DEBUG, ppp->dev,
                                                      "PPP: inbound frame "
@@ -1848,7 +1848,7 @@ ppp_receive_nonmp_frame(struct ppp *ppp, struct sk_buff *skb)
                                return;
                        }
                        if (!(ppp->active_filter &&
-                             SK_RUN_FILTER(ppp->active_filter, skb) == 0))
+                             BPF_PROG_RUN(ppp->active_filter, skb) == 0))
                                ppp->last_recv = jiffies;
                        __skb_pull(skb, 2);
                } else
@@ -2829,12 +2829,12 @@ static void ppp_destroy_interface(struct ppp *ppp)
 #endif /* CONFIG_PPP_MULTILINK */
 #ifdef CONFIG_PPP_FILTER
        if (ppp->pass_filter) {
-               sk_unattached_filter_destroy(ppp->pass_filter);
+               bpf_prog_destroy(ppp->pass_filter);
                ppp->pass_filter = NULL;
        }
 
        if (ppp->active_filter) {
-               sk_unattached_filter_destroy(ppp->active_filter);
+               bpf_prog_destroy(ppp->active_filter);
                ppp->active_filter = NULL;
        }
 #endif /* CONFIG_PPP_FILTER */
index d7be9b36bce67bd60e0549f86c0fb5c8f942f866..a1536d0d83a9700fe1281565c2b4fcb2f81134b0 100644 (file)
@@ -58,7 +58,7 @@ struct lb_priv_ex {
 };
 
 struct lb_priv {
-       struct sk_filter __rcu *fp;
+       struct bpf_prog __rcu *fp;
        lb_select_tx_port_func_t __rcu *select_tx_port_func;
        struct lb_pcpu_stats __percpu *pcpu_stats;
        struct lb_priv_ex *ex; /* priv extension */
@@ -174,14 +174,14 @@ static lb_select_tx_port_func_t *lb_select_tx_port_get_func(const char *name)
 static unsigned int lb_get_skb_hash(struct lb_priv *lb_priv,
                                    struct sk_buff *skb)
 {
-       struct sk_filter *fp;
+       struct bpf_prog *fp;
        uint32_t lhash;
        unsigned char *c;
 
        fp = rcu_dereference_bh(lb_priv->fp);
        if (unlikely(!fp))
                return 0;
-       lhash = SK_RUN_FILTER(fp, skb);
+       lhash = BPF_PROG_RUN(fp, skb);
        c = (char *) &lhash;
        return c[0] ^ c[1] ^ c[2] ^ c[3];
 }
@@ -271,8 +271,8 @@ static void __fprog_destroy(struct sock_fprog_kern *fprog)
 static int lb_bpf_func_set(struct team *team, struct team_gsetter_ctx *ctx)
 {
        struct lb_priv *lb_priv = get_lb_priv(team);
-       struct sk_filter *fp = NULL;
-       struct sk_filter *orig_fp = NULL;
+       struct bpf_prog *fp = NULL;
+       struct bpf_prog *orig_fp = NULL;
        struct sock_fprog_kern *fprog = NULL;
        int err;
 
@@ -281,7 +281,7 @@ static int lb_bpf_func_set(struct team *team, struct team_gsetter_ctx *ctx)
                                     ctx->data.bin_val.ptr);
                if (err)
                        return err;
-               err = sk_unattached_filter_create(&fp, fprog);
+               err = bpf_prog_create(&fp, fprog);
                if (err) {
                        __fprog_destroy(fprog);
                        return err;
@@ -300,7 +300,7 @@ static int lb_bpf_func_set(struct team *team, struct team_gsetter_ctx *ctx)
 
        if (orig_fp) {
                synchronize_rcu();
-               sk_unattached_filter_destroy(orig_fp);
+               bpf_prog_destroy(orig_fp);
        }
        return 0;
 }
index 7cb9b40e9a2f6c67a159d8cccdc05accc64ff67a..a5227ab8ccb17ebd4cf9ed7a55f80c352a2d6fb3 100644 (file)
@@ -296,7 +296,8 @@ enum {
 })
 
 /* Macro to invoke filter function. */
-#define SK_RUN_FILTER(filter, ctx)  (*filter->bpf_func)(ctx, filter->insnsi)
+#define SK_RUN_FILTER(filter, ctx) \
+       (*filter->prog->bpf_func)(ctx, filter->prog->insnsi)
 
 struct bpf_insn {
        __u8    code;           /* opcode */
@@ -323,12 +324,10 @@ struct sk_buff;
 struct sock;
 struct seccomp_data;
 
-struct sk_filter {
-       atomic_t                refcnt;
+struct bpf_prog {
        u32                     jited:1,        /* Is our filter JIT'ed? */
                                len:31;         /* Number of filter blocks */
        struct sock_fprog_kern  *orig_prog;     /* Original BPF program */
-       struct rcu_head         rcu;
        unsigned int            (*bpf_func)(const struct sk_buff *skb,
                                            const struct bpf_insn *filter);
        union {
@@ -338,25 +337,32 @@ struct sk_filter {
        };
 };
 
-static inline unsigned int sk_filter_size(unsigned int proglen)
+struct sk_filter {
+       atomic_t        refcnt;
+       struct rcu_head rcu;
+       struct bpf_prog *prog;
+};
+
+#define BPF_PROG_RUN(filter, ctx)  (*filter->bpf_func)(ctx, filter->insnsi)
+
+static inline unsigned int bpf_prog_size(unsigned int proglen)
 {
-       return max(sizeof(struct sk_filter),
-                  offsetof(struct sk_filter, insns[proglen]));
+       return max(sizeof(struct bpf_prog),
+                  offsetof(struct bpf_prog, insns[proglen]));
 }
 
 #define bpf_classic_proglen(fprog) (fprog->len * sizeof(fprog->filter[0]))
 
 int sk_filter(struct sock *sk, struct sk_buff *skb);
 
-void sk_filter_select_runtime(struct sk_filter *fp);
-void sk_filter_free(struct sk_filter *fp);
+void bpf_prog_select_runtime(struct bpf_prog *fp);
+void bpf_prog_free(struct bpf_prog *fp);
 
 int bpf_convert_filter(struct sock_filter *prog, int len,
                       struct bpf_insn *new_prog, int *new_len);
 
-int sk_unattached_filter_create(struct sk_filter **pfp,
-                               struct sock_fprog_kern *fprog);
-void sk_unattached_filter_destroy(struct sk_filter *fp);
+int bpf_prog_create(struct bpf_prog **pfp, struct sock_fprog_kern *fprog);
+void bpf_prog_destroy(struct bpf_prog *fp);
 
 int sk_attach_filter(struct sock_fprog *fprog, struct sock *sk);
 int sk_detach_filter(struct sock *sk);
@@ -369,7 +375,7 @@ bool sk_filter_charge(struct sock *sk, struct sk_filter *fp);
 void sk_filter_uncharge(struct sock *sk, struct sk_filter *fp);
 
 u64 __bpf_call_base(u64 r1, u64 r2, u64 r3, u64 r4, u64 r5);
-void bpf_int_jit_compile(struct sk_filter *fp);
+void bpf_int_jit_compile(struct bpf_prog *fp);
 
 #define BPF_ANC                BIT(15)
 
@@ -423,8 +429,8 @@ static inline void *bpf_load_pointer(const struct sk_buff *skb, int k,
 #include <linux/linkage.h>
 #include <linux/printk.h>
 
-void bpf_jit_compile(struct sk_filter *fp);
-void bpf_jit_free(struct sk_filter *fp);
+void bpf_jit_compile(struct bpf_prog *fp);
+void bpf_jit_free(struct bpf_prog *fp);
 
 static inline void bpf_jit_dump(unsigned int flen, unsigned int proglen,
                                u32 pass, void *image)
@@ -438,11 +444,11 @@ static inline void bpf_jit_dump(unsigned int flen, unsigned int proglen,
 #else
 #include <linux/slab.h>
 
-static inline void bpf_jit_compile(struct sk_filter *fp)
+static inline void bpf_jit_compile(struct bpf_prog *fp)
 {
 }
 
-static inline void bpf_jit_free(struct sk_filter *fp)
+static inline void bpf_jit_free(struct bpf_prog *fp)
 {
        kfree(fp);
 }
index 8e10f57f109f5a9eff472fe7a90dbce48e4a20dc..a0070c6dfaf88159a3742bce0aa8fe5d81c3b035 100644 (file)
@@ -180,8 +180,8 @@ struct ippp_struct {
   struct slcompress *slcomp;
 #endif
 #ifdef CONFIG_IPPP_FILTER
-  struct sk_filter *pass_filter;   /* filter for packets to pass */
-  struct sk_filter *active_filter; /* filter for pkts to reset idle */
+  struct bpf_prog *pass_filter;   /* filter for packets to pass */
+  struct bpf_prog *active_filter; /* filter for pkts to reset idle */
 #endif
   unsigned long debug;
   struct isdn_ppp_compressor *compressor,*decompressor;
index 2ec9fbcd06f9a489c0669521bf77728af69e0ed0..1fad2c27ac320d08f6f83ccbf7600b1c2768eb98 100644 (file)
@@ -6,14 +6,14 @@
 
 #define XT_BPF_MAX_NUM_INSTR   64
 
-struct sk_filter;
+struct bpf_prog;
 
 struct xt_bpf_info {
        __u16 bpf_program_num_elem;
        struct sock_filter bpf_program[XT_BPF_MAX_NUM_INSTR];
 
        /* only used in the kernel */
-       struct sk_filter *filter __attribute__((aligned(8)));
+       struct bpf_prog *filter __attribute__((aligned(8)));
 };
 
 #endif /*_XT_BPF_H */
index 188ac5ba390089c0f9a7af924277f169a2c52726..7f0dbcbb34af1559657adea42b7c726f145a3986 100644 (file)
@@ -73,15 +73,13 @@ noinline u64 __bpf_call_base(u64 r1, u64 r2, u64 r3, u64 r4, u64 r5)
 }
 
 /**
- *     __sk_run_filter - run a filter on a given context
- *     @ctx: buffer to run the filter on
- *     @insn: filter to apply
+ *     __bpf_prog_run - run eBPF program on a given context
+ *     @ctx: is the data we are operating on
+ *     @insn: is the array of eBPF instructions
  *
- * Decode and apply filter instructions to the skb->data. Return length to
- * keep, 0 for none. @ctx is the data we are operating on, @insn is the
- * array of filter instructions.
+ * Decode and execute eBPF instructions.
  */
-static unsigned int __sk_run_filter(void *ctx, const struct bpf_insn *insn)
+static unsigned int __bpf_prog_run(void *ctx, const struct bpf_insn *insn)
 {
        u64 stack[MAX_BPF_STACK / sizeof(u64)];
        u64 regs[MAX_BPF_REG], tmp;
@@ -508,29 +506,29 @@ load_byte:
                return 0;
 }
 
-void __weak bpf_int_jit_compile(struct sk_filter *prog)
+void __weak bpf_int_jit_compile(struct bpf_prog *prog)
 {
 }
 
 /**
- *     sk_filter_select_runtime - select execution runtime for BPF program
- *     @fp: sk_filter populated with internal BPF program
+ *     bpf_prog_select_runtime - select execution runtime for BPF program
+ *     @fp: bpf_prog populated with internal BPF program
  *
  * try to JIT internal BPF program, if JIT is not available select interpreter
- * BPF program will be executed via SK_RUN_FILTER() macro
+ * BPF program will be executed via BPF_PROG_RUN() macro
  */
-void sk_filter_select_runtime(struct sk_filter *fp)
+void bpf_prog_select_runtime(struct bpf_prog *fp)
 {
-       fp->bpf_func = (void *) __sk_run_filter;
+       fp->bpf_func = (void *) __bpf_prog_run;
 
        /* Probe if internal BPF can be JITed */
        bpf_int_jit_compile(fp);
 }
-EXPORT_SYMBOL_GPL(sk_filter_select_runtime);
+EXPORT_SYMBOL_GPL(bpf_prog_select_runtime);
 
 /* free internal BPF program */
-void sk_filter_free(struct sk_filter *fp)
+void bpf_prog_free(struct bpf_prog *fp)
 {
        bpf_jit_free(fp);
 }
-EXPORT_SYMBOL_GPL(sk_filter_free);
+EXPORT_SYMBOL_GPL(bpf_prog_free);
index 33a3a97e2b58c5ddefe0b76065e12fe5c05ebbae..2f3fa2cc2eac2bfaf4c3082fbe3d267898c2f915 100644 (file)
@@ -54,7 +54,7 @@
 struct seccomp_filter {
        atomic_t usage;
        struct seccomp_filter *prev;
-       struct sk_filter *prog;
+       struct bpf_prog *prog;
 };
 
 /* Limit any path through the tree to 256KB worth of instructions. */
@@ -187,7 +187,7 @@ static u32 seccomp_run_filters(int syscall)
         * value always takes priority (ignoring the DATA).
         */
        for (f = current->seccomp.filter; f; f = f->prev) {
-               u32 cur_ret = SK_RUN_FILTER(f->prog, (void *)&sd);
+               u32 cur_ret = BPF_PROG_RUN(f->prog, (void *)&sd);
 
                if ((cur_ret & SECCOMP_RET_ACTION) < (ret & SECCOMP_RET_ACTION))
                        ret = cur_ret;
@@ -260,7 +260,7 @@ static long seccomp_attach_filter(struct sock_fprog *fprog)
        if (!filter)
                goto free_prog;
 
-       filter->prog = kzalloc(sk_filter_size(new_len),
+       filter->prog = kzalloc(bpf_prog_size(new_len),
                               GFP_KERNEL|__GFP_NOWARN);
        if (!filter->prog)
                goto free_filter;
@@ -273,7 +273,7 @@ static long seccomp_attach_filter(struct sock_fprog *fprog)
        atomic_set(&filter->usage, 1);
        filter->prog->len = new_len;
 
-       sk_filter_select_runtime(filter->prog);
+       bpf_prog_select_runtime(filter->prog);
 
        /*
         * If there is an existing filter, make it the prev and don't drop its
@@ -337,7 +337,7 @@ void put_seccomp_filter(struct task_struct *tsk)
        while (orig && atomic_dec_and_test(&orig->usage)) {
                struct seccomp_filter *freeme = orig;
                orig = orig->prev;
-               sk_filter_free(freeme->prog);
+               bpf_prog_free(freeme->prog);
                kfree(freeme);
        }
 }
index 5f48623ee1a797188bc32f5aefc3cf00c00c696c..89e0345733bd33db9bc9ab1f6184a1d376ec9c89 100644 (file)
@@ -1761,9 +1761,9 @@ static int probe_filter_length(struct sock_filter *fp)
        return len + 1;
 }
 
-static struct sk_filter *generate_filter(int which, int *err)
+static struct bpf_prog *generate_filter(int which, int *err)
 {
-       struct sk_filter *fp;
+       struct bpf_prog *fp;
        struct sock_fprog_kern fprog;
        unsigned int flen = probe_filter_length(tests[which].u.insns);
        __u8 test_type = tests[which].aux & TEST_TYPE_MASK;
@@ -1773,7 +1773,7 @@ static struct sk_filter *generate_filter(int which, int *err)
                fprog.filter = tests[which].u.insns;
                fprog.len = flen;
 
-               *err = sk_unattached_filter_create(&fp, &fprog);
+               *err = bpf_prog_create(&fp, &fprog);
                if (tests[which].aux & FLAG_EXPECTED_FAIL) {
                        if (*err == -EINVAL) {
                                pr_cont("PASS\n");
@@ -1798,7 +1798,7 @@ static struct sk_filter *generate_filter(int which, int *err)
                break;
 
        case INTERNAL:
-               fp = kzalloc(sk_filter_size(flen), GFP_KERNEL);
+               fp = kzalloc(bpf_prog_size(flen), GFP_KERNEL);
                if (fp == NULL) {
                        pr_cont("UNEXPECTED_FAIL no memory left\n");
                        *err = -ENOMEM;
@@ -1809,7 +1809,7 @@ static struct sk_filter *generate_filter(int which, int *err)
                memcpy(fp->insnsi, tests[which].u.insns_int,
                       fp->len * sizeof(struct bpf_insn));
 
-               sk_filter_select_runtime(fp);
+               bpf_prog_select_runtime(fp);
                break;
        }
 
@@ -1817,21 +1817,21 @@ static struct sk_filter *generate_filter(int which, int *err)
        return fp;
 }
 
-static void release_filter(struct sk_filter *fp, int which)
+static void release_filter(struct bpf_prog *fp, int which)
 {
        __u8 test_type = tests[which].aux & TEST_TYPE_MASK;
 
        switch (test_type) {
        case CLASSIC:
-               sk_unattached_filter_destroy(fp);
+               bpf_prog_destroy(fp);
                break;
        case INTERNAL:
-               sk_filter_free(fp);
+               bpf_prog_free(fp);
                break;
        }
 }
 
-static int __run_one(const struct sk_filter *fp, const void *data,
+static int __run_one(const struct bpf_prog *fp, const void *data,
                     int runs, u64 *duration)
 {
        u64 start, finish;
@@ -1840,7 +1840,7 @@ static int __run_one(const struct sk_filter *fp, const void *data,
        start = ktime_to_us(ktime_get());
 
        for (i = 0; i < runs; i++)
-               ret = SK_RUN_FILTER(fp, data);
+               ret = BPF_PROG_RUN(fp, data);
 
        finish = ktime_to_us(ktime_get());
 
@@ -1850,7 +1850,7 @@ static int __run_one(const struct sk_filter *fp, const void *data,
        return ret;
 }
 
-static int run_one(const struct sk_filter *fp, struct bpf_test *test)
+static int run_one(const struct bpf_prog *fp, struct bpf_test *test)
 {
        int err_cnt = 0, i, runs = MAX_TESTRUNS;
 
@@ -1884,7 +1884,7 @@ static __init int test_bpf(void)
        int i, err_cnt = 0, pass_cnt = 0;
 
        for (i = 0; i < ARRAY_SIZE(tests); i++) {
-               struct sk_filter *fp;
+               struct bpf_prog *fp;
                int err;
 
                pr_info("#%d %s ", i, tests[i].descr);
index 6ac901613bee1b53d8e43401ff1c0aa2a4028ff4..d814b8a89d0f2f65efb95858f2e4fe26bf5a4ca9 100644 (file)
@@ -810,8 +810,8 @@ int bpf_check_classic(const struct sock_filter *filter, unsigned int flen)
 }
 EXPORT_SYMBOL(bpf_check_classic);
 
-static int sk_store_orig_filter(struct sk_filter *fp,
-                               const struct sock_fprog *fprog)
+static int bpf_prog_store_orig_filter(struct bpf_prog *fp,
+                                     const struct sock_fprog *fprog)
 {
        unsigned int fsize = bpf_classic_proglen(fprog);
        struct sock_fprog_kern *fkprog;
@@ -831,7 +831,7 @@ static int sk_store_orig_filter(struct sk_filter *fp,
        return 0;
 }
 
-static void sk_release_orig_filter(struct sk_filter *fp)
+static void bpf_release_orig_filter(struct bpf_prog *fp)
 {
        struct sock_fprog_kern *fprog = fp->orig_prog;
 
@@ -841,10 +841,16 @@ static void sk_release_orig_filter(struct sk_filter *fp)
        }
 }
 
+static void __bpf_prog_release(struct bpf_prog *prog)
+{
+       bpf_release_orig_filter(prog);
+       bpf_prog_free(prog);
+}
+
 static void __sk_filter_release(struct sk_filter *fp)
 {
-       sk_release_orig_filter(fp);
-       sk_filter_free(fp);
+       __bpf_prog_release(fp->prog);
+       kfree(fp);
 }
 
 /**
@@ -872,7 +878,7 @@ static void sk_filter_release(struct sk_filter *fp)
 
 void sk_filter_uncharge(struct sock *sk, struct sk_filter *fp)
 {
-       u32 filter_size = sk_filter_size(fp->len);
+       u32 filter_size = bpf_prog_size(fp->prog->len);
 
        atomic_sub(filter_size, &sk->sk_omem_alloc);
        sk_filter_release(fp);
@@ -883,7 +889,7 @@ void sk_filter_uncharge(struct sock *sk, struct sk_filter *fp)
  */
 bool sk_filter_charge(struct sock *sk, struct sk_filter *fp)
 {
-       u32 filter_size = sk_filter_size(fp->len);
+       u32 filter_size = bpf_prog_size(fp->prog->len);
 
        /* same check as in sock_kmalloc() */
        if (filter_size <= sysctl_optmem_max &&
@@ -895,10 +901,10 @@ bool sk_filter_charge(struct sock *sk, struct sk_filter *fp)
        return false;
 }
 
-static struct sk_filter *__sk_migrate_filter(struct sk_filter *fp)
+static struct bpf_prog *bpf_migrate_filter(struct bpf_prog *fp)
 {
        struct sock_filter *old_prog;
-       struct sk_filter *old_fp;
+       struct bpf_prog *old_fp;
        int err, new_len, old_len = fp->len;
 
        /* We are free to overwrite insns et al right here as it
@@ -927,7 +933,7 @@ static struct sk_filter *__sk_migrate_filter(struct sk_filter *fp)
 
        /* Expand fp for appending the new filter representation. */
        old_fp = fp;
-       fp = krealloc(old_fp, sk_filter_size(new_len), GFP_KERNEL);
+       fp = krealloc(old_fp, bpf_prog_size(new_len), GFP_KERNEL);
        if (!fp) {
                /* The old_fp is still around in case we couldn't
                 * allocate new memory, so uncharge on that one.
@@ -949,7 +955,7 @@ static struct sk_filter *__sk_migrate_filter(struct sk_filter *fp)
                 */
                goto out_err_free;
 
-       sk_filter_select_runtime(fp);
+       bpf_prog_select_runtime(fp);
 
        kfree(old_prog);
        return fp;
@@ -957,11 +963,11 @@ static struct sk_filter *__sk_migrate_filter(struct sk_filter *fp)
 out_err_free:
        kfree(old_prog);
 out_err:
-       __sk_filter_release(fp);
+       __bpf_prog_release(fp);
        return ERR_PTR(err);
 }
 
-static struct sk_filter *__sk_prepare_filter(struct sk_filter *fp)
+static struct bpf_prog *bpf_prepare_filter(struct bpf_prog *fp)
 {
        int err;
 
@@ -970,7 +976,7 @@ static struct sk_filter *__sk_prepare_filter(struct sk_filter *fp)
 
        err = bpf_check_classic(fp->insns, fp->len);
        if (err) {
-               __sk_filter_release(fp);
+               __bpf_prog_release(fp);
                return ERR_PTR(err);
        }
 
@@ -983,13 +989,13 @@ static struct sk_filter *__sk_prepare_filter(struct sk_filter *fp)
         * internal BPF translation for the optimized interpreter.
         */
        if (!fp->jited)
-               fp = __sk_migrate_filter(fp);
+               fp = bpf_migrate_filter(fp);
 
        return fp;
 }
 
 /**
- *     sk_unattached_filter_create - create an unattached filter
+ *     bpf_prog_create - create an unattached filter
  *     @pfp: the unattached filter that is created
  *     @fprog: the filter program
  *
@@ -998,23 +1004,21 @@ static struct sk_filter *__sk_prepare_filter(struct sk_filter *fp)
  * If an error occurs or there is insufficient memory for the filter
  * a negative errno code is returned. On success the return is zero.
  */
-int sk_unattached_filter_create(struct sk_filter **pfp,
-                               struct sock_fprog_kern *fprog)
+int bpf_prog_create(struct bpf_prog **pfp, struct sock_fprog_kern *fprog)
 {
        unsigned int fsize = bpf_classic_proglen(fprog);
-       struct sk_filter *fp;
+       struct bpf_prog *fp;
 
        /* Make sure new filter is there and in the right amounts. */
        if (fprog->filter == NULL)
                return -EINVAL;
 
-       fp = kmalloc(sk_filter_size(fprog->len), GFP_KERNEL);
+       fp = kmalloc(bpf_prog_size(fprog->len), GFP_KERNEL);
        if (!fp)
                return -ENOMEM;
 
        memcpy(fp->insns, fprog->filter, fsize);
 
-       atomic_set(&fp->refcnt, 1);
        fp->len = fprog->len;
        /* Since unattached filters are not copied back to user
         * space through sk_get_filter(), we do not need to hold
@@ -1022,23 +1026,23 @@ int sk_unattached_filter_create(struct sk_filter **pfp,
         */
        fp->orig_prog = NULL;
 
-       /* __sk_prepare_filter() already takes care of freeing
+       /* bpf_prepare_filter() already takes care of freeing
         * memory in case something goes wrong.
         */
-       fp = __sk_prepare_filter(fp);
+       fp = bpf_prepare_filter(fp);
        if (IS_ERR(fp))
                return PTR_ERR(fp);
 
        *pfp = fp;
        return 0;
 }
-EXPORT_SYMBOL_GPL(sk_unattached_filter_create);
+EXPORT_SYMBOL_GPL(bpf_prog_create);
 
-void sk_unattached_filter_destroy(struct sk_filter *fp)
+void bpf_prog_destroy(struct bpf_prog *fp)
 {
-       __sk_filter_release(fp);
+       __bpf_prog_release(fp);
 }
-EXPORT_SYMBOL_GPL(sk_unattached_filter_destroy);
+EXPORT_SYMBOL_GPL(bpf_prog_destroy);
 
 /**
  *     sk_attach_filter - attach a socket filter
@@ -1054,7 +1058,8 @@ int sk_attach_filter(struct sock_fprog *fprog, struct sock *sk)
 {
        struct sk_filter *fp, *old_fp;
        unsigned int fsize = bpf_classic_proglen(fprog);
-       unsigned int sk_fsize = sk_filter_size(fprog->len);
+       unsigned int bpf_fsize = bpf_prog_size(fprog->len);
+       struct bpf_prog *prog;
        int err;
 
        if (sock_flag(sk, SOCK_FILTER_LOCKED))
@@ -1064,29 +1069,36 @@ int sk_attach_filter(struct sock_fprog *fprog, struct sock *sk)
        if (fprog->filter == NULL)
                return -EINVAL;
 
-       fp = kmalloc(sk_fsize, GFP_KERNEL);
-       if (!fp)
+       prog = kmalloc(bpf_fsize, GFP_KERNEL);
+       if (!prog)
                return -ENOMEM;
 
-       if (copy_from_user(fp->insns, fprog->filter, fsize)) {
-               kfree(fp);
+       if (copy_from_user(prog->insns, fprog->filter, fsize)) {
+               kfree(prog);
                return -EFAULT;
        }
 
-       fp->len = fprog->len;
+       prog->len = fprog->len;
 
-       err = sk_store_orig_filter(fp, fprog);
+       err = bpf_prog_store_orig_filter(prog, fprog);
        if (err) {
-               kfree(fp);
+               kfree(prog);
                return -ENOMEM;
        }
 
-       /* __sk_prepare_filter() already takes care of freeing
+       /* bpf_prepare_filter() already takes care of freeing
         * memory in case something goes wrong.
         */
-       fp = __sk_prepare_filter(fp);
-       if (IS_ERR(fp))
-               return PTR_ERR(fp);
+       prog = bpf_prepare_filter(prog);
+       if (IS_ERR(prog))
+               return PTR_ERR(prog);
+
+       fp = kmalloc(sizeof(*fp), GFP_KERNEL);
+       if (!fp) {
+               __bpf_prog_release(prog);
+               return -ENOMEM;
+       }
+       fp->prog = prog;
 
        atomic_set(&fp->refcnt, 0);
 
@@ -1142,7 +1154,7 @@ int sk_get_filter(struct sock *sk, struct sock_filter __user *ubuf,
        /* We're copying the filter that has been originally attached,
         * so no conversion/decode needed anymore.
         */
-       fprog = filter->orig_prog;
+       fprog = filter->prog->orig_prog;
 
        ret = fprog->len;
        if (!len)
index 12ab7b4be60932700901f4f3fb9d159f3ddb7ebd..4eab4a94a59defa60d79988670681f2b3cdbcda5 100644 (file)
 #include <linux/filter.h>
 #include <linux/ptp_classify.h>
 
-static struct sk_filter *ptp_insns __read_mostly;
+static struct bpf_prog *ptp_insns __read_mostly;
 
 unsigned int ptp_classify_raw(const struct sk_buff *skb)
 {
-       return SK_RUN_FILTER(ptp_insns, skb);
+       return BPF_PROG_RUN(ptp_insns, skb);
 }
 EXPORT_SYMBOL_GPL(ptp_classify_raw);
 
@@ -189,5 +189,5 @@ void __init ptp_classifier_init(void)
                .len = ARRAY_SIZE(ptp_filter), .filter = ptp_filter,
        };
 
-       BUG_ON(sk_unattached_filter_create(&ptp_insns, &ptp_prog));
+       BUG_ON(bpf_prog_create(&ptp_insns, &ptp_prog));
 }
index 57d922320c59c782e034cac5180438e50c9ba5e4..ad704c757bb4a6e48c685fe0478ce6525b08a3ab 100644 (file)
@@ -68,7 +68,7 @@ int sock_diag_put_filterinfo(bool may_report_filterinfo, struct sock *sk,
        if (!filter)
                goto out;
 
-       fprog = filter->orig_prog;
+       fprog = filter->prog->orig_prog;
        flen = bpf_classic_proglen(fprog);
 
        attr = nla_reserve(skb, attrtype, flen);
index bbffdbdaf6031bef784042c064dded271d6a406f..dffee9d47ec4b3333271ac0dd54a306cf9943e3b 100644 (file)
@@ -28,7 +28,7 @@ static int bpf_mt_check(const struct xt_mtchk_param *par)
        program.len = info->bpf_program_num_elem;
        program.filter = info->bpf_program;
 
-       if (sk_unattached_filter_create(&info->filter, &program)) {
+       if (bpf_prog_create(&info->filter, &program)) {
                pr_info("bpf: check failed: parse error\n");
                return -EINVAL;
        }
@@ -40,13 +40,13 @@ static bool bpf_mt(const struct sk_buff *skb, struct xt_action_param *par)
 {
        const struct xt_bpf_info *info = par->matchinfo;
 
-       return SK_RUN_FILTER(info->filter, skb);
+       return BPF_PROG_RUN(info->filter, skb);
 }
 
 static void bpf_mt_destroy(const struct xt_mtdtor_param *par)
 {
        const struct xt_bpf_info *info = par->matchinfo;
-       sk_unattached_filter_destroy(info->filter);
+       bpf_prog_destroy(info->filter);
 }
 
 static struct xt_match bpf_mt_reg __read_mostly = {
index 13f64df2c710663f2753df92632c5575bc2fe317..0e30d58149dad95874b9023abb5febf4cab97bfa 100644 (file)
@@ -30,7 +30,7 @@ struct cls_bpf_head {
 };
 
 struct cls_bpf_prog {
-       struct sk_filter *filter;
+       struct bpf_prog *filter;
        struct sock_filter *bpf_ops;
        struct tcf_exts exts;
        struct tcf_result res;
@@ -54,7 +54,7 @@ static int cls_bpf_classify(struct sk_buff *skb, const struct tcf_proto *tp,
        int ret;
 
        list_for_each_entry(prog, &head->plist, link) {
-               int filter_res = SK_RUN_FILTER(prog->filter, skb);
+               int filter_res = BPF_PROG_RUN(prog->filter, skb);
 
                if (filter_res == 0)
                        continue;
@@ -92,7 +92,7 @@ static void cls_bpf_delete_prog(struct tcf_proto *tp, struct cls_bpf_prog *prog)
        tcf_unbind_filter(tp, &prog->res);
        tcf_exts_destroy(tp, &prog->exts);
 
-       sk_unattached_filter_destroy(prog->filter);
+       bpf_prog_destroy(prog->filter);
 
        kfree(prog->bpf_ops);
        kfree(prog);
@@ -161,7 +161,7 @@ static int cls_bpf_modify_existing(struct net *net, struct tcf_proto *tp,
        struct sock_filter *bpf_ops, *bpf_old;
        struct tcf_exts exts;
        struct sock_fprog_kern tmp;
-       struct sk_filter *fp, *fp_old;
+       struct bpf_prog *fp, *fp_old;
        u16 bpf_size, bpf_len;
        u32 classid;
        int ret;
@@ -193,7 +193,7 @@ static int cls_bpf_modify_existing(struct net *net, struct tcf_proto *tp,
        tmp.len = bpf_len;
        tmp.filter = bpf_ops;
 
-       ret = sk_unattached_filter_create(&fp, &tmp);
+       ret = bpf_prog_create(&fp, &tmp);
        if (ret)
                goto errout_free;
 
@@ -211,7 +211,7 @@ static int cls_bpf_modify_existing(struct net *net, struct tcf_proto *tp,
        tcf_exts_change(tp, &prog->exts, &exts);
 
        if (fp_old)
-               sk_unattached_filter_destroy(fp_old);
+               bpf_prog_destroy(fp_old);
        if (bpf_old)
                kfree(bpf_old);