From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-8.2 required=3.0 tests=HEADER_FROM_DIFFERENT_DOMAINS, INCLUDES_PATCH,MAILING_LIST_MULTI,SIGNED_OFF_BY,SPF_HELO_NONE,SPF_PASS, URIBL_BLOCKED,USER_AGENT_SANE_1 autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id E49D2C432C0 for ; Thu, 21 Nov 2019 12:37:40 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.kernel.org (Postfix) with ESMTP id C1FF12075E for ; Thu, 21 Nov 2019 12:37:40 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726529AbfKUMhk (ORCPT ); Thu, 21 Nov 2019 07:37:40 -0500 Received: from m9784.mail.qiye.163.com ([220.181.97.84]:3403 "EHLO m9784.mail.qiye.163.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726358AbfKUMhk (ORCPT ); Thu, 21 Nov 2019 07:37:40 -0500 Received: from [192.168.1.4] (unknown [116.237.146.20]) by m9784.mail.qiye.163.com (Hmail) with ESMTPA id 8398F41D4D; Thu, 21 Nov 2019 20:37:35 +0800 (CST) Subject: Re: [PATCH nf-next v2 2/4] netfilter: nf_flow_table_offload: add indr block setup support From: wenxu Cc: netfilter-devel@vger.kernel.org, Paul Blakey References: <1574330056-5377-1-git-send-email-wenxu@ucloud.cn> <1574330056-5377-3-git-send-email-wenxu@ucloud.cn> Message-ID: Date: Thu, 21 Nov 2019 20:37:15 +0800 User-Agent: Mozilla/5.0 (Windows NT 6.1; WOW64; rv:60.0) Gecko/20100101 Thunderbird/60.9.1 MIME-Version: 1.0 In-Reply-To: <1574330056-5377-3-git-send-email-wenxu@ucloud.cn> Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 8bit X-HM-Spam-Status: e1kfGhgUHx5ZQUtXWQgYFAkeWUFZVkpVSEpJS0tLS0hNSUhCTUlZV1koWU FJQjdXWS1ZQUlXWQkOFx4IWUFZNTQpNjo3JCkuNz5ZBg++ X-HM-Sender-Digest: e1kMHhlZQR0aFwgeV1kSHx4VD1lBWUc6Mgg6HTo5MTgxHw8VSTEsTQ4d PylPCxBVSlVKTkxPSEhCQ05OQk1LVTMWGhIXVQweFQMOOw4YFxQOH1UYFUVZV1kSC1lBWUpKTVVJ SExVSk9NVUlLWVdZCAFZQU5MT0k3Bg++ X-HM-Tid: 0a6e8df7c5002086kuqy8398f41d4d To: unlisted-recipients:; (no To-header on input) Sender: netfilter-devel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: netfilter-devel@vger.kernel.org 在 2019/11/21 17:54, wenxu@ucloud.cn 写道: > From: wenxu > > Nf flow table support indr-block setup. It makes flow table offload vlan > and tunnel device. > > Signed-off-by: wenxu > --- > v2: no change > > net/netfilter/nf_flow_table_offload.c | 89 ++++++++++++++++++++++++++++++++++- > 1 file changed, 88 insertions(+), 1 deletion(-) > > diff --git a/net/netfilter/nf_flow_table_offload.c b/net/netfilter/nf_flow_table_offload.c > index 2d92043..653866f 100644 > --- a/net/netfilter/nf_flow_table_offload.c > +++ b/net/netfilter/nf_flow_table_offload.c > @@ -7,6 +7,7 @@ > #include > #include > #include > +#include > #include > #include > #include > @@ -834,6 +835,24 @@ static int nf_flow_table_offload_cmd(struct nf_flowtable *flowtable, > return nf_flow_table_block_setup(flowtable, &bo, cmd); > } > > +static int nf_flow_table_indr_offload_cmd(struct nf_flowtable *flowtable, > + struct net_device *dev, > + enum flow_block_command cmd) > +{ > + struct netlink_ext_ack extack = {}; > + struct flow_block_offload bo; > + > + nf_flow_table_block_offload_init(&bo, dev_net(dev), cmd, flowtable, > + &extack); > + > + flow_indr_block_call(dev, &bo, cmd); > + > + if (list_empty(&bo.cb_list)) > + return -EOPNOTSUPP; > + > + return nf_flow_table_block_setup(flowtable, &bo, cmd); > +} > + > int nf_flow_table_offload_setup(struct nf_flowtable *flowtable, > struct net_device *dev, > enum flow_block_command cmd) > @@ -846,16 +865,82 @@ int nf_flow_table_offload_setup(struct nf_flowtable *flowtable, > if (dev->netdev_ops->ndo_setup_tc) > err = nf_flow_table_offload_cmd(flowtable, dev, cmd); > else > - err = -EOPNOTSUPP; > + err = nf_flow_table_indr_offload_cmd(flowtable, dev, cmd); > > return err; > } > EXPORT_SYMBOL_GPL(nf_flow_table_offload_setup); > > +static struct nf_flowtable *__nf_flow_table_offload_get(struct net_device *dev) > +{ > + struct nf_flowtable *n_flowtable; > + struct nft_flowtable *flowtable; > + struct net *net = dev_net(dev); > + struct nft_table *table; > + struct nft_hook *hook; > + > + list_for_each_entry(table, &net->nft.tables, list) { > + list_for_each_entry(flowtable, &table->flowtables, list) { > + list_for_each_entry(hook, &flowtable->hook_list, list) { > + if (hook->ops.dev != dev) > + continue; > + > + n_flowtable = &flowtable->data; > + return n_flowtable; > + } > + } > + } > + > + return NULL; > +} > + > +static void nf_flow_table_indr_block_ing_cmd(struct net_device *dev, > + struct nf_flowtable *flowtable, > + flow_indr_block_bind_cb_t *cb, > + void *cb_priv, > + enum flow_block_command cmd) > +{ > + struct netlink_ext_ack extack = {}; > + struct flow_block_offload bo; > + > + if (!flowtable) > + return; > + > + nf_flow_table_block_offload_init(&bo, dev_net(dev), cmd, flowtable, > + &extack); > + > + cb(dev, cb_priv, TC_SETUP_BLOCK, &bo); > + > + nf_flow_table_block_setup(flowtable, &bo, cmd); > +} > + > +static void nf_flow_table_indr_block_cb(struct net_device *dev, > + flow_indr_block_bind_cb_t *cb, > + void *cb_priv, > + enum flow_block_command cmd) > +{ > + struct net *net = dev_net(dev); > + struct nf_flowtable *flowtable; > + > + mutex_lock(&net->nft.commit_mutex); > + flowtable = __nf_flow_table_offload_get(dev); > + if (flowtable) > + nf_flow_table_indr_block_ing_cmd(dev, flowtable, cb, cb_priv, > + cmd); > + mutex_unlock(&net->nft.commit_mutex); > +} > + > +static struct flow_indr_block_ing_entry block_ing_entry = { > + .cb = nf_flow_table_indr_block_cb, > + .list = LIST_HEAD_INIT(block_ing_entry.list), > +}; > + > int nf_flow_table_offload_init(void) > { > INIT_WORK(&nf_flow_offload_work, flow_offload_work_handler); > > + flow_indr_add_block_ing_cb(&block_ing_entry); > + > return 0; > } > > @@ -864,6 +949,8 @@ void nf_flow_table_offload_exit(void) > struct flow_offload_work *offload, *next; > LIST_HEAD(offload_pending_list); > > + flow_indr_del_block_ing_cb(&block_ing_entry); > + > cancel_work_sync(&nf_flow_offload_work); > > list_for_each_entry_safe(offload, next, &offload_pending_list, list) {