public inbox for netdev@vger.kernel.org
 help / color / mirror / Atom feed
From: Martin KaFai Lau <kafai@fb.com>
To: Stanislav Fomichev <sdf@google.com>
Cc: netdev@vger.kernel.org, bpf@vger.kernel.org, ast@kernel.org,
	daniel@iogearbox.net, andrii@kernel.org
Subject: Re: [PATCH bpf-next v10 11/11] selftests/bpf: lsm_cgroup functional test
Date: Thu, 23 Jun 2022 15:36:31 -0700	[thread overview]
Message-ID: <20220623223631.uu3uakauseipsx5a@kafai-mbp> (raw)
In-Reply-To: <20220622160346.967594-12-sdf@google.com>

On Wed, Jun 22, 2022 at 09:03:46AM -0700, Stanislav Fomichev wrote:
> diff --git a/tools/testing/selftests/bpf/prog_tests/lsm_cgroup.c b/tools/testing/selftests/bpf/prog_tests/lsm_cgroup.c
> new file mode 100644
> index 000000000000..a96057ec7dd4
> --- /dev/null
> +++ b/tools/testing/selftests/bpf/prog_tests/lsm_cgroup.c
> @@ -0,0 +1,277 @@
> +// SPDX-License-Identifier: GPL-2.0
> +
> +#include <sys/types.h>
> +#include <sys/socket.h>
> +#include <test_progs.h>
> +#include <bpf/btf.h>
> +
> +#include "lsm_cgroup.skel.h"
> +#include "cgroup_helpers.h"
> +#include "network_helpers.h"
> +
> +static __u32 query_prog_cnt(int cgroup_fd, const char *attach_func)
> +{
> +	LIBBPF_OPTS(bpf_prog_query_opts, p);
> +	static struct btf *btf;
> +	int cnt = 0;
> +	int i;
> +
> +	ASSERT_OK(bpf_prog_query_opts(cgroup_fd, BPF_LSM_CGROUP, &p), "prog_query");
> +
> +	if (!attach_func)
> +		return p.prog_cnt;
> +
> +	/* When attach_func is provided, count the number of progs that
> +	 * attach to the given symbol.
> +	 */
> +
> +	if (!btf)
> +		btf = btf__load_vmlinux_btf();
This needs a btf__free().  Probably at the end of test_lsm_cgroup().

> +	if (!ASSERT_OK(libbpf_get_error(btf), "btf_vmlinux"))
> +		return -1;
> +
> +	p.prog_ids = malloc(sizeof(u32) * p.prog_cnt);
> +	p.prog_attach_flags = malloc(sizeof(u32) * p.prog_cnt);
and these mallocs too ?

> +	ASSERT_OK(bpf_prog_query_opts(cgroup_fd, BPF_LSM_CGROUP, &p), "prog_query");
> +
> +	for (i = 0; i < p.prog_cnt; i++) {
> +		struct bpf_prog_info info = {};
> +		__u32 info_len = sizeof(info);
> +		int fd;
> +
> +		fd = bpf_prog_get_fd_by_id(p.prog_ids[i]);
> +		ASSERT_GE(fd, 0, "prog_get_fd_by_id");
> +		ASSERT_OK(bpf_obj_get_info_by_fd(fd, &info, &info_len), "prog_info_by_fd");
> +		close(fd);
> +
> +		if (info.attach_btf_id ==
> +		    btf__find_by_name_kind(btf, attach_func, BTF_KIND_FUNC))
> +			cnt++;
> +	}
> +
> +	return cnt;
> +}
> +
> +static void test_lsm_cgroup_functional(void)
> +{
> +	DECLARE_LIBBPF_OPTS(bpf_prog_attach_opts, attach_opts);
> +	DECLARE_LIBBPF_OPTS(bpf_link_update_opts, update_opts);
> +	int cgroup_fd, cgroup_fd2, err, fd, prio;
> +	int listen_fd, client_fd, accepted_fd;
> +	struct lsm_cgroup *skel = NULL;
> +	int post_create_prog_fd2 = -1;
> +	int post_create_prog_fd = -1;
> +	int bind_link_fd2 = -1;
> +	int bind_prog_fd2 = -1;
> +	int alloc_prog_fd = -1;
> +	int bind_prog_fd = -1;
> +	int bind_link_fd = -1;
> +	int clone_prog_fd = -1;
> +	socklen_t socklen;
> +
> +	cgroup_fd = test__join_cgroup("/sock_policy");
> +	if (!ASSERT_GE(cgroup_fd, 0, "join_cgroup"))
> +		goto close_skel;
> +
> +	cgroup_fd2 = create_and_get_cgroup("/sock_policy2");
> +	if (!ASSERT_GE(cgroup_fd2, 0, "create second cgroup"))
cgroup_fd needs to close in this error case

> +		goto close_skel;

A valid cgroup_fd2 should be closed later also.

> +
> +	skel = lsm_cgroup__open_and_load();
> +	if (!ASSERT_OK_PTR(skel, "open_and_load"))
> +		goto close_cgroup;
> +
> +	post_create_prog_fd = bpf_program__fd(skel->progs.socket_post_create);
> +	post_create_prog_fd2 = bpf_program__fd(skel->progs.socket_post_create2);
> +	bind_prog_fd = bpf_program__fd(skel->progs.socket_bind);
> +	bind_prog_fd2 = bpf_program__fd(skel->progs.socket_bind2);
> +	alloc_prog_fd = bpf_program__fd(skel->progs.socket_alloc);
> +	clone_prog_fd = bpf_program__fd(skel->progs.socket_clone);
> +
> +	ASSERT_EQ(query_prog_cnt(cgroup_fd, "bpf_lsm_sk_alloc_security"), 0, "prog count");
> +	ASSERT_EQ(query_prog_cnt(cgroup_fd, NULL), 0, "total prog count");
> +	err = bpf_prog_attach(alloc_prog_fd, cgroup_fd, BPF_LSM_CGROUP, 0);
> +	if (!ASSERT_OK(err, "attach alloc_prog_fd"))
> +		goto detach_cgroup;
> +	ASSERT_EQ(query_prog_cnt(cgroup_fd, "bpf_lsm_sk_alloc_security"), 1, "prog count");
> +	ASSERT_EQ(query_prog_cnt(cgroup_fd, NULL), 1, "total prog count");
> +
> +	ASSERT_EQ(query_prog_cnt(cgroup_fd, "bpf_lsm_inet_csk_clone"), 0, "prog count");
> +	err = bpf_prog_attach(clone_prog_fd, cgroup_fd, BPF_LSM_CGROUP, 0);
> +	if (!ASSERT_OK(err, "attach clone_prog_fd"))
> +		goto detach_cgroup;
> +	ASSERT_EQ(query_prog_cnt(cgroup_fd, "bpf_lsm_inet_csk_clone"), 1, "prog count");
> +	ASSERT_EQ(query_prog_cnt(cgroup_fd, NULL), 2, "total prog count");
> +
> +	/* Make sure replacing works. */
> +
> +	ASSERT_EQ(query_prog_cnt(cgroup_fd, "bpf_lsm_socket_post_create"), 0, "prog count");
> +	err = bpf_prog_attach(post_create_prog_fd, cgroup_fd,
> +			      BPF_LSM_CGROUP, 0);
> +	if (!ASSERT_OK(err, "attach post_create_prog_fd"))
> +		goto close_cgroup;
> +	ASSERT_EQ(query_prog_cnt(cgroup_fd, "bpf_lsm_socket_post_create"), 1, "prog count");
> +	ASSERT_EQ(query_prog_cnt(cgroup_fd, NULL), 3, "total prog count");
> +
> +	attach_opts.replace_prog_fd = post_create_prog_fd;
> +	err = bpf_prog_attach_opts(post_create_prog_fd2, cgroup_fd,
> +				   BPF_LSM_CGROUP, &attach_opts);
> +	if (!ASSERT_OK(err, "prog replace post_create_prog_fd"))
> +		goto detach_cgroup;
> +	ASSERT_EQ(query_prog_cnt(cgroup_fd, "bpf_lsm_socket_post_create"), 1, "prog count");
> +	ASSERT_EQ(query_prog_cnt(cgroup_fd, NULL), 3, "total prog count");
> +
> +	/* Try the same attach/replace via link API. */
> +
> +	ASSERT_EQ(query_prog_cnt(cgroup_fd, "bpf_lsm_socket_bind"), 0, "prog count");
> +	bind_link_fd = bpf_link_create(bind_prog_fd, cgroup_fd,
> +				       BPF_LSM_CGROUP, NULL);
> +	if (!ASSERT_GE(bind_link_fd, 0, "link create bind_prog_fd"))
> +		goto detach_cgroup;
> +	ASSERT_EQ(query_prog_cnt(cgroup_fd, "bpf_lsm_socket_bind"), 1, "prog count");
> +	ASSERT_EQ(query_prog_cnt(cgroup_fd, NULL), 4, "total prog count");
> +
> +	update_opts.old_prog_fd = bind_prog_fd;
> +	update_opts.flags = BPF_F_REPLACE;
> +
> +	err = bpf_link_update(bind_link_fd, bind_prog_fd2, &update_opts);
> +	if (!ASSERT_OK(err, "link update bind_prog_fd"))
> +		goto detach_cgroup;
> +	ASSERT_EQ(query_prog_cnt(cgroup_fd, "bpf_lsm_socket_bind"), 1, "prog count");
> +	ASSERT_EQ(query_prog_cnt(cgroup_fd, NULL), 4, "total prog count");
> +
> +	/* Attach another instance of bind program to another cgroup.
> +	 * This should trigger the reuse of the trampoline shim (two
> +	 * programs attaching to the same btf_id).
> +	 */
> +
> +	ASSERT_EQ(query_prog_cnt(cgroup_fd, "bpf_lsm_socket_bind"), 1, "prog count");
> +	ASSERT_EQ(query_prog_cnt(cgroup_fd2, "bpf_lsm_socket_bind"), 0, "prog count");
> +	bind_link_fd2 = bpf_link_create(bind_prog_fd2, cgroup_fd2,
> +					BPF_LSM_CGROUP, NULL);
> +	if (!ASSERT_GE(bind_link_fd2, 0, "link create bind_prog_fd2"))
> +		goto detach_cgroup;
> +	ASSERT_EQ(query_prog_cnt(cgroup_fd2, "bpf_lsm_socket_bind"), 1, "prog count");
> +	ASSERT_EQ(query_prog_cnt(cgroup_fd, NULL), 4, "total prog count");
> +	ASSERT_EQ(query_prog_cnt(cgroup_fd2, NULL), 1, "total prog count");
> +
> +	/* AF_UNIX is prohibited. */
> +
> +	fd = socket(AF_UNIX, SOCK_STREAM, 0);
> +	ASSERT_LT(fd, 0, "socket(AF_UNIX)");
close on fd >=0 case.

> +
> +	/* AF_INET6 gets default policy (sk_priority). */
> +
> +	fd = socket(AF_INET6, SOCK_STREAM, 0);
> +	if (!ASSERT_GE(fd, 0, "socket(SOCK_STREAM)"))
> +		goto detach_cgroup;
> +
> +	prio = 0;
> +	socklen = sizeof(prio);
> +	ASSERT_GE(getsockopt(fd, SOL_SOCKET, SO_PRIORITY, &prio, &socklen), 0,
> +		  "getsockopt");
> +	ASSERT_EQ(prio, 123, "sk_priority");
> +
> +	close(fd);
> +
> +	/* TX-only AF_PACKET is allowed. */
> +
> +	ASSERT_LT(socket(AF_PACKET, SOCK_RAW, htons(ETH_P_ALL)), 0,
> +		  "socket(AF_PACKET, ..., ETH_P_ALL)");
> +
> +	fd = socket(AF_PACKET, SOCK_RAW, 0);
> +	ASSERT_GE(fd, 0, "socket(AF_PACKET, ..., 0)");
> +
> +	/* TX-only AF_PACKET can not be rebound. */
> +
> +	struct sockaddr_ll sa = {
> +		.sll_family = AF_PACKET,
> +		.sll_protocol = htons(ETH_P_ALL),
> +	};
> +	ASSERT_LT(bind(fd, (struct sockaddr *)&sa, sizeof(sa)), 0,
> +		  "bind(ETH_P_ALL)");
> +
> +	close(fd);
> +
> +	/* Trigger passive open. */
> +
> +	listen_fd = start_server(AF_INET6, SOCK_STREAM, "::1", 0, 0);
> +	ASSERT_GE(listen_fd, 0, "start_server");
> +	client_fd = connect_to_fd(listen_fd, 0);
> +	ASSERT_GE(client_fd, 0, "connect_to_fd");
> +	accepted_fd = accept(listen_fd, NULL, NULL);
> +	ASSERT_GE(accepted_fd, 0, "accept");
This listen/client/accept_fd needs a close.

> +
> +	prio = 0;
> +	socklen = sizeof(prio);
> +	ASSERT_GE(getsockopt(accepted_fd, SOL_SOCKET, SO_PRIORITY, &prio, &socklen), 0,
> +		  "getsockopt");
> +	ASSERT_EQ(prio, 234, "sk_priority");
> +
> +	/* These are replaced and never called. */
> +	ASSERT_EQ(skel->bss->called_socket_post_create, 0, "called_create");
> +	ASSERT_EQ(skel->bss->called_socket_bind, 0, "called_bind");
> +
> +	/* AF_INET6+SOCK_STREAM
> +	 * AF_PACKET+SOCK_RAW
> +	 * listen_fd
> +	 * client_fd
> +	 * accepted_fd
> +	 */
> +	ASSERT_EQ(skel->bss->called_socket_post_create2, 5, "called_create2");
> +
> +	/* start_server
> +	 * bind(ETH_P_ALL)
> +	 */
> +	ASSERT_EQ(skel->bss->called_socket_bind2, 2, "called_bind2");
> +	/* Single accept(). */
> +	ASSERT_EQ(skel->bss->called_socket_clone, 1, "called_clone");
> +
> +	/* AF_UNIX+SOCK_STREAM (failed)
> +	 * AF_INET6+SOCK_STREAM
> +	 * AF_PACKET+SOCK_RAW (failed)
> +	 * AF_PACKET+SOCK_RAW
> +	 * listen_fd
> +	 * client_fd
> +	 * accepted_fd
> +	 */
> +	ASSERT_EQ(skel->bss->called_socket_alloc, 7, "called_alloc");
> +
> +	/* Make sure other cgroup doesn't trigger the programs. */
> +
> +	if (!ASSERT_OK(join_cgroup(""), "join root cgroup"))
In my qemu setup, I am hitting this:
(cgroup_helpers.c:166: errno: Device or resource busy) Joining Cgroup
test_lsm_cgroup_functional:FAIL:join root cgroup unexpected error: 1 (errno 16)

A quick tracing leads to the non zero dst_cgrp->subtree_control
in cgroup_migrate_vet_dst().

This is likely due to the enable_all_controllers() in cgroup_helpers.c
that enabled the 'memory' controller in my setup.  Avoid this function
worked around the issue.  Do you know how to solve this ?

> +		goto detach_cgroup;
> +
> +	fd = socket(AF_INET6, SOCK_STREAM, 0);
> +	if (!ASSERT_GE(fd, 0, "socket(SOCK_STREAM)"))
> +		goto detach_cgroup;
> +
> +	prio = 0;
> +	socklen = sizeof(prio);
> +	ASSERT_GE(getsockopt(fd, SOL_SOCKET, SO_PRIORITY, &prio, &socklen), 0,
> +		  "getsockopt");
> +	ASSERT_EQ(prio, 0, "sk_priority");
> +
> +	close(fd);
> +
> +detach_cgroup:
> +	ASSERT_GE(bpf_prog_detach2(post_create_prog_fd2, cgroup_fd,
> +				   BPF_LSM_CGROUP), 0, "detach_create");
> +	close(bind_link_fd);
> +	/* Don't close bind_link_fd2, exercise cgroup release cleanup. */
> +	ASSERT_GE(bpf_prog_detach2(alloc_prog_fd, cgroup_fd,
> +				   BPF_LSM_CGROUP), 0, "detach_alloc");
> +	ASSERT_GE(bpf_prog_detach2(clone_prog_fd, cgroup_fd,
> +				   BPF_LSM_CGROUP), 0, "detach_clone");
> +
> +close_cgroup:
> +	close(cgroup_fd);
> +close_skel:
> +	lsm_cgroup__destroy(skel);
> +}
> +
> +void test_lsm_cgroup(void)
> +{
> +	if (test__start_subtest("functional"))
> +		test_lsm_cgroup_functional();
> +}

  reply	other threads:[~2022-06-23 22:37 UTC|newest]

Thread overview: 21+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2022-06-22 16:03 [PATCH bpf-next v10 00/11] bpf: cgroup_sock lsm flavor Stanislav Fomichev
2022-06-22 16:03 ` [PATCH bpf-next v10 01/11] bpf: add bpf_func_t and trampoline helpers Stanislav Fomichev
2022-06-22 16:03 ` [PATCH bpf-next v10 02/11] bpf: convert cgroup_bpf.progs to hlist Stanislav Fomichev
2022-06-22 16:03 ` [PATCH bpf-next v10 03/11] bpf: per-cgroup lsm flavor Stanislav Fomichev
2022-06-23 22:40   ` Martin KaFai Lau
2022-06-22 16:03 ` [PATCH bpf-next v10 04/11] bpf: minimize number of allocated lsm slots per program Stanislav Fomichev
2022-06-23 22:41   ` Martin KaFai Lau
2022-06-22 16:03 ` [PATCH bpf-next v10 05/11] bpf: implement BPF_PROG_QUERY for BPF_LSM_CGROUP Stanislav Fomichev
2022-06-22 16:03 ` [PATCH bpf-next v10 06/11] bpf: expose bpf_{g,s}etsockopt to lsm cgroup Stanislav Fomichev
2022-06-23 22:43   ` Martin KaFai Lau
2022-06-22 16:03 ` [PATCH bpf-next v10 07/11] tools/bpf: Sync btf_ids.h to tools Stanislav Fomichev
2022-06-23 22:54   ` Martin KaFai Lau
2022-06-22 16:03 ` [PATCH bpf-next v10 08/11] libbpf: add lsm_cgoup_sock type Stanislav Fomichev
2022-06-23 22:54   ` Martin KaFai Lau
2022-06-22 16:03 ` [PATCH bpf-next v10 09/11] libbpf: implement bpf_prog_query_opts Stanislav Fomichev
2022-06-23 22:58   ` Martin KaFai Lau
2022-06-22 16:03 ` [PATCH bpf-next v10 10/11] bpftool: implement cgroup tree for BPF_LSM_CGROUP Stanislav Fomichev
2022-06-23 22:58   ` Martin KaFai Lau
2022-06-22 16:03 ` [PATCH bpf-next v10 11/11] selftests/bpf: lsm_cgroup functional test Stanislav Fomichev
2022-06-23 22:36   ` Martin KaFai Lau [this message]
2022-06-23 23:48     ` Stanislav Fomichev

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20220623223631.uu3uakauseipsx5a@kafai-mbp \
    --to=kafai@fb.com \
    --cc=andrii@kernel.org \
    --cc=ast@kernel.org \
    --cc=bpf@vger.kernel.org \
    --cc=daniel@iogearbox.net \
    --cc=netdev@vger.kernel.org \
    --cc=sdf@google.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox