* [PATCH v2 0/2] ss: pretty-printing BPF socket-local storage
@ 2023-12-08 14:57 Quentin Deslandes
2023-12-08 14:57 ` [PATCH v2 1/2] ss: add support for " Quentin Deslandes
2023-12-08 14:57 ` [PATCH v2 2/2] ss: pretty-print " Quentin Deslandes
0 siblings, 2 replies; 6+ messages in thread
From: Quentin Deslandes @ 2023-12-08 14:57 UTC (permalink / raw)
To: netdev; +Cc: David Ahern, Martin KaFai Lau, Quentin Deslandes, kernel-team
BPF allows programs to store socket-specific data using
BPF_MAP_TYPE_SK_STORAGE maps. The data is attached to the socket itself,
and Martin added INET_DIAG_REQ_SK_BPF_STORAGES, so it can be fetched
using the INET_DIAG mechanism.
Currently, ss doesn't request the socket-local data, this patch aims to
fix this.
The first patch requests the socket-local data for the requested map ID
(--bpf-map-id=) or all the maps (--bpf-maps). It then prints the map_id
in a dedicated column.
Patch #2 uses libbpf and BTF to pretty print the map's content, like
`bpftool map dump` would do.
While I think it makes sense for ss to provide the socket-local storage
content for the sockets, it's difficult to conciliate the column-based
output of ss and having readable socket-local data. Hence, the
socket-local data is printed in a readable fashion over multiple lines
under its socket statistics, independently of the column-based approach.
Here is an example of ss' output with --bpf-maps:
[...]
ESTAB 340116 0 [...]
map_id: 114 [
(struct my_sk_storage){
.field_hh = (char)3,
(union){
.a = (int)17,
.b = (int)17,
},
}
]
Changes from v1:
* Remove the first patch from the series (fix) and submit it separately.
* Remove double allocation of struct rtattr.
* Close BPF map FDs on exit.
* If bpf_map_get_fd_by_id() fails with ENOENT, print an error message
and continue to the next map ID.
* Fix typo in new command line option documentation.
* Only use bpf_map_info.btf_value_type_id and ignore
bpf_map_info.btf_vmlinux_value_type_id (unused for socket-local storage).
* Use btf_dump__dump_type_data() instead of manually using BTF to
pretty-print socket-local storage data. This change alone divides the size
of the patch series by 2.
Quentin Deslandes (2):
ss: add support for BPF socket-local storage
ss: pretty-print BPF socket-local storage
misc/ss.c | 433 +++++++++++++++++++++++++++++++++++++++++++++++++++++-
1 file changed, 425 insertions(+), 8 deletions(-)
--
2.43.0
^ permalink raw reply [flat|nested] 6+ messages in thread* [PATCH v2 1/2] ss: add support for BPF socket-local storage 2023-12-08 14:57 [PATCH v2 0/2] ss: pretty-printing BPF socket-local storage Quentin Deslandes @ 2023-12-08 14:57 ` Quentin Deslandes 2023-12-12 22:49 ` Martin KaFai Lau 2023-12-08 14:57 ` [PATCH v2 2/2] ss: pretty-print " Quentin Deslandes 1 sibling, 1 reply; 6+ messages in thread From: Quentin Deslandes @ 2023-12-08 14:57 UTC (permalink / raw) To: netdev; +Cc: David Ahern, Martin KaFai Lau, Quentin Deslandes, kernel-team While sock_diag is able to return BPF socket-local storage in response to INET_DIAG_REQ_SK_BPF_STORAGES requests, ss doesn't request it. This change introduces the --bpf-maps and --bpf-map-id= options to request BPF socket-local storage for all SK_STORAGE maps, or only specific ones. The bigger part of this change will check the requested map IDs and ensure they are valid. A new column has been added named "Socket storage" to print a list of map ID a given socket has data defined for. This column is disabled unless --bpf-maps or --bpf-map-id= is used. Signed-off-by: Quentin Deslandes <qde@naccy.de> Co-authored-by: Martin KaFai Lau <martin.lau@kernel.org> --- misc/ss.c | 290 +++++++++++++++++++++++++++++++++++++++++++++++++++++- 1 file changed, 287 insertions(+), 3 deletions(-) diff --git a/misc/ss.c b/misc/ss.c index 01d29dc5..f979b61d 100644 --- a/misc/ss.c +++ b/misc/ss.c @@ -51,6 +51,11 @@ #include <linux/tls.h> #include <linux/mptcp.h> +#ifdef HAVE_LIBBPF +#include <bpf/bpf.h> +#include <bpf/libbpf.h> +#endif + #if HAVE_RPC #include <rpc/rpc.h> #include <rpc/xdr.h> @@ -101,6 +106,7 @@ enum col_id { COL_RADDR, COL_RSERV, COL_PROC, + COL_SKSTOR, COL_EXT, COL_MAX }; @@ -130,6 +136,7 @@ static struct column columns[] = { { ALIGN_RIGHT, "Peer Address:", " ", 0, 0, 0 }, { ALIGN_LEFT, "Port", "", 0, 0, 0 }, { ALIGN_LEFT, "Process", "", 0, 0, 0 }, + { ALIGN_LEFT, "Socket storage", "", 1, 0, 0 }, { ALIGN_LEFT, "", "", 0, 0, 0 }, }; @@ -3374,6 +3381,238 @@ static void parse_diag_msg(struct nlmsghdr *nlh, struct sockstat *s) memcpy(s->remote.data, r->id.idiag_dst, s->local.bytelen); } +#ifdef HAVE_LIBBPF + +#define MAX_NR_BPF_MAP_ID_OPTS 32 + +struct btf; + +static struct bpf_map_opts { + unsigned int nr_maps; + struct bpf_sk_storage_map_info { + unsigned int id; + int fd; + } maps[MAX_NR_BPF_MAP_ID_OPTS]; + bool show_all; + struct btf *kernel_btf; +} bpf_map_opts; + +static void bpf_map_opts_mixed_error(void) +{ + fprintf(stderr, + "ss: --bpf-maps and --bpf-map-id cannot be used together\n"); +} + +static int bpf_map_opts_add_all(void) +{ + unsigned int i; + unsigned int fd; + uint32_t id = 0; + int r; + + if (bpf_map_opts.nr_maps) { + bpf_map_opts_mixed_error(); + return -1; + } + + while (1) { + struct bpf_map_info info = {}; + uint32_t len = sizeof(info); + + r = bpf_map_get_next_id(id, &id); + if (r) { + if (errno == ENOENT) + break; + + fprintf(stderr, "ss: failed to fetch BPF map ID\n"); + goto err; + } + + fd = bpf_map_get_fd_by_id(id); + if (fd == -1) { + if (errno == -ENOENT) { + fprintf(stderr, "ss: missing BPF map ID %u, skipping", + id); + continue; + } + + fprintf(stderr, "ss: cannot get fd for BPF map ID %u%s\n", + id, errno == EPERM ? + ": missing root permissions, CAP_BPF, or CAP_SYS_ADMIN" : ""); + goto err; + } + + r = bpf_obj_get_info_by_fd(fd, &info, &len); + if (r) { + fprintf(stderr, "ss: failed to get info for BPF map ID %u\n", + id); + close(fd); + goto err; + } + + if (info.type != BPF_MAP_TYPE_SK_STORAGE) { + close(fd); + continue; + } + + if (bpf_map_opts.nr_maps == MAX_NR_BPF_MAP_ID_OPTS) { + fprintf(stderr, "ss: too many (> %u) BPF socket-local storage maps found, skipping map ID %u\n", + MAX_NR_BPF_MAP_ID_OPTS, id); + close(fd); + continue; + } + + bpf_map_opts.maps[bpf_map_opts.nr_maps].id = id; + bpf_map_opts.maps[bpf_map_opts.nr_maps++].fd = fd; + } + + bpf_map_opts.show_all = true; + + return 0; + +err: + for (i = 0; i < bpf_map_opts.nr_maps; ++i) + close(bpf_map_opts.maps[i].fd); + + return -1; +} + +static int bpf_map_opts_add_id(const char *optarg) +{ + struct bpf_map_info info = {}; + uint32_t len = sizeof(info); + size_t optarg_len; + unsigned long id; + unsigned int i; + char *end; + int fd; + int r; + + if (bpf_map_opts.show_all) { + bpf_map_opts_mixed_error(); + return -1; + } + + optarg_len = strlen(optarg); + id = strtoul(optarg, &end, 0); + if (end != optarg + optarg_len || id == 0 || id >= UINT32_MAX) { + fprintf(stderr, "ss: invalid BPF map ID %s\n", optarg); + return -1; + } + + for (i = 0; i < bpf_map_opts.nr_maps; i++) { + if (bpf_map_opts.maps[i].id == id) + return 0; + } + + if (bpf_map_opts.nr_maps == MAX_NR_BPF_MAP_ID_OPTS) { + fprintf(stderr, "ss: too many (> %u) BPF socket-local storage maps found, skipping map ID %lu\n", + MAX_NR_BPF_MAP_ID_OPTS, id); + return 0; + } + + fd = bpf_map_get_fd_by_id(id); + if (fd == -1) { + if (errno == -ENOENT) { + fprintf(stderr, "ss: missing BPF map ID %lu, skipping", + id); + return 0; + } + + fprintf(stderr, "ss: cannot get fd for BPF map ID %lu%s\n", + id, errno == EPERM ? + ": missing root permissions, CAP_BPF, or CAP_SYS_ADMIN" : ""); + return -1; + } + + r = bpf_obj_get_info_by_fd(fd, &info, &len); + if (r) { + fprintf(stderr, "ss: failed to get info for BPF map ID %lu\n", id); + close(fd); + return -1; + } + + if (info.type != BPF_MAP_TYPE_SK_STORAGE) { + fprintf(stderr, "ss: BPF map with ID %s has type '%s', expecting 'sk_storage'\n", + optarg, libbpf_bpf_map_type_str(info.type)); + close(fd); + return -1; + } + + bpf_map_opts.maps[bpf_map_opts.nr_maps].id = id; + bpf_map_opts.maps[bpf_map_opts.nr_maps++].fd = fd; + + return 0; +} + +static void bpf_map_opts_destroy(void) +{ + int i; + + for (i = 0; i < bpf_map_opts.nr_maps; ++i) + close(bpf_map_opts.maps[i].fd); +} + +static inline bool bpf_map_opts_is_enabled(void) +{ + return bpf_map_opts.nr_maps; +} + +static struct rtattr *bpf_map_opts_alloc_rta(void) +{ + size_t total_size = RTA_LENGTH(RTA_LENGTH(sizeof(int)) * bpf_map_opts.nr_maps); + struct rtattr *stgs_rta, *fd_rta; + unsigned int i; + void *buf; + + buf = malloc(total_size); + if (!buf) + return NULL; + + stgs_rta = buf; + stgs_rta->rta_type = INET_DIAG_REQ_SK_BPF_STORAGES | NLA_F_NESTED; + stgs_rta->rta_len = total_size; + + buf = RTA_DATA(stgs_rta); + for (i = 0; i < bpf_map_opts.nr_maps; i++) { + int *fd; + + fd_rta = buf; + fd_rta->rta_type = SK_DIAG_BPF_STORAGE_REQ_MAP_FD; + fd_rta->rta_len = RTA_LENGTH(sizeof(int)); + + fd = RTA_DATA(fd_rta); + *fd = bpf_map_opts.maps[i].fd; + + buf += fd_rta->rta_len; + } + + return stgs_rta; +} + +static void show_sk_bpf_storages(struct rtattr *bpf_stgs) +{ + struct rtattr *tb[SK_DIAG_BPF_STORAGE_MAX + 1], *bpf_stg; + unsigned int rem; + + for (bpf_stg = RTA_DATA(bpf_stgs), rem = RTA_PAYLOAD(bpf_stgs); + RTA_OK(bpf_stg, rem); bpf_stg = RTA_NEXT(bpf_stg, rem)) { + + if ((bpf_stg->rta_type & NLA_TYPE_MASK) != SK_DIAG_BPF_STORAGE) + continue; + + parse_rtattr_nested(tb, SK_DIAG_BPF_STORAGE_MAX, + (struct rtattr *)bpf_stg); + + if (tb[SK_DIAG_BPF_STORAGE_MAP_ID]) { + out("map_id:%u", + rta_getattr_u32(tb[SK_DIAG_BPF_STORAGE_MAP_ID])); + } + } +} + +#endif + static int inet_show_sock(struct nlmsghdr *nlh, struct sockstat *s) { @@ -3381,8 +3620,8 @@ static int inet_show_sock(struct nlmsghdr *nlh, struct inet_diag_msg *r = NLMSG_DATA(nlh); unsigned char v6only = 0; - parse_rtattr(tb, INET_DIAG_MAX, (struct rtattr *)(r+1), - nlh->nlmsg_len - NLMSG_LENGTH(sizeof(*r))); + parse_rtattr_flags(tb, INET_DIAG_MAX, (struct rtattr *)(r+1), + nlh->nlmsg_len - NLMSG_LENGTH(sizeof(*r)), NLA_F_NESTED); if (tb[INET_DIAG_PROTOCOL]) s->type = rta_getattr_u8(tb[INET_DIAG_PROTOCOL]); @@ -3479,6 +3718,11 @@ static int inet_show_sock(struct nlmsghdr *nlh, } sctp_ino = s->ino; + if (tb[INET_DIAG_SK_BPF_STORAGES]) { + field_set(COL_SKSTOR); + show_sk_bpf_storages(tb[INET_DIAG_SK_BPF_STORAGES]); + } + return 0; } @@ -3560,13 +3804,14 @@ static int sockdiag_send(int family, int fd, int protocol, struct filter *f) { struct sockaddr_nl nladdr = { .nl_family = AF_NETLINK }; DIAG_REQUEST(req, struct inet_diag_req_v2 r); + struct rtattr *bpf_stgs_rta = NULL; char *bc = NULL; int bclen; __u32 proto; struct msghdr msg; struct rtattr rta_bc; struct rtattr rta_proto; - struct iovec iov[5]; + struct iovec iov[6]; int iovlen = 1; if (family == PF_UNSPEC) @@ -3619,6 +3864,17 @@ static int sockdiag_send(int family, int fd, int protocol, struct filter *f) iovlen += 2; } + if (bpf_map_opts_is_enabled()) { + bpf_stgs_rta = bpf_map_opts_alloc_rta(); + if (!bpf_stgs_rta) { + fprintf(stderr, "ss: cannot alloc request for --bpf-map\n"); + return -1; + } + + iov[iovlen++] = (struct iovec){ bpf_stgs_rta, bpf_stgs_rta->rta_len }; + req.nlh.nlmsg_len += bpf_stgs_rta->rta_len; + } + msg = (struct msghdr) { .msg_name = (void *)&nladdr, .msg_namelen = sizeof(nladdr), @@ -3627,10 +3883,13 @@ static int sockdiag_send(int family, int fd, int protocol, struct filter *f) }; if (sendmsg(fd, &msg, 0) < 0) { + free(bpf_stgs_rta); close(fd); return -1; } + free(bpf_stgs_rta); + return 0; } @@ -5350,6 +5609,10 @@ static void _usage(FILE *dest) " --tos show tos and priority information\n" " --cgroup show cgroup information\n" " -b, --bpf show bpf filter socket information\n" +#ifdef HAVE_LIBBPF +" --bpf-maps show all BPF socket-local storage maps\n" +" --bpf-map-id=MAP-ID show a BPF socket-local storage map\n" +#endif " -E, --events continually display sockets as they are destroyed\n" " -Z, --context display task SELinux security contexts\n" " -z, --contexts display task and socket SELinux security contexts\n" @@ -5466,6 +5729,9 @@ static int scan_state(const char *state) #define OPT_INET_SOCKOPT 262 +#define OPT_BPF_MAPS 263 +#define OPT_BPF_MAP_ID 264 + static const struct option long_opts[] = { { "numeric", 0, 0, 'n' }, { "resolve", 0, 0, 'r' }, @@ -5510,6 +5776,10 @@ static const struct option long_opts[] = { { "mptcp", 0, 0, 'M' }, { "oneline", 0, 0, 'O' }, { "inet-sockopt", 0, 0, OPT_INET_SOCKOPT }, +#ifdef HAVE_LIBBPF + { "bpf-maps", 0, 0, OPT_BPF_MAPS}, + { "bpf-map-id", 1, 0, OPT_BPF_MAP_ID}, +#endif { 0 } }; @@ -5712,6 +5982,16 @@ int main(int argc, char *argv[]) case OPT_INET_SOCKOPT: show_inet_sockopt = 1; break; +#ifdef HAVE_LIBBPF + case OPT_BPF_MAPS: + if (bpf_map_opts_add_all()) + exit(1); + break; + case OPT_BPF_MAP_ID: + if (bpf_map_opts_add_id(optarg)) + exit(1); + break; +#endif case 'h': help(); case '?': @@ -5810,6 +6090,9 @@ int main(int argc, char *argv[]) if (!(current_filter.states & (current_filter.states - 1))) columns[COL_STATE].disabled = 1; + if (bpf_map_opts.nr_maps) + columns[COL_SKSTOR].disabled = 0; + if (show_header) print_header(); @@ -5845,6 +6128,7 @@ int main(int argc, char *argv[]) if (show_processes || show_threads || show_proc_ctx || show_sock_ctx) user_ent_destroy(); + bpf_map_opts_destroy(); render(); -- 2.43.0 ^ permalink raw reply related [flat|nested] 6+ messages in thread
* Re: [PATCH v2 1/2] ss: add support for BPF socket-local storage 2023-12-08 14:57 ` [PATCH v2 1/2] ss: add support for " Quentin Deslandes @ 2023-12-12 22:49 ` Martin KaFai Lau 0 siblings, 0 replies; 6+ messages in thread From: Martin KaFai Lau @ 2023-12-12 22:49 UTC (permalink / raw) To: Quentin Deslandes; +Cc: David Ahern, Martin KaFai Lau, kernel-team, netdev On 12/8/23 6:57 AM, Quentin Deslandes wrote: > +static inline bool bpf_map_opts_is_enabled(void) nit. This is the only "inline" usage in this file. I would avoid it and depend on the compiler to decide. > +{ > + return bpf_map_opts.nr_maps; > +} > + [ ... ] > static int inet_show_sock(struct nlmsghdr *nlh, > struct sockstat *s) > { > @@ -3381,8 +3620,8 @@ static int inet_show_sock(struct nlmsghdr *nlh, > struct inet_diag_msg *r = NLMSG_DATA(nlh); > unsigned char v6only = 0; > > - parse_rtattr(tb, INET_DIAG_MAX, (struct rtattr *)(r+1), > - nlh->nlmsg_len - NLMSG_LENGTH(sizeof(*r))); > + parse_rtattr_flags(tb, INET_DIAG_MAX, (struct rtattr *)(r+1), > + nlh->nlmsg_len - NLMSG_LENGTH(sizeof(*r)), NLA_F_NESTED); > > if (tb[INET_DIAG_PROTOCOL]) > s->type = rta_getattr_u8(tb[INET_DIAG_PROTOCOL]); > @@ -3479,6 +3718,11 @@ static int inet_show_sock(struct nlmsghdr *nlh, > } > sctp_ino = s->ino; > > + if (tb[INET_DIAG_SK_BPF_STORAGES]) { > + field_set(COL_SKSTOR); > + show_sk_bpf_storages(tb[INET_DIAG_SK_BPF_STORAGES]); > + } > + > return 0; > } > > @@ -3560,13 +3804,14 @@ static int sockdiag_send(int family, int fd, int protocol, struct filter *f) > { > struct sockaddr_nl nladdr = { .nl_family = AF_NETLINK }; > DIAG_REQUEST(req, struct inet_diag_req_v2 r); > + struct rtattr *bpf_stgs_rta = NULL; > char *bc = NULL; > int bclen; > __u32 proto; > struct msghdr msg; > struct rtattr rta_bc; > struct rtattr rta_proto; > - struct iovec iov[5]; > + struct iovec iov[6]; > int iovlen = 1; > > if (family == PF_UNSPEC) > @@ -3619,6 +3864,17 @@ static int sockdiag_send(int family, int fd, int protocol, struct filter *f) > iovlen += 2; > } > > + if (bpf_map_opts_is_enabled()) { This will have compiler error when HAVE_LIBBPF is not set. > + bpf_stgs_rta = bpf_map_opts_alloc_rta(); > + if (!bpf_stgs_rta) { > + fprintf(stderr, "ss: cannot alloc request for --bpf-map\n"); > + return -1; > + } > + > + iov[iovlen++] = (struct iovec){ bpf_stgs_rta, bpf_stgs_rta->rta_len }; > + req.nlh.nlmsg_len += bpf_stgs_rta->rta_len; > + } > + > msg = (struct msghdr) { > .msg_name = (void *)&nladdr, > .msg_namelen = sizeof(nladdr), [ ... ] > @@ -5712,6 +5982,16 @@ int main(int argc, char *argv[]) > case OPT_INET_SOCKOPT: > show_inet_sockopt = 1; > break; > +#ifdef HAVE_LIBBPF > + case OPT_BPF_MAPS: > + if (bpf_map_opts_add_all()) > + exit(1); > + break; > + case OPT_BPF_MAP_ID: > + if (bpf_map_opts_add_id(optarg)) > + exit(1); > + break; > +#endif > case 'h': > help(); > case '?': > @@ -5810,6 +6090,9 @@ int main(int argc, char *argv[]) > if (!(current_filter.states & (current_filter.states - 1))) > columns[COL_STATE].disabled = 1; > > + if (bpf_map_opts.nr_maps) same here when HAVE_LIBBPF is not set > + columns[COL_SKSTOR].disabled = 0; > + > if (show_header) > print_header(); > > @@ -5845,6 +6128,7 @@ int main(int argc, char *argv[]) > > if (show_processes || show_threads || show_proc_ctx || show_sock_ctx) > user_ent_destroy(); > + bpf_map_opts_destroy(); same here. A #ifdef is needed. Another option is to create an empty or always-return-false function. > > render(); > ^ permalink raw reply [flat|nested] 6+ messages in thread
* [PATCH v2 2/2] ss: pretty-print BPF socket-local storage 2023-12-08 14:57 [PATCH v2 0/2] ss: pretty-printing BPF socket-local storage Quentin Deslandes 2023-12-08 14:57 ` [PATCH v2 1/2] ss: add support for " Quentin Deslandes @ 2023-12-08 14:57 ` Quentin Deslandes 2023-12-12 22:58 ` Martin KaFai Lau 2023-12-13 17:46 ` Alan Maguire 1 sibling, 2 replies; 6+ messages in thread From: Quentin Deslandes @ 2023-12-08 14:57 UTC (permalink / raw) To: netdev; +Cc: David Ahern, Martin KaFai Lau, Quentin Deslandes, kernel-team ss is able to print the map ID(s) for which a given socket has BPF socket-local storage defined (using --bpf-maps or --bpf-map-id=). However, the actual content of the map remains hidden. This change aims to pretty-print the socket-local storage content following the socket details, similar to what `bpftool map dump` would do. The exact output format is inspired by drgn, while the BTF data processing is similar to bpftool's. ss will use libbpf's btf_dump__dump_type_data() to ease pretty-printing of binary data. This requires out_bpf_sk_storage_print_fn() as a print callback function used by btf_dump__dump_type_data(). vout() is also introduced, which is similar to out() but accepts a va_list as parameter. COL_SKSTOR's header is replaced with an empty string, as it doesn't need to be printed anymore; it's used as a "virtual" column to refer to the socket-local storage dump, which will be printed under the socket information. The column's width is fixed to 1, so it doesn't mess up ss' output. ss' output remains unchanged unless --bpf-maps or --bpf-map-id= is used, in which case each socket containing BPF local storage will be followed by the content of the storage before the next socket's info is displayed. Signed-off-by: Quentin Deslandes <qde@naccy.de> --- misc/ss.c | 157 +++++++++++++++++++++++++++++++++++++++++++++++++----- 1 file changed, 145 insertions(+), 12 deletions(-) diff --git a/misc/ss.c b/misc/ss.c index f979b61d..6051b694 100644 --- a/misc/ss.c +++ b/misc/ss.c @@ -51,8 +51,13 @@ #include <linux/tls.h> #include <linux/mptcp.h> +#ifdef HAVE_LIBBPF +#include <linux/btf.h> +#endif + #ifdef HAVE_LIBBPF #include <bpf/bpf.h> +#include <bpf/btf.h> #include <bpf/libbpf.h> #endif @@ -136,7 +141,7 @@ static struct column columns[] = { { ALIGN_RIGHT, "Peer Address:", " ", 0, 0, 0 }, { ALIGN_LEFT, "Port", "", 0, 0, 0 }, { ALIGN_LEFT, "Process", "", 0, 0, 0 }, - { ALIGN_LEFT, "Socket storage", "", 1, 0, 0 }, + { ALIGN_LEFT, "", "", 1, 0, 0 }, { ALIGN_LEFT, "", "", 0, 0, 0 }, }; @@ -1039,11 +1044,10 @@ static int buf_update(int len) } /* Append content to buffer as part of the current field */ -__attribute__((format(printf, 1, 2))) -static void out(const char *fmt, ...) +static void vout(const char *fmt, va_list args) { struct column *f = current_field; - va_list args; + va_list _args; char *pos; int len; @@ -1054,18 +1058,27 @@ static void out(const char *fmt, ...) buffer.head = buf_chunk_new(); again: /* Append to buffer: if we have a new chunk, print again */ + va_copy(_args, args); pos = buffer.cur->data + buffer.cur->len; - va_start(args, fmt); /* Limit to tail room. If we hit the limit, buf_update() will tell us */ len = vsnprintf(pos, buf_chunk_avail(buffer.tail), fmt, args); - va_end(args); if (buf_update(len)) goto again; } +__attribute__((format(printf, 1, 2))) +static void out(const char *fmt, ...) +{ + va_list args; + + va_start(args, fmt); + vout(fmt, args); + va_end(args); +} + static int print_left_spacing(struct column *f, int stored, int printed) { int s; @@ -1213,6 +1226,9 @@ static void render_calc_width(void) */ c->width = min(c->width, screen_width); + if (c == &columns[COL_SKSTOR]) + c->width = 1; + if (c->width) first = 0; } @@ -3392,6 +3408,8 @@ static struct bpf_map_opts { struct bpf_sk_storage_map_info { unsigned int id; int fd; + struct bpf_map_info info; + struct btf *btf; } maps[MAX_NR_BPF_MAP_ID_OPTS]; bool show_all; struct btf *kernel_btf; @@ -3403,6 +3421,22 @@ static void bpf_map_opts_mixed_error(void) "ss: --bpf-maps and --bpf-map-id cannot be used together\n"); } +static int bpf_maps_opts_load_btf(struct bpf_map_info *info, struct btf **btf) +{ + if (info->btf_value_type_id) { + *btf = btf__load_from_kernel_by_id(info->btf_id); + if (!*btf) { + fprintf(stderr, "ss: failed to load BTF for map ID %u\n", + info->id); + return -1; + } + } else { + *btf = NULL; + } + + return 0; +} + static int bpf_map_opts_add_all(void) { unsigned int i; @@ -3418,6 +3452,7 @@ static int bpf_map_opts_add_all(void) while (1) { struct bpf_map_info info = {}; uint32_t len = sizeof(info); + struct btf *btf; r = bpf_map_get_next_id(id, &id); if (r) { @@ -3462,8 +3497,18 @@ static int bpf_map_opts_add_all(void) continue; } + r = bpf_maps_opts_load_btf(&info, &btf); + if (r) { + fprintf(stderr, "ss: failed to get BTF data for BPF map ID: %u\n", + id); + close(fd); + goto err; + } + bpf_map_opts.maps[bpf_map_opts.nr_maps].id = id; - bpf_map_opts.maps[bpf_map_opts.nr_maps++].fd = fd; + bpf_map_opts.maps[bpf_map_opts.nr_maps].fd = fd; + bpf_map_opts.maps[bpf_map_opts.nr_maps].info = info; + bpf_map_opts.maps[bpf_map_opts.nr_maps++].btf = btf; } bpf_map_opts.show_all = true; @@ -3482,6 +3527,7 @@ static int bpf_map_opts_add_id(const char *optarg) struct bpf_map_info info = {}; uint32_t len = sizeof(info); size_t optarg_len; + struct btf *btf; unsigned long id; unsigned int i; char *end; @@ -3539,8 +3585,17 @@ static int bpf_map_opts_add_id(const char *optarg) return -1; } + r = bpf_maps_opts_load_btf(&info, &btf); + if (r) { + fprintf(stderr, "ss: failed to get BTF data for BPF map ID: %lu\n", + id); + return -1; + } + bpf_map_opts.maps[bpf_map_opts.nr_maps].id = id; - bpf_map_opts.maps[bpf_map_opts.nr_maps++].fd = fd; + bpf_map_opts.maps[bpf_map_opts.nr_maps].fd = fd; + bpf_map_opts.maps[bpf_map_opts.nr_maps].info = info; + bpf_map_opts.maps[bpf_map_opts.nr_maps++].btf = btf; return 0; } @@ -3549,8 +3604,23 @@ static void bpf_map_opts_destroy(void) { int i; - for (i = 0; i < bpf_map_opts.nr_maps; ++i) + for (i = 0; i < bpf_map_opts.nr_maps; ++i) { + btf__free(bpf_map_opts.maps[i].btf); close(bpf_map_opts.maps[i].fd); + } +} + +static const struct bpf_sk_storage_map_info *bpf_map_opts_get_info( + unsigned int map_id) +{ + unsigned int i; + + for (i = 0; i < bpf_map_opts.nr_maps; ++i) { + if (bpf_map_opts.maps[i].id == map_id) + return &bpf_map_opts.maps[i]; + } + + return NULL; } static inline bool bpf_map_opts_is_enabled(void) @@ -3590,10 +3660,63 @@ static struct rtattr *bpf_map_opts_alloc_rta(void) return stgs_rta; } +static void out_bpf_sk_storage_print_fn(void *ctx, const char *fmt, va_list args) +{ + vout(fmt, args); +} + +#define SK_STORAGE_INDENT_STR " " + +static void out_bpf_sk_storage(int map_id, const void *data, size_t len) +{ + uint32_t type_id; + const struct bpf_sk_storage_map_info *map_info; + struct btf_dump *dump; + struct btf_dump_type_data_opts opts = { + .sz = sizeof(struct btf_dump_type_data_opts), + .indent_str = SK_STORAGE_INDENT_STR, + .indent_level = 2, + .emit_zeroes = 1 + }; + struct btf_dump_opts dopts = { + .sz = sizeof(struct btf_dump_opts) + }; + int r; + + map_info = bpf_map_opts_get_info(map_id); + if (!map_info) { + fprintf(stderr, "map_id: %d: missing map info", map_id); + return; + } + + if (map_info->info.value_size != len) { + fprintf(stderr, "map_id: %d: invalid value size, expecting %u, got %lu\n", + map_id, map_info->info.value_size, len); + return; + } + + type_id = map_info->info.btf_value_type_id; + + dump = btf_dump__new(map_info->btf, out_bpf_sk_storage_print_fn, NULL, &dopts); + if (!dump) { + fprintf(stderr, "Failed to create btf_dump object\n"); + return; + } + + out(SK_STORAGE_INDENT_STR "map_id: %d [\n", map_id); + r = btf_dump__dump_type_data(dump, type_id, data, len, &opts); + if (r < 0) + out(SK_STORAGE_INDENT_STR SK_STORAGE_INDENT_STR "failed to dump data: %d", r); + out("\n" SK_STORAGE_INDENT_STR "]"); + + btf_dump__free(dump); +} + static void show_sk_bpf_storages(struct rtattr *bpf_stgs) { struct rtattr *tb[SK_DIAG_BPF_STORAGE_MAX + 1], *bpf_stg; - unsigned int rem; + unsigned int rem, map_id; + struct rtattr *value; for (bpf_stg = RTA_DATA(bpf_stgs), rem = RTA_PAYLOAD(bpf_stgs); RTA_OK(bpf_stg, rem); bpf_stg = RTA_NEXT(bpf_stg, rem)) { @@ -3605,8 +3728,13 @@ static void show_sk_bpf_storages(struct rtattr *bpf_stgs) (struct rtattr *)bpf_stg); if (tb[SK_DIAG_BPF_STORAGE_MAP_ID]) { - out("map_id:%u", - rta_getattr_u32(tb[SK_DIAG_BPF_STORAGE_MAP_ID])); + out("\n"); + + map_id = rta_getattr_u32(tb[SK_DIAG_BPF_STORAGE_MAP_ID]); + value = tb[SK_DIAG_BPF_STORAGE_MAP_VALUE]; + + out_bpf_sk_storage(map_id, RTA_DATA(value), + RTA_PAYLOAD(value)); } } } @@ -6000,6 +6128,11 @@ int main(int argc, char *argv[]) } } + if (oneline && (bpf_map_opts.nr_maps || bpf_map_opts.show_all)) { + fprintf(stderr, "ss: --oneline, --bpf-maps, and --bpf-map-id are incompatible\n"); + exit(-1); + } + if (show_processes || show_threads || show_proc_ctx || show_sock_ctx) user_ent_hash_build(); -- 2.43.0 ^ permalink raw reply related [flat|nested] 6+ messages in thread
* Re: [PATCH v2 2/2] ss: pretty-print BPF socket-local storage 2023-12-08 14:57 ` [PATCH v2 2/2] ss: pretty-print " Quentin Deslandes @ 2023-12-12 22:58 ` Martin KaFai Lau 2023-12-13 17:46 ` Alan Maguire 1 sibling, 0 replies; 6+ messages in thread From: Martin KaFai Lau @ 2023-12-12 22:58 UTC (permalink / raw) To: Quentin Deslandes; +Cc: David Ahern, Martin KaFai Lau, kernel-team, netdev On 12/8/23 6:57 AM, Quentin Deslandes wrote: > @@ -1039,11 +1044,10 @@ static int buf_update(int len) > } > > /* Append content to buffer as part of the current field */ > -__attribute__((format(printf, 1, 2))) > -static void out(const char *fmt, ...) > +static void vout(const char *fmt, va_list args) > { > struct column *f = current_field; > - va_list args; > + va_list _args; > char *pos; > int len; > > @@ -1054,18 +1058,27 @@ static void out(const char *fmt, ...) > buffer.head = buf_chunk_new(); > > again: /* Append to buffer: if we have a new chunk, print again */ > + va_copy(_args, args); > > pos = buffer.cur->data + buffer.cur->len; > - va_start(args, fmt); > > /* Limit to tail room. If we hit the limit, buf_update() will tell us */ > len = vsnprintf(pos, buf_chunk_avail(buffer.tail), fmt, args); hmm... based on the changes made in this function, I am pretty sure the intention is to pass the "_args" here instead of passing the "args". Please double check. > - va_end(args); > > if (buf_update(len)) > goto again; > } > > +__attribute__((format(printf, 1, 2))) > +static void out(const char *fmt, ...) > +{ > + va_list args; > + > + va_start(args, fmt); > + vout(fmt, args); > + va_end(args); > +} > + > static int print_left_spacing(struct column *f, int stored, int printed) > { > int s; > @@ -1213,6 +1226,9 @@ static void render_calc_width(void) > */ > c->width = min(c->width, screen_width); > > + if (c == &columns[COL_SKSTOR]) > + c->width = 1; > + > if (c->width) > first = 0; > } > @@ -3392,6 +3408,8 @@ static struct bpf_map_opts { > struct bpf_sk_storage_map_info { > unsigned int id; > int fd; > + struct bpf_map_info info; > + struct btf *btf; > } maps[MAX_NR_BPF_MAP_ID_OPTS]; > bool show_all; > struct btf *kernel_btf; > @@ -3403,6 +3421,22 @@ static void bpf_map_opts_mixed_error(void) > "ss: --bpf-maps and --bpf-map-id cannot be used together\n"); > } > > +static int bpf_maps_opts_load_btf(struct bpf_map_info *info, struct btf **btf) > +{ > + if (info->btf_value_type_id) { > + *btf = btf__load_from_kernel_by_id(info->btf_id); > + if (!*btf) { > + fprintf(stderr, "ss: failed to load BTF for map ID %u\n", > + info->id); > + return -1; > + } > + } else { > + *btf = NULL; > + } > + > + return 0; > +} > + > static int bpf_map_opts_add_all(void) > { > unsigned int i; > @@ -3418,6 +3452,7 @@ static int bpf_map_opts_add_all(void) > while (1) { > struct bpf_map_info info = {}; > uint32_t len = sizeof(info); > + struct btf *btf; > > r = bpf_map_get_next_id(id, &id); > if (r) { > @@ -3462,8 +3497,18 @@ static int bpf_map_opts_add_all(void) > continue; > } > > + r = bpf_maps_opts_load_btf(&info, &btf); > + if (r) { > + fprintf(stderr, "ss: failed to get BTF data for BPF map ID: %u\n", > + id); > + close(fd); > + goto err; > + } > + > bpf_map_opts.maps[bpf_map_opts.nr_maps].id = id; > - bpf_map_opts.maps[bpf_map_opts.nr_maps++].fd = fd; > + bpf_map_opts.maps[bpf_map_opts.nr_maps].fd = fd; > + bpf_map_opts.maps[bpf_map_opts.nr_maps].info = info; > + bpf_map_opts.maps[bpf_map_opts.nr_maps++].btf = btf; The "err:" path of this function needs a btf__free(). > } > > bpf_map_opts.show_all = true; > @@ -3482,6 +3527,7 @@ static int bpf_map_opts_add_id(const char *optarg) > struct bpf_map_info info = {}; > uint32_t len = sizeof(info); > size_t optarg_len; > + struct btf *btf; > unsigned long id; > unsigned int i; > char *end; > @@ -3539,8 +3585,17 @@ static int bpf_map_opts_add_id(const char *optarg) > return -1; > } > > + r = bpf_maps_opts_load_btf(&info, &btf); > + if (r) { > + fprintf(stderr, "ss: failed to get BTF data for BPF map ID: %lu\n", > + id); close(fd); > + return -1; > + } > + > bpf_map_opts.maps[bpf_map_opts.nr_maps].id = id; > - bpf_map_opts.maps[bpf_map_opts.nr_maps++].fd = fd; > + bpf_map_opts.maps[bpf_map_opts.nr_maps].fd = fd; > + bpf_map_opts.maps[bpf_map_opts.nr_maps].info = info; > + bpf_map_opts.maps[bpf_map_opts.nr_maps++].btf = btf; > > return 0; > } ^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: [PATCH v2 2/2] ss: pretty-print BPF socket-local storage 2023-12-08 14:57 ` [PATCH v2 2/2] ss: pretty-print " Quentin Deslandes 2023-12-12 22:58 ` Martin KaFai Lau @ 2023-12-13 17:46 ` Alan Maguire 1 sibling, 0 replies; 6+ messages in thread From: Alan Maguire @ 2023-12-13 17:46 UTC (permalink / raw) To: Quentin Deslandes, netdev; +Cc: David Ahern, Martin KaFai Lau, kernel-team On 08/12/2023 14:57, Quentin Deslandes wrote: > ss is able to print the map ID(s) for which a given socket has BPF > socket-local storage defined (using --bpf-maps or --bpf-map-id=). However, > the actual content of the map remains hidden. > > This change aims to pretty-print the socket-local storage content following > the socket details, similar to what `bpftool map dump` would do. The exact > output format is inspired by drgn, while the BTF data processing is similar > to bpftool's. > > ss will use libbpf's btf_dump__dump_type_data() to ease pretty-printing > of binary data. This requires out_bpf_sk_storage_print_fn() as a print > callback function used by btf_dump__dump_type_data(). vout() is also > introduced, which is similar to out() but accepts a va_list as > parameter. > > COL_SKSTOR's header is replaced with an empty string, as it doesn't need to > be printed anymore; it's used as a "virtual" column to refer to the > socket-local storage dump, which will be printed under the socket information. > The column's width is fixed to 1, so it doesn't mess up ss' output. > > ss' output remains unchanged unless --bpf-maps or --bpf-map-id= is used, > in which case each socket containing BPF local storage will be followed by > the content of the storage before the next socket's info is displayed. > this is great! one small idea below, but > Signed-off-by: Quentin Deslandes <qde@naccy.de> Reviewed-by: Alan Maguire <alan.maguire@oracle.com> > --- > misc/ss.c | 157 +++++++++++++++++++++++++++++++++++++++++++++++++----- > 1 file changed, 145 insertions(+), 12 deletions(-) > > diff --git a/misc/ss.c b/misc/ss.c > index f979b61d..6051b694 100644 > --- a/misc/ss.c > +++ b/misc/ss.c > @@ -51,8 +51,13 @@ > #include <linux/tls.h> > #include <linux/mptcp.h> > > +#ifdef HAVE_LIBBPF > +#include <linux/btf.h> > +#endif > + > #ifdef HAVE_LIBBPF > #include <bpf/bpf.h> > +#include <bpf/btf.h> > #include <bpf/libbpf.h> > #endif > > @@ -136,7 +141,7 @@ static struct column columns[] = { > { ALIGN_RIGHT, "Peer Address:", " ", 0, 0, 0 }, > { ALIGN_LEFT, "Port", "", 0, 0, 0 }, > { ALIGN_LEFT, "Process", "", 0, 0, 0 }, > - { ALIGN_LEFT, "Socket storage", "", 1, 0, 0 }, > + { ALIGN_LEFT, "", "", 1, 0, 0 }, > { ALIGN_LEFT, "", "", 0, 0, 0 }, > }; > > @@ -1039,11 +1044,10 @@ static int buf_update(int len) > } > > /* Append content to buffer as part of the current field */ > -__attribute__((format(printf, 1, 2))) > -static void out(const char *fmt, ...) > +static void vout(const char *fmt, va_list args) > { > struct column *f = current_field; > - va_list args; > + va_list _args; > char *pos; > int len; > > @@ -1054,18 +1058,27 @@ static void out(const char *fmt, ...) > buffer.head = buf_chunk_new(); > > again: /* Append to buffer: if we have a new chunk, print again */ > + va_copy(_args, args); > > pos = buffer.cur->data + buffer.cur->len; > - va_start(args, fmt); > > /* Limit to tail room. If we hit the limit, buf_update() will tell us */ > len = vsnprintf(pos, buf_chunk_avail(buffer.tail), fmt, args); > - va_end(args); > > if (buf_update(len)) > goto again; > } > > +__attribute__((format(printf, 1, 2))) > +static void out(const char *fmt, ...) > +{ > + va_list args; > + > + va_start(args, fmt); > + vout(fmt, args); > + va_end(args); > +} > + > static int print_left_spacing(struct column *f, int stored, int printed) > { > int s; > @@ -1213,6 +1226,9 @@ static void render_calc_width(void) > */ > c->width = min(c->width, screen_width); > > + if (c == &columns[COL_SKSTOR]) > + c->width = 1; > + > if (c->width) > first = 0; > } > @@ -3392,6 +3408,8 @@ static struct bpf_map_opts { > struct bpf_sk_storage_map_info { > unsigned int id; > int fd; > + struct bpf_map_info info; > + struct btf *btf; > } maps[MAX_NR_BPF_MAP_ID_OPTS]; > bool show_all; > struct btf *kernel_btf; > @@ -3403,6 +3421,22 @@ static void bpf_map_opts_mixed_error(void) > "ss: --bpf-maps and --bpf-map-id cannot be used together\n"); > } > > +static int bpf_maps_opts_load_btf(struct bpf_map_info *info, struct btf **btf) > +{ > + if (info->btf_value_type_id) { > + *btf = btf__load_from_kernel_by_id(info->btf_id); > + if (!*btf) { > + fprintf(stderr, "ss: failed to load BTF for map ID %u\n", > + info->id); > + return -1; > + } > + } else { > + *btf = NULL; > + } > + > + return 0; > +} > + > static int bpf_map_opts_add_all(void) > { > unsigned int i; > @@ -3418,6 +3452,7 @@ static int bpf_map_opts_add_all(void) > while (1) { > struct bpf_map_info info = {}; > uint32_t len = sizeof(info); > + struct btf *btf; > > r = bpf_map_get_next_id(id, &id); > if (r) { > @@ -3462,8 +3497,18 @@ static int bpf_map_opts_add_all(void) > continue; > } > > + r = bpf_maps_opts_load_btf(&info, &btf); > + if (r) { > + fprintf(stderr, "ss: failed to get BTF data for BPF map ID: %u\n", > + id); > + close(fd); > + goto err; > + } > + > bpf_map_opts.maps[bpf_map_opts.nr_maps].id = id; > - bpf_map_opts.maps[bpf_map_opts.nr_maps++].fd = fd; > + bpf_map_opts.maps[bpf_map_opts.nr_maps].fd = fd; > + bpf_map_opts.maps[bpf_map_opts.nr_maps].info = info; > + bpf_map_opts.maps[bpf_map_opts.nr_maps++].btf = btf; > } > > bpf_map_opts.show_all = true; > @@ -3482,6 +3527,7 @@ static int bpf_map_opts_add_id(const char *optarg) > struct bpf_map_info info = {}; > uint32_t len = sizeof(info); > size_t optarg_len; > + struct btf *btf; > unsigned long id; > unsigned int i; > char *end; > @@ -3539,8 +3585,17 @@ static int bpf_map_opts_add_id(const char *optarg) > return -1; > } > > + r = bpf_maps_opts_load_btf(&info, &btf); > + if (r) { > + fprintf(stderr, "ss: failed to get BTF data for BPF map ID: %lu\n", > + id); > + return -1; > + } > + > bpf_map_opts.maps[bpf_map_opts.nr_maps].id = id; > - bpf_map_opts.maps[bpf_map_opts.nr_maps++].fd = fd; > + bpf_map_opts.maps[bpf_map_opts.nr_maps].fd = fd; > + bpf_map_opts.maps[bpf_map_opts.nr_maps].info = info; > + bpf_map_opts.maps[bpf_map_opts.nr_maps++].btf = btf; > > return 0; > } > @@ -3549,8 +3604,23 @@ static void bpf_map_opts_destroy(void) > { > int i; > > - for (i = 0; i < bpf_map_opts.nr_maps; ++i) > + for (i = 0; i < bpf_map_opts.nr_maps; ++i) { > + btf__free(bpf_map_opts.maps[i].btf); > close(bpf_map_opts.maps[i].fd); > + } > +} > + > +static const struct bpf_sk_storage_map_info *bpf_map_opts_get_info( > + unsigned int map_id) > +{ > + unsigned int i; > + > + for (i = 0; i < bpf_map_opts.nr_maps; ++i) { > + if (bpf_map_opts.maps[i].id == map_id) > + return &bpf_map_opts.maps[i]; > + } > + > + return NULL; > } > > static inline bool bpf_map_opts_is_enabled(void) > @@ -3590,10 +3660,63 @@ static struct rtattr *bpf_map_opts_alloc_rta(void) > return stgs_rta; > } > > +static void out_bpf_sk_storage_print_fn(void *ctx, const char *fmt, va_list args) > +{ > + vout(fmt, args); > +} > + > +#define SK_STORAGE_INDENT_STR " " > + > +static void out_bpf_sk_storage(int map_id, const void *data, size_t len) > +{ > + uint32_t type_id; > + const struct bpf_sk_storage_map_info *map_info; > + struct btf_dump *dump; > + struct btf_dump_type_data_opts opts = { > + .sz = sizeof(struct btf_dump_type_data_opts), > + .indent_str = SK_STORAGE_INDENT_STR, > + .indent_level = 2, > + .emit_zeroes = 1 > + }; > + struct btf_dump_opts dopts = { > + .sz = sizeof(struct btf_dump_opts) > + }; > + int r; > + > + map_info = bpf_map_opts_get_info(map_id); > + if (!map_info) { > + fprintf(stderr, "map_id: %d: missing map info", map_id); > + return; > + } > + > + if (map_info->info.value_size != len) { > + fprintf(stderr, "map_id: %d: invalid value size, expecting %u, got %lu\n", > + map_id, map_info->info.value_size, len); > + return; > + } > + > + type_id = map_info->info.btf_value_type_id; > + > + dump = btf_dump__new(map_info->btf, out_bpf_sk_storage_print_fn, NULL, &dopts); > + if (!dump) { > + fprintf(stderr, "Failed to create btf_dump object\n"); > + return; > + } > + > + out(SK_STORAGE_INDENT_STR "map_id: %d [\n", map_id); > + r = btf_dump__dump_type_data(dump, type_id, data, len, &opts); > + if (r < 0) > + out(SK_STORAGE_INDENT_STR SK_STORAGE_INDENT_STR "failed to dump data: %d", r); > + out("\n" SK_STORAGE_INDENT_STR "]"); > + > + btf_dump__free(dump); > +} > + > static void show_sk_bpf_storages(struct rtattr *bpf_stgs) > { > struct rtattr *tb[SK_DIAG_BPF_STORAGE_MAX + 1], *bpf_stg; > - unsigned int rem; > + unsigned int rem, map_id; > + struct rtattr *value; > > for (bpf_stg = RTA_DATA(bpf_stgs), rem = RTA_PAYLOAD(bpf_stgs); > RTA_OK(bpf_stg, rem); bpf_stg = RTA_NEXT(bpf_stg, rem)) { > @@ -3605,8 +3728,13 @@ static void show_sk_bpf_storages(struct rtattr *bpf_stgs) > (struct rtattr *)bpf_stg); > > if (tb[SK_DIAG_BPF_STORAGE_MAP_ID]) { > - out("map_id:%u", > - rta_getattr_u32(tb[SK_DIAG_BPF_STORAGE_MAP_ID])); > + out("\n"); > + > + map_id = rta_getattr_u32(tb[SK_DIAG_BPF_STORAGE_MAP_ID]); > + value = tb[SK_DIAG_BPF_STORAGE_MAP_VALUE]; > + > + out_bpf_sk_storage(map_id, RTA_DATA(value), > + RTA_PAYLOAD(value)); > } > } > } > @@ -6000,6 +6128,11 @@ int main(int argc, char *argv[]) > } > } > > + if (oneline && (bpf_map_opts.nr_maps || bpf_map_opts.show_all)) { > + fprintf(stderr, "ss: --oneline, --bpf-maps, and --bpf-map-id are incompatible\n"); > + exit(-1); > + } > + I guess it would be possible to provide oneline output if we used compact-mode + omit-zeros BTF data dumping. Did you try that or is the output just too cluttered? > if (show_processes || show_threads || show_proc_ctx || show_sock_ctx) > user_ent_hash_build(); > ^ permalink raw reply [flat|nested] 6+ messages in thread
end of thread, other threads:[~2023-12-13 17:46 UTC | newest] Thread overview: 6+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2023-12-08 14:57 [PATCH v2 0/2] ss: pretty-printing BPF socket-local storage Quentin Deslandes 2023-12-08 14:57 ` [PATCH v2 1/2] ss: add support for " Quentin Deslandes 2023-12-12 22:49 ` Martin KaFai Lau 2023-12-08 14:57 ` [PATCH v2 2/2] ss: pretty-print " Quentin Deslandes 2023-12-12 22:58 ` Martin KaFai Lau 2023-12-13 17:46 ` Alan Maguire
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).