1 // SPDX-License-Identifier: GPL-2.0-only
2 /* Copyright (c) 2025 Google LLC */
3 #include <linux/bpf.h>
4 #include <linux/btf_ids.h>
5 #include <linux/dma-buf.h>
6 #include <linux/kernel.h>
7 #include <linux/seq_file.h>
8
9 struct dmabuf_iter_priv {
10 /*
11 * If this pointer is non-NULL, the buffer's refcount is elevated to
12 * prevent destruction between stop/start. If reading is not resumed and
13 * start is never called again, then dmabuf_iter_seq_fini drops the
14 * reference when the iterator is released.
15 */
16 struct dma_buf *dmabuf;
17 };
18
dmabuf_iter_seq_start(struct seq_file * seq,loff_t * pos)19 static void *dmabuf_iter_seq_start(struct seq_file *seq, loff_t *pos)
20 {
21 struct dmabuf_iter_priv *p = seq->private;
22
23 if (*pos) {
24 struct dma_buf *dmabuf = p->dmabuf;
25
26 if (!dmabuf)
27 return NULL;
28
29 /*
30 * Always resume from where we stopped, regardless of the value
31 * of pos.
32 */
33 p->dmabuf = NULL;
34 return dmabuf;
35 }
36
37 return dma_buf_iter_begin();
38 }
39
dmabuf_iter_seq_next(struct seq_file * seq,void * v,loff_t * pos)40 static void *dmabuf_iter_seq_next(struct seq_file *seq, void *v, loff_t *pos)
41 {
42 struct dma_buf *dmabuf = v;
43
44 ++*pos;
45
46 return dma_buf_iter_next(dmabuf);
47 }
48
49 struct bpf_iter__dmabuf {
50 __bpf_md_ptr(struct bpf_iter_meta *, meta);
51 __bpf_md_ptr(struct dma_buf *, dmabuf);
52 };
53
__dmabuf_seq_show(struct seq_file * seq,void * v,bool in_stop)54 static int __dmabuf_seq_show(struct seq_file *seq, void *v, bool in_stop)
55 {
56 struct bpf_iter_meta meta = {
57 .seq = seq,
58 };
59 struct bpf_iter__dmabuf ctx = {
60 .meta = &meta,
61 .dmabuf = v,
62 };
63 struct bpf_prog *prog = bpf_iter_get_info(&meta, in_stop);
64
65 if (prog)
66 return bpf_iter_run_prog(prog, &ctx);
67
68 return 0;
69 }
70
dmabuf_iter_seq_show(struct seq_file * seq,void * v)71 static int dmabuf_iter_seq_show(struct seq_file *seq, void *v)
72 {
73 return __dmabuf_seq_show(seq, v, false);
74 }
75
dmabuf_iter_seq_stop(struct seq_file * seq,void * v)76 static void dmabuf_iter_seq_stop(struct seq_file *seq, void *v)
77 {
78 struct dma_buf *dmabuf = v;
79
80 if (dmabuf) {
81 struct dmabuf_iter_priv *p = seq->private;
82
83 p->dmabuf = dmabuf;
84 }
85 }
86
87 static const struct seq_operations dmabuf_iter_seq_ops = {
88 .start = dmabuf_iter_seq_start,
89 .next = dmabuf_iter_seq_next,
90 .stop = dmabuf_iter_seq_stop,
91 .show = dmabuf_iter_seq_show,
92 };
93
bpf_iter_dmabuf_show_fdinfo(const struct bpf_iter_aux_info * aux,struct seq_file * seq)94 static void bpf_iter_dmabuf_show_fdinfo(const struct bpf_iter_aux_info *aux,
95 struct seq_file *seq)
96 {
97 seq_puts(seq, "dmabuf iter\n");
98 }
99
dmabuf_iter_seq_init(void * priv,struct bpf_iter_aux_info * aux)100 static int dmabuf_iter_seq_init(void *priv, struct bpf_iter_aux_info *aux)
101 {
102 struct dmabuf_iter_priv *p = (struct dmabuf_iter_priv *)priv;
103
104 p->dmabuf = NULL;
105 return 0;
106 }
107
dmabuf_iter_seq_fini(void * priv)108 static void dmabuf_iter_seq_fini(void *priv)
109 {
110 struct dmabuf_iter_priv *p = (struct dmabuf_iter_priv *)priv;
111
112 if (p->dmabuf)
113 dma_buf_put(p->dmabuf);
114 }
115
116 static const struct bpf_iter_seq_info dmabuf_iter_seq_info = {
117 .seq_ops = &dmabuf_iter_seq_ops,
118 .init_seq_private = dmabuf_iter_seq_init,
119 .fini_seq_private = dmabuf_iter_seq_fini,
120 .seq_priv_size = sizeof(struct dmabuf_iter_priv),
121 };
122
123 static struct bpf_iter_reg bpf_dmabuf_reg_info = {
124 .target = "dmabuf",
125 .feature = BPF_ITER_RESCHED,
126 .show_fdinfo = bpf_iter_dmabuf_show_fdinfo,
127 .ctx_arg_info_size = 1,
128 .ctx_arg_info = {
129 { offsetof(struct bpf_iter__dmabuf, dmabuf),
130 PTR_TO_BTF_ID_OR_NULL },
131 },
132 .seq_info = &dmabuf_iter_seq_info,
133 };
134
DEFINE_BPF_ITER_FUNC(dmabuf,struct bpf_iter_meta * meta,struct dma_buf * dmabuf)135 DEFINE_BPF_ITER_FUNC(dmabuf, struct bpf_iter_meta *meta, struct dma_buf *dmabuf)
136 BTF_ID_LIST_SINGLE(bpf_dmabuf_btf_id, struct, dma_buf)
137
138 static int __init dmabuf_iter_init(void)
139 {
140 bpf_dmabuf_reg_info.ctx_arg_info[0].btf_id = bpf_dmabuf_btf_id[0];
141 return bpf_iter_reg_target(&bpf_dmabuf_reg_info);
142 }
143
144 late_initcall(dmabuf_iter_init);
145
146 struct bpf_iter_dmabuf {
147 /*
148 * opaque iterator state; having __u64 here allows to preserve correct
149 * alignment requirements in vmlinux.h, generated from BTF
150 */
151 __u64 __opaque[1];
152 } __aligned(8);
153
154 /* Non-opaque version of bpf_iter_dmabuf */
155 struct bpf_iter_dmabuf_kern {
156 struct dma_buf *dmabuf;
157 } __aligned(8);
158
159 __bpf_kfunc_start_defs();
160
bpf_iter_dmabuf_new(struct bpf_iter_dmabuf * it)161 __bpf_kfunc int bpf_iter_dmabuf_new(struct bpf_iter_dmabuf *it)
162 {
163 struct bpf_iter_dmabuf_kern *kit = (void *)it;
164
165 BUILD_BUG_ON(sizeof(*kit) > sizeof(*it));
166 BUILD_BUG_ON(__alignof__(*kit) != __alignof__(*it));
167
168 kit->dmabuf = NULL;
169 return 0;
170 }
171
bpf_iter_dmabuf_next(struct bpf_iter_dmabuf * it)172 __bpf_kfunc struct dma_buf *bpf_iter_dmabuf_next(struct bpf_iter_dmabuf *it)
173 {
174 struct bpf_iter_dmabuf_kern *kit = (void *)it;
175
176 if (kit->dmabuf)
177 kit->dmabuf = dma_buf_iter_next(kit->dmabuf);
178 else
179 kit->dmabuf = dma_buf_iter_begin();
180
181 return kit->dmabuf;
182 }
183
bpf_iter_dmabuf_destroy(struct bpf_iter_dmabuf * it)184 __bpf_kfunc void bpf_iter_dmabuf_destroy(struct bpf_iter_dmabuf *it)
185 {
186 struct bpf_iter_dmabuf_kern *kit = (void *)it;
187
188 if (kit->dmabuf)
189 dma_buf_put(kit->dmabuf);
190 }
191
192 __bpf_kfunc_end_defs();
193