Lines Matching +full:- +full:affinity

1 // SPDX-License-Identifier: GPL-2.0-only
3 * Copyright (C) 2011-2017, Red Hat Inc, Arnaldo Carvalho de Melo <acme@redhat.com>
5 * Parts came from evlist.c builtin-{top,stat,record}.c, see those files for further
34 len = bitmap_scnprintf(mask->bits, mask->nbits, buf, MASK_SIZE);
36 pr_debug("%p: %s mask[%zd]: %s\n", mask, tag, mask->nbits, buf);
41 return perf_mmap__mmap_len(&map->core);
73 return map->aio.nr_cblocks > 0;
79 map->aio.data[idx] = mmap(NULL, mmap__mmap_len(map), PROT_READ|PROT_WRITE,
81 if (map->aio.data[idx] == MAP_FAILED) {
82 map->aio.data[idx] = NULL;
83 return -1;
91 if (map->aio.data[idx]) {
92 munmap(map->aio.data[idx], mmap__mmap_len(map));
93 map->aio.data[idx] = NULL;
97 static int perf_mmap__aio_bind(struct mmap *map, int idx, struct perf_cpu cpu, int affinity)
105 if (affinity != PERF_AFFINITY_SYS && cpu__max_node() > 1) {
106 data = map->aio.data[idx];
112 return -1;
116 pr_err("Failed to bind [%p-%p] AIO buffer to node %lu: error %m\n",
118 err = -1;
128 map->aio.data[idx] = malloc(mmap__mmap_len(map));
129 if (map->aio.data[idx] == NULL)
130 return -1;
137 zfree(&(map->aio.data[idx]));
141 struct perf_cpu cpu __maybe_unused, int affinity __maybe_unused)
151 map->aio.nr_cblocks = mp->nr_cblocks;
152 if (map->aio.nr_cblocks) {
153 map->aio.aiocb = calloc(map->aio.nr_cblocks, sizeof(struct aiocb *));
154 if (!map->aio.aiocb) {
156 return -1;
158 map->aio.cblocks = calloc(map->aio.nr_cblocks, sizeof(struct aiocb));
159 if (!map->aio.cblocks) {
161 return -1;
163 map->aio.data = calloc(map->aio.nr_cblocks, sizeof(void *));
164 if (!map->aio.data) {
166 return -1;
169 for (i = 0; i < map->aio.nr_cblocks; ++i) {
171 if (ret == -1) {
173 return -1;
175 ret = perf_mmap__aio_bind(map, i, map->core.cpu, mp->affinity);
176 if (ret == -1)
177 return -1;
179 * Use cblock.aio_fildes value different from -1
184 map->aio.cblocks[i].aio_fildes = -1;
188 * are kept in separate per-prio queues and adding
189 * a new request will iterate thru shorter per-prio
193 prio = delta_max - i;
194 map->aio.cblocks[i].aio_reqprio = prio >= 0 ? prio : 0;
205 for (i = 0; i < map->aio.nr_cblocks; ++i)
207 if (map->aio.data)
208 zfree(&map->aio.data);
209 zfree(&map->aio.cblocks);
210 zfree(&map->aio.aiocb);
231 bitmap_free(map->affinity_mask.bits);
233 zstd_fini(&map->zstd_data);
236 if (map->data != NULL) {
237 munmap(map->data, mmap__mmap_len(map));
238 map->data = NULL;
240 auxtrace_mmap__munmap(&map->auxtrace_mmap);
256 __set_bit(cpu.cpu, mask->bits);
263 map->affinity_mask.nbits = cpu__max_cpu().cpu;
264 map->affinity_mask.bits = bitmap_zalloc(map->affinity_mask.nbits);
265 if (!map->affinity_mask.bits)
266 return -1;
268 if (mp->affinity == PERF_AFFINITY_NODE && cpu__max_node() > 1)
269 build_node_mask(cpu__get_node(map->core.cpu), &map->affinity_mask);
270 else if (mp->affinity == PERF_AFFINITY_CPU)
271 __set_bit(map->core.cpu.cpu, map->affinity_mask.bits);
278 if (perf_mmap__mmap(&map->core, &mp->core, fd, cpu)) {
281 return -1;
284 if (mp->affinity != PERF_AFFINITY_SYS &&
286 pr_debug2("failed to alloc mmap affinity mask, error %d\n",
288 return -1;
292 mmap_cpu_mask__scnprintf(&map->affinity_mask, "mmap");
294 map->core.flush = mp->flush;
296 if (zstd_init(&map->zstd_data, mp->comp_level)) {
298 return -1;
301 if (mp->comp_level && !perf_mmap__aio_enabled(map)) {
302 map->data = mmap(NULL, mmap__mmap_len(map), PROT_READ|PROT_WRITE,
304 if (map->data == MAP_FAILED) {
307 map->data = NULL;
308 return -1;
312 if (auxtrace_mmap__mmap(&map->auxtrace_mmap,
313 &mp->auxtrace_mp, map->core.base, fd))
314 return -1;
322 u64 head = perf_mmap__read_head(&md->core);
323 unsigned char *data = md->core.base + page_size;
328 rc = perf_mmap__read_init(&md->core);
330 return (rc == -EAGAIN) ? 1 : -1;
332 size = md->core.end - md->core.start;
334 if ((md->core.start & md->core.mask) + size != (md->core.end & md->core.mask)) {
335 buf = &data[md->core.start & md->core.mask];
336 size = md->core.mask + 1 - (md->core.start & md->core.mask);
337 md->core.start += size;
340 rc = -1;
345 buf = &data[md->core.start & md->core.mask];
346 size = md->core.end - md->core.start;
347 md->core.start += size;
350 rc = -1;
354 md->core.prev = head;
355 perf_mmap__consume(&md->core);