Lines Matching full:thread
70 int idx, thread; in perf_evsel__alloc_fd() local
73 for (thread = 0; thread < nthreads; thread++) { in perf_evsel__alloc_fd()
74 int *fd = FD(evsel, idx, thread); in perf_evsel__alloc_fd()
100 static int get_group_fd(struct perf_evsel *evsel, int cpu_map_idx, int thread, int *group_fd) in get_group_fd() argument
117 fd = FD(leader, cpu_map_idx, thread); in get_group_fd()
130 int idx, thread, err = 0; in perf_evsel__open() local
161 for (thread = 0; thread < threads->nr; thread++) { in perf_evsel__open()
164 evsel_fd = FD(evsel, idx, thread); in perf_evsel__open()
170 err = get_group_fd(evsel, idx, thread, &group_fd); in perf_evsel__open()
175 threads->map[thread].pid, in perf_evsel__open()
195 int thread; in perf_evsel__close_fd_cpu() local
197 for (thread = 0; thread < xyarray__max_y(evsel->fd); ++thread) { in perf_evsel__close_fd_cpu()
198 int *fd = FD(evsel, cpu_map_idx, thread); in perf_evsel__close_fd_cpu()
238 int idx, thread; in perf_evsel__munmap() local
244 for (thread = 0; thread < xyarray__max_y(evsel->fd); thread++) { in perf_evsel__munmap()
245 int *fd = FD(evsel, idx, thread); in perf_evsel__munmap()
250 perf_mmap__munmap(MMAP(evsel, idx, thread)); in perf_evsel__munmap()
260 int ret, idx, thread; in perf_evsel__mmap() local
273 for (thread = 0; thread < xyarray__max_y(evsel->fd); thread++) { in perf_evsel__mmap()
274 int *fd = FD(evsel, idx, thread); in perf_evsel__mmap()
281 map = MMAP(evsel, idx, thread); in perf_evsel__mmap()
295 void *perf_evsel__mmap_base(struct perf_evsel *evsel, int cpu_map_idx, int thread) in perf_evsel__mmap_base() argument
297 int *fd = FD(evsel, cpu_map_idx, thread); in perf_evsel__mmap_base()
299 if (fd == NULL || *fd < 0 || MMAP(evsel, cpu_map_idx, thread) == NULL) in perf_evsel__mmap_base()
302 return MMAP(evsel, cpu_map_idx, thread)->base; in perf_evsel__mmap_base()
335 int thread, struct perf_counts_values *count) in perf_evsel__read_group() argument
338 int *fd = FD(evsel, cpu_map_idx, thread); in perf_evsel__read_group()
400 int perf_evsel__read(struct perf_evsel *evsel, int cpu_map_idx, int thread, in perf_evsel__read() argument
404 int *fd = FD(evsel, cpu_map_idx, thread); in perf_evsel__read()
414 return perf_evsel__read_group(evsel, cpu_map_idx, thread, count); in perf_evsel__read()
416 if (MMAP(evsel, cpu_map_idx, thread) && in perf_evsel__read()
418 !perf_mmap__read_self(MMAP(evsel, cpu_map_idx, thread), count)) in perf_evsel__read()
429 int cpu_map_idx, int thread) in perf_evsel__ioctl() argument
431 int *fd = FD(evsel, cpu_map_idx, thread); in perf_evsel__ioctl()
443 int thread; in perf_evsel__run_ioctl() local
445 for (thread = 0; thread < xyarray__max_y(evsel->fd); thread++) { in perf_evsel__run_ioctl()
446 int err = perf_evsel__ioctl(evsel, ioc, arg, cpu_map_idx, thread); in perf_evsel__run_ioctl()
460 int perf_evsel__enable_thread(struct perf_evsel *evsel, int thread) in perf_evsel__enable_thread() argument
467 err = perf_evsel__ioctl(evsel, PERF_EVENT_IOC_ENABLE, NULL, idx, thread); in perf_evsel__enable_thread()