/linux-6.14.4/tools/lib/perf/include/perf/ |
D | cpumap.h | 57 LIBPERF_API int perf_cpu_map__nr(const struct perf_cpu_map *cpus); 89 (idx) < perf_cpu_map__nr(cpus); \ 94 (idx) < perf_cpu_map__nr(cpus); \ 99 for ((idx) = 0; (idx) < perf_cpu_map__nr(cpus); (idx)++)
|
/linux-6.14.4/tools/perf/tests/ |
D | cpumap.c | 40 TEST_ASSERT_VAL("wrong nr", perf_cpu_map__nr(map) == 20); in process_event_mask() 68 TEST_ASSERT_VAL("wrong nr", perf_cpu_map__nr(map) == 2); in process_event_cpus() 94 TEST_ASSERT_VAL("wrong nr", perf_cpu_map__nr(map) == 256); in process_event_range_cpus() 166 TEST_ASSERT_VAL("failed to merge map: bad nr", perf_cpu_map__nr(a) == nr); in __test__cpu_map_merge() 214 TEST_ASSERT_EQUAL("failed to intersect map: bad nr", perf_cpu_map__nr(c), nr); in __test__cpu_map_intersect()
|
D | event_update.c | 72 TEST_ASSERT_VAL("wrong cpus", perf_cpu_map__nr(map) == 3); in process_event_cpus()
|
/linux-6.14.4/tools/perf/bench/ |
D | futex-wake.c | 116 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, size, cpuset); in block_threads() 167 params.nthreads = perf_cpu_map__nr(cpu); in bench_futex_wake()
|
D | evlist-open-close.c | 74 cnt += evsel->core.threads->nr * perf_cpu_map__nr(evsel->core.cpus); in evlist__count_evsel_fds() 154 printf(" Number of cpus:\t%d\n", perf_cpu_map__nr(evlist->core.user_requested_cpus)); in bench_evlist_open_close__run()
|
D | futex-lock-pi.c | 148 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, size, cpuset); in create_threads() 190 params.nthreads = perf_cpu_map__nr(cpu); in bench_futex_lock_pi()
|
D | futex-hash.c | 156 params.nthreads = perf_cpu_map__nr(cpu); in bench_futex_hash() 190 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, size, cpuset); in bench_futex_hash()
|
D | futex-requeue.c | 143 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, size, cpuset); in block_threads() 192 params.nthreads = perf_cpu_map__nr(cpu); in bench_futex_requeue()
|
D | futex-wake-parallel.c | 167 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, size, cpuset); in block_threads() 272 params.nthreads = perf_cpu_map__nr(cpu); in bench_futex_wake_parallel()
|
D | epoll-ctl.c | 264 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, in do_threads() 350 nthreads = perf_cpu_map__nr(cpu); in bench_epoll_ctl()
|
D | epoll-wait.c | 353 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, in do_threads() 474 nthreads = perf_cpu_map__nr(cpu) - 1; in bench_epoll_wait()
|
/linux-6.14.4/tools/lib/perf/ |
D | libperf.map | 9 perf_cpu_map__nr;
|
D | evlist.c | 341 int nr_cpus = perf_cpu_map__nr(evlist->all_cpus); in perf_evlist__alloc_pollfd() 567 int nr_cpus = perf_cpu_map__nr(evlist->all_cpus); in mmap_per_thread() 609 int nr_cpus = perf_cpu_map__nr(evlist->all_cpus); in mmap_per_cpu() 641 nr_mmaps = perf_cpu_map__nr(evlist->all_cpus); in perf_evlist__nr_mmaps()
|
D | evsel.c | 146 perf_evsel__alloc_fd(evsel, perf_cpu_map__nr(cpus), threads->nr) < 0) in perf_evsel__open() 493 for (i = 0; i < perf_cpu_map__nr(evsel->cpus) && !err; i++) in perf_evsel__apply_filter()
|
/linux-6.14.4/tools/perf/util/ |
D | cpumap.c | 241 struct cpu_aggr_map *c = cpu_aggr_map__empty_new(perf_cpu_map__nr(cpus)); in cpu_aggr_map__new() 265 if (c->nr != perf_cpu_map__nr(cpus)) { in cpu_aggr_map__new() 608 for (i = 0; i < perf_cpu_map__nr(map) + 1; i++) { in cpu_map__snprint() 610 bool last = i == perf_cpu_map__nr(map); in cpu_map__snprint()
|
D | counts.c | 63 evsel->counts = perf_counts__new(perf_cpu_map__nr(cpus), nthreads); in evsel__alloc_counts()
|
D | cpumap.h | 71 return perf_cpu_map__nr(cpus) == 1 && perf_cpu_map__cpu(cpus, 0).cpu == -1; in cpu_map__is_dummy()
|
D | tool_pmu.c | 208 evsel->start_times = xyarray__new(perf_cpu_map__nr(cpus), in evsel__tool_pmu_prepare_open() 357 *result = perf_cpu_map__nr(online); in tool_pmu__read_event()
|
D | top.c | 97 nr_cpus = perf_cpu_map__nr(top->evlist->core.user_requested_cpus); in perf_top__header_snprintf()
|
D | bpf_ftrace.c | 45 ncpus = perf_cpu_map__nr(ftrace->evlist->core.user_requested_cpus); in perf_ftrace__latency_prepare_bpf()
|
D | bpf_counter.c | 381 *filter_entry_cnt = perf_cpu_map__nr(evsel__cpus(evsel)); in bperf_check_target() 614 num_cpu = perf_cpu_map__nr(all_cpu_map); in bperf_sync_counters()
|
D | bpf_off_cpu.c | 152 ncpus = perf_cpu_map__nr(evlist->core.user_requested_cpus); in off_cpu_prepare()
|
/linux-6.14.4/tools/lib/perf/tests/ |
D | test-evlist.c | 77 for (idx = 0; idx < perf_cpu_map__nr(cpus); idx++) { in test_stat_cpu() 414 __T("failed count", count >= perf_cpu_map__nr(cpus)); in test_mmap_cpus()
|
D | test-evsel.c | 39 for (idx = 0; idx < perf_cpu_map__nr(cpus); idx++) { in test_stat_cpu()
|
/linux-6.14.4/tools/perf/arch/arm64/util/ |
D | arm-spe.c | 76 ARM_SPE_CPU_PRIV_MAX * perf_cpu_map__nr(cpu_map); in arm_spe_info_priv_size() 152 auxtrace_info->priv[ARM_SPE_CPUS_NUM] = perf_cpu_map__nr(cpu_map); in arm_spe_info_fill()
|