Home
last modified time | relevance | path

Searched refs:cpu (Results 1 – 25 of 2962) sorted by relevance

12345678910>>...119

/linux/drivers/cpufreq/
H A Dintel_pstate.c21 #include <linux/cpu.h>
33 #include <asm/cpu.h>
100 * to account for cpu idle period
136 * Stores the per cpu model P state limits and current P state.
192 * @cpu: CPU number for this instance data
230 int cpu; member
279 * @get_cpu_scaling: Get frequency scaling factor for a given cpu
288 int (*get_max)(int cpu);
289 int (*get_max_physical)(int cpu);
290 int (*get_min)(int cpu);
362 intel_pstate_set_itmt_prio(int cpu) intel_pstate_set_itmt_prio() argument
405 intel_pstate_get_cppc_guaranteed(int cpu) intel_pstate_get_cppc_guaranteed() argument
420 intel_pstate_cppc_get_scaling(int cpu) intel_pstate_cppc_get_scaling() argument
437 intel_pstate_set_itmt_prio(int cpu) intel_pstate_set_itmt_prio() argument
444 struct cpudata *cpu; intel_pstate_init_acpi_perf_limits() local
500 struct cpudata *cpu; intel_pstate_exit_perf_limits() local
524 intel_pstate_get_cppc_guaranteed(int cpu) intel_pstate_get_cppc_guaranteed() argument
529 intel_pstate_cppc_get_scaling(int cpu) intel_pstate_cppc_get_scaling() argument
535 intel_pstate_freq_to_hwp_rel(struct cpudata * cpu,int freq,unsigned int relation) intel_pstate_freq_to_hwp_rel() argument
554 intel_pstate_freq_to_hwp(struct cpudata * cpu,int freq) intel_pstate_freq_to_hwp() argument
570 intel_pstate_hybrid_hwp_adjust(struct cpudata * cpu) intel_pstate_hybrid_hwp_adjust() argument
618 struct cpudata *cpu = all_cpu_data[0]; min_perf_pct_min() local
720 intel_pstate_set_epp(struct cpudata * cpu,u32 epp) intel_pstate_set_epp() argument
797 struct cpudata *cpu = all_cpu_data[policy->cpu]; store_energy_performance_preference() local
882 struct cpudata *cpu = all_cpu_data[policy->cpu]; show_base_frequency() local
917 hybrid_get_cpu_type(unsigned int cpu) hybrid_get_cpu_type() argument
953 hybrid_has_l3(unsigned int cpu) hybrid_has_l3() argument
992 hybrid_register_perf_domain(unsigned int cpu) hybrid_register_perf_domain() argument
1022 unsigned int cpu; hybrid_register_all_perf_domains() local
1028 hybrid_update_perf_domain(struct cpudata * cpu) hybrid_update_perf_domain() argument
1034 hybrid_register_perf_domain(unsigned int cpu) hybrid_register_perf_domain() argument
1036 hybrid_update_perf_domain(struct cpudata * cpu) hybrid_update_perf_domain() argument
1039 hybrid_set_cpu_capacity(struct cpudata * cpu) hybrid_set_cpu_capacity() argument
1059 hybrid_get_capacity_perf(struct cpudata * cpu) hybrid_get_capacity_perf() argument
1074 struct cpudata *cpu = all_cpu_data[cpunum]; hybrid_set_capacity_of_cpus() local
1088 struct cpudata *cpu = all_cpu_data[cpunum]; hybrid_update_cpu_capacity_scaling() local
1186 __intel_pstate_get_hwp_cap(struct cpudata * cpu) __intel_pstate_get_hwp_cap() argument
1196 intel_pstate_get_hwp_cap(struct cpudata * cpu) intel_pstate_get_hwp_cap() argument
1214 hybrid_update_capacity(struct cpudata * cpu) hybrid_update_capacity() argument
1259 intel_pstate_hwp_set(unsigned int cpu) intel_pstate_hwp_set() argument
1323 intel_pstate_hwp_offline(struct cpudata * cpu) intel_pstate_hwp_offline() argument
1411 intel_pstate_hwp_reenable(struct cpudata * cpu) intel_pstate_hwp_reenable() argument
1419 struct cpudata *cpu = all_cpu_data[policy->cpu]; intel_pstate_suspend() local
1433 struct cpudata *cpu = all_cpu_data[policy->cpu]; intel_pstate_resume() local
1459 int cpu; intel_pstate_update_policies() local
1479 intel_pstate_update_max_freq(int cpu) intel_pstate_update_max_freq() argument
1501 int cpu; intel_pstate_update_limits_for_all() local
1551 struct cpudata *cpu; show_turbo_pct() local
1573 struct cpudata *cpu; show_num_pstates() local
1631 struct cpudata *cpu = all_cpu_data[0]; store_no_turbo() local
1647 update_cpu_qos_request(int cpu,enum freq_qos_req_type type) update_cpu_qos_request() argument
2155 core_get_min_pstate(int cpu) core_get_min_pstate() argument
2163 core_get_max_pstate_physical(int cpu) core_get_max_pstate_physical() argument
2171 core_get_tdp_ratio(int cpu,u64 plat_info) core_get_tdp_ratio() argument
2204 core_get_max_pstate(int cpu) core_get_max_pstate() argument
2239 core_get_turbo_pstate(int cpu) core_get_turbo_pstate() argument
2262 knl_get_turbo_pstate(int cpu) knl_get_turbo_pstate() argument
2275 hwp_get_cpu_scaling(int cpu) hwp_get_cpu_scaling() argument
2300 intel_pstate_set_pstate(struct cpudata * cpu,int pstate) intel_pstate_set_pstate() argument
2313 intel_pstate_set_min_pstate(struct cpudata * cpu) intel_pstate_set_min_pstate() argument
2318 intel_pstate_get_cpu_pstates(struct cpudata * cpu) intel_pstate_get_cpu_pstates() argument
2369 intel_pstate_hwp_boost_up(struct cpudata * cpu) intel_pstate_hwp_boost_up() argument
2416 intel_pstate_hwp_boost_down(struct cpudata * cpu) intel_pstate_hwp_boost_down() argument
2432 intel_pstate_update_util_hwp_local(struct cpudata * cpu,u64 time) intel_pstate_update_util_hwp_local() argument
2464 struct cpudata *cpu = container_of(data, struct cpudata, update_util); intel_pstate_update_util_hwp() local
2472 intel_pstate_calc_avg_perf(struct cpudata * cpu) intel_pstate_calc_avg_perf() argument
2479 intel_pstate_sample(struct cpudata * cpu,u64 time) intel_pstate_sample() argument
2521 get_avg_frequency(struct cpudata * cpu) get_avg_frequency() argument
2526 get_avg_pstate(struct cpudata * cpu) get_avg_pstate() argument
2532 get_target_pstate(struct cpudata * cpu) get_target_pstate() argument
2567 intel_pstate_prepare_request(struct cpudata * cpu,int pstate) intel_pstate_prepare_request() argument
2575 intel_pstate_update_pstate(struct cpudata * cpu,int pstate) intel_pstate_update_pstate() argument
2584 intel_pstate_adjust_pstate(struct cpudata * cpu) intel_pstate_adjust_pstate() argument
2610 struct cpudata *cpu = container_of(data, struct cpudata, update_util); intel_pstate_update_util() local
2745 struct cpudata *cpu; intel_pstate_init_cpu() local
2787 struct cpudata *cpu = all_cpu_data[cpu_num]; intel_pstate_set_update_util_hook() local
2804 intel_pstate_clear_update_util_hook(unsigned int cpu) intel_pstate_clear_update_util_hook() argument
2816 intel_pstate_get_max_freq(struct cpudata * cpu) intel_pstate_get_max_freq() argument
2822 intel_pstate_update_perf_limits(struct cpudata * cpu,unsigned int policy_min,unsigned int policy_max) intel_pstate_update_perf_limits() argument
2887 struct cpudata *cpu; intel_pstate_set_policy() local
2936 intel_pstate_adjust_policy_max(struct cpudata * cpu,struct cpufreq_policy_data * policy) intel_pstate_adjust_policy_max() argument
2948 intel_pstate_verify_cpu_policy(struct cpudata * cpu,struct cpufreq_policy_data * policy) intel_pstate_verify_cpu_policy() argument
2974 struct cpudata *cpu = all_cpu_data[policy->cpu]; intel_cpufreq_cpu_offline() local
2999 struct cpudata *cpu = all_cpu_data[policy->cpu]; intel_pstate_cpu_online() local
3035 struct cpudata *cpu; __intel_pstate_cpu_init() local
3076 struct cpudata *cpu = all_cpu_data[policy->cpu]; intel_pstate_cpu_init() local
3100 struct cpudata *cpu = all_cpu_data[policy->cpu]; intel_cpufreq_verify_policy() local
3124 intel_cpufreq_trace(struct cpudata * cpu,unsigned int trace_type,int old_pstate) intel_cpufreq_trace() argument
3146 intel_cpufreq_hwp_update(struct cpudata * cpu,u32 min,u32 max,u32 desired,bool fast_switch) intel_cpufreq_hwp_update() argument
3170 intel_cpufreq_perf_ctl_update(struct cpudata * cpu,u32 target_pstate,bool fast_switch) intel_cpufreq_perf_ctl_update() argument
3184 struct cpudata *cpu = all_cpu_data[policy->cpu]; intel_cpufreq_update_pstate() local
3210 struct cpudata *cpu = all_cpu_data[policy->cpu]; intel_cpufreq_target() local
3232 struct cpudata *cpu = all_cpu_data[policy->cpu]; intel_cpufreq_fast_switch() local
3247 struct cpudata *cpu = all_cpu_data[cpunum]; intel_cpufreq_adjust_perf() local
3290 struct cpudata *cpu; intel_cpufreq_cpu_init() local
3379 struct cpudata *cpu = all_cpu_data[policy->cpu]; intel_cpufreq_suspend() local
3414 unsigned int cpu; intel_pstate_driver_cleanup() local
[all...]
/linux/drivers/base/
H A Darch_topology.c74 int cpu; in topology_set_scale_freq_source() local
85 for_each_cpu(cpu, cpus) { in topology_set_scale_freq_source()
86 sfd = rcu_dereference(*per_cpu_ptr(&sft_data, cpu)); in topology_set_scale_freq_source()
90 rcu_assign_pointer(per_cpu(sft_data, cpu), data); in topology_set_scale_freq_source()
91 cpumask_set_cpu(cpu, &scale_freq_counters_mask); in topology_set_scale_freq_source()
105 int cpu; in topology_clear_scale_freq_source() local
109 for_each_cpu(cpu, cpus) { in topology_clear_scale_freq_source()
110 sfd = rcu_dereference(*per_cpu_ptr(&sft_data, cpu)); in topology_clear_scale_freq_source()
113 rcu_assign_pointer(per_cpu(sft_data, cpu), NULL); in topology_clear_scale_freq_source()
114 cpumask_clear_cpu(cpu, &scale_freq_counters_mask); in topology_clear_scale_freq_source()
[all …]
H A Dcacheinfo.c26 #define ci_cacheinfo(cpu) (&per_cpu(ci_cpu_cacheinfo, cpu)) argument
27 #define cache_leaves(cpu) (ci_cacheinfo(cpu)->num_leaves) argument
28 #define per_cpu_cacheinfo(cpu) (ci_cacheinfo(cpu)->info_list) argument
29 #define per_cpu_cacheinfo_idx(cpu, idx) \ argument
30 (per_cpu_cacheinfo(cpu) + (idx))
35 struct cpu_cacheinfo *get_cpu_cacheinfo(unsigned int cpu) in get_cpu_cacheinfo() argument
37 return ci_cacheinfo(cpu); in get_cpu_cacheinfo()
58 bool last_level_cache_is_valid(unsigned int cpu) in last_level_cache_is_valid() argument
62 if (!cache_leaves(cpu) || !per_cpu_cacheinfo(cpu)) in last_level_cache_is_valid()
65 llc = per_cpu_cacheinfo_idx(cpu, cache_leaves(cpu) - 1); in last_level_cache_is_valid()
[all …]
/linux/arch/powerpc/kernel/
H A Dsmp.c278 void smp_muxed_ipi_set_message(int cpu, int msg) in smp_muxed_ipi_set_message() argument
280 struct cpu_messages *info = &per_cpu(ipi_message, cpu); in smp_muxed_ipi_set_message()
290 void smp_muxed_ipi_message_pass(int cpu, int msg) in smp_muxed_ipi_message_pass() argument
292 smp_muxed_ipi_set_message(cpu, msg); in smp_muxed_ipi_message_pass()
298 smp_ops->cause_ipi(cpu); in smp_muxed_ipi_message_pass()
352 static inline void do_message_pass(int cpu, int msg) in do_message_pass() argument
355 smp_ops->message_pass(cpu, msg); in do_message_pass()
358 smp_muxed_ipi_message_pass(cpu, msg); in do_message_pass()
362 void arch_smp_send_reschedule(int cpu) in arch_smp_send_reschedule() argument
365 do_message_pass(cpu, PPC_MSG_RESCHEDULE); in arch_smp_send_reschedule()
[all …]
H A Dtau_6xx.c55 static void set_thresholds(unsigned long cpu) in set_thresholds() argument
60 mtspr(SPRN_THRM1, THRM1_THRES(tau[cpu].low) | THRM1_V | maybe_tie | THRM1_TID); in set_thresholds()
63 mtspr(SPRN_THRM2, THRM1_THRES(tau[cpu].high) | THRM1_V | maybe_tie); in set_thresholds()
66 static void TAUupdate(int cpu) in TAUupdate() argument
77 if (tau[cpu].low >= step_size) { in TAUupdate()
78 tau[cpu].low -= step_size; in TAUupdate()
79 tau[cpu].high -= (step_size - window_expand); in TAUupdate()
81 tau[cpu].grew = 1; in TAUupdate()
88 if (tau[cpu].high <= 127 - step_size) { in TAUupdate()
89 tau[cpu].low += (step_size - window_expand); in TAUupdate()
[all …]
H A Dwatchdog.c149 int cpu = raw_smp_processor_id(); in wd_lockup_ipi() local
152 pr_emerg("CPU %d Hard LOCKUP\n", cpu); in wd_lockup_ipi()
154 cpu, tb, per_cpu(wd_timer_tb, cpu), in wd_lockup_ipi()
155 tb_to_ns(tb - per_cpu(wd_timer_tb, cpu)) / 1000000); in wd_lockup_ipi()
183 static bool set_cpu_stuck(int cpu) in set_cpu_stuck() argument
185 cpumask_set_cpu(cpu, &wd_smp_cpus_stuck); in set_cpu_stuck()
186 cpumask_clear_cpu(cpu, &wd_smp_cpus_pending); in set_cpu_stuck()
201 static void watchdog_smp_panic(int cpu) in watchdog_smp_panic() argument
214 if (cpumask_test_cpu(cpu, &wd_smp_cpus_pending)) in watchdog_smp_panic()
221 if (c == cpu) in watchdog_smp_panic()
[all …]
/linux/arch/arm/boot/dts/intel/axm/
H A Daxm5516-cpus.dtsi13 cpu-map {
16 cpu = <&CPU0>;
19 cpu = <&CPU1>;
22 cpu = <&CPU2>;
25 cpu = <&CPU3>;
30 cpu = <&CPU4>;
33 cpu = <&CPU5>;
36 cpu = <&CPU6>;
39 cpu = <&CPU7>;
44 cpu = <&CPU8>;
[all …]
/linux/include/linux/
H A Dtopology.h91 static inline int cpu_to_node(int cpu) in cpu_to_node() argument
93 return per_cpu(numa_node, cpu); in cpu_to_node()
105 static inline void set_cpu_numa_node(int cpu, int node) in set_cpu_numa_node() argument
107 per_cpu(numa_node, cpu) = node; in set_cpu_numa_node()
148 static inline int cpu_to_mem(int cpu) in cpu_to_mem() argument
150 return per_cpu(_numa_mem_, cpu); in cpu_to_mem()
155 static inline void set_cpu_numa_mem(int cpu, int node) in set_cpu_numa_mem() argument
157 per_cpu(_numa_mem_, cpu) = node; in set_cpu_numa_mem()
172 static inline int cpu_to_mem(int cpu) in cpu_to_mem() argument
174 return cpu_to_node(cpu); in cpu_to_mem()
[all …]
H A Dcpumask.h135 static __always_inline void cpu_max_bits_warn(unsigned int cpu, unsigned int bits) in cpu_max_bits_warn() argument
138 WARN_ON_ONCE(cpu >= bits); in cpu_max_bits_warn()
143 static __always_inline unsigned int cpumask_check(unsigned int cpu) in cpumask_check() argument
145 cpu_max_bits_warn(cpu, small_cpumask_bits); in cpumask_check()
146 return cpu; in cpumask_check()
379 #define for_each_cpu(cpu, mask) \ argument
380 for_each_set_bit(cpu, cpumask_bits(mask), small_cpumask_bits)
392 #define for_each_cpu_wrap(cpu, mask, start) \ argument
393 for_each_set_bit_wrap(cpu, cpumask_bits(mask), small_cpumask_bits, start)
409 #define for_each_cpu_and(cpu, mask1, mask2) \ argument
[all …]
/linux/arch/um/kernel/
H A Dsmp.c32 void arch_smp_send_reschedule(int cpu) in arch_smp_send_reschedule() argument
34 os_send_ipi(cpu, UML_IPI_RES); in arch_smp_send_reschedule()
37 void arch_send_call_function_single_ipi(int cpu) in arch_send_call_function_single_ipi() argument
39 os_send_ipi(cpu, UML_IPI_CALL_SINGLE); in arch_send_call_function_single_ipi()
44 int cpu; in arch_send_call_function_ipi_mask() local
46 for_each_cpu(cpu, mask) in arch_send_call_function_ipi_mask()
47 os_send_ipi(cpu, UML_IPI_CALL); in arch_send_call_function_ipi_mask()
52 int cpu, me = smp_processor_id(); in smp_send_stop() local
54 for_each_online_cpu(cpu) { in smp_send_stop()
55 if (cpu == me) in smp_send_stop()
[all …]
/linux/tools/perf/tests/
H A Dtopology.c75 struct perf_cpu cpu; in check_cpu_topology()
86 * cpu is a NULL pointer. in check_cpu_topology()
104 if (!env->cpu && strncmp(env->arch, "s390", 4) && strncmp(env->arch, "aarch64", 7)) in check_cpu_topology()
111 * test if physical_package_id returns -1 for cpu from perf_cpu_map. in check_cpu_topology()
118 TEST_ASSERT_VAL("Session header CPU map not set", env->cpu); in check_cpu_topology()
121 cpu.cpu = i; in check_cpu_topology()
122 if (!perf_cpu_map__has(map, cpu)) in check_cpu_topology()
125 env->cpu[i].core_id, in check_cpu_topology()
126 env->cpu[ in check_cpu_topology()
72 struct perf_cpu cpu; check_cpu_topology() local
[all...]
/linux/tools/testing/selftests/cpu-hotplug/
H A Dcpu-on-off-test.sh27 if ! ls $SYSFS/devices/system/cpu/cpu*/online > /dev/null 2>&1; then
28 echo $msg cpu hotplug is not supported >&2
33 online_cpus=`cat $SYSFS/devices/system/cpu/online`
37 echo "$msg: since there is only one cpu: $online_cpus"
41 present_cpus=`cat $SYSFS/devices/system/cpu/present`
47 offline_cpus=`cat $SYSFS/devices/system/cpu/offline`
63 for cpu in $SYSFS/devices/system/cpu/cpu*; d
[all...]
/linux/arch/arm/mach-meson/
H A Dplatsmp.c38 static struct reset_control *meson_smp_get_core_reset(int cpu) in meson_smp_get_core_reset() argument
40 struct device_node *np = of_get_cpu_node(cpu, 0); in meson_smp_get_core_reset()
45 static void meson_smp_set_cpu_ctrl(int cpu, bool on_off) in meson_smp_set_cpu_ctrl() argument
50 val |= BIT(cpu); in meson_smp_set_cpu_ctrl()
52 val &= ~BIT(cpu); in meson_smp_set_cpu_ctrl()
116 static void meson_smp_begin_secondary_boot(unsigned int cpu) in meson_smp_begin_secondary_boot() argument
125 sram_base + MESON_SMP_SRAM_CPU_CTRL_ADDR_REG(cpu)); in meson_smp_begin_secondary_boot()
131 scu_cpu_power_enable(scu_base, cpu); in meson_smp_begin_secondary_boot()
134 static int meson_smp_finalize_secondary_boot(unsigned int cpu) in meson_smp_finalize_secondary_boot() argument
139 while (readl(sram_base + MESON_SMP_SRAM_CPU_CTRL_ADDR_REG(cpu))) { in meson_smp_finalize_secondary_boot()
[all …]
/linux/Documentation/translations/zh_CN/scheduler/
H A Dsched-bwc.rst24 达“配额”微秒的CPU时间。当cgroup中的线程可运行时,该配额以时间片段的方式被分配到每个cpu
29 它以需求为基础被转移到cpu-local“筒仓”,在每次更新中转移的数量是可调整的,被描述为“片“(时
65 配额、周期和突发是在cpu子系统内通过cgroupfs管理的。
69 :ref:`Documentation/admin-guide/cgroup-v2.rst <cgroup-v2-cpu>`.
71 - cpu.cfs_quota_us:在一个时期内补充的运行时间(微秒)。
72 - cpu.cfs_period_us:一个周期的长度(微秒)。
73 - cpu.stat: 输出节流统计数据[下面进一步解释]
74 - cpu.cfs_burst_us:最大累积运行时间(微秒)。
78 cpu.cfs_period_us=100ms
79 cpu.cfs_quota_us=-1
[all …]
/linux/arch/arm/mach-bcm/
H A Dplatsmp-brcmstb.c59 static int per_cpu_sw_state_rd(u32 cpu) in per_cpu_sw_state_rd() argument
61 sync_cache_r(SHIFT_PERCPU_PTR(&per_cpu_sw_state, per_cpu_offset(cpu))); in per_cpu_sw_state_rd()
62 return per_cpu(per_cpu_sw_state, cpu); in per_cpu_sw_state_rd()
65 static void per_cpu_sw_state_wr(u32 cpu, int val) in per_cpu_sw_state_wr() argument
68 per_cpu(per_cpu_sw_state, cpu) = val; in per_cpu_sw_state_wr()
69 sync_cache_w(SHIFT_PERCPU_PTR(&per_cpu_sw_state, per_cpu_offset(cpu))); in per_cpu_sw_state_wr()
72 static inline void per_cpu_sw_state_wr(u32 cpu, int val) { } in per_cpu_sw_state_wr() argument
75 static void __iomem *pwr_ctrl_get_base(u32 cpu) in pwr_ctrl_get_base() argument
78 base += (cpu_logical_map(cpu) * 4); in pwr_ctrl_get_base()
82 static u32 pwr_ctrl_rd(u32 cpu) in pwr_ctrl_rd() argument
[all …]
/linux/arch/powerpc/include/asm/
H A Dsmp.h35 extern int cpu_to_chip_id(int cpu);
45 void (*message_pass)(int cpu, int msg);
47 void (*cause_ipi)(int cpu);
49 int (*cause_nmi_ipi)(int cpu);
68 extern int smp_send_nmi_ipi(int cpu, void (*fn)(struct pt_regs *), u64 delay_us);
69 extern int smp_send_safe_nmi_ipi(int cpu, void (*fn)(struct pt_regs *), u64 delay_us);
79 void generic_cpu_die(unsigned int cpu);
80 void generic_set_cpu_dead(unsigned int cpu);
81 void generic_set_cpu_up(unsigned int cpu);
82 int generic_check_cpu_restart(unsigned int cpu);
[all …]
H A Dtopology.h48 static inline void update_numa_cpu_lookup_table(unsigned int cpu, int node) in update_numa_cpu_lookup_table() argument
50 numa_cpu_lookup_table[cpu] = node; in update_numa_cpu_lookup_table()
53 static inline int early_cpu_to_node(int cpu) in early_cpu_to_node() argument
57 nid = numa_cpu_lookup_table[cpu]; in early_cpu_to_node()
69 extern void map_cpu_to_node(int cpu, int node);
71 extern void unmap_cpu_from_node(unsigned long cpu);
76 static inline int early_cpu_to_node(int cpu) { return 0; } in early_cpu_to_node() argument
90 static inline void update_numa_cpu_lookup_table(unsigned int cpu, int node) {} in update_numa_cpu_lookup_table() argument
105 static inline void map_cpu_to_node(int cpu, int node) {} in map_cpu_to_node() argument
107 static inline void unmap_cpu_from_node(unsigned long cpu) {} in unmap_cpu_from_node() argument
[all …]
/linux/tools/power/cpupower/utils/
H A Dcpufreq-info.c42 if (strstr(value, "cpu ")) in count_cpus()
44 if (sscanf(value, "cpu%d ", &cpunr) != 1) in count_cpus()
51 /* cpu count starts from 0, on error return 1 (UP) */ in count_cpus()
58 unsigned int cpu, nr_cpus; in proc_cpufreq_output() local
67 for (cpu = 0; cpu < nr_cpus; cpu++) { in proc_cpufreq_output()
68 policy = cpufreq_get_policy(cpu); in proc_cpufreq_output()
72 if (cpufreq_get_hardware_limits(cpu, &min, &max)) { in proc_cpufreq_output()
79 cpu , polic in proc_cpufreq_output()
125 get_boost_mode_x86(unsigned int cpu) get_boost_mode_x86() argument
207 get_boost_mode_generic(unsigned int cpu) get_boost_mode_generic() argument
221 get_boost_mode(unsigned int cpu) get_boost_mode() argument
250 get_freq_kernel(unsigned int cpu,unsigned int human) get_freq_kernel() argument
269 get_freq_hardware(unsigned int cpu,unsigned int human) get_freq_hardware() argument
292 get_hardware_limits(unsigned int cpu,unsigned int human) get_hardware_limits() argument
315 get_driver(unsigned int cpu) get_driver() argument
329 get_policy(unsigned int cpu) get_policy() argument
351 get_available_governors(unsigned int cpu) get_available_governors() argument
374 get_affected_cpus(unsigned int cpu) get_affected_cpus() argument
395 get_related_cpus(unsigned int cpu) get_related_cpus() argument
416 get_freq_stats(unsigned int cpu,unsigned int human) get_freq_stats() argument
441 get_epp(unsigned int cpu,bool interactive) get_epp() argument
458 get_latency(unsigned int cpu,unsigned int human) get_latency() argument
481 get_perf_cap(unsigned int cpu) get_perf_cap() argument
490 debug_output_one(unsigned int cpu) debug_output_one() argument
548 unsigned int cpu = 0; cmd_freq_info() local
[all...]
/linux/arch/x86/xen/
H A Dsmp.c32 void xen_smp_intr_free(unsigned int cpu) in xen_smp_intr_free() argument
34 kfree(per_cpu(xen_resched_irq, cpu).name); in xen_smp_intr_free()
35 per_cpu(xen_resched_irq, cpu).name = NULL; in xen_smp_intr_free()
36 if (per_cpu(xen_resched_irq, cpu).irq >= 0) { in xen_smp_intr_free()
37 unbind_from_irqhandler(per_cpu(xen_resched_irq, cpu).irq, NULL); in xen_smp_intr_free()
38 per_cpu(xen_resched_irq, cpu).irq = -1; in xen_smp_intr_free()
40 kfree(per_cpu(xen_callfunc_irq, cpu).name); in xen_smp_intr_free()
41 per_cpu(xen_callfunc_irq, cpu).name = NULL; in xen_smp_intr_free()
42 if (per_cpu(xen_callfunc_irq, cpu).irq >= 0) { in xen_smp_intr_free()
43 unbind_from_irqhandler(per_cpu(xen_callfunc_irq, cpu).irq, NULL); in xen_smp_intr_free()
[all …]
H A Dsmp_pv.c59 int cpu; in cpu_bringup() local
71 cpu = smp_processor_id(); in cpu_bringup()
72 identify_secondary_cpu(cpu); in cpu_bringup()
73 set_cpu_sibling_map(cpu); in cpu_bringup()
79 notify_cpu_starting(cpu); in cpu_bringup()
81 set_cpu_online(cpu, true); in cpu_bringup()
95 void xen_smp_intr_free_pv(unsigned int cpu) in xen_smp_intr_free_pv() argument
97 kfree(per_cpu(xen_irq_work, cpu).name); in xen_smp_intr_free_pv()
98 per_cpu(xen_irq_work, cpu).name = NULL; in xen_smp_intr_free_pv()
99 if (per_cpu(xen_irq_work, cpu).irq >= 0) { in xen_smp_intr_free_pv()
[all …]
/linux/arch/arm/mach-tegra/
H A Dplatsmp.c36 static void tegra_secondary_init(unsigned int cpu) in tegra_secondary_init() argument
38 cpumask_set_cpu(cpu, &tegra_cpu_init_mask); in tegra_secondary_init()
42 static int tegra20_boot_secondary(unsigned int cpu, struct task_struct *idle) in tegra20_boot_secondary() argument
44 cpu = cpu_logical_map(cpu); in tegra20_boot_secondary()
54 tegra_put_cpu_in_reset(cpu); in tegra20_boot_secondary()
62 flowctrl_write_cpu_halt(cpu, 0); in tegra20_boot_secondary()
64 tegra_enable_cpu_clock(cpu); in tegra20_boot_secondary()
65 flowctrl_write_cpu_csr(cpu, 0); /* Clear flow controller CSR. */ in tegra20_boot_secondary()
66 tegra_cpu_out_of_reset(cpu); in tegra20_boot_secondary()
70 static int tegra30_boot_secondary(unsigned int cpu, struct task_struct *idle) in tegra30_boot_secondary() argument
[all …]
/linux/tools/perf/util/
H A Dcpumap.c75 * Special treatment for -1, which is not real cpu number, in cpu_map__from_entries()
79 if (data->cpus_data.cpu[i] == (u16) -1) { in cpu_map__from_entries()
80 RC_CHK_ACCESS(map)->map[i].cpu = -1; in cpu_map__from_entries()
81 } else if (data->cpus_data.cpu[i] < INT16_MAX) { in cpu_map__from_entries()
82 RC_CHK_ACCESS(map)->map[i].cpu = (int16_t) data->cpus_data.cpu[i]; in cpu_map__from_entries()
84 pr_err("Invalid cpumap entry %u\n", data->cpus_data.cpu[i]); in cpu_map__from_entries()
110 int cpu; in cpu_map__from_mask() local
113 for_each_set_bit(cpu, local_copy, 64) { in cpu_map__from_mask()
114 if (cpu in cpu_map__from_mask()
140 for (int cpu = data->range_cpu_data.start_cpu; cpu <= data->range_cpu_data.end_cpu; cpu_map__from_range() local
206 cpu__get_topology_int(int cpu,const char * name,int * value) cpu__get_topology_int() argument
216 cpu__get_socket_id(struct perf_cpu cpu) cpu__get_socket_id() argument
222 aggr_cpu_id__socket(struct perf_cpu cpu,void * data __maybe_unused) aggr_cpu_id__socket() argument
258 struct perf_cpu cpu; cpu_aggr_map__new() local
300 cpu__get_die_id(struct perf_cpu cpu) cpu__get_die_id() argument
307 aggr_cpu_id__die(struct perf_cpu cpu,void * data) aggr_cpu_id__die() argument
330 cpu__get_cluster_id(struct perf_cpu cpu) cpu__get_cluster_id() argument
337 aggr_cpu_id__cluster(struct perf_cpu cpu,void * data) aggr_cpu_id__cluster() argument
354 cpu__get_core_id(struct perf_cpu cpu) cpu__get_core_id() argument
360 aggr_cpu_id__core(struct perf_cpu cpu,void * data) aggr_cpu_id__core() argument
379 aggr_cpu_id__cpu(struct perf_cpu cpu,void * data) aggr_cpu_id__cpu() argument
393 aggr_cpu_id__node(struct perf_cpu cpu,void * data __maybe_unused) aggr_cpu_id__node() argument
401 aggr_cpu_id__global(struct perf_cpu cpu,void * data __maybe_unused) aggr_cpu_id__global() argument
544 cpu__get_node(struct perf_cpu cpu) cpu__get_node() argument
577 unsigned int cpu, mem; cpu__setup_cpunode_map() local
635 struct perf_cpu cpu = { .cpu = INT16_MAX }; cpu_map__snprint() local
704 for (int cpu = last_cpu.cpu / 4 * 4; cpu >= 0; cpu -= 4) { cpu_map__snprint_mask() local
[all...]
/linux/arch/s390/kernel/
H A Dsmp.c167 int cpu; in pcpu_find_address() local
169 for_each_cpu(cpu, mask) in pcpu_find_address()
170 if (per_cpu(pcpu_devices, cpu).address == address) in pcpu_find_address()
171 return &per_cpu(pcpu_devices, cpu); in pcpu_find_address()
183 static int pcpu_alloc_lowcore(struct pcpu *pcpu, int cpu) in pcpu_alloc_lowcore() argument
199 lc->cpu_nr = cpu; in pcpu_alloc_lowcore()
200 lc->spinlock_lockval = arch_spin_lockval(cpu); in pcpu_alloc_lowcore()
207 if (abs_lowcore_map(cpu, lc, true)) in pcpu_alloc_lowcore()
209 lowcore_ptr[cpu] = lc; in pcpu_alloc_lowcore()
223 static void pcpu_free_lowcore(struct pcpu *pcpu, int cpu) in pcpu_free_lowcore() argument
[all …]
/linux/kernel/
H A Dsmpboot.c28 struct task_struct *idle_thread_get(unsigned int cpu) in idle_thread_get() argument
30 struct task_struct *tsk = per_cpu(idle_threads, cpu); in idle_thread_get()
48 static __always_inline void idle_init(unsigned int cpu) in idle_init() argument
50 struct task_struct *tsk = per_cpu(idle_threads, cpu); in idle_init()
53 tsk = fork_idle(cpu); in idle_init()
55 pr_err("SMP: fork_idle() failed for CPU %u\n", cpu); in idle_init()
57 per_cpu(idle_threads, cpu) = tsk; in idle_init()
66 unsigned int cpu, boot_cpu; in idle_threads_init() local
70 for_each_possible_cpu(cpu) { in idle_threads_init()
71 if (cpu != boot_cpu) in idle_threads_init()
[all …]
/linux/arch/arm64/boot/dts/amd/
H A Delba-16core.dtsi11 cpu-map {
13 core0 { cpu = <&cpu0>; };
14 core1 { cpu = <&cpu1>; };
15 core2 { cpu = <&cpu2>; };
16 core3 { cpu = <&cpu3>; };
20 core0 { cpu = <&cpu4>; };
21 core1 { cpu = <&cpu5>; };
22 core2 { cpu = <&cpu6>; };
23 core3 { cpu = <&cpu7>; };
27 core0 { cpu = <&cpu8>; };
[all …]

12345678910>>...119