| /linux/drivers/cpufreq/ |
| H A D | intel_pstate.c | 21 #include <linux/cpu.h> 33 #include <asm/cpu.h> 100 * to account for cpu idle period 136 * Stores the per cpu model P state limits and current P state. 192 * @cpu: CPU number for this instance data 230 int cpu; member 279 * @get_cpu_scaling: Get frequency scaling factor for a given cpu 288 int (*get_max)(int cpu); 289 int (*get_max_physical)(int cpu); 290 int (*get_min)(int cpu); 362 intel_pstate_set_itmt_prio(int cpu) intel_pstate_set_itmt_prio() argument 405 intel_pstate_get_cppc_guaranteed(int cpu) intel_pstate_get_cppc_guaranteed() argument 420 intel_pstate_cppc_get_scaling(int cpu) intel_pstate_cppc_get_scaling() argument 437 intel_pstate_set_itmt_prio(int cpu) intel_pstate_set_itmt_prio() argument 444 struct cpudata *cpu; intel_pstate_init_acpi_perf_limits() local 500 struct cpudata *cpu; intel_pstate_exit_perf_limits() local 524 intel_pstate_get_cppc_guaranteed(int cpu) intel_pstate_get_cppc_guaranteed() argument 529 intel_pstate_cppc_get_scaling(int cpu) intel_pstate_cppc_get_scaling() argument 535 intel_pstate_freq_to_hwp_rel(struct cpudata * cpu,int freq,unsigned int relation) intel_pstate_freq_to_hwp_rel() argument 554 intel_pstate_freq_to_hwp(struct cpudata * cpu,int freq) intel_pstate_freq_to_hwp() argument 570 intel_pstate_hybrid_hwp_adjust(struct cpudata * cpu) intel_pstate_hybrid_hwp_adjust() argument 618 struct cpudata *cpu = all_cpu_data[0]; min_perf_pct_min() local 720 intel_pstate_set_epp(struct cpudata * cpu,u32 epp) intel_pstate_set_epp() argument 797 struct cpudata *cpu = all_cpu_data[policy->cpu]; store_energy_performance_preference() local 882 struct cpudata *cpu = all_cpu_data[policy->cpu]; show_base_frequency() local 917 hybrid_get_cpu_type(unsigned int cpu) hybrid_get_cpu_type() argument 953 hybrid_has_l3(unsigned int cpu) hybrid_has_l3() argument 992 hybrid_register_perf_domain(unsigned int cpu) hybrid_register_perf_domain() argument 1022 unsigned int cpu; hybrid_register_all_perf_domains() local 1028 hybrid_update_perf_domain(struct cpudata * cpu) hybrid_update_perf_domain() argument 1034 hybrid_register_perf_domain(unsigned int cpu) hybrid_register_perf_domain() argument 1036 hybrid_update_perf_domain(struct cpudata * cpu) hybrid_update_perf_domain() argument 1039 hybrid_set_cpu_capacity(struct cpudata * cpu) hybrid_set_cpu_capacity() argument 1059 hybrid_get_capacity_perf(struct cpudata * cpu) hybrid_get_capacity_perf() argument 1074 struct cpudata *cpu = all_cpu_data[cpunum]; hybrid_set_capacity_of_cpus() local 1088 struct cpudata *cpu = all_cpu_data[cpunum]; hybrid_update_cpu_capacity_scaling() local 1186 __intel_pstate_get_hwp_cap(struct cpudata * cpu) __intel_pstate_get_hwp_cap() argument 1196 intel_pstate_get_hwp_cap(struct cpudata * cpu) intel_pstate_get_hwp_cap() argument 1214 hybrid_update_capacity(struct cpudata * cpu) hybrid_update_capacity() argument 1259 intel_pstate_hwp_set(unsigned int cpu) intel_pstate_hwp_set() argument 1323 intel_pstate_hwp_offline(struct cpudata * cpu) intel_pstate_hwp_offline() argument 1411 intel_pstate_hwp_reenable(struct cpudata * cpu) intel_pstate_hwp_reenable() argument 1419 struct cpudata *cpu = all_cpu_data[policy->cpu]; intel_pstate_suspend() local 1433 struct cpudata *cpu = all_cpu_data[policy->cpu]; intel_pstate_resume() local 1459 int cpu; intel_pstate_update_policies() local 1479 intel_pstate_update_max_freq(int cpu) intel_pstate_update_max_freq() argument 1501 int cpu; intel_pstate_update_limits_for_all() local 1551 struct cpudata *cpu; show_turbo_pct() local 1573 struct cpudata *cpu; show_num_pstates() local 1631 struct cpudata *cpu = all_cpu_data[0]; store_no_turbo() local 1647 update_cpu_qos_request(int cpu,enum freq_qos_req_type type) update_cpu_qos_request() argument 2155 core_get_min_pstate(int cpu) core_get_min_pstate() argument 2163 core_get_max_pstate_physical(int cpu) core_get_max_pstate_physical() argument 2171 core_get_tdp_ratio(int cpu,u64 plat_info) core_get_tdp_ratio() argument 2204 core_get_max_pstate(int cpu) core_get_max_pstate() argument 2239 core_get_turbo_pstate(int cpu) core_get_turbo_pstate() argument 2262 knl_get_turbo_pstate(int cpu) knl_get_turbo_pstate() argument 2275 hwp_get_cpu_scaling(int cpu) hwp_get_cpu_scaling() argument 2300 intel_pstate_set_pstate(struct cpudata * cpu,int pstate) intel_pstate_set_pstate() argument 2313 intel_pstate_set_min_pstate(struct cpudata * cpu) intel_pstate_set_min_pstate() argument 2318 intel_pstate_get_cpu_pstates(struct cpudata * cpu) intel_pstate_get_cpu_pstates() argument 2369 intel_pstate_hwp_boost_up(struct cpudata * cpu) intel_pstate_hwp_boost_up() argument 2416 intel_pstate_hwp_boost_down(struct cpudata * cpu) intel_pstate_hwp_boost_down() argument 2432 intel_pstate_update_util_hwp_local(struct cpudata * cpu,u64 time) intel_pstate_update_util_hwp_local() argument 2464 struct cpudata *cpu = container_of(data, struct cpudata, update_util); intel_pstate_update_util_hwp() local 2472 intel_pstate_calc_avg_perf(struct cpudata * cpu) intel_pstate_calc_avg_perf() argument 2479 intel_pstate_sample(struct cpudata * cpu,u64 time) intel_pstate_sample() argument 2521 get_avg_frequency(struct cpudata * cpu) get_avg_frequency() argument 2526 get_avg_pstate(struct cpudata * cpu) get_avg_pstate() argument 2532 get_target_pstate(struct cpudata * cpu) get_target_pstate() argument 2567 intel_pstate_prepare_request(struct cpudata * cpu,int pstate) intel_pstate_prepare_request() argument 2575 intel_pstate_update_pstate(struct cpudata * cpu,int pstate) intel_pstate_update_pstate() argument 2584 intel_pstate_adjust_pstate(struct cpudata * cpu) intel_pstate_adjust_pstate() argument 2610 struct cpudata *cpu = container_of(data, struct cpudata, update_util); intel_pstate_update_util() local 2745 struct cpudata *cpu; intel_pstate_init_cpu() local 2787 struct cpudata *cpu = all_cpu_data[cpu_num]; intel_pstate_set_update_util_hook() local 2804 intel_pstate_clear_update_util_hook(unsigned int cpu) intel_pstate_clear_update_util_hook() argument 2816 intel_pstate_get_max_freq(struct cpudata * cpu) intel_pstate_get_max_freq() argument 2822 intel_pstate_update_perf_limits(struct cpudata * cpu,unsigned int policy_min,unsigned int policy_max) intel_pstate_update_perf_limits() argument 2887 struct cpudata *cpu; intel_pstate_set_policy() local 2936 intel_pstate_adjust_policy_max(struct cpudata * cpu,struct cpufreq_policy_data * policy) intel_pstate_adjust_policy_max() argument 2948 intel_pstate_verify_cpu_policy(struct cpudata * cpu,struct cpufreq_policy_data * policy) intel_pstate_verify_cpu_policy() argument 2974 struct cpudata *cpu = all_cpu_data[policy->cpu]; intel_cpufreq_cpu_offline() local 2999 struct cpudata *cpu = all_cpu_data[policy->cpu]; intel_pstate_cpu_online() local 3035 struct cpudata *cpu; __intel_pstate_cpu_init() local 3076 struct cpudata *cpu = all_cpu_data[policy->cpu]; intel_pstate_cpu_init() local 3100 struct cpudata *cpu = all_cpu_data[policy->cpu]; intel_cpufreq_verify_policy() local 3124 intel_cpufreq_trace(struct cpudata * cpu,unsigned int trace_type,int old_pstate) intel_cpufreq_trace() argument 3146 intel_cpufreq_hwp_update(struct cpudata * cpu,u32 min,u32 max,u32 desired,bool fast_switch) intel_cpufreq_hwp_update() argument 3170 intel_cpufreq_perf_ctl_update(struct cpudata * cpu,u32 target_pstate,bool fast_switch) intel_cpufreq_perf_ctl_update() argument 3184 struct cpudata *cpu = all_cpu_data[policy->cpu]; intel_cpufreq_update_pstate() local 3210 struct cpudata *cpu = all_cpu_data[policy->cpu]; intel_cpufreq_target() local 3232 struct cpudata *cpu = all_cpu_data[policy->cpu]; intel_cpufreq_fast_switch() local 3247 struct cpudata *cpu = all_cpu_data[cpunum]; intel_cpufreq_adjust_perf() local 3290 struct cpudata *cpu; intel_cpufreq_cpu_init() local 3379 struct cpudata *cpu = all_cpu_data[policy->cpu]; intel_cpufreq_suspend() local 3414 unsigned int cpu; intel_pstate_driver_cleanup() local [all...] |
| /linux/drivers/base/ |
| H A D | arch_topology.c | 74 int cpu; in topology_set_scale_freq_source() local 85 for_each_cpu(cpu, cpus) { in topology_set_scale_freq_source() 86 sfd = rcu_dereference(*per_cpu_ptr(&sft_data, cpu)); in topology_set_scale_freq_source() 90 rcu_assign_pointer(per_cpu(sft_data, cpu), data); in topology_set_scale_freq_source() 91 cpumask_set_cpu(cpu, &scale_freq_counters_mask); in topology_set_scale_freq_source() 105 int cpu; in topology_clear_scale_freq_source() local 109 for_each_cpu(cpu, cpus) { in topology_clear_scale_freq_source() 110 sfd = rcu_dereference(*per_cpu_ptr(&sft_data, cpu)); in topology_clear_scale_freq_source() 113 rcu_assign_pointer(per_cpu(sft_data, cpu), NULL); in topology_clear_scale_freq_source() 114 cpumask_clear_cpu(cpu, &scale_freq_counters_mask); in topology_clear_scale_freq_source() [all …]
|
| H A D | cacheinfo.c | 26 #define ci_cacheinfo(cpu) (&per_cpu(ci_cpu_cacheinfo, cpu)) argument 27 #define cache_leaves(cpu) (ci_cacheinfo(cpu)->num_leaves) argument 28 #define per_cpu_cacheinfo(cpu) (ci_cacheinfo(cpu)->info_list) argument 29 #define per_cpu_cacheinfo_idx(cpu, idx) \ argument 30 (per_cpu_cacheinfo(cpu) + (idx)) 35 struct cpu_cacheinfo *get_cpu_cacheinfo(unsigned int cpu) in get_cpu_cacheinfo() argument 37 return ci_cacheinfo(cpu); in get_cpu_cacheinfo() 58 bool last_level_cache_is_valid(unsigned int cpu) in last_level_cache_is_valid() argument 62 if (!cache_leaves(cpu) || !per_cpu_cacheinfo(cpu)) in last_level_cache_is_valid() 65 llc = per_cpu_cacheinfo_idx(cpu, cache_leaves(cpu) - 1); in last_level_cache_is_valid() [all …]
|
| /linux/arch/powerpc/kernel/ |
| H A D | smp.c | 278 void smp_muxed_ipi_set_message(int cpu, int msg) in smp_muxed_ipi_set_message() argument 280 struct cpu_messages *info = &per_cpu(ipi_message, cpu); in smp_muxed_ipi_set_message() 290 void smp_muxed_ipi_message_pass(int cpu, int msg) in smp_muxed_ipi_message_pass() argument 292 smp_muxed_ipi_set_message(cpu, msg); in smp_muxed_ipi_message_pass() 298 smp_ops->cause_ipi(cpu); in smp_muxed_ipi_message_pass() 352 static inline void do_message_pass(int cpu, int msg) in do_message_pass() argument 355 smp_ops->message_pass(cpu, msg); in do_message_pass() 358 smp_muxed_ipi_message_pass(cpu, msg); in do_message_pass() 362 void arch_smp_send_reschedule(int cpu) in arch_smp_send_reschedule() argument 365 do_message_pass(cpu, PPC_MSG_RESCHEDULE); in arch_smp_send_reschedule() [all …]
|
| H A D | tau_6xx.c | 55 static void set_thresholds(unsigned long cpu) in set_thresholds() argument 60 mtspr(SPRN_THRM1, THRM1_THRES(tau[cpu].low) | THRM1_V | maybe_tie | THRM1_TID); in set_thresholds() 63 mtspr(SPRN_THRM2, THRM1_THRES(tau[cpu].high) | THRM1_V | maybe_tie); in set_thresholds() 66 static void TAUupdate(int cpu) in TAUupdate() argument 77 if (tau[cpu].low >= step_size) { in TAUupdate() 78 tau[cpu].low -= step_size; in TAUupdate() 79 tau[cpu].high -= (step_size - window_expand); in TAUupdate() 81 tau[cpu].grew = 1; in TAUupdate() 88 if (tau[cpu].high <= 127 - step_size) { in TAUupdate() 89 tau[cpu].low += (step_size - window_expand); in TAUupdate() [all …]
|
| H A D | watchdog.c | 149 int cpu = raw_smp_processor_id(); in wd_lockup_ipi() local 152 pr_emerg("CPU %d Hard LOCKUP\n", cpu); in wd_lockup_ipi() 154 cpu, tb, per_cpu(wd_timer_tb, cpu), in wd_lockup_ipi() 155 tb_to_ns(tb - per_cpu(wd_timer_tb, cpu)) / 1000000); in wd_lockup_ipi() 183 static bool set_cpu_stuck(int cpu) in set_cpu_stuck() argument 185 cpumask_set_cpu(cpu, &wd_smp_cpus_stuck); in set_cpu_stuck() 186 cpumask_clear_cpu(cpu, &wd_smp_cpus_pending); in set_cpu_stuck() 201 static void watchdog_smp_panic(int cpu) in watchdog_smp_panic() argument 214 if (cpumask_test_cpu(cpu, &wd_smp_cpus_pending)) in watchdog_smp_panic() 221 if (c == cpu) in watchdog_smp_panic() [all …]
|
| /linux/arch/arm/boot/dts/intel/axm/ |
| H A D | axm5516-cpus.dtsi | 13 cpu-map { 16 cpu = <&CPU0>; 19 cpu = <&CPU1>; 22 cpu = <&CPU2>; 25 cpu = <&CPU3>; 30 cpu = <&CPU4>; 33 cpu = <&CPU5>; 36 cpu = <&CPU6>; 39 cpu = <&CPU7>; 44 cpu = <&CPU8>; [all …]
|
| /linux/include/linux/ |
| H A D | topology.h | 91 static inline int cpu_to_node(int cpu) in cpu_to_node() argument 93 return per_cpu(numa_node, cpu); in cpu_to_node() 105 static inline void set_cpu_numa_node(int cpu, int node) in set_cpu_numa_node() argument 107 per_cpu(numa_node, cpu) = node; in set_cpu_numa_node() 148 static inline int cpu_to_mem(int cpu) in cpu_to_mem() argument 150 return per_cpu(_numa_mem_, cpu); in cpu_to_mem() 155 static inline void set_cpu_numa_mem(int cpu, int node) in set_cpu_numa_mem() argument 157 per_cpu(_numa_mem_, cpu) = node; in set_cpu_numa_mem() 172 static inline int cpu_to_mem(int cpu) in cpu_to_mem() argument 174 return cpu_to_node(cpu); in cpu_to_mem() [all …]
|
| H A D | cpumask.h | 135 static __always_inline void cpu_max_bits_warn(unsigned int cpu, unsigned int bits) in cpu_max_bits_warn() argument 138 WARN_ON_ONCE(cpu >= bits); in cpu_max_bits_warn() 143 static __always_inline unsigned int cpumask_check(unsigned int cpu) in cpumask_check() argument 145 cpu_max_bits_warn(cpu, small_cpumask_bits); in cpumask_check() 146 return cpu; in cpumask_check() 379 #define for_each_cpu(cpu, mask) \ argument 380 for_each_set_bit(cpu, cpumask_bits(mask), small_cpumask_bits) 392 #define for_each_cpu_wrap(cpu, mask, start) \ argument 393 for_each_set_bit_wrap(cpu, cpumask_bits(mask), small_cpumask_bits, start) 409 #define for_each_cpu_and(cpu, mask1, mask2) \ argument [all …]
|
| /linux/arch/um/kernel/ |
| H A D | smp.c | 32 void arch_smp_send_reschedule(int cpu) in arch_smp_send_reschedule() argument 34 os_send_ipi(cpu, UML_IPI_RES); in arch_smp_send_reschedule() 37 void arch_send_call_function_single_ipi(int cpu) in arch_send_call_function_single_ipi() argument 39 os_send_ipi(cpu, UML_IPI_CALL_SINGLE); in arch_send_call_function_single_ipi() 44 int cpu; in arch_send_call_function_ipi_mask() local 46 for_each_cpu(cpu, mask) in arch_send_call_function_ipi_mask() 47 os_send_ipi(cpu, UML_IPI_CALL); in arch_send_call_function_ipi_mask() 52 int cpu, me = smp_processor_id(); in smp_send_stop() local 54 for_each_online_cpu(cpu) { in smp_send_stop() 55 if (cpu == me) in smp_send_stop() [all …]
|
| /linux/tools/perf/tests/ |
| H A D | topology.c | 75 struct perf_cpu cpu; in check_cpu_topology() 86 * cpu is a NULL pointer. in check_cpu_topology() 104 if (!env->cpu && strncmp(env->arch, "s390", 4) && strncmp(env->arch, "aarch64", 7)) in check_cpu_topology() 111 * test if physical_package_id returns -1 for cpu from perf_cpu_map. in check_cpu_topology() 118 TEST_ASSERT_VAL("Session header CPU map not set", env->cpu); in check_cpu_topology() 121 cpu.cpu = i; in check_cpu_topology() 122 if (!perf_cpu_map__has(map, cpu)) in check_cpu_topology() 125 env->cpu[i].core_id, in check_cpu_topology() 126 env->cpu[ in check_cpu_topology() 72 struct perf_cpu cpu; check_cpu_topology() local [all...] |
| /linux/tools/testing/selftests/cpu-hotplug/ |
| H A D | cpu-on-off-test.sh | 27 if ! ls $SYSFS/devices/system/cpu/cpu*/online > /dev/null 2>&1; then 28 echo $msg cpu hotplug is not supported >&2 33 online_cpus=`cat $SYSFS/devices/system/cpu/online` 37 echo "$msg: since there is only one cpu: $online_cpus" 41 present_cpus=`cat $SYSFS/devices/system/cpu/present` 47 offline_cpus=`cat $SYSFS/devices/system/cpu/offline` 63 for cpu in $SYSFS/devices/system/cpu/cpu*; d [all...] |
| /linux/arch/arm/mach-meson/ |
| H A D | platsmp.c | 38 static struct reset_control *meson_smp_get_core_reset(int cpu) in meson_smp_get_core_reset() argument 40 struct device_node *np = of_get_cpu_node(cpu, 0); in meson_smp_get_core_reset() 45 static void meson_smp_set_cpu_ctrl(int cpu, bool on_off) in meson_smp_set_cpu_ctrl() argument 50 val |= BIT(cpu); in meson_smp_set_cpu_ctrl() 52 val &= ~BIT(cpu); in meson_smp_set_cpu_ctrl() 116 static void meson_smp_begin_secondary_boot(unsigned int cpu) in meson_smp_begin_secondary_boot() argument 125 sram_base + MESON_SMP_SRAM_CPU_CTRL_ADDR_REG(cpu)); in meson_smp_begin_secondary_boot() 131 scu_cpu_power_enable(scu_base, cpu); in meson_smp_begin_secondary_boot() 134 static int meson_smp_finalize_secondary_boot(unsigned int cpu) in meson_smp_finalize_secondary_boot() argument 139 while (readl(sram_base + MESON_SMP_SRAM_CPU_CTRL_ADDR_REG(cpu))) { in meson_smp_finalize_secondary_boot() [all …]
|
| /linux/Documentation/translations/zh_CN/scheduler/ |
| H A D | sched-bwc.rst | 24 达“配额”微秒的CPU时间。当cgroup中的线程可运行时,该配额以时间片段的方式被分配到每个cpu 29 它以需求为基础被转移到cpu-local“筒仓”,在每次更新中转移的数量是可调整的,被描述为“片“(时 65 配额、周期和突发是在cpu子系统内通过cgroupfs管理的。 69 :ref:`Documentation/admin-guide/cgroup-v2.rst <cgroup-v2-cpu>`. 71 - cpu.cfs_quota_us:在一个时期内补充的运行时间(微秒)。 72 - cpu.cfs_period_us:一个周期的长度(微秒)。 73 - cpu.stat: 输出节流统计数据[下面进一步解释] 74 - cpu.cfs_burst_us:最大累积运行时间(微秒)。 78 cpu.cfs_period_us=100ms 79 cpu.cfs_quota_us=-1 [all …]
|
| /linux/arch/arm/mach-bcm/ |
| H A D | platsmp-brcmstb.c | 59 static int per_cpu_sw_state_rd(u32 cpu) in per_cpu_sw_state_rd() argument 61 sync_cache_r(SHIFT_PERCPU_PTR(&per_cpu_sw_state, per_cpu_offset(cpu))); in per_cpu_sw_state_rd() 62 return per_cpu(per_cpu_sw_state, cpu); in per_cpu_sw_state_rd() 65 static void per_cpu_sw_state_wr(u32 cpu, int val) in per_cpu_sw_state_wr() argument 68 per_cpu(per_cpu_sw_state, cpu) = val; in per_cpu_sw_state_wr() 69 sync_cache_w(SHIFT_PERCPU_PTR(&per_cpu_sw_state, per_cpu_offset(cpu))); in per_cpu_sw_state_wr() 72 static inline void per_cpu_sw_state_wr(u32 cpu, int val) { } in per_cpu_sw_state_wr() argument 75 static void __iomem *pwr_ctrl_get_base(u32 cpu) in pwr_ctrl_get_base() argument 78 base += (cpu_logical_map(cpu) * 4); in pwr_ctrl_get_base() 82 static u32 pwr_ctrl_rd(u32 cpu) in pwr_ctrl_rd() argument [all …]
|
| /linux/arch/powerpc/include/asm/ |
| H A D | smp.h | 35 extern int cpu_to_chip_id(int cpu); 45 void (*message_pass)(int cpu, int msg); 47 void (*cause_ipi)(int cpu); 49 int (*cause_nmi_ipi)(int cpu); 68 extern int smp_send_nmi_ipi(int cpu, void (*fn)(struct pt_regs *), u64 delay_us); 69 extern int smp_send_safe_nmi_ipi(int cpu, void (*fn)(struct pt_regs *), u64 delay_us); 79 void generic_cpu_die(unsigned int cpu); 80 void generic_set_cpu_dead(unsigned int cpu); 81 void generic_set_cpu_up(unsigned int cpu); 82 int generic_check_cpu_restart(unsigned int cpu); [all …]
|
| H A D | topology.h | 48 static inline void update_numa_cpu_lookup_table(unsigned int cpu, int node) in update_numa_cpu_lookup_table() argument 50 numa_cpu_lookup_table[cpu] = node; in update_numa_cpu_lookup_table() 53 static inline int early_cpu_to_node(int cpu) in early_cpu_to_node() argument 57 nid = numa_cpu_lookup_table[cpu]; in early_cpu_to_node() 69 extern void map_cpu_to_node(int cpu, int node); 71 extern void unmap_cpu_from_node(unsigned long cpu); 76 static inline int early_cpu_to_node(int cpu) { return 0; } in early_cpu_to_node() argument 90 static inline void update_numa_cpu_lookup_table(unsigned int cpu, int node) {} in update_numa_cpu_lookup_table() argument 105 static inline void map_cpu_to_node(int cpu, int node) {} in map_cpu_to_node() argument 107 static inline void unmap_cpu_from_node(unsigned long cpu) {} in unmap_cpu_from_node() argument [all …]
|
| /linux/tools/power/cpupower/utils/ |
| H A D | cpufreq-info.c | 42 if (strstr(value, "cpu ")) in count_cpus() 44 if (sscanf(value, "cpu%d ", &cpunr) != 1) in count_cpus() 51 /* cpu count starts from 0, on error return 1 (UP) */ in count_cpus() 58 unsigned int cpu, nr_cpus; in proc_cpufreq_output() local 67 for (cpu = 0; cpu < nr_cpus; cpu++) { in proc_cpufreq_output() 68 policy = cpufreq_get_policy(cpu); in proc_cpufreq_output() 72 if (cpufreq_get_hardware_limits(cpu, &min, &max)) { in proc_cpufreq_output() 79 cpu , polic in proc_cpufreq_output() 125 get_boost_mode_x86(unsigned int cpu) get_boost_mode_x86() argument 207 get_boost_mode_generic(unsigned int cpu) get_boost_mode_generic() argument 221 get_boost_mode(unsigned int cpu) get_boost_mode() argument 250 get_freq_kernel(unsigned int cpu,unsigned int human) get_freq_kernel() argument 269 get_freq_hardware(unsigned int cpu,unsigned int human) get_freq_hardware() argument 292 get_hardware_limits(unsigned int cpu,unsigned int human) get_hardware_limits() argument 315 get_driver(unsigned int cpu) get_driver() argument 329 get_policy(unsigned int cpu) get_policy() argument 351 get_available_governors(unsigned int cpu) get_available_governors() argument 374 get_affected_cpus(unsigned int cpu) get_affected_cpus() argument 395 get_related_cpus(unsigned int cpu) get_related_cpus() argument 416 get_freq_stats(unsigned int cpu,unsigned int human) get_freq_stats() argument 441 get_epp(unsigned int cpu,bool interactive) get_epp() argument 458 get_latency(unsigned int cpu,unsigned int human) get_latency() argument 481 get_perf_cap(unsigned int cpu) get_perf_cap() argument 490 debug_output_one(unsigned int cpu) debug_output_one() argument 548 unsigned int cpu = 0; cmd_freq_info() local [all...] |
| /linux/arch/x86/xen/ |
| H A D | smp.c | 32 void xen_smp_intr_free(unsigned int cpu) in xen_smp_intr_free() argument 34 kfree(per_cpu(xen_resched_irq, cpu).name); in xen_smp_intr_free() 35 per_cpu(xen_resched_irq, cpu).name = NULL; in xen_smp_intr_free() 36 if (per_cpu(xen_resched_irq, cpu).irq >= 0) { in xen_smp_intr_free() 37 unbind_from_irqhandler(per_cpu(xen_resched_irq, cpu).irq, NULL); in xen_smp_intr_free() 38 per_cpu(xen_resched_irq, cpu).irq = -1; in xen_smp_intr_free() 40 kfree(per_cpu(xen_callfunc_irq, cpu).name); in xen_smp_intr_free() 41 per_cpu(xen_callfunc_irq, cpu).name = NULL; in xen_smp_intr_free() 42 if (per_cpu(xen_callfunc_irq, cpu).irq >= 0) { in xen_smp_intr_free() 43 unbind_from_irqhandler(per_cpu(xen_callfunc_irq, cpu).irq, NULL); in xen_smp_intr_free() [all …]
|
| H A D | smp_pv.c | 59 int cpu; in cpu_bringup() local 71 cpu = smp_processor_id(); in cpu_bringup() 72 identify_secondary_cpu(cpu); in cpu_bringup() 73 set_cpu_sibling_map(cpu); in cpu_bringup() 79 notify_cpu_starting(cpu); in cpu_bringup() 81 set_cpu_online(cpu, true); in cpu_bringup() 95 void xen_smp_intr_free_pv(unsigned int cpu) in xen_smp_intr_free_pv() argument 97 kfree(per_cpu(xen_irq_work, cpu).name); in xen_smp_intr_free_pv() 98 per_cpu(xen_irq_work, cpu).name = NULL; in xen_smp_intr_free_pv() 99 if (per_cpu(xen_irq_work, cpu).irq >= 0) { in xen_smp_intr_free_pv() [all …]
|
| /linux/arch/arm/mach-tegra/ |
| H A D | platsmp.c | 36 static void tegra_secondary_init(unsigned int cpu) in tegra_secondary_init() argument 38 cpumask_set_cpu(cpu, &tegra_cpu_init_mask); in tegra_secondary_init() 42 static int tegra20_boot_secondary(unsigned int cpu, struct task_struct *idle) in tegra20_boot_secondary() argument 44 cpu = cpu_logical_map(cpu); in tegra20_boot_secondary() 54 tegra_put_cpu_in_reset(cpu); in tegra20_boot_secondary() 62 flowctrl_write_cpu_halt(cpu, 0); in tegra20_boot_secondary() 64 tegra_enable_cpu_clock(cpu); in tegra20_boot_secondary() 65 flowctrl_write_cpu_csr(cpu, 0); /* Clear flow controller CSR. */ in tegra20_boot_secondary() 66 tegra_cpu_out_of_reset(cpu); in tegra20_boot_secondary() 70 static int tegra30_boot_secondary(unsigned int cpu, struct task_struct *idle) in tegra30_boot_secondary() argument [all …]
|
| /linux/tools/perf/util/ |
| H A D | cpumap.c | 75 * Special treatment for -1, which is not real cpu number, in cpu_map__from_entries() 79 if (data->cpus_data.cpu[i] == (u16) -1) { in cpu_map__from_entries() 80 RC_CHK_ACCESS(map)->map[i].cpu = -1; in cpu_map__from_entries() 81 } else if (data->cpus_data.cpu[i] < INT16_MAX) { in cpu_map__from_entries() 82 RC_CHK_ACCESS(map)->map[i].cpu = (int16_t) data->cpus_data.cpu[i]; in cpu_map__from_entries() 84 pr_err("Invalid cpumap entry %u\n", data->cpus_data.cpu[i]); in cpu_map__from_entries() 110 int cpu; in cpu_map__from_mask() local 113 for_each_set_bit(cpu, local_copy, 64) { in cpu_map__from_mask() 114 if (cpu in cpu_map__from_mask() 140 for (int cpu = data->range_cpu_data.start_cpu; cpu <= data->range_cpu_data.end_cpu; cpu_map__from_range() local 206 cpu__get_topology_int(int cpu,const char * name,int * value) cpu__get_topology_int() argument 216 cpu__get_socket_id(struct perf_cpu cpu) cpu__get_socket_id() argument 222 aggr_cpu_id__socket(struct perf_cpu cpu,void * data __maybe_unused) aggr_cpu_id__socket() argument 258 struct perf_cpu cpu; cpu_aggr_map__new() local 300 cpu__get_die_id(struct perf_cpu cpu) cpu__get_die_id() argument 307 aggr_cpu_id__die(struct perf_cpu cpu,void * data) aggr_cpu_id__die() argument 330 cpu__get_cluster_id(struct perf_cpu cpu) cpu__get_cluster_id() argument 337 aggr_cpu_id__cluster(struct perf_cpu cpu,void * data) aggr_cpu_id__cluster() argument 354 cpu__get_core_id(struct perf_cpu cpu) cpu__get_core_id() argument 360 aggr_cpu_id__core(struct perf_cpu cpu,void * data) aggr_cpu_id__core() argument 379 aggr_cpu_id__cpu(struct perf_cpu cpu,void * data) aggr_cpu_id__cpu() argument 393 aggr_cpu_id__node(struct perf_cpu cpu,void * data __maybe_unused) aggr_cpu_id__node() argument 401 aggr_cpu_id__global(struct perf_cpu cpu,void * data __maybe_unused) aggr_cpu_id__global() argument 544 cpu__get_node(struct perf_cpu cpu) cpu__get_node() argument 577 unsigned int cpu, mem; cpu__setup_cpunode_map() local 635 struct perf_cpu cpu = { .cpu = INT16_MAX }; cpu_map__snprint() local 704 for (int cpu = last_cpu.cpu / 4 * 4; cpu >= 0; cpu -= 4) { cpu_map__snprint_mask() local [all...] |
| /linux/arch/s390/kernel/ |
| H A D | smp.c | 167 int cpu; in pcpu_find_address() local 169 for_each_cpu(cpu, mask) in pcpu_find_address() 170 if (per_cpu(pcpu_devices, cpu).address == address) in pcpu_find_address() 171 return &per_cpu(pcpu_devices, cpu); in pcpu_find_address() 183 static int pcpu_alloc_lowcore(struct pcpu *pcpu, int cpu) in pcpu_alloc_lowcore() argument 199 lc->cpu_nr = cpu; in pcpu_alloc_lowcore() 200 lc->spinlock_lockval = arch_spin_lockval(cpu); in pcpu_alloc_lowcore() 207 if (abs_lowcore_map(cpu, lc, true)) in pcpu_alloc_lowcore() 209 lowcore_ptr[cpu] = lc; in pcpu_alloc_lowcore() 223 static void pcpu_free_lowcore(struct pcpu *pcpu, int cpu) in pcpu_free_lowcore() argument [all …]
|
| /linux/kernel/ |
| H A D | smpboot.c | 28 struct task_struct *idle_thread_get(unsigned int cpu) in idle_thread_get() argument 30 struct task_struct *tsk = per_cpu(idle_threads, cpu); in idle_thread_get() 48 static __always_inline void idle_init(unsigned int cpu) in idle_init() argument 50 struct task_struct *tsk = per_cpu(idle_threads, cpu); in idle_init() 53 tsk = fork_idle(cpu); in idle_init() 55 pr_err("SMP: fork_idle() failed for CPU %u\n", cpu); in idle_init() 57 per_cpu(idle_threads, cpu) = tsk; in idle_init() 66 unsigned int cpu, boot_cpu; in idle_threads_init() local 70 for_each_possible_cpu(cpu) { in idle_threads_init() 71 if (cpu != boot_cpu) in idle_threads_init() [all …]
|
| /linux/arch/arm64/boot/dts/amd/ |
| H A D | elba-16core.dtsi | 11 cpu-map { 13 core0 { cpu = <&cpu0>; }; 14 core1 { cpu = <&cpu1>; }; 15 core2 { cpu = <&cpu2>; }; 16 core3 { cpu = <&cpu3>; }; 20 core0 { cpu = <&cpu4>; }; 21 core1 { cpu = <&cpu5>; }; 22 core2 { cpu = <&cpu6>; }; 23 core3 { cpu = <&cpu7>; }; 27 core0 { cpu = <&cpu8>; }; [all …]
|