| /linux/kernel/rcu/ |
| H A D | tree_nocb.h | 20 static inline bool rcu_current_is_nocb_kthread(struct rcu_data *rdp) in rcu_current_is_nocb_kthread() argument 23 if (!rdp->nocb_cb_kthread || !rdp->nocb_gp_kthread) in rcu_current_is_nocb_kthread() 26 if (current == rdp->nocb_cb_kthread || current == rdp->nocb_gp_kthread) in rcu_current_is_nocb_kthread() 92 static void rcu_nocb_bypass_lock(struct rcu_data *rdp) in rcu_nocb_bypass_lock() argument 93 __acquires(&rdp->nocb_bypass_lock) in rcu_nocb_bypass_lock() 96 if (raw_spin_trylock(&rdp->nocb_bypass_lock)) in rcu_nocb_bypass_lock() 102 WARN_ON_ONCE(smp_processor_id() != rdp->cpu); in rcu_nocb_bypass_lock() 103 raw_spin_lock(&rdp in rcu_nocb_bypass_lock() 110 rcu_nocb_bypass_trylock(struct rcu_data * rdp) rcu_nocb_bypass_trylock() argument 119 rcu_nocb_bypass_unlock(struct rcu_data * rdp) rcu_nocb_bypass_unlock() argument 130 rcu_nocb_lock(struct rcu_data * rdp) rcu_nocb_lock() argument 142 rcu_nocb_unlock(struct rcu_data * rdp) rcu_nocb_unlock() argument 154 rcu_nocb_unlock_irqrestore(struct rcu_data * rdp,unsigned long flags) rcu_nocb_unlock_irqrestore() argument 166 rcu_lockdep_assert_cblist_protected(struct rcu_data * rdp) rcu_lockdep_assert_cblist_protected() argument 193 __wake_nocb_gp(struct rcu_data * rdp_gp,struct rcu_data * rdp,bool force,unsigned long flags) __wake_nocb_gp() argument 228 wake_nocb_gp(struct rcu_data * rdp,bool force) wake_nocb_gp() argument 266 wake_nocb_gp_defer(struct rcu_data * rdp,int waketype,const char * reason) wake_nocb_gp_defer() argument 308 rcu_nocb_do_flush_bypass(struct rcu_data * rdp,struct rcu_head * rhp_in,unsigned long j,bool lazy) rcu_nocb_do_flush_bypass() argument 352 rcu_nocb_flush_bypass(struct rcu_data * rdp,struct rcu_head * rhp,unsigned long j,bool lazy) rcu_nocb_flush_bypass() argument 366 rcu_nocb_try_flush_bypass(struct rcu_data * rdp,unsigned long j) rcu_nocb_try_flush_bypass() argument 393 rcu_nocb_try_bypass(struct rcu_data * rdp,struct rcu_head * rhp,bool * was_alldone,unsigned long flags,bool lazy) rcu_nocb_try_bypass() argument 526 __call_rcu_nocb_wake(struct rcu_data * rdp,bool was_alldone,unsigned long flags) __call_rcu_nocb_wake() argument 595 call_rcu_nocb(struct rcu_data * rdp,struct rcu_head * head,rcu_callback_t func,unsigned long flags,bool lazy) call_rcu_nocb() argument 607 nocb_gp_toggle_rdp(struct rcu_data * rdp_gp,struct rcu_data * rdp) nocb_gp_toggle_rdp() argument 660 struct rcu_data *rdp, *rdp_toggling = NULL; nocb_gp_wait() local 858 struct rcu_data *rdp = arg; rcu_nocb_gp_kthread() local 868 nocb_cb_wait_cond(struct rcu_data * rdp) nocb_cb_wait_cond() argument 877 nocb_cb_wait(struct rcu_data * rdp) nocb_cb_wait() argument 946 struct rcu_data *rdp = arg; rcu_nocb_cb_kthread() local 958 rcu_nocb_need_deferred_wakeup(struct rcu_data * rdp,int level) rcu_nocb_need_deferred_wakeup() argument 964 do_nocb_deferred_wakeup_common(struct rcu_data * rdp_gp,struct rcu_data * rdp,int level,unsigned long flags) do_nocb_deferred_wakeup_common() argument 988 struct rcu_data *rdp = timer_container_of(rdp, t, nocb_timer); do_nocb_deferred_wakeup_timer() local 1003 do_nocb_deferred_wakeup(struct rcu_data * rdp) do_nocb_deferred_wakeup() argument 1021 rcu_nocb_queue_toggle_rdp(struct rcu_data * rdp) rcu_nocb_queue_toggle_rdp() argument 1039 rcu_nocb_rdp_deoffload_wait_cond(struct rcu_data * rdp) rcu_nocb_rdp_deoffload_wait_cond() argument 1056 rcu_nocb_rdp_deoffload(struct rcu_data * rdp) rcu_nocb_rdp_deoffload() argument 1112 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); rcu_nocb_cpu_deoffload() local 1134 rcu_nocb_rdp_offload_wait_cond(struct rcu_data * rdp) rcu_nocb_rdp_offload_wait_cond() argument 1146 rcu_nocb_rdp_offload(struct rcu_data * rdp) rcu_nocb_rdp_offload() argument 1180 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); rcu_nocb_cpu_offload() local 1218 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); lazy_rcu_shrink_count() local 1253 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); lazy_rcu_shrink_scan() local 1291 struct rcu_data *rdp; rcu_init_nohz() local 1354 rcu_boot_init_nocb_percpu_data(struct rcu_data * rdp) rcu_boot_init_nocb_percpu_data() argument 1375 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); rcu_spawn_cpu_nocb_kthread() local 1453 struct rcu_data *rdp; rcu_organize_nocb_kthreads() local 1527 show_rcu_nocb_gp_state(struct rcu_data * rdp) show_rcu_nocb_gp_state() argument 1551 show_rcu_nocb_state(struct rcu_data * rdp) show_rcu_nocb_state() argument 1625 rcu_nocb_lock(struct rcu_data * rdp) rcu_nocb_lock() argument 1630 rcu_nocb_unlock(struct rcu_data * rdp) rcu_nocb_unlock() argument 1635 rcu_nocb_unlock_irqrestore(struct rcu_data * rdp,unsigned long flags) rcu_nocb_unlock_irqrestore() argument 1642 rcu_lockdep_assert_cblist_protected(struct rcu_data * rdp) rcu_lockdep_assert_cblist_protected() argument 1660 wake_nocb_gp(struct rcu_data * rdp,bool force) wake_nocb_gp() argument 1665 rcu_nocb_flush_bypass(struct rcu_data * rdp,struct rcu_head * rhp,unsigned long j,bool lazy) rcu_nocb_flush_bypass() argument 1671 call_rcu_nocb(struct rcu_data * rdp,struct rcu_head * head,rcu_callback_t func,unsigned long flags,bool lazy) call_rcu_nocb() argument 1677 __call_rcu_nocb_wake(struct rcu_data * rdp,bool was_empty,unsigned long flags) __call_rcu_nocb_wake() argument 1683 rcu_boot_init_nocb_percpu_data(struct rcu_data * rdp) rcu_boot_init_nocb_percpu_data() argument 1687 rcu_nocb_need_deferred_wakeup(struct rcu_data * rdp,int level) rcu_nocb_need_deferred_wakeup() argument 1692 do_nocb_deferred_wakeup(struct rcu_data * rdp) do_nocb_deferred_wakeup() argument 1701 show_rcu_nocb_state(struct rcu_data * rdp) show_rcu_nocb_state() argument [all...] |
| H A D | tree.c | 86 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); in rcu_get_gpwrap_count() local 88 return READ_ONCE(rdp->gpwrap_count); in rcu_get_gpwrap_count() 162 static void rcu_report_exp_rdp(struct rcu_data *rdp); 163 static void rcu_report_qs_rdp(struct rcu_data *rdp); 164 static void check_cb_ovld_locked(struct rcu_data *rdp, struct rcu_node *rnp); 165 static bool rcu_rdp_is_offloaded(struct rcu_data *rdp); 166 static bool rcu_rdp_cpu_online(struct rcu_data *rdp); 231 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); in rcu_get_n_cbs_cpu() 233 if (rcu_segcblist_is_enabled(&rdp->cblist)) in rcu_get_n_cbs_cpu() 234 return rcu_segcblist_n_cbs(&rdp in rcu_get_n_cbs_cpu() 230 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); rcu_get_n_cbs_cpu() local 311 rcu_watching_snap_stopped_since(struct rcu_data * rdp,int snap) rcu_watching_snap_stopped_since() argument 600 struct rcu_data *rdp = this_cpu_ptr(&rcu_data); rcu_irq_work_resched() local 663 struct rcu_data *rdp = this_cpu_ptr(&rcu_data); __rcu_irq_enter_check_tick() local 720 rcu_disable_urgency_upon_qs(struct rcu_data * rdp) rcu_disable_urgency_upon_qs() argument 801 rcu_gpnum_ovf(struct rcu_node * rnp,struct rcu_data * rdp) rcu_gpnum_ovf() argument 818 rcu_watching_snap_save(struct rcu_data * rdp) rcu_watching_snap_save() argument 854 rcu_watching_snap_recheck(struct rcu_data * rdp) rcu_watching_snap_recheck() argument 991 trace_rcu_this_gp(struct rcu_node * rnp,struct rcu_data * rdp,unsigned long gp_seq_req,const char * s) trace_rcu_this_gp() argument 1015 rcu_start_this_gp(struct rcu_node * rnp_start,struct rcu_data * rdp,unsigned long gp_seq_req) rcu_start_this_gp() argument 1093 struct rcu_data *rdp = this_cpu_ptr(&rcu_data); rcu_future_gp_cleanup() local 1142 rcu_accelerate_cbs(struct rcu_node * rnp,struct rcu_data * rdp) rcu_accelerate_cbs() argument 1189 rcu_accelerate_cbs_unlocked(struct rcu_node * rnp,struct rcu_data * rdp) rcu_accelerate_cbs_unlocked() argument 1218 rcu_advance_cbs(struct rcu_node * rnp,struct rcu_data * rdp) rcu_advance_cbs() argument 1242 rcu_advance_cbs_nowake(struct rcu_node * rnp,struct rcu_data * rdp) rcu_advance_cbs_nowake() argument 1272 __note_gp_changes(struct rcu_node * rnp,struct rcu_data * rdp) __note_gp_changes() argument 1321 note_gp_changes(struct rcu_data * rdp) note_gp_changes() argument 1808 struct rcu_data *rdp; rcu_gp_init() local 2145 struct rcu_data *rdp; rcu_gp_cleanup() local 2431 rcu_report_qs_rdp(struct rcu_data * rdp) rcu_report_qs_rdp() argument 2486 rcu_check_quiescent_state(struct rcu_data * rdp) rcu_check_quiescent_state() argument 2528 rcu_do_batch(struct rcu_data * rdp) rcu_do_batch() argument 2720 force_qs_rnp(int (* f)(struct rcu_data * rdp)) force_qs_rnp() argument 2750 struct rcu_data *rdp; force_qs_rnp() local 2825 struct rcu_data *rdp = raw_cpu_ptr(&rcu_data); rcu_core() local 2986 rcutree_enqueue(struct rcu_data * rdp,struct rcu_head * head,rcu_callback_t func) rcutree_enqueue() argument 2997 call_rcu_core(struct rcu_data * rdp,struct rcu_head * head,rcu_callback_t func,unsigned long flags) call_rcu_core() argument 3053 check_cb_ovld_locked(struct rcu_data * rdp,struct rcu_node * rnp) check_cb_ovld_locked() argument 3076 check_cb_ovld(struct rcu_data * rdp) check_cb_ovld() argument 3095 struct rcu_data *rdp; __call_rcu_common() local 3450 struct rcu_data *rdp; start_poll_synchronize_rcu_common() local 3661 struct rcu_data *rdp = this_cpu_ptr(&rcu_data); rcu_pending() local 3742 rcu_barrier_entrain(struct rcu_data * rdp) rcu_barrier_entrain() argument 3782 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); rcu_barrier_handler() local 3810 struct rcu_data *rdp; rcu_barrier() local 3993 rcu_rdp_cpu_online(struct rcu_data * rdp) rcu_rdp_cpu_online() argument 4000 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); rcu_cpu_online() local 4022 struct rcu_data *rdp; rcu_lockdep_current_cpu_online() local 4132 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); rcu_boot_init_percpu_data() local 4232 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); rcutree_prepare_cpu() local 4281 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); rcu_cpu_beenfullyonline() local 4293 struct rcu_data *rdp; rcutree_online_cpu() local 4326 struct rcu_data *rdp; rcutree_report_cpu_starting() local 4384 struct rcu_data *rdp = this_cpu_ptr(&rcu_data); rcutree_report_cpu_dead() local 4440 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); rcutree_migrate_callbacks() local 4508 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); rcutree_dying_cpu() local 4524 struct rcu_data *rdp; rcutree_offline_cpu() local 4572 struct rcu_data *rdp = this_cpu_ptr(&rcu_data); rcu_spawn_gp_kthread() local [all...] |
| H A D | tree.h | 164 * equal to rdp->gp_seq. 167 unsigned long gp_seq; /* Track rdp->gp_seq counter */ 266 struct rcu_data *nocb_toggling_rdp; /* rdp queued for (de-)offloading */ 270 /* GP rdp takes GP-end wakeups. */ 490 static void rcu_preempt_deferred_qs_init(struct rcu_data *rdp); 493 static bool rcu_is_callbacks_kthread(struct rcu_data *rdp); 498 static void zero_cpu_stall_ticks(struct rcu_data *rdp); 502 static bool wake_nocb_gp(struct rcu_data *rdp); 503 static bool rcu_nocb_flush_bypass(struct rcu_data *rdp, struct rcu_head *rhp, 505 static void call_rcu_nocb(struct rcu_data *rdp, struc 527 rcu_nocb_lock_irqsave(rdp,flags) global() argument 534 rcu_nocb_lock_irqsave(rdp,flags) global() argument [all...] |
| /linux/drivers/net/ethernet/amd/ |
| H A D | a2065.c | 135 ll->rdp = (leptr & 0xFFFF); in load_csrs() 137 ll->rdp = leptr >> 16; in load_csrs() 139 ll->rdp = lp->busmaster_regval; in load_csrs() 224 ll->rdp = LE_C0_INIT; in init_restart_lance() 227 for (i = 0; (i < 100) && !(ll->rdp & (LE_C0_ERR | LE_C0_IDON)); i++) in init_restart_lance() 229 if ((i == 100) || (ll->rdp & LE_C0_ERR)) { in init_restart_lance() 230 pr_err("unopened after %d ticks, csr0=%04x\n", i, ll->rdp); in init_restart_lance() 235 ll->rdp = LE_C0_IDON; in init_restart_lance() 236 ll->rdp = LE_C0_INEA | LE_C0_STRT; in init_restart_lance() 265 ll->rdp = LE_C0_RINT | LE_C0_INEA; in lance_rx() [all …]
|
| H A D | a2065.h | 41 unsigned short rdp; /* Register Data Port */ member
|
| /linux/drivers/net/ethernet/cavium/liquidio/ |
| H A D | octeon_nic.c | 38 struct octeon_instr_rdp *rdp; in octeon_alloc_soft_command_resp() local 54 rdp = (struct octeon_instr_rdp *)&sc->cmd.cmd3.rdp; in octeon_alloc_soft_command_resp() 60 rdp = (struct octeon_instr_rdp *)&sc->cmd.cmd2.rdp; in octeon_alloc_soft_command_resp() 68 rdp->pcie_port = oct->pcie_port; in octeon_alloc_soft_command_resp() 69 rdp->rlen = rdatasize; in octeon_alloc_soft_command_resp()
|
| H A D | request_manager.c | 590 struct octeon_instr_rdp *rdp; in octeon_prepare_soft_command() local 631 rdp = (struct octeon_instr_rdp *)&sc->cmd.cmd3.rdp; in octeon_prepare_soft_command() 632 rdp->pcie_port = oct->pcie_port; in octeon_prepare_soft_command() 633 rdp->rlen = sc->rdatasize; in octeon_prepare_soft_command() 668 rdp = (struct octeon_instr_rdp *)&sc->cmd.cmd2.rdp; in octeon_prepare_soft_command() 669 rdp->pcie_port = oct->pcie_port; in octeon_prepare_soft_command() 670 rdp->rlen = sc->rdatasize; in octeon_prepare_soft_command()
|
| H A D | octeon_iq.h | 221 u64 rdp; member 248 u64 rdp; member
|
| /linux/tools/rcu/ |
| H A D | rcu-cbs.py | 42 rdp = per_cpu_ptr(rdp0, cpu); variable 43 len = rdp.cblist.len.value_();
|
| /linux/drivers/pnp/isapnp/ |
| H A D | core.c | 219 int rdp = isapnp_rdp; in isapnp_next_rdp() local 226 while (rdp <= 0x3ff) { in isapnp_next_rdp() 231 if ((rdp < 0x280 || rdp > 0x380) in isapnp_next_rdp() 232 && request_region(rdp, 1, "ISAPnP")) { in isapnp_next_rdp() 233 isapnp_rdp = rdp; in isapnp_next_rdp() 234 old_rdp = rdp; in isapnp_next_rdp() 237 rdp += RDP_STEP; in isapnp_next_rdp()
|
| /linux/sound/pcmcia/pdaudiocf/ |
| H A D | pdaudiocf_irq.c | 250 int size, off, cont, rdp, wdp; in pdacf_threaded_irq() local 258 rdp = inw(chip->port + PDAUDIOCF_REG_RDP); in pdacf_threaded_irq() 260 size = wdp - rdp; in pdacf_threaded_irq()
|
| /linux/arch/arm64/boot/dts/qcom/ |
| H A D | ipq9574-rdp454.dts | 11 #include "ipq9574-rdp-common.dtsi"
|
| H A D | ipq9574-rdp453.dts | 11 #include "ipq9574-rdp-common.dtsi"
|
| H A D | ipq9574-rdp449.dts | 11 #include "ipq9574-rdp-common.dtsi"
|
| H A D | ipq5332-rdp474.dts | 10 #include "ipq5332-rdp-common.dtsi"
|
| H A D | ipq9574-rdp418.dts | 11 #include "ipq9574-rdp-common.dtsi"
|
| H A D | ipq5332-rdp442.dts | 10 #include "ipq5332-rdp-common.dtsi"
|
| H A D | ipq5332-rdp468.dts | 10 #include "ipq5332-rdp-common.dtsi"
|
| H A D | ipq5332-rdp441.dts | 10 #include "ipq5332-rdp-common.dtsi"
|
| /linux/arch/x86/include/asm/ |
| H A D | user_64.h | 58 __u64 rdp; member
|
| /linux/Documentation/RCU/Design/Memory-Ordering/ |
| H A D | Tree-RCU-Memory-Ordering.rst | 205 4 struct rcu_data *rdp = this_cpu_ptr(&rcu_data); 210 9 if (rcu_rdp_is_offloaded(rdp)) 215 14 if (tne != rdp->tick_nohz_enabled_snap) { 216 15 if (!rcu_segcblist_empty(&rdp->cblist)) 218 17 rdp->tick_nohz_enabled_snap = tne; 228 27 if (rdp->last_accelerate == jiffies) 230 29 rdp->last_accelerate = jiffies; 231 30 if (rcu_segcblist_pend_cbs(&rdp->cblist)) { 232 31 rnp = rdp->mynode; 234 33 needwake = rcu_accelerate_cbs(rnp, rdp);
|
| /linux/include/soc/fsl/ |
| H A D | cpm.h | 21 __be32 rdp; /* Internal */ member
|
| /linux/arch/x86/include/uapi/asm/ |
| H A D | sigcontext.h | 156 __u64 rdp; member
|
| /linux/arch/x86/include/asm/fpu/ |
| H A D | types.h | 44 u64 rdp; /* Data Pointer */ member
|
| /linux/drivers/i2c/busses/ |
| H A D | i2c-cpm.c | 56 uint rdp; /* Internal */ member 157 out_be32(&i2c_ram->rdp, 0); in cpm_reset_i2c_params()
|