Home
last modified time | relevance | path

Searched refs:rdp (Results 1 – 25 of 39) sorted by relevance

12

/linux/kernel/rcu/
H A Dtree_nocb.h20 static inline bool rcu_current_is_nocb_kthread(struct rcu_data *rdp) in rcu_current_is_nocb_kthread() argument
23 if (!rdp->nocb_cb_kthread || !rdp->nocb_gp_kthread) in rcu_current_is_nocb_kthread()
26 if (current == rdp->nocb_cb_kthread || current == rdp->nocb_gp_kthread) in rcu_current_is_nocb_kthread()
92 static void rcu_nocb_bypass_lock(struct rcu_data *rdp) in rcu_nocb_bypass_lock() argument
93 __acquires(&rdp->nocb_bypass_lock) in rcu_nocb_bypass_lock()
96 if (raw_spin_trylock(&rdp->nocb_bypass_lock)) in rcu_nocb_bypass_lock()
102 WARN_ON_ONCE(smp_processor_id() != rdp->cpu); in rcu_nocb_bypass_lock()
103 raw_spin_lock(&rdp in rcu_nocb_bypass_lock()
110 rcu_nocb_bypass_trylock(struct rcu_data * rdp) rcu_nocb_bypass_trylock() argument
119 rcu_nocb_bypass_unlock(struct rcu_data * rdp) rcu_nocb_bypass_unlock() argument
130 rcu_nocb_lock(struct rcu_data * rdp) rcu_nocb_lock() argument
142 rcu_nocb_unlock(struct rcu_data * rdp) rcu_nocb_unlock() argument
154 rcu_nocb_unlock_irqrestore(struct rcu_data * rdp,unsigned long flags) rcu_nocb_unlock_irqrestore() argument
166 rcu_lockdep_assert_cblist_protected(struct rcu_data * rdp) rcu_lockdep_assert_cblist_protected() argument
193 __wake_nocb_gp(struct rcu_data * rdp_gp,struct rcu_data * rdp,bool force,unsigned long flags) __wake_nocb_gp() argument
228 wake_nocb_gp(struct rcu_data * rdp,bool force) wake_nocb_gp() argument
266 wake_nocb_gp_defer(struct rcu_data * rdp,int waketype,const char * reason) wake_nocb_gp_defer() argument
308 rcu_nocb_do_flush_bypass(struct rcu_data * rdp,struct rcu_head * rhp_in,unsigned long j,bool lazy) rcu_nocb_do_flush_bypass() argument
352 rcu_nocb_flush_bypass(struct rcu_data * rdp,struct rcu_head * rhp,unsigned long j,bool lazy) rcu_nocb_flush_bypass() argument
366 rcu_nocb_try_flush_bypass(struct rcu_data * rdp,unsigned long j) rcu_nocb_try_flush_bypass() argument
393 rcu_nocb_try_bypass(struct rcu_data * rdp,struct rcu_head * rhp,bool * was_alldone,unsigned long flags,bool lazy) rcu_nocb_try_bypass() argument
526 __call_rcu_nocb_wake(struct rcu_data * rdp,bool was_alldone,unsigned long flags) __call_rcu_nocb_wake() argument
595 call_rcu_nocb(struct rcu_data * rdp,struct rcu_head * head,rcu_callback_t func,unsigned long flags,bool lazy) call_rcu_nocb() argument
607 nocb_gp_toggle_rdp(struct rcu_data * rdp_gp,struct rcu_data * rdp) nocb_gp_toggle_rdp() argument
660 struct rcu_data *rdp, *rdp_toggling = NULL; nocb_gp_wait() local
858 struct rcu_data *rdp = arg; rcu_nocb_gp_kthread() local
868 nocb_cb_wait_cond(struct rcu_data * rdp) nocb_cb_wait_cond() argument
877 nocb_cb_wait(struct rcu_data * rdp) nocb_cb_wait() argument
946 struct rcu_data *rdp = arg; rcu_nocb_cb_kthread() local
958 rcu_nocb_need_deferred_wakeup(struct rcu_data * rdp,int level) rcu_nocb_need_deferred_wakeup() argument
964 do_nocb_deferred_wakeup_common(struct rcu_data * rdp_gp,struct rcu_data * rdp,int level,unsigned long flags) do_nocb_deferred_wakeup_common() argument
988 struct rcu_data *rdp = timer_container_of(rdp, t, nocb_timer); do_nocb_deferred_wakeup_timer() local
1003 do_nocb_deferred_wakeup(struct rcu_data * rdp) do_nocb_deferred_wakeup() argument
1021 rcu_nocb_queue_toggle_rdp(struct rcu_data * rdp) rcu_nocb_queue_toggle_rdp() argument
1039 rcu_nocb_rdp_deoffload_wait_cond(struct rcu_data * rdp) rcu_nocb_rdp_deoffload_wait_cond() argument
1056 rcu_nocb_rdp_deoffload(struct rcu_data * rdp) rcu_nocb_rdp_deoffload() argument
1112 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); rcu_nocb_cpu_deoffload() local
1134 rcu_nocb_rdp_offload_wait_cond(struct rcu_data * rdp) rcu_nocb_rdp_offload_wait_cond() argument
1146 rcu_nocb_rdp_offload(struct rcu_data * rdp) rcu_nocb_rdp_offload() argument
1180 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); rcu_nocb_cpu_offload() local
1218 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); lazy_rcu_shrink_count() local
1253 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); lazy_rcu_shrink_scan() local
1291 struct rcu_data *rdp; rcu_init_nohz() local
1354 rcu_boot_init_nocb_percpu_data(struct rcu_data * rdp) rcu_boot_init_nocb_percpu_data() argument
1375 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); rcu_spawn_cpu_nocb_kthread() local
1453 struct rcu_data *rdp; rcu_organize_nocb_kthreads() local
1527 show_rcu_nocb_gp_state(struct rcu_data * rdp) show_rcu_nocb_gp_state() argument
1551 show_rcu_nocb_state(struct rcu_data * rdp) show_rcu_nocb_state() argument
1625 rcu_nocb_lock(struct rcu_data * rdp) rcu_nocb_lock() argument
1630 rcu_nocb_unlock(struct rcu_data * rdp) rcu_nocb_unlock() argument
1635 rcu_nocb_unlock_irqrestore(struct rcu_data * rdp,unsigned long flags) rcu_nocb_unlock_irqrestore() argument
1642 rcu_lockdep_assert_cblist_protected(struct rcu_data * rdp) rcu_lockdep_assert_cblist_protected() argument
1660 wake_nocb_gp(struct rcu_data * rdp,bool force) wake_nocb_gp() argument
1665 rcu_nocb_flush_bypass(struct rcu_data * rdp,struct rcu_head * rhp,unsigned long j,bool lazy) rcu_nocb_flush_bypass() argument
1671 call_rcu_nocb(struct rcu_data * rdp,struct rcu_head * head,rcu_callback_t func,unsigned long flags,bool lazy) call_rcu_nocb() argument
1677 __call_rcu_nocb_wake(struct rcu_data * rdp,bool was_empty,unsigned long flags) __call_rcu_nocb_wake() argument
1683 rcu_boot_init_nocb_percpu_data(struct rcu_data * rdp) rcu_boot_init_nocb_percpu_data() argument
1687 rcu_nocb_need_deferred_wakeup(struct rcu_data * rdp,int level) rcu_nocb_need_deferred_wakeup() argument
1692 do_nocb_deferred_wakeup(struct rcu_data * rdp) do_nocb_deferred_wakeup() argument
1701 show_rcu_nocb_state(struct rcu_data * rdp) show_rcu_nocb_state() argument
[all...]
H A Dtree.c86 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); in rcu_get_gpwrap_count() local
88 return READ_ONCE(rdp->gpwrap_count); in rcu_get_gpwrap_count()
162 static void rcu_report_exp_rdp(struct rcu_data *rdp);
163 static void rcu_report_qs_rdp(struct rcu_data *rdp);
164 static void check_cb_ovld_locked(struct rcu_data *rdp, struct rcu_node *rnp);
165 static bool rcu_rdp_is_offloaded(struct rcu_data *rdp);
166 static bool rcu_rdp_cpu_online(struct rcu_data *rdp);
231 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); in rcu_get_n_cbs_cpu()
233 if (rcu_segcblist_is_enabled(&rdp->cblist)) in rcu_get_n_cbs_cpu()
234 return rcu_segcblist_n_cbs(&rdp in rcu_get_n_cbs_cpu()
230 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); rcu_get_n_cbs_cpu() local
311 rcu_watching_snap_stopped_since(struct rcu_data * rdp,int snap) rcu_watching_snap_stopped_since() argument
600 struct rcu_data *rdp = this_cpu_ptr(&rcu_data); rcu_irq_work_resched() local
663 struct rcu_data *rdp = this_cpu_ptr(&rcu_data); __rcu_irq_enter_check_tick() local
720 rcu_disable_urgency_upon_qs(struct rcu_data * rdp) rcu_disable_urgency_upon_qs() argument
801 rcu_gpnum_ovf(struct rcu_node * rnp,struct rcu_data * rdp) rcu_gpnum_ovf() argument
818 rcu_watching_snap_save(struct rcu_data * rdp) rcu_watching_snap_save() argument
854 rcu_watching_snap_recheck(struct rcu_data * rdp) rcu_watching_snap_recheck() argument
991 trace_rcu_this_gp(struct rcu_node * rnp,struct rcu_data * rdp,unsigned long gp_seq_req,const char * s) trace_rcu_this_gp() argument
1015 rcu_start_this_gp(struct rcu_node * rnp_start,struct rcu_data * rdp,unsigned long gp_seq_req) rcu_start_this_gp() argument
1093 struct rcu_data *rdp = this_cpu_ptr(&rcu_data); rcu_future_gp_cleanup() local
1142 rcu_accelerate_cbs(struct rcu_node * rnp,struct rcu_data * rdp) rcu_accelerate_cbs() argument
1189 rcu_accelerate_cbs_unlocked(struct rcu_node * rnp,struct rcu_data * rdp) rcu_accelerate_cbs_unlocked() argument
1218 rcu_advance_cbs(struct rcu_node * rnp,struct rcu_data * rdp) rcu_advance_cbs() argument
1242 rcu_advance_cbs_nowake(struct rcu_node * rnp,struct rcu_data * rdp) rcu_advance_cbs_nowake() argument
1272 __note_gp_changes(struct rcu_node * rnp,struct rcu_data * rdp) __note_gp_changes() argument
1321 note_gp_changes(struct rcu_data * rdp) note_gp_changes() argument
1808 struct rcu_data *rdp; rcu_gp_init() local
2145 struct rcu_data *rdp; rcu_gp_cleanup() local
2431 rcu_report_qs_rdp(struct rcu_data * rdp) rcu_report_qs_rdp() argument
2486 rcu_check_quiescent_state(struct rcu_data * rdp) rcu_check_quiescent_state() argument
2528 rcu_do_batch(struct rcu_data * rdp) rcu_do_batch() argument
2720 force_qs_rnp(int (* f)(struct rcu_data * rdp)) force_qs_rnp() argument
2750 struct rcu_data *rdp; force_qs_rnp() local
2825 struct rcu_data *rdp = raw_cpu_ptr(&rcu_data); rcu_core() local
2986 rcutree_enqueue(struct rcu_data * rdp,struct rcu_head * head,rcu_callback_t func) rcutree_enqueue() argument
2997 call_rcu_core(struct rcu_data * rdp,struct rcu_head * head,rcu_callback_t func,unsigned long flags) call_rcu_core() argument
3053 check_cb_ovld_locked(struct rcu_data * rdp,struct rcu_node * rnp) check_cb_ovld_locked() argument
3076 check_cb_ovld(struct rcu_data * rdp) check_cb_ovld() argument
3095 struct rcu_data *rdp; __call_rcu_common() local
3450 struct rcu_data *rdp; start_poll_synchronize_rcu_common() local
3661 struct rcu_data *rdp = this_cpu_ptr(&rcu_data); rcu_pending() local
3742 rcu_barrier_entrain(struct rcu_data * rdp) rcu_barrier_entrain() argument
3782 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); rcu_barrier_handler() local
3810 struct rcu_data *rdp; rcu_barrier() local
3993 rcu_rdp_cpu_online(struct rcu_data * rdp) rcu_rdp_cpu_online() argument
4000 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); rcu_cpu_online() local
4022 struct rcu_data *rdp; rcu_lockdep_current_cpu_online() local
4132 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); rcu_boot_init_percpu_data() local
4232 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); rcutree_prepare_cpu() local
4281 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); rcu_cpu_beenfullyonline() local
4293 struct rcu_data *rdp; rcutree_online_cpu() local
4326 struct rcu_data *rdp; rcutree_report_cpu_starting() local
4384 struct rcu_data *rdp = this_cpu_ptr(&rcu_data); rcutree_report_cpu_dead() local
4440 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); rcutree_migrate_callbacks() local
4508 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); rcutree_dying_cpu() local
4524 struct rcu_data *rdp; rcutree_offline_cpu() local
4572 struct rcu_data *rdp = this_cpu_ptr(&rcu_data); rcu_spawn_gp_kthread() local
[all...]
H A Dtree.h164 * equal to rdp->gp_seq.
167 unsigned long gp_seq; /* Track rdp->gp_seq counter */
266 struct rcu_data *nocb_toggling_rdp; /* rdp queued for (de-)offloading */
270 /* GP rdp takes GP-end wakeups. */
490 static void rcu_preempt_deferred_qs_init(struct rcu_data *rdp);
493 static bool rcu_is_callbacks_kthread(struct rcu_data *rdp);
498 static void zero_cpu_stall_ticks(struct rcu_data *rdp);
502 static bool wake_nocb_gp(struct rcu_data *rdp);
503 static bool rcu_nocb_flush_bypass(struct rcu_data *rdp, struct rcu_head *rhp,
505 static void call_rcu_nocb(struct rcu_data *rdp, struc
527 rcu_nocb_lock_irqsave(rdp,flags) global() argument
534 rcu_nocb_lock_irqsave(rdp,flags) global() argument
[all...]
/linux/drivers/net/ethernet/amd/
H A Da2065.c135 ll->rdp = (leptr & 0xFFFF); in load_csrs()
137 ll->rdp = leptr >> 16; in load_csrs()
139 ll->rdp = lp->busmaster_regval; in load_csrs()
224 ll->rdp = LE_C0_INIT; in init_restart_lance()
227 for (i = 0; (i < 100) && !(ll->rdp & (LE_C0_ERR | LE_C0_IDON)); i++) in init_restart_lance()
229 if ((i == 100) || (ll->rdp & LE_C0_ERR)) { in init_restart_lance()
230 pr_err("unopened after %d ticks, csr0=%04x\n", i, ll->rdp); in init_restart_lance()
235 ll->rdp = LE_C0_IDON; in init_restart_lance()
236 ll->rdp = LE_C0_INEA | LE_C0_STRT; in init_restart_lance()
265 ll->rdp = LE_C0_RINT | LE_C0_INEA; in lance_rx()
[all …]
H A Da2065.h41 unsigned short rdp; /* Register Data Port */ member
/linux/drivers/net/ethernet/cavium/liquidio/
H A Docteon_nic.c38 struct octeon_instr_rdp *rdp; in octeon_alloc_soft_command_resp() local
54 rdp = (struct octeon_instr_rdp *)&sc->cmd.cmd3.rdp; in octeon_alloc_soft_command_resp()
60 rdp = (struct octeon_instr_rdp *)&sc->cmd.cmd2.rdp; in octeon_alloc_soft_command_resp()
68 rdp->pcie_port = oct->pcie_port; in octeon_alloc_soft_command_resp()
69 rdp->rlen = rdatasize; in octeon_alloc_soft_command_resp()
H A Drequest_manager.c590 struct octeon_instr_rdp *rdp; in octeon_prepare_soft_command() local
631 rdp = (struct octeon_instr_rdp *)&sc->cmd.cmd3.rdp; in octeon_prepare_soft_command()
632 rdp->pcie_port = oct->pcie_port; in octeon_prepare_soft_command()
633 rdp->rlen = sc->rdatasize; in octeon_prepare_soft_command()
668 rdp = (struct octeon_instr_rdp *)&sc->cmd.cmd2.rdp; in octeon_prepare_soft_command()
669 rdp->pcie_port = oct->pcie_port; in octeon_prepare_soft_command()
670 rdp->rlen = sc->rdatasize; in octeon_prepare_soft_command()
H A Docteon_iq.h221 u64 rdp; member
248 u64 rdp; member
/linux/tools/rcu/
H A Drcu-cbs.py42 rdp = per_cpu_ptr(rdp0, cpu); variable
43 len = rdp.cblist.len.value_();
/linux/drivers/pnp/isapnp/
H A Dcore.c219 int rdp = isapnp_rdp; in isapnp_next_rdp() local
226 while (rdp <= 0x3ff) { in isapnp_next_rdp()
231 if ((rdp < 0x280 || rdp > 0x380) in isapnp_next_rdp()
232 && request_region(rdp, 1, "ISAPnP")) { in isapnp_next_rdp()
233 isapnp_rdp = rdp; in isapnp_next_rdp()
234 old_rdp = rdp; in isapnp_next_rdp()
237 rdp += RDP_STEP; in isapnp_next_rdp()
/linux/sound/pcmcia/pdaudiocf/
H A Dpdaudiocf_irq.c250 int size, off, cont, rdp, wdp; in pdacf_threaded_irq() local
258 rdp = inw(chip->port + PDAUDIOCF_REG_RDP); in pdacf_threaded_irq()
260 size = wdp - rdp; in pdacf_threaded_irq()
/linux/arch/arm64/boot/dts/qcom/
H A Dipq9574-rdp454.dts11 #include "ipq9574-rdp-common.dtsi"
H A Dipq9574-rdp453.dts11 #include "ipq9574-rdp-common.dtsi"
H A Dipq9574-rdp449.dts11 #include "ipq9574-rdp-common.dtsi"
H A Dipq5332-rdp474.dts10 #include "ipq5332-rdp-common.dtsi"
H A Dipq9574-rdp418.dts11 #include "ipq9574-rdp-common.dtsi"
H A Dipq5332-rdp442.dts10 #include "ipq5332-rdp-common.dtsi"
H A Dipq5332-rdp468.dts10 #include "ipq5332-rdp-common.dtsi"
H A Dipq5332-rdp441.dts10 #include "ipq5332-rdp-common.dtsi"
/linux/arch/x86/include/asm/
H A Duser_64.h58 __u64 rdp; member
/linux/Documentation/RCU/Design/Memory-Ordering/
H A DTree-RCU-Memory-Ordering.rst205 4 struct rcu_data *rdp = this_cpu_ptr(&rcu_data);
210 9 if (rcu_rdp_is_offloaded(rdp))
215 14 if (tne != rdp->tick_nohz_enabled_snap) {
216 15 if (!rcu_segcblist_empty(&rdp->cblist))
218 17 rdp->tick_nohz_enabled_snap = tne;
228 27 if (rdp->last_accelerate == jiffies)
230 29 rdp->last_accelerate = jiffies;
231 30 if (rcu_segcblist_pend_cbs(&rdp->cblist)) {
232 31 rnp = rdp->mynode;
234 33 needwake = rcu_accelerate_cbs(rnp, rdp);
/linux/include/soc/fsl/
H A Dcpm.h21 __be32 rdp; /* Internal */ member
/linux/arch/x86/include/uapi/asm/
H A Dsigcontext.h156 __u64 rdp; member
/linux/arch/x86/include/asm/fpu/
H A Dtypes.h44 u64 rdp; /* Data Pointer */ member
/linux/drivers/i2c/busses/
H A Di2c-cpm.c56 uint rdp; /* Internal */ member
157 out_be32(&i2c_ram->rdp, 0); in cpm_reset_i2c_params()

12