Home
last modified time | relevance | path

Searched full:locked (Results 1 – 25 of 1524) sorted by relevance

12345678910>>...61

/linux/tools/testing/selftests/net/forwarding/
H A Dbridge_locked_port.sh102 bridge link set dev $swp1 locked on
112 bridge link set dev $swp1 locked off
118 log_test "Locked port ipv4"
133 bridge link set dev $swp1 locked on
142 bridge link set dev $swp1 locked off
150 log_test "Locked port vlan"
161 bridge link set dev $swp1 locked on
170 bridge link set dev $swp1 locked off
176 log_test "Locked port ipv6"
187 bridge link set dev $swp1 learning on locked on
[all …]
/linux/Documentation/i2c/
H A Di2c-topology.rst40 mux-locked or parent-locked muxes.
43 Mux-locked muxes
46 Mux-locked muxes does not lock the entire parent adapter during the
48 adapter are locked. Mux-locked muxes are mostly interesting if the
50 their tasks. Since the parent adapter is not fully locked during the
55 Mux-locked Example
62 | root |--+--| locked | '--------'
87 Mux-locked caveats
90 When using a mux-locked mux, be aware of the following restrictions:
93 If you build a topology with a mux-locked mux being the parent
[all …]
/linux/fs/resctrl/
H A Dpseudo_lock.c29 * pseudo-locked regions.
86 * @minor: The minor number of the device representing pseudo-locked region
89 * pseudo-locked region it belongs to. This is done by matching the minor
90 * number of the device to the pseudo-locked region it belongs.
92 * Minor numbers are assigned at the time a pseudo-locked region is associated
95 * Return: On success return pointer to resource group owning the pseudo-locked
113 * @list: Entry within the @pm_reqs list for a pseudo-locked region
134 * @plr: Pseudo-locked region
188 * All content of the pseudo-locked region is reset - any memory allocated
211 * Called after user provided a schemata to be pseudo-locked. From the
[all …]
/linux/kernel/locking/
H A Dqspinlock_paravirt.h92 try_cmpxchg_acquire(&lock->locked, &old, _Q_LOCKED_VAL)) { in pv_hybrid_queued_unfair_trylock()
124 return !READ_ONCE(lock->locked) && in trylock_clear_pending()
142 * Try to clear pending bit & set locked bit in trylock_clear_pending()
286 * Wait for node->locked to become true, halt the vcpu after a short spin.
299 if (READ_ONCE(node->locked)) in pv_wait_node()
309 * Order pn->state vs pn->locked thusly: in pv_wait_node()
311 * [S] pn->state = VCPU_HALTED [S] next->locked = 1 in pv_wait_node()
313 * [L] pn->locked [RmW] pn->state = VCPU_HASHED in pv_wait_node()
319 if (!READ_ONCE(node->locked)) { in pv_wait_node()
333 * If the locked flag is still not set after wakeup, it is a in pv_wait_node()
[all …]
H A Dmcs_spinlock.h53 * on this node->locked until the previous lock holder sets the node->locked
62 node->locked = 0;
74 * Lock acquired, don't need to set node->locked to 1. Threads in mcs_spin_lock()
75 * only spin on its own node->locked value for lock acquisition. in mcs_spin_lock()
77 * and does not proceed to spin on its own node->locked, this in mcs_spin_lock()
79 * audit lock status, then set node->locked value here. in mcs_spin_lock()
86 arch_mcs_spin_lock_contended(&node->locked); in mcs_spin_lock()
110 arch_mcs_spin_unlock_contended(&next->locked); in mcs_spin_unlock()
20 int locked; /* 1 if lock acquired */ global() member
/linux/include/drm/
H A Ddrm_exec.h29 * @num_objects: number of objects locked
39 * @objects: array of the locked objects
49 * @prelocked: already locked GEM object due to contention
59 * Return: Pointer to the locked object corresponding to @index if
60 * index is within the number of locked objects. NULL otherwise.
69 * drm_exec_for_each_locked_object - iterate over all the locked objects
74 * Iterate over all the locked GEM objects inside the drm_exec object.
80 * drm_exec_for_each_locked_object_reverse - iterate over all the locked
86 * Iterate over all the locked GEM objects inside the drm_exec object in
95 * drm_exec_until_all_locked - loop until all GEM objects are locked
[all …]
/linux/scripts/coccinelle/locks/
H A Ddouble_lock.cocci17 @locked@
42 position p1 != locked.p1;
43 position locked.p;
45 expression x <= locked.E1;
46 expression E,locked.E1;
65 expression x <= locked.E1;
66 expression locked.E1;
69 position locked.p,p1,p2;
H A Dcall_kern.cocci39 @locked exists@
74 @depends on locked && patch@
81 @depends on locked && !patch@
90 p1 << locked.p1;
91 p2 << locked.p2;
101 p1 << locked.p1;
102 p2 << locked.p2;
/linux/tools/testing/selftests/mm/
H A Dmlock2-tests.c92 #define LOCKED "lo" macro
133 bool locked; in is_vma_lock_on_fault() local
136 locked = is_vmflag_set(addr, LOCKED); in is_vma_lock_on_fault()
137 if (!locked) in is_vma_lock_on_fault()
153 bool locked; in lock_check() local
156 locked = is_vmflag_set(addr, LOCKED); in lock_check()
157 if (!locked) in lock_check()
168 if (is_vmflag_set((unsigned long)map, LOCKED)) { in unlock_lock_check()
169 ksft_print_msg("VMA flag %s is present on page 1 after unlock\n", LOCKED); in unlock_lock_check()
191 ksft_test_result(lock_check((unsigned long)map), "%s: Locked\n", __func__); in test_mlock_lock()
[all …]
/linux/drivers/media/dvb-frontends/
H A Dlgs8gxx.c295 static int lgs8gxx_is_locked(struct lgs8gxx_state *priv, u8 *locked) in lgs8gxx_is_locked() argument
308 *locked = ((t & 0x80) == 0x80) ? 1 : 0; in lgs8gxx_is_locked()
310 *locked = ((t & 0xC0) == 0xC0) ? 1 : 0; in lgs8gxx_is_locked()
315 static int lgs8gxx_wait_ca_lock(struct lgs8gxx_state *priv, u8 *locked) in lgs8gxx_wait_ca_lock() argument
331 *locked = (ret == 0) ? 1 : 0; in lgs8gxx_wait_ca_lock()
359 u8 *locked) in lgs8gxx_autolock_gi() argument
384 err = lgs8gxx_wait_ca_lock(priv, locked); in lgs8gxx_autolock_gi()
385 if (err || !(*locked)) in lgs8gxx_autolock_gi()
393 *locked = 0; in lgs8gxx_autolock_gi()
403 u8 locked = 0, tmp_gi; in lgs8gxx_auto_detect() local
[all …]
/linux/drivers/net/ethernet/intel/ice/
H A Dice_dcb_lib.h18 int ice_dcb_sw_dflt_cfg(struct ice_pf *pf, bool ets_willing, bool locked);
24 ice_pf_dcb_cfg(struct ice_pf *pf, struct ice_dcbx_cfg *new_cfg, bool locked);
26 void ice_pf_dcb_recfg(struct ice_pf *pf, bool locked);
28 int ice_init_pf_dcb(struct ice_pf *pf, bool locked);
96 ice_init_pf_dcb(struct ice_pf *pf, bool __always_unused locked) in ice_init_pf_dcb() argument
105 bool __always_unused locked) in ice_pf_dcb_cfg() argument
134 static inline void ice_pf_dcb_recfg(struct ice_pf *pf, bool locked) { } in ice_pf_dcb_recfg() argument
/linux/drivers/watchdog/
H A Dintel_oc_wdt.c45 bool locked; member
62 if (oc_wdt->locked) in intel_oc_wdt_start()
126 oc_wdt->locked = !!(val & INTEL_OC_WDT_CTL_LCK); in intel_oc_wdt_setup()
136 if (oc_wdt->locked) { in intel_oc_wdt_setup()
143 * If we are locked read the current timeout value in intel_oc_wdt_setup()
150 "Register access locked, heartbeat fixed at: %u s\n", in intel_oc_wdt_setup()
153 } else if (oc_wdt->locked) { in intel_oc_wdt_setup()
155 * In case the watchdog is disabled and locked there in intel_oc_wdt_setup()
/linux/drivers/thermal/intel/
H A Dintel_quark_dts_thermal.c100 bool locked; member
129 if (!aux_entry->locked) { in soc_dts_enable()
136 pr_info("DTS is locked. Cannot enable DTS\n"); in soc_dts_enable()
157 if (!aux_entry->locked) { in soc_dts_disable()
165 pr_info("DTS is locked. Cannot disable DTS\n"); in soc_dts_disable()
206 if (aux_entry->locked) { in update_trip_temp()
316 if (!aux_entry->locked) { in free_soc_dts()
345 /* Check if DTS register is locked */ in alloc_soc_dts()
351 aux_entry->locked = !!(out & QRK_DTS_LOCK_BIT); in alloc_soc_dts()
353 /* Store DTS default state if DTS registers are not locked */ in alloc_soc_dts()
[all …]
/linux/tools/testing/selftests/bpf/progs/
H A Dbpf_arena_spin_lock.h37 u8 locked; member
52 u8 locked; member
64 int locked; member
78 * 0- 7: locked byte
90 #define _Q_LOCKED_MASK _Q_SET_MASK(LOCKED)
204 WRITE_ONCE(lock->locked, _Q_LOCKED_VAL); in set_locked()
252 * Wait for in-progress pending->locked hand-overs with a bounded in arena_spin_lock_slowpath()
299 * store-release that clears the locked bit and create lock in arena_spin_lock_slowpath()
305 (void)smp_cond_load_acquire_label(&lock->locked, !VAL, release_err); in arena_spin_lock_slowpath()
345 node->locked = 0; in arena_spin_lock_slowpath()
[all …]
/linux/drivers/media/common/siano/
H A Dsmscoreapi.h677 u32 is_rf_locked; /* 0 - not locked, 1 - locked */
678 u32 is_demod_locked; /* 0 - not locked, 1 - locked */
802 u32 is_rf_locked; /* 0 - not locked, 1 - locked */
803 u32 is_demod_locked; /* 0 - not locked, 1 - locked */
816 u32 modem_state; /* 0 - Acquisition, 1 - Locked */
844 u32 is_rf_locked; /* 0 - not locked, 1 - locked */
845 u32 is_demod_locked; /* 0 - not locked, 1 - locked */
858 u32 modem_state; /* 0 - Acquisition, 1 - Locked */
942 u32 is_demod_locked; /* 0 - not locked, 1 - locked */
946 u32 is_rf_locked; /* 0 - not locked, 1 - locked */
[all …]
/linux/mm/
H A Dgup.c1084 * FOLL_NOWAIT, the mmap_lock may be released. If it is, *@locked will be set
1089 int *locked) in faultin_page() argument
1134 *locked = 0; in faultin_page()
1156 *locked = 0; in faultin_page()
1309 * @locked: whether we're still with the mmap_lock held
1344 * be released. If this happens *@locked will be set to 0 on return.
1357 int *locked) in __get_user_pages() argument
1429 PTR_ERR(page) == -EMLINK, locked); in __get_user_pages()
1637 * Locking: (*locked == 1) means that the mmap_lock has already been acquired by
1639 * set (*locked
1653 __get_user_pages_locked(struct mm_struct * mm,unsigned long start,unsigned long nr_pages,struct page ** pages,int * locked,unsigned int flags) __get_user_pages_locked() argument
1814 populate_vma_page_range(struct vm_area_struct * vma,unsigned long start,unsigned long end,int * locked) populate_vma_page_range() argument
1888 faultin_page_range(struct mm_struct * mm,unsigned long start,unsigned long end,bool write,int * locked) faultin_page_range() argument
1930 int locked = 0; __mm_populate() local
1981 __get_user_pages_locked(struct mm_struct * mm,unsigned long start,unsigned long nr_pages,struct page ** pages,int * locked,unsigned int foll_flags) __get_user_pages_locked() argument
2187 get_dump_page(unsigned long addr,int * locked) get_dump_page() argument
2469 __gup_longterm_locked(struct mm_struct * mm,unsigned long start,unsigned long nr_pages,struct page ** pages,int * locked,unsigned int gup_flags) __gup_longterm_locked() argument
2500 is_valid_gup_args(struct page ** pages,int * locked,unsigned int * gup_flags_p,unsigned int to_set) is_valid_gup_args() argument
2606 get_user_pages_remote(struct mm_struct * mm,unsigned long start,unsigned long nr_pages,unsigned int gup_flags,struct page ** pages,int * locked) get_user_pages_remote() argument
2624 get_user_pages_remote(struct mm_struct * mm,unsigned long start,unsigned long nr_pages,unsigned int gup_flags,struct page ** pages,int * locked) get_user_pages_remote() argument
2647 int locked = 1; get_user_pages() local
2675 int locked = 0; get_user_pages_unlocked() local
3191 int locked = 0; gup_fast_fallback() local
3356 pin_user_pages_remote(struct mm_struct * mm,unsigned long start,unsigned long nr_pages,unsigned int gup_flags,struct page ** pages,int * locked) pin_user_pages_remote() argument
3390 int locked = 1; pin_user_pages() local
3410 int locked = 0; pin_user_pages_unlocked() local
[all...]
H A Dmmap_lock.c58 bool exclusive; /* Are we exclusively locked? */ in __vma_enter_locked()
202 * locked result to avoid performance overhead, in which case we fall back to in vma_start_read()
204 * False locked result is possible if mm_lock_seq overflows or if vma gets in vma_start_read()
220 * Check before locking. A race might cause false locked result. in vma_start_read()
252 * Overflow of vm_lock_seq/mm_lock_seq might produce false locked result. in lock_vma_under_rcu()
325 * locked and we know it hasn't already been isolated. in lock_next_vma()
331 /* Check if the vma we locked is the right one. */ in lock_next_vma()
102 int locked; __vma_start_write() local
/linux/Documentation/filesystems/
H A Ddirectory-locking.rst44 * decide which of the source and target need to be locked.
45 The source needs to be locked if it's a non-directory, target - if it's
49 are non-directories - the source because it wouldn't need to be locked
66 to be read, modified or removed by method will be locked by the caller.
89 dcache trees. Lookup is already holding the parent locked. If alias is
102 all we change is the view in dcache. Moreover, holding a directory locked
185 Each operation in the minimal cycle must have locked at least
204 In other words, we have a cross-directory rename that locked
213 Consider the order in which directories are locked by the
215 Dn and D1 would have to be among those, with Dn locked before D1.
[all …]
/linux/include/linux/
H A Dww_mutex.h70 * @first_lock_dep_map: fake lockdep_map for first locked ww_mutex.
72 * lockdep requires the lockdep_map for the first locked ww_mutex
75 * fake locked ww_mutex lockdep map between ww_acquire_init() and
107 * It is not allowed to initialize an already locked mutex.
180 * code and clearly designated the acquire phase from actually using the locked
246 * locked. The mutex must first be initialized (or statically defined) before it
247 * can be locked. memset()-ing the mutex to 0 is not allowed. The mutex must be
279 * locked. The mutex must first be initialized (or statically defined) before it
280 * can be locked. memset()-ing the mutex to 0 is not allowed. The mutex must be
371 * use of the mutex is forbidden. The mutex must not be locked whe in ww_mutex_destroy()
[all...]
/linux/include/asm-generic/
H A Dqspinlock_types.h25 u8 locked; member
40 u8 locked; member
55 * 0- 7: locked byte
62 * 0- 7: locked byte
71 #define _Q_LOCKED_MASK _Q_SET_MASK(LOCKED)
/linux/arch/arc/include/asm/
H A Dspinlock.h23 " breq %[val], %[LOCKED], 1b \n" /* spin while LOCKED */ in arch_spin_lock()
24 " scond %[LOCKED], [%[slock]] \n" /* acquire */ in arch_spin_lock()
29 [LOCKED] "r" (__ARCH_SPIN_LOCK_LOCKED__) in arch_spin_lock()
50 " breq %[val], %[LOCKED], 4f \n" /* already LOCKED, just bail */ in arch_spin_trylock()
51 " scond %[LOCKED], [%[slock]] \n" /* acquire */ in arch_spin_trylock()
59 [LOCKED] "r" (__ARCH_SPIN_LOCK_LOCKED__) in arch_spin_trylock()
95 " brls %[val], %[WR_LOCKED], 1b\n" /* <= 0: spin while write locked */ in arch_read_lock()
115 " brls %[val], %[WR_LOCKED], 4f\n" /* <= 0: already write locked, bail */ in arch_read_trylock()
/linux/drivers/video/fbdev/
H A Dep93xx-fb.c27 #define EP93XXFB_VLINES_TOTAL 0x0000 /* SW locked */
28 #define EP93XXFB_VSYNC 0x0004 /* SW locked */
29 #define EP93XXFB_VACTIVE 0x0008 /* SW locked */
30 #define EP93XXFB_VBLANK 0x0228 /* SW locked */
31 #define EP93XXFB_VCLK 0x000c /* SW locked */
34 #define EP93XXFB_HCLKS_TOTAL 0x0010 /* SW locked */
35 #define EP93XXFB_HSYNC 0x0014 /* SW locked */
36 #define EP93XXFB_HACTIVE 0x0018 /* SW locked */
37 #define EP93XXFB_HBLANK 0x022c /* SW locked */
38 #define EP93XXFB_HCLK 0x001c /* SW locked */
[all …]
/linux/tools/perf/pmu-events/arch/x86/broadwellde/
H A Dother.json32 "BriefDescription": "Cycles when L1 and L2 are locked due to UC or split lock",
36 …L1 and L2 are locked due to a UC lock or split lock. A lock is asserted in case of locked memory a…
/linux/tools/perf/pmu-events/arch/x86/broadwellx/
H A Dother.json32 "BriefDescription": "Cycles when L1 and L2 are locked due to UC or split lock",
36 …L1 and L2 are locked due to a UC lock or split lock. A lock is asserted in case of locked memory a…
/linux/tools/perf/pmu-events/arch/x86/broadwell/
H A Dother.json32 "BriefDescription": "Cycles when L1 and L2 are locked due to UC or split lock",
36 …L1 and L2 are locked due to a UC lock or split lock. A lock is asserted in case of locked memory a…

12345678910>>...61