/linux/kernel/ |
H A D | audit_tree.c | 53 * association between mark and chunk is protected by hash_lock and 68 * tree.chunks anchors chunk.owners[].list hash_lock 70 * chunk.trees anchors tree.same_root hash_lock 72 * a hash function. RCU, hash_lock 210 static __cacheline_aligned_in_smp DEFINE_SPINLOCK(hash_lock); 225 /* hash_lock & mark->group->mark_mutex is held by caller */ 284 assert_spin_locked(&hash_lock); in replace_mark_chunk() 357 * whether it didn't change while we've dropped hash_lock. in untag_chunk() 365 spin_lock(&hash_lock); in untag_chunk() 369 spin_unlock(&hash_lock); in untag_chunk() [all...] |
/linux/net/netfilter/ |
H A D | xt_RATEEST.c | 23 struct mutex hash_lock; member 68 mutex_lock(&xn->hash_lock); in xt_rateest_lookup() 70 mutex_unlock(&xn->hash_lock); in xt_rateest_lookup() 79 mutex_lock(&xn->hash_lock); in xt_rateest_put() 89 mutex_unlock(&xn->hash_lock); in xt_rateest_put() 123 mutex_lock(&xn->hash_lock); in xt_rateest_tg_checkentry() 126 mutex_unlock(&xn->hash_lock); in xt_rateest_tg_checkentry() 165 mutex_unlock(&xn->hash_lock); in xt_rateest_tg_checkentry() 171 mutex_unlock(&xn->hash_lock); in xt_rateest_tg_checkentry() 214 mutex_init(&xn->hash_lock); in xt_rateest_net_init()
|
/linux/drivers/md/ |
H A D | dm-region-hash.c | 68 rwlock_t hash_lock; member 202 rwlock_init(&rh->hash_lock); in dm_region_hash_create() 307 write_lock_irq(&rh->hash_lock); in __rh_alloc() 322 write_unlock_irq(&rh->hash_lock); in __rh_alloc() 333 read_unlock(&rh->hash_lock); in __rh_find() 335 read_lock(&rh->hash_lock); in __rh_find() 346 read_lock(&rh->hash_lock); in dm_rh_get_state() 348 read_unlock(&rh->hash_lock); in dm_rh_get_state() 417 read_lock(&rh->hash_lock); in dm_rh_mark_nosync() 419 read_unlock(&rh->hash_lock); in dm_rh_mark_nosync() [all …]
|
/linux/net/bridge/ |
H A D | br_fdb.c | 226 lockdep_assert_held_once(&br->hash_lock); in br_fdb_find() 374 spin_lock_bh(&br->hash_lock); in br_fdb_find_delete_local() 379 spin_unlock_bh(&br->hash_lock); in br_fdb_find_delete_local() 463 spin_lock_bh(&br->hash_lock); in br_fdb_changeaddr() 495 spin_unlock_bh(&br->hash_lock); in br_fdb_changeaddr() 504 spin_lock_bh(&br->hash_lock); in br_fdb_change_mac_address() 530 spin_unlock_bh(&br->hash_lock); in br_fdb_change_mac_address() 566 spin_lock_bh(&br->hash_lock); in br_fdb_cleanup() 569 spin_unlock_bh(&br->hash_lock); in br_fdb_cleanup() 607 spin_lock_bh(&br->hash_lock); in br_fdb_flush() [all …]
|
H A D | br_device.c | 511 spin_lock_init(&br->hash_lock); in br_dev_setup()
|
/linux/fs/hfs/ |
H A D | bnode.c | 269 spin_lock(&tree->hash_lock); in __hfs_bnode_create() 278 spin_unlock(&tree->hash_lock); in __hfs_bnode_create() 283 spin_unlock(&tree->hash_lock); in __hfs_bnode_create() 324 spin_lock(&tree->hash_lock); in hfs_bnode_find() 328 spin_unlock(&tree->hash_lock); in hfs_bnode_find() 334 spin_unlock(&tree->hash_lock); in hfs_bnode_find() 417 spin_lock(&tree->hash_lock); in hfs_bnode_create() 419 spin_unlock(&tree->hash_lock); in hfs_bnode_create() 468 if (!atomic_dec_and_lock(&node->refcnt, &tree->hash_lock)) in hfs_bnode_put() 478 spin_unlock(&tree->hash_lock); in hfs_bnode_put() [all …]
|
H A D | inode.c | 93 spin_lock(&tree->hash_lock); in hfs_release_folio() 103 spin_unlock(&tree->hash_lock); in hfs_release_folio() 107 spin_lock(&tree->hash_lock); in hfs_release_folio() 119 spin_unlock(&tree->hash_lock); in hfs_release_folio()
|
H A D | btree.h | 47 spinlock_t hash_lock; member
|
H A D | btree.c | 32 spin_lock_init(&tree->hash_lock); in hfs_btree_open()
|
/linux/drivers/net/ppp/ |
H A D | pppoe.c | 104 rwlock_t hash_lock; member 228 read_lock_bh(&pn->hash_lock); in get_item() 232 read_unlock_bh(&pn->hash_lock); in get_item() 261 write_lock_bh(&pn->hash_lock); in delete_item() 263 write_unlock_bh(&pn->hash_lock); in delete_item() 279 write_lock_bh(&pn->hash_lock); in pppoe_flush_dev() 303 write_unlock_bh(&pn->hash_lock); in pppoe_flush_dev() 323 write_lock_bh(&pn->hash_lock); in pppoe_flush_dev() 327 write_unlock_bh(&pn->hash_lock); in pppoe_flush_dev() 684 write_lock_bh(&pn->hash_lock); in pppoe_connect() [all …]
|
/linux/net/sunrpc/ |
H A D | cache.c | 86 /* Must be called under cd->hash_lock */ in sunrpc_begin_cache_remove_entry() 116 spin_lock(&detail->hash_lock); in sunrpc_cache_add_entry() 120 lockdep_is_held(&detail->hash_lock)) { in sunrpc_cache_add_entry() 131 spin_unlock(&detail->hash_lock); in sunrpc_cache_add_entry() 139 spin_unlock(&detail->hash_lock); in sunrpc_cache_add_entry() 212 spin_lock(&detail->hash_lock); in sunrpc_cache_update() 216 spin_unlock(&detail->hash_lock); in sunrpc_cache_update() 220 spin_unlock(&detail->hash_lock); in sunrpc_cache_update() 231 spin_lock(&detail->hash_lock); in sunrpc_cache_update() 238 spin_unlock(&detail->hash_lock); in sunrpc_cache_update() [all...] |
/linux/fs/hfsplus/ |
H A D | bnode.c | 419 spin_lock(&tree->hash_lock); in __hfs_bnode_create() 427 spin_unlock(&tree->hash_lock); in __hfs_bnode_create() 433 spin_unlock(&tree->hash_lock); in __hfs_bnode_create() 474 spin_lock(&tree->hash_lock); in hfs_bnode_find() 478 spin_unlock(&tree->hash_lock); in hfs_bnode_find() 485 spin_unlock(&tree->hash_lock); in hfs_bnode_find() 568 spin_lock(&tree->hash_lock); in hfs_bnode_create() 570 spin_unlock(&tree->hash_lock); in hfs_bnode_create() 619 if (!atomic_dec_and_lock(&node->refcnt, &tree->hash_lock)) in hfs_bnode_put() 629 spin_unlock(&tree->hash_lock); in hfs_bnode_put() [all …]
|
H A D | inode.c | 89 spin_lock(&tree->hash_lock); in hfsplus_release_folio() 99 spin_unlock(&tree->hash_lock); in hfsplus_release_folio() 104 spin_lock(&tree->hash_lock); in hfsplus_release_folio() 116 spin_unlock(&tree->hash_lock); in hfsplus_release_folio()
|
H A D | btree.c | 147 spin_lock_init(&tree->hash_lock); in hfs_btree_open()
|
H A D | hfsplus_fs.h | 96 spinlock_t hash_lock; member
|
/linux/drivers/md/dm-vdo/ |
H A D | dedupe.h | 30 struct hash_lock; 55 struct hash_lock *lock_array;
|
H A D | statistics.h | 240 struct hash_lock_statistics hash_lock; member
|
H A D | message-stats.c | 370 write_hash_lock_statistics("hashLock : ", &stats->hash_lock, ", ", buf, maxlen); in write_vdo_statistics()
|
/linux/Documentation/bpf/ |
H A D | map_lru_hash_update.dot | 12 hash_lock [shape=rectangle,fillcolor=3,label="hashtab lock"] 152 lru_lock -> hash_lock [style=invis] 153 hash_lock -> remote_lock [style=invis]
|
/linux/kernel/time/ |
H A D | posix-timers.c | 50 static DEFINE_SPINLOCK(hash_lock); 81 hlist_for_each_entry_rcu(timer, head, t_hash, lockdep_is_held(&hash_lock)) { in __posix_timers_find() 108 spin_lock(&hash_lock); in posix_timer_add() 117 spin_unlock(&hash_lock); in posix_timer_add() 120 spin_unlock(&hash_lock); in posix_timer_add() 376 spin_lock(&hash_lock); in posix_timer_unhash_and_free() 378 spin_unlock(&hash_lock); in posix_timer_unhash_and_free()
|
/linux/drivers/net/vxlan/ |
H A D | vxlan_core.c | 523 spin_lock_bh(&vxlan->hash_lock[h]); in vxlan_fdb_replay() 535 spin_unlock_bh(&vxlan->hash_lock[h]); in vxlan_fdb_replay() 540 spin_unlock_bh(&vxlan->hash_lock[h]); in vxlan_fdb_replay() 557 spin_lock_bh(&vxlan->hash_lock[h]); in vxlan_fdb_clear_offload() 562 spin_unlock_bh(&vxlan->hash_lock[h]); in vxlan_fdb_clear_offload() 1273 spin_lock_bh(&vxlan->hash_lock[hash_index]); in vxlan_fdb_add() 1278 spin_unlock_bh(&vxlan->hash_lock[hash_index]); in vxlan_fdb_add() 1339 spin_lock_bh(&vxlan->hash_lock[hash_index]); in vxlan_fdb_delete() 1342 spin_unlock_bh(&vxlan->hash_lock[hash_index]); in vxlan_fdb_delete() 1489 spin_lock(&vxlan->hash_lock[hash_index]); in vxlan_snoop() [all …]
|
H A D | vxlan_vnifilter.c | 485 spin_lock_bh(&vxlan->hash_lock[hash_index]); in vxlan_update_default_fdb_entry() 497 spin_unlock_bh(&vxlan->hash_lock[hash_index]); in vxlan_update_default_fdb_entry() 510 spin_unlock_bh(&vxlan->hash_lock[hash_index]); in vxlan_update_default_fdb_entry()
|
/linux/include/linux/sunrpc/ |
H A D | cache.h | 74 spinlock_t hash_lock; member
|
/linux/include/net/ |
H A D | vxlan.h | 298 spinlock_t hash_lock[FDB_HASH_SIZE]; member
|
/linux/Documentation/admin-guide/device-mapper/ |
H A D | vdo-design.rst | 433 7. The data_vio reserves or joins a struct hash_lock, which manages all of 441 and sets itself as the new hash lock's "agent." The hash_lock pool is
|