Home
last modified time | relevance | path

Searched refs:alloc_lock (Results 1 – 23 of 23) sorted by relevance

/linux/drivers/uio/
H A Duio_dmem_genirq.c39 struct mutex alloc_lock; member
55 mutex_lock(&priv->alloc_lock); in uio_dmem_genirq_open()
68 mutex_unlock(&priv->alloc_lock); in uio_dmem_genirq_open()
84 mutex_lock(&priv->alloc_lock); in uio_dmem_genirq_release()
99 mutex_unlock(&priv->alloc_lock); in uio_dmem_genirq_release()
200 mutex_init(&priv->alloc_lock); in uio_dmem_genirq_probe()
/linux/mm/
H A Dswap_slots.c148 mutex_init(&cache->alloc_lock); in alloc_swap_slot_cache()
156 * We initialized alloc_lock and free_lock earlier. We use in alloc_swap_slot_cache()
176 mutex_lock(&cache->alloc_lock); in drain_slots_cache_cpu()
184 mutex_unlock(&cache->alloc_lock); in drain_slots_cache_cpu()
322 * mutex cache->alloc_lock. in folio_alloc_swap()
330 mutex_lock(&cache->alloc_lock); in folio_alloc_swap()
341 mutex_unlock(&cache->alloc_lock); in folio_alloc_swap()
H A Dvmscan.c2860 lockdep_assert_held(&task->alloc_lock); in lru_gen_migrate_mm()
/linux/drivers/misc/
H A Dhpilo.c547 spin_lock_irqsave(&hw->alloc_lock, flags); in ilo_close()
549 spin_unlock_irqrestore(&hw->alloc_lock, flags); in ilo_close()
594 spin_lock_irqsave(&hw->alloc_lock, flags); in ilo_open()
597 spin_unlock_irqrestore(&hw->alloc_lock, flags); in ilo_open()
603 spin_lock_irqsave(&hw->alloc_lock, flags); in ilo_open()
605 spin_unlock_irqrestore(&hw->alloc_lock, flags); in ilo_open()
651 spin_lock(&hw->alloc_lock); in ilo_isr()
656 spin_unlock(&hw->alloc_lock); in ilo_isr()
676 spin_unlock(&hw->alloc_lock); in ilo_isr()
806 spin_lock_init(&ilo_hw->alloc_lock); in ilo_probe()
H A Dhpilo.h61 spinlock_t alloc_lock; member
/linux/init/
H A Dinit_task.c126 .alloc_lock = __SPIN_LOCK_UNLOCKED(init_task.alloc_lock),
161 &init_task.alloc_lock),
/linux/include/linux/sched/
H A Dtask.h223 spin_lock(&p->alloc_lock); in task_lock()
228 spin_unlock(&p->alloc_lock); in task_unlock()
/linux/drivers/infiniband/hw/hfi1/
H A Dpio.c700 spin_lock_init(&sc->alloc_lock); in sc_alloc()
887 spin_lock_irq(&sc->alloc_lock); in sc_disable()
931 spin_unlock_irq(&sc->alloc_lock); in sc_disable()
1261 spin_lock_irqsave(&sc->alloc_lock, flags); in sc_enable()
1332 spin_unlock_irqrestore(&sc->alloc_lock, flags); in sc_enable()
1387 spin_lock_irqsave(&sc->alloc_lock, flags); in sc_stop()
1391 spin_unlock_irqrestore(&sc->alloc_lock, flags); in sc_stop()
1420 spin_lock_irqsave(&sc->alloc_lock, flags); in sc_buffer_alloc()
1422 spin_unlock_irqrestore(&sc->alloc_lock, flags); in sc_buffer_alloc()
1431 spin_unlock_irqrestore(&sc->alloc_lock, flags); in sc_buffer_alloc()
[all …]
H A Dpio.h73 spinlock_t alloc_lock ____cacheline_aligned_in_smp;
/linux/include/linux/
H A Dswap_slots.h15 struct mutex alloc_lock; /* protects slots, nr, cur */ member
H A Dioprio.h66 lockdep_assert_held(&p->alloc_lock); in __get_task_ioprio()
H A Dsched.h1177 spinlock_t alloc_lock;
1261 /* Protected by ->alloc_lock: */
1270 /* cg_list protected by css_set_lock and tsk->alloc_lock: */
1297 /* Protected by alloc_lock: */
1157 spinlock_t alloc_lock; global() member
/linux/drivers/net/ethernet/mellanox/mlx5/core/
H A Dcmd.c164 spin_lock_irqsave(&cmd->alloc_lock, flags); in cmd_alloc_index()
171 spin_unlock_irqrestore(&cmd->alloc_lock, flags); in cmd_alloc_index()
178 lockdep_assert_held(&cmd->alloc_lock); in cmd_free_index()
192 spin_lock_irqsave(&cmd->alloc_lock, flags); in cmd_ent_put()
203 spin_unlock_irqrestore(&cmd->alloc_lock, flags); in cmd_ent_put()
1021 spin_lock_irqsave(&cmd->alloc_lock, flags); in cmd_work_handler()
1024 spin_unlock_irqrestore(&cmd->alloc_lock, flags); in cmd_work_handler()
1778 spin_lock_irqsave(&dev->cmd.alloc_lock, flags); in mlx5_cmd_trigger_completions()
1791 spin_unlock_irqrestore(&dev->cmd.alloc_lock, flags); in mlx5_cmd_trigger_completions()
1800 spin_unlock_irqrestore(&dev->cmd.alloc_lock, flags); in mlx5_cmd_trigger_completions()
[all …]
/linux/fs/bcachefs/
H A Dbtree_cache.c602 if (bc->alloc_lock == current) { in bch2_btree_cache_cannibalize_unlock()
604 bc->alloc_lock = NULL; in bch2_btree_cache_cannibalize_unlock()
616 if (try_cmpxchg(&bc->alloc_lock, &old, current) || old == current) in bch2_btree_cache_cannibalize_lock()
628 if (try_cmpxchg(&bc->alloc_lock, &old, current) || old == current) { in bch2_btree_cache_cannibalize_lock()
761 if (bc->alloc_lock == current) { in bch2_btree_node_mem_alloc()
1361 prt_printf(out, "cannibalize lock:\t%p\n", bc->alloc_lock); in bch2_btree_cache_to_text()
H A Dbtree_types.h187 struct task_struct *alloc_lock; member
/linux/arch/powerpc/kvm/
H A Dbook3s_64_vio.c227 mutex_lock(&stt->alloc_lock); in kvm_spapr_get_tce_page()
235 mutex_unlock(&stt->alloc_lock); in kvm_spapr_get_tce_page()
328 mutex_init(&stt->alloc_lock); in kvm_vm_ioctl_create_spapr_tce()
/linux/security/yama/
H A Dyama_lsm.c81 assert_spin_locked(&target->alloc_lock); /* for target->comm */ in report_access()
/linux/arch/powerpc/include/asm/
H A Dkvm_host.h198 struct mutex alloc_lock; member
/linux/include/linux/mlx5/
H A Ddriver.h316 spinlock_t alloc_lock; member
/linux/drivers/net/ethernet/mellanox/mlx4/
H A Dresource_tracker.c321 spin_lock(&res_alloc->alloc_lock); in mlx4_grant_resource()
374 spin_unlock(&res_alloc->alloc_lock); in mlx4_grant_resource()
390 spin_lock(&res_alloc->alloc_lock); in mlx4_release_resource()
419 spin_unlock(&res_alloc->alloc_lock); in mlx4_release_resource()
551 spin_lock_init(&res_alloc->alloc_lock); in mlx4_init_resource_tracker()
H A Dmlx4.h543 spinlock_t alloc_lock; /* protect quotas */ member
/linux/kernel/
H A Dfork.c2266 spin_lock_init(&p->alloc_lock); in copy_process()
2314 seqcount_spinlock_init(&p->mems_allowed_seq, &p->alloc_lock); in copy_process()
/linux/Documentation/admin-guide/cgroup-v1/
H A Dcgroups.rst535 - while holding the task's alloc_lock (via task_lock())