Home
last modified time | relevance | path

Searched refs:mres (Results 1 – 11 of 11) sorted by relevance

/linux/drivers/vdpa/mlx5/core/
H A Dmr.c667 struct mlx5_vdpa_mr_resources *mres; in mlx5_vdpa_mr_gc_handler() local
671 mres = container_of(work, struct mlx5_vdpa_mr_resources, gc_dwork_ent.work); in mlx5_vdpa_mr_gc_handler()
673 if (atomic_read(&mres->shutdown)) { in mlx5_vdpa_mr_gc_handler()
674 mutex_lock(&mres->lock); in mlx5_vdpa_mr_gc_handler()
675 } else if (!mutex_trylock(&mres->lock)) { in mlx5_vdpa_mr_gc_handler()
676 queue_delayed_work(mres->wq_gc, &mres->gc_dwork_ent, in mlx5_vdpa_mr_gc_handler()
681 mvdev = container_of(mres, struct mlx5_vdpa_dev, mres); in mlx5_vdpa_mr_gc_handler()
683 list_for_each_entry_safe(mr, tmp, &mres->mr_gc_list_head, mr_list) { in mlx5_vdpa_mr_gc_handler()
687 mutex_unlock(&mres->lock); in mlx5_vdpa_mr_gc_handler()
693 struct mlx5_vdpa_mr_resources *mres = &mvdev->mres; in _mlx5_vdpa_put_mr() local
[all …]
H A Dmlx5_vdpa.h107 struct mlx5_vdpa_mr_resources mres; member
/linux/fs/ocfs2/dlm/
H A Ddlmrecovery.c51 static void dlm_init_migratable_lockres(struct dlm_migratable_lockres *mres,
56 struct dlm_migratable_lockres *mres,
62 struct dlm_migratable_lockres *mres);
863 struct dlm_migratable_lockres *mres; in dlm_request_all_locks_worker() local
874 mres = (struct dlm_migratable_lockres *)data; in dlm_request_all_locks_worker()
911 ret = dlm_send_one_lockres(dlm, res, mres, reco_master, in dlm_request_all_locks_worker()
1094 struct dlm_migratable_lockres *mres, in dlm_send_mig_lockres_msg() argument
1099 u64 mig_cookie = be64_to_cpu(mres->mig_cookie); in dlm_send_mig_lockres_msg()
1100 int mres_total_locks = be32_to_cpu(mres->total_locks); in dlm_send_mig_lockres_msg()
1102 u8 orig_flags = mres->flags, in dlm_send_mig_lockres_msg()
[all …]
H A Ddlmmaster.c2534 struct dlm_migratable_lockres *mres = NULL; in dlm_migrate_lockres() local
2552 mres = (struct dlm_migratable_lockres *) __get_free_page(GFP_NOFS); in dlm_migrate_lockres()
2553 if (!mres) { in dlm_migrate_lockres()
2637 ret = dlm_send_one_lockres(dlm, res, mres, target, in dlm_migrate_lockres()
2729 if (mres) in dlm_migrate_lockres()
2730 free_page((unsigned long)mres); in dlm_migrate_lockres()
H A Ddlmcommon.h991 struct dlm_migratable_lockres *mres,
/linux/drivers/mmc/host/
H A Dmmc_spi.c1075 int mres; in mmc_spi_set_ios() local
1079 mres = spi_setup(host->spi); in mmc_spi_set_ios()
1080 if (mres < 0) in mmc_spi_set_ios()
1098 if (mres == 0) { in mmc_spi_set_ios()
1100 mres = spi_setup(host->spi); in mmc_spi_set_ios()
1101 if (mres < 0) in mmc_spi_set_ios()
/linux/drivers/vdpa/mlx5/net/
H A Dmlx5_vnet.c942 vq_mr = mvdev->mres.mr[mvdev->mres.group2asid[MLX5_VDPA_DATAVQ_GROUP]]; in create_virtqueue()
946 vq_desc_mr = mvdev->mres.mr[mvdev->mres.group2asid[MLX5_VDPA_DATAVQ_DESC_GROUP]]; in create_virtqueue()
954 vq_mr = mvdev->mres.mr[mvdev->mres.group2asid[MLX5_VDPA_DATAVQ_GROUP]]; in create_virtqueue()
958 vq_desc_mr = mvdev->mres.mr[mvdev->mres.group2asid[MLX5_VDPA_DATAVQ_DESC_GROUP]]; in create_virtqueue()
1355 vq_mr = mvdev->mres.mr[mvdev->mres.group2asid[MLX5_VDPA_DATAVQ_GROUP]]; in fill_modify_virtqueue_cmd()
1364 desc_mr = mvdev->mres.mr[mvdev->mres.group2asid[MLX5_VDPA_DATAVQ_DESC_GROUP]]; in fill_modify_virtqueue_cmd()
1382 unsigned int asid = mvdev->mres.group2asid[MLX5_VDPA_DATAVQ_GROUP]; in modify_virtqueue_end()
1383 struct mlx5_vdpa_mr *vq_mr = mvdev->mres.mr[asid]; in modify_virtqueue_end()
1391 unsigned int asid = mvdev->mres.group2asid[MLX5_VDPA_DATAVQ_DESC_GROUP]; in modify_virtqueue_end()
1392 struct mlx5_vdpa_mr *desc_mr = mvdev->mres.mr[asid]; in modify_virtqueue_end()
[all …]
/linux/sound/core/
H A Dpcm_compat.c57 struct snd_mask mres[5]; /* reserved masks */ member
/linux/tools/include/uapi/sound/
H A Dasound.h
/linux/include/uapi/sound/
H A Dasound.h412 struct snd_mask mres[5]; /* reserved masks */ member
/linux/tools/perf/trace/beauty/include/uapi/sound/
H A Dasound.h412 struct snd_mask mres[5]; /* reserved masks */ member