| /linux/samples/damon/ |
| H A D | mtier.c | 53 static struct damon_ctx *ctxs[2]; variable 179 ctxs[0] = ctx; in damon_sample_mtier_start() 182 damon_destroy_ctx(ctxs[0]); in damon_sample_mtier_start() 185 ctxs[1] = ctx; in damon_sample_mtier_start() 186 return damon_start(ctxs, 2, true); in damon_sample_mtier_start() 191 damon_stop(ctxs, 2); in damon_sample_mtier_stop() 192 damon_destroy_ctx(ctxs[0]); in damon_sample_mtier_stop() 193 damon_destroy_ctx(ctxs[1]); in damon_sample_mtier_stop()
|
| /linux/tools/testing/selftests/drivers/net/hw/ |
| H A D | rss_ctx.py | 395 ctxs = cfg.ethnl.rss_get({}, dump=True) 396 tuples = [(c['header']['dev-name'], c.get('context', -1)) for c in ctxs] 402 for data in ctxs: 414 ctxs = cfg.ethnl.rss_get({'header': {'dev-name': cfg.ifname}}, dump=True) 415 tuples = [(c['header']['dev-name'], c.get('context', -1)) for c in ctxs] 423 ctxs = cfg.ethnl.rss_get({'start-context': 1}, dump=True) 424 tuples = [(c['header']['dev-name'], c.get('context', -1)) for c in ctxs] 430 ctxs = cfg.ethnl.rss_get({'header': {'dev-name': cfg.ifname}, 'start-context': 1}, dump=True) 431 ctx_tuples = set([(c['header']['dev-name'], c.get('context', -1)) for c in ctxs]) 868 ctxs = cfg.ethnl.rss_get({'header': {'dev-name': cfg.ifname}}, dump=True) [all …]
|
| /linux/lib/crypto/tests/ |
| H A D | hash-test-template.h | 368 struct hash_irq_test2_hash_ctx ctxs[IRQ_TEST_NUM_BUFFERS]; member 380 for (ctx = &state->ctxs[0]; ctx < &state->ctxs[ARRAY_SIZE(state->ctxs)]; in hash_irq_test2_func() 385 if (WARN_ON_ONCE(ctx == &state->ctxs[ARRAY_SIZE(state->ctxs)])) { in hash_irq_test2_func()
|
| /linux/drivers/iommu/arm/arm-smmu/ |
| H A D | qcom_iommu.c | 54 struct qcom_iommu_ctx *ctxs[]; /* indexed by asid */ member 87 return qcom_iommu->ctxs[asid]; in to_ctx() 575 WARN_ON(qcom_iommu->ctxs[asid] == NULL)) in qcom_iommu_of_xlate() 730 qcom_iommu->ctxs[ctx->asid] = ctx; in qcom_iommu_ctx_probe() 742 qcom_iommu->ctxs[ctx->asid] = NULL; in qcom_iommu_ctx_remove() 788 qcom_iommu = devm_kzalloc(dev, struct_size(qcom_iommu, ctxs, max_asid + 1), in qcom_iommu_device_probe()
|
| /linux/tools/testing/selftests/damon/ |
| H A D | sysfs.py | 195 ctxs = kdamonds.kdamonds[0].contexts 197 assert_true(len(ctxs) == len(dump), 'ctxs length', dump) 198 for idx, ctx in enumerate(ctxs):
|
| /linux/block/ |
| H A D | blk-mq.c | 1803 struct blk_mq_ctx *ctx = hctx->ctxs[bitnr]; in flush_busy_ctx() 1837 struct blk_mq_ctx *ctx = hctx->ctxs[bitnr]; in dispatch_rq_from_ctx() 4059 hctx->ctxs = kmalloc_array_node(nr_cpu_ids, sizeof(void *), in blk_mq_alloc_hctx() 4061 if (!hctx->ctxs) in blk_mq_alloc_hctx() 4078 kfree(hctx->ctxs); in blk_mq_alloc_hctx() 4225 hctx->ctxs[hctx->nr_ctx++] = ctx; in blk_mq_map_swqueue() 4362 struct blk_mq_ctxs *ctxs; in blk_mq_alloc_ctxs() local 4365 ctxs = kzalloc_obj(*ctxs); in blk_mq_alloc_ctxs() 4366 if (!ctxs) in blk_mq_alloc_ctxs() 4369 ctxs->queue_ctx = alloc_percpu(struct blk_mq_ctx); in blk_mq_alloc_ctxs() [all …]
|
| H A D | blk-mq.h | 30 struct blk_mq_ctxs *ctxs; member
|
| H A D | blk-mq-sched.c | 202 return hctx->ctxs[idx]; in blk_mq_next_ctx()
|
| /linux/mm/damon/ |
| H A D | core.c | 1364 int damon_start(struct damon_ctx **ctxs, int nr_ctxs, bool exclusive) in damon_start() argument 1377 err = __damon_start(ctxs[i]); in damon_start() 1419 int damon_stop(struct damon_ctx **ctxs, int nr_ctxs) in damon_stop() argument 1425 err = __damon_stop(ctxs[i]); in damon_stop()
|
| /linux/drivers/android/binder/ |
| H A D | process.rs | 1544 let ctxs = crate::context::get_all_contexts()?; in ioctl_freeze() localVariable 1545 for ctx in ctxs { in ioctl_freeze()
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_gt_sriov_pf_debugfs.c | 504 DEFINE_SRIOV_GT_CONFIG_DEBUGFS_ATTRIBUTE(ctxs, u32, "%llu\n");
|
| /linux/kernel/ |
| H A D | workqueue.c | 6975 LIST_HEAD(ctxs); in workqueue_apply_unbound_cpumask() 6992 list_add_tail(&ctx->list, &ctxs); in workqueue_apply_unbound_cpumask() 6995 list_for_each_entry_safe(ctx, n, &ctxs, list) { in workqueue_apply_unbound_cpumask()
|