| /linux/lib/ |
| H A D | interval_tree_test.c | 16 __param(int, nnodes, 100, "Number of nodes in the interval tree"); 48 for (i = 0; i < nnodes; i++) { in init() 77 for (j = 0; j < nnodes; j++) in basic_check() 79 for (j = 0; j < nnodes; j++) in basic_check() 102 for (j = 0; j < nnodes; j++) in search_check() 124 for (j = 0; j < nnodes; j++) in search_check() 140 intxn1 = bitmap_alloc(nnodes, GFP_KERNEL); in intersection_range_check() 146 intxn2 = bitmap_alloc(nnodes, GFP_KERNEL); in intersection_range_check() 156 for (j = 0; j < nnodes; j++) in intersection_range_check() 171 bitmap_zero(intxn1, nnodes); in intersection_range_check() [all …]
|
| H A D | rbtree_test.c | 14 __param(int, nnodes, 100, "Number of nodes in the rb-tree"); 153 for (i = 0; i < nnodes; i++) { in init() 256 for (j = 0; j < nnodes; j++) in basic_check() 258 for (j = 0; j < nnodes; j++) in basic_check() 272 for (j = 0; j < nnodes; j++) in basic_check() 274 for (j = 0; j < nnodes; j++) in basic_check() 285 for (i = 0; i < nnodes; i++) in basic_check() 325 for (i = 0; i < nnodes; i++) in basic_check() 331 for (j = 0; j < nnodes; j++) { in basic_check() 335 for (j = 0; j < nnodes; j++) { in basic_check() [all …]
|
| /linux/fs/btrfs/ |
| H A D | ulist.c | 52 ulist->nnodes = 0; in ulist_init() 158 BUG_ON(ulist->nnodes == 0); in ulist_rbtree_erase() 159 ulist->nnodes--; in ulist_rbtree_erase() 233 ulist->nnodes++; in ulist_add_merge()
|
| H A D | ulist.h | 40 unsigned long nnodes; member
|
| H A D | qgroup.c | 2916 if (!roots || roots->nnodes == 0) in maybe_fs_roots() 2954 nr_new_roots = new_roots->nnodes; in btrfs_qgroup_account_extent() 2959 nr_old_roots = old_roots->nnodes; in btrfs_qgroup_account_extent()
|
| H A D | backref.c | 1918 const unsigned long prev_ref_count = ctx->refs.nnodes; in btrfs_is_data_extent_shared() 1965 if ((ctx->refs.nnodes - prev_ref_count) > 1) in btrfs_is_data_extent_shared()
|
| /linux/drivers/clk/zynqmp/ |
| H A D | clkc.c | 412 u32 *nnodes) in __zynqmp_clock_get_topology() argument 421 topology[*nnodes].type = type; in __zynqmp_clock_get_topology() 422 topology[*nnodes].flag = FIELD_GET(CLK_TOPOLOGY_FLAGS, in __zynqmp_clock_get_topology() 424 topology[*nnodes].type_flag = in __zynqmp_clock_get_topology() 427 topology[*nnodes].custom_type_flag = in __zynqmp_clock_get_topology() 430 (*nnodes)++; in __zynqmp_clock_get_topology()
|
| /linux/drivers/vfio/pci/pds/ |
| H A D | dirty.c | 243 struct rb_root_cached *ranges, u32 nnodes, in pds_vfio_dirty_enable() argument 253 u32 num_ranges = nnodes; in pds_vfio_dirty_enable() 283 vfio_combine_iova_ranges(ranges, nnodes, max_regions); in pds_vfio_dirty_enable() 618 struct rb_root_cached *ranges, u32 nnodes, in pds_vfio_dma_logging_start() argument 628 err = pds_vfio_dirty_enable(pds_vfio, ranges, nnodes, page_size); in pds_vfio_dma_logging_start()
|
| H A D | dirty.h | 38 struct rb_root_cached *ranges, u32 nnodes,
|
| /linux/drivers/vfio/ |
| H A D | vfio_main.c | 1063 u32 nnodes; in vfio_ioctl_device_feature_logging_start() local 1078 nnodes = control.num_ranges; in vfio_ioctl_device_feature_logging_start() 1079 if (!nnodes) in vfio_ioctl_device_feature_logging_start() 1082 if (nnodes > LOG_MAX_RANGES) in vfio_ioctl_device_feature_logging_start() 1086 nodes = kmalloc_objs(struct interval_tree_node, nnodes); in vfio_ioctl_device_feature_logging_start() 1090 for (i = 0; i < nnodes; i++) { in vfio_ioctl_device_feature_logging_start() 1118 ret = device->log_ops->log_start(device, &root, nnodes, in vfio_ioctl_device_feature_logging_start()
|
| /linux/drivers/vfio/pci/mlx5/ |
| H A D | cmd.h | 243 struct rb_root_cached *ranges, u32 nnodes, u64 *page_size);
|
| H A D | cmd.c | 913 struct rb_root_cached *ranges, u32 nnodes) in mlx5vf_create_tracker() argument 922 u32 num_ranges = nnodes; in mlx5vf_create_tracker() 933 vfio_combine_iova_ranges(ranges, nnodes, max_num_range); in mlx5vf_create_tracker() 1495 struct rb_root_cached *ranges, u32 nnodes, in mlx5vf_start_page_tracker() argument 1584 err = mlx5vf_create_tracker(mdev, mvdev, ranges, nnodes); in mlx5vf_start_page_tracker()
|
| /linux/mm/ |
| H A D | mempolicy.c | 2245 unsigned int target, nnodes; in interleave_nid() local 2249 nnodes = read_once_policy_nodemask(pol, &nodemask); in interleave_nid() 2250 if (!nnodes) in interleave_nid() 2252 target = ilx % nnodes; in interleave_nid() 2639 int nnodes, node; in alloc_pages_bulk_weighted_interleave() local 2651 nnodes = read_once_policy_nodemask(pol, &nodes); in alloc_pages_bulk_weighted_interleave() 2655 if (!nnodes) in alloc_pages_bulk_weighted_interleave() 2710 for (i = 0; i < nnodes; i++) { in alloc_pages_bulk_weighted_interleave()
|
| /linux/kernel/time/ |
| H A D | timer_migration.c | 1958 unsigned int nnodes = num_possible_nodes(); in tmigr_init() local 1983 cpus_per_node = DIV_ROUND_UP(ncpus, nnodes); in tmigr_init() 1990 nodelvl = DIV_ROUND_UP(order_base_2(nnodes), in tmigr_init()
|
| /linux/include/linux/ |
| H A D | vfio.h | 251 struct rb_root_cached *ranges, u32 nnodes, u64 *page_size);
|
| /linux/Documentation/bpf/ |
| H A D | map_lru_hash_update.dot | 44 label="Able to free\ntarget_free\nnodes?"]
|
| /linux/samples/vfio-mdev/ |
| H A D | mtty.c | 1289 u32 nnodes, u64 *page_size) in mtty_log_start() argument
|