/linux/lib/ |
H A D | interval_tree.c | 30 struct interval_tree_node *cur = state->nodes[1]; in interval_tree_span_iter_next_gap() 32 state->nodes[0] = cur; in interval_tree_span_iter_next_gap() 34 if (cur->last > state->nodes[0]->last) in interval_tree_span_iter_next_gap() 35 state->nodes[0] = cur; in interval_tree_span_iter_next_gap() 38 } while (cur && (state->nodes[0]->last >= cur->start || in interval_tree_span_iter_next_gap() 39 state->nodes[0]->last + 1 == cur->start)); in interval_tree_span_iter_next_gap() 40 state->nodes[1] = cur; in interval_tree_span_iter_next_gap() 50 iter->nodes[0] = NULL; in interval_tree_span_iter_first() 51 iter->nodes[1] = in interval_tree_span_iter_first() 53 if (!iter->nodes[1]) { in interval_tree_span_iter_first() [all …]
|
H A D | objpool.c | 21 int nodes, void *context, in objpool_init_percpu_slot() argument 30 for (i = 0; i < nodes; i++) { in objpool_init_percpu_slot() 56 int nodes, size, rc; in objpool_init_percpu_slots() local 63 nodes = nr_objs / pool->nr_possible_cpus; in objpool_init_percpu_slots() 65 nodes++; in objpool_init_percpu_slots() 69 pool->obj_size * nodes; in objpool_init_percpu_slots() 96 rc = objpool_init_percpu_slot(pool, slot, nodes, context, objinit); in objpool_init_percpu_slots()
|
/linux/drivers/net/ethernet/wangxun/txgbe/ |
H A D | txgbe_phy.c | 25 struct txgbe_nodes *nodes = &txgbe->nodes; in txgbe_swnodes_register() local 32 snprintf(nodes->gpio_name, sizeof(nodes->gpio_name), "txgbe_gpio-%x", id); in txgbe_swnodes_register() 33 snprintf(nodes->i2c_name, sizeof(nodes->i2c_name), "txgbe_i2c-%x", id); in txgbe_swnodes_register() 34 snprintf(nodes->sfp_name, sizeof(nodes->sfp_name), "txgbe_sfp-%x", id); in txgbe_swnodes_register() 35 snprintf(nodes->phylink_name, sizeof(nodes->phylink_name), "txgbe_phylink-%x", id); in txgbe_swnodes_register() 37 swnodes = nodes->swnodes; in txgbe_swnodes_register() 46 nodes->gpio_props[0] = PROPERTY_ENTRY_STRING("pinctrl-names", "default"); in txgbe_swnodes_register() 47 swnodes[SWNODE_GPIO] = NODE_PROP(nodes->gpio_name, nodes->gpio_props); in txgbe_swnodes_register() 48 nodes->gpio0_ref[0] = SOFTWARE_NODE_REFERENCE(&swnodes[SWNODE_GPIO], 0, GPIO_ACTIVE_HIGH); in txgbe_swnodes_register() 49 nodes->gpio1_ref[0] = SOFTWARE_NODE_REFERENCE(&swnodes[SWNODE_GPIO], 1, GPIO_ACTIVE_HIGH); in txgbe_swnodes_register() [all …]
|
/linux/drivers/media/pci/intel/ |
H A D | ipu-bridge.c | 324 struct software_node *nodes = sensor->swnodes; in ipu_bridge_create_fwnode_properties() local 330 SOFTWARE_NODE_REFERENCE(&nodes[SWNODE_IVSC_SENSOR_ENDPOINT]); in ipu_bridge_create_fwnode_properties() 332 SOFTWARE_NODE_REFERENCE(&nodes[SWNODE_IVSC_IPU_ENDPOINT]); in ipu_bridge_create_fwnode_properties() 334 SOFTWARE_NODE_REFERENCE(&nodes[SWNODE_SENSOR_ENDPOINT]); in ipu_bridge_create_fwnode_properties() 336 SOFTWARE_NODE_REFERENCE(&nodes[SWNODE_IPU_ENDPOINT]); in ipu_bridge_create_fwnode_properties() 361 SOFTWARE_NODE_REFERENCE(&nodes[SWNODE_IPU_ENDPOINT]); in ipu_bridge_create_fwnode_properties() 363 SOFTWARE_NODE_REFERENCE(&nodes[SWNODE_SENSOR_ENDPOINT]); in ipu_bridge_create_fwnode_properties() 435 struct software_node *nodes = sensor->swnodes; in ipu_bridge_init_swnode_group() local 437 sensor->group[SWNODE_SENSOR_HID] = &nodes[SWNODE_SENSOR_HID]; in ipu_bridge_init_swnode_group() 438 sensor->group[SWNODE_SENSOR_PORT] = &nodes[SWNODE_SENSOR_PORT]; in ipu_bridge_init_swnode_group() [all …]
|
/linux/fs/bcachefs/ |
H A D | btree_node_scan.c | 42 …tic void found_btree_nodes_to_text(struct printbuf *out, struct bch_fs *c, found_btree_nodes nodes) in found_btree_nodes_to_text() argument 45 darray_for_each(nodes, i) { in found_btree_nodes_to_text() 221 if (darray_push(&f->nodes, n)) in try_read_btree_node() 372 if (f->nodes.nr) in bch2_scan_for_btree_nodes() 381 if (!f->nodes.nr) { in bch2_scan_for_btree_nodes() 390 found_btree_nodes_to_text(&buf, c, f->nodes); in bch2_scan_for_btree_nodes() 394 sort(f->nodes.data, f->nodes.nr, sizeof(f->nodes.data[0]), found_btree_node_cmp_cookie, NULL); in bch2_scan_for_btree_nodes() 397 darray_for_each(f->nodes, i) { in bch2_scan_for_btree_nodes() 398 struct found_btree_node *prev = dst ? f->nodes.data + dst - 1 : NULL; in bch2_scan_for_btree_nodes() 409 f->nodes.data[dst++] = *i; in bch2_scan_for_btree_nodes() [all …]
|
/linux/Documentation/sphinx/ |
H A D | rstFlatTable.py | 45 from docutils import nodes 94 class rowSpan(nodes.General, nodes.Element): pass # pylint: disable=C0103,C0321 95 class colSpan(nodes.General, nodes.Element): pass # pylint: disable=C0103,C0321 117 nodes.literal_block(self.block_text, self.block_text), 122 node = nodes.Element() # anonymous container for parsing 156 table = nodes.table() 157 tgroup = nodes.tgroup(cols=len(colwidths)) 162 colspec = nodes.colspec(colwidth=colwidth) 176 thead = nodes.thead() 181 tbody = nodes.tbody() [all …]
|
H A D | automarkup.py | 7 from docutils import nodes 113 repl.append(nodes.Text(t[done:m.start()])) 123 repl.append(nodes.Text(t[done:])) 146 target_text = nodes.Text(match.group(0)) 157 lit_text = nodes.literal(classes=['xref', 'c', 'c-func']) 206 target_text = nodes.Text(match.group(0)) 217 lit_text = nodes.literal(classes=['xref', 'c', class_str[match.re]]) 271 return nodes.Text(match.group(0)) 290 return nodes.reference('', nodes.Text(text), 300 if not isinstance(node, nodes.Text) or isinstance(node.parent, nodes.literal): [all …]
|
H A D | kfigure.py | 56 from docutils import nodes 61 from sphinx.util.nodes import clean_astext 90 node = nodes.literal_block(data, data) 438 class kernel_image(nodes.image): 457 if len(result) == 2 or isinstance(result[0], nodes.system_message): 475 class kernel_figure(nodes.figure): 494 if len(result) == 2 or isinstance(result[0], nodes.system_message): 541 img_node = nodes.image(node.rawsource, **node.attributes) 550 class kernel_render(nodes.General, nodes.Inline, nodes.Element): 597 literal_node = nodes.literal_block(code, code) [all …]
|
/linux/drivers/md/persistent-data/ |
H A D | dm-btree-spine.c | 128 s->nodes[0] = NULL; in init_ro_spine() 129 s->nodes[1] = NULL; in init_ro_spine() 137 unlock_block(s->info, s->nodes[i]); in exit_ro_spine() 145 unlock_block(s->info, s->nodes[0]); in ro_step() 146 s->nodes[0] = s->nodes[1]; in ro_step() 150 r = bn_read_lock(s->info, new_child, s->nodes + s->count); in ro_step() 161 unlock_block(s->info, s->nodes[s->count]); in ro_pop() 169 block = s->nodes[s->count - 1]; in ro_node() 187 unlock_block(s->info, s->nodes[i]); in exit_shadow_spine() 196 unlock_block(s->info, s->nodes[0]); in shadow_step() [all …]
|
/linux/arch/arm/mach-sunxi/ |
H A D | mc_smp.c | 702 int (*get_smp_nodes)(struct sunxi_mc_smp_nodes *nodes); 706 static void __init sunxi_mc_smp_put_nodes(struct sunxi_mc_smp_nodes *nodes) in sunxi_mc_smp_put_nodes() argument 708 of_node_put(nodes->prcm_node); in sunxi_mc_smp_put_nodes() 709 of_node_put(nodes->cpucfg_node); in sunxi_mc_smp_put_nodes() 710 of_node_put(nodes->sram_node); in sunxi_mc_smp_put_nodes() 711 of_node_put(nodes->r_cpucfg_node); in sunxi_mc_smp_put_nodes() 712 memset(nodes, 0, sizeof(*nodes)); in sunxi_mc_smp_put_nodes() 715 static int __init sun9i_a80_get_smp_nodes(struct sunxi_mc_smp_nodes *nodes) in sun9i_a80_get_smp_nodes() argument 717 nodes->prcm_node = of_find_compatible_node(NULL, NULL, in sun9i_a80_get_smp_nodes() 719 if (!nodes->prcm_node) { in sun9i_a80_get_smp_nodes() [all …]
|
/linux/tools/perf/tests/ |
H A D | mem2node.c | 50 struct memory_node nodes[3]; in test__mem2node() local 52 .memory_nodes = (struct memory_node *) &nodes[0], in test__mem2node() 53 .nr_memory_nodes = ARRAY_SIZE(nodes), in test__mem2node() 58 for (i = 0; i < ARRAY_SIZE(nodes); i++) { in test__mem2node() 59 nodes[i].node = test_nodes[i].node; in test__mem2node() 60 nodes[i].size = 10; in test__mem2node() 63 (nodes[i].set = get_bitmap(test_nodes[i].map, 10))); in test__mem2node() 75 for (i = 0; i < ARRAY_SIZE(nodes); i++) in test__mem2node() 76 zfree(&nodes[i].set); in test__mem2node()
|
/linux/mm/ |
H A D | mempolicy.c | 219 int (*create)(struct mempolicy *pol, const nodemask_t *nodes); 220 void (*rebind)(struct mempolicy *pol, const nodemask_t *nodes); 236 static int mpol_new_nodemask(struct mempolicy *pol, const nodemask_t *nodes) in mpol_new_nodemask() argument 238 if (nodes_empty(*nodes)) in mpol_new_nodemask() 240 pol->nodes = *nodes; in mpol_new_nodemask() 244 static int mpol_new_preferred(struct mempolicy *pol, const nodemask_t *nodes) in mpol_new_preferred() argument 246 if (nodes_empty(*nodes)) in mpol_new_preferred() 249 nodes_clear(pol->nodes); in mpol_new_preferred() 250 node_set(first_node(*nodes), pol->nodes); in mpol_new_preferred() 263 const nodemask_t *nodes, struct nodemask_scratch *nsc) in mpol_set_nodemask() argument [all …]
|
/linux/Documentation/devicetree/bindings/cpu/ |
H A D | cpu-topology.txt | 20 For instance in a system where CPUs support SMT, "cpu" nodes represent all 22 In systems where SMT is not supported "cpu" nodes represent all cores present 25 CPU topology bindings allow one to associate cpu nodes with hierarchical groups 27 tree nodes. 32 The cpu nodes, as per bindings defined in [4], represent the devices that 35 A topology description containing phandles to cpu nodes that are not compliant 44 nodes are listed. 60 The cpu-map node's child nodes can be: 62 - one or more cluster nodes or 63 - one or more socket nodes in a multi-socket system [all …]
|
/linux/drivers/gpu/drm/tests/ |
H A D | drm_mm_test.c | 193 struct drm_mm_node nodes[2]; in drm_test_mm_debug() local 200 memset(nodes, 0, sizeof(nodes)); in drm_test_mm_debug() 201 nodes[0].start = 512; in drm_test_mm_debug() 202 nodes[0].size = 1024; in drm_test_mm_debug() 203 KUNIT_ASSERT_FALSE_MSG(test, drm_mm_reserve_node(&mm, &nodes[0]), in drm_test_mm_debug() 205 nodes[0].start, nodes[0].size); in drm_test_mm_debug() 207 nodes[1].size = 1024; in drm_test_mm_debug() 208 nodes[1].start = 4096 - 512 - nodes[1].size; in drm_test_mm_debug() 209 KUNIT_ASSERT_FALSE_MSG(test, drm_mm_reserve_node(&mm, &nodes[1]), in drm_test_mm_debug() 211 nodes[0].start, nodes[0].size); in drm_test_mm_debug()
|
/linux/fs/btrfs/ |
H A D | ctree.c | 166 if (!p->nodes[i]) in btrfs_release_path() 169 btrfs_tree_unlock_rw(p->nodes[i], p->locks[i]); in btrfs_release_path() 172 free_extent_buffer(p->nodes[i]); in btrfs_release_path() 173 p->nodes[i] = NULL; in btrfs_release_path() 866 mid = path->nodes[level]; in balance_level() 874 parent = path->nodes[level + 1]; in balance_level() 917 path->nodes[level] = NULL; in balance_level() 1086 path->nodes[level] = left; in balance_level() 1100 btrfs_node_blockptr(path->nodes[level], path->slots[level])) in balance_level() 1108 if (path->nodes[level] != left) in balance_level() [all …]
|
H A D | inode-item.c | 101 return btrfs_find_name_in_ext_backref(path->nodes[0], path->slots[0], in btrfs_lookup_inode_extref() 141 extref = btrfs_find_name_in_ext_backref(path->nodes[0], path->slots[0], in btrfs_del_inode_extref() 149 leaf = path->nodes[0]; in btrfs_del_inode_extref() 210 ref = btrfs_find_name_in_backref(path->nodes[0], path->slots[0], name); in btrfs_del_inode_ref() 216 leaf = path->nodes[0]; in btrfs_del_inode_ref() 278 if (btrfs_find_name_in_ext_backref(path->nodes[0], in btrfs_insert_inode_extref() 290 leaf = path->nodes[0]; in btrfs_insert_inode_extref() 295 btrfs_set_inode_extref_name_len(path->nodes[0], extref, name->len); in btrfs_insert_inode_extref() 296 btrfs_set_inode_extref_index(path->nodes[0], extref, index); in btrfs_insert_inode_extref() 297 btrfs_set_inode_extref_parent(path->nodes[0], extref, ref_objectid); in btrfs_insert_inode_extref() [all …]
|
H A D | file-item.c | 180 leaf = path->nodes[0]; in btrfs_insert_hole_extent() 220 leaf = path->nodes[0]; in btrfs_lookup_csum() 295 if (path->nodes[0]) { in search_csum_tree() 296 item = btrfs_item_ptr(path->nodes[0], path->slots[0], in search_csum_tree() 298 btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); in search_csum_tree() 299 itemsize = btrfs_item_size(path->nodes[0], path->slots[0]); in search_csum_tree() 316 btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); in search_csum_tree() 317 itemsize = btrfs_item_size(path->nodes[0], path->slots[0]); in search_csum_tree() 326 read_extent_buffer(path->nodes[0], dst, (unsigned long)item, in search_csum_tree() 494 leaf = path->nodes[0]; in btrfs_lookup_csums_list() [all …]
|
/linux/drivers/platform/surface/ |
H A D | surface_aggregator_registry.c | 490 const struct software_node **nodes; in ssam_platform_hub_probe() local 497 nodes = (const struct software_node **)acpi_device_get_match_data(&pdev->dev); in ssam_platform_hub_probe() 498 if (!nodes) { in ssam_platform_hub_probe() 508 nodes = (const struct software_node **)match->data; in ssam_platform_hub_probe() 509 if (!nodes) in ssam_platform_hub_probe() 524 status = software_node_register_node_group(nodes); in ssam_platform_hub_probe() 530 software_node_unregister_node_group(nodes); in ssam_platform_hub_probe() 539 software_node_unregister_node_group(nodes); in ssam_platform_hub_probe() 542 platform_set_drvdata(pdev, nodes); in ssam_platform_hub_probe() 548 const struct software_node **nodes = platform_get_drvdata(pdev); in ssam_platform_hub_remove() local [all …]
|
/linux/sound/hda/ |
H A D | hdac_sysfs.c | 16 struct kobject **nodes; member 328 if (tree->nodes) { in widget_tree_free() 329 for (p = tree->nodes; *p; p++) in widget_tree_free() 331 kfree(tree->nodes); in widget_tree_free() 377 tree->nodes = kcalloc(codec->num_nodes + 1, sizeof(*tree->nodes), in widget_tree_create() 379 if (!tree->nodes) in widget_tree_create() 384 &tree->nodes[i]); in widget_tree_create() 439 tree->nodes = kcalloc(num_nodes + 1, sizeof(*tree->nodes), GFP_KERNEL); in hda_widget_sysfs_reinit() 440 if (!tree->nodes) { in hda_widget_sysfs_reinit() 448 free_widget_node(codec->widgets->nodes[i], in hda_widget_sysfs_reinit() [all …]
|
/linux/arch/sparc/kernel/ |
H A D | cpumap.c | 53 struct cpuinfo_node nodes[] __counted_by(total_nodes); 197 new_tree = kzalloc(struct_size(new_tree, nodes, n), GFP_ATOMIC); in build_cpuinfo_tree() 211 node = &new_tree->nodes[n]; in build_cpuinfo_tree() 252 node = &new_tree->nodes[level_rover[level]]; in build_cpuinfo_tree() 277 node = &new_tree->nodes[n]; in build_cpuinfo_tree() 299 struct cpuinfo_node *node = &t->nodes[node_index]; in increment_rover() 302 top_level = t->nodes[root_index].level; in increment_rover() 314 node = &t->nodes[node->parent_index]; in increment_rover() 340 for (level = t->nodes[root_index].level; level < CPUINFO_LVL_MAX; in iterate_cpu() 342 new_index = t->nodes[index].rover; in iterate_cpu() [all …]
|
/linux/tools/testing/selftests/bpf/progs/ |
H A D | rbtree.c | 134 struct node_data *nodes[3][2] = {{NULL, NULL}, {NULL, NULL}, {NULL, NULL}}; in rbtree_add_and_remove_array() local 141 nodes[i][j] = bpf_obj_new(typeof(*nodes[i][j])); in rbtree_add_and_remove_array() 142 if (!nodes[i][j]) in rbtree_add_and_remove_array() 144 nodes[i][j]->key = i * 2 + j; in rbtree_add_and_remove_array() 151 bpf_rbtree_add(&groot_array[i], &nodes[i][j]->node, less); in rbtree_add_and_remove_array() 153 bpf_rbtree_add(&groot_array_one[0], &nodes[2][j]->node, less); in rbtree_add_and_remove_array() 154 res1 = bpf_rbtree_remove(&groot_array[0], &nodes[0][0]->node); in rbtree_add_and_remove_array() 155 res2 = bpf_rbtree_remove(&groot_array[1], &nodes[1][0]->node); in rbtree_add_and_remove_array() 156 res3 = bpf_rbtree_remove(&groot_array_one[0], &nodes[2][0]->node); in rbtree_add_and_remove_array() 182 if (nodes[i][j]) in rbtree_add_and_remove_array() [all …]
|
/linux/Documentation/admin-guide/mm/ |
H A D | numa_memory_policy.rst | 17 which is an administrative mechanism for restricting the nodes from which 40 allocations across all nodes with "sufficient" memory, so as 164 an optional set of nodes. The mode determines the behavior of the 166 and the optional set of nodes can be viewed as the arguments to the 188 does not use the optional set of nodes. 190 It is an error for the set of nodes specified for this policy to 195 nodes specified by the policy. Memory will be allocated from 202 allocation fails, the kernel will search other nodes, in order 222 page granularity, across the nodes specified in the policy. 227 Interleave mode indexes the set of nodes specified by the [all …]
|
/linux/Documentation/driver-api/md/ |
H A D | md-cluster.rst | 54 node may write to those sectors. This is used when a new nodes 60 Each node has to communicate with other nodes when starting or ending 70 Normally all nodes hold a concurrent-read lock on this device. 75 Messages can be broadcast to all nodes, and the sender waits for all 76 other nodes to acknowledge the message before proceeding. Only one 87 informs other nodes that the metadata has 94 informs other nodes that a resync is initiated or 104 informs other nodes that a device is being added to 128 The DLM LVB is used to communicate within nodes of the cluster. There 145 acknowledged by all nodes in the cluster. The BAST of the resource [all …]
|
/linux/drivers/clk/zynqmp/ |
H A D | clk-zynqmp.h | 73 const struct clock_topology *nodes); 78 const struct clock_topology *nodes); 84 const struct clock_topology *nodes); 89 const struct clock_topology *nodes); 95 const struct clock_topology *nodes);
|
/linux/Documentation/filesystems/ |
H A D | ubifs-authentication.rst | 80 - *Index*: an on-flash B+ tree where the leaf nodes contain filesystem data 98 Basic on-flash UBIFS entities are called *nodes*. UBIFS knows different types 99 of nodes. Eg. data nodes (``struct ubifs_data_node``) which store chunks of file 100 contents or inode nodes (``struct ubifs_ino_node``) which represent VFS inodes. 101 Almost all types of nodes share a common header (``ubifs_ch``) containing basic 104 and some less important node types like padding nodes which are used to pad 108 as *wandering tree*, where only the changed nodes are re-written and previous 121 a dirty-flag which marks nodes that have to be persisted the next time the 126 on-flash filesystem structures like the index. On every commit, the TNC nodes 135 any changes (in form of inode nodes, data nodes etc.) between commits [all …]
|