/Linux-v5.10/drivers/gpu/drm/selftests/ |
D | test-drm_mm.c | 268 struct drm_mm_node nodes[2]; in igt_debug() local 277 memset(nodes, 0, sizeof(nodes)); in igt_debug() 278 nodes[0].start = 512; in igt_debug() 279 nodes[0].size = 1024; in igt_debug() 280 ret = drm_mm_reserve_node(&mm, &nodes[0]); in igt_debug() 283 nodes[0].start, nodes[0].size); in igt_debug() 287 nodes[1].size = 1024; in igt_debug() 288 nodes[1].start = 4096 - 512 - nodes[1].size; in igt_debug() 289 ret = drm_mm_reserve_node(&mm, &nodes[1]); in igt_debug() 292 nodes[1].start, nodes[1].size); in igt_debug() [all …]
|
/Linux-v5.10/Documentation/sphinx/ |
D | rstFlatTable.py | 47 from docutils import nodes 104 class rowSpan(nodes.General, nodes.Element): pass # pylint: disable=C0103,C0321 105 class colSpan(nodes.General, nodes.Element): pass # pylint: disable=C0103,C0321 127 nodes.literal_block(self.block_text, self.block_text), 132 node = nodes.Element() # anonymous container for parsing 166 table = nodes.table() 167 tgroup = nodes.tgroup(cols=len(colwidths)) 172 colspec = nodes.colspec(colwidth=colwidth) 186 thead = nodes.thead() 191 tbody = nodes.tbody() [all …]
|
D | automarkup.py | 7 from docutils import nodes 106 repl.append(nodes.Text(t[done:m.start()])) 116 repl.append(nodes.Text(t[done:])) 132 target_text = nodes.Text(match.group(0)) 136 lit_text = nodes.literal(classes=['xref', 'c', class_s]) 183 target_text = nodes.Text(match.group(0)) 187 lit_text = nodes.literal(classes=['xref', 'c', class_str[match.re]]) 240 return nodes.Text(match.group(0)) 252 for para in doctree.traverse(nodes.paragraph): 253 for node in para.traverse(nodes.Text): [all …]
|
D | kfigure.py | 54 from docutils import nodes 60 from sphinx.util.nodes import clean_astext 104 node = nodes.literal_block(data, data) 346 class kernel_image(nodes.image): 365 if len(result) == 2 or isinstance(result[0], nodes.system_message): 383 class kernel_figure(nodes.figure): 402 if len(result) == 2 or isinstance(result[0], nodes.system_message): 449 img_node = nodes.image(node.rawsource, **node.attributes) 458 class kernel_render(nodes.General, nodes.Inline, nodes.Element): 505 literal_node = nodes.literal_block(code, code) [all …]
|
/Linux-v5.10/tools/perf/tests/ |
D | mem2node.c | 49 struct memory_node nodes[3]; in test__mem2node() local 51 .memory_nodes = (struct memory_node *) &nodes[0], in test__mem2node() 52 .nr_memory_nodes = ARRAY_SIZE(nodes), in test__mem2node() 57 for (i = 0; i < ARRAY_SIZE(nodes); i++) { in test__mem2node() 58 nodes[i].node = test_nodes[i].node; in test__mem2node() 59 nodes[i].size = 10; in test__mem2node() 62 (nodes[i].set = get_bitmap(test_nodes[i].map, 10))); in test__mem2node() 74 for (i = 0; i < ARRAY_SIZE(nodes); i++) in test__mem2node() 75 zfree(&nodes[i].set); in test__mem2node()
|
/Linux-v5.10/arch/arm/mach-sunxi/ |
D | mc_smp.c | 703 int (*get_smp_nodes)(struct sunxi_mc_smp_nodes *nodes); 707 static void __init sunxi_mc_smp_put_nodes(struct sunxi_mc_smp_nodes *nodes) in sunxi_mc_smp_put_nodes() argument 709 of_node_put(nodes->prcm_node); in sunxi_mc_smp_put_nodes() 710 of_node_put(nodes->cpucfg_node); in sunxi_mc_smp_put_nodes() 711 of_node_put(nodes->sram_node); in sunxi_mc_smp_put_nodes() 712 of_node_put(nodes->r_cpucfg_node); in sunxi_mc_smp_put_nodes() 713 memset(nodes, 0, sizeof(*nodes)); in sunxi_mc_smp_put_nodes() 716 static int __init sun9i_a80_get_smp_nodes(struct sunxi_mc_smp_nodes *nodes) in sun9i_a80_get_smp_nodes() argument 718 nodes->prcm_node = of_find_compatible_node(NULL, NULL, in sun9i_a80_get_smp_nodes() 720 if (!nodes->prcm_node) { in sun9i_a80_get_smp_nodes() [all …]
|
/Linux-v5.10/drivers/md/persistent-data/ |
D | dm-btree-spine.c | 131 s->nodes[0] = NULL; in init_ro_spine() 132 s->nodes[1] = NULL; in init_ro_spine() 140 unlock_block(s->info, s->nodes[i]); in exit_ro_spine() 149 unlock_block(s->info, s->nodes[0]); in ro_step() 150 s->nodes[0] = s->nodes[1]; in ro_step() 154 r = bn_read_lock(s->info, new_child, s->nodes + s->count); in ro_step() 165 unlock_block(s->info, s->nodes[s->count]); in ro_pop() 173 block = s->nodes[s->count - 1]; in ro_node() 191 unlock_block(s->info, s->nodes[i]); in exit_shadow_spine() 203 unlock_block(s->info, s->nodes[0]); in shadow_step() [all …]
|
/Linux-v5.10/mm/ |
D | mempolicy.c | 176 int (*create)(struct mempolicy *pol, const nodemask_t *nodes); 177 void (*rebind)(struct mempolicy *pol, const nodemask_t *nodes); 193 static int mpol_new_interleave(struct mempolicy *pol, const nodemask_t *nodes) in mpol_new_interleave() argument 195 if (nodes_empty(*nodes)) in mpol_new_interleave() 197 pol->v.nodes = *nodes; in mpol_new_interleave() 201 static int mpol_new_preferred(struct mempolicy *pol, const nodemask_t *nodes) in mpol_new_preferred() argument 203 if (!nodes) in mpol_new_preferred() 205 else if (nodes_empty(*nodes)) in mpol_new_preferred() 208 pol->v.preferred_node = first_node(*nodes); in mpol_new_preferred() 212 static int mpol_new_bind(struct mempolicy *pol, const nodemask_t *nodes) in mpol_new_bind() argument [all …]
|
D | mmzone.c | 46 static inline int zref_in_nodemask(struct zoneref *zref, nodemask_t *nodes) in zref_in_nodemask() argument 49 return node_isset(zonelist_node_idx(zref), *nodes); in zref_in_nodemask() 58 nodemask_t *nodes) in __next_zones_zonelist() argument 64 if (unlikely(nodes == NULL)) in __next_zones_zonelist() 69 (z->zone && !zref_in_nodemask(z, nodes))) in __next_zones_zonelist()
|
/Linux-v5.10/fs/btrfs/ |
D | inode-item.c | 93 return btrfs_find_name_in_ext_backref(path->nodes[0], path->slots[0], in btrfs_lookup_inode_extref() 135 extref = btrfs_find_name_in_ext_backref(path->nodes[0], path->slots[0], in btrfs_del_inode_extref() 143 leaf = path->nodes[0]; in btrfs_del_inode_extref() 207 ref = btrfs_find_name_in_backref(path->nodes[0], path->slots[0], name, in btrfs_del_inode_ref() 214 leaf = path->nodes[0]; in btrfs_del_inode_ref() 277 if (btrfs_find_name_in_ext_backref(path->nodes[0], in btrfs_insert_inode_extref() 289 leaf = path->nodes[0]; in btrfs_insert_inode_extref() 295 btrfs_set_inode_extref_name_len(path->nodes[0], extref, name_len); in btrfs_insert_inode_extref() 296 btrfs_set_inode_extref_index(path->nodes[0], extref, index); in btrfs_insert_inode_extref() 297 btrfs_set_inode_extref_parent(path->nodes[0], extref, ref_objectid); in btrfs_insert_inode_extref() [all …]
|
D | ctree.c | 102 if (!p->nodes[i]) in btrfs_release_path() 105 btrfs_tree_unlock_rw(p->nodes[i], p->locks[i]); in btrfs_release_path() 108 free_extent_buffer(p->nodes[i]); in btrfs_release_path() 109 p->nodes[i] = NULL; in btrfs_release_path() 1836 mid = path->nodes[level]; in balance_level() 1845 parent = path->nodes[level + 1]; in balance_level() 1885 path->nodes[level] = NULL; in balance_level() 2017 path->nodes[level] = left; in balance_level() 2031 btrfs_node_blockptr(path->nodes[level], path->slots[level])) in balance_level() 2039 if (path->nodes[level] != left) in balance_level() [all …]
|
D | ulist.c | 49 INIT_LIST_HEAD(&ulist->nodes); in ulist_init() 66 list_for_each_entry_safe(node, next, &ulist->nodes, list) { in ulist_release() 70 INIT_LIST_HEAD(&ulist->nodes); in ulist_release() 212 list_add_tail(&node->list, &ulist->nodes); in ulist_add_merge() 265 if (list_empty(&ulist->nodes)) in ulist_next() 267 if (uiter->cur_list && uiter->cur_list->next == &ulist->nodes) in ulist_next() 272 uiter->cur_list = ulist->nodes.next; in ulist_next()
|
/Linux-v5.10/lib/ |
D | interval_tree_test.c | 24 static struct interval_tree_node *nodes = NULL; variable 49 nodes[i].start = a; in init() 50 nodes[i].last = b; in init() 68 nodes = kmalloc_array(nnodes, sizeof(struct interval_tree_node), in interval_tree_test_init() 70 if (!nodes) in interval_tree_test_init() 75 kfree(nodes); in interval_tree_test_init() 88 interval_tree_insert(nodes + j, &root); in interval_tree_test_init() 90 interval_tree_remove(nodes + j, &root); in interval_tree_test_init() 102 interval_tree_insert(nodes + j, &root); in interval_tree_test_init() 124 kfree(nodes); in interval_tree_test_init()
|
D | rbtree_test.c | 28 static struct test_node *nodes = NULL; variable 153 nodes[i].key = prandom_u32_state(&rnd); in init() 154 nodes[i].val = prandom_u32_state(&rnd); in init() 248 nodes = kmalloc_array(nnodes, sizeof(*nodes), GFP_KERNEL); in rbtree_test_init() 249 if (!nodes) in rbtree_test_init() 261 insert(nodes + j, &root); in rbtree_test_init() 263 erase(nodes + j, &root); in rbtree_test_init() 277 insert_cached(nodes + j, &root); in rbtree_test_init() 279 erase_cached(nodes + j, &root); in rbtree_test_init() 290 insert(nodes + i, &root); in rbtree_test_init() [all …]
|
/Linux-v5.10/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_vram_mgr.c | 261 struct drm_mm_node *nodes = mem->mm_node; in amdgpu_vram_mgr_bo_visible_size() local 271 for (usage = 0; nodes && pages; pages -= nodes->size, nodes++) in amdgpu_vram_mgr_bo_visible_size() 272 usage += amdgpu_vram_mgr_vis_size(adev, nodes); in amdgpu_vram_mgr_bo_visible_size() 317 struct drm_mm_node *nodes; in amdgpu_vram_mgr_new() local 353 nodes = kvmalloc_array((uint32_t)num_nodes, sizeof(*nodes), in amdgpu_vram_mgr_new() 355 if (!nodes) { in amdgpu_vram_mgr_new() 371 r = drm_mm_insert_node_in_range(mm, &nodes[i], pages, in amdgpu_vram_mgr_new() 378 vis_usage += amdgpu_vram_mgr_vis_size(adev, &nodes[i]); in amdgpu_vram_mgr_new() 379 amdgpu_vram_mgr_virt_start(mem, &nodes[i]); in amdgpu_vram_mgr_new() 390 r = drm_mm_insert_node_in_range(mm, &nodes[i], in amdgpu_vram_mgr_new() [all …]
|
/Linux-v5.10/security/selinux/ss/ |
D | conditional.c | 34 struct cond_expr_node *node = &expr->nodes[i]; in cond_evaluate_expr() 105 avnode = node->true_list.nodes[i]; in evaluate_cond_node() 113 avnode = node->false_list.nodes[i]; in evaluate_cond_node() 142 kfree(node->expr.nodes); in cond_node_destroy() 144 kfree(node->true_list.nodes); in cond_node_destroy() 145 kfree(node->false_list.nodes); in cond_node_destroy() 293 if (other->nodes[i] == node_ptr) { in cond_insertf() 338 list->nodes = kcalloc(len, sizeof(*list->nodes), GFP_KERNEL); in cond_read_av_list() 339 if (!list->nodes) in cond_read_av_list() 345 data.dst = &list->nodes[i]; in cond_read_av_list() [all …]
|
/Linux-v5.10/Documentation/devicetree/bindings/cpu/ |
D | cpu-topology.txt | 20 For instance in a system where CPUs support SMT, "cpu" nodes represent all 22 In systems where SMT is not supported "cpu" nodes represent all cores present 25 CPU topology bindings allow one to associate cpu nodes with hierarchical groups 27 tree nodes. 32 The cpu nodes, as per bindings defined in [4], represent the devices that 35 A topology description containing phandles to cpu nodes that are not compliant 44 nodes are listed. 60 The cpu-map node's child nodes can be: 62 - one or more cluster nodes or 63 - one or more socket nodes in a multi-socket system [all …]
|
/Linux-v5.10/Documentation/devicetree/bindings/usb/ |
D | usb-device.txt | 7 Four types of device-tree nodes are defined: "host-controller nodes" 8 representing USB host controllers, "device nodes" representing USB devices, 9 "interface nodes" representing USB interfaces and "combined nodes" 20 Required properties for device nodes: 30 Required properties for device nodes with interface nodes: 35 Required properties for interface nodes: 49 Required properties for combined nodes: 59 Required properties for hub nodes with device nodes: 64 Required properties for host-controller nodes with device nodes:
|
/Linux-v5.10/drivers/clk/zynqmp/ |
D | clk-zynqmp.h | 39 const struct clock_topology *nodes); 44 const struct clock_topology *nodes); 50 const struct clock_topology *nodes); 55 const struct clock_topology *nodes); 61 const struct clock_topology *nodes);
|
/Linux-v5.10/Documentation/devicetree/bindings/pinctrl/ |
D | meson,pinctrl.txt | 21 === GPIO sub-nodes === 26 Required properties for sub-nodes are: 35 === Other sub-nodes === 37 Child nodes without the "gpio-controller" represent some desired 38 configuration for a pin or a group. Those nodes can be pinmux nodes or 39 configuration nodes. 41 Required properties for pinmux nodes are: 48 Required properties for configuration nodes: 51 Configuration nodes support the following generic properties, as
|
/Linux-v5.10/arch/sparc/kernel/ |
D | cpumap.c | 53 struct cpuinfo_node nodes[]; member 197 new_tree = kzalloc(struct_size(new_tree, nodes, n), GFP_ATOMIC); in build_cpuinfo_tree() 211 node = &new_tree->nodes[n]; in build_cpuinfo_tree() 252 node = &new_tree->nodes[level_rover[level]]; in build_cpuinfo_tree() 277 node = &new_tree->nodes[n]; in build_cpuinfo_tree() 299 struct cpuinfo_node *node = &t->nodes[node_index]; in increment_rover() 302 top_level = t->nodes[root_index].level; in increment_rover() 314 node = &t->nodes[node->parent_index]; in increment_rover() 340 for (level = t->nodes[root_index].level; level < CPUINFO_LVL_MAX; in iterate_cpu() 342 new_index = t->nodes[index].rover; in iterate_cpu() [all …]
|
/Linux-v5.10/sound/hda/ |
D | hdac_sysfs.c | 16 struct kobject **nodes; member 328 if (tree->nodes) { in widget_tree_free() 329 for (p = tree->nodes; *p; p++) in widget_tree_free() 331 kfree(tree->nodes); in widget_tree_free() 375 tree->nodes = kcalloc(codec->num_nodes + 1, sizeof(*tree->nodes), in widget_tree_create() 377 if (!tree->nodes) in widget_tree_create() 382 &tree->nodes[i]); in widget_tree_create() 437 tree->nodes = kcalloc(num_nodes + 1, sizeof(*tree->nodes), GFP_KERNEL); in hda_widget_sysfs_reinit() 438 if (!tree->nodes) { in hda_widget_sysfs_reinit() 446 free_widget_node(codec->widgets->nodes[i], in hda_widget_sysfs_reinit() [all …]
|
/Linux-v5.10/Documentation/admin-guide/mm/ |
D | numa_memory_policy.rst | 19 which is an administrative mechanism for restricting the nodes from which 42 allocations across all nodes with "sufficient" memory, so as 166 an optional set of nodes. The mode determines the behavior of the 168 and the optional set of nodes can be viewed as the arguments to the 190 does not use the optional set of nodes. 192 It is an error for the set of nodes specified for this policy to 197 nodes specified by the policy. Memory will be allocated from 204 allocation fails, the kernel will search other nodes, in order 224 page granularity, across the nodes specified in the policy. 229 Interleave mode indexes the set of nodes specified by the [all …]
|
/Linux-v5.10/Documentation/driver-api/md/ |
D | md-cluster.rst | 54 node may write to those sectors. This is used when a new nodes 60 Each node has to communicate with other nodes when starting or ending 70 Normally all nodes hold a concurrent-read lock on this device. 75 Messages can be broadcast to all nodes, and the sender waits for all 76 other nodes to acknowledge the message before proceeding. Only one 87 informs other nodes that the metadata has 94 informs other nodes that a resync is initiated or 104 informs other nodes that a device is being added to 128 The DLM LVB is used to communicate within nodes of the cluster. There 145 acknowledged by all nodes in the cluster. The BAST of the resource [all …]
|
/Linux-v5.10/drivers/dma/ |
D | uniphier-xdmac.c | 83 struct uniphier_xdmac_desc_node nodes[]; member 139 src_addr = xd->nodes[xd->cur_node].src; in uniphier_xdmac_chan_start() 140 dst_addr = xd->nodes[xd->cur_node].dst; in uniphier_xdmac_chan_start() 141 its = xd->nodes[xd->cur_node].burst_size; in uniphier_xdmac_chan_start() 142 tnum = xd->nodes[xd->cur_node].nr_burst; in uniphier_xdmac_chan_start() 294 xd = kzalloc(struct_size(xd, nodes, nr), GFP_NOWAIT); in uniphier_xdmac_prep_dma_memcpy() 300 xd->nodes[i].src = src; in uniphier_xdmac_prep_dma_memcpy() 301 xd->nodes[i].dst = dst; in uniphier_xdmac_prep_dma_memcpy() 302 xd->nodes[i].burst_size = burst_size; in uniphier_xdmac_prep_dma_memcpy() 303 xd->nodes[i].nr_burst = len / burst_size; in uniphier_xdmac_prep_dma_memcpy() [all …]
|