/Linux-v4.19/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_vram_mgr.c | 111 struct drm_mm_node *nodes = mem->mm_node; in amdgpu_vram_mgr_bo_visible_size() local 121 for (usage = 0; nodes && pages; pages -= nodes->size, nodes++) in amdgpu_vram_mgr_bo_visible_size() 122 usage += amdgpu_vram_mgr_vis_size(adev, nodes); in amdgpu_vram_mgr_bo_visible_size() 145 struct drm_mm_node *nodes; in amdgpu_vram_mgr_new() local 166 nodes = kvmalloc_array(num_nodes, sizeof(*nodes), in amdgpu_vram_mgr_new() 168 if (!nodes) in amdgpu_vram_mgr_new() 187 r = drm_mm_insert_node_in_range(mm, &nodes[i], in amdgpu_vram_mgr_new() 194 usage += nodes[i].size << PAGE_SHIFT; in amdgpu_vram_mgr_new() 195 vis_usage += amdgpu_vram_mgr_vis_size(adev, &nodes[i]); in amdgpu_vram_mgr_new() 200 start = nodes[i].start + nodes[i].size; in amdgpu_vram_mgr_new() [all …]
|
/Linux-v4.19/drivers/gpu/drm/selftests/ |
D | test-drm_mm.c | 266 struct drm_mm_node nodes[2]; in igt_debug() local 275 memset(nodes, 0, sizeof(nodes)); in igt_debug() 276 nodes[0].start = 512; in igt_debug() 277 nodes[0].size = 1024; in igt_debug() 278 ret = drm_mm_reserve_node(&mm, &nodes[0]); in igt_debug() 281 nodes[0].start, nodes[0].size); in igt_debug() 285 nodes[1].size = 1024; in igt_debug() 286 nodes[1].start = 4096 - 512 - nodes[1].size; in igt_debug() 287 ret = drm_mm_reserve_node(&mm, &nodes[1]); in igt_debug() 290 nodes[1].start, nodes[1].size); in igt_debug() [all …]
|
/Linux-v4.19/Documentation/sphinx/ |
D | rstFlatTable.py | 47 from docutils import nodes 104 class rowSpan(nodes.General, nodes.Element): pass # pylint: disable=C0103,C0321 105 class colSpan(nodes.General, nodes.Element): pass # pylint: disable=C0103,C0321 127 nodes.literal_block(self.block_text, self.block_text), 132 node = nodes.Element() # anonymous container for parsing 166 table = nodes.table() 167 tgroup = nodes.tgroup(cols=len(colwidths)) 172 colspec = nodes.colspec(colwidth=colwidth) 186 thead = nodes.thead() 191 tbody = nodes.tbody() [all …]
|
D | kfigure.py | 54 from docutils import nodes 60 from sphinx.util.nodes import clean_astext 102 node = nodes.literal_block(data, data) 340 class kernel_image(nodes.image): 359 if len(result) == 2 or isinstance(result[0], nodes.system_message): 377 class kernel_figure(nodes.figure): 396 if len(result) == 2 or isinstance(result[0], nodes.system_message): 443 img_node = nodes.image(node.rawsource, **node.attributes) 452 class kernel_render(nodes.General, nodes.Inline, nodes.Element): 499 literal_node = nodes.literal_block(code, code) [all …]
|
/Linux-v4.19/tools/perf/tests/ |
D | mem2node.c | 43 struct memory_node nodes[3]; in test__mem2node() local 45 .memory_nodes = (struct memory_node *) &nodes[0], in test__mem2node() 46 .nr_memory_nodes = ARRAY_SIZE(nodes), in test__mem2node() 51 for (i = 0; i < ARRAY_SIZE(nodes); i++) { in test__mem2node() 52 nodes[i].node = test_nodes[i].node; in test__mem2node() 53 nodes[i].size = 10; in test__mem2node() 56 (nodes[i].set = get_bitmap(test_nodes[i].map, 10))); in test__mem2node() 68 for (i = 0; i < ARRAY_SIZE(nodes); i++) in test__mem2node() 69 free(nodes[i].set); in test__mem2node()
|
/Linux-v4.19/mm/ |
D | mempolicy.c | 150 int (*create)(struct mempolicy *pol, const nodemask_t *nodes); 151 void (*rebind)(struct mempolicy *pol, const nodemask_t *nodes); 167 static int mpol_new_interleave(struct mempolicy *pol, const nodemask_t *nodes) in mpol_new_interleave() argument 169 if (nodes_empty(*nodes)) in mpol_new_interleave() 171 pol->v.nodes = *nodes; in mpol_new_interleave() 175 static int mpol_new_preferred(struct mempolicy *pol, const nodemask_t *nodes) in mpol_new_preferred() argument 177 if (!nodes) in mpol_new_preferred() 179 else if (nodes_empty(*nodes)) in mpol_new_preferred() 182 pol->v.preferred_node = first_node(*nodes); in mpol_new_preferred() 186 static int mpol_new_bind(struct mempolicy *pol, const nodemask_t *nodes) in mpol_new_bind() argument [all …]
|
D | mmzone.c | 46 static inline int zref_in_nodemask(struct zoneref *zref, nodemask_t *nodes) in zref_in_nodemask() argument 49 return node_isset(zonelist_node_idx(zref), *nodes); in zref_in_nodemask() 58 nodemask_t *nodes) in __next_zones_zonelist() argument 64 if (unlikely(nodes == NULL)) in __next_zones_zonelist() 69 (z->zone && !zref_in_nodemask(z, nodes))) in __next_zones_zonelist()
|
/Linux-v4.19/arch/arm/mach-sunxi/ |
D | mc_smp.c | 696 int (*get_smp_nodes)(struct sunxi_mc_smp_nodes *nodes); 700 static void __init sunxi_mc_smp_put_nodes(struct sunxi_mc_smp_nodes *nodes) in sunxi_mc_smp_put_nodes() argument 702 of_node_put(nodes->prcm_node); in sunxi_mc_smp_put_nodes() 703 of_node_put(nodes->cpucfg_node); in sunxi_mc_smp_put_nodes() 704 of_node_put(nodes->sram_node); in sunxi_mc_smp_put_nodes() 705 of_node_put(nodes->r_cpucfg_node); in sunxi_mc_smp_put_nodes() 706 memset(nodes, 0, sizeof(*nodes)); in sunxi_mc_smp_put_nodes() 709 static int __init sun9i_a80_get_smp_nodes(struct sunxi_mc_smp_nodes *nodes) in sun9i_a80_get_smp_nodes() argument 711 nodes->prcm_node = of_find_compatible_node(NULL, NULL, in sun9i_a80_get_smp_nodes() 713 if (!nodes->prcm_node) { in sun9i_a80_get_smp_nodes() [all …]
|
/Linux-v4.19/drivers/md/persistent-data/ |
D | dm-btree-spine.c | 131 s->nodes[0] = NULL; in init_ro_spine() 132 s->nodes[1] = NULL; in init_ro_spine() 140 unlock_block(s->info, s->nodes[i]); in exit_ro_spine() 151 unlock_block(s->info, s->nodes[0]); in ro_step() 152 s->nodes[0] = s->nodes[1]; in ro_step() 156 r = bn_read_lock(s->info, new_child, s->nodes + s->count); in ro_step() 167 unlock_block(s->info, s->nodes[s->count]); in ro_pop() 175 block = s->nodes[s->count - 1]; in ro_node() 193 unlock_block(s->info, s->nodes[i]); in exit_shadow_spine() 205 unlock_block(s->info, s->nodes[0]); in shadow_step() [all …]
|
/Linux-v4.19/fs/btrfs/ |
D | inode-item.c | 101 if (!btrfs_find_name_in_ext_backref(path->nodes[0], path->slots[0], in btrfs_lookup_inode_extref() 145 if (!btrfs_find_name_in_ext_backref(path->nodes[0], path->slots[0], in btrfs_del_inode_extref() 153 leaf = path->nodes[0]; in btrfs_del_inode_extref() 216 if (!btrfs_find_name_in_backref(path->nodes[0], path->slots[0], in btrfs_del_inode_ref() 222 leaf = path->nodes[0]; in btrfs_del_inode_ref() 285 if (btrfs_find_name_in_ext_backref(path->nodes[0], in btrfs_insert_inode_extref() 297 leaf = path->nodes[0]; in btrfs_insert_inode_extref() 303 btrfs_set_inode_extref_name_len(path->nodes[0], extref, name_len); in btrfs_insert_inode_extref() 304 btrfs_set_inode_extref_index(path->nodes[0], extref, index); in btrfs_insert_inode_extref() 305 btrfs_set_inode_extref_parent(path->nodes[0], extref, ref_objectid); in btrfs_insert_inode_extref() [all …]
|
D | ctree.c | 45 if (!p->nodes[i] || !p->locks[i]) in btrfs_set_path_blocking() 47 btrfs_set_lock_blocking_rw(p->nodes[i], p->locks[i]); in btrfs_set_path_blocking() 78 if (p->nodes[i] && p->locks[i]) { in btrfs_clear_path_blocking() 79 btrfs_clear_lock_blocking_rw(p->nodes[i], p->locks[i]); in btrfs_clear_path_blocking() 112 if (!p->nodes[i]) in btrfs_release_path() 115 btrfs_tree_unlock_rw(p->nodes[i], p->locks[i]); in btrfs_release_path() 118 free_extent_buffer(p->nodes[i]); in btrfs_release_path() 119 p->nodes[i] = NULL; in btrfs_release_path() 1821 mid = path->nodes[level]; in balance_level() 1830 parent = path->nodes[level + 1]; in balance_level() [all …]
|
D | tree-log.c | 355 u32 dst_size = btrfs_item_size_nr(path->nodes[0], in overwrite_item() 375 dst_ptr = btrfs_item_ptr_offset(path->nodes[0], path->slots[0]); in overwrite_item() 376 read_extent_buffer(path->nodes[0], dst_copy, dst_ptr, in overwrite_item() 402 item = btrfs_item_ptr(path->nodes[0], path->slots[0], in overwrite_item() 404 nbytes = btrfs_inode_nbytes(path->nodes[0], item); in overwrite_item() 449 found_size = btrfs_item_size_nr(path->nodes[0], in overwrite_item() 459 dst_ptr = btrfs_item_ptr_offset(path->nodes[0], in overwrite_item() 479 struct extent_buffer *dst_eb = path->nodes[0]; in overwrite_item() 503 S_ISDIR(btrfs_inode_mode(path->nodes[0], dst_item))) { in overwrite_item() 505 saved_i_size = btrfs_inode_size(path->nodes[0], in overwrite_item() [all …]
|
D | ulist.c | 49 INIT_LIST_HEAD(&ulist->nodes); in ulist_init() 66 list_for_each_entry_safe(node, next, &ulist->nodes, list) { in ulist_release() 70 INIT_LIST_HEAD(&ulist->nodes); in ulist_release() 212 list_add_tail(&node->list, &ulist->nodes); in ulist_add_merge() 265 if (list_empty(&ulist->nodes)) in ulist_next() 267 if (uiter->cur_list && uiter->cur_list->next == &ulist->nodes) in ulist_next() 272 uiter->cur_list = ulist->nodes.next; in ulist_next()
|
/Linux-v4.19/lib/ |
D | interval_tree_test.c | 23 static struct interval_tree_node *nodes = NULL; variable 48 nodes[i].start = a; in init() 49 nodes[i].last = b; in init() 67 nodes = kmalloc_array(nnodes, sizeof(struct interval_tree_node), in interval_tree_test_init() 69 if (!nodes) in interval_tree_test_init() 74 kfree(nodes); in interval_tree_test_init() 87 interval_tree_insert(nodes + j, &root); in interval_tree_test_init() 89 interval_tree_remove(nodes + j, &root); in interval_tree_test_init() 101 interval_tree_insert(nodes + j, &root); in interval_tree_test_init() 123 kfree(nodes); in interval_tree_test_init()
|
D | rbtree_test.c | 27 static struct test_node *nodes = NULL; variable 168 nodes[i].key = prandom_u32_state(&rnd); in init() 169 nodes[i].val = prandom_u32_state(&rnd); in init() 250 nodes = kmalloc_array(nnodes, sizeof(*nodes), GFP_KERNEL); in rbtree_test_init() 251 if (!nodes) in rbtree_test_init() 263 insert(nodes + j, &root); in rbtree_test_init() 265 erase(nodes + j, &root); in rbtree_test_init() 279 insert_cached(nodes + j, &root); in rbtree_test_init() 281 erase_cached(nodes + j, &root); in rbtree_test_init() 292 insert(nodes + i, &root); in rbtree_test_init() [all …]
|
/Linux-v4.19/Documentation/devicetree/bindings/usb/ |
D | usb-device.txt | 7 Four types of device-tree nodes are defined: "host-controller nodes" 8 representing USB host controllers, "device nodes" representing USB devices, 9 "interface nodes" representing USB interfaces and "combined nodes" 20 Required properties for device nodes: 30 Required properties for device nodes with interface nodes: 35 Required properties for interface nodes: 49 Required properties for combined nodes: 59 Required properties for hub nodes with device nodes: 64 Required properties for host-controller nodes with device nodes:
|
/Linux-v4.19/Documentation/devicetree/bindings/arm/ |
D | topology.txt | 16 The cpu nodes (bindings defined in [1]) represent the devices that 22 For instance in a system where CPUs support SMT, "cpu" nodes represent all 24 In systems where SMT is not supported "cpu" nodes represent all cores present 27 ARM topology bindings allow one to associate cpu nodes with hierarchical groups 29 tree nodes. 39 A topology description containing phandles to cpu nodes that are not compliant 48 nodes are listed. 64 The cpu-map node's child nodes can be: 66 - one or more cluster nodes 70 The cpu-map node can only contain three types of child nodes: [all …]
|
/Linux-v4.19/Documentation/devicetree/bindings/pinctrl/ |
D | meson,pinctrl.txt | 18 === GPIO sub-nodes === 23 Required properties for sub-nodes are: 32 === Other sub-nodes === 34 Child nodes without the "gpio-controller" represent some desired 35 configuration for a pin or a group. Those nodes can be pinmux nodes or 36 configuration nodes. 38 Required properties for pinmux nodes are: 45 Required properties for configuration nodes: 48 Configuration nodes support the generic properties "bias-disable",
|
/Linux-v4.19/sound/hda/ |
D | hdac_sysfs.c | 16 struct kobject **nodes; member 328 if (tree->nodes) { in widget_tree_free() 329 for (p = tree->nodes; *p; p++) in widget_tree_free() 331 kfree(tree->nodes); in widget_tree_free() 375 tree->nodes = kcalloc(codec->num_nodes + 1, sizeof(*tree->nodes), in widget_tree_create() 377 if (!tree->nodes) in widget_tree_create() 382 &tree->nodes[i]); in widget_tree_create() 434 tree->nodes = kcalloc(num_nodes + 1, sizeof(*tree->nodes), GFP_KERNEL); in hda_widget_sysfs_reinit() 435 if (!tree->nodes) { in hda_widget_sysfs_reinit() 443 free_widget_node(codec->widgets->nodes[i], in hda_widget_sysfs_reinit() [all …]
|
/Linux-v4.19/arch/sparc/kernel/ |
D | cpumap.c | 53 struct cpuinfo_node nodes[0]; member 212 node = &new_tree->nodes[n]; in build_cpuinfo_tree() 253 node = &new_tree->nodes[level_rover[level]]; in build_cpuinfo_tree() 278 node = &new_tree->nodes[n]; in build_cpuinfo_tree() 300 struct cpuinfo_node *node = &t->nodes[node_index]; in increment_rover() 303 top_level = t->nodes[root_index].level; in increment_rover() 315 node = &t->nodes[node->parent_index]; in increment_rover() 341 for (level = t->nodes[root_index].level; level < CPUINFO_LVL_MAX; in iterate_cpu() 343 new_index = t->nodes[index].rover; in iterate_cpu() 369 for (i = 0; i < cpuinfo_tree->nodes[0].num_cpus; i++) in _cpu_map_rebuild() [all …]
|
/Linux-v4.19/Documentation/admin-guide/mm/ |
D | numa_memory_policy.rst | 19 which is an administrative mechanism for restricting the nodes from which 42 allocations across all nodes with "sufficient" memory, so as 166 an optional set of nodes. The mode determines the behavior of the 168 and the optional set of nodes can be viewed as the arguments to the 190 does not use the optional set of nodes. 192 It is an error for the set of nodes specified for this policy to 197 nodes specified by the policy. Memory will be allocated from 204 allocation fails, the kernel will search other nodes, in order 224 page granularity, across the nodes specified in the policy. 229 Interleave mode indexes the set of nodes specified by the [all …]
|
/Linux-v4.19/kernel/irq/ |
D | affinity.c | 85 int n, nodes = 0; in get_nodes_in_cpumask() local 91 nodes++; in get_nodes_in_cpumask() 94 return nodes; in get_nodes_in_cpumask() 104 int n, nodes, cpus_per_vec, extra_vecs, done = 0; in irq_build_affinity_masks() local 112 nodes = get_nodes_in_cpumask(node_to_cpumask, cpu_mask, &nodemsk); in irq_build_affinity_masks() 118 if (numvecs <= nodes) { in irq_build_affinity_masks() 133 vecs_per_node = (numvecs - (curvec - affd->pre_vectors)) / nodes; in irq_build_affinity_masks() 162 --nodes; in irq_build_affinity_masks()
|
/Linux-v4.19/Documentation/md/ |
D | md-cluster.txt | 47 node may write to those sectors. This is used when a new nodes 52 Each node has to communicate with other nodes when starting or ending 61 Normally all nodes hold a concurrent-read lock on this device. 65 Messages can be broadcast to all nodes, and the sender waits for all 66 other nodes to acknowledge the message before proceeding. Only one 73 3.1.1 METADATA_UPDATED: informs other nodes that the metadata has 78 3.1.2 RESYNCING: informs other nodes that a resync is initiated or 85 3.1.3 NEWDISK: informs other nodes that a device is being added to 101 The DLM LVB is used to communicate within nodes of the cluster. There 112 acknowledged by all nodes in the cluster. The BAST of the resource [all …]
|
/Linux-v4.19/arch/mips/sgi-ip27/ |
D | Kconfig | 10 The nodes of Origin, Onyx, Fuel and Tezro systems can be configured 11 in either N-Modes which allows for more nodes or M-Mode which allows 18 The nodes of Origin, Onyx, Fuel and Tezro systems can be configured 19 in either N-Modes which allows for more nodes or M-Mode which allows 39 nodes in a NUMA cluster. This trades memory for speed. 46 across multiple nodes in a NUMA cluster. This trades memory for
|
/Linux-v4.19/Documentation/devicetree/bindings/i2c/ |
D | i2c-mux.txt | 4 numbered uniquely in a device dependent manner. The nodes for an i2c bus 15 For i2c multiplexers/switches that have child nodes that are a mixture 16 of both i2c child busses and other child nodes, the 'i2c-mux' subnode 25 Required properties for i2c child bus nodes: 30 Optional properties for i2c child bus nodes: 32 - Child nodes conforming to i2c bus binding
|