Home
last modified time | relevance | path

Searched refs:nodes (Results 1 – 25 of 1078) sorted by relevance

12345678910>>...44

/Linux-v4.19/drivers/gpu/drm/amd/amdgpu/
Damdgpu_vram_mgr.c111 struct drm_mm_node *nodes = mem->mm_node; in amdgpu_vram_mgr_bo_visible_size() local
121 for (usage = 0; nodes && pages; pages -= nodes->size, nodes++) in amdgpu_vram_mgr_bo_visible_size()
122 usage += amdgpu_vram_mgr_vis_size(adev, nodes); in amdgpu_vram_mgr_bo_visible_size()
145 struct drm_mm_node *nodes; in amdgpu_vram_mgr_new() local
166 nodes = kvmalloc_array(num_nodes, sizeof(*nodes), in amdgpu_vram_mgr_new()
168 if (!nodes) in amdgpu_vram_mgr_new()
187 r = drm_mm_insert_node_in_range(mm, &nodes[i], in amdgpu_vram_mgr_new()
194 usage += nodes[i].size << PAGE_SHIFT; in amdgpu_vram_mgr_new()
195 vis_usage += amdgpu_vram_mgr_vis_size(adev, &nodes[i]); in amdgpu_vram_mgr_new()
200 start = nodes[i].start + nodes[i].size; in amdgpu_vram_mgr_new()
[all …]
/Linux-v4.19/drivers/gpu/drm/selftests/
Dtest-drm_mm.c266 struct drm_mm_node nodes[2]; in igt_debug() local
275 memset(nodes, 0, sizeof(nodes)); in igt_debug()
276 nodes[0].start = 512; in igt_debug()
277 nodes[0].size = 1024; in igt_debug()
278 ret = drm_mm_reserve_node(&mm, &nodes[0]); in igt_debug()
281 nodes[0].start, nodes[0].size); in igt_debug()
285 nodes[1].size = 1024; in igt_debug()
286 nodes[1].start = 4096 - 512 - nodes[1].size; in igt_debug()
287 ret = drm_mm_reserve_node(&mm, &nodes[1]); in igt_debug()
290 nodes[1].start, nodes[1].size); in igt_debug()
[all …]
/Linux-v4.19/Documentation/sphinx/
DrstFlatTable.py47 from docutils import nodes
104 class rowSpan(nodes.General, nodes.Element): pass # pylint: disable=C0103,C0321
105 class colSpan(nodes.General, nodes.Element): pass # pylint: disable=C0103,C0321
127 nodes.literal_block(self.block_text, self.block_text),
132 node = nodes.Element() # anonymous container for parsing
166 table = nodes.table()
167 tgroup = nodes.tgroup(cols=len(colwidths))
172 colspec = nodes.colspec(colwidth=colwidth)
186 thead = nodes.thead()
191 tbody = nodes.tbody()
[all …]
Dkfigure.py54 from docutils import nodes
60 from sphinx.util.nodes import clean_astext
102 node = nodes.literal_block(data, data)
340 class kernel_image(nodes.image):
359 if len(result) == 2 or isinstance(result[0], nodes.system_message):
377 class kernel_figure(nodes.figure):
396 if len(result) == 2 or isinstance(result[0], nodes.system_message):
443 img_node = nodes.image(node.rawsource, **node.attributes)
452 class kernel_render(nodes.General, nodes.Inline, nodes.Element):
499 literal_node = nodes.literal_block(code, code)
[all …]
/Linux-v4.19/tools/perf/tests/
Dmem2node.c43 struct memory_node nodes[3]; in test__mem2node() local
45 .memory_nodes = (struct memory_node *) &nodes[0], in test__mem2node()
46 .nr_memory_nodes = ARRAY_SIZE(nodes), in test__mem2node()
51 for (i = 0; i < ARRAY_SIZE(nodes); i++) { in test__mem2node()
52 nodes[i].node = test_nodes[i].node; in test__mem2node()
53 nodes[i].size = 10; in test__mem2node()
56 (nodes[i].set = get_bitmap(test_nodes[i].map, 10))); in test__mem2node()
68 for (i = 0; i < ARRAY_SIZE(nodes); i++) in test__mem2node()
69 free(nodes[i].set); in test__mem2node()
/Linux-v4.19/mm/
Dmempolicy.c150 int (*create)(struct mempolicy *pol, const nodemask_t *nodes);
151 void (*rebind)(struct mempolicy *pol, const nodemask_t *nodes);
167 static int mpol_new_interleave(struct mempolicy *pol, const nodemask_t *nodes) in mpol_new_interleave() argument
169 if (nodes_empty(*nodes)) in mpol_new_interleave()
171 pol->v.nodes = *nodes; in mpol_new_interleave()
175 static int mpol_new_preferred(struct mempolicy *pol, const nodemask_t *nodes) in mpol_new_preferred() argument
177 if (!nodes) in mpol_new_preferred()
179 else if (nodes_empty(*nodes)) in mpol_new_preferred()
182 pol->v.preferred_node = first_node(*nodes); in mpol_new_preferred()
186 static int mpol_new_bind(struct mempolicy *pol, const nodemask_t *nodes) in mpol_new_bind() argument
[all …]
Dmmzone.c46 static inline int zref_in_nodemask(struct zoneref *zref, nodemask_t *nodes) in zref_in_nodemask() argument
49 return node_isset(zonelist_node_idx(zref), *nodes); in zref_in_nodemask()
58 nodemask_t *nodes) in __next_zones_zonelist() argument
64 if (unlikely(nodes == NULL)) in __next_zones_zonelist()
69 (z->zone && !zref_in_nodemask(z, nodes))) in __next_zones_zonelist()
/Linux-v4.19/arch/arm/mach-sunxi/
Dmc_smp.c696 int (*get_smp_nodes)(struct sunxi_mc_smp_nodes *nodes);
700 static void __init sunxi_mc_smp_put_nodes(struct sunxi_mc_smp_nodes *nodes) in sunxi_mc_smp_put_nodes() argument
702 of_node_put(nodes->prcm_node); in sunxi_mc_smp_put_nodes()
703 of_node_put(nodes->cpucfg_node); in sunxi_mc_smp_put_nodes()
704 of_node_put(nodes->sram_node); in sunxi_mc_smp_put_nodes()
705 of_node_put(nodes->r_cpucfg_node); in sunxi_mc_smp_put_nodes()
706 memset(nodes, 0, sizeof(*nodes)); in sunxi_mc_smp_put_nodes()
709 static int __init sun9i_a80_get_smp_nodes(struct sunxi_mc_smp_nodes *nodes) in sun9i_a80_get_smp_nodes() argument
711 nodes->prcm_node = of_find_compatible_node(NULL, NULL, in sun9i_a80_get_smp_nodes()
713 if (!nodes->prcm_node) { in sun9i_a80_get_smp_nodes()
[all …]
/Linux-v4.19/drivers/md/persistent-data/
Ddm-btree-spine.c131 s->nodes[0] = NULL; in init_ro_spine()
132 s->nodes[1] = NULL; in init_ro_spine()
140 unlock_block(s->info, s->nodes[i]); in exit_ro_spine()
151 unlock_block(s->info, s->nodes[0]); in ro_step()
152 s->nodes[0] = s->nodes[1]; in ro_step()
156 r = bn_read_lock(s->info, new_child, s->nodes + s->count); in ro_step()
167 unlock_block(s->info, s->nodes[s->count]); in ro_pop()
175 block = s->nodes[s->count - 1]; in ro_node()
193 unlock_block(s->info, s->nodes[i]); in exit_shadow_spine()
205 unlock_block(s->info, s->nodes[0]); in shadow_step()
[all …]
/Linux-v4.19/fs/btrfs/
Dinode-item.c101 if (!btrfs_find_name_in_ext_backref(path->nodes[0], path->slots[0], in btrfs_lookup_inode_extref()
145 if (!btrfs_find_name_in_ext_backref(path->nodes[0], path->slots[0], in btrfs_del_inode_extref()
153 leaf = path->nodes[0]; in btrfs_del_inode_extref()
216 if (!btrfs_find_name_in_backref(path->nodes[0], path->slots[0], in btrfs_del_inode_ref()
222 leaf = path->nodes[0]; in btrfs_del_inode_ref()
285 if (btrfs_find_name_in_ext_backref(path->nodes[0], in btrfs_insert_inode_extref()
297 leaf = path->nodes[0]; in btrfs_insert_inode_extref()
303 btrfs_set_inode_extref_name_len(path->nodes[0], extref, name_len); in btrfs_insert_inode_extref()
304 btrfs_set_inode_extref_index(path->nodes[0], extref, index); in btrfs_insert_inode_extref()
305 btrfs_set_inode_extref_parent(path->nodes[0], extref, ref_objectid); in btrfs_insert_inode_extref()
[all …]
Dctree.c45 if (!p->nodes[i] || !p->locks[i]) in btrfs_set_path_blocking()
47 btrfs_set_lock_blocking_rw(p->nodes[i], p->locks[i]); in btrfs_set_path_blocking()
78 if (p->nodes[i] && p->locks[i]) { in btrfs_clear_path_blocking()
79 btrfs_clear_lock_blocking_rw(p->nodes[i], p->locks[i]); in btrfs_clear_path_blocking()
112 if (!p->nodes[i]) in btrfs_release_path()
115 btrfs_tree_unlock_rw(p->nodes[i], p->locks[i]); in btrfs_release_path()
118 free_extent_buffer(p->nodes[i]); in btrfs_release_path()
119 p->nodes[i] = NULL; in btrfs_release_path()
1821 mid = path->nodes[level]; in balance_level()
1830 parent = path->nodes[level + 1]; in balance_level()
[all …]
Dtree-log.c355 u32 dst_size = btrfs_item_size_nr(path->nodes[0], in overwrite_item()
375 dst_ptr = btrfs_item_ptr_offset(path->nodes[0], path->slots[0]); in overwrite_item()
376 read_extent_buffer(path->nodes[0], dst_copy, dst_ptr, in overwrite_item()
402 item = btrfs_item_ptr(path->nodes[0], path->slots[0], in overwrite_item()
404 nbytes = btrfs_inode_nbytes(path->nodes[0], item); in overwrite_item()
449 found_size = btrfs_item_size_nr(path->nodes[0], in overwrite_item()
459 dst_ptr = btrfs_item_ptr_offset(path->nodes[0], in overwrite_item()
479 struct extent_buffer *dst_eb = path->nodes[0]; in overwrite_item()
503 S_ISDIR(btrfs_inode_mode(path->nodes[0], dst_item))) { in overwrite_item()
505 saved_i_size = btrfs_inode_size(path->nodes[0], in overwrite_item()
[all …]
Dulist.c49 INIT_LIST_HEAD(&ulist->nodes); in ulist_init()
66 list_for_each_entry_safe(node, next, &ulist->nodes, list) { in ulist_release()
70 INIT_LIST_HEAD(&ulist->nodes); in ulist_release()
212 list_add_tail(&node->list, &ulist->nodes); in ulist_add_merge()
265 if (list_empty(&ulist->nodes)) in ulist_next()
267 if (uiter->cur_list && uiter->cur_list->next == &ulist->nodes) in ulist_next()
272 uiter->cur_list = ulist->nodes.next; in ulist_next()
/Linux-v4.19/lib/
Dinterval_tree_test.c23 static struct interval_tree_node *nodes = NULL; variable
48 nodes[i].start = a; in init()
49 nodes[i].last = b; in init()
67 nodes = kmalloc_array(nnodes, sizeof(struct interval_tree_node), in interval_tree_test_init()
69 if (!nodes) in interval_tree_test_init()
74 kfree(nodes); in interval_tree_test_init()
87 interval_tree_insert(nodes + j, &root); in interval_tree_test_init()
89 interval_tree_remove(nodes + j, &root); in interval_tree_test_init()
101 interval_tree_insert(nodes + j, &root); in interval_tree_test_init()
123 kfree(nodes); in interval_tree_test_init()
Drbtree_test.c27 static struct test_node *nodes = NULL; variable
168 nodes[i].key = prandom_u32_state(&rnd); in init()
169 nodes[i].val = prandom_u32_state(&rnd); in init()
250 nodes = kmalloc_array(nnodes, sizeof(*nodes), GFP_KERNEL); in rbtree_test_init()
251 if (!nodes) in rbtree_test_init()
263 insert(nodes + j, &root); in rbtree_test_init()
265 erase(nodes + j, &root); in rbtree_test_init()
279 insert_cached(nodes + j, &root); in rbtree_test_init()
281 erase_cached(nodes + j, &root); in rbtree_test_init()
292 insert(nodes + i, &root); in rbtree_test_init()
[all …]
/Linux-v4.19/Documentation/devicetree/bindings/usb/
Dusb-device.txt7 Four types of device-tree nodes are defined: "host-controller nodes"
8 representing USB host controllers, "device nodes" representing USB devices,
9 "interface nodes" representing USB interfaces and "combined nodes"
20 Required properties for device nodes:
30 Required properties for device nodes with interface nodes:
35 Required properties for interface nodes:
49 Required properties for combined nodes:
59 Required properties for hub nodes with device nodes:
64 Required properties for host-controller nodes with device nodes:
/Linux-v4.19/Documentation/devicetree/bindings/arm/
Dtopology.txt16 The cpu nodes (bindings defined in [1]) represent the devices that
22 For instance in a system where CPUs support SMT, "cpu" nodes represent all
24 In systems where SMT is not supported "cpu" nodes represent all cores present
27 ARM topology bindings allow one to associate cpu nodes with hierarchical groups
29 tree nodes.
39 A topology description containing phandles to cpu nodes that are not compliant
48 nodes are listed.
64 The cpu-map node's child nodes can be:
66 - one or more cluster nodes
70 The cpu-map node can only contain three types of child nodes:
[all …]
/Linux-v4.19/Documentation/devicetree/bindings/pinctrl/
Dmeson,pinctrl.txt18 === GPIO sub-nodes ===
23 Required properties for sub-nodes are:
32 === Other sub-nodes ===
34 Child nodes without the "gpio-controller" represent some desired
35 configuration for a pin or a group. Those nodes can be pinmux nodes or
36 configuration nodes.
38 Required properties for pinmux nodes are:
45 Required properties for configuration nodes:
48 Configuration nodes support the generic properties "bias-disable",
/Linux-v4.19/sound/hda/
Dhdac_sysfs.c16 struct kobject **nodes; member
328 if (tree->nodes) { in widget_tree_free()
329 for (p = tree->nodes; *p; p++) in widget_tree_free()
331 kfree(tree->nodes); in widget_tree_free()
375 tree->nodes = kcalloc(codec->num_nodes + 1, sizeof(*tree->nodes), in widget_tree_create()
377 if (!tree->nodes) in widget_tree_create()
382 &tree->nodes[i]); in widget_tree_create()
434 tree->nodes = kcalloc(num_nodes + 1, sizeof(*tree->nodes), GFP_KERNEL); in hda_widget_sysfs_reinit()
435 if (!tree->nodes) { in hda_widget_sysfs_reinit()
443 free_widget_node(codec->widgets->nodes[i], in hda_widget_sysfs_reinit()
[all …]
/Linux-v4.19/arch/sparc/kernel/
Dcpumap.c53 struct cpuinfo_node nodes[0]; member
212 node = &new_tree->nodes[n]; in build_cpuinfo_tree()
253 node = &new_tree->nodes[level_rover[level]]; in build_cpuinfo_tree()
278 node = &new_tree->nodes[n]; in build_cpuinfo_tree()
300 struct cpuinfo_node *node = &t->nodes[node_index]; in increment_rover()
303 top_level = t->nodes[root_index].level; in increment_rover()
315 node = &t->nodes[node->parent_index]; in increment_rover()
341 for (level = t->nodes[root_index].level; level < CPUINFO_LVL_MAX; in iterate_cpu()
343 new_index = t->nodes[index].rover; in iterate_cpu()
369 for (i = 0; i < cpuinfo_tree->nodes[0].num_cpus; i++) in _cpu_map_rebuild()
[all …]
/Linux-v4.19/Documentation/admin-guide/mm/
Dnuma_memory_policy.rst19 which is an administrative mechanism for restricting the nodes from which
42 allocations across all nodes with "sufficient" memory, so as
166 an optional set of nodes. The mode determines the behavior of the
168 and the optional set of nodes can be viewed as the arguments to the
190 does not use the optional set of nodes.
192 It is an error for the set of nodes specified for this policy to
197 nodes specified by the policy. Memory will be allocated from
204 allocation fails, the kernel will search other nodes, in order
224 page granularity, across the nodes specified in the policy.
229 Interleave mode indexes the set of nodes specified by the
[all …]
/Linux-v4.19/kernel/irq/
Daffinity.c85 int n, nodes = 0; in get_nodes_in_cpumask() local
91 nodes++; in get_nodes_in_cpumask()
94 return nodes; in get_nodes_in_cpumask()
104 int n, nodes, cpus_per_vec, extra_vecs, done = 0; in irq_build_affinity_masks() local
112 nodes = get_nodes_in_cpumask(node_to_cpumask, cpu_mask, &nodemsk); in irq_build_affinity_masks()
118 if (numvecs <= nodes) { in irq_build_affinity_masks()
133 vecs_per_node = (numvecs - (curvec - affd->pre_vectors)) / nodes; in irq_build_affinity_masks()
162 --nodes; in irq_build_affinity_masks()
/Linux-v4.19/Documentation/md/
Dmd-cluster.txt47 node may write to those sectors. This is used when a new nodes
52 Each node has to communicate with other nodes when starting or ending
61 Normally all nodes hold a concurrent-read lock on this device.
65 Messages can be broadcast to all nodes, and the sender waits for all
66 other nodes to acknowledge the message before proceeding. Only one
73 3.1.1 METADATA_UPDATED: informs other nodes that the metadata has
78 3.1.2 RESYNCING: informs other nodes that a resync is initiated or
85 3.1.3 NEWDISK: informs other nodes that a device is being added to
101 The DLM LVB is used to communicate within nodes of the cluster. There
112 acknowledged by all nodes in the cluster. The BAST of the resource
[all …]
/Linux-v4.19/arch/mips/sgi-ip27/
DKconfig10 The nodes of Origin, Onyx, Fuel and Tezro systems can be configured
11 in either N-Modes which allows for more nodes or M-Mode which allows
18 The nodes of Origin, Onyx, Fuel and Tezro systems can be configured
19 in either N-Modes which allows for more nodes or M-Mode which allows
39 nodes in a NUMA cluster. This trades memory for speed.
46 across multiple nodes in a NUMA cluster. This trades memory for
/Linux-v4.19/Documentation/devicetree/bindings/i2c/
Di2c-mux.txt4 numbered uniquely in a device dependent manner. The nodes for an i2c bus
15 For i2c multiplexers/switches that have child nodes that are a mixture
16 of both i2c child busses and other child nodes, the 'i2c-mux' subnode
25 Required properties for i2c child bus nodes:
30 Optional properties for i2c child bus nodes:
32 - Child nodes conforming to i2c bus binding

12345678910>>...44