/Linux-v6.6/tools/perf/tests/ |
D | topology.c | 1 // SPDX-License-Identifier: GPL-2.0 14 #define TEMPL "/tmp/perf-test-XXXXXX" 26 return -1; in get_temp() 44 session->evlist = evlist__new_default(); in session_write_header() 45 TEST_ASSERT_VAL("can't get evlist", session->evlist); in session_write_header() 47 perf_header__set_feat(&session->header, HEADER_CPU_TOPOLOGY); in session_write_header() 48 perf_header__set_feat(&session->header, HEADER_NRCPUS); in session_write_header() 49 perf_header__set_feat(&session->header, HEADER_ARCH); in session_write_header() 51 session->header.data_size += DATA_SIZE; in session_write_header() 54 !perf_session__write_header(session, session->evlist, data.file.fd, true)); in session_write_header() [all …]
|
D | cpumap.c | 1 // SPDX-License-Identifier: GPL-2.0 6 #include "util/synthetic-events.h" 19 struct perf_record_cpu_map *map_event = &event->cpu_map; in process_event_mask() 21 struct perf_cpu_map *map; in process_event_mask() local 24 data = &map_event->data; in process_event_mask() 26 TEST_ASSERT_VAL("wrong type", data->type == PERF_CPU_MAP__MASK); in process_event_mask() 28 long_size = data->mask32_data.long_size; in process_event_mask() 32 TEST_ASSERT_VAL("wrong nr", data->mask32_data.nr == 1); in process_event_mask() 34 TEST_ASSERT_VAL("wrong cpu", perf_record_cpu_map_data__test_bit(0, data)); in process_event_mask() 35 TEST_ASSERT_VAL("wrong cpu", !perf_record_cpu_map_data__test_bit(1, data)); in process_event_mask() [all …]
|
/Linux-v6.6/arch/powerpc/platforms/cell/ |
D | cbe_regs.c | 1 // SPDX-License-Identifier: GPL-2.0-only 19 #include <asm/cell-regs.h> 22 * Current implementation uses "cpu" nodes. We build our own mapping 23 * array of cpu numbers to cpu nodes locally for now to allow interrupt 25 * we implement cpu hotplug, we'll have to install an appropriate notifier 26 * in order to release references to the cpu going away 48 static cpumask_t cbe_local_mask[MAX_CBE] = { [0 ... MAX_CBE-1] = {CPU_BITS_NONE} }; 64 if (np->data) in cbe_find_map() 65 return np->data; in cbe_find_map() 67 /* walk up path until cpu or be node was found */ in cbe_find_map() [all …]
|
/Linux-v6.6/tools/perf/util/ |
D | cpumap.c | 1 // SPDX-License-Identifier: GPL-2.0 22 * CPU number. 34 return (data->mask32_data.long_size == 4) in perf_record_cpu_map_data__test_bit() 35 ? (bit_word32 < data->mask32_data.nr) && in perf_record_cpu_map_data__test_bit() 36 (data->mask32_data.mask[bit_word32] & bit_mask32) != 0 in perf_record_cpu_map_data__test_bit() 37 : (bit_word64 < data->mask64_data.nr) && in perf_record_cpu_map_data__test_bit() 38 (data->mask64_data.mask[bit_word64] & bit_mask64) != 0; in perf_record_cpu_map_data__test_bit() 41 /* Read ith mask value from data into the given 64-bit sized bitmap */ 46 if (data->mask32_data.long_size == 4) in perf_record_cpu_map_data__read_one_mask() 47 bitmap[0] = data->mask32_data.mask[i]; in perf_record_cpu_map_data__read_one_mask() [all …]
|
D | mmap.c | 1 // SPDX-License-Identifier: GPL-2.0-only 3 * Copyright (C) 2011-2017, Red Hat Inc, Arnaldo Carvalho de Melo <acme@redhat.com> 5 * Parts came from evlist.c builtin-{top,stat,record}.c, see those files for further 34 len = bitmap_scnprintf(mask->bits, mask->nbits, buf, MASK_SIZE); in mmap_cpu_mask__scnprintf() 36 pr_debug("%p: %s mask[%zd]: %s\n", mask, tag, mask->nbits, buf); in mmap_cpu_mask__scnprintf() 39 size_t mmap__mmap_len(struct mmap *map) in mmap__mmap_len() argument 41 return perf_mmap__mmap_len(&map->core); in mmap__mmap_len() 71 static int perf_mmap__aio_enabled(struct mmap *map) in perf_mmap__aio_enabled() argument 73 return map->aio.nr_cblocks > 0; in perf_mmap__aio_enabled() 77 static int perf_mmap__aio_alloc(struct mmap *map, int idx) in perf_mmap__aio_alloc() argument [all …]
|
/Linux-v6.6/tools/lib/perf/ |
D | cpumap.c | 1 // SPDX-License-Identifier: GPL-2.0-only 13 void perf_cpu_map__set_nr(struct perf_cpu_map *map, int nr_cpus) in perf_cpu_map__set_nr() argument 15 RC_CHK_ACCESS(map)->nr = nr_cpus; in perf_cpu_map__set_nr() 24 cpus->nr = nr_cpus; in perf_cpu_map__alloc() 25 refcount_set(&cpus->refcnt, 1); in perf_cpu_map__alloc() 35 RC_CHK_ACCESS(cpus)->map[0].cpu = -1; in perf_cpu_map__dummy_new() 40 static void cpu_map__delete(struct perf_cpu_map *map) in cpu_map__delete() argument 42 if (map) { in cpu_map__delete() 43 WARN_ONCE(refcount_read(perf_cpu_map__refcnt(map)) != 0, in cpu_map__delete() 45 RC_CHK_FREE(map); in cpu_map__delete() [all …]
|
D | evlist.c | 1 // SPDX-License-Identifier: GPL-2.0 30 INIT_LIST_HEAD(&evlist->entries); in perf_evlist__init() 31 evlist->nr_entries = 0; in perf_evlist__init() 32 fdarray__init(&evlist->pollfd, 64); in perf_evlist__init() 39 if (evsel->system_wide) { in __perf_evlist__propagate_maps() 40 /* System wide: set the cpu map of the evsel to all online CPUs. */ in __perf_evlist__propagate_maps() 41 perf_cpu_map__put(evsel->cpus); in __perf_evlist__propagate_maps() 42 evsel->cpus = perf_cpu_map__new(NULL); in __perf_evlist__propagate_maps() 43 } else if (evlist->has_user_cpus && evsel->is_pmu_core) { in __perf_evlist__propagate_maps() 48 perf_cpu_map__put(evsel->cpus); in __perf_evlist__propagate_maps() [all …]
|
/Linux-v6.6/Documentation/devicetree/bindings/cpu/ |
D | cpu-topology.txt | 2 CPU topology binding description 6 1 - Introduction 12 - socket 13 - cluster 14 - core 15 - thread 18 symmetric multi-threading (SMT) is supported or not. 20 For instance in a system where CPUs support SMT, "cpu" nodes represent all 21 threads existing in the system and map to the hierarchy level "thread" above. 22 In systems where SMT is not supported "cpu" nodes represent all cores present [all …]
|
/Linux-v6.6/kernel/bpf/ |
D | cpumap.c | 1 // SPDX-License-Identifier: GPL-2.0-only 8 * DOC: cpu map 9 * The 'cpumap' is primarily used as a backend map for XDP BPF helper 13 * this map type redirects raw XDP frames to another CPU. The remote 14 * CPU will do SKB-allocation and call the normal network stack. 20 * basically allows for 10G wirespeed pre-filtering via bpf. 38 /* General idea: XDP packets getting XDP redirected to another CPU, 39 * will maximum be stored/queued for one driver ->poll() call. It is 41 * same CPU. Thus, cpu_map_flush operation can deduct via this_cpu_ptr() 45 #define CPU_MAP_BULK_SIZE 8 /* 8 == one cacheline on 64-bit archs */ [all …]
|
D | hashtab.c | 1 // SPDX-License-Identifier: GPL-2.0-only 2 /* Copyright (c) 2011-2014 PLUMgrid, http://plumgrid.com 46 * from sys_bpf(). BPF recursion is prevented by incrementing the per CPU 50 * by pinning the task to the current CPU and incrementing the recursion 51 * protection across the map operation. 72 * it is only safe to use raw spinlock for preallocated hash map on a RT kernel, 74 * after hash map was fully converted to use bpf_mem_alloc, there will be 75 * non-synchronous memory allocation for non-preallocated hash map, so it is 84 #define HASHTAB_MAP_LOCK_MASK (HASHTAB_MAP_LOCK_COUNT - 1) 87 struct bpf_map map; member [all …]
|
D | local_storage.c | 1 // SPDX-License-Identifier: GPL-2.0 2 #include <linux/bpf-cgroup.h> 16 #include "../cgroup/cgroup-internal.h" 22 struct bpf_map map; member 29 static struct bpf_cgroup_storage_map *map_to_storage(struct bpf_map *map) in map_to_storage() argument 31 return container_of(map, struct bpf_cgroup_storage_map, map); in map_to_storage() 34 static bool attach_type_isolated(const struct bpf_map *map) in attach_type_isolated() argument 36 return map->key_size == sizeof(struct bpf_cgroup_storage_key); in attach_type_isolated() 39 static int bpf_cgroup_storage_key_cmp(const struct bpf_cgroup_storage_map *map, in bpf_cgroup_storage_key_cmp() argument 42 if (attach_type_isolated(&map->map)) { in bpf_cgroup_storage_key_cmp() [all …]
|
D | arraymap.c | 1 // SPDX-License-Identifier: GPL-2.0-only 2 /* Copyright (c) 2011-2014 PLUMgrid, http://plumgrid.com 26 for (i = 0; i < array->map.max_entries; i++) { in bpf_array_free_percpu() 27 free_percpu(array->pptrs[i]); in bpf_array_free_percpu() 37 for (i = 0; i < array->map.max_entries; i++) { in bpf_array_alloc_percpu() 38 ptr = bpf_map_alloc_percpu(&array->map, array->elem_size, 8, in bpf_array_alloc_percpu() 42 return -ENOMEM; in bpf_array_alloc_percpu() 44 array->pptrs[i] = ptr; in bpf_array_alloc_percpu() 54 bool percpu = attr->map_type == BPF_MAP_TYPE_PERCPU_ARRAY; in array_map_alloc_check() 58 if (attr->max_entries == 0 || attr->key_size != 4 || in array_map_alloc_check() [all …]
|
/Linux-v6.6/tools/lib/perf/include/perf/ |
D | cpumap.h | 1 /* SPDX-License-Identifier: GPL-2.0 */ 9 /** A wrapper around a CPU to avoid confusion with the perf_cpu_map's map's indices. */ 11 int cpu; member 22 * perf_cpu_map__dummy_new - a map with a singular "any CPU"/dummy -1 value. 28 LIBPERF_API struct perf_cpu_map *perf_cpu_map__get(struct perf_cpu_map *map); 33 LIBPERF_API void perf_cpu_map__put(struct perf_cpu_map *map); 37 * perf_cpu_map__empty - is map either empty or the "any CPU"/dummy value. 39 LIBPERF_API bool perf_cpu_map__empty(const struct perf_cpu_map *map); 40 LIBPERF_API struct perf_cpu perf_cpu_map__max(const struct perf_cpu_map *map); 41 LIBPERF_API bool perf_cpu_map__has(const struct perf_cpu_map *map, struct perf_cpu cpu); [all …]
|
/Linux-v6.6/drivers/clocksource/ |
D | ingenic-timer.c | 1 // SPDX-License-Identifier: GPL-2.0 14 #include <linux/mfd/ingenic-tcu.h> 23 #include <dt-bindings/clock/ingenic,tcu.h> 32 unsigned int cpu; member 40 struct regmap *map; member 56 regmap_read(tcu->map, TCU_REG_TCNTc(tcu->cs_channel), &count); in ingenic_tcu_timer_read() 69 return container_of(timer, struct ingenic_tcu, timers[timer->cpu]); in to_ingenic_tcu() 83 regmap_write(tcu->map, TCU_REG_TECR, BIT(timer->channel)); in ingenic_tcu_cevt_set_state_shutdown() 95 return -EINVAL; in ingenic_tcu_cevt_set_next() 97 regmap_write(tcu->map, TCU_REG_TDFRc(timer->channel), next); in ingenic_tcu_cevt_set_next() [all …]
|
/Linux-v6.6/lib/ |
D | cpu_rmap.c | 1 // SPDX-License-Identifier: GPL-2.0-only 3 * cpu_rmap.c: CPU affinity reverse-map support 13 * objects with CPU affinities. This can be seen as a reverse-map of 14 * CPU affinity. However, we do not assume that the object affinities 17 * CPU topology. 21 * alloc_cpu_rmap - allocate CPU affinity reverse-map 28 unsigned int cpu; in alloc_cpu_rmap() local 39 rmap = kzalloc(obj_offset + size * sizeof(rmap->obj[0]), flags); in alloc_cpu_rmap() 43 kref_init(&rmap->refcount); in alloc_cpu_rmap() 44 rmap->obj = (void **)((char *)rmap + obj_offset); in alloc_cpu_rmap() [all …]
|
/Linux-v6.6/samples/bpf/ |
D | map_perf_test_user.c | 1 // SPDX-License-Identifier: GPL-2.0-only 82 static void test_hash_prealloc(int cpu) in test_hash_prealloc() argument 90 printf("%d:hash_map_perf pre-alloc %lld events per sec\n", in test_hash_prealloc() 91 cpu, max_cnt * 1000000000ll / (time_get_ns() - start_time)); in test_hash_prealloc() 106 * It is fine that the user requests for a map with in pre_test_lru_hash_lookup() 108 * may return not found. For LRU map, we are not interested in pre_test_lru_hash_lookup() 109 * in such small map performance. in pre_test_lru_hash_lookup() 120 static void do_test_lru(enum test_type test, int cpu) in do_test_lru() argument 129 if (test == INNER_LRU_HASH_PREALLOC && cpu) { in do_test_lru() 130 /* If CPU is not 0, create inner_lru hash map and insert the fd in do_test_lru() [all …]
|
/Linux-v6.6/Documentation/bpf/ |
D | map_hash.rst | 1 .. SPDX-License-Identifier: GPL-2.0-only 3 .. Copyright (C) 2022-2023 Isovalent, Inc. 10 - ``BPF_MAP_TYPE_HASH`` was introduced in kernel version 3.19 11 - ``BPF_MAP_TYPE_PERCPU_HASH`` was introduced in version 4.6 12 - Both ``BPF_MAP_TYPE_LRU_HASH`` and ``BPF_MAP_TYPE_LRU_PERCPU_HASH`` 16 purpose hash map storage. Both the key and the value can be structs, 20 to the max_entries limit that you specify. Hash maps use pre-allocation 22 used to disable pre-allocation when it is too memory expensive. 25 CPU. The per-cpu values are stored internally in an array. 32 shared across CPUs but it is possible to request a per CPU LRU list with [all …]
|
D | map_cpumap.rst | 1 .. SPDX-License-Identifier: GPL-2.0-only 9 - ``BPF_MAP_TYPE_CPUMAP`` was introduced in kernel version 4.15 11 .. kernel-doc:: kernel/bpf/cpumap.c 12 :doc: cpu map 14 An example use-case for this map type is software based Receive Side Scaling (RSS). 16 The CPUMAP represents the CPUs in the system indexed as the map-key, and the 17 map-value is the config setting (per CPUMAP entry). Each CPUMAP entry has a dedicated 18 kernel thread bound to the given CPU to represent the remote CPU execution unit. 21 on the remote CPU. This allows an XDP program to split its processing across 22 multiple CPUs. For example, a scenario where the initial CPU (that sees/receives [all …]
|
D | map_cgroup_storage.rst | 1 .. SPDX-License-Identifier: GPL-2.0-only 8 The ``BPF_MAP_TYPE_CGROUP_STORAGE`` map type represents a local fix-sized 13 The map provide a local storage at the cgroup that the BPF program is attached 19 ``BPF_MAP_TYPE_CGROUP_STORAGE`` map type. Some of its behaviors was changed in 25 The map uses key of type of either ``__u64 cgroup_inode_id`` or 38 map will share the same storage. Otherwise, if the type is 44 void *bpf_get_local_storage(void *map, u64 flags) 75 Userspace accessing map declared above:: 80 __u32 map_lookup(struct bpf_map *map, __u64 cgrp, enum bpf_attach_type type) 87 bpf_map_lookup_elem(bpf_map__fd(map), &key, &value); [all …]
|
/Linux-v6.6/drivers/gpu/drm/i915/gem/selftests/ |
D | i915_gem_coherency.c | 2 * SPDX-License-Identifier: MIT 27 void *map; in cpu_set() local 28 u32 *cpu; in cpu_set() local 31 i915_gem_object_lock(ctx->obj, NULL); in cpu_set() 32 err = i915_gem_object_prepare_write(ctx->obj, &needs_clflush); in cpu_set() 36 page = i915_gem_object_get_page(ctx->obj, offset >> PAGE_SHIFT); in cpu_set() 37 map = kmap_atomic(page); in cpu_set() 38 cpu = map + offset_in_page(offset); in cpu_set() 41 drm_clflush_virt_range(cpu, sizeof(*cpu)); in cpu_set() 43 *cpu = v; in cpu_set() [all …]
|
/Linux-v6.6/arch/sh/kernel/cpu/sh4/ |
D | sq.c | 1 // SPDX-License-Identifier: GPL-2.0 3 * arch/sh/kernel/cpu/sh4/sq.c 5 * General management API for SH-4 integrated Store Queues 7 * Copyright (C) 2001 - 2006 Paul Mundt 11 #include <linux/cpu.h> 23 #include <cpu/sq.h> 50 * sq_flush_range - Flush (prefetch) a specific SQ range 62 for (len >>= 5; len--; sq += 8) in sq_flush_range() 70 static inline void sq_mapping_list_add(struct sq_mapping *map) in sq_mapping_list_add() argument 78 p = &tmp->next; in sq_mapping_list_add() [all …]
|
/Linux-v6.6/arch/arm/kernel/ |
D | devtree.c | 1 // SPDX-License-Identifier: GPL-2.0-only 25 #include <asm/mach-types.h> 40 if (of_property_read_string(node, "enable-method", &method)) in set_smp_ops_by_method() 43 for (; m->method; m++) in set_smp_ops_by_method() 44 if (!strcmp(m->method, method)) { in set_smp_ops_by_method() 45 smp_set_ops(m->ops); in set_smp_ops_by_method() 60 * arm_dt_init_cpu_maps - Function retrieves cpu nodes from the device tree 61 * and builds the cpu logical map array containing MPIDR values related to 64 * Updates the cpu possible mask with the number of parsed cpu nodes 69 * Temp logical map is initialized with UINT_MAX values that are in arm_dt_init_cpu_maps() [all …]
|
/Linux-v6.6/Documentation/ABI/stable/ |
D | sysfs-devices-system-cpu | 1 What: /sys/devices/system/cpu/dscr_default 2 Date: 13-May-2014 6 /sys/devices/system/cpu/cpuN/dscr on all CPUs. 9 all per-CPU defaults at the same time. 12 What: /sys/devices/system/cpu/cpu[0-9]+/dscr 13 Date: 13-May-2014 17 a CPU. 22 on any CPU where it executes (overriding the value described 27 What: /sys/devices/system/cpu/cpuX/topology/physical_package_id 33 What: /sys/devices/system/cpu/cpuX/topology/die_id [all …]
|
/Linux-v6.6/tools/lib/perf/include/internal/ |
D | evsel.h | 1 /* SPDX-License-Identifier: GPL-2.0 */ 15 * Per fd, to map back from PERF_SAMPLE_ID to evsel, only used when there are 25 * queues for each CPU (per-cpu tracing) or task (per-thread tracing). 30 struct perf_cpu cpu; member 33 /* Guest machine pid and VCPU, valid only if machine_pid is non-zero */ 44 /** The commonly used cpu map of CPUs the event should be opened upon, etc. */ 47 * The cpu map read from the PMU. For core PMUs this is the list of all 49 * cpu map for opening the event on, for example, the first CPU on a 64 * system_wide is for events that need to be on every CPU, irrespective 66 * dummy event. Map propagation will set cpus for this event to all CPUs [all …]
|
/Linux-v6.6/arch/arc/plat-axs10x/ |
D | axs10x.c | 1 // SPDX-License-Identifier: GPL-2.0-only 5 * Copyright (C) 2013-15 Synopsys, Inc. (www.synopsys.com) 11 #include <asm/asm-offsets.h> 30 * Peripherals on CPU Card and Mother Board are wired to cpu intc via in axs10x_enable_gpio_intc_wire() 33 * --------------------- in axs10x_enable_gpio_intc_wire() 34 * | snps,arc700-intc | in axs10x_enable_gpio_intc_wire() 35 * --------------------- in axs10x_enable_gpio_intc_wire() 37 * ------------------- ------------------- in axs10x_enable_gpio_intc_wire() 38 * | snps,dw-apb-gpio | | snps,dw-apb-gpio | in axs10x_enable_gpio_intc_wire() 39 * ------------------- ------------------- in axs10x_enable_gpio_intc_wire() [all …]
|